Unnamed: 0
int64 0
7.24k
| id
int64 1
7.28k
| raw_text
stringlengths 9
124k
| vw_text
stringlengths 12
15k
|
---|---|---|---|
4,800 | 5,346 | Sequence to Sequence Learning
with Neural Networks
Ilya Sutskever
Google
[email protected]
Oriol Vinyals
Google
[email protected]
Quoc V. Le
Google
[email protected]
Abstract
Deep Neural Networks (DNNs) are powerful models that have achieved excellent performance on difficult learning tasks. Although DNNs work well whenever
large labeled training sets are available, they cannot be used to map sequences to
sequences. In this paper, we present a general end-to-end approach to sequence
learning that makes minimal assumptions on the sequence structure. Our method
uses a multilayered Long Short-Term Memory (LSTM) to map the input sequence
to a vector of a fixed dimensionality, and then another deep LSTM to decode the
target sequence from the vector. Our main result is that on an English to French
translation task from the WMT-14 dataset, the translations produced by the LSTM
achieve a BLEU score of 34.8 on the entire test set, where the LSTM?s BLEU
score was penalized on out-of-vocabulary words. Additionally, the LSTM did not
have difficulty on long sentences. For comparison, a phrase-based SMT system
achieves a BLEU score of 33.3 on the same dataset. When we used the LSTM
to rerank the 1000 hypotheses produced by the aforementioned SMT system, its
BLEU score increases to 36.5, which is close to the previous state of the art. The
LSTM also learned sensible phrase and sentence representations that are sensitive
to word order and are relatively invariant to the active and the passive voice. Finally, we found that reversing the order of the words in all source sentences (but
not target sentences) improved the LSTM?s performance markedly, because doing
so introduced many short term dependencies between the source and the target
sentence which made the optimization problem easier.
1
Introduction
Deep Neural Networks (DNNs) are extremely powerful machine learning models that achieve excellent performance on difficult problems such as speech recognition [13, 7] and visual object recognition [19, 6, 21, 20]. DNNs are powerful because they can perform arbitrary parallel computation
for a modest number of steps. A surprising example of the power of DNNs is their ability to sort
N N -bit numbers using only 2 hidden layers of quadratic size [27]. So, while neural networks are
related to conventional statistical models, they learn an intricate computation. Furthermore, large
DNNs can be trained with supervised backpropagation whenever the labeled training set has enough
information to specify the network?s parameters. Thus, if there exists a parameter setting of a large
DNN that achieves good results (for example, because humans can solve the task very rapidly),
supervised backpropagation will find these parameters and solve the problem.
Despite their flexibility and power, DNNs can only be applied to problems whose inputs and targets
can be sensibly encoded with vectors of fixed dimensionality. It is a significant limitation, since
many important problems are best expressed with sequences whose lengths are not known a-priori.
For example, speech recognition and machine translation are sequential problems. Likewise, question answering can also be seen as mapping a sequence of words representing the question to a
1
sequence of words representing the answer. It is therefore clear that a domain-independent method
that learns to map sequences to sequences would be useful.
Sequences pose a challenge for DNNs because they require that the dimensionality of the inputs and
outputs is known and fixed. In this paper, we show that a straightforward application of the Long
Short-Term Memory (LSTM) architecture [16] can solve general sequence to sequence problems.
The idea is to use one LSTM to read the input sequence, one timestep at a time, to obtain large fixeddimensional vector representation, and then to use another LSTM to extract the output sequence
from that vector (fig. 1). The second LSTM is essentially a recurrent neural network language model
[28, 23, 30] except that it is conditioned on the input sequence. The LSTM?s ability to successfully
learn on data with long range temporal dependencies makes it a natural choice for this application
due to the considerable time lag between the inputs and their corresponding outputs (fig. 1).
There have been a number of related attempts to address the general sequence to sequence learning
problem with neural networks. Our approach is closely related to Kalchbrenner and Blunsom [18]
who were the first to map the entire input sentence to vector, and is very similar to Cho et al. [5].
Graves [10] introduced a novel differentiable attention mechanism that allows neural networks to
focus on different parts of their input, and an elegant variant of this idea was successfully applied
to machine translation by Bahdanau et al. [2]. The Connectionist Sequence Classification is another
popular technique for mapping sequences to sequences with neural networks, although it assumes a
monotonic alignment between the inputs and the outputs [11].
Figure 1: Our model reads an input sentence ?ABC? and produces ?WXYZ? as the output sentence. The
model stops making predictions after outputting the end-of-sentence token. Note that the LSTM reads the
input sentence in reverse, because doing so introduces many short term dependencies in the data that make the
optimization problem much easier.
The main result of this work is the following. On the WMT?14 English to French translation task,
we obtained a BLEU score of 34.81 by directly extracting translations from an ensemble of 5 deep
LSTMs (with 380M parameters each) using a simple left-to-right beam-search decoder. This is
by far the best result achieved by direct translation with large neural networks. For comparison,
the BLEU score of a SMT baseline on this dataset is 33.30 [29]. The 34.81 BLEU score was
achieved by an LSTM with a vocabulary of 80k words, so the score was penalized whenever the
reference translation contained a word not covered by these 80k. This result shows that a relatively
unoptimized neural network architecture which has much room for improvement outperforms a
mature phrase-based SMT system.
Finally, we used the LSTM to rescore the publicly available 1000-best lists of the SMT baseline on
the same task [29]. By doing so, we obtained a BLEU score of 36.5, which improves the baseline
by 3.2 BLEU points and is close to the previous state-of-the-art (which is 37.0 [9]).
Surprisingly, the LSTM did not suffer on very long sentences, despite the recent experience of other
researchers with related architectures [26]. We were able to do well on long sentences because we
reversed the order of words in the source sentence but not the target sentences in the training and test
set. By doing so, we introduced many short term dependencies that made the optimization problem
much simpler (see sec. 2 and 3.3). As a result, SGD could learn LSTMs that had no trouble with
long sentences. The simple trick of reversing the words in the source sentence is one of the key
technical contributions of this work.
A useful property of the LSTM is that it learns to map an input sentence of variable length into
a fixed-dimensional vector representation. Given that translations tend to be paraphrases of the
source sentences, the translation objective encourages the LSTM to find sentence representations
that capture their meaning, as sentences with similar meanings are close to each other while different
2
sentences meanings will be far. A qualitative evaluation supports this claim, showing that our model
is aware of word order and is fairly invariant to the active and passive voice.
2
The model
The Recurrent Neural Network (RNN) [31, 28] is a natural generalization of feedforward neural
networks to sequences. Given a sequence of inputs (x1 , . . . , xT ), a standard RNN computes a
sequence of outputs (y1 , . . . , yT ) by iterating the following equation:
ht = sigm W hx xt + W hh ht?1
yt
=
W yh ht
The RNN can easily map sequences to sequences whenever the alignment between the inputs the
outputs is known ahead of time. However, it is not clear how to apply an RNN to problems whose
input and the output sequences have different lengths with complicated and non-monotonic relationships.
A simple strategy for general sequence learning is to map the input sequence to a fixed-sized vector
using one RNN, and then to map the vector to the target sequence with another RNN (this approach
has also been taken by Cho et al. [5]). While it could work in principle since the RNN is provided
with all the relevant information, it would be difficult to train the RNNs due to the resulting long
term dependencies [14, 4] (figure 1) [16, 15]. However, the Long Short-Term Memory (LSTM) [16]
is known to learn problems with long range temporal dependencies, so an LSTM may succeed in
this setting.
The goal of the LSTM is to estimate the conditional probability p(y1 , . . . , yT ? |x1 , . . . , xT ) where
(x1 , . . . , xT ) is an input sequence and y1 , . . . , yT ? is its corresponding output sequence whose length
T ? may differ from T . The LSTM computes this conditional probability by first obtaining the fixeddimensional representation v of the input sequence (x1 , . . . , xT ) given by the last hidden state of the
LSTM, and then computing the probability of y1 , . . . , yT ? with a standard LSTM-LM formulation
whose initial hidden state is set to the representation v of x1 , . . . , xT :
?
p(y1 , . . . , yT ? |x1 , . . . , xT ) =
T
Y
p(yt |v, y1 , . . . , yt?1 )
(1)
t=1
In this equation, each p(yt |v, y1 , . . . , yt?1 ) distribution is represented with a softmax over all the
words in the vocabulary. We use the LSTM formulation from Graves [10]. Note that we require that
each sentence ends with a special end-of-sentence symbol ?<EOS>?, which enables the model to
define a distribution over sequences of all possible lengths. The overall scheme is outlined in figure
1, where the shown LSTM computes the representation of ?A?, ?B?, ?C?, ?<EOS>? and then uses
this representation to compute the probability of ?W?, ?X?, ?Y?, ?Z?, ?<EOS>?.
Our actual models differ from the above description in three important ways. First, we used two
different LSTMs: one for the input sequence and another for the output sequence, because doing
so increases the number model parameters at negligible computational cost and makes it natural to
train the LSTM on multiple language pairs simultaneously [18]. Second, we found that deep LSTMs
significantly outperformed shallow LSTMs, so we chose an LSTM with four layers. Third, we found
it extremely valuable to reverse the order of the words of the input sentence. So for example, instead
of mapping the sentence a, b, c to the sentence ?, ?, ?, the LSTM is asked to map c, b, a to ?, ?, ?,
where ?, ?, ? is the translation of a, b, c. This way, a is in close proximity to ?, b is fairly close to
?, and so on, a fact that makes it easy for SGD to ?establish communication? between the input and
the output. We found this simple data transformation to greatly boost the performance of the LSTM.
3
Experiments
We applied our method to the WMT?14 English to French MT task in two ways. We used it to
directly translate the input sentence without using a reference SMT system and we it to rescore the
n-best lists of an SMT baseline. We report the accuracy of these translation methods, present sample
translations, and visualize the resulting sentence representation.
3
3.1
Dataset details
We used the WMT?14 English to French dataset. We trained our models on a subset of 12M sentences consisting of 348M French words and 304M English words, which is a clean ?selected?
subset from [29]. We chose this translation task and this specific training set subset because of the
public availability of a tokenized training and test set together with 1000-best lists from the baseline
SMT [29].
As typical neural language models rely on a vector representation for each word, we used a fixed
vocabulary for both languages. We used 160,000 of the most frequent words for the source language
and 80,000 of the most frequent words for the target language. Every out-of-vocabulary word was
replaced with a special ?UNK? token.
3.2
Decoding and Rescoring
The core of our experiments involved training a large deep LSTM on many sentence pairs. We
trained it by maximizing the log probability of a correct translation T given the source sentence S,
so the training objective is
X
1/|S|
log p(T |S)
(T,S)?S
where S is the training set. Once training is complete, we produce translations by finding the most
likely translation according to the LSTM:
T? = arg max p(T |S)
T
(2)
We search for the most likely translation using a simple left-to-right beam search decoder which
maintains a small number B of partial hypotheses, where a partial hypothesis is a prefix of some
translation. At each timestep we extend each partial hypothesis in the beam with every possible
word in the vocabulary. This greatly increases the number of the hypotheses so we discard all but
the B most likely hypotheses according to the model?s log probability. As soon as the ?<EOS>?
symbol is appended to a hypothesis, it is removed from the beam and is added to the set of complete
hypotheses. While this decoder is approximate, it is simple to implement. Interestingly, our system
performs well even with a beam size of 1, and a beam of size 2 provides most of the benefits of beam
search (Table 1).
We also used the LSTM to rescore the 1000-best lists produced by the baseline system [29]. To
rescore an n-best list, we computed the log probability of every hypothesis with our LSTM and took
an even average with their score and the LSTM?s score.
3.3
Reversing the Source Sentences
While the LSTM is capable of solving problems with long term dependencies, we discovered that
the LSTM learns much better when the source sentences are reversed (the target sentences are not
reversed). By doing so, the LSTM?s test perplexity dropped from 5.8 to 4.7, and the test BLEU
scores of its decoded translations increased from 25.9 to 30.6.
While we do not have a complete explanation to this phenomenon, we believe that it is caused by
the introduction of many short term dependencies to the dataset. Normally, when we concatenate a
source sentence with a target sentence, each word in the source sentence is far from its corresponding
word in the target sentence. As a result, the problem has a large ?minimal time lag? [17]. By
reversing the words in the source sentence, the average distance between corresponding words in
the source and target language is unchanged. However, the first few words in the source language
are now very close to the first few words in the target language, so the problem?s minimal time lag is
greatly reduced. Thus, backpropagation has an easier time ?establishing communication? between
the source sentence and the target sentence, which in turn results in substantially improved overall
performance.
Initially, we believed that reversing the input sentences would only lead to more confident predictions in the early parts of the target sentence and to less confident predictions in the later parts. However, LSTMs trained on reversed source sentences did much better on long sentences than LSTMs
4
trained on the raw source sentences (see sec. 3.7), which suggests that reversing the input sentences
results in LSTMs with better memory utilization.
3.4
Training details
We found that the LSTM models are fairly easy to train. We used deep LSTMs with 4 layers,
with 1000 cells at each layer and 1000 dimensional word embeddings, with an input vocabulary
of 160,000 and an output vocabulary of 80,000. We found deep LSTMs to significantly outperform
shallow LSTMs, where each additional layer reduced perplexity by nearly 10%, possibly due to their
much larger hidden state. We used a naive softmax over 80,000 words at each output. The resulting
LSTM has 380M parameters of which 64M are pure recurrent connections (32M for the ?encoder?
LSTM and 32M for the ?decoder? LSTM). The complete training details are given below:
? We initialized all of the LSTM?s parameters with the uniform distribution between -0.08
and 0.08
? We used stochastic gradient descent without momentum, with a fixed learning rate of 0.7.
After 5 epochs, we begun halving the learning rate every half epoch. We trained our models
for a total of 7.5 epochs.
? We used batches of 128 sequences for the gradient and divided it the size of the batch
(namely, 128).
? Although LSTMs tend to not suffer from the vanishing gradient problem, they can have
exploding gradients. Thus we enforced a hard constraint on the norm of the gradient [10,
25] by scaling it when its norm exceeded a threshold. For each training batch, we compute
s = kgk2 , where g is the gradient divided by 128. If s > 5, we set g = 5g
s .
? Different sentences have different lengths. Most sentences are short (e.g., length 20-30)
but some sentences are long (e.g., length > 100), so a minibatch of 128 randomly chosen
training sentences will have many short sentences and few long sentences, and as a result,
much of the computation in the minibatch is wasted. To address this problem, we made
sure that all sentences within a minibatch were roughly of the same length, which a 2x
speedup.
3.5
Parallelization
A C++ implementation of deep LSTM with the configuration from the previous section on a single GPU processes a speed of approximately 1,700 words per second. This was too slow for our
purposes, so we parallelized our model using an 8-GPU machine. Each layer of the LSTM was
executed on a different GPU and communicated its activations to the next GPU (or layer) as soon
as they were computed. Our models have 4 layers of LSTMs, each of which resides on a separate
GPU. The remaining 4 GPUs were used to parallelize the softmax, so each GPU was responsible
for multiplying by a 1000 ? 20000 matrix. The resulting implementation achieved a speed of 6,300
(both English and French) words per second with a minibatch size of 128. Training took about a ten
days with this implementation.
3.6
Experimental Results
We used the cased BLEU score [24] to evaluate the quality of our translations. We computed our
BLEU scores using multi-bleu.pl1 on the tokenized predictions and ground truth. This way
of evaluating the BELU score is consistent with [5] and [2], and reproduces the 33.3 score of [29].
However, if we evaluate the state of the art system of [9] (whose predictions can be downloaded
from statmt.org\matrix) in this manner, we get 37.0, which is greater than the 35.8 reported
by statmt.org\matrix.
The results are presented in tables 1 and 2. Our best results are obtained with an ensemble of
LSTMs that differ in their random initializations and in the random order of minibatches. While the
decoded translations of the LSTM ensemble do not beat the state of the art, it is the first time that
a pure neural translation system outperforms a phrase-based SMT baseline on a large MT task by
1
There several variants of the BLEU score, and each variant is defined with a perl script.
5
Method
Bahdanau et al. [2]
Baseline System [29]
Single forward LSTM, beam size 12
Single reversed LSTM, beam size 12
Ensemble of 5 reversed LSTMs, beam size 1
Ensemble of 2 reversed LSTMs, beam size 12
Ensemble of 5 reversed LSTMs, beam size 2
Ensemble of 5 reversed LSTMs, beam size 12
test BLEU score (ntst14)
28.45
33.30
26.17
30.59
33.00
33.27
34.50
34.81
Table 1: The performance of the LSTM on WMT?14 English to French test set (ntst14). Note that
an ensemble of 5 LSTMs with a beam of size 2 is cheaper than of a single LSTM with a beam of
size 12.
Method
Baseline System [29]
Cho et al. [5]
State of the art [9]
Rescoring the baseline 1000-best with a single forward LSTM
Rescoring the baseline 1000-best with a single reversed LSTM
Rescoring the baseline 1000-best with an ensemble of 5 reversed LSTMs
Oracle Rescoring of the Baseline 1000-best lists
test BLEU score (ntst14)
33.30
34.54
37.0
35.61
35.85
36.5
?45
Table 2: Methods that use neural networks together with an SMT system on the WMT?14 English
to French test set (ntst14).
a sizeable margin, despite its inability to handle out-of-vocabulary words. The LSTM is within 0.5
BLEU points of the previous state of the art by rescoring the 1000-best list of the baseline system.
3.7
Performance on long sentences
We were surprised to discover that the LSTM did well on long sentences, which is shown quantitatively in figure 3. Table 3 presents several examples of long sentences and their translations.
3.8
Model Analysis
15
I was given a card by her in the garden
4
Mary admires John
3
2
In the garden , she gave me a card
She gave me a card in the garden
10
Mary is in love with John
5
1
0
?1
0
Mary respects John
John admires Mary
?5
John is in love with Mary
?2
She was given a card by me in the garden
In the garden , I gave her a card
?3
?10
?4
?5
?6
?8
?15
John respects Mary
?6
?4
?2
0
2
4
6
8
?20
?15
10
I gave her a card in the garden
?10
?5
0
5
10
15
20
Figure 2: The figure shows a 2-dimensional PCA projection of the LSTM hidden states that are obtained
after processing the phrases in the figures. The phrases are clustered by meaning, which in these examples is
primarily a function of word order, which would be difficult to capture with a bag-of-words model. Notice that
both clusters have similar internal structure.
One of the attractive features of our model is its ability to turn a sequence of words into a vector
of fixed dimensionality. Figure 2 visualizes some of the learned representations. The figure clearly
shows that the representations are sensitive to the order of words, while being fairly insensitive to the
6
Type
Our model
Sentence
Ulrich UNK , membre du conseil d? administration du constructeur automobile Audi ,
affirme qu? il s? agit d? une pratique courante depuis des ann?ees pour que les t?el?ephones
portables puissent e? tre collect?es avant les r?eunions du conseil d? administration afin qu? ils
ne soient pas utilis?es comme appareils d? e? coute a` distance .
Ulrich Hackenberg , membre du conseil d? administration du constructeur automobile Audi ,
d?eclare que la collecte des t?el?ephones portables avant les r?eunions du conseil , afin qu? ils
ne puissent pas e? tre utilis?es comme appareils d? e? coute a` distance , est une pratique courante
depuis des ann?ees .
? Les t?el?ephones cellulaires , qui sont vraiment une question , non seulement parce qu? ils
pourraient potentiellement causer des interf?erences avec les appareils de navigation , mais
nous savons , selon la FCC , qu? ils pourraient interf?erer avec les tours de t?el?ephone cellulaire
lorsqu? ils sont dans l? air ? , dit UNK .
? Les t?el?ephones portables sont v?eritablement un probl`eme , non seulement parce qu? ils
pourraient e? ventuellement cr?eer des interf?erences avec les instruments de navigation , mais
parce que nous savons , d? apr`es la FCC , qu? ils pourraient perturber les antennes-relais de
t?el?ephonie mobile s? ils sont utilis?es a` bord ? , a d?eclar?e Rosenker .
Avec la cr?emation , il y a un ? sentiment de violence contre le corps d? un e? tre cher ? ,
qui sera ? r?eduit a` une pile de cendres ? en tr`es peu de temps au lieu d? un processus de
d?ecomposition ? qui accompagnera les e? tapes du deuil ? .
Il y a , avec la cr?emation , ? une violence faite au corps aim?e ? ,
qui va e? tre ? r?eduit a` un tas de cendres ? en tr`es peu de temps , et non apr`es un processus de
d?ecomposition , qui ? accompagnerait les phases du deuil ? .
Truth
Our model
Truth
Our model
Truth
Table 3: A few examples of long translations produced by the LSTM alongside the ground truth
translations. The reader can verify that the translations are sensible using Google translate.
LSTM (34.8)
baseline (33.3)
40
40
35
BLEU score
BLEU score
35
30
25
20
LSTM (34.8)
baseline (33.3)
30
25
478
12
17
22
28
35
test sentences sorted by their length
20
0
79
500
1000
1500
2000
2500
3000
test sentences sorted by average word frequency rank
3500
Figure 3: The left plot shows the performance of our system as a function of sentence length, where the
x-axis corresponds to the test sentences sorted by their length and is marked by the actual sequence lengths.
There is no degradation on sentences with less than 35 words, there is only a minor degradation on the longest
sentences. The right plot shows the LSTM?s performance on sentences with progressively more rare words,
where the x-axis corresponds to the test sentences sorted by their ?average word frequency rank?.
replacement of an active voice with a passive voice. The two-dimensional projections are obtained
using PCA.
4
Related work
There is a large body of work on applications of neural networks to machine translation. So far,
the simplest and most effective way of applying an RNN-Language Model (RNNLM) [23] or a
7
Feedforward Neural Network Language Model (NNLM) [3] to an MT task is by rescoring the nbest lists of a strong MT baseline [22], which reliably improves translation quality.
More recently, researchers have begun to look into ways of including information about the source
language into the NNLM. Examples of this work include Auli et al. [1], who combine an NNLM
with a topic model of the input sentence, which improves rescoring performance. Devlin et al. [8]
followed a similar approach, but they incorporated their NNLM into the decoder of an MT system
and used the decoder?s alignment information to provide the NNLM with the most useful words in
the input sentence. Their approach was highly successful and it achieved large improvements over
their baseline.
Our work is closely related to Kalchbrenner and Blunsom [18], who were the first to map the input
sentence into a vector and then back to a sentence, although they map sentences to vectors using
convolutional neural networks, which lose the ordering of the words. Similarly to this work, Cho et
al. [5] used an LSTM-like RNN architecture to map sentences into vectors and back, although their
primary focus was on integrating their neural network into an SMT system. Bahdanau et al. [2] also
attempted direct translations with a neural network that used an attention mechanism to overcome
the poor performance on long sentences experienced by Cho et al. [5] and achieved encouraging
results. Likewise, Pouget-Abadie et al. [26] attempted to address the memory problem of Cho et
al. [5] by translating pieces of the source sentence in way that produces smooth translations, which
is similar to a phrase-based approach. We suspect that they could achieve similar improvements by
simply training their networks on reversed source sentences.
End-to-end training is also the focus of Hermann et al. [12], whose model represents the inputs and
outputs by feedforward networks, and map them to similar points in space. However, their approach
cannot generate translations directly: to get a translation, they need to do a look up for closest vector
in the pre-computed database of sentences, or to rescore a sentence.
5
Conclusion
In this work, we showed that a large deep LSTM with a limited vocabulary can outperform a standard SMT-based system whose vocabulary is unlimited on a large-scale MT task. The success of
our simple LSTM-based approach on MT suggests that it should do well on many other sequence
learning problems, provided they have enough training data.
We were surprised by the extent of the improvement obtained by reversing the words in the source
sentences. We conclude that it is important to find a problem encoding that has the greatest number
of short term dependencies, as they make the learning problem much simpler. In particular, while
we were unable to train a standard RNN on the non-reversed translation problem (shown in fig. 1),
we believe that a standard RNN should be easily trainable when the source sentences are reversed
(although we did not verify it experimentally).
We were also surprised by the ability of the LSTM to correctly translate very long sentences. We
were initially convinced that the LSTM would fail on long sentences due to its limited memory,
and other researchers reported poor performance on long sentences with a model similar to ours
[5, 2, 26]. And yet, LSTMs trained on the reversed dataset had little difficulty translating long
sentences.
Most importantly, we demonstrated that a simple, straightforward and a relatively unoptimized approach can outperform a mature SMT system, so further work will likely lead to even greater translation accuracies. These results suggest that our approach will likely do well on other challenging
sequence to sequence problems.
6
Acknowledgments
We thank Samy Bengio, Jeff Dean, Matthieu Devin, Geoffrey Hinton, Nal Kalchbrenner, Thang Luong, Wolfgang Macherey, Rajat Monga, Vincent Vanhoucke, Peng Xu, Wojciech Zaremba, and the Google Brain team
for useful comments and discussions.
8
References
[1] M. Auli, M. Galley, C. Quirk, and G. Zweig. Joint language and translation modeling with recurrent
neural networks. In EMNLP, 2013.
[2] D. Bahdanau, K. Cho, and Y. Bengio. Neural machine translation by jointly learning to align and translate.
arXiv preprint arXiv:1409.0473, 2014.
[3] Y. Bengio, R. Ducharme, P. Vincent, and C. Jauvin. A neural probabilistic language model. In Journal of
Machine Learning Research, pages 1137?1155, 2003.
[4] Y. Bengio, P. Simard, and P. Frasconi. Learning long-term dependencies with gradient descent is difficult.
IEEE Transactions on Neural Networks, 5(2):157?166, 1994.
[5] K. Cho, B. Merrienboer, C. Gulcehre, F. Bougares, H. Schwenk, and Y. Bengio. Learning phrase representations using RNN encoder-decoder for statistical machine translation. In Arxiv preprint arXiv:1406.1078,
2014.
[6] D. Ciresan, U. Meier, and J. Schmidhuber. Multi-column deep neural networks for image classification.
In CVPR, 2012.
[7] G. E. Dahl, D. Yu, L. Deng, and A. Acero. Context-dependent pre-trained deep neural networks for large
vocabulary speech recognition. IEEE Transactions on Audio, Speech, and Language Processing - Special
Issue on Deep Learning for Speech and Language Processing, 2012.
[8] J. Devlin, R. Zbib, Z. Huang, T. Lamar, R. Schwartz, and J. Makhoul. Fast and robust neural network
joint models for statistical machine translation. In ACL, 2014.
[9] Nadir Durrani, Barry Haddow, Philipp Koehn, and Kenneth Heafield. Edinburgh?s phrase-based machine
translation systems for wmt-14. In WMT, 2014.
[10] A. Graves. Generating sequences with recurrent neural networks. In Arxiv preprint arXiv:1308.0850,
2013.
[11] A. Graves, S. Fern?andez, F. Gomez, and J. Schmidhuber. Connectionist temporal classification: labelling
unsegmented sequence data with recurrent neural networks. In ICML, 2006.
[12] K. M. Hermann and P. Blunsom. Multilingual distributed representations without word alignment. In
ICLR, 2014.
[13] G. Hinton, L. Deng, D. Yu, G. Dahl, A. Mohamed, N. Jaitly, A. Senior, V. Vanhoucke, P. Nguyen,
T. Sainath, and B. Kingsbury. Deep neural networks for acoustic modeling in speech recognition. IEEE
Signal Processing Magazine, 2012.
[14] S. Hochreiter. Untersuchungen zu dynamischen neuronalen netzen. Master?s thesis, Institut fur Informatik, Technische Universitat, Munchen, 1991.
[15] S. Hochreiter, Y. Bengio, P. Frasconi, and J. Schmidhuber. Gradient flow in recurrent nets: the difficulty
of learning long-term dependencies, 2001.
[16] S. Hochreiter and J. Schmidhuber. Long short-term memory. Neural Computation, 1997.
[17] S. Hochreiter and J. Schmidhuber. LSTM can solve hard long time lag problems. 1997.
[18] N. Kalchbrenner and P. Blunsom. Recurrent continuous translation models. In EMNLP, 2013.
[19] A. Krizhevsky, I. Sutskever, and G. E. Hinton. ImageNet classification with deep convolutional neural
networks. In NIPS, 2012.
[20] Q.V. Le, M.A. Ranzato, R. Monga, M. Devin, K. Chen, G.S. Corrado, J. Dean, and A.Y. Ng. Building
high-level features using large scale unsupervised learning. In ICML, 2012.
[21] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition.
Proceedings of the IEEE, 1998.
[22] T. Mikolov. Statistical Language Models based on Neural Networks. PhD thesis, Brno University of
Technology, 2012.
[23] T. Mikolov, M. Karafi?at, L. Burget, J. Cernock`y, and S. Khudanpur. Recurrent neural network based
language model. In INTERSPEECH, pages 1045?1048, 2010.
[24] K. Papineni, S. Roukos, T. Ward, and W. J. Zhu. BLEU: a method for automatic evaluation of machine
translation. In ACL, 2002.
[25] R. Pascanu, T. Mikolov, and Y. Bengio. On the difficulty of training recurrent neural networks. arXiv
preprint arXiv:1211.5063, 2012.
[26] J. Pouget-Abadie, D. Bahdanau, B. van Merrienboer, K. Cho, and Y. Bengio. Overcoming the
curse of sentence length for neural machine translation using automatic segmentation. arXiv preprint
arXiv:1409.1257, 2014.
[27] A. Razborov. On small depth threshold circuits. In Proc. 3rd Scandinavian Workshop on Algorithm
Theory, 1992.
[28] D. Rumelhart, G. E. Hinton, and R. J. Williams. Learning representations by back-propagating errors.
Nature, 323(6088):533?536, 1986.
[29] H. Schwenk. University le mans. http://www-lium.univ-lemans.fr/?schwenk/cslm_
joint_paper/, 2014. [Online; accessed 03-September-2014].
[30] M. Sundermeyer, R. Schluter, and H. Ney. LSTM neural networks for language modeling. In INTERSPEECH, 2010.
[31] P. Werbos. Backpropagation through time: what it does and how to do it. Proceedings of IEEE, 1990.
9
| 5346 |@word norm:2 sgd:2 tr:2 initial:1 configuration:1 score:21 ours:1 interestingly:1 prefix:1 document:1 outperforms:2 com:3 surprising:1 activation:1 yet:1 gpu:6 john:6 devin:2 concatenate:1 enables:1 plot:2 sont:4 progressively:1 half:1 selected:1 une:5 schluter:1 vanishing:1 short:11 core:1 provides:1 rescoring:8 pascanu:1 philipp:1 org:2 simpler:2 accessed:1 kingsbury:1 direct:2 surprised:3 qualitative:1 combine:1 manner:1 peng:1 intricate:1 roughly:1 pour:1 love:2 multi:2 brain:1 peu:2 actual:2 encouraging:1 little:1 curse:1 provided:2 discover:1 circuit:1 cher:1 what:1 substantially:1 finding:1 transformation:1 temporal:3 every:4 fcc:2 zaremba:1 sensibly:1 schwartz:1 utilization:1 normally:1 negligible:1 dropped:1 despite:3 encoding:1 establishing:1 parallelize:1 approximately:1 blunsom:4 rnns:1 chose:2 initialization:1 au:2 acl:2 suggests:2 collect:1 rescore:5 challenging:1 heafield:1 limited:2 range:2 acknowledgment:1 responsible:1 lecun:1 implement:1 backpropagation:4 communicated:1 rnn:12 nnlm:5 significantly:2 projection:2 burget:1 word:42 eer:1 integrating:1 pre:2 suggest:1 get:2 cannot:2 close:6 acero:1 context:1 applying:1 www:1 conventional:1 map:13 demonstrated:1 yt:10 maximizing:1 dean:2 straightforward:2 attention:2 serum:1 processus:2 sainath:1 williams:1 pure:2 pouget:2 matthieu:1 importantly:1 handle:1 razborov:1 target:14 netzen:1 decode:1 magazine:1 us:2 samy:1 hypothesis:9 jaitly:1 trick:1 pa:2 rumelhart:1 recognition:6 werbos:1 labeled:2 database:1 preprint:5 capture:2 ranzato:1 ordering:1 removed:1 valuable:1 asked:1 trained:8 solving:1 easily:2 joint:2 schwenk:3 represented:1 sigm:1 train:4 univ:1 fast:1 effective:1 que:3 eos:4 kalchbrenner:4 whose:8 encoded:1 lag:4 solve:4 larger:1 ducharme:1 cvpr:1 koehn:1 encoder:2 ability:4 ward:1 jointly:1 online:1 sequence:47 differentiable:1 net:1 took:2 outputting:1 fr:1 frequent:2 dans:1 relevant:1 rapidly:1 translate:4 flexibility:1 achieve:3 papineni:1 description:1 sutskever:2 cluster:1 produce:3 generating:1 object:1 recurrent:10 propagating:1 pose:1 quirk:1 minor:1 strong:1 differ:3 hermann:2 closely:2 correct:1 stochastic:1 human:1 translating:2 public:1 require:2 dnns:8 hx:1 zbib:1 generalization:1 clustered:1 andez:1 avec:5 merrienboer:2 proximity:1 ground:2 mapping:3 visualize:1 claim:1 lm:1 achieves:2 early:1 purpose:1 proc:1 outperformed:1 bag:1 lose:1 sensitive:2 successfully:2 clearly:1 aim:1 cr:3 mobile:1 focus:3 improvement:4 she:3 rank:2 longest:1 fur:1 greatly:3 baseline:18 dependent:1 el:6 jauvin:1 entire:2 initially:2 hidden:5 her:3 dnn:1 unoptimized:2 statmt:2 issue:1 overall:2 aforementioned:1 classification:4 unk:3 arg:1 priori:1 art:6 softmax:3 fairly:4 special:3 aware:1 once:1 frasconi:2 ng:1 thang:1 untersuchungen:1 represents:1 look:2 yu:2 nearly:1 icml:2 unsupervised:1 connectionist:2 report:1 quantitatively:1 few:4 primarily:1 randomly:1 simultaneously:1 cheaper:1 lium:1 replaced:1 phase:1 consisting:1 replacement:1 attempt:1 highly:1 evaluation:2 alignment:4 introduces:1 navigation:2 capable:1 partial:3 experience:1 modest:1 institut:1 initialized:1 minimal:3 increased:1 column:1 modeling:3 lamar:1 phrase:9 cost:1 technische:1 subset:3 rare:1 tour:1 uniform:1 krizhevsky:1 successful:1 galley:1 too:1 universitat:1 reported:2 dependency:11 answer:1 cho:9 confident:2 lstm:67 probabilistic:1 decoding:1 together:2 ilya:1 thesis:2 huang:1 possibly:1 emnlp:2 luong:1 simard:1 wojciech:1 de:16 sizeable:1 sec:2 availability:1 caused:1 piece:1 later:1 script:1 wolfgang:1 doing:6 sort:1 maintains:1 parallel:1 complicated:1 kgk2:1 contribution:1 appended:1 air:1 publicly:1 accuracy:2 il:11 afin:2 who:3 likewise:2 ensemble:9 convolutional:2 raw:1 vincent:2 produced:4 fern:1 informatik:1 multiplying:1 researcher:3 visualizes:1 whenever:4 frequency:2 involved:1 cased:1 mohamed:1 stop:1 violence:2 dataset:7 dynamischen:1 popular:1 begun:2 dimensionality:4 improves:3 segmentation:1 back:3 exceeded:1 ta:1 supervised:2 day:1 specify:1 improved:2 formulation:2 erences:2 furthermore:1 tre:4 lstms:21 unsegmented:1 google:8 minibatch:4 french:8 quality:2 believe:2 mary:6 building:1 verify:2 read:3 attractive:1 interspeech:2 encourages:1 complete:4 performs:1 passive:3 meaning:4 image:1 novel:1 recently:1 admires:2 mt:7 insensitive:1 extend:1 bougares:1 significant:1 probl:1 automatic:2 rd:1 outlined:1 similarly:1 language:19 had:2 wmt:8 scandinavian:1 haddow:1 align:1 closest:1 recent:1 pl1:1 showed:1 reverse:2 discard:1 perplexity:2 corp:2 schmidhuber:5 success:1 seen:1 additional:1 greater:2 deng:2 parallelized:1 barry:1 exploding:1 signal:1 corrado:1 multiple:1 smooth:1 technical:1 believed:1 long:27 zweig:1 divided:2 va:1 prediction:5 variant:3 halving:1 essentially:1 arxiv:10 monga:2 achieved:6 cell:1 beam:15 hochreiter:4 source:22 parallelization:1 markedly:1 sure:1 smt:13 tend:2 elegant:1 bahdanau:5 mature:2 suspect:1 comment:1 flow:1 extracting:1 ee:2 feedforward:3 bengio:9 enough:2 easy:2 embeddings:1 gave:4 architecture:4 ciresan:1 idea:2 devlin:2 haffner:1 administration:3 pca:2 sentiment:1 suffer:2 speech:6 tape:1 deep:15 useful:4 iterating:1 clear:2 covered:1 ten:1 dit:1 reduced:2 simplest:1 generate:1 outperform:3 http:1 notice:1 per:2 correctly:1 sundermeyer:1 key:1 four:1 threshold:2 clean:1 nal:1 ht:3 dahl:2 kenneth:1 timestep:2 wasted:1 eme:1 enforced:1 powerful:3 master:1 reader:1 scaling:1 qui:5 bit:1 layer:8 followed:1 gomez:1 quadratic:1 oracle:1 ahead:1 constraint:1 unlimited:1 speed:2 extremely:2 mikolov:3 relatively:3 gpus:1 speedup:1 according:2 poor:2 makhoul:1 brno:1 karafi:1 shallow:2 qu:7 making:1 temp:2 quoc:1 invariant:2 taken:1 equation:2 turn:2 mechanism:2 hh:1 fail:1 instrument:1 end:7 lieu:1 gulcehre:1 available:2 apply:1 munchen:1 ney:1 batch:3 voice:4 assumes:1 remaining:1 include:1 trouble:1 establish:1 unchanged:1 objective:2 question:3 added:1 strategy:1 primary:1 september:1 gradient:9 iclr:1 reversed:15 distance:3 separate:1 card:6 unable:1 decoder:7 sensible:2 thank:1 me:3 topic:1 portable:3 extent:1 bleu:20 tokenized:2 length:14 relationship:1 difficult:5 executed:1 ilyasu:1 implementation:3 reliably:1 perform:1 descent:2 beat:1 hinton:4 communication:2 incorporated:1 team:1 y1:7 discovered:1 auli:2 arbitrary:1 paraphrase:1 overcoming:1 introduced:3 pair:2 namely:1 meier:1 sentence:84 connection:1 imagenet:1 acoustic:1 learned:2 boost:1 nip:1 address:3 able:1 alongside:1 below:1 challenge:1 perl:1 max:1 memory:7 explanation:1 garden:6 including:1 power:2 greatest:1 difficulty:4 natural:3 rely:1 cernock:1 zhu:1 representing:2 scheme:1 conseil:4 technology:1 ne:2 axis:2 extract:1 naive:1 epoch:3 graf:4 macherey:1 rerank:1 limitation:1 rnnlm:1 geoffrey:1 fixeddimensional:2 downloaded:1 vanhoucke:2 consistent:1 principle:1 neuronalen:1 ulrich:2 roukos:1 translation:43 pile:1 penalized:2 token:2 surprisingly:1 last:1 soon:2 english:8 avant:2 convinced:1 senior:1 benefit:1 edinburgh:1 overcome:1 distributed:1 vocabulary:12 evaluating:1 van:1 qvl:1 computes:3 resides:1 forward:2 made:3 depth:1 nguyen:1 far:4 transaction:2 approximate:1 multilingual:1 reproduces:1 active:3 conclude:1 search:4 un:6 continuous:1 table:6 additionally:1 learn:4 nature:1 robust:1 obtaining:1 du:8 excellent:2 automobile:2 bottou:1 domain:1 did:5 apr:2 main:2 multilayered:1 x1:6 body:1 fig:3 xu:1 en:2 slow:1 experienced:1 momentum:1 decoded:2 answering:1 yh:1 third:1 learns:3 xt:7 specific:1 zu:1 showing:1 symbol:2 list:8 abadie:2 exists:1 workshop:1 sequential:1 phd:1 labelling:1 conditioned:1 margin:1 chen:1 easier:3 simply:1 likely:5 visual:1 vinyals:2 expressed:1 contained:1 khudanpur:1 monotonic:2 corresponds:2 truth:5 abc:1 minibatches:1 succeed:1 conditional:2 sized:1 goal:1 sorted:4 ann:2 marked:1 room:1 jeff:1 man:1 considerable:1 hard:2 experimentally:1 typical:1 except:1 reversing:7 degradation:2 total:1 experimental:1 e:8 la:5 est:1 attempted:2 internal:1 support:1 audi:2 inability:1 rajat:1 oriol:1 evaluate:2 audio:1 trainable:1 phenomenon:1 |
4,801 | 5,347 | How transferable are features in deep neural
networks?
Jason Yosinski,1 Jeff Clune,2 Yoshua Bengio,3 and Hod Lipson4
1
Dept. Computer Science, Cornell University
2
Dept. Computer Science, University of Wyoming
3
Dept. Computer Science & Operations Research, University of Montreal
4
Dept. Mechanical & Aerospace Engineering, Cornell University
Abstract
Many deep neural networks trained on natural images exhibit a curious phenomenon in common: on the first layer they learn features similar to Gabor filters
and color blobs. Such first-layer features appear not to be specific to a particular
dataset or task, but general in that they are applicable to many datasets and tasks.
Features must eventually transition from general to specific by the last layer of
the network, but this transition has not been studied extensively. In this paper we
experimentally quantify the generality versus specificity of neurons in each layer
of a deep convolutional neural network and report a few surprising results. Transferability is negatively affected by two distinct issues: (1) the specialization of
higher layer neurons to their original task at the expense of performance on the
target task, which was expected, and (2) optimization difficulties related to splitting networks between co-adapted neurons, which was not expected. In an example network trained on ImageNet, we demonstrate that either of these two issues
may dominate, depending on whether features are transferred from the bottom,
middle, or top of the network. We also document that the transferability of features decreases as the distance between the base task and target task increases, but
that transferring features even from distant tasks can be better than using random
features. A final surprising result is that initializing a network with transferred
features from almost any number of layers can produce a boost to generalization
that lingers even after fine-tuning to the target dataset.
1
Introduction
Modern deep neural networks exhibit a curious phenomenon: when trained on images, they all tend
to learn first-layer features that resemble either Gabor filters or color blobs. The appearance of these
filters is so common that obtaining anything else on a natural image dataset causes suspicion of
poorly chosen hyperparameters or a software bug. This phenomenon occurs not only for different
datasets, but even with very different training objectives, including supervised image classification
(Krizhevsky et al., 2012), unsupervised density learning (Lee et al., 2009), and unsupervised learning of sparse representations (Le et al., 2011).
Because finding these standard features on the first layer seems to occur regardless of the exact cost
function and natural image dataset, we call these first-layer features general. On the other hand, we
know that the features computed by the last layer of a trained network must depend greatly on the
chosen dataset and task. For example, in a network with an N-dimensional softmax output layer that
has been successfully trained toward a supervised classification objective, each output unit will be
specific to a particular class. We thus call the last-layer features specific. These are intuitive notions
of general and specific for which we will provide more rigorous definitions below. If first-layer
1
features are general and last-layer features are specific, then there must be a transition from general
to specific somewhere in the network. This observation raises a few questions:
? Can we quantify the degree to which a particular layer is general or specific?
? Does the transition occur suddenly at a single layer, or is it spread out over several layers?
? Where does this transition take place: near the first, middle, or last layer of the network?
We are interested in the answers to these questions because, to the extent that features within a
network are general, we will be able to use them for transfer learning (Caruana, 1995; Bengio
et al., 2011; Bengio, 2011). In transfer learning, we first train a base network on a base dataset and
task, and then we repurpose the learned features, or transfer them, to a second target network to
be trained on a target dataset and task. This process will tend to work if the features are general,
meaning suitable to both base and target tasks, instead of specific to the base task.
When the target dataset is significantly smaller than the base dataset, transfer learning can be a
powerful tool to enable training a large target network without overfitting; Recent studies have
taken advantage of this fact to obtain state-of-the-art results when transferring from higher layers
(Donahue et al., 2013a; Zeiler and Fergus, 2013; Sermanet et al., 2014), collectively suggesting that
these layers of neural networks do indeed compute features that are fairly general. These results
further emphasize the importance of studying the exact nature and extent of this generality.
The usual transfer learning approach is to train a base network and then copy its first n layers to
the first n layers of a target network. The remaining layers of the target network are then randomly
initialized and trained toward the target task. One can choose to backpropagate the errors from
the new task into the base (copied) features to fine-tune them to the new task, or the transferred
feature layers can be left frozen, meaning that they do not change during training on the new task.
The choice of whether or not to fine-tune the first n layers of the target network depends on the
size of the target dataset and the number of parameters in the first n layers. If the target dataset is
small and the number of parameters is large, fine-tuning may result in overfitting, so the features
are often left frozen. On the other hand, if the target dataset is large or the number of parameters is
small, so that overfitting is not a problem, then the base features can be fine-tuned to the new task
to improve performance. Of course, if the target dataset is very large, there would be little need to
transfer because the lower level filters could just be learned from scratch on the target dataset. We
compare results from each of these two techniques ? fine-tuned features or frozen features ? in
the following sections.
In this paper we make several contributions:
1. We define a way to quantify the degree to which a particular layer is general or specific, namely,
how well features at that layer transfer from one task to another (Section 2). We then train pairs
of convolutional neural networks on the ImageNet dataset and characterize the layer-by-layer
transition from general to specific (Section 4), which yields the following four results.
2. We experimentally show two separate issues that cause performance degradation when using transferred features without fine-tuning: (i) the specificity of the features themselves, and
(ii) optimization difficulties due to splitting the base network between co-adapted neurons on
neighboring layers. We show how each of these two effects can dominate at different layers of
the network. (Section 4.1)
3. We quantify how the performance benefits of transferring features decreases the more dissimilar
the base task and target task are. (Section 4.2)
4. On the relatively large ImageNet dataset, we find lower performance than has been previously
reported for smaller datasets (Jarrett et al., 2009) when using features computed from random
lower-layer weights vs. trained weights. We compare random weights to transferred weights?
both frozen and fine-tuned?and find the transferred weights perform better. (Section 4.3)
5. Finally, we find that initializing a network with transferred features from almost any number
of layers can produce a boost to generalization performance after fine-tuning to a new dataset.
This is particularly surprising because the effect of having seen the first dataset persists even
after extensive fine-tuning. (Section 4.1)
2
2
Generality vs. Specificity Measured as Transfer Performance
We have noted the curious tendency of Gabor filters and color blobs to show up in the first layer of
neural networks trained on natural images. In this study, we define the degree of generality of a set
of features learned on task A as the extent to which the features can be used for another task B. It
is important to note that this definition depends on the similarity between A and B. We create pairs
of classification tasks A and B by constructing pairs of non-overlapping subsets of the ImageNet
dataset.1 These subsets can be chosen to be similar to or different from each other.
To create tasks A and B, we randomly split the 1000 ImageNet classes into two groups each containing 500 classes and approximately half of the data, or about 645,000 examples each. We train
one eight-layer convolutional network on A and another on B. These networks, which we call baseA
and baseB, are shown in the top two rows of Figure 1. We then choose a layer n from {1, 2, . . . , 7}
and train several new networks. In the following explanation and in Figure 1, we use layer n = 3 as
the example layer chosen. First, we define and train the following two networks:
? A selffer network B3B: the first 3 layers are copied from baseB and frozen. The five higher
layers (4?8) are initialized randomly and trained on dataset B. This network is a control for the
next transfer network. (Figure 1, row 3)
? A transfer network A3B: the first 3 layers are copied from baseA and frozen. The five higher
layers (4?8) are initialized randomly and trained toward dataset B. Intuitively, here we copy
the first 3 layers from a network trained on dataset A and then learn higher layer features on top
of them to classify a new target dataset B. If A3B performs as well as baseB, there is evidence
that the third-layer features are general, at least with respect to B. If performance suffers, there
is evidence that the third-layer features are specific to A. (Figure 1, row 4)
We repeated this process for all n in {1, 2, . . . , 7}2 and in both directions (i.e. AnB and BnA). In
the above two networks, the transferred layers are frozen. We also create versions of the above two
networks where the transferred layers are fine-tuned:
? A selffer network B3B+ : just like B3B, but where all layers learn.
? A transfer network A3B+ : just like A3B, but where all layers learn.
To create base and target datasets that are similar to each other, we randomly assign half of the 1000
ImageNet classes to A and half to B. ImageNet contains clusters of similar classes, particularly dogs
and cats, like these 13 classes from the biological family Felidae: {tabby cat, tiger cat, Persian cat,
Siamese cat, Egyptian cat, mountain lion, lynx, leopard, snow leopard, jaguar, lion, tiger, cheetah}.
On average, A and B will each contain approximately 6 or 7 of these felid classes, meaning that
base networks trained on each dataset will have features at all levels that help classify some types
of felids. When generalizing to the other dataset, we would expect that the new high-level felid
detectors trained on top of old low-level felid detectors would work well. Thus A and B are similar
when created by randomly assigning classes to each, and we expect that transferred features will
perform better than when A and B are less similar.
Fortunately, in ImageNet we are also provided with a hierarchy of parent classes. This information
allowed us to create a special split of the dataset into two halves that are as semantically different
from each other as possible: with dataset A containing only man-made entities and B containing
natural entities. The split is not quite even, with 551 classes in the man-made group and 449 in the
natural group. Further details of this split and the classes in each half are given in the supplementary
material. In Section 4.2 we will show that features transfer more poorly (i.e. they are more specific)
when the datasets are less similar.
1
The ImageNet dataset, as released in the Large Scale Visual Recognition Challenge 2012 (ILSVRC2012)
(Deng et al., 2009) contains 1,281,167 labeled training images and 50,000 test images, with each image labeled
with one of 1000 classes.
2
Note that n = 8 doesn?t make sense in either case: B8B is just baseB, and A8B would not work because
it is never trained on B.
3
WA1
WA2
WA3
WA4
WA5
WA6
WA7
WA8
input
A
WB1
WB2
WB3
WB4
WB5
input
B
WB1
WB2
WB3
or
or
or
WA1
WA2
WA3
or
or
or
WB6
WB7
labels
A
baseA
labels
B
baseB
WB8
B3B
and
B3B+
A3B
and
A3B+
Figure 1: Overview of the experimental treatments and controls. Top two rows: The base networks
are trained using standard supervised backprop on only half of the ImageNet dataset (first row: A
half, second row: B half). The labeled rectangles (e.g. WA1 ) represent the weight vector learned for
that layer, with the color indicating which dataset the layer was originally trained on. The vertical,
ellipsoidal bars between weight vectors represent the activations of the network at each layer. Third
row: In the selffer network control, the first n weight layers of the network (in this example, n = 3)
are copied from a base network (e.g. one trained on dataset B), the upper 8 ? n layers are randomly
initialized, and then the entire network is trained on that same dataset (in this example, dataset B).
The first n layers are either locked during training (?frozen? selffer treatment B3B) or allowed to
learn (?fine-tuned? selffer treatment B3B+ ). This treatment reveals the occurrence of fragile coadaptation, when neurons on neighboring layers co-adapt during training in such a way that cannot
be rediscovered when one layer is frozen. Fourth row: The transfer network experimental treatment
is the same as the selffer treatment, except that the first n layers are copied from a network trained
on one dataset (e.g. A) and then the entire network is trained on the other dataset (e.g. B). This
treatment tests the extent to which the features on layer n are general or specific.
3
Experimental Setup
Since Krizhevsky et al. (2012) won the ImageNet 2012 competition, there has been much interest
and work toward tweaking hyperparameters of large convolutional models. However, in this study
we aim not to maximize absolute performance, but rather to study transfer results on a well-known
architecture. We use the reference implementation provided by Caffe (Jia et al., 2014) so that our
results will be comparable, extensible, and useful to a large number of researchers. Further details of
the training setup (learning rates, etc.) are given in the supplementary material, and code and parameter files to reproduce these experiments are available at http://yosinski.com/transfer.
4
Results and Discussion
We performed three sets of experiments. The main experiment has random A/B splits and is discussed in Section 4.1. Section 4.2 presents an experiment with the man-made/natural split. Section 4.3 describes an experiment with random weights.
4
0.66
Top-1 accuracy (higher is better)
0.64
0.62
0.60
0.58
baseB
selffer BnB
selffer BnB +
transfer AnB
transfer AnB +
0.56
0.54
0.52
0
1
2
3
4
5
6
7
5: Transfer + fine-tuning improves generalization
Top-1 accuracy (higher is better)
0.64
3: Fine-tuning recovers co-adapted interactions
0.62
2: Performance drops
due to fragile
co-adaptation
4: Performance
drops due to
representation
specificity
0.60
0.58
0.56
0.54
0
1
2
3
4
5
Layer n at which network is chopped and retrained
6
7
Figure 2: The results from this paper?s main experiment. Top: Each marker in the figure represents
the average accuracy over the validation set for a trained network. The white circles above n =
0 represent the accuracy of baseB. There are eight points, because we tested on four separate
random A/B splits. Each dark blue dot represents a BnB network. Light blue points represent
BnB+ networks, or fine-tuned versions of BnB. Dark red diamonds are AnB networks, and light
red diamonds are the fine-tuned AnB+ versions. Points are shifted slightly left or right for visual
clarity. Bottom: Lines connecting the means of each treatment. Numbered descriptions above each
line refer to which interpretation from Section 4.1 applies.
4.1
Similar Datasets: Random A/B splits
The results of all A/B transfer learning experiments on randomly split (i.e. similar) datasets are
shown3 in Figure 2. The results yield many different conclusions. In each of the following interpretations, we compare the performance to the base case (white circles and dotted line in Figure 2).
3
AnA networks and BnB networks are statistically equivalent, because in both cases a network is trained
on 500 random classes. To simplify notation we label these BnB networks. Similarly, we have aggregated the
statistically identical BnA and AnB networks and just call them AnB.
5
1. The white baseB circles show that a network trained to classify a random subset of 500 classes
attains a top-1 accuracy of 0.625, or 37.5% error. This error is lower than the 42.5% top-1 error
attained on the 1000-class network. While error might have been higher because the network is
trained on only half of the data, which could lead to more overfitting, the net result is that error is
lower because there are only 500 classes, so there are only half as many ways to make mistakes.
2. The dark blue BnB points show a curious behavior. As expected, performance at layer one is
the same as the baseB points. That is, if we learn eight layers of features, save the first layer of
learned Gabor features and color blobs, reinitialize the whole network, and retrain it toward the
same task, it does just as well. This result also holds true for layer 2. However, layers 3, 4, 5,
and 6, particularly 4 and 5, exhibit worse performance. This performance drop is evidence that
the original network contained fragile co-adapted features on successive layers, that is, features
that interact with each other in a complex or fragile way such that this co-adaptation could not be
relearned by the upper layers alone. Gradient descent was able to find a good solution the first
time, but this was only possible because the layers were jointly trained. By layer 6 performance
is nearly back to the base level, as is layer 7. As we get closer and closer to the final, 500-way
softmax output layer 8, there is less to relearn, and apparently relearning these one or two layers
is simple enough for gradient descent to find a good solution. Alternately, we may say that
there is less co-adaptation of features between layers 6 & 7 and between 7 & 8 than between
previous layers. To our knowledge it has not been previously observed in the literature that such
optimization difficulties may be worse in the middle of a network than near the bottom or top.
3. The light blue BnB+ points show that when the copied, lower-layer features also learn on the
target dataset (which here is the same as the base dataset), performance is similar to the base
case. Such fine-tuning thus prevents the performance drop observed in the BnB networks.
4. The dark red AnB diamonds show the effect we set out to measure in the first place: the transferability of features from one network to another at each layer. Layers one and two transfer almost
perfectly from A to B, giving evidence that, at least for these two tasks, not only are the first-layer
Gabor and color blob features general, but the second layer features are general as well. Layer
three shows a slight drop, and layers 4-7 show a more significant drop in performance. Thanks
to the BnB points, we can tell that this drop is from a combination of two separate effects: the
drop from lost co-adaptation and the drop from features that are less and less general. On layers
3, 4, and 5, the first effect dominates, whereas on layers 6 and 7 the first effect diminishes and
the specificity of representation dominates the drop in performance.
Although examples of successful feature transfer have been reported elsewhere in the literature
(Girshick et al., 2013; Donahue et al., 2013b), to our knowledge these results have been limited
to noticing that transfer from a given layer is much better than the alternative of training strictly
on the target task, i.e. noticing that the AnB points at some layer are much better than training
all layers from scratch. We believe this is the first time that (1) the extent to which transfer is
successful has been carefully quantified layer by layer, and (2) that these two separate effects
have been decoupled, showing that each effect dominates in part of the regime.
5. The light red AnB+ diamonds show a particularly surprising effect: that transferring features
and then fine-tuning them results in networks that generalize better than those trained directly on
the target dataset. Previously, the reason one might want to transfer learned features is to enable
training without overfitting on small target datasets, but this new result suggests that transferring
features will boost generalization performance even if the target dataset is large. Note that this
effect should not be attributed to the longer total training time (450k base iterations + 450k finetuned iterations for AnB+ vs. 450k for baseB), because the BnB+ networks are also trained
for the same longer length of time and do not exhibit this same performance improvement.
Thus, a plausible explanation is that even after 450k iterations of fine-tuning (beginning with
completely random top layers), the effects of having seen the base dataset still linger, boosting
generalization performance. It is surprising that this effect lingers through so much retraining.
This generalization improvement seems not to depend much on how much of the first network
we keep to initialize the second network: keeping anywhere from one to seven layers produces
improved performance, with slightly better performance as we keep more layers. The average
boost across layers 1 to 7 is 1.6% over the base case, and the average if we keep at least five
layers is 2.1%.4 The degree of performance boost is shown in Table 1.
4
We aggregate performance over several layers because each point is computationally expensive to obtain
(9.5 days on a GPU), so at the time of publication we have few data points per layer. The aggregation is
6
Table 1: Performance boost of AnB+ over controls, averaged over different ranges of layers.
layers
aggregated
1-7
3-7
5-7
4.2
mean boost
over
baseB
1.6%
1.8%
2.1%
mean boost
over
selffer BnB+
1.4%
1.4%
1.7%
Dissimilar Datasets: Splitting Man-made and Natural Classes Into Separate Datasets
As mentioned previously, the effectiveness of feature transfer is expected to decline as the base and
target tasks become less similar. We test this hypothesis by comparing transfer performance on
similar datasets (the random A/B splits discussed above) to that on dissimilar datasets, created by
assigning man-made object classes to A and natural object classes to B. This man-made/natural split
creates datasets as dissimilar as possible within the ImageNet dataset.
The upper-left subplot of Figure 3 shows the accuracy of a baseA and baseB network (white circles)
and BnA and AnB networks (orange hexagons). Lines join common target tasks. The upper of the
two lines contains those networks trained toward the target task containing natural categories (baseB
and AnB). These networks perform better than those trained toward the man-made categories, which
may be due to having only 449 classes instead of 551, or simply being an easier task, or both.
4.3
Random Weights
We also compare to random, untrained weights because Jarrett et al. (2009) showed ? quite strikingly ? that the combination of random convolutional filters, rectification, pooling, and local normalization can work almost as well as learned features. They reported this result on relatively small
networks of two or three learned layers and on the smaller Caltech-101 dataset (Fei-Fei et al., 2004).
It is natural to ask whether or not the nearly optimal performance of random filters they report carries
over to a deeper network trained on a larger dataset.
The upper-right subplot of Figure 3 shows the accuracy obtained when using random filters for the
first n layers for various choices of n. Performance falls off quickly in layers 1 and 2, and then
drops to near-chance levels for layers 3+, which suggests that getting random weights to work in
convolutional neural networks may not be as straightforward as it was for the smaller network size
and smaller dataset used by Jarrett et al. (2009). However, the comparison is not straightforward.
Whereas our networks have max pooling and local normalization on layers 1 and 2, just as Jarrett
et al. (2009) did, we use a different nonlinearity (relu(x) instead of abs(tanh(x))), different layer
sizes and number of layers, as well as other differences. Additionally, their experiment only considered two layers of random weights. The hyperparameter and architectural choices of our network
collectively provide one new datapoint, but it may well be possible to tweak layer sizes and random
initialization details to enable much better performance for random weights.5
The bottom subplot of Figure 3 shows the results of the experiments of the previous two sections
after subtracting the performance of their individual base cases. These normalized performances
are plotted across the number of layers n that are either random or were trained on a different,
base dataset. This comparison makes two things apparent. First, the transferability gap when using
frozen features grows more quickly as n increases for dissimilar tasks (hexagons) than similar tasks
(diamonds), with a drop by the final layer for similar tasks of only 8% vs. 25% for dissimilar tasks.
Second, transferring even from a distant task is better than using random filters. One possible reason
this latter result may differ from Jarrett et al. (2009) is because their fully-trained (non-random)
networks were overfitting more on the smaller Caltech-101 dataset than ours on the larger ImageNet
informative, however, because the performance at each layer is based on different random draws of the upper
layer initialization weights. Thus, the fact that layers 5, 6, and 7 result in almost identical performance across
random draws suggests that multiple runs at a given layer would result in similar performance.
5
For example, the training loss of the network with three random layers failed to converge, producing only
chance-level validation performance. Much better convergence may be possible with different hyperparameters.
7
Man-made/Natural split
Top-1 accuracy
0.7
0.5
0.6
0.4
0.5
0.3
0.4
0.2
0.1
0.3
0
Relative top-1 accuracy (higher is better)
Random, untrained filters
0.6
1
2
3
4
5
6
0.0
7
0
1
2
3
4
5
6
7
0.00
?0.05
?0.10
?0.15
reference
mean AnB, random splits
mean AnB, m/n split
random features
?0.20
?0.25
?0.30
0
1
2
3
4
5
Layer n at which network is chopped and retrained
6
7
Figure 3: Performance degradation vs. layer. Top left: Degradation when transferring between dissimilar tasks (from man-made classes of ImageNet to natural classes or vice versa). The upper line
connects networks trained to the ?natural? target task, and the lower line connects those trained toward the ?man-made? target task. Top right: Performance when the first n layers consist of random,
untrained weights. Bottom: The top two plots compared to the random A/B split from Section 4.1
(red diamonds), all normalized by subtracting their base level performance.
dataset, making their random filters perform better by comparison. In the supplementary material,
we provide an extra experiment indicating the extent to which our networks are overfit.
5
Conclusions
We have demonstrated a method for quantifying the transferability of features from each layer of
a neural network, which reveals their generality or specificity. We showed how transferability is
negatively affected by two distinct issues: optimization difficulties related to splitting networks in
the middle of fragilely co-adapted layers and the specialization of higher layer features to the original
task at the expense of performance on the target task. We observed that either of these two issues
may dominate, depending on whether features are transferred from the bottom, middle, or top of
the network. We also quantified how the transferability gap grows as the distance between tasks
increases, particularly when transferring higher layers, but found that even features transferred from
distant tasks are better than random weights. Finally, we found that initializing with transferred
features can improve generalization performance even after substantial fine-tuning on a new task,
which could be a generally useful technique for improving deep neural network performance.
Acknowledgments
The authors would like to thank Kyunghyun Cho and Thomas Fuchs for helpful discussions, Joost
Huizinga, Anh Nguyen, and Roby Velez for editing, as well as funding from the NASA Space
Technology Research Fellowship (JY), DARPA project W911NF-12-1-0449, NSERC, Ubisoft, and
CIFAR (YB is a CIFAR Fellow).
8
References
Bengio, Y. (2011). Deep learning of representations for unsupervised and transfer learning. In JMLR W&CP:
Proc. Unsupervised and Transfer Learning.
Bengio, Y., Bastien, F., Bergeron, A., Boulanger-Lewandowski, N., Breuel, T., Chherawala, Y., Cisse, M., C?ot?e,
M., Erhan, D., Eustache, J., Glorot, X., Muller, X., Pannetier Lebeuf, S., Pascanu, R., Rifai, S., Savard, F.,
and Sicard, G. (2011). Deep learners benefit more from out-of-distribution examples. In JMLR W&CP:
Proc. AISTATS?2011.
Caruana, R. (1995). Learning many related tasks at the same time with backpropagation. pages 657?664,
Cambridge, MA. MIT Press.
Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L. (2009). ImageNet: A Large-Scale Hierarchical
Image Database. In CVPR09.
Donahue, J., Jia, Y., Vinyals, O., Hoffman, J., Zhang, N., Tzeng, E., and Darrell, T. (2013a). Decaf:
A deep convolutional activation feature for generic visual recognition. Technical report, arXiv preprint
arXiv:1310.1531.
Donahue, J., Jia, Y., Vinyals, O., Hoffman, J., Zhang, N., Tzeng, E., and Darrell, T. (2013b). Decaf: A deep
convolutional activation feature for generic visual recognition. arXiv preprint arXiv:1310.1531.
Fei-Fei, L., Fergus, R., and Perona, P. (2004). Learning generative visual models from few training examples:
An incremental Bayesian approach tested on 101 object categories. In Conference on Computer Vision and
Pattern Recognition Workshop (CVPR 2004), page 178.
Girshick, R., Donahue, J., Darrell, T., and Malik, J. (2013). Rich feature hierarchies for accurate object detection and semantic segmentation. arXiv preprint arXiv:1311.2524.
Jarrett, K., Kavukcuoglu, K., Ranzato, M., and LeCun, Y. (2009). What is the best multi-stage architecture for
object recognition? In Proc. International Conference on Computer Vision (ICCV?09), pages 2146?2153.
IEEE.
Jia, Y., Shelhamer, E., Donahue, J., Karayev, S., Long, J., Girshick, R., Guadarrama, S., and Darrell, T. (2014).
Caffe: Convolutional architecture for fast feature embedding. arXiv preprint arXiv:1408.5093.
Krizhevsky, A., Sutskever, I., and Hinton, G. (2012). ImageNet classification with deep convolutional neural
networks. In Advances in Neural Information Processing Systems 25 (NIPS?2012).
Le, Q. V., Karpenko, A., Ngiam, J., and Ng, A. Y. (2011). ICA with reconstruction cost for efficient overcomplete feature learning. In J. Shawe-Taylor, R. Zemel, P. Bartlett, F. Pereira, and K. Weinberger, editors,
Advances in Neural Information Processing Systems 24, pages 1017?1025.
Lee, H., Grosse, R., Ranganath, R., and Ng, A. Y. (2009). Convolutional deep belief networks for scalable
unsupervised learning of hierarchical representations. Montreal, Canada.
Sermanet, P., Eigen, D., Zhang, X., Mathieu, M., Fergus, R., and LeCun, Y. (2014). Overfeat: Integrated recognition, localization and detection using convolutional networks. In International Conference on Learning
Representations (ICLR 2014). CBLS.
Zeiler, M. D. and Fergus, R. (2013). Visualizing and understanding convolutional networks. Technical Report
Arxiv 1311.2901.
9
| 5347 |@word version:3 middle:5 seems:2 retraining:1 carry:1 coadaptation:1 contains:3 tuned:7 document:1 ours:1 guadarrama:1 transferability:7 com:1 surprising:5 comparing:1 activation:3 assigning:2 must:3 gpu:1 distant:3 informative:1 drop:12 plot:1 v:5 alone:1 half:10 generative:1 beginning:1 boosting:1 pascanu:1 successive:1 zhang:3 five:3 become:1 ica:1 indeed:1 expected:4 themselves:1 behavior:1 cheetah:1 multi:1 little:1 provided:2 project:1 notation:1 anh:1 what:1 mountain:1 finding:1 wa2:2 fellow:1 control:4 unit:1 appear:1 producing:1 engineering:1 persists:1 local:2 mistake:1 approximately:2 might:2 initialization:2 studied:1 quantified:2 suggests:3 co:10 repurpose:1 limited:1 locked:1 jarrett:6 statistically:2 averaged:1 range:1 acknowledgment:1 lecun:2 lost:1 backpropagation:1 gabor:5 significantly:1 a3b:6 bergeron:1 tweaking:1 specificity:6 numbered:1 get:1 cannot:1 equivalent:1 demonstrated:1 straightforward:2 regardless:1 splitting:4 dominate:3 lewandowski:1 embedding:1 notion:1 cvpr09:1 target:31 hierarchy:2 exact:2 hypothesis:1 recognition:6 particularly:5 finetuned:1 expensive:1 labeled:3 database:1 bottom:6 observed:3 preprint:4 initializing:3 ilsvrc2012:1 ranzato:1 decrease:2 mentioned:1 substantial:1 jaguar:1 trained:35 depend:2 raise:1 negatively:2 creates:1 localization:1 learner:1 completely:1 strikingly:1 joost:1 darpa:1 cat:6 various:1 train:6 distinct:2 fast:1 zemel:1 tell:1 aggregate:1 caffe:2 lynx:1 quite:2 supplementary:3 plausible:1 larger:2 say:1 apparent:1 cvpr:1 jointly:1 final:3 blob:5 advantage:1 frozen:10 net:1 breuel:1 karayev:1 reconstruction:1 subtracting:2 interaction:1 adaptation:4 karpenko:1 neighboring:2 poorly:2 bug:1 intuitive:1 description:1 competition:1 getting:1 sutskever:1 parent:1 cluster:1 convergence:1 darrell:4 produce:3 incremental:1 object:5 help:1 depending:2 montreal:2 measured:1 resemble:1 lingers:2 quantify:4 differ:1 direction:1 snow:1 filter:11 enable:3 ana:1 material:3 backprop:1 assign:1 generalization:7 biological:1 leopard:2 strictly:1 tabby:1 hold:1 considered:1 released:1 diminishes:1 proc:3 applicable:1 label:3 tanh:1 vice:1 create:5 successfully:1 tool:1 hoffman:2 bna:3 mit:1 aim:1 rather:1 cornell:2 publication:1 clune:1 improvement:2 greatly:1 rigorous:1 attains:1 sense:1 helpful:1 entire:2 transferring:8 integrated:1 perona:1 reproduce:1 interested:1 issue:5 classification:4 overfeat:1 art:1 softmax:2 fairly:1 special:1 initialize:1 orange:1 tzeng:2 never:1 having:3 ng:2 identical:2 represents:2 unsupervised:5 nearly:2 yoshua:1 report:4 simplify:1 few:4 modern:1 randomly:8 individual:1 connects:2 ab:1 detection:2 interest:1 rediscovered:1 light:4 accurate:1 closer:2 decoupled:1 old:1 taylor:1 initialized:4 circle:4 plotted:1 overcomplete:1 girshick:3 classify:3 extensible:1 w911nf:1 caruana:2 cost:2 tweak:1 subset:3 krizhevsky:3 successful:2 characterize:1 reported:3 answer:1 cho:1 thanks:1 density:1 international:2 lee:2 off:1 dong:1 connecting:1 quickly:2 containing:4 choose:2 worse:2 relearned:1 li:2 suggesting:1 depends:2 performed:1 jason:1 apparently:1 red:5 aggregation:1 jia:4 contribution:1 accuracy:9 convolutional:13 yield:2 generalize:1 bayesian:1 kavukcuoglu:1 researcher:1 detector:2 datapoint:1 suffers:1 definition:2 attributed:1 recovers:1 dataset:47 treatment:8 ask:1 color:6 knowledge:2 improves:1 segmentation:1 carefully:1 back:1 nasa:1 higher:11 originally:1 supervised:3 attained:1 day:1 improved:1 editing:1 yb:1 generality:5 just:7 anywhere:1 stage:1 relearn:1 hand:2 overfit:1 overlapping:1 marker:1 believe:1 grows:2 effect:12 contain:1 true:1 normalized:2 kyunghyun:1 semantic:1 white:4 visualizing:1 during:3 transferable:1 anything:1 noted:1 won:1 demonstrate:1 performs:1 cp:2 image:10 meaning:3 funding:1 common:3 overview:1 discussed:2 yosinski:2 interpretation:2 slight:1 velez:1 refer:1 significant:1 versa:1 cambridge:1 tuning:11 similarly:1 nonlinearity:1 shawe:1 dot:1 similarity:1 longer:2 etc:1 base:26 recent:1 showed:2 muller:1 caltech:2 seen:2 fortunately:1 subplot:3 deng:2 aggregated:2 maximize:1 converge:1 ii:1 persian:1 siamese:1 multiple:1 technical:2 adapt:1 long:1 cifar:2 jy:1 scalable:1 vision:2 arxiv:9 iteration:3 represent:4 normalization:2 whereas:2 want:1 fine:20 chopped:2 fellowship:1 else:1 extra:1 ot:1 file:1 pooling:2 tend:2 thing:1 effectiveness:1 call:4 curious:4 near:3 bengio:5 split:15 enough:1 relu:1 architecture:3 perfectly:1 decline:1 rifai:1 fragile:4 whether:4 specialization:2 fuchs:1 bartlett:1 cause:2 deep:11 useful:2 generally:1 tune:2 dark:4 ellipsoidal:1 extensively:1 category:3 http:1 shifted:1 dotted:1 per:1 blue:4 hyperparameter:1 affected:2 group:3 four:2 clarity:1 rectangle:1 run:1 noticing:2 powerful:1 fourth:1 linger:1 place:2 almost:5 family:1 architectural:1 wa1:3 draw:2 comparable:1 layer:122 hexagon:2 copied:6 adapted:5 occur:2 fei:6 software:1 relatively:2 transferred:13 combination:2 smaller:6 describes:1 slightly:2 across:3 making:1 intuitively:1 iccv:1 taken:1 computationally:1 rectification:1 previously:4 eventually:1 know:1 studying:1 available:1 operation:1 eight:3 hierarchical:2 generic:2 occurrence:1 save:1 alternative:1 weinberger:1 eigen:1 original:3 thomas:1 top:18 remaining:1 zeiler:2 somewhere:1 giving:1 boulanger:1 suddenly:1 objective:2 malik:1 question:2 occurs:1 reinitialize:1 usual:1 exhibit:4 gradient:2 iclr:1 distance:2 separate:5 thank:1 entity:2 seven:1 extent:6 toward:8 reason:2 savard:1 code:1 length:1 sermanet:2 setup:2 expense:2 implementation:1 perform:4 diamond:6 upper:7 vertical:1 neuron:5 observation:1 datasets:13 descent:2 hinton:1 anb:16 retrained:2 canada:1 namely:1 mechanical:1 pair:3 extensive:1 dog:1 imagenet:16 aerospace:1 egyptian:1 learned:8 boost:8 alternately:1 nip:1 able:2 bar:1 below:1 lion:2 pattern:1 regime:1 challenge:1 including:1 max:1 explanation:2 belief:1 suitable:1 natural:15 difficulty:4 improve:2 technology:1 cisse:1 suspicion:1 mathieu:1 created:2 literature:2 understanding:1 relative:1 fully:1 expect:2 loss:1 versus:1 validation:2 shelhamer:1 degree:4 editor:1 row:8 course:1 elsewhere:1 last:5 copy:2 keeping:1 deeper:1 fall:1 absolute:1 sparse:1 benefit:2 transition:6 rich:1 doesn:1 author:1 made:10 nguyen:1 erhan:1 ranganath:1 emphasize:1 keep:3 overfitting:6 reveals:2 fergus:4 table:2 additionally:1 scratch:2 learn:8 transfer:28 nature:1 obtaining:1 improving:1 interact:1 ngiam:1 untrained:3 complex:1 constructing:1 did:1 aistats:1 spread:1 main:2 whole:1 hyperparameters:3 repeated:1 allowed:2 retrain:1 join:1 grosse:1 pereira:1 jmlr:2 third:3 donahue:6 specific:14 bastien:1 showing:1 evidence:4 dominates:3 consist:1 glorot:1 socher:1 workshop:1 importance:1 decaf:2 hod:1 relearning:1 gap:2 easier:1 backpropagate:1 generalizing:1 simply:1 appearance:1 visual:5 failed:1 prevents:1 vinyals:2 contained:1 nserc:1 collectively:2 applies:1 chance:2 ma:1 quantifying:1 jeff:1 man:10 change:1 experimentally:2 tiger:2 except:1 semantically:1 degradation:3 total:1 ubisoft:1 tendency:1 experimental:3 indicating:2 latter:1 dissimilar:7 bnb:13 dept:4 tested:2 phenomenon:3 |
4,802 | 5,348 | Convolutional Kernel Networks
Julien Mairal, Piotr Koniusz, Zaid Harchaoui, and Cordelia Schmid
Inria?
[email protected]
Abstract
An important goal in visual recognition is to devise image representations that are
invariant to particular transformations. In this paper, we address this goal with a
new type of convolutional neural network (CNN) whose invariance is encoded by
a reproducing kernel. Unlike traditional approaches where neural networks are
learned either to represent data or for solving a classification task, our network
learns to approximate the kernel feature map on training data.
Such an approach enjoys several benefits over classical ones. First, by teaching CNNs to be invariant, we obtain simple network architectures that achieve a
similar accuracy to more complex ones, while being easy to train and robust to
overfitting. Second, we bridge a gap between the neural network literature and
kernels, which are natural tools to model invariance. We evaluate our methodology on visual recognition tasks where CNNs have proven to perform well, e.g.,
digit recognition with the MNIST dataset, and the more challenging CIFAR-10
and STL-10 datasets, where our accuracy is competitive with the state of the art.
1
Introduction
We have recently seen a revival of attention given to convolutional neural networks (CNNs) [22]
due to their high performance for large-scale visual recognition tasks [15, 21, 30]. The architecture
of CNNs is relatively simple and consists of successive layers organized in a hierarchical fashion;
each layer involves convolutions with learned filters followed by a pointwise non-linearity and a
downsampling operation called ?feature pooling?. The resulting image representation has been empirically observed to be invariant to image perturbations and to encode complex visual patterns [33],
which are useful properties for visual recognition. Training CNNs remains however difficult since
high-capacity networks may involve billions of parameters to learn, which requires both high computational power, e.g., GPUs, and appropriate regularization techniques [18, 21, 30].
The exact nature of invariance that CNNs exhibit is also not precisely understood. Only recently, the
invariance of related architectures has been characterized; this is the case for the wavelet scattering
transform [8] or the hierarchical models of [7]. Our work revisits convolutional neural networks,
but we adopt a significantly different approach than the traditional one. Indeed, we use kernels [26],
which are natural tools to model invariance [14]. Inspired by the hierarchical kernel descriptors
of [2], we propose a reproducing kernel that produces multi-layer image representations.
Our main contribution is an approximation scheme called convolutional kernel network (CKN) to
make the kernel approach computationally feasible. Our approach is a new type of unsupervised
convolutional neural network that is trained to approximate the kernel map. Interestingly, our network uses non-linear functions that resemble rectified linear units [1, 30], even though they were not
handcrafted and naturally emerge from an approximation scheme of the Gaussian kernel map.
By bridging a gap between kernel methods and neural networks, we believe that we are opening
a fruitful research direction for the future. Our network is learned without supervision since the
?
LEAR team, Inria Grenoble, Laboratoire Jean Kuntzmann, CNRS, Univ. Grenoble Alpes, France.
1
label information is only used subsequently in a support vector machine (SVM). Yet, we achieve
competitive results on several datasets such as MNIST [22], CIFAR-10 [20] and STL-10 [13] with
simple architectures, few parameters to learn, and no data augmentation. Open-source code for
learning our convolutional kernel networks is available on the first author?s webpage.
1.1
Related Work
There have been several attempts to build kernel-based methods that mimic deep neural networks;
we only review here the ones that are most related to our approach.
Arc-cosine kernels. Kernels for building deep large-margin classifiers have been introduced
in [10]. The multilayer arc-cosine kernel is built by successive kernel compositions, and each layer
relies on an integral representation. Similarly, our kernels rely on an integral representation, and
enjoy a multilayer construction. However, in contrast to arc-cosine kernels: (i) we build our sequence of kernels by convolutions, using local information over spatial neighborhoods (as opposed
to compositions, using global information); (ii) we propose a new training procedure for learning a
compact representation of the kernel in a data-dependent manner.
Multilayer derived kernels. Kernels with invariance properties for visual recognition have been
proposed in [7]. Such kernels are built with a parameterized ?neural response? function, which consists in computing the maximal response of a base kernel over a local neighborhood. Multiple layers
are then built by iteratively renormalizing the response kernels and pooling using neural response
functions. Learning is performed by plugging the obtained kernel in an SVM. In contrast to [7], we
propagate information up, from lower to upper layers, by using sequences of convolutions. Furthermore, we propose a simple and effective data-dependent way to learn a compact representation of
our kernels and show that we obtain near state-of-the-art performance on several benchmarks.
Hierarchical kernel descriptors. The kernels proposed in [2, 3] produce multilayer image representations for visual recognition tasks. We discuss in details these kernels in the next section: our
paper generalizes them and establishes a strong link with convolutional neural networks.
2
Convolutional Multilayer Kernels
The convolutional multilayer kernel is a generalization of the hierarchical kernel descriptors introduced in computer vision [2, 3]. The kernel produces a sequence of image representations that are
built on top of each other in a multilayer fashion. Each layer can be interpreted as a non-linear transformation of the previous one with additional spatial invariance. We call these layers image feature
maps1 , and formally define them as follows:
Definition 1. An image feature map ? is a function ? : ? ? H, where ? is a (usually discrete)
subset of [0, 1]d representing normalized ?coordinates? in the image and H is a Hilbert space.
For all practical examples in this paper, ? is a two-dimensional grid and corresponds to different
locations in a two-dimensional image. In other words, ? is a set of pixel coordinates. Given z
in ?, the point ?(z) represents some characteristics of the image at location z, or in a neighborhood
of z. For instance, a color image of size m ? n with three channels, red, green, and blue, may be
represented by an initial feature map ?0 : ?0 ? H0 , where ?0 is an m ? n regular grid, H0 is the
Euclidean space R3 , and ?0 provides the color pixel values. With the multilayer scheme, non-trivial
feature maps will be obtained subsequently, which will encode more complex image characteristics.
With this terminology in hand, we now introduce the convolutional kernel, first, for a single layer.
Definition 2 (Convolutional Kernel with Single Layer). Let us consider two images represented
by two image feature maps, respectively ? and ?? : ? ? H, where ? is a set of pixel locations,
and H is a Hilbert space. The one-layer convolutional kernel between ? and ?? is defined as
2
XX
2
? 1 z?z? k ? 12 k?(z)?
?
?
?? (z? )k
2 e 2?
H,
K(?, ?? ) :=
k?(z)kH k?? (z? )kH e 2?2 k
(1)
z?? z? ??
1
In the kernel literature, ?feature map? denotes the mapping between data points and their representation in
a reproducing kernel Hilbert space (RKHS) [26]. Here, feature maps refer to spatial maps representing local
image characteristics at everly location, as usual in the neural network literature [22].
2
where ? and ? are smoothing parameters of Gaussian kernels, and ?(z)
?
:= (1/ k?(z)kH ) ?(z)
? ?
if ?(z) 6= 0 and ?(z)
?
= 0 otherwise. Similarly, ?? (z ) is a normalized version of ?? (z? ).2
It is easy to show that the kernel K is positive definite (see Appendix A). It consists of a sum of
pairwise comparisons between the image features ?(z) and ?? (z? ) computed at all spatial locations z
and z? in ?. To be significant in the sum, a comparison needs the corresponding z and z? to be
close in ?, and the normalized features ?(z)
?
and ??? (z? ) to be close in the feature space H. The
parameters ? and ? respectively control these two definitions of ?closeness?. Indeed, when ? is
large, the kernel K is invariant to the positions z and z? but when ? is small, only features placed
at the same location z = z? are compared to each other. Therefore, the role of ? is to control how
much the kernel is locally shift-invariant. Next, we will show how to go beyond one single layer,
but before that, we present concrete examples of simple input feature maps ?0 : ?0 ? H0 .
Gradient map. Assume that H0 = R2 and that ?0 (z) provides the two-dimensional gradient of the
image at pixel z, which is often computed with first-order differences along each dimension. Then,
the quantity k?0 (z)kH0 is the gradient intensity, and ??0 (z) is its orientation, which can be characterized by a particular angle?that is, there exists ? in [0; 2?] such that ??0 (z) = [cos(?), sin(?)]. The
resulting kernel K is exactly the kernel descriptor introduced in [2, 3] for natural image patches.
Patch map. In that setting, ?0 associates to a location z an image patch of size m ? m centered
at z. Then, the space H0 is simply Rm?m , and ??0 (z) is a contrast-normalized version of the patch,
which is a useful transformation for visual recognition according to classical findings in computer
vision [19]. When the image is encoded with three color channels, patches are of size m ? m ? 3.
We now define the multilayer convolutional kernel, generalizing some ideas of [2].
Definition 3 (Multilayer Convolutional Kernel). Let us consider a set ?k?1 ? [0, 1]d and a Hilbert
space Hk?1 . We build a new set ?k and a new Hilbert space Hk as follows:
(i) choose a patch shape Pk defined as a bounded symmetric subset of [?1, 1]d , and a set of coordinates ?k such that for all location zk in ?k , the patch {zk } + Pk is a subset of ?k?1 ;3 In other
words, each coordinate zk in ?k corresponds to a valid patch in ?k?1 centered at zk .
(ii) define the convolutional kernel Kk on the ?patch? feature maps Pk ? Hk?1 , by replacing
in (1): ? by Pk , H by Hk?1 , and ?, ? by appropriate smoothing parameters ?k , ?k . We denote
by Hk the Hilbert space for which the positive definite kernel Kk is reproducing.
An image represented by a feature map ?k?1 : ?k?1 ? Hk?1 at layer k?1 is now encoded in the k-th
layer as ?k : ?k ? Hk , where for all zk in ?k , ?k (zk ) is the representation in Hk of the patch
feature map z 7? ?k?1 (zk + z) for z in Pk .
Concretely, the kernel Kk between two patches of ?k?1 and ??k?1 at respective locations zk and z?k is
2
2
X X
? 12 kz?z? k ? 12 k?
?k?1 (zk +z)??
??k?1 (z?k +z? )k
2
k?k?1 (zk + z)k k??k?1 (z?k + z? )k e 2?k
, (2)
e 2?k
z?Pk z? ?Pk
where k.k is the Hilbertian norm of Hk?1 . In Figure 1(a), we illustrate the interactions between the
sets of coordinates ?k , patches Pk , and feature spaces Hk across layers. For two-dimensional grids,
a typical patch shape is a square, for example P := {?1/n, 0, 1/n} ? {?1/n, 0, 1/n} for a 3 ? 3
patch in an image of size n ? n. Information encoded in the k-th layer differs from the (k?1)-th one
in two aspects: first, each point ?k (zk ) in layer k contains information about several points from
the (k?1)-th layer and can possibly represent larger patterns; second, the new feature map is more
locally shift-invariant than the previous one due to the term involving the parameter ?k in (2).
The multilayer convolutional kernel slightly differs from the hierarchical kernel descriptors of [2]
but exploits similar ideas. Bo et al. [2] define indeed several ad hoc kernels for representing local
information in images, such as gradient, color, or shape. These kernels are close to the one defined
in (1) but with a few variations. Some of them do not use normalized features ?(z),
?
and these kernels
use different weighting strategies for the summands of (1) that are specialized to the image modality,
e.g., color, or gradient, whereas we use the same weight k?(z)kH k?? (z? )kH for all kernels. The
generic formulation (1) that we propose may be useful per se, but our main contribution comes in
the next section, where we use the kernel as a new tool for learning convolutional neural networks.
R
P
When ? is not discrete, the notation
in (1) should be replaced by the Lebesgue integral in the paper.
3
For two sets A and B, the Minkowski sum A + B is defined as {a + b : a ? A, b ? B}.
2
3
?2 (z2 ) ? H2
?k (z)
?2
??k
{z2 } + P2
?1
?k?1
?1 (z1 ) ? H1
Gaussian filtering
+ downsampling
= pooling
?k (zk?1 )
pk
convolution
+ non-linearity
?
{zk?1 }+Pk?1
?0 (z0 ) ? H0
{z1 } + P1
?0
??k?1
?k?1 (z)
?k?1 (zk?1 )
(patch extraction)
(b) Zoom between layer k?1 and k of the CKN.
(a) Hierarchy of image feature maps.
Figure 1: Left: concrete representation of the successive layers for the multilayer convolutional
kernel. Right: one layer of the convolutional neural network that approximates the kernel.
3
Training Invariant Convolutional Kernel Networks
Generic schemes have been proposed for approximating a non-linear kernel with a linear one, such
as the Nystr?om method and its variants [5, 31], or random sampling techniques in the Fourier domain for shift-invariant kernels [24]. In the context of convolutional multilayer kernels, such an
approximation is critical because computing the full kernel matrix on a database of images is computationally infeasible, even for a moderate number of images (? 10 000) and moderate number of
layers. For this reason, Bo et al. [2] use the Nystr?om method for their hierarchical kernel descriptors.
In this section, we show that when the coordinate sets ?k are two-dimensional regular grids, a
natural approximation for the multilayer convolutional kernel consists of a sequence of spatial convolutions with learned filters, pointwise non-linearities, and pooling operations, as illustrated in
Figure 1(b). More precisely, our scheme approximates the kernel map of K defined in (1) at layer k
by finite-dimensional spatial maps ?k : ??k ? Rpk , where ??k is a set of coordinates related to ?k ,
and pk is a positive integer controlling the quality of the approximation. Consider indeed two images
represented at layer k by image feature maps ?k and ??k , respectively. Then,
(A) the corresponding maps ?k and ?k? are learned such that K(?k?1 , ??k?1 ) ? h?k , ?k? i, where h., .i
?
is the Euclidean inner-product acting as if ?k and ?k? were vectors in R|?k |pk ;
(B) the set ??k is linked to ?k by the relation ??k = ?k + Pk? where Pk? is a patch shape, and
?
the quantities ?k (zk ) in Hk admit finite-dimensional approximations ?k (zk ) in R|Pk |pk ; as
illustrated in Figure 1(b), ?k (zk ) is a patch from ?k centered at location zk with shape Pk? ;
(C) an activation map ?k : ?k?1 7? Rpk is computed from ?k?1 by convolution with pk filters
followed by a non-linearity. The subsequent map ?k is obtained from ?k by a pooling operation.
We call this approximation scheme a convolutional kernel network (CKN). In comparison to CNNs,
our approach enjoys similar benefits such as efficient prediction at test time, and involves the same
set of hyper-parameters: number of layers, numbers of filters pk at layer k, shape Pk? of the filters,
sizes of the feature maps. The other parameters ?k , ?k can be automatically chosen, as discussed
later. Training a CKN can be argued to be as simple as training a CNN in an unsupervised manner [25] since we will show that the main difference is in the cost function that is optimized.
3.1
Fast Approximation of the Gaussian Kernel
A key component of our formulation is the Gaussian kernel. We start by approximating it by a linear
operation with learned filters followed by a pointwise non-linearity. Our starting point is the next
lemma, which can be obtained after a simple calculation.
4
Lemma 1 (Linear expansion of the Gaussian Kernel). For all x and x? in Rm , and ? > 0,
m2 Z
2
?
2
1
1
2
? 2?12 kx?x? k22
=
(3)
e? ?2 kx?wk2 e? ?2 kx ?wk2 dw.
e
2
??
w?Rm
?
2
2
The lemma gives us a mapping of any x in Rm to the function w 7? Ce?(1/? )kx?wk2 in L2 (Rm ),
where the kernel is linear, and C is the constant in front of the integral. To obtain a finite-dimensional
representation, we need to approximate the integral with a weighted finite sum, which is a classical
problem arising in statistics (see [29] and chapter 8 of [6]). Then, we consider two different cases.
Small dimension, m ? 2. When the data lives in a compact set of Rm , the integral in (3) can be
approximated by uniform sampling over a large enough
set. We choose such a strategy for two types
2
2
? 2?12 kz?z? k
? 1
?(z)?
?
?
?? (z? )k
2 ; (ii) the terms e ( 2? 2 )k
H
of kernels from Eq. (1): (i) the spatial kernels e
when ? is the ?gradient map? presented in Section 2. In the latter case, H = R2 and ?(z)
?
is the
gradient orientation. We typically sample a few orientations as explained in Section 4.
Higher dimensions. To prevent the curse of dimensionality, we learn to approximate the kernel on
training data, which is intrinsically low-dimensional. We optimize importance weights ? = [?l ]pl=1
in Rp+ and sampling points W = [wl ]pl=1 in Rm?p on n training pairs (xi , yi )i=1,...,n in Rm ? Rm :
X
p
n
2
X
1
? ?12 kxi ?wl k22 ? ?12 kyi ?wl k22
? 2?12 kxi ?yi k22
.
(4)
?
e
?
e
e
min
l
m?p
n i=1
??Rp
+ ,W?R
l=1
Interestingly, we may already draw some links with neural networks. When applied to unit-norm
vectors xi and yi , problem (4) produces sampling points wl whose norm is close to one. After
2
2
?
learning, a new unit-norm point x in Rm is mapped to the vector [ ?l e?(1/? )kx?wl k2 ]pl=1 in Rp ,
which may be written as [f (wl? x)]pl=1 , assuming that the norm of wl is always one, where f is the
2
function u 7? e(2/? )(u?1) for u = wl? x in [?1, 1]. Therefore, the finite-dimensional representation
of x only involves a linear operation followed by a non-linearity, as in typical neural networks. In
Figure 2, we show that the shape of f resembles the ?rectified linear unit? function [30].
2
f (u)
-1
0
f (u) = e(2/? )(u?1)
f (u) = max(u, 0)
1
u
Figure 2: In dotted red, we plot the ?rectified linear unit? function u 7? max(u, 0). In blue, we plot
non-linear functions of our network for typical values of ? that we use in our experiments.
3.2
Approximating the Multilayer Convolutional Kernel
We have now all the tools in hand to build our convolutional kernel network. We start by making assumptions on the input data, and then present the learning scheme and its approximation principles.
The zeroth layer. We assume that the input data is a finite-dimensional map ?0 : ??0 ? Rp0 , and
that ?0 : ?0 ? H0 ?extracts? patches from ?0 . Formally, there exists a patch shape P0? such that
?
??0 = ?0 + P0? , H0 = Rp0 |P0 | , and for all z0 in ?0 , ?0 (z0 ) is a patch of ?0 centered at z0 . Then,
property (B) described at the beginning of Section 3 is satisfied for k = 0 by choosing ?0 = ?0 .
The examples of input feature maps given earlier satisfy this finite-dimensional assumption: for the
gradient map, ?0 is the gradient of the image along each direction, with p0 = 2, P0? = {0} is a 1?1
patch, ?0 = ??0 , and ?0 = ?0 ; for the patch map, ?0 is the input image, say with p0 = 3 for RGB data.
The convolutional kernel network. The zeroth layer being characterized, we present in Algorithms 1 and 2 the subsequent layers and how to learn their parameters in a feedforward manner. It
is interesting to note that the input parameters of the algorithm are exactly the same as a CNN?that
is, number of layers and filters, sizes of the patches and feature maps (obtained here via the subsampling factor). Ultimately, CNNs and CKNs only differ in the cost function that is optimized for
learning the filters and in the choice of non-linearities. As we show next, there exists a link between
the parameters of a CKN and those of a convolutional multilayer kernel.
5
Algorithm 1 Convolutional kernel network - learning the parameters of the k-th layer.
1
2
input ?k?1
, ?k?1
, . . . : ??k?1 ? Rpk?1 (sequence of (k?1)-th maps obtained from training images);
?
Pk?1 (patch shape); pk (number of filters); n (number of training pairs);
?
1
2
1: extract at random n pairs (xi , yi ) of patches with shape Pk?1
from the maps ?k?1
, ?k?1
, . . .;
2: if not provided by the user, set ?k to the 0.1 quantile of the data (kxi ? yi k2 )n
;
i=1
?
3: unsupervised learning: optimize (4) to obtain the filters Wk in R|Pk?1 |pk?1 ?pk and ? k in Rpk ;
output Wk , ? k , and ?k (smoothing parameter);
Algorithm 2 Convolutional kernel network - computing the k-th map form the (k?1)-th one.
?
input ?k?1 : ??k?1 ? Rpk?1 (input map); Pk?1
(patch shape); ?k ? 1 (subsampling factor); pk (numk
k
ber of filters); ?k (smoothing parameter); Wk = [wkl ]pl=1
and ? k = [?kl ]pl=1
(layer parameters);
1: convolution and non-linearity: define the activation map ?k : ?k?1 ? Rpk as
?k?1 (z)?wkl k2 pk
? 12 k?
?
?
2
,
(5)
?k : z 7? k?k?1 (z)k2
?kl e k
l=1
?
where ?k?1 (z) is a vector representing a patch from ?k?1 centered at z with shape Pk?1
, and the
?
vector ?k?1 (z) is an ?2 -normalized version of ?k?1 (z). This operation can be interpreted as a
spatial convolution of the map ?k?1 with the filters wkl followed by pointwise non-linearities;
2: set ?k to be ?k times the spacing between two pixels in ?k?1 ;
3: feature pooling: ??k is obtained by subsampling ?k?1 by a factor ?k and we define a new map
?k : ??k ? Rpk obtained from ?k by linear pooling with Gaussian weights:
?k : z 7?
X ? 12 ku?zk22
p
e ?k
2/?
?k (u).
(6)
u??k?1
output ?k : ??k ? Rpk (new map);
Approximation principles. We proceed recursively to show that the kernel approximation property (A) is satisfied; we assume that (B) holds at layer k?1, and then, we show that (A) and (B) also
hold at layer k. This is sufficient for our purpose since we have previously assumed (B) for the zeroth layer. Given two images feature maps ?k?1 and ??k?1 , we start by approximating K(?k?1 , ??k?1 )
by replacing ?k?1 (z) and ??k?1 (z? ) by their finite-dimensional approximations provided by (B):
2
X
?k?1 (z)??
?? (z? )k2
? 12 kz?z? k ? 12 k?
k?1
?
2
2
e 2?k
. (7)
k?k?1 (z)k2 k?k?1
(z? )k2 e 2?k
K(?k?1 , ??k?1 ) ?
z,z? ??k?1
Then, we use the finite-dimensional approximation of the Gaussian kernel involving ?k and
2
X
? 12 kz?z? k
2
?k (z)? ?k? (z? )e 2?k
,
K(?k?1 , ??k?1 ) ?
(8)
z,z? ??k?1
where ?k is defined in (5) and ?k? is defined similarly by replacing ?? by ??? . Finally, we approximate
the remaining Gaussian kernel by uniform sampling on ??k , following Section 3.1. After exchanging
sums and grouping appropriate terms together, we obtain the new approximation
? X
X
2
? 12 kz? ?uk ?
? 12 kz?uk22
2 X
?
?
?
?
2
k
k
K(?k?1 , ?k?1 ) ?
e
?k (z)
?k (z ) ,
(9)
e
?
?
?
u??k
z ??k?1
z??k?1
where the constant 2/? comes from the multiplication of the constant 2/(??k2 ) from (3) and the
weight ?k2 of uniform sampling orresponding to the square of the distance between two pixels of ??k .4
As a result, the right-hand side is exactly h?k , ?k? i, where ?k is defined in (6), giving us property (A).
It remains to show that property (B) also holds, specifically that the quantity (2) can be approximated
by the Euclidean inner-product h?k (zk ), ?k? (z?k )i with the patches ?k (zk ) and ?k? (z?k ) of shape Pk? ;
we assume for that purpose that Pk? is a subsampled version of the patch shape Pk by a factor ?k .
4
The choice of ?k in Algorithm 2 is driven by signal processing principles. The feature pooling step can
indeed be interpreted as a downsampling operation that reduces the resolution of the map from ?k?1 to ?k by
using a Gaussian anti-aliasing filter, whose role is to reduce frequencies above the Nyquist limit.
6
We remark that the kernel (2) is the same as (1) applied to layer k?1 by replacing ?k?1 by {zk }+Pk .
By doing the same substitution in (9), we immediately obtain an approximation of (2). Then, all
Gaussian terms are negligible forPall u and
each other?say
Pz that are far fromP
P when ku?zk2 ? 2?k .
Thus, we may replace the sums u??? z,z? ?{zk }+Pk by u?{zk }+P ? z,z? ??k?1 , which has the
k
k
same set of ?non-negligible? terms. This yields exactly the approximation h?k (zk ), ?k? (z?k )i.
Optimization. Regarding problem (4), stochastic gradient descent (SGD) may be used since a
potentially infinite amount of training data is available. However, we have preferred to use L-BFGSB [9] on 300 000 pairs of randomly selected training data points, and initialize W with the K-means
algorithm. L-BFGS-B is a parameter-free state-of-the-art batch method, which is not as fast as SGD
but much easier to use. We always run the L-BFGS-B algorithm for 4 000 iterations, which seems
to ensure convergence to a stationary point. Our goal is to demonstrate the preliminary performance
of a new type of convolutional network, and we leave as future work any speed improvement.
4
Experiments
We now present experiments that were performed using Matlab and an L-BFGS-B solver [9] interfaced by Stephen Becker. Each image is represented by the last map ? k of the CKN, which is used
in a linear SVM implemented in the software package LibLinear [16]. These representations are
centered, rescaled to have unit ?2 -norm on average, and the regularization parameter of the SVM is
always selected on a validation set or by 5-fold cross-validation in the range 2i , i = ?15 . . . , 15.
The patches Pk? are typically small; we tried the sizes m ? m with m = 3, 4, 5 for the first
layer, and m = 2, 3 for the upper ones. The number of filters pk in our experiments is in the
set {50, 100, 200, 400, 800}. The downsampling factor ?k is always chosen to be 2 between two consecutive layers, whereas the last layer is downsampled to produce final maps ?k of a small size?say,
2
?
?
5?5 or 4?4. For the gradient map ?0 , we approximate the Gaussian kernel e(1/?1 )k?0 (z)??0 (z )kH0
by uniformly sampling p1 = 12 orientations, setting ?1 = 2?/p1 . Finally, we also use a small off?
set ? to prevent numerical instabilities in the normalization steps ?(z)
= ?(z)/ max(k?(z)k2 , ?).
4.1
Discovering the Structure of Natural Image Patches
Unsupervised learning was first used for discovering the underlying structure of natural image
patches by Olshausen and Field [23]. Without making any a priori assumption about the data except a parsimony principle, the method is able to produce small prototypes that resemble Gabor
wavelets?that is, spatially localized oriented basis functions. The results were found impressive by
the scientific community and their work received substantial attention. It is also known that such
results can also be achieved with CNNs [25]. We show in this section that this is also the case for
convolutional kernel networks, even though they are not explicitly trained to reconstruct data.
Following [23], we randomly select a database of 300 000 whitened natural image patches of
size 12 ? 12 and learn p = 256 filters W using the formulation (4). We initialize W with Gaussian
random noise without performing the K-means step, in order to ensure that the output we obtain is
not an artifact of the initialization. In Figure 3, we display the filters associated to the top-128 largest
weights ?l . Among the 256 filters, 197 exhibit interpretable Gabor-like structures and the rest was
less interpretable. To the best of our knowledge, this is the first time that the explicit kernel map of
the Gaussian kernel for whitened natural image patches is shown to be related to Gabor wavelets.
4.2
Digit Classification on MNIST
The MNIST dataset [22] consists of 60 000 images of handwritten digits for training and 10 000
for testing. We use two types of initial maps in our networks: the ?patch map?, denoted by CNKPM and the ?gradient map?, denoted by CNK-GM. We follow the evaluation methodology of [25]
Figure 3: Filters obtained by the first layer of the convolutional kernel network on natural images.
7
Tr.
size
300
1K
2K
5K
10K
20K
40K
60K
CNN
[25]
7.18
3.21
2.53
1.52
0.85
0.76
0.65
0.53
Scat-1
[8]
4.7
2.3
1.3
1.03
0.88
0.79
0.74
0.70
Scat-2
[8]
5.6
2.6
1.8
1.4
1
0.58
0.53
0.4
CKN-GM1
(12/50)
4.39
2.60
1.85
1.41
1.17
0.89
0.68
0.58
CKN-GM2
(12/400)
4.24
2.05
1.51
1.21
0.88
0.60
0.51
0.39
CKN-PM1
(200)
5.98
3.23
1.97
1.41
1.18
0.83
0.64
0.63
CKN-PM2
(50/200)
4.15
2.76
2.28
1.56
1.10
0.77
0.58
0.53
[32]
[18]
[19]
0.47
NA
NA
NA
NA
NA
NA
NA
0.45
0.53
Table 1: Test error in % for various approaches on the MNIST dataset without data augmentation.
The numbers in parentheses represent the size p1 and p2 of the feature maps at each layer.
for comparison when varying the training set size. We select the regularization parameter of the
SVM by 5-fold cross validation when the training size is smaller than 20 000, or otherwise, we
keep 10 0000 examples from the training set for validation. We report in Table 1 the results obtained
for four simple architectures. CKN-GM1 is the simplest one: its second layer uses 3 ? 3 patches and
only p2 = 50 filters, resulting in a network with 5 400 parameters. Yet, it achieves an outstanding
performance of 0.58% error on the full dataset. The best performing, CKN-GM2, is similar to
CKN-GM1 but uses p2 = 400 filters. When working with raw patches, two layers (CKN-PM2)
gives better results than one layer. More details about the network architectures are provided in the
supplementary material. In general, our method achieves a state-of-the-art accuracy for this task
since lower error rates have only been reported by using data augmentation [11].
4.3
Visual Recognition on CIFAR-10 and STL-10
We now move to the more challenging datasets CIFAR-10 [20] and STL-10 [13]. We select the
best architectures on a validation set of 10 000 examples from the training set for CIFAR-10, and
by 5-fold cross-validation on STL-10. We report in Table 2 results for CKN-GM, defined in the
previous section, without exploiting color information, and CKN-PM when working on raw RGB
patches whose mean color is subtracted. The best selected models have always two layers, with 800
filters for the top layer. Since CKN-PM and CKN-GM exploit a different information, we also report
a combination of such two models, CKN-CO, by concatenating normalized image representations
together. The standard deviations for STL-10 was always below 0.7%. Our approach appears to
be competitive with the state of the art, especially on STL-10 where only one method does better
than ours, despite the fact that our models only use 2 layers and require learning few parameters.
Note that better results than those reported in Table 2 have been obtained in the literature by using
either data augmentation (around 90% on CIFAR-10 for [18, 30]), or external data (around 70% on
STL-10 for [28]). We are planning to investigate similar data manipulations in the future.
Method
CIFAR-10
STL-10
[12]
82.0
60.1
[27]
82.2
58.7
[18]
88.32
NA
[13]
79.6
51.5
[4]
NA
64.5
[17]
83.96
62.3
[32]
84.87
NA
CKN-GM
74.84
60.04
CKN-PM
78.30
60.25
CKN-CO
82.18
62.32
Table 2: Classification accuracy in % on CIFAR-10 and STL-10 without data augmentation.
5
Conclusion
In this paper, we have proposed a new methodology for combining kernels and convolutional neural
networks. We show that mixing the ideas of these two concepts is fruitful, since we achieve near
state-of-the-art performance on several datasets such as MNIST, CIFAR-10, and STL10, with simple
architectures and no data augmentation. Some challenges regarding our work are left open for the
future. The first one is the use of supervision to better approximate the kernel for the prediction task.
The second consists in leveraging the kernel interpretation of our convolutional neural networks to
better understand the theoretical properties of the feature spaces that these networks produce.
Acknowledgments
This work was partially supported by grants from ANR (project MACARON ANR-14-CE23-000301), MSR-Inria joint centre, European Research Council (project ALLEGRO), CNRS-Mastodons
program (project GARGANTUA), and the LabEx PERSYVAL-Lab (ANR-11-LABX-0025).
8
References
[1] Y. Bengio. Learning deep architectures for AI. Found. Trends Mach. Learn., 2009.
[2] L. Bo, K. Lai, X. Ren, and D. Fox. Object recognition with hierarchical kernel descriptors. In Proc.
CVPR, 2011.
[3] L. Bo, X. Ren, and D. Fox. Kernel descriptors for visual recognition. In Adv. NIPS, 2010.
[4] L. Bo, X. Ren, and D. Fox. Unsupervised feature learning for RGB-D based object recognition. In
Experimental Robotics, 2013.
[5] L. Bo and C. Sminchisescu. Efficient match kernel between sets of features for visual recognition. In Adv.
NIPS, 2009.
[6] L. Bottou, O. Chapelle, D. DeCoste, and J. Weston. Large-Scale Kernel Machines (Neural Information
Processing). The MIT Press, 2007.
[7] J. V. Bouvrie, L. Rosasco, and T. Poggio. On invariance in hierarchical models. In Adv. NIPS, 2009.
[8] J. Bruna and S. Mallat. Invariant scattering convolution networks. IEEE T. Pattern Anal., 35(8):1872?
1886, 2013.
[9] R. H. Byrd, P. Lu, J. Nocedal, and C. Zhu. A limited memory algorithm for bound constrained optimization. SIAM J. Sci. Comput., 16(5):1190?1208, 1995.
[10] Y. Cho and L. K. Saul. Large-margin classification in infinite neural networks. Neural Comput., 22(10),
2010.
[11] D. Ciresan, U. Meier, and J. Schmidhuber. Multi-column deep neural networks for image classification.
In Proc. CVPR, 2012.
[12] A. Coates and A. Y. Ng. Selecting receptive fields in deep networks. In Adv. NIPS, 2011.
[13] A. Coates, A. Y. Ng, and H. Lee. An analysis of single-layer networks in unsupervised feature learning.
In Proc. AISTATS, 2011.
[14] D. Decoste and B. Sch?olkopf. Training invariant support vector machines. Mach. Learn., 46(1-3):161?
190, 2002.
[15] J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng, and T. Darrell. DeCAF: A deep convolutional activation feature for generic visual recognition. preprint arXiv:1310.1531, 2013.
[16] R.-E. Fan, K.-W. Chang, C.-J. Hsieh, X.-R. Wang, and C.-J. Lin. LIBLINEAR: A library for large linear
classification. J. Mach. Learn. Res., 9:1871?1874, 2008.
[17] R. Gens and P. Domingos. Discriminative learning of sum-product networks. In Adv. NIPS, 2012.
[18] I. J. Goodfellow, D. Warde-Farley, M. Mirza, A. Courville, and Y. Bengio. Maxout networks. In Proc.
ICML, 2013.
[19] K. Jarrett, K. Kavukcuoglu, M. Ranzato, and Y. LeCun. What is the best multi-stage architecture for
object recognition? In Proc. ICCV, 2009.
[20] A. Krizhevsky and G. Hinton. Learning multiple layers of features from tiny images. Tech. Rep., 2009.
[21] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural
networks. In Adv. NIPS, 2012.
[22] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition.
P. IEEE, 86(11):2278?2324, 1998.
[23] B. A. Olshausen and D. J. Field. Emergence of simple-cell receptive field properties by learning a sparse
code for natural images. Nature, 381(6583):607?609, 1996.
[24] A. Rahimi and B. Recht. Random features for large-scale kernel machines. In Adv. NIPS, 2007.
[25] M. Ranzato, F.-J. Huang, Y-L. Boureau, and Y. LeCun. Unsupervised learning of invariant feature hierarchies with applications to object recognition. In Proc. CVPR, 2007.
[26] J. Shawe-Taylor and N. Cristianini. Kernel methods for pattern analysis. 2004.
[27] K. Sohn and H. Lee. Learning invariant representations with local transformations. In Proc. ICML, 2012.
[28] K. Swersky, J. Snoek, and R. P. Adams. Multi-task Bayesian optimization. In Adv. NIPS, 2013.
[29] G. Wahba. Spline models for observational data. SIAM, 1990.
[30] L. Wan, M. D. Zeiler, S. Zhang, Y. LeCun, and R. Fergus. Regularization of neural networks using
dropconnect. In Proc. ICML, 2013.
[31] C. Williams and M. Seeger. Using the Nystr?om method to speed up kernel machines. In Adv. NIPS, 2001.
[32] M. D. Zeiler and R. Fergus. Stochastic pooling for regularization of deep convolutional neural networks.
In Proc. ICLR, 2013.
[33] M. D. Zeiler and R. Fergus. Visualizing and understanding convolutional networks. In Proc. ECCV, 2014.
9
| 5348 |@word msr:1 cnn:4 version:4 norm:6 seems:1 open:2 propagate:1 rgb:3 tried:1 p0:6 hsieh:1 sgd:2 nystr:3 tr:1 recursively:1 liblinear:2 initial:2 substitution:1 contains:1 selecting:1 rkhs:1 interestingly:2 ours:1 document:1 z2:2 activation:3 yet:2 written:1 subsequent:2 ckns:1 numerical:1 shape:14 zaid:1 plot:2 interpretable:2 stationary:1 selected:3 discovering:2 beginning:1 provides:2 location:10 successive:3 zhang:2 along:2 consists:6 introduce:1 manner:3 pairwise:1 snoek:1 indeed:5 p1:4 planning:1 multi:4 aliasing:1 inspired:1 automatically:1 byrd:1 curse:1 decoste:2 solver:1 provided:3 xx:1 linearity:9 project:3 bounded:1 notation:1 underlying:1 what:1 interpreted:3 parsimony:1 finding:1 transformation:4 gm1:3 exactly:4 classifier:1 rm:10 k2:10 control:2 unit:6 uk:1 enjoy:1 grant:1 positive:3 before:1 understood:1 local:5 negligible:2 limit:1 despite:1 mach:3 inria:4 zeroth:3 initialization:1 resembles:1 wk2:3 challenging:2 co:3 limited:1 range:1 jarrett:1 practical:1 acknowledgment:1 lecun:4 testing:1 definite:2 differs:2 digit:3 procedure:1 significantly:1 gabor:3 word:2 regular:2 downsampled:1 close:4 context:1 instability:1 optimize:2 fruitful:2 map:49 go:1 attention:2 starting:1 williams:1 resolution:1 alpes:1 immediately:1 m2:1 dw:1 coordinate:7 variation:1 construction:1 hierarchy:2 controlling:1 user:1 exact:1 gm:4 mallat:1 us:3 domingo:1 goodfellow:1 associate:1 trend:1 recognition:17 approximated:2 database:2 observed:1 role:2 preprint:1 wang:1 rpk:8 revival:1 adv:9 ranzato:2 rescaled:1 substantial:1 warde:1 cristianini:1 ultimately:1 trained:2 solving:1 basis:1 joint:1 represented:5 chapter:1 various:1 train:1 univ:1 fast:2 effective:1 hyper:1 neighborhood:3 h0:8 choosing:1 whose:4 encoded:4 jean:1 larger:1 supplementary:1 say:3 cvpr:3 otherwise:2 reconstruct:1 anr:3 statistic:1 transform:1 emergence:1 final:1 hoc:1 sequence:5 propose:4 interaction:1 maximal:1 product:3 fr:1 combining:1 gen:1 mixing:1 achieve:3 cnk:1 stl10:1 kh:5 olkopf:1 billion:1 webpage:1 convergence:1 exploiting:1 darrell:1 sutskever:1 produce:7 renormalizing:1 adam:1 leave:1 object:4 illustrate:1 received:1 eq:1 p2:4 strong:1 implemented:1 involves:3 resemble:2 come:2 differ:1 direction:2 cnns:9 filter:21 subsequently:2 centered:6 stochastic:2 observational:1 material:1 argued:1 require:1 generalization:1 preliminary:1 pl:6 hold:3 around:2 mapping:2 achieves:2 adopt:1 consecutive:1 purpose:2 proc:10 label:1 bridge:1 council:1 largest:1 wl:8 establishes:1 tool:4 weighted:1 hoffman:1 mit:1 gaussian:14 always:6 varying:1 encode:2 derived:1 improvement:1 hk:11 contrast:3 tech:1 seeger:1 zk22:1 dependent:2 cnrs:2 typically:2 relation:1 france:1 pixel:6 classification:7 orientation:4 among:1 denoted:2 hilbertian:1 priori:1 art:6 spatial:8 smoothing:4 initialize:2 tzeng:1 field:4 constrained:1 cordelia:1 piotr:1 extraction:1 sampling:7 ng:2 represents:1 unsupervised:7 icml:3 future:4 mimic:1 report:3 mirza:1 spline:1 opening:1 few:4 grenoble:2 randomly:2 oriented:1 zoom:1 subsampled:1 replaced:1 lebesgue:1 attempt:1 investigate:1 evaluation:1 farley:1 integral:6 pm2:2 poggio:1 respective:1 fox:3 euclidean:3 taylor:1 re:1 theoretical:1 instance:1 column:1 earlier:1 exchanging:1 cost:2 deviation:1 subset:3 uniform:3 krizhevsky:2 gargantua:1 front:1 reported:2 kxi:3 cho:1 recht:1 siam:2 lee:2 off:1 together:2 concrete:2 na:10 augmentation:6 satisfied:2 opposed:1 choose:2 possibly:1 rosasco:1 huang:1 wan:1 dropconnect:1 admit:1 external:1 bfgs:3 persyval:1 wk:3 satisfy:1 explicitly:1 ad:1 performed:2 h1:1 later:1 lab:1 linked:1 doing:1 red:2 competitive:3 start:3 jia:1 contribution:2 om:3 square:2 accuracy:4 convolutional:39 descriptor:8 characteristic:3 yield:1 interfaced:1 pm1:1 handwritten:1 raw:2 kavukcuoglu:1 bayesian:1 mastodon:1 ren:3 lu:1 rectified:3 definition:4 frequency:1 naturally:1 associated:1 dataset:4 intrinsically:1 color:7 knowledge:1 dimensionality:1 organized:1 hilbert:6 appears:1 scattering:2 higher:1 follow:1 methodology:3 response:4 formulation:3 though:2 furthermore:1 stage:1 hand:3 working:2 replacing:4 quality:1 artifact:1 scientific:1 believe:1 olshausen:2 building:1 k22:4 normalized:7 concept:1 regularization:5 spatially:1 symmetric:1 iteratively:1 illustrated:2 visualizing:1 sin:1 lastname:1 cosine:3 demonstrate:1 image:45 recently:2 specialized:1 empirically:1 handcrafted:1 discussed:1 interpretation:1 approximates:2 refer:1 composition:2 significant:1 ai:1 grid:4 pm:3 similarly:3 teaching:1 centre:1 shawe:1 chapelle:1 bruna:1 supervision:2 impressive:1 summands:1 base:1 moderate:2 driven:1 manipulation:1 schmidhuber:1 rep:1 life:1 yi:5 devise:1 seen:1 additional:1 signal:1 ii:3 stephen:1 multiple:2 harchaoui:1 full:2 reduces:1 rahimi:1 match:1 characterized:3 calculation:1 cross:3 cifar:9 lin:1 lai:1 gm2:2 plugging:1 parenthesis:1 prediction:2 involving:2 variant:1 multilayer:16 vision:2 whitened:2 arxiv:1 iteration:1 kernel:102 represent:3 normalization:1 achieved:1 robotics:1 cell:1 whereas:2 spacing:1 laboratoire:1 scat:2 source:1 modality:1 sch:1 rest:1 unlike:1 pooling:9 leveraging:1 call:2 integer:1 near:2 feedforward:1 bengio:3 easy:2 enough:1 architecture:10 ciresan:1 wahba:1 inner:2 idea:3 reduce:1 regarding:2 prototype:1 ce23:1 haffner:1 shift:3 bridging:1 becker:1 nyquist:1 proceed:1 remark:1 matlab:1 deep:8 useful:3 se:1 involve:1 amount:1 locally:2 sohn:1 simplest:1 coates:2 dotted:1 arising:1 per:1 blue:2 discrete:2 key:1 four:1 terminology:1 kyi:1 prevent:2 ce:1 nocedal:1 sum:7 run:1 angle:1 parameterized:1 package:1 swersky:1 patch:38 draw:1 appendix:1 layer:49 bound:1 followed:5 display:1 courville:1 fold:3 fan:1 precisely:2 software:1 aspect:1 fourier:1 speed:2 min:1 minkowski:1 performing:2 relatively:1 gpus:1 according:1 combination:1 across:1 slightly:1 smaller:1 rp0:2 making:2 explained:1 invariant:12 iccv:1 computationally:2 remains:2 previously:1 discus:1 r3:1 zk2:1 generalizes:1 operation:7 available:2 hierarchical:9 appropriate:3 generic:3 subtracted:1 batch:1 rp:3 top:3 denotes:1 subsampling:3 remaining:1 ensure:2 zeiler:3 exploit:2 kuntzmann:1 giving:1 quantile:1 build:4 especially:1 approximating:4 classical:3 move:1 already:1 quantity:3 strategy:2 receptive:2 usual:1 traditional:2 exhibit:2 gradient:13 iclr:1 distance:1 link:3 mapped:1 sci:1 capacity:1 uk22:1 trivial:1 reason:1 assuming:1 code:2 pointwise:4 kk:3 downsampling:4 difficult:1 potentially:1 bouvrie:1 anal:1 perform:1 upper:2 convolution:9 datasets:4 arc:3 benchmark:1 finite:9 descent:1 anti:1 hinton:2 team:1 wkl:3 perturbation:1 reproducing:4 community:1 intensity:1 introduced:3 pair:4 meier:1 kl:2 z1:2 optimized:2 imagenet:1 learned:6 nip:9 address:1 beyond:1 able:1 usually:1 pattern:4 firstname:1 below:1 challenge:1 program:1 built:4 green:1 max:3 memory:1 power:1 critical:1 natural:10 rely:1 fromp:1 zhu:1 representing:4 scheme:7 library:1 julien:1 extract:2 schmid:1 review:1 literature:4 l2:1 understanding:1 multiplication:1 interesting:1 filtering:1 proven:1 localized:1 validation:6 h2:1 labex:1 sufficient:1 ckn:22 principle:4 tiny:1 eccv:1 placed:1 last:2 free:1 supported:1 infeasible:1 enjoys:2 side:1 understand:1 ber:1 saul:1 emerge:1 sparse:1 benefit:2 dimension:3 valid:1 kz:6 author:1 concretely:1 far:1 approximate:7 compact:3 preferred:1 keep:1 global:1 overfitting:1 koniusz:1 mairal:1 assumed:1 xi:3 allegro:1 discriminative:1 fergus:3 table:5 learn:9 nature:2 robust:1 channel:2 zk:24 ku:2 sminchisescu:1 expansion:1 bottou:2 complex:3 european:1 domain:1 aistats:1 pk:37 main:3 revisits:1 noise:1 fashion:2 position:1 explicit:1 concatenating:1 comput:2 weighting:1 learns:1 wavelet:3 donahue:1 z0:4 r2:2 pz:1 svm:5 stl:10 closeness:1 kh0:2 exists:3 mnist:6 grouping:1 macaron:1 importance:1 decaf:1 margin:2 kx:5 gap:2 easier:1 boureau:1 generalizing:1 simply:1 visual:12 vinyals:1 partially:1 bo:6 chang:1 corresponds:2 relies:1 labx:1 weston:1 goal:3 lear:1 maxout:1 replace:1 feasible:1 typical:3 specifically:1 infinite:2 uniformly:1 acting:1 except:1 lemma:3 called:2 invariance:8 experimental:1 formally:2 select:3 support:2 latter:1 outstanding:1 evaluate:1 |
4,803 | 5,349 | Learning Deep Features for Scene Recognition
using Places Database
Bolei Zhou1 , Agata Lapedriza1,3 , Jianxiong Xiao2 , Antonio Torralba1 , and Aude Oliva1
1
Massachusetts Institute of Technology
2
Princeton University
3
Universitat Oberta de Catalunya
Abstract
Scene recognition is one of the hallmark tasks of computer vision, allowing definition of a context for object recognition. Whereas the tremendous recent progress
in object recognition tasks is due to the availability of large datasets like ImageNet
and the rise of Convolutional Neural Networks (CNNs) for learning high-level features, performance at scene recognition has not attained the same level of success.
This may be because current deep features trained from ImageNet are not competitive enough for such tasks. Here, we introduce a new scene-centric database called
Places with over 7 million labeled pictures of scenes. We propose new methods
to compare the density and diversity of image datasets and show that Places is as
dense as other scene datasets and has more diversity. Using CNN, we learn deep
features for scene recognition tasks, and establish new state-of-the-art results on
several scene-centric datasets. A visualization of the CNN layers? responses allows us to show differences in the internal representations of object-centric and
scene-centric networks.
1
Introduction
Understanding the world in a single glance is one of the most accomplished feats of the human brain:
it takes only a few tens of milliseconds to recognize the category of an object or environment, emphasizing an important role of feedforward processing in visual recognition. One of the mechanisms
subtending efficient human visual recognition is our capacity to learn and remember a diverse set of
places and exemplars [11]; by sampling the world several times per second, our neural architecture
constantly registers new inputs even for a very short time, reaching an exposure to millions of natural images within just a year. How much would an artificial system have to learn before reaching
the scene recognition abilities of a human being?
Besides the exposure to a dense and rich variety of natural images, one important property of the
primate brain is its hierarchical organization in layers of increasing processing complexity, an architecture that has inspired Convolutional Neural Networks or CNNs [2, 14]. These architectures
together with recent large databases (e.g., ImageNet [3]) have obtained astonishing performance on
object classification tasks [12, 5, 20]. However, the baseline performance reached by these networks
on scene classification tasks is within the range of performance based on hand-designed features
and sophisticated classifiers [24, 21, 4]. Here, we show that one of the reasons for this discrepancy
is that the higher-level features learned by object-centric versus scene-centric CNNs are different:
iconic images of objects do not contain the richness and diversity of visual information that pictures
of scenes and environments provide for learning to recognize them.
Here we introduce Places, a scene-centric image dataset 60 times larger than the SUN database [24].
With this database and a standard CNN architecture, we establish new baselines of accuracies on
1
various scene datasets (Scene15 [17, 13], MIT Indoor67 [19], SUN database [24], and SUN Attribute
Database [18]), significantly outperforming the results obtained by the deep features from the same
network architecture trained with ImageNet1 .
The paper is organized as follows: in Section 2 we introduce the Places database and describe the
collection procedure. In Section 3 we compare Places with the other two large image datasets:
SUN [24] and ImageNet [3]. We perform experiments on Amazon Mechanical Turk (AMT) to compare these 3 datasets in terms of density and diversity. In Section 4 we show new scene classification
performance when training deep features from millions of labeled scene images. Finally, we visualize the units? responses at different layers of the CNNs, demonstrating that an object-centric network
(using ImageNet [12]) and a scene-centric network (using Places) learn different features.
2
Places Database
The first benchmark for scene classification was the Scene15 database [13] based on [17]. This
dataset contains only 15 scene categories with a few hundred images per class, where current classifiers are saturating this dataset nearing human performance at 95%. The MIT Indoor67 database
[19] has 67 categories on indoor places. The SUN database [24] was introduced to provide a wide
coverage of scene categories. It is composed of 397 categories containing more than 100 images per
category.
Despite those efforts, all these scene-centric datasets are small in comparison with current object
datasets such as ImageNet (note that ImageNet also contains scene categories but in a very small
proportion as is shown in Fig. 2). Complementary to ImageNet (mostly object-centric), we present
here a scene-centric database, that we term the Places database. As now, Places contain more than 7
million images from 476 place categories, making it the largest image database of scenes and places
so far and the first scene-centric database competitive enough to train algorithms that require huge
amounts of data, such as CNNs.
2.1
Building the Places Database
Since the SUN database [24] has a rich scene taxonomy, the Places database has inherited the same
list of scene categories. To generate the query of image URL, 696 common adjectives (messy, spare,
sunny, desolate, etc), manually selected from a list of popular adjectives in English, are combined
with each scene category name and are sent to three image search engines (Google Images, Bing
Images, and Flickr). Adding adjectives to the queries allows us to download a larger number of
images than what is available in ImageNet and to increase the diversity of visual appearances. We
then remove duplicated URLs and download the raw images with unique URLs. To date, more than
40 million images have been downloaded. Only color images of 200?200 pixels or larger are kept.
PCA-based duplicate removal is conducted within each scene category in the Places database and
across the same scene category in the SUN database, which ensures that Places and the SUN do not
contain the same images, allowing us to combine the two datasets.
The images that survive this initial selection are sent to Amazon Mechanical Turk for two rounds
of individual image annotation. For a given category name, its definition as in [24], is shown at
the top of a screen, with a question like is this a living room scene? A single image at a time
is shown centered in a large window, and workers are asked to press a Yes or No key. For the
first round of labeling, the default answer is set to No, requiring the worker to actively pick up the
positive images. The positive images resulting from the first round annotation are further sent for a
second round annotation, in which the default answer is set to Yes (to pick up the remaining negative
images). In each HIT(one assignment for each worker), 750 downloaded images are included for
annotation, and an additional 30 positive samples and 30 negative samples with ground truth from
the SUN database are also randomly injected as control. Valid HITs kept for further analyses require
an accuracy of 90% or higher on these control images. After the two rounds of annotation, and as this
paper is published, 7,076,580 images from 476 scene categories are included in the Places database.
Fig. 1 shows image samples obtained with some of the adjectives used in the queries.
1
The database and pre-trained networks are available at http://places.csail.mit.edu
2
spare bedroom
wooded kitchen
teenage bedroom
messy kitchen
romantic bedroom
darkest forest path
stylish kitchen
rocky coast
wintering forest path
misty coast
greener forest path
sunny coast
Figure 1: Image samples from the scene categories grouped by their queried adjectives.
100000
Places
ImageNet
SUN
10000
100
bridge
cemetery
tower
train railway
canyon
pond
fountain
castle
lighthouse
valley
harbor
skyscraper
aquarium
palace
arch
highway
bedroom
creek
botanical garden
restaurant
kitchen
ocean
railroad track
river
baseball field
rainforest
stadium baseball
art gallery
office building
golf course
mansion
staircase
windmill
coast
stadium football
parking lot
basilica
building facade
lobby
abbey
vegetable garden
volcano
amusement park
shed
herb garden
alley
pasture
marsh
raft
dock
playground
mountain
hotel room
sea cliff
courtyard
badlands
office
boardwalk
desert sand
patio
living room
runway
plaza
sky
motel
underwater coral reef
driveway
dining room
train station platform
hospital
viaduct
forest path
construction site
campsite
mausoleum
music studio
mountain snowy
basement
cottage garden
boat deck
coffee shop
pagoda
shower
classroom
ballroom
corn field
parlor
yard
hot spring
kitchenette
art studio
butte
orchard
gas station
forest road
corridor
closet
fire station
dam
ski slope
field wild
ski resort
iceberg
fairway
phone booth
swamp
airport terminal
auditorium
wheat field
wind farm
bookstore
fire escape
supermarket
bar
water tower
rice paddy
cockpit
home office
crosswalk
bakery shop
bayou
veranda
slum
formal garden
chalet
ruin
attic
track outdoor
clothing store
tree farm
residential neighborhood
courthouse
restaurant patio
engine room
market outdoor
excavation
inn outdoor
trench
schoolhouse
conference room
pavilion
aqueduct
temple east asia
conference center
hospital room
rock arch
racecourse
shopfront
topiary garden
field cultivated
church outdoor
pulpit
museum indoor
dinette home
ice cream parlor
gift shop
boxing ring
laundromat
nursery
martial arts gym
swimming pool outdoor
food court
cathedral outdoor
reception
temple south asia
amphitheater
medina
pantry
galley
apartment building outdoor
watering hole
islet
banquet hall
crevasse
jail cell
candy store
kindergarden classroom
dorm room
bowling alley
ice skating rink outdoor
garbage dump
assembly line
picnic area
locker room
monastery outdoor
game room
kasbah
hotel outdoor
bus interior
doorway outdoor
television studio
butchers shop
waiting room
bamboo forest
restaurant kitchen
subway station platform
desert vegetation
beauty salon
rope bridge
stage indoor
snowfield
cafeteria
shoe shop
sandbar
igloo
1000
Figure 2: Comparison of the number of images per scene category in three databases.
We made 2 subsets of Places that will be used across the paper as benchmarks. The first one is Places
205, with the 205 categories with more than 5000 images. Fig. 2 compares the number of images in
Places 205 with ImageNet and SUN. Note that ImageNet only has 128 of the 205 categories, while
SUN contains all of them (we will call this set SUN 205, and it has, at least, 50 images per category).
The second subset of Places used in this paper is Places 88. It contains the 88 common categories
with ImageNet such that there are at least 1000 images in ImageNet. We call the corresponding
subsets SUN 88 and ImageNet 88.
3
Comparing Scene-centric Databases
Despite the importance of benchmarks and training datasets in computer vision, comparing datasets
is still an open problem. Even datasets covering the same visual classes have notable differences
providing different generalization performance when used to train a classifier [23]. Beyond the
number of images and categories, there are aspects that are important but difficult to quantify, like
the variability in camera poses, in decoration styles or in the objects that appear in the scene.
Although the quality of a database will be task dependent, it is reasonable to assume that a good
database should be dense (with a high degree of data concentration), and diverse (it should include
a high variability of appearances and viewpoints). Both quantities, density and diversity, are hard to
estimate in image sets, as they assume some notion of similarity between images which, in general,
is not well defined. Two images of scenes can be considered similar if they contain similar objects,
and the objects are in similar spatial configurations and pose, and have similar decoration styles.
However, this notion is loose and subjective so it is hard to answer the question are these two images
similar? For this reason, we define relative measures for comparing datasets in terms of density and
diversity that only require ranking similarities. In this section we will compare the densities and
diversities of SUN, ImageNet and Places using these relative measures.
3
3.1
Relative Density and Diversity
Density is a measure of data concentration. We assume that, in an image set, high density is equivalent to the fact that images have, in general, similar neighbors. Given two databases A and B, relative
density aims to measure which one of the two sets has the most similar nearest neighbors. Let a1
be a random image from set A and b1 from set B and let us take their respective nearest neighbors
in each set, a2 from A and b2 from B. If A is denser than B, then it would be more likely that a1
and a2 are closer to each other than b1 and b2 . From this idea we define the relative density as
DenB (A) = p (d(a1 , a2 ) < d(b1 , b2 )), where d(a1 , a2 ) is a distance measure between two images
(small distance implies high similarity). With this definition of relative density we have that A is
denser than B if, and only if, DenB (A) > DenA (B). This definition can be extended to an arbitrary
number of datasets, A1 , ..., AN :
DenA2 ,...,AN (A1 ) = p(d(a11 , a12 ) < min d(ai1 , ai2 ))
i=2:N
(1)
where ai1 ? Ai are randomly selected and ai2 ? Ai are near neighbors of their respective ai1 .
The quality of a dataset can not be measured just by its density. Imagine, for instance, a dataset
composed of 100,000 images all taken within the same bedroom. This dataset would have a very
high density but a very low diversity as all the images would look very similar. An ideal dataset,
expected to generalize well, should have high diversity as well.
There are several measures of diversity, most of them frequently used in biology to characterize the
richness of an ecosystem (see [9] for a review). In this section, we will use a measure inspired by
Simpson index of diversity [22]. Simpson index measures the probability that two random individuals from an ecosystem belong to the same species. It is a measure of how well distributed are the
individuals across different species in an ecosystem and it is related to the entropy of the distribution. Extending this measure for evaluating the diversity of images within a category is non-trivial if
there are no annotations of sub-categories. For this reason, we propose to measure relative diversity
of image datasets A and B based on this idea: if set A is more diverse than set B, then two random
images from set B are more likely to be visually similar than two random samples from A. Then,
the diversity of A with respect to B can be defined as DivB (A) = 1 ? p(d(a1 , a2 ) < d(b1 , b2 )),
where a1 , a2 ? A and b1 , b2 ? B are randomly selected. With this definition of relative diversity we
have that A is more diverse than B if, and only if, DivB (A) > DivA (B). For an arbitrary number of
datasets, A1 , ..., AN :
DivA2 ,...,AN (A1 ) = 1 ? p(d(a11 , a12 ) < min d(ai1 , ai2 ))
i=2:N
(2)
where ai1 , ai2 ? Ai are randomly selected.
3.2
Experimental Results
We measured the relative densities and diversities between SUN, ImageNet and Places using AMT.
Both measures used the same experimental interface: workers were presented with different pairs
of images and they had to select the pair that contained the most similar images. We observed that
different annotators are consistent in deciding whether a pair of images is more similar than another
pair of images.
In these experiments, the only difference when estimating density and diversity is how the pairs are
generated. For the diversity experiment, the pairs are randomly sampled from each database. Each
trial is composed of 4 pairs from each database, giving a total of 12 pairs to chose from. We used
4 pairs per database to increase the chances of finding a similar pair and avoiding users having to
skip trials. AMT workers had to select the most similar pair on each trial. We ran 40 trials per
category and two observers per trial, for the 88 categories in common between ImageNet, SUN and
Places databases. Fig. 3a shows some examples of pairs from one of the density experiments.The
pair selected by AMT workers as being more similar is highlighted.
For the density experiments, we selected pairs that were more likely to be visually similar. This
would require first finding the true nearest neighbor of each image, which would be experimentally
costly. Instead we used visual similarity as measured by using the Euclidean distance between
the Gist descriptor [17] of two images. Each pair of images was composed from one randomly
selected image and its 5-th nearest neighbor using Gist (we ignored the first 4 neighbors to avoid
4
ImageNet Places
ImageNet Places
1
Places
0.9
0.8
ImageNet
Diversity
0.7
0.6
0.5
SUN
0.4
0.3
SUN
SUN
0.2
0.1
a)
b)
c)
0
0
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Density
1
Figure 3: a) Examples of pairs for the diversity experiment. b) Examples of pairs for the density
experiment. c) Scatter plot of relative diversity vs. relative density per each category and dataset.
Test on SUN 88
70
70
Test on ImageNet Scene 88
Test on Places 88
55
50
50
40
30
Train on Places 88 [69.5]
Train on SUN 88 [63.3]
Train on ImageNet 88 [62.8]
20
10 0
10
a)
10
1
10
2
10
3
Number of training samples per category
Classification accuracy
60
Classification accuracy
Classification accuracy
60
50
40
30
Train on ImageNet 88 [65.6]
Train on Places 88 [60.3]
Train on SUN 88 [49.2]
20
10
10 0
10
4
b)
10
1
10
2
10
3
Number of training samples per category
45
40
35
30
25
Train on Places 88 [54.2]
Train on ImageNet 88 [44.6]
Train on SUN 88 [37.0]
20
15
10
4
10 0
10
c)
10
1
10
2
10
3
10
4
Number of training samples per category
Figure 4: Cross dataset generalization of training on the 88 common scenes between Places, SUN
and ImageNet then testing on the 88 common scenes from: a) SUN, b) ImageNet and c) Places
database.
near duplicates, which would give a wrong sense of high density). In this case we also show 12
pairs of images at each trial, but run 25 trials per category instead of 40 to avoid duplicate queries.
Fig. 3b shows some examples of pairs per one of the density experiments and also the selected pair
is highlighted. Notice that in the density experiment (where we computed neighbors) the pairs look,
in general, more similar than in the diversity experiment.
Fig. 3c shows a scatter plot of relative diversity vs. relative density for all the 88 categories and the
three databases. The point of crossing between the two black lines indicates the point where all the
results should fall if all the datasets were identical in terms of diversity and density. The figure also
shows the average of the density and diversity over all categories for each dataset.
In terms of density, the three datasets are, on average, very similar. However, there is a larger
variation in terms of diversity, showing Places to be the most diverse of the three datasets. The
average relative diversity on each dataset is 0.83 for Places, 0.67 for ImageNet and 0.50 for SUN.
In the experiment, users selected pairs from the SUN database to be the closest to each other 50%
of the time, while the pairs from the Places database were judged to be the most similar only on
17% of the trials. The categories with the largest variation in diversity across the three datasets are
playground, veranda and waiting room.
3.3
Cross Dataset Generalization
As discussed in [23], training and testing across different datasets generally results in a drop of
performance due to the dataset bias problem. In this case, the bias between datasets is due, among
other factors, to the differences in the density and diversity between the three datasets. Fig. 4 shows
the classification results obtained from the training and testing on different permutations of the 3
datasets. For these results we use the features extracted from a pre-trained ImageNet-CNN and a
linear SVM. In all three cases training and testing on the same dataset provides the best performance
for a fixed number of training examples. As the Places database is very large, it achieves the best
performance on two of the test sets when all the training data is used. In the next section we will
show that a CNN network trained using the Places database achieves a significant improvement over
scene-centered benchmarks in comparison with a network trained using ImageNet.
5
Table 1: Classification accuracy on the test set of Places 205 and the test set of SUN 205.
Places 205 SUN 205
Places-CNN
50.0%
66.2%
ImageNet CNN feature+SVM
40.8%
49.6%
4
Training Neural Network for Scene Recognition and Deep Features
Deep convolutional neural networks have obtained impressive classification performance on the
ImageNet benchmark [12]. For the training of Places-CNN, we randomly select 2,448,873 images
from 205 categories of Places (referred to as Places 205) as the train set, with minimum 5,000 and
maximum 15,000 images per category. The validation set contains 100 images per category and the
test set contains 200 images per category (a total of 41,000 images). Places-CNN is trained using
the Caffe package on a GPU NVIDIA Tesla K40. It took about 6 days to finish 300,000 iterations of
training. The network architecture of Places-CNN is the same as the one used in the Caffe reference
network [10]. The Caffe reference network, which is trained on 1.2 million images of ImageNet
(ILSVRC 2012), has approximately the same architecture as the network proposed by [12]. We call
the Caffe reference network as ImageNet-CNN in the following comparison experiments.
4.1
Visualization of the Deep Features
Through the visualization of the responses of the units for various levels of network layers, we
can have a better understanding of the differences between the ImageNet-CNN and Places-CNN
given that they share the same architecture. Fig.5 visualizes the learned representation of the units
at the Conv 1, Pool 2, Pool 5, and FC 7 layers of the two networks. Whereas Conv 1 units can
be directly visualized (they capture the oriented edges and opponent colors from both networks),
we use the mean image method to visualize the units of the higher layers: we first combine the
test set of ImageNet LSVRC2012 (100,000 images) and SUN397 (108,754 images) as the input
for both networks; then we sort all these images based on the activation response of each unit at
each layer; finally we average the top 100 images with the largest responses for each unit as a kind
of receptive field (RF) visualization of each unit. To compare the units from the two networks,
Fig. 5 displays mean images sorted by their first principal component. Despite the simplicity of
the method, the units in both networks exhibit many differences starting from Pool 2. From Pool
2 to Pool 5 and FC 7, gradually the units in ImageNet-CNN have RFs that look like object-blobs,
while units in Places-CNN have more RFs that look like landscapes with more spatial structures.
These learned unit structures are closely relevant to the differences of the training data. In future
work, it will be fascinating to relate the similarity and differences of the RF at different layers of
the object-centric network and scene-centric network with the known object-centered and scenecentered neural cortical pathways identified in the human brain (for a review, [16]). In the next
section we will show that these two networks (only differing in the training sets) yield very different
performances on a variety of recognition benchmarks.
4.2
Results on Places 205 and SUN 205
After the Places-CNN is trained, we use the final layer output (Soft-max) of the network to classify
images in the test set of Places 205 and SUN 205. The classification result is listed in Table 1. As
a baseline comparison, we show the results of a linear SVM trained on ImageNet-CNN features
of 5000 images per category in Places 205 and 50 images per category in SUN 205 respectively.
Places-CNN performs much better. We further compute the performance of the Places-CNN in the
terms of the top-5 error rate (one test sample is counted as misclassified if the ground-truth label is
not among the top 5 predicted labels of the model). The top-5 error rate for the test set of the Places
205 is 18.9%, while the top-5 error rate for the test set of SUN 205 is 8.1%.
4.3
Generic Deep Features for Visual Recognition
We use the responses from the trained CNN as generic features for visual recognition tasks. Responses from the higher-level layers of CNN have proven to be effective generic features with stateof-the-art performance on various image datasets [5, 20]. Thus we evaluate performance of the
6
Pool 2
Pool 5
FC 7
Pla c es -CNN
ImageNet-CNN
Conv 1
Figure 5: Visualization of the units? receptive fields at different layers for the ImageNet-CNN and
Places-CNN. Conv 1 units contains 96 filters. The Pool 2 feature map is 13?13?256; The Pool 5
feature map is 6?6?256; The FC 7 feature map is 4096?1. Subset of units at each layer are shown.
Table 2: Classification accuracy/precision on scene-centric databases and object-centric databases
for the Places-CNN feature and ImageNet-CNN feature. The classifier in all the experiments is a
linear SVM with the same parameters for the two features.
SUN397
MIT Indoor67
Scene15
SUN Attribute
Places-CNN feature
54.32?0.14
68.24
90.19?0.34
91.29
ImageNet-CNN feature 42.61?0.16
56.79
84.23?0.37
89.85
Caltech101
Caltech256
Action40
Event8
Places-CNN feature
65.18?0.88
45.59?0.31
42.86?0.25
94.12?0.99
ImageNet-CNN feature 87.22?0.92
67.23?0.27
54.92?0.33
94.42?0.76
deep features from the Places-CNN on the following scene and object benchmarks: SUN397 [24],
MIT Indoor67 [19], Scene15 [13], SUN Attribute [18], Caltech101 [7], Caltech256 [8], Stanford
Action40 [25], and UIUC Event8 [15]. All the experiments follow the standards in those papers 2 .
As a comparison, we evaluate the deep feature?s performance from the ImageNet-CNN on those
same benchmarks. Places-CNN and ImageNet-CNN have exactly the same network architecture,
but they are trained on scene-centric data and object-centric data respectively. We use the deep
features from the response of the Fully Connected Layer (FC) 7 of the CNNs, which is the final fully
connected layer before producing the class predictions. There is only a minor difference between
the feature of FC 7 and the feature of FC 6 layer [5]. The deep feature for each image is a 4096dimensional vector.
Table 2 summarizes the classification accuracy on various datasets for the ImageNet-CNN feature
and the Places-CNN feature. Fig.6 plots the classification accuracy for different visual features
on SUN397 database and SUN Attribute dataset. The classifier is a linear SVM with the same
default parameters for the two deep features (C=1) [6]. The Places-CNN feature shows impressive
performance on scene classification benchmarks, outperforming the current state-of-the-art methods
for SUN397 (47.20% [21]) and for MIT Indoor67 (66.87% [4]). On the other hand, the ImageNetCNN feature shows better performance on object-related databases. Importantly, our comparison
2
Detailed experimental setups are included in the supplementary materials.
7
Benchmark on SUN397 Dataset
Benchmark on SUN Attribute Dataset
70
Combined kernel [37.5]
HoG2x2 [26.3]
0.9
DenseSIFT [23.5]
60
Texton [21.6]
Gist [16.3]
0.85
LBP [14.7]
50
ImageNet?CNN [42.6]
Average Precision
Classification accuracy
Places?CNN [54.3]
40
30
0.8
0.75
0.7
Places?CNN [0.912]
ImageNet?CNN [0.898]
20
Combined kernel [0.879]
0.65
HoG2x2 [0.848]
Self?similarity [0.820]
10
Geometric Color Hist [0.783]
0.6
Gist [0.799]
0
1
5
10
20
0.55
1/1
50
Number of training samples per category
5/5
20/20
50/50
150/150
Number of training samples per attribute (positive/negative)
Figure 6: Classification accuracy on the SUN397 Dataset and average precision on the SUN Attribute Dataset with increasing size of training samples for the ImageNet-CNN feature and the
Places-CNN feature. Results of other hand-designed features/kernels are fetched from [24] and
[18] respectively.
Table 3: Classification accuracy/precision on various databases for Hybrid-CNN feature. The numbers in bold indicate the results outperform the ImageNet-CNN feature or Places-CNN feature.
SUN397
53.86?0.21
MIT Indoor67
70.80
Scene15
91.59?0.48
SUN Attribute
91.56
Caltech101
84.79?0.66
Caltech256
65.06?0.25
Action40
55.28?0.64
Event8
94.22?0.78
shows that Places-CNN and ImageNet-CNN have complementary strengths on scene-centric tasks
and object-centric tasks, as expected from the benchmark datasets used to train these networks.
Furthermore, we follow the same experimental setting of train and test split in [1] to fine tune
Places-CNN on SUN397: the fine-tuned Places-CNN achieves the accuracy of 56.2%, compared to
the accuracy of 52.2% achieved by the fine-tuned ImageNet-CNN in [1]. Note that the final output
of the fine-tuned CNN is directly used to predict scene category.
Additionally, we train a Hybrid-CNN, by combining the training set of Places-CNN and training set
of ImageNet-CNN. We remove the overlapping scene categories from the training set of ImageNet,
and then the training set of Hybrid-CNN has 3.5 million images from 1183 categories. HybridCNN is trained over 700,000 iterations, under the same network architecture of Places-CNN and
ImageNet-CNN. The accuracy on the validation set is 52.3%. We evaluate the deep feature (FC
7) from Hybrid-CNN on benchmarks shown in Table 3. Combining the two datasets yields an
additional increase in performance for a few benchmarks.
5
Conclusion
Deep convolutional neural networks are designed to benefit and learn from massive amounts of data.
We introduce a new benchmark with millions of labeled images, the Places database, designed to
represent places and scenes found in the real world. We introduce a novel measure of density and
diversity, and show the usefulness of these quantitative measures for estimating dataset biases and
comparing different datasets. We demonstrate that object-centric and scene-centric neural networks
differ in their internal representations, by introducing a simple visualization of the receptive fields
of CNN units. Finally, we provide the state-of-the-art performance using our deep features on all
the current scene benchmarks.
Acknowledgement. Thanks to Aditya Khosla for valuable discussions. This work is supported by the National
Science Foundation under Grant No. 1016862 to A.O, ONR MURI N000141010933 to A.T, as well as MIT
Big Data Initiative at CSAIL, Google and Xerox Awards, a hardware donation from NVIDIA Corporation, to
A.O and A.T., Intel and Google awards to J.X, and grant TIN2012-38187-C03-02 to A.L. This work is also supported by the Intelligence Advanced Research Projects Activity (IARPA) via Air Force Research Laboratory,
contract FA8650-12-C-7211 to A.T. The U.S. Government is authorized to reproduce and distribute reprints for
Governmental purposes notwithstanding any copyright annotation thereon. Disclaimer: The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the
official policies or endorsements, either expressed or implied, of IARPA, AFRL, or the U.S. Government.
8
References
[1] P. Agrawal, R. Girshick, and J. Malik. Analyzing the performance of multilayer neural networks for
object recognition. In Proc. ECCV. 2014.
R in Machine Learning, 2009.
[2] Y. Bengio. Learning deep architectures for ai. Foundations and trends
[3] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. Imagenet: A large-scale hierarchical image
database. In Proc. CVPR, 2009.
[4] C. Doersch, A. Gupta, and A. A. Efros. Mid-level visual element discovery as discriminative mode
seeking. In In Advances in Neural Information Processing Systems, 2013.
[5] J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng, and T. Darrell. Decaf: A deep convolutional activation feature for generic visual recognition. 2014.
[6] R.-E. Fan, K.-W. Chang, C.-J. Hsieh, X.-R. Wang, and C.-J. Lin. LIBLINEAR: A library for large linear
classification. 2008.
[7] L. Fei-Fei, R. Fergus, and P. Perona. Learning generative visual models from few training examples: An
incremental bayesian approach tested on 101 object categories. Computer Vision and Image Understanding, 2007.
[8] G. Griffin, A. Holub, and P. Perona. Caltech-256 object category dataset. 2007.
[9] C. Heip, P. Herman, and K. Soetaert. Indices of diversity and evenness. Oceanis, 1998.
[10] Y. Jia. Caffe: An open source convolutional architecture for fast feature embedding. http://caffe.
berkeleyvision.org/, 2013.
[11] T. Konkle, T. F. Brady, G. A. Alvarez, and A. Oliva. Scene memory is more detailed than you think: the
role of categories in visual long-term memory. Psych Science, 2010.
[12] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural
networks. In In Advances in Neural Information Processing Systems, 2012.
[13] S. Lazebnik, C. Schmid, and J. Ponce. Beyond bags of features: Spatial pyramid matching for recognizing
natural scene categories. In Proc. CVPR, 2006.
[14] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel. Backpropagation applied to handwritten zip code recognition. Neural computation, 1989.
[15] L.-J. Li and L. Fei-Fei. What, where and who? classifying events by scene and object recognition. In
Proc. ICCV, 2007.
[16] A. Oliva. Scene perception (chapter 51). The New Visual Neurosciences, 2013.
[17] A. Oliva and A. Torralba. Modeling the shape of the scene: A holistic representation of the spatial
envelope. Int?l Journal of Computer Vision, 2001.
[18] G. Patterson and J. Hays. Sun attribute database: Discovering, annotating, and recognizing scene attributes. In Proc. CVPR, 2012.
[19] A. Quattoni and A. Torralba. Recognizing indoor scenes. In Proc. CVPR, 2009.
[20] A. S. Razavian, H. Azizpour, J. Sullivan, and S. Carlsson. Cnn features off-the-shelf: an astounding
baseline for recognition. arXiv preprint arXiv:1403.6382, 2014.
[21] J. S?anchez, F. Perronnin, T. Mensink, and J. Verbeek. Image classification with the fisher vector: Theory
and practice. Int?l Journal of Computer Vision, 2013.
[22] E. H. Simpson. Measurement of diversity. Nature, 1949.
[23] A. Torralba and A. A. Efros. Unbiased look at dataset bias. In Proc. CVPR, 2011.
[24] J. Xiao, J. Hays, K. A. Ehinger, A. Oliva, and A. Torralba. Sun database: Large-scale scene recognition
from abbey to zoo. In Proc. CVPR, 2010.
[25] B. Yao, X. Jiang, A. Khosla, A. L. Lin, L. Guibas, and L. Fei-Fei. Human action recognition by learning
bases of action attributes and parts. In Proc. ICCV, 2011.
9
| 5349 |@word trial:8 cnn:62 proportion:1 open:2 hsieh:1 pick:2 liblinear:1 initial:1 configuration:1 contains:7 tuned:3 subjective:1 current:5 comparing:4 activation:2 scatter:2 gpu:1 candy:1 shape:1 remove:2 designed:4 gist:4 plot:3 drop:1 v:2 intelligence:1 selected:9 generative:1 discovering:1 short:1 provides:1 org:1 zhang:1 corridor:1 initiative:1 combine:2 wild:1 pathway:1 introduce:5 expected:2 market:1 frequently:1 uiuc:1 brain:3 terminal:1 inspired:2 food:1 window:1 increasing:2 gift:1 conv:4 estimating:2 project:1 snowy:1 pasture:1 what:2 mountain:2 kind:1 interpreted:1 teenage:1 psych:1 differing:1 finding:2 playground:2 corporation:1 brady:1 remember:1 sky:1 quantitative:1 shed:1 exactly:1 classifier:5 hit:2 wrong:1 control:2 unit:17 grant:2 appear:1 producing:1 thereon:1 positive:4 before:2 ice:2 despite:3 analyzing:1 cliff:1 jiang:1 path:4 approximately:1 reception:1 sandbar:1 chose:1 black:1 range:1 unique:1 camera:1 pla:1 testing:4 lecun:1 practice:1 trench:1 backpropagation:1 sullivan:1 procedure:1 area:1 significantly:1 matching:1 pre:2 road:1 interior:1 selection:1 valley:1 judged:1 context:1 dam:1 scene15:5 equivalent:1 map:3 center:1 exposure:2 starting:1 amazon:2 simplicity:1 fountain:1 importantly:1 embedding:1 swamp:1 notion:2 underwater:1 variation:2 construction:1 imagine:1 user:2 massive:1 runway:1 crossing:1 trend:1 recognition:20 element:1 skating:1 muri:1 database:48 labeled:3 observed:1 role:2 caltech256:3 preprint:1 wang:1 capture:1 apartment:1 ensures:1 wheat:1 richness:2 sun:42 k40:1 connected:2 valuable:1 ran:1 environment:2 complexity:1 messy:2 asked:1 cafeteria:1 trained:13 astonishing:1 baseball:2 basement:1 patterson:1 various:5 chapter:1 train:17 fast:1 describe:1 effective:1 artificial:1 query:4 zhou1:1 labeling:1 dorm:1 neighborhood:1 aquarium:1 caffe:6 larger:4 stanford:1 denser:2 supplementary:1 cvpr:6 annotating:1 football:1 ability:1 think:1 farm:2 highlighted:2 final:3 pavilion:1 blob:1 agrawal:1 inn:1 dining:1 propose:2 rock:1 took:1 facade:1 relevant:1 wooded:1 combining:2 date:1 holistic:1 konkle:1 sutskever:1 darrell:1 extending:1 sea:1 a11:2 incremental:1 ring:1 object:26 donation:1 pose:2 measured:3 exemplar:1 nearest:4 minor:1 progress:1 coverage:1 predicted:1 skip:1 decoration:2 implies:1 quantify:1 indicate:1 differ:1 closely:1 attribute:11 cnns:6 filter:1 centered:3 human:6 a12:2 material:1 spare:2 sand:1 require:4 government:2 generalization:3 clothing:1 hall:1 ground:2 ruin:1 considered:1 visually:2 deciding:1 bakery:1 predict:1 visualize:2 guibas:1 efros:2 achieves:3 abbey:2 a2:6 torralba:4 purpose:1 proc:9 bag:1 label:2 jackel:1 bridge:2 highway:1 largest:3 grouped:1 hubbard:1 hoffman:1 mit:8 aim:1 reaching:2 agata:1 avoid:2 shelf:1 yard:1 beauty:1 railroad:1 office:3 closet:1 azizpour:1 ponce:1 iconic:1 improvement:1 indicates:1 baseline:4 sense:1 dependent:1 perronnin:1 perona:2 butcher:1 misclassified:1 reproduce:1 pixel:1 classification:21 among:2 stateof:1 art:7 platform:2 airport:1 tzeng:1 spatial:4 field:8 amusement:1 having:1 sampling:1 manually:1 biology:1 identical:1 park:1 look:5 survive:1 discrepancy:1 future:1 duplicate:3 few:4 escape:1 alley:2 palace:1 randomly:7 composed:4 museum:1 recognize:2 national:1 individual:3 oriented:1 kitchen:5 astounding:1 botanical:1 fire:2 organization:1 huge:1 simpson:3 ai1:5 henderson:1 copyright:1 edge:1 closer:1 worker:6 respective:2 tree:1 laundromat:1 euclidean:1 girshick:1 instance:1 classify:1 soft:1 herb:1 modeling:1 temple:2 assignment:1 subtending:1 introducing:1 subset:4 hundred:1 usefulness:1 krizhevsky:1 recognizing:3 galley:1 conducted:1 universitat:1 characterize:1 answer:3 combined:3 thanks:1 density:29 river:1 csail:2 contract:1 dong:1 off:1 pool:10 together:1 yao:1 containing:1 nearing:1 evenness:1 castle:1 resort:1 style:2 actively:1 li:3 distribute:1 de:1 diversity:34 b2:5 bold:1 availability:1 int:2 notable:1 register:1 ranking:1 view:1 lot:1 wind:1 observer:1 razavian:1 reached:1 competitive:2 sort:1 inherited:1 annotation:7 slope:1 jia:2 disclaimer:1 air:1 accuracy:15 convolutional:7 descriptor:1 who:1 yield:2 landscape:1 yes:2 generalize:1 raw:1 bayesian:1 handwritten:1 zoo:1 published:1 salon:1 visualizes:1 n000141010933:1 quattoni:1 flickr:1 rink:1 definition:5 hotel:2 turk:2 sampled:1 dataset:22 duplicated:1 massachusetts:1 popular:1 color:3 organized:1 classroom:2 holub:1 sophisticated:1 centric:24 islet:1 attained:1 higher:4 day:1 follow:2 asia:2 response:8 afrl:1 alvarez:1 mensink:1 furthermore:1 just:2 stage:1 arch:2 hand:3 overlapping:1 glance:1 google:3 mode:1 quality:2 aude:1 boardwalk:1 name:2 building:4 contain:4 requiring:1 staircase:1 true:1 unbiased:1 ai2:4 laboratory:1 round:5 bowling:1 game:1 self:1 covering:1 berkeleyvision:1 rope:1 demonstrate:1 performs:1 interface:1 hallmark:1 image:81 coast:4 marsh:1 novel:1 lazebnik:1 common:5 iceberg:1 shower:1 million:8 belong:1 discussed:1 vegetation:1 ecosystem:3 significant:1 measurement:1 queried:1 ai:4 reef:1 doersch:1 had:2 similarity:6 impressive:2 etc:1 base:1 closest:1 recent:2 phone:1 lighthouse:1 store:2 nvidia:2 hay:2 outperforming:2 success:1 onr:1 accomplished:1 caltech:1 minimum:1 additional:2 zip:1 deng:1 catalunya:1 living:2 patio:2 cross:2 bolei:1 lin:2 long:1 windmill:1 award:2 a1:10 prediction:1 verbeek:1 oliva:4 multilayer:1 vision:5 arxiv:2 iteration:2 kernel:3 represent:1 texton:1 achieved:1 cell:1 pyramid:1 whereas:2 lbp:1 fine:4 source:1 envelope:1 south:1 sent:3 cream:1 call:3 near:2 ideal:1 feedforward:1 split:1 enough:2 bengio:1 variety:2 harbor:1 restaurant:3 finish:1 bedroom:5 architecture:12 identified:1 idea:2 court:1 golf:1 whether:1 pca:1 url:3 effort:1 fa8650:1 action:2 deep:20 antonio:1 garbage:1 ignored:1 stadium:2 vegetable:1 generally:1 listed:1 detailed:2 amount:2 cathedral:1 tune:1 mid:1 ten:1 hardware:1 visualized:1 category:47 generate:1 http:2 outperform:1 millisecond:1 notice:1 governmental:1 neuroscience:1 per:21 track:2 diverse:5 nursery:1 waiting:2 key:1 subway:1 demonstrating:1 canyon:1 kept:2 swimming:1 year:1 residential:1 run:1 volcano:1 package:1 injected:1 you:1 place:79 reasonable:1 monastery:1 home:2 fetched:1 endorsement:1 griffin:1 summarizes:1 layer:15 display:1 fan:1 fascinating:1 plaza:1 activity:1 strength:1 fei:8 scene:62 aspect:1 min:2 spring:1 corn:1 xerox:1 orchard:1 sunny:2 across:5 primate:1 making:1 gradually:1 iccv:2 dock:1 taken:1 fairway:1 visualization:6 bing:1 indoor67:6 bus:1 loose:1 mechanism:1 available:2 boxing:1 opponent:1 xiao2:1 denker:1 hierarchical:2 generic:4 ocean:1 darkest:1 gym:1 top:6 remaining:1 include:1 assembly:1 music:1 giving:1 coral:1 coffee:1 establish:2 implied:1 pond:1 malik:1 question:2 quantity:1 seeking:1 receptive:3 concentration:2 costly:1 exhibit:1 distance:3 capacity:1 tower:2 trivial:1 reason:3 gallery:1 water:1 besides:1 code:1 index:3 providing:1 difficult:1 mostly:1 setup:1 taxonomy:1 relate:1 locker:1 negative:3 rise:1 ski:2 policy:1 perform:1 allowing:2 anchez:1 datasets:30 benchmark:16 howard:1 gas:1 extended:1 variability:2 hinton:1 station:4 arbitrary:2 igloo:1 download:2 introduced:1 pair:23 mechanical:2 imagenet:57 engine:2 learned:3 herein:1 tremendous:1 boser:1 beyond:2 bar:1 perception:1 indoor:4 herman:1 adjective:5 rf:4 max:1 memory:2 garden:6 hot:1 event:1 natural:3 hybrid:4 force:1 boat:1 advanced:1 representing:1 shop:5 amphitheater:1 technology:1 rocky:1 library:1 picture:2 martial:1 reprint:1 church:1 schmid:1 supermarket:1 review:2 understanding:3 geometric:1 removal:1 acknowledgement:1 discovery:1 carlsson:1 relative:14 fully:2 permutation:1 proven:1 versus:1 skyscraper:1 annotator:1 validation:2 foundation:2 downloaded:2 c03:1 degree:1 consistent:1 xiao:1 viewpoint:1 classifying:1 share:1 romantic:1 eccv:1 course:1 caltech101:3 supported:2 english:1 formal:1 bias:4 institute:1 wide:1 neighbor:8 fall:1 distributed:1 benefit:1 default:3 cortical:1 world:3 valid:1 rich:2 evaluating:1 author:1 collection:1 made:1 counted:1 far:1 feat:1 lobby:1 hist:1 b1:5 auditorium:1 doorway:1 discriminative:1 fergus:1 search:1 khosla:2 table:6 additionally:1 learn:5 nature:1 forest:6 necessarily:1 official:1 dense:3 big:1 iarpa:2 tesla:1 complementary:2 fig:10 site:1 dump:1 referred:1 intel:1 screen:1 ehinger:1 precision:4 sub:1 medina:1 parking:1 outdoor:11 donahue:1 emphasizing:1 bookstore:1 showing:1 list:2 svm:5 gupta:1 socher:1 adding:1 importance:1 decaf:1 notwithstanding:1 studio:3 hole:1 television:1 booth:1 authorized:1 torralba1:1 entropy:1 fc:8 appearance:2 likely:3 deck:1 shoe:1 visual:14 vinyals:1 aditya:1 saturating:1 contained:2 expressed:1 chang:1 truth:2 chance:1 constantly:1 amt:4 rice:1 extracted:1 sorted:1 sun397:9 room:12 fisher:1 hard:2 experimentally:1 included:3 principal:1 called:1 hospital:2 specie:2 total:2 experimental:4 e:1 east:1 desert:2 select:3 ilsvrc:1 internal:2 jianxiong:1 evaluate:3 princeton:1 tested:1 avoiding:1 |
4,804 | 535 | Unsupervised learning
of distributions on binary vectors
using two layer networks
David Haussler
Computer and Information Sciences
University of California Santa Cruz
Santa Cruz , CA 95064
Yoav Freund?
Computer and Information Sciences
University of California Santa Cruz
Santa Cruz, CA 95064
Abstract
We study a particular type of Boltzmann machine with a bipartite graph structure called a harmonium. Our interest is in using such a machine to model a probability distribution on binary input
vectors . We analyze the class of probability distributions that can be modeled by such machines.
showing that for each n ~ 1 this class includes arbitrarily good appwximations to any distribution
on the set of all n-vectors of binary inputs. We then present two learning algorithms for these
machines .. The first learning algorithm is the standard gradient ascent heuristic for computing
maximum likelihood estimates for the parameters (i.e. weights and thresholds) of the modeL Here
we give a closed form for this gradient that is significantly easier to compute than the corresponding
gradient for the general Boltzmann machine . The second learning algorithm is a greedy method
that creates the hidden units and computes their weights one at a time. This method is a variant
of the standard method for projection pursuit density estimation . We give experimental results for
these learning methods on synthetic data and natural data from the domain of handwritten digits.
1
Introduction
=
Let us suppose that each example in our in put data is a binary vector i
{x I, ... , x n } E {? l}n. and that
each such example is generated independently at random according some unknown distribution on {?l}n.
This situation arises. for instance. when each example consists of (possibly noisy) measurements of n different
binary attributes of a randomly selected object . In such a situation, unsupervised learning can be usefully
defined as using the input data to find a good model of the unknown distribution on {? l}n and thereby
learning the structure in the data.
The process of learning an unknown distribution from examples is usually called denszty estimation or
parameter estimation in statistics, depending on the nature of the class of distributions used as models.
Connectionist models of this type include Bayes networks (14). mixture models [3.13], and Markov random
fields [14,8]. Network models based on the notion of energy minimization such as Hopfield nets [9] and
Boltzmann machines [1] can also be used as models of probability distributions .
? yoavGcis.ucsc.edu
912
Unsupervised learning of distributions on binary vectors using 2-layer networks
The models defined by Hopfield networks are a special case of the more general Markov random field models
in which the local interactions are restricted to symmetric pairwise interactions between components of
the input. Boltzmann machines also use only pairwise interactions, but in addition they include hidden
units, which correspond to unobserved variables. These unobserved variables interact with the observed
variables represented by components of the input vector. The overall distribution on the set of possible
input vectors is defined as the marginal distribution induced on the components of the input vector by the
Markov random field over all variables, both observed and hidden . While the Hopfield network is relatively
well understood, it is limited in the types of distributions that it can model. On the other hand, Boltzmann
machines are universal in the sense that they are powerful enough to model any distribution (to any degree
of approximation), but the mathematical analysis of their capabilities is often intractable. Moreover, the
standard learning algorithm for the Boltzmann machine, a gradient ascent heuristic to compute the maximum
likelihood estimates for the weights and thresholds, requires repeated stochastic approximation, which results
in unacceptably slow learning. I In this work we attempt to narrow the gap between Hopfield networks and
Boltzmann machines by finding a model that will be powerful enough to be universal, 2 yet simple enough
to be analyzable and computationally efficient. 3 We have found such a model in a minor variant of the
special type of Boltzmann machine defined by Smolensky in his harmony theory [16][Ch.6J. This special type
of Boltzmann machine is defined by a network with a simple bipartite graph structure, which he called a
harmonium.
The harmonium consists of two types of units: input units, each of which holds one component of the input
vector, and hidden units, representing hidden variables. There is a weighted connection between each input
unit and each hidden unit, and no connections between input units or between hidden units (see Figure (1)) .
The presence of the hidden units induces dependencies, or correlations, between the variables modeled by
input units . To illustrate the kind of model that results, consider the distribution of people that visit a
specific coffee shop on Sunday. Let each of the n input variables represent the presence (+ 1) or absence (-1)
of a particular person that Sunday. These random variables are clearly not independent, e.g. if Fred's wife
and daughter are there, it is more likely that Fred is there , if you see three members of the golf club, you
expect to see other members of the golf club, if Bill is there you are unlikely to see Brenda there, etc. This
situation can be modeled by a harmonium model in which each hidden variable represents the presence or
absence of a social group . The weights connecting a hidden unit and an ipput unit measure the tendency of
the corresponding person to be associated with the corresponding group . In this coffee shop situation, several
social groups may be present at one time , exerting a combined influence on the distribution of customers.
This can be mo'deled easily with the harmonium , but is difficult to model using Bayes networks or mixture
models . <4
2
The Model
Let us begin by formalizing the harmonium model. To model a distribution on {?I}" we will use n input
units and some number m ~ 0 of hidden units. These units are connected in a bipartite graph as illustrated
in Figure (I) .
The random variables represented by the input units each take values in {+ I , -I}, while the hidden variables,
represented by the hidden units, take values in to, I} . The state of the machine is defined by the values
of these random variables. Define i
(XI," " xn) E {?l}n to be the state of the input units , and h
(hi , ... , hm ) E {O,l}m to be the state of the hidden units .
=
=
The connection weights between the input ~nits and the ith hidden unit are denoted 5 by w(') E R n and the
bias of the ith hidden unit is denoted by 9(') E R. The parameter vector ~
{(w(l),O(l?, . .. ,(w(m),o(m?))
=
lOne possible solution to this is tbe mean-field approximation [15], discussed furtber in section 2 below.
'In (4) we show tbat any distribution over (?1)" can be approximated to within any desired accuracy by a
harmonium model using 2" bidden units.
lSee also otber work relating Bayes nets and Boltzmann machines [12,1] .
t Noisy-OR gates have been introduced in the framework of Bayes Networks to allow for such combinations.
However, using this in networks with hidden units has not been studied, to the best of our knowledge.
~In (16)[Ch .6J, binary connection weights are used . Here we use real-valued weights .
913
914
Freund and Haussler
Hidden Units
Input
m=3
Units
2:1
2:2
2:3
2:4
2:5
Figure 1: The bipartite graph of the harmonium
defines the entire network, and thus also the probability model induced by the network. For a given
energy of a. state configuration of hidden and input units is defined to be
,p, the
m
E(i, hl~)
=-
L(w(i) . i
+ 8(i?)h i
(1)
i=!
and the probability of a configuration is
Pr(i,hl?l)
1
- l.) where Z =
= -Ze-E(Z,h
~-
L.,e-;.E(Z,h l .).
z,;;
Summing over h, it is easy to show that in the general case the probability distribution over possible state
vectors on the input units is given by
This product form is particular to the harmonium structure, and does not hold for general Boltzmann
machines. Product form distribution models have been used for density estimation in Projection Pursuit
[10,6,5] . We shall look further into this relationship in section 5.
3
Discussion of the model
The right hand side of Equation (2) has a simple intuitive interpretation . The ith factor in the product
corresponds to the hidden variable h. and is an increasing function of the dot product between i and the
weight vector of the ith hidden unit. Hence an input vector i will tend to have large probability when it is
in the direction of one of the weight vectors WCi) (i .e. when wei) . i is large). and small probability otherwise.
This is the way that the hidden variables can be seen to exert their" influence"; each corresponds to a.
preferred or "prototypical" direction in space .
The next to the last formula. in Equation (2) shows that the harmonium model can be written as a mixture
of 2m distributions of the form
~ exp (f)W(i) .i + 8('?)h.) ,
Z(h)
i=!
Unsupervised learning of distributions on binary vectors using 2-layer networks
where ii E to, l}m and Z(Ii) is the appropriate normalization factor. It is easily verified that each of these
distributions is in fact a product of n Bernoulli distributions on {+l, -l}, one for each input variable Xj.
Hence the harmonium model can be interpreted as a kind of mixture model. However, the number of
components in the mixture represented by a harmonium is exponential in the number of hidden units.
It is interesting to compare the class of harmonium models to the standard class of models defined by a
mixture of products of Bernoulli distributions. The same bipartite graph described in Figure (1) can be
used to define a standard mixture model. Assign each of the m hidden units a weight vector <.i;) and a
probability Pi such that I:~l Pi
1. To generate an example, choose one of the hidden units according to
te.;J(?) ?I. where Zi
the distribution defined by the Pi'S, and then choose the vector i according to P;(i)
is the appropriate normalization factor so that LIE{?I}" P;(i) 1. We thus get the distribution
=
=
=
m
P(i)
= L Pi e
i=1
W(') I
(3)
Z;
This form for presenting the standard mixture model emphasizes the similarity between this model and the
harmonium model. A vector i will have large probability if the dot product <.ii) ?x is large for some 1 :s i :s m
(so long as Pi is not too small). However, unlike the standard mixture model, the harmonium model allows
more than one hidden variable to be +1 for any generated example. This means that several hidden influences
can combine in the generation of a single example, because several hidden variables can be +1 at the same
time. To see why this is useful, consider the coffee shop example given in the introduction . At any moment
of time it is reasonable to find severa/social groups of people sitting in the shop . The harmonium model will
have a natural representation for this situation, while in order for the standard mixture model to describe
it accurately, a hidden variable has to be assigned to each combination of social groups that is likely to be
found in the shop at the same time. In such cases the harmonium model is exponentially more succinct than
the standard mixture model.
4
Learning by gradient ascent on the log-likelihood
We now suppose that we are given a sample consisting of a set 5 of vectors in {? l}n drawn independently
at random froro some unknown distribution . Our goal is use the sample 5 to find a good model for this
unknown distribution using a harmonium with m hidden units, if possible. The method we investigate here
is the method of maximum likelihood estimation using gradient ascent . The goal of learning is thus reduced
to finding the set of parameters for the harmonium that maximize the (log of the) probability of the set
of examples S. In fact, this gives the standard learning algorithm for general Boltzmann machines. For
a general Boltzmann machine this would require stochastic estimation of the parameters. As stochastic
estimation is very time-consuming, the result is that learning is very slow . In this section we show that
stochastic estimation need not be used for the harmonium model.
= {;{
I), ;(2), ... ,?(N)}, given a particular setting
From (2), the log likelihood of a sample of input vectors 5
{(w(l), 0(1? ?. ..? (w(m) , Oem?~} of the parameters of the model is:
?J
=
.
.
10g-hkehhood(?J)
=Lin Pr(i!?J) = Lm ( L In(l + e'"
-(.)
IES
.=1
(.)
H'
)) -
N In Z .
(4)
IES
Taking the gradient of the log-likelihood results in the following formula for the jth component of wei)
{}
~i) log-likelihood(?) = L x, 1 + e-(W~') 1+9(.1)
wJ
-
IES
N
L
Pr(il?J)x , 1 + e-(W!.IH,(,I)
(5)
IE!:!}"
A similar formula holds for the derivative of the bias term.
The purpose of the clamped and unclamped phases in the Boltzmann machine learning algorithm is to
approximate these two terms. In general, this requires stochastic methods. However , here the clamped term
is easy to calculate, it requires summing a logistic type function over all training examples. The same term
915
916
Freund and Haussler
is obtained by making the mean field approximation for the clamped phase in the general algorithm [15],
which is exact in this case. It is more difficult to compute the sleep phase term, as it is an explicit sum over
the entire input space, and within each term of this sum there is an implicit sum over the entire space of
configurations of hidden units in the factor Pr(i!,p) . However, again taking advantage of the special structure
of the harmonium, We can reduce this sleep phase gradient term to a sum only over the configurations of the
hidden units, yielding for each component of w(i)
8(i)log-likelibood(?l)
8w j
= L: Zj 1 + e-(W~')'I+I('?
les
where
Pr(hl?l)
=
-
N
L
Pr(hl?l)h i tanh(E hkWy?
he{O,I}"
(6)
k=1
exp(L~1 hi9(i? 0;=1 cosh(L~l hiW}i?
.
E.ii'e{o,I}" exp(E~1 h;9(i? OJ: 1 cosh(L~1 h;wJ'})]
Direct computation of (6) is fast for small m in contrast to the case for general Boltzmann machines (we
have performed experiments with m $ 10). However, for large m it is not possible to compute all 2m
terms. There is a way to avoid this exponential explosion if we can assume that a small number of terms
dominate the sums. If, for instance, we assume that the probability that more than k hidden units are
acti ve (+ I) at the same time is negligibly small we can get a good approximation by computing only O( mk)
terms . Alternately, if we are not sure which states of the hidden units have non-negligible probability, we
can dynamically search, as part of the learning process, for the significant terms in the sum . This way we
get an algorithm that is always accurate, and is efficient when the number of significant terms is small. In
the extreme case where we assume that only one hidden unit is active at a time (i.e. k = 1), the harmonium
model essentially reduces to the standard mixture model as discussed is section 3. For larger k, this type of
assumption provides a middle ground between the generality of the harmonium model and the simplicity of
the mixture model.
5
Projection Pursuit methods
A statistical method that has a close relationship with the harmonium model is the Projection Pursuit (PP)
technique [10,6 i5). The use of projection pursuit in the context of neural networks has been studied by
several researchers (e.g. [11]). Most of the work is in exploratory projection pursuit and projection pursuit
regreSSIOn. In this paper we are interested in projection pursuit dellslty estimation. Here PP avoids the
exponential blowup of the standard gradient ascent technique, and also has that advantage that the number
m of hidden units is estimated from the sample as well, rather than being specified in advance.
Projection pursuit density estimation [6] is based on several types of analysis, using the central limit theorem,
that lead to the following general conclusion. If i E R" is a random vector for which the different coordinates
are Independent, and w E R" is a vector from the n dimellsiollal ullit sphere, then the distribution of the
projectIon w? i is close to gaussian for most w. Thus searching for those directions wfor which the projection
of a sample is most non-gaussian is a way for detecting dependencies between the coordinates in high
dimensional distributions . Several "projection-indices" have been studied in the literature for measuring the
"non-gaussianity" of projection, each enhancing different properties of the projected distribution . In order
to find more than one projection direction, several methods of "structure elimination" have been devised .
These methods transform the sample in such a way that the the direction in which non-gaussianity has been
detected appears to be gaussian, thus enabling the algorithm to detect non-gaussian projections that would
otherwise be obscured. The search for a description of the distribution of a sample in terms of its projections
can be formalized in the context of maximal likelihood density estimation [6] . In order to create a formal
relation between the harmonium model and projection pursuit, we define a variant of the model that defines
a density over R" instead of a distribution over {?l}". Based on this form we devise a projection index and
a structure removal method that are the basis of the following learning algorithm (described fully in [4])
? Initialization
Set So to be the input sample.
Set Po to be the initial distribution (Gaussian).
Unsupervised learning of distributions on binary vectors using 2-layer networks
? Iteration
Repeat the following steps for i
1,2 . . . until no single-variable harmonium model has a significantly
higher likelihood than the Gaussian distribution with respect to Si'
1. Perform an estimate-maximize (EM) [2) search on the log-likelihood of a single hidden variable
model on the sample Si-I . Denote by 8i and wei) the parameters found by the search, and create
a new hidden unit with associated binary r. v. hi with these weights and bias.
2. Transform Si-l into Si using the following structure removal procedure.
For each example; E S'_1 compute the probability that the hidden variable h; found in the last
step is 1 on this input:
P(h;
1)
(1 + e-<I.+W(') .I))-I
=
= =
Flip a coin that has probability of "head" equal to P(h;
add; - WCi) to S; else add; to Si.
3. Set Pie;) to be Pi_l(i)Z,l (1
6
= 1).
If the coin turns out "head" then
+ el,+W(').I).
Experimental work
We have carried out several experiments to test the performance of unsupervised learning using the harmonium model. These are not, at this stage, extensive experimental comparisons, but they do provide initial
insights into the issues regarding our learning algorithms and the use of the harmonium model for learning
real world tasks .
The first set of experiments studies two methods for learning the harmonium model. The first is the gradient
ascent method, and the second is the projection pursuit method . The experiments in this set were performed
on synthetically generated data. The input consisted of binary vectors of 64 bits that represent 8 x 8 binary
images. The images are synthesized using a harmonium model with 10 hidden units whose weights were set
as in Figure (2,a) . The ultimate goal of the learning algorithms was to retrieve the model that generated
the data . To measure the quality of the models generated by the algorithms we use three different measures.
The likelihood.of the model, 6 the fraction of correct predictions the model makes when used to predict the
value of a single input bit given all the other bits, and the performance of the model when used to reconstruct
the inpu t from the most probable state of the hidden units. 7 All experiments use a test set and a train set,
each containing 1000 examples. The gradient ascent method used a standard momentum term, and typically
needed about 1000 epochs to stabilize. In the projection pursuit algorithm, 4 iterations of EM per hidden
unit proved sufficient to find a stable solution . The results are summarized in the following table and in
Figure (2) .
gradient ascent for 1000 epochs
proJectIOn pursuit
ProJection pursuit followed by
gradient ascent for 100 epochs
ProJection pursuit followed by
gradient ascent for 1000 epochs
true model
likelihood
train
test
0.399 0.425
0.799 0.802
single bit predictlOn
train
test
0.098
0.100
0.119
0.114
Input reconstructIOn
train
test
0.311
0.338
0.475
0.480
0.411
0.430
0.091
0.089
0.315
0.334
0.377
0.372
0.405
0.404
0.071
0.062
0.082
0.071
0.261
0.252
0.287
0.283
Looking at the table and Figure (2), and taking into account execution times, it appears that gradient
ascent is slow but eventually finds much of the underlying structure in the distribution, although several
of the hidden units (see units 1,2,6,7, counting from the left, in Figure (2,a? have no obvious relation to
the true model, In contrast, PP is fast and finds all of the features of the true model albeit sometimes
aWe present the negation of the log-likelihood, scaled so that the uniform distribution will have likelihood 1.0
1More precisely, for each input unit I we compute the probability p. that it has value +1. Then (or example
(XI, . . . ,1' .. ), we measure - L:~.I log,(1/2 + %,(p, - 1/2? .
917
918
Freund and Haussler
......
.... ..
?????
.
'
(b)
.. .
.???????
.. .... .
.
' "
.:~
.. .... . ? ....? .
.....
? ....... e
. ['::-,":.t~.
. ?cc. ? 0 t:~::~ ~ :~~: :
.. CICJ? ??
. . . ..
..
.. ....
?
..
.. .. . ? -OQ '
......
. .......
,
"
~~:::.,::':::~~~
~~ggg:~~
<d,
??
??
??
D ... ? . . . ?
CID
.C- ? ?? ?
.. ?
..
?
??
..
?
?
...
.
.
...
. . . ? . ? p . . ..............
~~oq2: 2.~
~a.aa.OO '
~"'I .~~?
?
? . ?
? :-oj1:. ? . : :,,~t ~.
?08000a ~ . ::DO:O.
.
..
. . . . . .. ???.0 . ??. ????
.... . ... ...... -0-.
.:.: .. . :.?::8?. :.:?:. ::
???
??
CODa
...
O ? CO? ~~
? ?
~.~ ~ OOOOOJ'
.. . . . . . . DO
.. D?
..
????
? ? ? ??
..
.. .... .. .. ....
.. . . . . . . . ? 0
. . .
......... .....
... .............
: ???
.
.
??
.
E;~:?::: ::.::: : ~ :o~~ : ~~~~O::";'
(c)
.....
...
........
.
..
.
....
.
..... .
: .:; .
'
? .. . . .... .. . . ........... .
' " ...
...0? ???
. ... .. ..
?. . . . . ... .......
'
.. ... ....
~
.. ~ ~.: ~~: ~.:
... . .. ?? .. .?..
..?.
. ' " ....
.... . !O,:
...... ...... . . .. ... .
...
.. . . . .
? . . . ...
..
a..
DOD
:;; '.
-0- .
. . . . D" .
. ........
.
. ....
?.
?? ? .0
..
? ........ D
? 0" ,. .. .. ??
....... . . .
.. ,
aD ???? ?
. ?? gaDa
I
[XJDOQD . . . . . . . . .
?D
.. . . . .
.. . . . . . . .
?
??
????
. ... .
..
.
? .. . .. D ... .
? D..
.
..
.... III
~~~c:-~g
.??
0
.a
?.
. . : .?
. ~~~C?...?
' . . . . . '"
. ? ? ? ? DO?
---
)D....
. . ........ ... ... . ... ... . .. . . .? .. ... . .... ... .
.., .....
?
...
I
... .
.
..
?
?
?
?
?
?
?
-a ??
I
?
D?
..
........
?? . . . . . .
.. ;::~:::
..
.
...
...
I ' '' , ? ? ? ? .
aOODaD
" ???? ?
?
?
??
?
???????
~. .. .
? ..
..
.
?
???
. . . ..
......... ..
..
?~?-:7?: . ::7~~;: ~~::::.::;.~:::.
wrnrmwlli1iJ~lli1l1J
Figure 2: The weight vectors of the models in the synthetic data experiments. Each matrix represents the
64 weights of one hidden unit. The square above the matrix represents the units bias. positive weights are
displayed as full squares and negative weights as empty squares, the area of the square is proportional to
the absolute value of the weight. (a) The weights in the model found by gradient ascent alone. (b) The
weights in the model found by projection pursuit alone. (c) The weights in the model used for generating
the data. (d) The weights in the model found by projection pursuit followed by gradient ascent. For this
last model we also show the histograms of the projection of the examples on the directions defined by those
weight vectors; the bimodality expected from projection pursuit analysis is evident .
in combinations, However, the error measurements show that something is still missing from the models
found by our implementation of PP. Following PP by a gradient ascent phase seems to give the best of both
algoflthms, finding a good approximation after only 140 epochs (40 PP + 100 gradient) and recovering the
true model almost exactly after 1040 epochs.
In the second set of experiments we compare the performance of the harmonium model to that of the mixture
model. The comparison uses real world data extracted from the NIST handwritten data base 8, Examples
are 16 x 16 binary images (see Figure (3)). We use 60 hidden units to model the distribution in both of the
models . Because of the large number of hidden units we cannot use gradient ascent learning and instead
use projection pursuit. For the same reason it was not possible to compute the likelihood of the harmonium
model and only the other two measures of error were used . Each test was run several times to get accuracy
bounds on the measurements. The results are summarized in the following table
Mixture model
HarmOnium model
smgle bIt predictIon
train
test
0.185 ? 0.005 0.258 ? 0.005
0.20 ? 0.01
0.21 ? om
anput reconstructIon
test
train
0.518 ? 0.002 0.715 ? 0.002
0.66 ? 0.03
0.63 ? 0.05
In Figure (4) we show some typical weight vectors found for the mixture model and for the harmonium
model, it is clear that while the mixture model finds weIghts that are some kind of average prototypes of
complete digits, the harmonium model finds weights that correspond to local features such as lines and
contrasts. There is a small but definite improvement in the errors of the harmonium model with respect to
the errors of the mixture model. As the experiments on synthetic data have shown that PP does not reach
INIST Special Database 1, HWDB RelI-l.l, May 1990.
Unsupervised learning of distributions on binary vectors using 2-layer networks
Figure 3: A few examples from the handwritten digits sample.
..............
.:~:I~~~;~L!i:
ll:~;~~! .l~.
!::::~ !:?i:::::
......?....
..
'
~~:
,~:::
::',
' . ' ? ? ? ? ? ? ? ? I"
?
.. "
. ..
???? 0
?? ??
'"
:
...:.?????????
,' ..
......
?.t ?
~::::::.::r::.
.:lIIl"!i:::::;
::::=,II~ : ::: ;:
~. :"~''';:::~::
'0'
eo. ? ?
....
?? 1. so ? ? ?
? ???? ? ??
?????? II . ???
~'~~~~~~'
.'
.. ~
............
.II... ..,
........
:,ii~~:!:~i
........
-, ....
;.II'?~ .~~:a::
III::;:~,;I~:
!!i:H'
.......:i:::;:
..
~
Figure 4: Typical weight vectors found by the mixture model (left) and the harmonium model (right)
optimal solutions by itself we expect the advantage of the harmonium model over the mixture model will
increase further by using improved learning methods. Of course, the harmonium model is a very general
distribution model and is not specifically tuned to the domain of handwritten digit images, thus it cannot be
compared to models specifically developed to capture structures in this domain. However, the experimental
results supports our claim that the harmonium model is a simple and tractable mathematical model for
describing distributions in which several correlation patterns combine to generate each individual example .
References
[1] D. H. Ackley, G. E. Hinton, ILnd T. J. Sejnowski. A learning algorithm for Boltzmann machines. Cognitive
Science, 9:}47-169 , 1985.
(2) A. Dempster, N. Laird, CLnd D. Rubin . Maximum likelihood from incomplete data via the EM algorithm . J.
ROil. Stall!t. Soc. 8, 39:1-38, 1977.
[3] B. Everitt CLnd D. HlLnd . Finite mixture di,tributlon!. Chapman CLnd Hall. 1981.
[4J Y. Freund CLnd D. HlLussler. Unsupervised learning of distributions on binary vectors using two Ia.yer networks .
Technical Report UCSC-CRL-9I-20, Univ. of Calif. Computer ReselLrch Lab, Santa Cruz, CA, 1992 (To appear).
(5) J. H. Friedman . Exploratory projection pursuit . J. Amer. Stot.Assoc., 82(397) :599-608, Mar . 1987.
(6) J . H. Friedman, W.Stuetzle, ILnd A. Schroeder. Projection pursuit density estimation. I. Amer. Stat.Auoc.,
79:599-608, 1984.
(7) H. Ge(ner ILnd J. Peu!. On the probabilistic semantics of connectionist networks. Technical Report CSD-870033,
UCLA Computer Science Deputment, July 1987.
[8J S. Geman and D. Geman. Stochutic reluations, Gibbs distributions ILnd the BayesilLn restoration of imlLges.
IEEE Tron!. on Pattern Analll!i! and Machine Intelligence, 6:721-742, 1984.
[9J J. Hopfield. Neural networks and physical systems with emergent collective computationallLbilities. Proc. Natl .
Acad Sci. USA, 79:2554-2558, Apr. 1982.
[10J P. Huber. Projection pursuit (witb discussion) . Ann. Stat., 13:435-525, 1985.
(11) N. lntrator. FelLture extraction using an unsupervised neural network . In D. Touretzky, J. Ellman, T. Sejnowski,
and G. Hinton, editors, Proceeding! 0/ the 1990 COnnectlOnI!t Model! Summer School, pages 310-318. Morgan
KlLufmlLnn, San Mateo, CA., 1990.
(12) R. M. Neal. Leuning stochastic feedforwlLrd networks. Technical report, Deputment of Computer Science,
UniverSity of Toronto, Nov. 1990.
(13) S. NowlCLn . Ma.ximum likelihood competitive learning. In D. Touretsky, editor, Advance!
Proceumg Sy!teml, volume 2, pages 514-582. Morgan Kau(mlLnn, 1990 .
In
Neurolinformation
[14J J. Peul. Probabi/i!hc Retuoning in Intelligent Sy~tem!. Morgan KlLufmann, 1988.
(15] C. Peterson and J. R. Anderson. A mean field theory learnillg algorithm (or neural networks. Complex SIiItem! ,
1:995-1019,1987.
(16) D. E. Rumelhart CLnd J . L. McClelland. Parallel Distributed Proceulng: ExploratIon!
Cognition . Volume 1,' FoundatIon!. MIT Press, Cambridge, Mass., 1986.
In
the Mlcro!tructure of
919
| 535 |@word middle:1 seems:1 thereby:1 moment:1 initial:2 configuration:4 tuned:1 si:5 yet:1 written:1 cruz:5 alone:2 greedy:1 selected:1 intelligence:1 unacceptably:1 ith:4 detecting:1 provides:1 severa:1 club:2 toronto:1 mathematical:2 ucsc:2 direct:1 consists:2 acti:1 combine:2 pairwise:2 huber:1 expected:1 blowup:1 peu:1 increasing:1 begin:1 moreover:1 formalizing:1 underlying:1 awe:1 mass:1 kind:3 interpreted:1 developed:1 lone:1 unobserved:2 finding:3 usefully:1 exactly:1 scaled:1 assoc:1 unit:50 appear:1 positive:1 negligible:1 understood:1 local:2 ner:1 limit:1 acad:1 exert:1 initialization:1 studied:3 mateo:1 dynamically:1 co:1 limited:1 definite:1 digit:4 stuetzle:1 procedure:1 area:1 universal:2 significantly:2 projection:31 get:4 cannot:2 close:2 put:1 context:2 influence:3 bill:1 customer:1 missing:1 nit:1 independently:2 simplicity:1 formalized:1 insight:1 haussler:4 dominate:1 his:1 retrieve:1 lsee:1 notion:1 exploratory:2 coordinate:2 searching:1 suppose:2 exact:1 us:1 rumelhart:1 approximated:1 ze:1 geman:2 database:1 observed:2 negligibly:1 ackley:1 capture:1 calculate:1 wj:2 connected:1 dempster:1 harmonium:40 creates:1 bipartite:5 basis:1 easily:2 po:1 hopfield:5 emergent:1 represented:4 train:6 univ:1 fast:2 describe:1 sejnowski:2 detected:1 tbat:1 whose:1 heuristic:2 larger:1 valued:1 otherwise:2 reconstruct:1 statistic:1 transform:2 noisy:2 itself:1 laird:1 otber:1 advantage:3 net:2 reconstruction:2 interaction:3 product:7 maximal:1 intuitive:1 description:1 empty:1 generating:1 bimodality:1 object:1 depending:1 illustrate:1 oo:1 stat:2 school:1 minor:1 soc:1 recovering:1 direction:6 correct:1 attribute:1 stochastic:6 exploration:1 elimination:1 require:1 assign:1 probable:1 hold:3 hall:1 ground:1 exp:3 cognition:1 predict:1 mo:1 lm:1 claim:1 purpose:1 estimation:12 proc:1 harmony:1 tanh:1 ellman:1 create:2 weighted:1 minimization:1 mit:1 clearly:1 always:1 sunday:2 gaussian:6 rather:1 avoid:1 unclamped:1 improvement:1 bernoulli:2 likelihood:17 contrast:3 sense:1 detect:1 el:1 unlikely:1 entire:3 typically:1 hidden:46 relation:2 interested:1 hiw:1 semantics:1 overall:1 issue:1 denoted:2 special:5 marginal:1 field:6 equal:1 extraction:1 chapman:1 represents:3 look:1 unsupervised:9 tem:1 connectionist:2 report:3 intelligent:1 few:1 randomly:1 ve:1 individual:1 phase:5 consisting:1 negation:1 attempt:1 friedman:2 interest:1 investigate:1 mixture:21 extreme:1 yielding:1 natl:1 accurate:1 explosion:1 incomplete:1 calif:1 desired:1 obscured:1 mk:1 instance:2 measuring:1 yoav:1 restoration:1 uniform:1 dod:1 too:1 dependency:2 synthetic:3 combined:1 person:2 density:6 ie:1 probabilistic:1 connecting:1 again:1 central:1 containing:1 choose:2 possibly:1 cognitive:1 derivative:1 account:1 summarized:2 stabilize:1 includes:1 gaussianity:2 ad:1 performed:2 closed:1 lab:1 analyze:1 competitive:1 bayes:4 capability:1 parallel:1 om:1 il:1 square:4 accuracy:2 sy:2 correspond:2 sitting:1 handwritten:4 accurately:1 emphasizes:1 researcher:1 cc:1 reach:1 touretzky:1 energy:2 pp:7 obvious:1 associated:2 di:1 proved:1 knowledge:1 exerting:1 appears:2 higher:1 wei:3 improved:1 amer:2 mar:1 generality:1 anderson:1 implicit:1 stage:1 correlation:2 until:1 hand:2 defines:2 logistic:1 quality:1 usa:1 consisted:1 true:4 hence:2 assigned:1 symmetric:1 neal:1 illustrated:1 ll:1 coda:1 presenting:1 evident:1 complete:1 tron:1 bidden:1 image:4 physical:1 exponentially:1 volume:2 discussed:2 he:2 interpretation:1 relating:1 synthesized:1 stot:1 measurement:3 significant:2 cambridge:1 gibbs:1 everitt:1 witb:1 dot:2 stable:1 similarity:1 etc:1 add:2 base:1 something:1 smgle:1 binary:15 arbitrarily:1 devise:1 seen:1 morgan:3 eo:1 maximize:2 july:1 ii:9 full:1 reduces:1 technical:3 long:1 lin:1 sphere:1 devised:1 visit:1 y:3 inpu:1 prediction:2 variant:3 regression:1 essentially:1 enhancing:1 iteration:2 represent:2 normalization:2 sometimes:1 histogram:1 addition:1 else:1 unlike:1 ascent:15 sure:1 induced:2 tend:1 member:2 oq:1 presence:3 synthetically:1 counting:1 iii:2 enough:3 easy:2 xj:1 zi:1 stall:1 reduce:1 regarding:1 prototype:1 golf:2 ultimate:1 ullit:1 useful:1 santa:5 clear:1 cosh:2 induces:1 mcclelland:1 reduced:1 generate:2 zj:1 wci:2 estimated:1 per:1 shall:1 group:5 threshold:2 drawn:1 verified:1 graph:5 fraction:1 tbe:1 wife:1 sum:6 run:1 powerful:2 you:3 i5:1 almost:1 reasonable:1 bit:5 layer:5 hi:2 bound:1 followed:3 summer:1 sleep:2 schroeder:1 precisely:1 ucla:1 relatively:1 according:3 combination:3 em:3 making:1 hl:4 restricted:1 pr:6 computationally:1 equation:2 liil:1 turn:1 eventually:1 describing:1 needed:1 flip:1 ge:1 tractable:1 pursuit:22 appropriate:2 coin:2 gate:1 include:2 oem:1 reli:1 coffee:3 cid:1 gradient:20 sci:1 reason:1 touretsky:1 modeled:3 relationship:2 index:2 difficult:2 pie:1 negative:1 daughter:1 implementation:1 collective:1 boltzmann:16 unknown:5 perform:1 clnd:5 markov:3 enabling:1 nist:1 finite:1 displayed:1 situation:5 hinton:2 looking:1 head:2 david:1 introduced:1 specified:1 extensive:1 connection:4 california:2 learnillg:1 narrow:1 alternately:1 usually:1 below:1 pattern:2 smolensky:1 oj:1 ia:1 natural:2 representing:1 shop:5 carried:1 hm:1 brenda:1 epoch:6 literature:1 removal:2 freund:5 fully:1 expect:2 prototypical:1 interesting:1 generation:1 oj1:1 proportional:1 foundation:1 degree:1 sufficient:1 cicj:1 rubin:1 ggg:1 editor:2 pi:5 course:1 repeat:1 last:3 jth:1 bias:4 allow:1 side:1 formal:1 peterson:1 taking:3 absolute:1 distributed:1 xn:1 fred:2 avoids:1 world:2 computes:1 kau:1 projected:1 san:1 social:4 approximate:1 nov:1 preferred:1 active:1 summing:2 consuming:1 xi:2 search:4 why:1 table:3 nature:1 ca:4 interact:1 hc:1 complex:1 tructure:1 domain:3 apr:1 csd:1 succinct:1 repeated:1 slow:3 analyzable:1 momentum:1 explicit:1 exponential:3 lie:1 clamped:3 formula:3 theorem:1 ximum:1 specific:1 showing:1 intractable:1 ih:1 albeit:1 te:1 execution:1 yer:1 gap:1 easier:1 likely:2 ch:2 corresponds:2 aa:1 extracted:1 ma:1 goal:3 ann:1 lntrator:1 absence:2 crl:1 typical:2 specifically:2 called:3 experimental:4 tendency:1 people:2 support:1 arises:1 |
4,805 | 5,350 | Learning to Discover
Efficient Mathematical Identities
Wojciech Zaremba
Dept. of Computer Science
Courant Institute
New York Unviersity
Karol Kurach
Google Zurich &
Dept. of Computer Science
University of Warsaw
Rob Fergus
Dept. of Computer Science
Courant Institute
New York Unviersity
Abstract
In this paper we explore how machine learning techniques can be applied to the
discovery of efficient mathematical identities. We introduce an attribute grammar framework for representing symbolic expressions. Given a grammar of math
operators, we build trees that combine them in different ways, looking for compositions that are analytically equivalent to a target expression but of lower computational complexity. However, as the space of trees grows exponentially with the
complexity of the target expression, brute force search is impractical for all but
the simplest of expressions. Consequently, we introduce two novel learning approaches that are able to learn from simpler expressions to guide the tree search.
The first of these is a simple n-gram model, the other being a recursive neuralnetwork. We show how these approaches enable us to derive complex identities,
beyond reach of brute-force search, or human derivation.
1
Introduction
Machine learning approaches have proven highly effective for statistical pattern recognition problems, such as those encountered in speech or vision. However, their use in symbolic settings has
been limited. In this paper, we explore how learning can be applied to the discovery of mathematical
identities. Specifically, we propose methods for finding computationally efficient versions of a given
target expression. That is, finding a new expression which computes an identical result to the target,
but has a lower complexity (in time and/or space).
We introduce a framework based on attribute grammars [14] that allows symbolic expressions to be
expressed as a sequence of grammar rules. Brute-force enumeration of all valid rule combinations
allows us to discover efficient versions of the target, including those too intricate to be discovered by
human manipulation. But for complex target expressions this strategy quickly becomes intractable,
due to the exponential number of combinations that must be explored. In practice, a random search
within the grammar tree is used to avoid memory problems, but the chance of finding a matching
solution becomes vanishingly small for complex targets.
To overcome this limitation, we use machine learning to produce a search strategy for the grammar
trees that selectively explores branches likely (under the model) to yield a solution. The training
data for the model comes from solutions discovered for simpler target expressions. We investigate
several different learning approaches. The first group are n-gram models, which learn pairs, triples
etc. of expressions that were part of previously discovered solutions, thus hopefully might be part
of the solution for the current target. We also train a recursive neural network (RNN) that operates
within the grammar trees. This model is first pretrained to learn a continuous representation for
symbolic expressions. Then, using this representation we learn to predict the next grammar rule to
add to the current expression to yield an efficient version of the target.
Through the use of learning, we are able to dramatically widen the complexity and
scope of expressions that can be handled in our framework. We show examples of (i) O n3 target expressions
which can be computed in O n2 time (e.g. see Examples 1 & 2), and (ii) cases where naive eval1
uation of the target would require exponential time, but can be computed in O n2 or O n3 time.
The majority of these examples are too complex to be found manually or by exhaustive search and,
as far as we are aware, are previously undiscovered. All code and evaluation data can be found at
https://github.com/kkurach/math_learning.
In summary our contributions are:
? A novel grammar framework for finding efficient versions of symbolic expressions.
? Showing how machine learning techniques can be integrated into this framework, and
demonstrating how training models on simpler expressions can help which the discovery
of more complex ones.
? A novel application of a recursive neural-network to learn a continuous representation of
mathematical structures, making the symbolic domain accessible to many other learning
approaches.
? The discovery of many new mathematical identities which offer a significant reduction in
computational complexity for certain expressions.
n?m
Example 1: Assume we are given matrices A
, BP? Rm?p
. We wish to compute the
P? R
n Pm Pp
target expression: sum(sum(A*B)), i.e. : n,p AB = i=1 j=1 k=1 Ai,j Bj,k which
naively takes O(nmp) time. Our framework is able to discover an efficient version of the
formula, that computes the same result in O(n(m + p)) time: sum((sum(A, 1) * B)?, 1).
Our framework builds grammar trees that explore valid compositions of expressions from the
grammar, using a search strategy. In this example, the naive strategy of randomly choosing
permissible rules suffices and we can find another tree which matches the target expression in
reasonable time. Below, we show trees for (i) the original expression and (ii) the efficient
formula which avoids the use of a matrix-matrix multiply operation, hence is efficient to
compute.
???????
Example 2: Consider the target expression: sum(sum((A*B)k )), where k = 6. For an
expression of this degree, there are 9785 possible grammar trees and the naive strategy used in
Example 1 breaks down. We therefore learn a search strategy, training a model on successful
trees from simpler expressions, such as those for k = 2, 3, 4, 5. Our learning approaches capture
the common structure within the solutions, evident below, so can find an efficient O (nm)
expression for this target:
k = 2: sum((((((sum(A, 1)) * B) * A) * B)?), 1)
k = 3: sum((((((((sum(A, 1)) * B) * A) * B) * A) * B)?), 1)
k = 4: sum((((((((((sum(A, 1)) * B) * A) * B) * A) * B) * A) * B)?), 1)
k = 5: sum((((((((((((sum(A, 1)) * B) * A) * B) * A) * B) * A) * B) * A) * B)?), 1)
k = 6: sum(((((((((((((sum(A, 1) * B) * A) * B) *A) * B) * A) * B)* A) * B) * A) * B)?), 1)
1.1
Related work
The problem addressed in this paper overlaps with the areas of theorem proving [5, 9, 11], program
induction [18, 28] and probabilistic programming [12, 20]. These domains involve the challenging
issues of undecidability, the halting problem, and a massive space of potential computation. However, we limit our domain to computation of polynomials with fixed degree k, where undecidability
and the halting problem are not present, and the space of computation is manageable (i.e. it grows
exponentially, but not super-exponentially). Symbolic computation engines, such as Maple [6] and
Mathematica [27] are capable of simplifying expressions by collecting terms but do not explicitly
seek versions of lower complexity. Furthermore, these systems are rule based and do not use learning approaches, the major focus of this paper. In general, there has been very little exploration of
statistical machine learning techniques in these fields, one of the few attempts being the recent work
of Bridge et al. [4] who use learning to select between different heuristics for 1st order reasoning. In
contrast, our approach does not use hand-designed heuristics, instead learning them automatically
from the results of simpler expressions.
2
Rule
Input
Matrix-matrix multiply
Matrix-element multiply
Matrix-vector multiply
Matrix transpose
Column sum
Row sum
Column repeat
Row repeat
Element repeat
X
X
X
X
X
X
X
X
X
?
?
?
?
?
?
?
?
?
Output
Rn?m , Y ? Rm?p
Rn?m , Y ? Rn?m
Rn?m , Y ? Rm?1
Rn?m
Rn?m
Rn?m
Rn?1
R1?m
R1?1
Z
Z
Z
Z
Z
Z
Z
Z
Z
?
?
?
?
?
?
?
?
?
Rn?p
Rn?m
Rn?n
Rm?n
Rn?1
R1?m
Rn?m
Rn?m
Rn?m
Computation
Complexity
Z
Z
Z
Z
Z
Z
Z
Z
Z
O (nmp)
O (nm)
O (nm)
O (nm)
O (nm)
O (nm)
O (nm)
O (nm)
O (nm)
=
=
=
=
=
=
=
=
=
X * Y
X .* Y
X * Y
XT
sum(X,1)
sum(X,2)
repmat(X,1,m)
repmat(X,n,1)
repmat(X,n,m)
Table 1: The grammar G used in our experiments.
The attribute grammar, originally developed in 1968 by Knuth [14] in context of compiler construction, has been successfully used as a tool for design and formal specification. In our work, we
apply attribute grammars to a search and optimization problem. This has previously been explored
in a range of domains: from well-known algorithmic problems like knapsack packing [19], through
bioinformatics [26] to music [10]. However, we are not aware of any previous work related to discovering mathematical formulas using grammars, and learning in such framework. The closest work
to ours can be found in [7] which involves searching over the space of algorithms and the grammar
attributes also represent computational complexity.
Classical techniques in natural language processing make extensive use of grammars, for example
to parse sentences and translate between languages. In this paper, we borrow techniques from NLP
and apply them to symbolic computation. In particular, we make use of an n-gram model over
mathematical operations, inspired by n-gram language models. Recursive neural networks have
also been recently used in NLP, for example by Luong et al. [15] and Socher et al. [22, 23], as well
as generic knowledge representation Bottou [2]. In particular, Socher et al. [23], apply them to parse
trees for sentiment analysis. By contrast, we apply them to trees of symbolic expressions. Our work
also has similarities to Bowman [3] who shows that a recursive network can learn simple logical
predicates.
Our demonstration of continuous embeddings for symbolic expressions has parallels with the embeddings used in NLP for words and sentence structure, for example, Collobert & Weston [8], Mnih
& Hinton [17], Turian et al. [25] and Mikolov et al. [16].
2
Problem Statement
Problem Definition: We are given a symbolic target expression T that combines a set of variables V
to produce an output O, i.e. O = T(V). We seek an alternate expression S, such that S(V) = T(V),
but has lower computational complexity, i.e. O (S) < O (T).
In this paper we consider the restricted setting where: (i) T is a homogeneous polynomial of degree
k ? , (ii) V contains a single matrix or vector A and (iii) O is a scalar. While these assumptions may
seem quite restrictive, they still permit a rich family of expressions for our algorithm to explore.
For example, by combining multiple polynomial terms, an efficient Taylor series approximation
can be found for expressions involving trigonometric or exponential operators. Regarding (ii), our
framework can easily handle multiple variables, e.g. Figure 1, which shows expressions using two
matrices, A and B. However, the rest of the paper considers targets based on a single variable. In
Section 8, we discuss these restrictions further.
Notation: We adopt Matlab-style syntax for expressions.
3
Attribute Grammar
We first define an attribute grammar G, which contains a set of mathematical operations, each with
an associated complexity (the attribute). Since T contains exclusively polynomials, we use the
grammar rules listed in Table 1.
Using these rules we can develop trees that combine rules to form expressions involving V, which
for the purposes of this paper is a single matrix A. Since we know T involves expressions of degree
?
I.e. It only contains terms of degree k. E.g. ab + a2 + ac is a homogeneous polynomial of degree 2, but
a2 + b is not homogeneous (b is of degree 1, but a2 is of degree 2).
3
k, each tree must use A exactly k times. Furthermore, since the output is a scalar, each tree must
also compute a scalar quantity. These two constraints limit the depth of each tree. For some targets
T whose complexity is only O (() n3 ), we remove the matrix-matrix multiply rule, thus ensuring
that if any solution is found its complexity is at most O (() n2 ) (see Section 7.2 for more details).
Examples of trees are shown in Fig. 1. The search strategy for determining which rules to combine
is addressed in Section 6.
4
Representation of Symbolic Expressions
We need an efficient way to check if the expression produced by a given tree, or combination of trees
(see Section 5), matches T. The conventional approach would be to perform this check symbolically,
but this is too slow for our purposes and is not amenable to integration with learning methods. We
therefore explore two alternate approaches.
4.1 Numerical Representation
In this representation, each expression is represented by its evaluation of a randomly drawn set of
N points, where N is large (typically 1000). More precisely, for each variable in V, N different
copies are made, each populated with randomly drawn elements. The target expression evaluates
each of these copies, producing a scalar value for each, so yielding a vector t of length N which
uniquely characterizes T. Formally, tn = T(Vn ). We call this numerical vector t the descriptor
of the symbolic expression T. The size of the descriptor N , must be sufficiently large to ensure
that different expressions are not mapped to the same descriptor. Furthermore, when the descriptors
are used in the linear system of Eqn. 5 below, N must also be greater than the number of linear
equations. Any expression S formed by the grammar can be used to evaluate each Vn to produce
another N -length descriptor vector s, which can then be compared to t. If the two match, then
S(V) = T(V).
In practice, using floating point values can result in numerical issues that prevent t and s matching,
even if the two expressions are equivalent. We therefore use an integer-based descriptor in the form
of Zp ? , where p is a large prime number. This prevents both rounding issues as well as numerical
overflow.
4.2 Learned Representation
We now consider how to learn a continuous representation for symbolic expressions, that is learn a
projection ? which maps expressions S to l-dimensional vectors: ?(S) ? Rl . We use a recursive
neural network (RNN) to do this, in a similar fashion to Socher et al. [23] for natural language
and Bowman et al. [3] for logical expressions. This potentially allows many symbolic tasks to be
performed by machine learning techniques, in the same way that the word-vectors (e.g.[8] and [16])
enable many NLP tasks to be posed a learning problems.
We first create a dataset of symbolic expressions, spanning the space of all valid expressions up to
degree k. We then group them into clusters of equivalent expressions (using the numerical representation to check for equality),
and give each
cluster a discrete label 1 . . . C. For example, A, (AT )T
P P
P P
might have label 1, and i j Ai,j , j i Ai,j might have label 2 and so on. For k = 6, the
dataset consists of C = 1687 classes, examples of which are show in Fig. 1. Each class is split
80/20 into train/test sets.
We then train a recursive neural network (RNN) to classify a grammar tree into one of the C clusters.
Instead of representing each grammar rule by its underlying arithmetic, we parameterize it by a
weight matrix or tensor (for operations with one or two inputs, respectively) and use this to learn
the concept of each operation, as part of the network. A vector a ? Rl , where l = 30? is used
to represent each input variable. Working along the grammar tree, each operation in S evolves this
vector via matrix/tensor multiplications (preserving its length) until the entire expression is parsed,
resulting in a single vector ?(S) of length l, which is passed to the classifier to determine the class
of the expression, and hence which other expressions it is equivalent to.
Fig. 2 shows this procedure for two different expressions. Consider the first expression S = (A. ?
A)0 ? sum(A, 2). The first operation here is .?, which is implemented in the RNN by taking the
?
Integers modulo p
This was selected by cross-validation to control the capacity of the RNN, since it directly controls the
number of parameters in the model.
?
4
two (identical) vectors a and applies a weight tensor W3 (of size l ? l ? l, so that the output is
also size l), followed by a rectified-linear non-linearity. The output of this stage is this max((W3 ?
a) ? a, 0). This vector is presented to the next operation, a matrix transpose, whose output is thus
max(W2 ? max((W3 ? a) ? a, 0), 0). Applying the remaining operations produces a final output:
?(S) = max((W4 ? max(W2 ? max((W3 ? a) ? a, 0), 0)) ? max(W1 ? a, 0)). This is presented to a
C-way softmax classifier to predict the class of the expression. The weights W are trained using a
cross-entropy loss and backpropagation.
(((sum((sum((A * (A?)), 1)), 2)) * ((A * (((sum((A?), 1)) * A)?))?)) * A)
(sum(((sum((A * (A?)), 2)) * ((sum((A?), 1)) * (A * ((A?) * A)))), 1))
(((sum(A, 1)) * (((sum(A, 2)) * (sum(A, 1)))?)) * (A * ((A?) * A)))
((((sum((sum((A * (A?)), 1)), 2)) * ((sum((A?), 1)) * (A * ((A?) * A))))?)?)
((sum(A, 1)) * (((A?) * (A * ((A?) * ((sum(A, 2)) * (sum(A, 1))))))?))
((sum((sum((A * (A?)), 1)), 2)) * ((sum((A?), 1)) * (A * ((A?) * A))))
(((sum((sum((A * (A?)), 1)), 2)) * ((sum((A?), 1)) * A)) * ((A?) * A))
((A?) * ((sum(A, 2)) * ((sum((A?), 1)) * (A * (((sum((A?), 1)) * A)?)))))
(sum(((A?) * ((sum(A, 2)) * ((sum((A?), 1)) * (A * ((A?) * A))))), 2))
((((sum(A, 2)) * ((sum((A?), 1)) * A))?) * (A * (((sum((A?), 1)) * A)?)))
(((sum((A?), 1)) * (A * ((A?) * ((sum(A, 2)) * ((sum((A?), 1)) * A)))))?)
((((sum((A?), 1)) * A)?) * ((sum((A?), 1)) * (A * (((sum((A?), 1)) * A)?))))
(((A * ((A?) * ((sum(A, 2)) * ((sum((A?), 1)) * A))))?) * (sum(A, 2)))
(((A?) * ((sum(A, 2)) * ((sum((A?), 1)) * A))) * (sum(((A?) * A), 2)))
(b) Class B
(a) Class A
Figure 1: Samples from two classes of degree k = 6 in our dataset of expressions, used to learn
a continuous representation of symbolic expressions via an RNN. Each line represents a different
expression, but those in the same class are equivalent to one another.
(a) (A. ? A)0 ? sum(A, 2)
(b) (A0 . ? A0 ) ? sum(A, 2)
.
.
Figure 2: Our RNN applied to two expressions. The matrix A is represented by a fixed random
vector a (of length l = 30). Each operation in the expression applies a different matrix (for single
input operations) or tensor (for dual inputs, e.g. matrix-element multiplication) to this vector. After
each operation, a rectified-linear non-linearity is applied. The weight matrices/tensors for each
operation are shared across different expressions. The final vector is passed to a softmax classifier
(not shown) to predict which class they belong to. In this example, both expressions are equivalent,
thus should be mapped to the same class.
When training the RNN, there are several important details that are crucial to obtaining high classification accuracy:
? The weights should be initialized to the identity, plus a small amount of Gaussian noise
added to all elements. The identity allows information to flow the full length of the network,
up to the classifier regardless of its depth [21]. Without this, the RNN overfits badly,
producing test accuracies of ? 1%.
? Rectified linear units work much better in this setting than tanh activation functions.
? We learn using a curriculum [1], starting with the simplest expressions of low degree and
slowly increasing k.
? The weight matrix in the softmax classifier has much larger (?100) learning rate than the
rest of the layers. This encourages the representation to stay still even when targets are
replaced, for example, as we move to harder examples.
? As well as updating the weights of the RNN, we also update the initial value of a (i.e we
backpropagate to the input also).
When the RNN-based representation is employed for identity discovery (see Section 6.3), the vector
?(S) is used directly (i.e. the C-way softmax used in training is removed from the network).
5
Linear Combinations of Trees
For simple targets, an expression that matches the target may be contained within a single grammar
tree. But more complex expressions typically require a linear combination of expressions from
different trees.
5
To handle this, we can use the integer-based descriptors for each tree in a linear system and solve
for a match to the target descriptor (if one exists). Given a set of M trees, each with its own integer
descriptor vector f , we form an M by N linear system of equations and solve it:
F w = t mod Zp
where F = [f1 , . . . , fM ] holds the tree representations, w is the weighting on each of the trees
and t is the target representation. The system is solved using Gaussian elimination, where addition
and multiplication is performed modulo p. The number of solutions can vary: (a) there can be no
solution, which means that no linear combination of the current set of trees can match the target
expression. If all possible trees have been enumerated, then this implies the target expression is
outside the scope of the grammar. (b) There can be one or more solutions, meaning that some
combination of the current set of trees yields a match to the target expression.
6
Search Strategy
So far, we have proposed a grammar which defines the computations that are permitted (like a
programming language grammar), but it gives no guidance as to how explore the space of possible
expressions. Neither do the representations we introduced help ? they simply allow us to determine
if an expression matches or not. We now describe how to efficiently explore the space by learning
which paths are likely to yield a match.
Our framework uses two components: a scheduler, and a strategy. The scheduler is fixed, and traverses space of expressions according to recommendations given by the selected strategy (e.g. ?Random? or ?n-gram? or ?RNN?). The strategy assesses which of the possible grammar rules is likely
to lead to a solution, given the current expression. Starting with the variables V (in our case a single
element A, or more generally, the elements A, B etc.), at each step the scheduler receives scores
for each rule from the strategy and picks the one with the highest score. This continues until the
expression reaches degree k and the tree is complete. We then run the linear solver to see if a linear
combination of the existing set of trees matches the target. If not, the scheduler starts again with
a new tree, initialized with the set of variables V. The n-gram and RNN strategies are learned in
an
fashion, starting with simple target expressions (i.e. those of low degree k, such as
P incremental
T
become training examples used to improve the
ij AA ). Once solutions to these are found, they
P
strategy, needed for tackling harder targets (e.g. ij AAT A).
6.1 Random Strategy
The random strategy involves no learning, thus assigns equal scores to all valid grammar rules,
hence the scheduler randomly picks which expression to try at each step. For simple targets, this
strategy may succeed as the scheduler may stumble upon a match to the target within a reasonable
time-frame. But for complex target expressions of high degree k, the search space is huge and the
approach fails.
6.2 n-gram
In this strategy, we simply count how often subtrees of depth n occur in solutions to previously
solved targets. As the number of different subtrees of depth n is large, the counts become very
sparse as n grows. Due to this, we use a weighted linear combination of the score from all depths
up to n. We found an effective weighting to be 10k , where k is the depth of the tree.
6.3 Recursive Neural Network
Section 4.2 showed how to use an RNN to learn a continuous representation of grammar trees. Recall
that the RNN ? maps expressions to continuous vectors: ?(S) ? Rl . To build a search strategy from
this, we train a softmax layer on top of the RNN to predict which rule should be applied to the current
expression (or expressions, since some rules have two inputs), so that we match the target.
Formally, we have two current branches b1 and b2 (each corresponding to an expression) and wish
to predict the root operation r that joins them (e.g. .?) from among the valid grammar rules (|r|
in total). We first use the previously trained RNN to compute ?(b1 ) and ?(b2 ). These are then
presented to a |r|-way softmax layer (whose weight matrix U is of size 2l ? |r|). If only one branch
exists, then b2 is set to a fixed random vector. The training data for U comes from trees that give
efficient solutions to targets of lower degree k (i.e. simpler targets). Training of the softmax layer
is performed by stochastic gradient descent. We use dropout [13] as the network has a tendency to
overfit and repeat exactly the same expressions for the next value of k. Thus, instead of training on
exactly ?(b1 ) and ?(b2 ), we drop activations as we propagate toward the top of the tree (the same
6
fraction for each depth), which encourages the RNN to capture more local structures. At test time,
the probabilities from the softmax become the scores used by the scheduler.
7
Experiments
We first show results relating to the learned representation for symbolic expressions (Section 4.2).
Then we demonstrate our framework discovering efficient identities. For brevity, the identities discovered are listed in the supplementary material [29].
7.1 Expression Classification using Learned Representation
Table 2 shows the accuracy of the RNN model on expressions of varying degree, ranging from k = 3
to k = 6. The difficulty of the task can be appreciated by looking at the examples in Fig. 1. The low
error rate of ? 5%, despite the use of a simple softmax classifier, demonstrates the effectiveness of
our learned representation.
Test accuracy
Number of classes
Number of expressions
Degree k = 3
100% ? 0%
12
126
Degree k = 4
96.9% ? 1.5%
125
1520
Degree k = 5
94.7% ? 1.0%
970
13038
Degree k = 6
95.3% ? 0.7%
1687
24210
Table 2: Accuracy of predictions using our learned symbolic representation (averaged over 10 different initializations). As the degree increases tasks becomes more challenging, because number of
classes grows, and computation trees become deeper. However our dataset grows larger too (training
uses 80% of examples).
7.2
Efficient Identity Discovery
In our experiments we consider 5 different families of expressions, chosen to fall within the scope
of our grammar rules:
P
P
1. ( AAT )k : A is an Rn?n matrix. The k-th term is i,j (AAT )bk/2c for even k
P
P
P
and i,j (AAT )bk/2c A , for odd k. E.g. for k = 2 : i,j AAT ; for k = 3 : i,j AAT A;
P
for k = 4 : i,j AAT AAT etc. Naive evaluation is O kn3 .
P
n?n
2. P
( (A. ? A)AT )k : A is an RP
matrix and let B = A. ? A. The k-th
Pterm is
T bk/2c
T
bk/2c
(BA
)
for
even
k
and
(BA
B)
,
for
odd
k.
E.g.
for
k
=
2
:
i,j
i,j
i,j (A.?
P
P
T
T
T
A)A ; for k = 3 : i,j (A. ? A)A (A. ? A); for k = 4 : i,j (A. ? A)A (A. ? A)AT etc.
Naive evaluation is O kn3 .
P
3. Symk : Elementary symmetric polynomials. A is a vector in Rn?1 . For k = 1 : i Ai , for
P
P
k = 2 : i<j Ai Aj , for k = 3 : i<j<k Ai Aj Ak , etc. Naive evaluation is O nk .
n?1
4. (RBM-1)
. v is a binary n-vector. The k-th term is:
k : A is a vector in R
P
T
k
(v
A)
.
Naive
evaluation
is O (2n ).
v?{0,1}n
5. (RBM-2)k : Taylor series terms for the partition function of
Pan RBM. A is a matrix in
Rn?n . v and h are a binary n-vectors. The k-th term is v?{0,1}n ,h?{0,1}n (v T Ah)k .
Naive evaluation is O 22n .
Note that (i) for all families, the expressions yield a scalar output; (ii) the families are ordered in
rough order of ?difficulty?; (iii) we are not aware of any
Pprevious exploration
P of these expressions,
except for Symk , which is well studied [24]. For the ( AAT )k and ( (A. ? A)AT )k families
we remove the matrix-multiply rule from the grammar, thus ensuring
that if any solution
is found
it will be efficient since the remaining rules are at most O kn2 , rather than O kn3 . The other
families use the full grammar, given in Table
1. However, the limited set of rules means that if any
solution is found, it can at most be O n3 , rather than exponential in n, as the naive evaluations
would be. For each family, we apply our framework, using the three different search strategies
introduced in Section 6. For each run we impose a fixed cut-off time of 10 minutes? beyond which
we terminate the search. At each value of k, we repeat the experiments 10 times with different
random initializations and count the number of runs that find an efficient solution. Any non-zero
count is deemed a success, since each identity only needs to be discovered once. However, in Fig. 3,
we show the fraction of successful runs, which gives a sense of how quickly the identity was found.
?
Running on a 3Ghz 16-core Intel Xeon. Changing the cut-off has little effect on the plots, since the search
space grows exponentially fast.
7
We start with k = 2 and increase up to k = 15, using the solutions from previous values of k as
training data for the current degree. The search space quickly grows with k, as shown in Table 3.
Fig. 3 shows results for four of the families. We use n-grams for n = 1 . . . 5, as well as the RNN with
two different dropout rates (0.125 and 0.3). The learning approaches generally do much better than
the random strategy for large values of k, with the 3-gram, 4-gram and 5-gram models outperforming
the RNN.
For the first two families, the 3-gram model reliably finds solutions. These solutions involve repetition of a local patterns (e.g. Example 2), which can easily be captured with n-gram models. However, patterns that don?t have a simple repetitive structure are much more difficult to generalize. The
(RBM-2)k family is the most challenging, involving a double exponential sum, and the solutions
have highly complex trees (see supplementary material [29]). In this case, none of our approaches
performed better than the random strategy and no solutions were discovered for k > 5. However,
the k = 5 solution was found by the RNN consistently faster than the random strategy (100 ? 12 vs
438 ? 77 secs).
(
( AA T ) )
k
( A. * A ) A T ) )
(
1
k
Sym k
1
0.8
0.8
0.8
0.8
0.7
0.7
0.7
0.7
0.6
0.6
0.6
RNN0.3
RNN0.13
0.4
1?gram
2?gram
3?gram
4?gram
5?gram
Random
0.3
0.2
0.1
0
2
3
4
5
0.5
RNN0.3
RNN0.13
0.4
1?gram
2?gram
3?gram
4?gram
5?gram
Random
0.3
0.2
0.1
6
7
8
k
9
10 11 12 13 14 15
0
2
3
4
5
0.5
RNN0.3
RNN0.13
0.4
1?gram
2?gram
3?gram
4?gram
5?gram
Random
0.3
0.2
0.1
6
7
8
k
9
10 11 12 13 14 15
0
2
p(Success)
0.9
p(Success)
0.9
0.5
3
4
5
( RBM-1)
1
0.9
p(Success)
p(Success)
1
0.9
0.6
0.5
RNN0.3
RNN0.13
0.4
1?gram
2?gram
3?gram
4?gram
5?gram
Random
0.3
0.2
0.1
6
7
8
k
9
10 11 12 13 14 15
k
0
2
3
4
5
6
7
8
k
9
10 11 12 13 14 15
Figure 3: Evaluation on four different families of expressions. As the degree k increases, we
see that the random strategy consistently fails but the learning approaches can still find solutions
(i.e. p(Success) is non-zero). Best viewed in electronic form.
# Terms ? O n2
3
# Terms ? O n
k=2
39
41
k=3
171
187
k=4
687
790
k=5
2628
3197
k=6
9785
10k+
k = 7 and higher
Out of memory
Table 3: The number of possible expressions for different degrees k.
8
Discussion
We have introduced a framework based on a grammar of symbolic operations for discovering mathematical identities. Through the novel application of learning methods, we have shown how the
exploration of the search space can be learned from previously successful solutions to simpler expressions. This allows us to discover complex expressions that random or brute-force strategies
cannot find (the identities are given in the supplementary material [29]).
Some of the families considered in this paper are close to expressions often encountered in machine
learning. For example, dropout involves an exponential sum over binary masks, which is related to
the RBM-1 family. Also, the partition function of an RBM can be approximated by the RBM-2
family. Hence the identities we have discovered could potentially be used to give a closed-form
version of dropout, or compute the RBM partition function efficiently (i.e. in polynomial time).
Additionally, the automatic nature of our system naturally lends itself to integration with compilers,
or other optimization tools, where it could replace computations with efficient versions thereof.
Our framework could potentially be applied to more general settings, to discover novel formulae in
broader areas of mathematics. To realize this, additional grammar rules, e.g. involving recursion or
trigonometric functions would be needed. However, this would require a more complex scheduler
to determine when to terminate a given grammar tree. Also, it is surprising that a recursive neural
network can generate an effective continuous representation for symbolic expressions. This could
have broad applicability in allowing machine learning tools to be applied to symbolic computation.
The problem addressed in this paper involves discrete search within a combinatorially large space
? a core problem with AI. Our successful use of machine learning to guide the search gives hope
that similar techniques might be effective in other AI tasks where combinatorial explosions are
encountered.
Acknowledgements
The authors would like to thank Facebook and Microsoft Research for their support.
8
References
[1] Y. Bengio, J. Louradour, R. Collobert, and J. Weston. Curriculum learning. In ICML, 2009.
[2] L. Bottou. From machine learning to machine reasoning. Machine Learning, 94(2):133?149, 2014.
[3] S. R. Bowman. Can recursive neural tensor networks learn logical reasoning?
arXiv preprint
arXiv:1312.6192, 2013.
[4] J. P. Bridge, S. B. Holden, and L. C. Paulson. Machine learning for first-order theorem proving. Journal
of Automated Reasoning, 53:141?172, August 2014.
[5] C.-L. Chang. Symbolic logic and mechanical theorem proving. Academic Press, 1973.
[6] B. W. Char, K. O. Geddes, G. H. Gonnet, B. L. Leong, M. B. Monagan, and S. M. Watt. Maple V library
reference manual, volume 199. Springer-verlag New York, 1991.
[7] G. Cheung and S. McCanne. An attribute grammar based framework for machine-dependent computational optimization of media processing algorithms. In ICIP, volume 2, pages 797?801. IEEE, 1999.
[8] R. Collobert and J. Weston. A unified architecture for natural language processing: deep neural networks
with multitask learning. In ICML, 2008.
[9] S. A. Cook. The complexity of theorem-proving procedures. In Proceedings of the third annual ACM
symposium on Theory of computing, pages 151?158. ACM, 1971.
[10] M. Desainte-Catherine and K. Barbar. Using attribute grammars to find solutions for musical equational
programs. ACM SIGPLAN Notices, 29(9):56?63, 1994.
[11] M. Fitting. First-order logic and automated theorem proving. Springer, 1996.
[12] N. Goodman, V. Mansinghka, D. Roy, K. Bonawitz, and D. Tarlow. Church: a language for generative
models. arXiv:1206.3255, 2012.
[13] G. E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, and R. R. Salakhutdinov. Improving neural
networks by preventing co-adaptation of feature detectors. arXiv:1207.0580, 2012.
[14] D. E. Knuth. Semantics of context-free languages. Mathematical systems theory, 2(2):127?145, 1968.
[15] M.-T. Luong, R. Socher, and C. D. Manning. Better word representations with recursive neural networks
for morphology. In CoNLL, 2013.
[16] T. Mikolov, K. Chen, G. Corrado, and J. Dean. Efficient estimation of word representations in vector
space. arXiv:1301.3781, 2013.
[17] A. Mnih and G. E. Hinton. A scalable hierarchical distributed language model. In NIPS, 2009.
[18] P. Nordin. Evolutionary program induction of binary machine code and its applications. Krehl Munster,
1997.
[19] M. ONeill, R. Cleary, and N. Nikolov. Solving knapsack problems with attribute grammars. In Proceedings of the Third Grammatical Evolution Workshop (GEWS04). Citeseer, 2004.
[20] A. Pfeffer. Practical probabilistic programming. In Inductive Logic Programming, pages 2?3. Springer,
2011.
[21] A. M. Saxe, J. L. McClelland, and S. Ganguli. Exact solutions to the nonlinear dynamics of learning in
deep linear neural networks. arXiv:1312.6120, 2013.
[22] R. Socher, C. D. Manning, and A. Y. Ng. Learning continuous phrase representations and syntactic
parsing with recursive neural networks. Proceedings of the NIPS-2010 Deep Learning and Unsupervised
Feature Learning Workshop, pages 1?9, 2010.
[23] R. Socher, A. Perelygin, J. Y. Wu, J. Chuang, C. D. Manning, A. Y. Ng, and C. P. Potts. Recursive deep
models for semantic compositionality over a sentiment treebank. In EMNLP, 2013.
[24] R. P. Stanley. Enumerative combinatorics. Number 49. Cambridge university press, 2011.
[25] J. Turian, L. Ratinov, and Y. Bengio. Word representations: a simple and general method for semisupervised learning. In ACL, 2010.
[26] J. Waldisp?uhl, B. Behzadi, and J.-M. Steyaert. An approximate matching algorithm for finding (sub-)
optimal sequences in s-attributed grammars. Bioinformatics, 18(suppl 2):S250?S259, 2002.
[27] S. Wolfram. The mathematica book, volume 221. Wolfram Media Champaign, Illinois, 1996.
[28] M. L. Wong and K. S. Leung. Evolutionary program induction directed by logic grammars. Evolutionary
Computation, 5(2):143?180, 1997.
[29] W. Zaremba, K. Kurach, and R. Fergus. Learning to discover efficient mathematical identities. arXiv
preprint arXiv:1406.1584 (http://arxiv.org/abs/1406.1584), 2014.
9
| 5350 |@word multitask:1 version:8 manageable:1 polynomial:7 seek:2 propagate:1 simplifying:1 citeseer:1 pick:2 cleary:1 harder:2 reduction:1 initial:1 contains:4 series:2 exclusively:1 score:5 ours:1 undiscovered:1 existing:1 current:8 com:1 surprising:1 activation:2 tackling:1 must:5 parsing:1 realize:1 numerical:5 partition:3 remove:2 designed:1 drop:1 update:1 plot:1 v:1 generative:1 discovering:3 selected:2 cook:1 core:2 wolfram:2 tarlow:1 math:1 traverse:1 org:1 simpler:7 mathematical:11 bowman:3 along:1 become:4 symposium:1 consists:1 combine:4 fitting:1 introduce:3 mask:1 intricate:1 morphology:1 inspired:1 salakhutdinov:1 automatically:1 little:2 enumeration:1 solver:1 increasing:1 becomes:3 discover:6 notation:1 underlying:1 linearity:2 medium:2 developed:1 unified:1 finding:5 impractical:1 collecting:1 zaremba:2 exactly:3 rm:4 classifier:6 demonstrates:1 brute:4 control:2 unit:1 producing:2 aat:9 local:2 limit:2 despite:1 ak:1 path:1 might:4 plus:1 acl:1 initialization:2 studied:1 challenging:3 co:1 limited:2 range:1 averaged:1 directed:1 practical:1 recursive:13 practice:2 backpropagation:1 procedure:2 area:2 rnn:22 w4:1 matching:3 projection:1 word:5 symbolic:23 cannot:1 close:1 operator:2 context:2 applying:1 wong:1 restriction:1 equivalent:6 conventional:1 map:2 dean:1 maple:2 regardless:1 starting:3 assigns:1 rule:23 borrow:1 proving:5 searching:1 handle:2 target:38 construction:1 massive:1 modulo:2 programming:4 homogeneous:3 us:2 exact:1 element:7 roy:1 recognition:1 approximated:1 updating:1 continues:1 cut:2 pfeffer:1 preprint:2 solved:2 capture:2 parameterize:1 barbar:1 removed:1 highest:1 complexity:13 dynamic:1 trained:2 solving:1 upon:1 packing:1 easily:2 represented:2 derivation:1 train:4 fast:1 effective:4 describe:1 choosing:1 outside:1 exhaustive:1 quite:1 heuristic:2 whose:3 posed:1 larger:2 solve:2 supplementary:3 grammar:44 syntactic:1 itself:1 final:2 sequence:2 propose:1 vanishingly:1 adaptation:1 combining:1 translate:1 trigonometric:2 sutskever:1 cluster:3 double:1 r1:3 zp:2 produce:4 karol:1 incremental:1 help:2 derive:1 develop:1 ac:1 ij:2 odd:2 mansinghka:1 implemented:1 involves:5 come:2 implies:1 attribute:11 stochastic:1 exploration:3 human:2 saxe:1 enable:2 char:1 elimination:1 material:3 require:3 suffices:1 f1:1 elementary:1 enumerated:1 hold:1 sufficiently:1 considered:1 warsaw:1 scope:3 predict:5 bj:1 algorithmic:1 major:1 vary:1 adopt:1 a2:3 purpose:2 estimation:1 label:3 tanh:1 combinatorial:1 bridge:2 combinatorially:1 repetition:1 create:1 successfully:1 tool:3 weighted:1 hope:1 rough:1 gaussian:2 super:1 rather:2 avoid:1 varying:1 broader:1 focus:1 consistently:2 potts:1 check:3 contrast:2 sense:1 ganguli:1 dependent:1 leung:1 holden:1 integrated:1 typically:2 entire:1 a0:2 semantics:1 issue:3 dual:1 classification:2 among:1 integration:2 softmax:9 uhl:1 field:1 aware:3 once:2 equal:1 ng:2 manually:1 identical:2 represents:1 broad:1 icml:2 unsupervised:1 few:1 widen:1 randomly:4 undecidability:2 floating:1 replaced:1 microsoft:1 ab:3 attempt:1 huge:1 highly:2 investigate:1 multiply:6 mnih:2 evaluation:9 yielding:1 amenable:1 subtrees:2 capable:1 explosion:1 tree:42 taylor:2 initialized:2 guidance:1 column:2 classify:1 xeon:1 phrase:1 applicability:1 krizhevsky:1 successful:4 predicate:1 rounding:1 too:4 st:1 explores:1 accessible:1 stay:1 kn3:3 probabilistic:2 off:2 quickly:3 eval1:1 w1:1 again:1 nm:9 slowly:1 emnlp:1 luong:2 book:1 style:1 wojciech:1 halting:2 potential:1 b2:4 sec:1 combinatorics:1 explicitly:1 collobert:3 performed:4 break:1 try:1 root:1 closed:1 overfits:1 characterizes:1 compiler:2 start:2 parallel:1 contribution:1 ass:1 formed:1 accuracy:5 descriptor:9 who:2 efficiently:2 musical:1 yield:5 generalize:1 produced:1 none:1 rectified:3 ah:1 detector:1 reach:2 manual:1 facebook:1 definition:1 evaluates:1 pp:1 mathematica:2 thereof:1 naturally:1 associated:1 rbm:9 attributed:1 dataset:4 logical:3 recall:1 knowledge:1 stanley:1 originally:1 courant:2 higher:1 permitted:1 furthermore:3 stage:1 until:2 overfit:1 hand:1 eqn:1 working:1 parse:2 receives:1 nonlinear:1 hopefully:1 kn2:1 google:1 defines:1 gonnet:1 aj:2 grows:7 semisupervised:1 effect:1 concept:1 evolution:1 analytically:1 hence:4 equality:1 inductive:1 symmetric:1 semantic:1 uniquely:1 encourages:2 syntax:1 evident:1 complete:1 demonstrate:1 tn:1 reasoning:4 meaning:1 ranging:1 novel:5 recently:1 common:1 rl:3 exponentially:4 volume:3 belong:1 relating:1 significant:1 composition:2 cambridge:1 ai:8 automatic:1 pm:1 populated:1 mathematics:1 illinois:1 language:9 specification:1 similarity:1 etc:5 add:1 closest:1 own:1 recent:1 showed:1 prime:1 manipulation:1 catherine:1 certain:1 verlag:1 binary:4 success:6 outperforming:1 neuralnetwork:1 preserving:1 captured:1 greater:1 additional:1 impose:1 employed:1 determine:3 corrado:1 nikolov:1 ii:5 branch:3 multiple:2 arithmetic:1 full:2 champaign:1 match:12 faster:1 academic:1 offer:1 cross:2 dept:3 ensuring:2 prediction:1 involving:4 scalable:1 vision:1 arxiv:9 repetitive:1 represent:2 suppl:1 addition:1 addressed:3 crucial:1 permissible:1 w2:2 rest:2 goodman:1 flow:1 mod:1 seem:1 effectiveness:1 call:1 integer:4 leong:1 iii:2 embeddings:2 split:1 bengio:2 automated:2 w3:4 architecture:1 fm:1 regarding:1 kurach:2 expression:98 handled:1 passed:2 sentiment:2 speech:1 york:3 matlab:1 deep:4 dramatically:1 generally:2 involve:2 listed:2 amount:1 mcclelland:1 simplest:2 http:2 generate:1 notice:1 discrete:2 group:2 four:2 demonstrating:1 drawn:2 changing:1 prevent:1 neither:1 symbolically:1 fraction:2 sum:67 ratinov:1 run:4 family:14 reasonable:2 electronic:1 vn:2 wu:1 conll:1 dropout:4 layer:4 followed:1 encountered:3 annual:1 badly:1 occur:1 constraint:1 precisely:1 bp:1 n3:4 sigplan:1 mikolov:2 according:1 alternate:2 combination:9 watt:1 manning:3 across:1 pan:1 rob:1 making:1 evolves:1 restricted:1 computationally:1 equation:2 zurich:1 previously:6 discus:1 count:4 needed:2 know:1 operation:15 permit:1 apply:5 hierarchical:1 generic:1 rp:1 knapsack:2 original:1 chuang:1 top:2 remaining:2 nlp:4 ensure:1 running:1 paulson:1 music:1 parsed:1 restrictive:1 build:3 overflow:1 classical:1 tensor:6 move:1 added:1 quantity:1 strategy:25 evolutionary:3 gradient:1 lends:1 thank:1 mapped:2 capacity:1 majority:1 enumerative:1 considers:1 spanning:1 induction:3 toward:1 unviersity:2 code:2 length:6 demonstration:1 difficult:1 statement:1 potentially:3 perelygin:1 ba:2 design:1 reliably:1 perform:1 allowing:1 pterm:1 descent:1 hinton:3 looking:2 frame:1 discovered:7 rn:18 august:1 compositionality:1 introduced:3 bk:4 pair:1 mechanical:1 extensive:1 sentence:2 icip:1 engine:1 learned:7 nip:2 able:3 beyond:2 below:3 pattern:3 program:4 including:1 memory:2 max:7 overlap:1 natural:3 force:4 difficulty:2 curriculum:2 recursion:1 representing:2 improve:1 github:1 library:1 deemed:1 church:1 naive:9 discovery:6 acknowledgement:1 multiplication:3 determining:1 stumble:1 loss:1 limitation:1 proven:1 triple:1 validation:1 degree:24 treebank:1 row:2 summary:1 repeat:5 transpose:2 copy:2 sym:1 free:1 appreciated:1 guide:2 formal:1 allow:1 deeper:1 institute:2 fall:1 taking:1 sparse:1 ghz:1 distributed:1 overcome:1 depth:7 grammatical:1 gram:33 valid:5 avoids:1 computes:2 rich:1 author:1 made:1 preventing:1 far:2 approximate:1 logic:4 b1:3 fergus:2 don:1 search:20 continuous:9 table:7 bonawitz:1 additionally:1 learn:14 terminate:2 nature:1 obtaining:1 improving:1 bottou:2 complex:10 domain:4 louradour:1 noise:1 turian:2 n2:4 fig:6 intel:1 join:1 fashion:2 slow:1 fails:2 sub:1 scheduler:8 wish:2 exponential:6 weighting:2 third:2 theorem:5 formula:4 uation:1 down:1 xt:1 minute:1 showing:1 explored:2 intractable:1 naively:1 socher:6 exists:2 workshop:2 knuth:2 nmp:2 nk:1 chen:1 backpropagate:1 entropy:1 simply:2 explore:7 likely:3 prevents:1 expressed:1 contained:1 ordered:1 scalar:5 pretrained:1 recommendation:1 applies:2 chang:1 aa:2 springer:3 chance:1 acm:3 weston:3 succeed:1 identity:17 viewed:1 cheung:1 consequently:1 shared:1 replace:1 specifically:1 except:1 operates:1 total:1 tendency:1 selectively:1 select:1 formally:2 support:1 brevity:1 bioinformatics:2 evaluate:1 srivastava:1 |
4,806 | 5,351 | Searching for Higgs Boson Decay Modes
with Deep Learning
Pierre Baldi
Department of Computer Science
University of California, Irvine
Irvine, CA 92617
[email protected]
Peter Sadowski
Department of Computer Science
University of California, Irvine
Irvine, CA 92617
[email protected]
Daniel Whiteson
Department of Physics and Astronomy
University of California, Irvine
Irvine, CA 92617 Address
[email protected]
Abstract
Particle colliders enable us to probe the fundamental nature of matter by observing exotic particles produced by high-energy collisions. Because the experimental measurements from these collisions are necessarily incomplete and imprecise,
machine learning algorithms play a major role in the analysis of experimental
data. The high-energy physics community typically relies on standardized machine learning software packages for this analysis, and devotes substantial effort
towards improving statistical power by hand-crafting high-level features derived
from the raw collider measurements. In this paper, we train artificial neural networks to detect the decay of the Higgs boson to tau leptons on a dataset of 82 million simulated collision events. We demonstrate that deep neural network architectures are particularly well-suited for this task with the ability to automatically
discover high-level features from the data and increase discovery significance.
1
Introduction
The Higgs boson was observed for the first time in 2011-2012 and will be the central object of
study when the Large Hadron Collider (LHC) comes back online to collect new data in 2015. The
observation of the Higgs boson in ZZ, ??, and W W decay modes at the LHC confirms its role
in electroweak symmetry-breaking [1, 2]. However, to establish that it also provides the interaction
which gives mass to the fundamental fermions, it must be demonstrated that the Higgs boson couples
to fermions through direct decay modes. Of the available modes, the most promising is the decay
to a pair of tau leptons (? ), which balances a modest branching ratio with manageable backgrounds.
From the measurements collected in 2011-2012, the LHC collaborations report data consistent with
h ? ? ? decays, but without statistical power to cross the 5? threshold, the standard for claims of
discovery in high-energy physics.
Machine learning plays a major role in processing data at particle colliders. This occurs at two levels:
the online filtering of streaming detector measurements, and the offline analysis of data once it has
been recorded [3], which is the focus of this work. Machine learning classifiers learn to distinguish
between different types of collision events by training on simulated data from sophisticated MonteCarlo programs. Single-hidden-layer, shallow neural networks are one of the primary techniques
used for this analysis, and standardized implementations are included in the prevalent multi-variate
1
? "!
g
??
??
H
#"?
??
?"
g
?
+
#+
? "!
q
??
??
#"?
??
Z
?"
q?
?+
#+
Figure 1: Diagrams for the signal gg ? h ? ? ? ? `? ??`+ ?? and the dominant background
q q? ? Z ? ? ? ? `? ??`+ ??.
analysis software tools used by physicists. Efforts to increase statistical power tend to focus on
developing new features for use with the existing machine learning classifiers ? these high-level
features are non-linear functions of the low-level measurements, derived using knowledge of the
underlying physical processes.
However, the abundance of labeled simulation training data and the complex underlying structure
make this an ideal application for large, deep neural networks. In this work, we show how deep
neural networks can simplify and improve the analysis of high-energy physics data by automatically
learning high-level features from the data. We begin by describing the nature of the data and explaining the difference between the low-level and high-level features used by physicists. Then we
demonstrate that deep neural networks increase the statistical power of this analysis even without
the help of manually-derived high-level features.
2
Data
Collisions of protons at the LHC annhiliate the proton constituents, quarks and gluons. In a small
fraction of collisions, a new heavy state of matter is formed, such as a Higgs or Z boson. Such states
are very unstable and decay rapidly and successively into lighter particles until stable particles are
produced. In the case of Higgs boson production, the process is:
gg ? H ? ? + ? ?
(1)
followed by the subsequent decay of the ? leptons into lighter leptons (e and ?) and pairs of neutrinos
(?), see Fig. 1.
The point of collision is surrounded by concentric layers of detectors that measure the momentum
and direction of the final stable particles. The intermediate states are not observable, such that two
different processes with the same set of final stable particles can be difficult to distinguish. For
example, Figure 1 shows how the process q q? ? Z ? ? + ? ? yields the identical list of particles as
a process that produces the Higgs boson.
The primary approach to distinguish between two processes with identical final state particles is
via the momentum and direction of the particles, which contain information about the identity of
the intermediate state. With perfect measurement resolution and complete information of final state
2
Fraction of Events
Fraction of Events
Fraction of Events
0.4
0.2
0.15
0.1
0.4
0.2
0.2
0.05
0
0
50
100
150
0
-4
200
Lepton 1 p [GeV]
-2
0
0.15
0.1
0.05
0
-4
0.8
0.6
0.4
0
2
4
50
0
100
150
200
Lepton 2 p [GeV]
T
0.2
0.15
0.1
0.05
0.2
-2
0
0
4
Fraction of Events
Fraction of Events
Fraction of Events
0.2
2
Lepton 1 ?
T
0
1
2
Lepton 2 ?
3
4
N jets
0
0
50
100
150
200
Missing Trans. Mom [GeV]
Figure 2: Low-level input features from basic kinematic quantities in `` + p6 T events for simulated
signal (black) and background (red) benchmark events. Shown are the distributions of transverse
momenta (pT ) of each observed particle as well as the imbalance of momentum in the final state.
Momentum angular information for each observed particle is also available to the network, but is
not shown, as the one-dimensional projections have little information.
particles B and C, we could calculate the invariant mass of the short-lived intermediate state A in
the process A ? B + C, via:
m2A = m2B+C = (EB + EC )2 ? |(pB + pC )|2
(2)
However, finite measurement resolution and escaping neutrinos (which are invisible to the detectors)
make it impossible to calculate the intermediate state mass precisely. Instead, the momentum and
direction of the final state particles are studied. This is done using simulated collisions from sophisticated Monte Carlo programs [4, 5, 6] that have been carefully tuned to provide highly faithful
descriptions of the collider data. Machine learning classifiers are trained on the simulated data to
recognize small differences in these processes, then the trained classifiers are used to analyze the
experimental data.
2.1
Low-level features
There are ten low-level features that comprise the essential measurements provided by the detectors:
? The three-dimensional momenta, p, of the charged leptons;
? The imbalance of momentum (6pT ) in the final state transverse to the beam direction, due to
unobserved or mismeasured particles;
? The number and momenta of particle ?jets? due to radiation of gluons or quarks.
Distributions of these features are given in Fig. 2.
2.2
High-level features
There is a vigorous effort in the physics community to construct non-linear combinations of these
low-level features that improve discrimination between Higgs-boson production and Z-boson production. High-level features that have been considered include:
? Axial missing momentum, p6 T ? p`+ `? ;
3
? Scalar sum of the observed momenta, |p`+ | + |p`? | + |6pT | +
P
i
|pjet |;
i
? Relative missing momentum, p6 T if ??(p, p6 T ) ? ?/2, and p6 T ? sin(??(p, p6 T ) if
??(p, p6 T ) < ?/2, where p is the momentum of any charged lepton or jet;
? Difference in lepton azimuthal angles, ??(`+ , `? );
? Difference in lepton polar angles, ??(`+ , `? );
p
? Angular distance between leptons, ?R = (??)2 + (??)2 ;
? Invariant mass of the two leptons, m`+ `? ;
? Missing mass, mMMC [7];
? Sphericity and transverse sphericity;
? Invariant mass of all visible objects (leptons and jets).
Distributions of these features are given in Fig. 3.
3
3.1
Methods
Current approach
Standard machine learning techniques in high-energy physics include methods such as boosted decision trees and single-layer feed-forward neural networks. The TMVA package [8] contains a
standardized implementation of these techniques that is widely-used by physicists. However, we
have found that our own hyperparameter-optimized, single-layer neural networks perform better
than the TMVA implementations. Therefore, we use our own hyperparameter-optimized shallow
neural networks trained on fast graphics processors as a benchmark for comparison.
3.2
Deep learning
Deep neural networks can automatically learn a complex hierarchy of non-linear features from data.
Training deep networks often requires additional computation and a careful selection of hyperparameters, but these difficulties have diminished substantially with the advent of inexpensive graphics processing hardware. We demonstrate here that deep neural networks provide a practical tool
for learning deep feature hierarchies and improving classifier accuracy while reducing the need for
physicists to carefully derive new features by hand. Many exploratory experiments were carried
with different architectures, training protocols, and hyperparameter optimization strategies. Some
of these experiments are still ongoing and, for conciseness, we report only the main results obtained
so far.
3.3
Hyperparameter optimization
Hyperparameters were optimized separately for shallow and deep neural networks. Shallow network
hyperparameters were chosen from combinations of the parameters listed in Table 1, while deep
network hyperparameters were chosen from combinations of those listed in Table 2. These were
selected based on classification performance (cross-entropy error) on the validation set, using the
full set of available features: 10 low-level features plus 15 high-level features. The best architectures
were the largest ones: a deep network with 300 hidden units in each of five hidden layers and an
initial learning rate of 0.03, and a shallow network with 15000 hidden units and an initial learning
rate of 0.01. These neural networks have approximately the same number of tunable parameters,
with 369,301 parameters in the deep network and 405,001 parameters in the shallow network.
Table 1: Hyperparameter options for shallow networks.
Hyperparameter
Options
Hidden units
100, 300, 1000, 15000
Initial learning rate 0.03, 0.01, 0.003, 0.001
4
0.15
0.1
Fraction of Events
Fraction of Events
Fraction of Events
0.2
0.3
0.2
0.1
0.05
0
-100
-50
0
50
0
0
100
100
200
300
400
0
0
500
20
40
60
0.4
0.3
0.2
80
100
MET_rel
Fraction of Events
0.1
0.1
Sum PT
Fraction of Events
Fraction of Events
0.15
0.15
0.05
Axial MET
0.2
0.2
0.3
0.2
0.1
0.05
0
0.1
-2
0
0
0
2
1
2
3
0.15
0.1
0.2
0.15
0.1
1
2
0
0
3
50
100
0.4
0.2
0.15
0.1
0.05
0.2
0.2
0.4
0.6
0.8
0.1
0.2
0.4
0.6
0
0
1
0.8
1
Spher
0.2
0.15
0.1
0.05
100
200
300
Spher T
MMC
Fraction of Events
0
0
0.15
0
0
150
Fraction of Events
Fraction of Events
Fraction of Events
0.6
150
0.2
Pall
T
Pt l1/Ptl2
0.8
100
0.05
0.05
0.05
0
0
50
mll
Fraction of Events
0.2
0
0
4
? R(ll)
Fraction of Events
Fraction of Events
? ?(ll)
0
0
100
200
300
mvis
0.2
0.1
0
-4
-2
0
2
4
? ?(l,l)
Figure 3: Distribution of high-level input features from invariant mass calculations in `?jjb?b events
for simulated signal (black) and background (red) events.
3.4
Training details
The problem is a basic classification task with two classes. The data set is balanced and contains
82 million examples. A validation set of 1 million examples was randomly set aside for tuning the
hyperparameters. Different cross validation strategies were used with little influence on the results
reported since these are obtained in a regime far away from overfitting.
5
Table 2: Hyperparameter options for deep networks.
Hyperparameter
Options
Number of layers
3,4,5,6
Hidden units per layer 100, 300
Initial learning rate
0.03, 0.01, 0.003
The following neural network hyperparameters were predetermined without optimization. The tanh
activation function was used for all hidden units, while the the logistic function was used for the
output. Weights were initialized from a normal distribution with zero mean and standard deviation
0.1 in the first layer, 0.001 in the output layer, and ?1k for all other hidden layers, where k was the
number of units in the previous layer. Gradient computations were made on mini-batches of size
100. A momentum term increased linearly over the first 25 epochs from 0.5 to 0.99, then remained
constant. The learning rate decayed by a factor of 1.0000002 every batch update until it reached a
minimum of 10?6 . All networks were trained for 50 epochs.
Computations were performed using machines with 16 Intel Xeon cores, an NVIDIA Tesla C2070
graphics processor, and 64 GB memory. Training was performed using the Theano and Pylearn2
software libraries [9, 10].
4
Results
The performance of each neural network architecture in terms of the Area Under the signal-rejection
Curve (AUC) is shown in Table 3. As expected, the shallow neural networks (one hidden layer)
perform better with the high-level features than the low-level features alone; the high-level features
were specifically designed to discriminate between the two hypotheses. However, this difference
disappears in deep neural networks, and in fact performance is better with the 10 low-level features
than with the 15 high-level features alone. This, along with the fact that the complete set of features
always performs best, suggests that there is information in the low-level measurements that is not
captured by the high-level features, and that the deep networks are exploiting this information.
Table 3: Comparison of performance for neural network architectures: shallow neural networks
(NN), and deep neural networks (DN) with different numbers of hidden units and layers. Each network architecture was trained on three sets of input features: low-level features, high-level features,
and the complete set of features. The table displays the test set AUC and the expected significance
of a discovery (in units of Gaussian ?) for 100 signal events and 5000 background events with a 5%
relative uncertainty.
AUC
Technique
Low-level High-level Complete
NN 300
0.788
0.792
0.798
NN 1000
0.788
0.792
0.798
NN 15000
0.788
0.792
0.798
DN 3-layer 0.796
0.794
0.801
DN 4-layer 0.797
0.797
0.802
DN 5-layer 0.798
0.798
0.803
DN 6-layer 0.799
0.797
0.803
Discovery significance
Technique
Low-level High-level Complete
NN 15000
1.7?
2.0?
2.0?
DN 6-layer 2.1?
2.2?
2.2?
The best networks are trained with the complete set of features, which provides both the raw measurements and the physicist?s domain knowledge. Figure 4 plots the empirical distribution of predictions (neural network output) for the test samples from each class, and shows how both the shallow
and deep networks trained on the complete feature set are more confident about their correct predictions.
6
NN lo-level
NN hi-level
NN lo+hi-level
0.0
0.2
0.4
0.6
Prediction
0.8
1.0
0.8
1.0
DN lo-level
DN hi-level
DN lo+hi-level
0.0
0.2
0.4
0.6
Prediction
Figure 4: Empirical distribution of predictions for signal events (solid) and background events
(dashed) from the test set.
Figure 5 shows how the AUC translates into discovery significance [11]. On this metric too, the sixlayer deep network trained on the low-level features outperforms the best shallow network (15000
hidden units) trained with the best feature set.
5
Discussion
While deep learning has led to significant advances in computer vision, speech, and natural language
processing, it is clearly useful for a wide range of applications, including a host of applications in the
natural sciences. The problems in high-energy physics are particularly suitable for deep learning,
having large data sets with complex underlying structure. Our results show that deep neural networks
provide a powerful and practical approach to analyzing particle collider data, and that the high-level
features learned from the data by deep neural networks increase the statistical power more than
the common high-level features handcrafted by the physicists. While the improvements may seem
small, they are very significant, especially when considering the billion-dollar cost of accelerator
experiments.
These preliminary experiments demonstrate the advantages of deep neural networks, but we have
not yet pushed the limits of what deep learning can do for this application. The deep architectures
in this work have less than 500,000 parameters and have not even begun to overfit the training data.
7
Shallow networks
Deep networks
all inputs
human-assisted
0.5
raw inputs
1.0
all inputs
1.5
human-assisted
2.0
raw inputs
Discovery significance (?)
2.5
0.0
Figure 5: Comparison of discovery significance for the traditional learning method (left) and the
deep learning method (right) using the low-level features, the high-level features, and the complete
set of features.
Experiments with larger architectures, including ensembles, with a variety of shapes and neuron
types, are currently in progress.
Since the high-level features are derived from the low-level features, it is interesting to note that one
could train a regression neural network to learn this relationship. Such a network would then be able
to predict the physicist-derived features from the low-level measurements. Some of these high-level
features may be more difficult to compute than others, requiring neural networks of a particular size
and depth, and it would be interesting to analyze the complexity of the high-level features in this
way. We are in the process of training such regression networks which could then be incorporated
into a larger prediction architecture, either by freezing their weights, or by allowing them to learn
further. In combination, these deep learning approaches should yield a system ready to sift through
the new Large Hadron Collider data in 2015.
References
[1] Aad, G., Abajyan, T., et al. A particle consistent with the higgs boson observed with the
ATLAS detector at the large hadron collider. Science, 338(6114):1576?1582, December 2012.
ISSN 0036-8075, 1095-9203. doi:10.1126/science.1232005. PMID: 23258888.
[2] Abbaneo, D., Abbiendi, G., et al. A new boson with a mass of 125 GeV observed with the
CMS experiment at the large hadron collider. Science, 338(6114):1569?1575, December 2012.
ISSN 0036-8075, 1095-9203. doi:10.1126/science.1230816. PMID: 23258887.
[3] Denby, B. Neural networks in high energy physics: A ten year perspective. Computer
Physics Communications, 119(23):219?231, June 1999. ISSN 0010-4655. doi:10.1016/
S0010-4655(98)00199-4.
[4] Alwall, J. et al. MadGraph 5 : Going Beyond. JHEP, 1106:128, 2011. doi:10.1007/
JHEP06(2011)128.
[5] Sjostrand, T. et al. PYTHIA 6.4 physics and manual. JHEP, 05:026, 2006.
[6] Ovyn, S., Rouby, X., et al. DELPHES, a framework for fast simulation of a generic collider
experiment. 2009.
[7] Elagin, A., Murat, P., et al. A New Mass Reconstruction Technique for Resonances Decaying
to di-tau. Nucl.Instrum.Meth., A654:481?489, 2011. doi:10.1016/j.nima.2011.07.009.
[8] Hocker, A. et al. TMVA - Toolkit for Multivariate Data Analysis. PoS, ACAT:040, 2007.
[9] Bergstra, J., Breuleux, O., et al. Theano: a CPU and GPU math expression compiler. In
Proceedings of the Python for Scientific Computing Conference (SciPy). Austin, TX, June
2010. Oral Presentation.
8
[10] Goodfellow, I. J., Warde-Farley, D., et al. Pylearn2: a machine learning research library. arXiv
e-print 1308.4214, August 2013.
[11] Cowan, G., Cranmer, K., et al. Asymptotic formulae for likelihood-based tests of new physics.
Eur.Phys.J., C71:1554, 2011. doi:10.1140/epjc/s10052-011-1554-0.
9
| 5351 |@word manageable:1 confirms:1 azimuthal:1 simulation:2 solid:1 initial:4 contains:2 denby:1 daniel:2 tuned:1 outperforms:1 existing:1 current:1 activation:1 yet:1 must:1 gpu:1 visible:1 subsequent:1 predetermined:1 shape:1 designed:1 plot:1 update:1 atlas:1 discrimination:1 aside:1 alone:2 selected:1 short:1 core:1 provides:2 math:1 five:1 along:1 dn:9 direct:1 fermion:2 baldi:1 expected:2 multi:1 automatically:3 little:2 cpu:1 considering:1 begin:1 discover:1 exotic:1 underlying:3 provided:1 mass:9 advent:1 what:1 cm:1 substantially:1 astronomy:1 unobserved:1 every:1 classifier:5 unit:9 limit:1 physicist:7 analyzing:1 approximately:1 black:2 plus:1 eb:1 studied:1 collect:1 suggests:1 range:1 faithful:1 practical:2 area:1 empirical:2 projection:1 imprecise:1 selection:1 impossible:1 influence:1 demonstrated:1 missing:4 charged:2 resolution:2 scipy:1 searching:1 exploratory:1 pt:5 play:2 hierarchy:2 lighter:2 hypothesis:1 goodfellow:1 particularly:2 quark:2 labeled:1 observed:6 role:3 calculate:2 substantial:1 balanced:1 complexity:1 warde:1 trained:9 oral:1 po:1 tx:1 train:2 pmid:2 fast:2 monte:1 doi:6 artificial:1 widely:1 larger:2 ability:1 final:7 online:2 advantage:1 reconstruction:1 interaction:1 uci:3 rapidly:1 description:1 constituent:1 exploiting:1 billion:1 produce:1 perfect:1 object:2 help:1 derive:1 mmc:1 radiation:1 boson:12 axial:2 progress:1 come:1 met:1 collider:10 direction:4 correct:1 human:2 enable:1 preliminary:1 assisted:2 considered:1 ic:1 normal:1 predict:1 claim:1 major:2 polar:1 tanh:1 currently:1 largest:1 tool:2 sphericity:2 clearly:1 always:1 gaussian:1 boosted:1 derived:5 focus:2 june:2 improvement:1 prevalent:1 likelihood:1 detect:1 dollar:1 streaming:1 nn:8 typically:1 hidden:11 going:1 classification:2 resonance:1 once:1 comprise:1 construct:1 having:1 zz:1 manually:1 identical:2 report:2 others:1 simplify:1 jhep:2 randomly:1 lhc:4 recognize:1 mll:1 highly:1 kinematic:1 farley:1 pc:1 modest:1 tree:1 incomplete:1 initialized:1 increased:1 xeon:1 cost:1 deviation:1 graphic:3 too:1 reported:1 confident:1 eur:1 decayed:1 fundamental:2 physic:11 central:1 recorded:1 successively:1 bergstra:1 matter:2 devotes:1 higgs:10 performed:2 observing:1 analyze:2 red:2 reached:1 decaying:1 option:4 compiler:1 formed:1 accuracy:1 ensemble:1 yield:2 raw:4 produced:2 carlo:1 processor:2 detector:5 phys:1 manual:1 inexpensive:1 energy:7 conciseness:1 di:1 couple:1 irvine:6 dataset:1 tunable:1 begun:1 knowledge:2 sophisticated:2 carefully:2 back:1 feed:1 done:1 angular:2 p6:7 until:2 overfit:1 hand:2 freezing:1 mode:4 logistic:1 scientific:1 neutrino:2 contain:1 requiring:1 sin:1 ll:2 branching:1 auc:4 gg:2 complete:8 demonstrate:4 invisible:1 performs:1 l1:1 common:1 physical:1 handcrafted:1 million:3 measurement:11 significant:2 tuning:1 particle:18 language:1 toolkit:1 stable:3 dominant:1 multivariate:1 own:2 perspective:1 nvidia:1 captured:1 minimum:1 additional:1 hadron:4 signal:6 dashed:1 full:1 jet:4 calculation:1 cross:3 host:1 prediction:6 basic:2 regression:2 vision:1 metric:1 arxiv:1 beam:1 background:6 separately:1 diagram:1 breuleux:1 tend:1 cowan:1 december:2 seem:1 ideal:1 intermediate:4 variety:1 variate:1 pfbaldi:1 architecture:9 escaping:1 translates:1 expression:1 gb:1 effort:3 peter:2 speech:1 gev:4 deep:30 useful:1 collision:8 listed:2 ten:2 hardware:1 per:1 hyperparameter:8 threshold:1 pb:1 fraction:20 sum:2 year:1 package:2 angle:2 uncertainty:1 powerful:1 decision:1 pushed:1 layer:18 hi:4 followed:1 distinguish:3 display:1 precisely:1 software:3 department:3 developing:1 combination:4 shallow:12 invariant:4 theano:2 describing:1 montecarlo:1 available:3 probe:1 away:1 generic:1 pierre:1 batch:2 standardized:3 include:2 pall:1 especially:1 establish:1 crafting:1 print:1 quantity:1 occurs:1 strategy:2 primary:2 traditional:1 gradient:1 distance:1 simulated:6 collected:1 unstable:1 issn:3 relationship:1 mini:1 ratio:1 balance:1 difficult:2 m2b:1 implementation:3 lived:1 murat:1 perform:2 allowing:1 imbalance:2 observation:1 neuron:1 benchmark:2 finite:1 acat:1 incorporated:1 communication:1 august:1 community:2 concentric:1 transverse:3 pair:2 optimized:3 proton:2 california:3 learned:1 pylearn2:2 trans:1 address:1 able:1 beyond:1 regime:1 program:2 including:2 tau:3 memory:1 power:5 event:29 suitable:1 difficulty:1 natural:2 nucl:1 meth:1 improve:2 library:2 disappears:1 carried:1 ready:1 mom:1 epoch:2 discovery:7 python:1 relative:2 asymptotic:1 accelerator:1 interesting:2 filtering:1 validation:3 consistent:2 surrounded:1 collaboration:1 heavy:1 production:3 lo:4 austin:1 offline:1 aad:1 explaining:1 wide:1 cranmer:1 curve:1 depth:1 forward:1 made:1 ec:1 far:2 observable:1 overfitting:1 table:7 promising:1 nature:2 learn:4 ca:3 symmetry:1 improving:2 whiteson:1 necessarily:1 complex:3 protocol:1 domain:1 significance:6 main:1 linearly:1 hyperparameters:6 tesla:1 fig:3 intel:1 momentum:14 breaking:1 abundance:1 sadowski:2 remained:1 formula:1 sift:1 list:1 decay:8 essential:1 rejection:1 suited:1 entropy:1 vigorous:1 led:1 scalar:1 relies:1 lepton:15 identity:1 presentation:1 careful:1 towards:1 nima:1 included:1 diminished:1 specifically:1 reducing:1 instrum:1 discriminate:1 experimental:3 ongoing:1 |
4,807 | 5,352 | Semi-supervised Learning with
Deep Generative Models
?
Diederik P. Kingma? , Danilo J. Rezende? , Shakir Mohamed? , Max Welling?
Machine Learning Group, Univ. of Amsterdam, {D.P.Kingma, M.Welling}@uva.nl
?
Google Deepmind, {danilor, shakir}@google.com
Abstract
The ever-increasing size of modern data sets combined with the difficulty of obtaining label information has made semi-supervised learning one of the problems
of significant practical importance in modern data analysis. We revisit the approach to semi-supervised learning with generative models and develop new models that allow for effective generalisation from small labelled data sets to large
unlabelled ones. Generative approaches have thus far been either inflexible, inefficient or non-scalable. We show that deep generative models and approximate
Bayesian inference exploiting recent advances in variational methods can be used
to provide significant improvements, making generative approaches highly competitive for semi-supervised learning.
1
Introduction
Semi-supervised learning considers the problem of classification when only a small subset of the
observations have corresponding class labels. Such problems are of immense practical interest in a
wide range of applications, including image search (Fergus et al., 2009), genomics (Shi and Zhang,
2011), natural language parsing (Liang, 2005), and speech analysis (Liu and Kirchhoff, 2013), where
unlabelled data is abundant, but obtaining class labels is expensive or impossible to obtain for the
entire data set. The question that is then asked is: how can properties of the data be used to improve
decision boundaries and to allow for classification that is more accurate than that based on classifiers
constructed using the labelled data alone. In this paper we answer this question by developing
probabilistic models for inductive and transductive semi-supervised learning by utilising an explicit
model of the data density, building upon recent advances in deep generative models and scalable
variational inference (Kingma and Welling, 2014; Rezende et al., 2014).
Amongst existing approaches, the simplest algorithm for semi-supervised learning is based on a
self-training scheme (Rosenberg et al., 2005) where the the model is bootstrapped with additional
labelled data obtained from its own highly confident predictions; this process being repeated until
some termination condition is reached. These methods are heuristic and prone to error since they
can reinforce poor predictions. Transductive SVMs (TSVM) (Joachims, 1999) extend SVMs with
the aim of max-margin classification while ensuring that there are as few unlabelled observations
near the margin as possible. These approaches have difficulty extending to large amounts of unlabelled data, and efficient optimisation in this setting is still an open problem. Graph-based methods
are amongst the most popular and aim to construct a graph connecting similar observations; label
information propagates through the graph from labelled to unlabelled nodes by finding the minimum
energy (MAP) configuration (Blum et al., 2004; Zhu et al., 2003). Graph-based approaches are sensitive to the graph structure and require eigen-analysis of the graph Laplacian, which limits the scale
to which these methods can be applied ? though efficient spectral methods are now available (Fergus et al., 2009). Neural network-based approaches combine unsupervised and supervised learning
For an updated version of this paper, please see http://arxiv.org/abs/1406.5298
1
by training feed-forward classifiers with an additional penalty from an auto-encoder or other unsupervised embedding of the data (Ranzato and Szummer, 2008; Weston et al., 2012). The Manifold
Tangent Classifier (MTC) (Rifai et al., 2011) trains contrastive auto-encoders (CAEs) to learn the
manifold on which the data lies, followed by an instance of TangentProp to train a classifier that is
approximately invariant to local perturbations along the manifold. The idea of manifold learning
using graph-based methods has most recently been combined with kernel (SVM) methods in the Atlas RBF model (Pitelis et al., 2014) and provides amongst most competitive performance currently
available.
In this paper, we instead, choose to exploit the power of generative models, which recognise the
semi-supervised learning problem as a specialised missing data imputation task for the classification problem. Existing generative approaches based on models such as Gaussian mixture or hidden
Markov models (Zhu, 2006), have not been very successful due to the need for a large number
of mixtures components or states to perform well. More recent solutions have used non-parametric
density models, either based on trees (Kemp et al., 2003) or Gaussian processes (Adams and Ghahramani, 2009), but scalability and accurate inference for these approaches is still lacking. Variational
approximations for semi-supervised clustering have also been explored previously (Li et al., 2009;
Wang et al., 2009).
Thus, while a small set of generative approaches have been previously explored, a generalised and
scalable probabilistic approach for semi-supervised learning is still lacking. It is this gap that we
address through the following contributions:
? We describe a new framework for semi-supervised learning with generative models, employing rich parametric density estimators formed by the fusion of probabilistic modelling
and deep neural networks.
? We show for the first time how variational inference can be brought to bear upon the problem of semi-supervised classification. In particular, we develop a stochastic variational
inference algorithm that allows for joint optimisation of both model and variational parameters, and that is scalable to large datasets.
? We demonstrate the performance of our approach on a number of data sets providing stateof-the-art results on benchmark problems.
? We show qualitatively generative semi-supervised models learn to separate the data classes
(content types) from the intra-class variabilities (styles), allowing in a very straightforward
fashion to simulate analogies of images on a variety of datasets.
2
Deep Generative Models for Semi-supervised Learning
We are faced with data that appear as pairs (X, Y) = {(x1 , y1 ), . . . , (xN , yN )}, with the i-th observation xi ? RD and the corresponding class label yi ? {1, . . . , L}. Observations will have
corresponding latent variables, which we denote by zi . We will omit the index i whenever it is clear
that we are referring to terms associated with a single data point. In semi-supervised classification,
only a subset of the observations have corresponding class labels; we refer to the empirical distribution over the labelled and unlabelled subsets as pel (x, y) and peu (x), respectively. We now develop
models for semi-supervised learning that exploit generative descriptions of the data to improve upon
the classification performance that would be obtained using the labelled data alone.
Latent-feature discriminative model (M1): A commonly used approach is to construct a model
that provides an embedding or feature representation of the data. Using these features, a separate
classifier is thereafter trained. The embeddings allow for a clustering of related observations in a
latent feature space that allows for accurate classification, even with a limited number of labels.
Instead of a linear embedding, or features obtained from a regular auto-encoder, we construct a
deep generative model of the data that is able to provide a more robust set of latent features. The
generative model we use is:
p(z) = N (z|0, I);
p? (x|z) = f (x; z, ?),
(1)
where f (x; z, ?) is a suitable likelihood function (e.g., a Gaussian or Bernoulli distribution) whose
probabilities are formed by a non-linear transformation, with parameters ?, of a set of latent variables z. This non-linear transformation is essential to allow for higher moments of the data to be
captured by the density model, and we choose these non-linear functions to be deep neural networks.
2
Approximate samples from the posterior distribution over the latent variables p(z|x) are used as features to train a classifier that predicts class labels y, such as a (transductive) SVM or multinomial
regression. Using this approach, we can now perform classification in a lower dimensional space
since we typically use latent variables whose dimensionality is much less than that of the observations. These low dimensional embeddings should now also be more easily separable since we
make use of independent latent Gaussian posteriors whose parameters are formed by a sequence of
non-linear transformations of the data. This simple approach results in improved performance for
SVMs, and we demonstrate this in section 4.
Generative semi-supervised model (M2): We propose a probabilistic model that describes the data
as being generated by a latent class variable y in addition to a continuous latent variable z. The data
is explained by the generative process:
p(y) = Cat(y|?);
p(z) = N (z|0, I);
p? (x|y, z) = f (x; y, z, ?),
(2)
where Cat(y|?) is the multinomial distribution, the class labels y are treated as latent variables if
no class label is available and z are additional latent variables. These latent variables are marginally
independent and allow us, in case of digit generation for example, to separate the class specification from the writing style of the digit. As before, f (x; y, z, ?) is a suitable likelihood function,
e.g., a Bernoulli or Gaussian distribution, parameterised by a non-linear transformation of the latent
variables. In our experiments, we choose deep neural networks as this non-linear function. Since
most labels y are unobserved, we integrate over the class of any unlabelled data during the inference process, thus performing classification as inference. Predictions for any missing labels are
obtained from the inferred posterior distribution p? (y|x). This model can also be seen as a hybrid
continuous-discrete mixture model where the different mixture components share parameters.
Stacked generative semi-supervised model (M1+M2): We can combine these two approaches by
first learning a new latent representation z1 using the generative model from M1, and subsequently
learning a generative semi-supervised model M2, using embeddings from z1 instead of the raw data
x. The result is a deep generative model with two layers of stochastic variables: p? (x, y, z1 , z2 ) =
p(y)p(z2 )p? (z1 |y, z2 )p? (x|z1 ), where the priors p(y) and p(z2 ) equal those of y and z above, and
both p? (z1 |y, z2 ) and p? (x|z1 ) are parameterised as deep neural networks.
3
Scalable Variational Inference
3.1 Lower Bound Objective
In all our models, computation of the exact posterior distribution is intractable due to the nonlinear,
non-conjugate dependencies between the random variables. To allow for tractable and scalable
inference and parameter learning, we exploit recent advances in variational inference (Kingma and
Welling, 2014; Rezende et al., 2014). For all the models described, we introduce a fixed-form
distribution q? (z|x) with parameters ? that approximates the true posterior distribution p(z|x). We
then follow the variational principle to derive a lower bound on the marginal likelihood of the model
? this bound forms our objective function and ensures that our approximate posterior is as close as
possible to the true posterior.
We construct the approximate posterior distribution q? (?) as an inference or recognition model,
which has become a popular approach for efficient variational inference (Dayan, 2000; Kingma and
Welling, 2014; Rezende et al., 2014; Stuhlm?uller et al., 2013). Using an inference network, we avoid
the need to compute per data point variational parameters, but can instead compute a set of global
variational parameters ?. This allows us to amortise the cost of inference by generalising between
the posterior estimates for all latent variables through the parameters of the inference network, and
allows for fast inference at both training and testing time (unlike with VEM, in which we repeat
the generalized E-step optimisation for every test data point). An inference network is introduced
for all latent variables, and we parameterise them as deep neural networks whose outputs form the
parameters of the distribution q? (?). For the latent-feature discriminative model (M1), we use a
Gaussian inference network q? (z|x) for the latent variable z. For the generative semi-supervised
model (M2), we introduce an inference model for each of the latent variables z and y, which we we
assume has a factorised form q? (z, y|x) = q? (z|x)q? (y|x), specified as Gaussian and multinomial
distributions respectively.
M1: q? (z|x) = N (z|?? (x), diag(? 2? (x))),
(3)
M2: q? (z|y, x) = N (z|?? (y, x), diag(? 2? (x))); q? (y|x) = Cat(y|? ? (x)),
3
(4)
where ? ? (x) is a vector of standard deviations, ? ? (x) is a probability vector, and the functions
?? (x), ? ? (x) and ? ? (x) are represented as MLPs.
3.1.1
Latent Feature Discriminative Model Objective
For this model, the variational bound J (x) on the marginal likelihood for a single data point is:
log p? (x) ? Eq? (z|x) [log p? (x|z)] ? KL[q? (z|x)kp? (z)] = ?J (x),
(5)
The inference network q? (z|x) (3) is used during training of the model using both the labelled and
unlabelled data sets. This approximate posterior is then used as a feature extractor for the labelled
data set, and the features used for training the classifier.
3.1.2
Generative Semi-supervised Model Objective
For this model, we have two cases to consider. In the first case, the label corresponding to a data
point is observed and the variational bound is a simple extension of equation (5):
log p? (x, y) ? Eq? (z|x,y) [log p? (x|y, z) + log p? (y) + log p(z) ? log q? (z|x, y)] = ?L(x, y), (6)
For the case where the label is missing, it is treated as a latent variable over which we perform
posterior inference and the resulting bound for handling data points with an unobserved label y is:
log p? (x) ? Eq? (y,z|x) [log p? (x|y, z) + log p? (y) + log p(z) ? log q? (y, z|x)]
X
=
q? (y|x)(?L(x, y)) + H(q? (y|x)) = ?U(x).
y
The bound on the marginal likelihood for the entire dataset is now:
X
X
J =
L(x, y) +
U(x)
(x,y)?e
pl
x?e
pu
(7)
(8)
The distribution q? (y|x) (4) for the missing labels has the form a discriminative classifier, and
we can use this knowledge to construct the best classifier possible as our inference model. This
distribution is also used at test time for predictions of any unseen data.
In the objective function (8), the label predictive distribution q? (y|x) contributes only to the second
term relating to the unlabelled data, which is an undesirable property if we wish to use this distribution as a classifier. Ideally, all model and variational parameters should learn in all cases. To remedy
this, we add a classification loss to (8), such that the distribution q? (y|x) also learns from labelled
data. The extended objective function is:
J ? = J + ? ? Epel (x,y) [? log q? (y|x)] ,
(9)
where the hyper-parameter ? controls the relative weight between generative and purely discriminative learning. We use ? = 0.1 ? N in all experiments. While we have obtained this objective function
by motivating the need for all model components to learn at all times, the objective 9 can also be
derived directly using the variational principle by instead performing inference over the parameters
? of the categorical distribution, using a symmetric Dirichlet prior over these parameterss.
3.2
Optimisation
The bounds in equations (5) and (9) provide a unified objective function for optimisation of both
the parameters ? and ? of the generative and inference models, respectively. This optimisation can
be done jointly, without resort to the variational EM algorithm, by using deterministic reparameterisations of the expectations in the objective function, combined with Monte Carlo approximation ?
referred to in previous work as stochastic gradient variational Bayes (SGVB) (Kingma and Welling,
2014) or as stochastic backpropagation (Rezende et al., 2014). We describe the core strategy for the
latent-feature discriminative model M1, since the same computations are used for the generative
semi-supervised model.
When the prior p(z) is a spherical Gaussian distribution p(z) = N (z|0, I) and the variational distribution q? (z|x) is also a Gaussian distribution as in (3), the KL term in equation (5) can be computed
4
Algorithm 1 Learning in model M1
while generativeTraining() do
D ? getRandomMiniBatch()
zi ? qP
? (zi |xi ) ?xi ? D
J ? n J (xi )
?J
(g? , g? ) ? ( ?J
?? , ?? )
(?, ?) ? (?, ?) + ?(g? , g? )
end while
while discriminativeTraining() do
D ? getLabeledRandomMiniBatch()
zi ? q? (zi |xi ) ?{xi , yi } ? D
trainClassifier({zi , yi } )
end while
Algorithm 2 Learning in model M2
while training() do
D ? getRandomMiniBatch()
yi ? q? (yi |xi ) ?{xi , yi } ?
/O
zi ? q? (zi |yi , xi )
J ? ? eq. (9)
?
?L?
(g? , g? ) ? ( ?L
?? , ?? )
(?, ?) ? (?, ?) + ?(g? , g? )
end while
analytically and the log-likelihood term can be rewritten, using the location-scale transformation for
the Gaussian distribution, as:
Eq? (z|x) [log p? (x|z)] = EN (|0,I) log p? (x|?? (x) + ? ? (x) ) ,
(10)
where indicates the element-wise product. While the expectation (10) still cannot be solved
analytically, its gradients with respect to the generative parameters ? and variational parameters ?
can be efficiently computed as expectations of simple gradients:
?{?,?} Eq? (z|x) [log p? (x|z)] = EN (|0,I) ?{?,?} log p? (x|?? (x) + ? ? (x) ) .
(11)
The gradients of the loss (9) for model M2 can be computed by a direct application of the chain rule
and by noting that the conditional bound L(xn , y) contains the same type of terms as the loss (9).
The gradients of the latter can then be efficiently estimated using (11) .
During optimization we use the estimated gradients in conjunction with standard stochastic gradientbased optimization methods such as SGD, RMSprop or AdaGrad (Duchi et al., 2010). This results
in parameter updates of the form: (? t+1 , ?t+1 ) ? (? t , ?t ) + ?t (gt? , gt? ), where ? is a diagonal
preconditioning matrix that adaptively scales the gradients for faster minimization. The training procedure for models M1 and M2 are summarised in algorithms 1 and 2, respectively. Our experimental
results were obtained using AdaGrad.
3.3
Computational Complexity
The overall algorithmic complexity of a single joint update of the parameters (?, ?) for M1 using the
estimator (11) is CM1 = M SCMLP where M is the minibatch size used , S is the number of samples
of the random variate , and CMLP is the cost of an evaluation of the MLPs in the conditional
distributions p? (x|z) and q? (z|x). The cost CMLP is of the form O(KD2 ) where K is the total
number of layers and D is the average dimension of the layers of the MLPs in the model. Training
M1 also requires training a supervised classifier, whose algorithmic complexity, if it is a neural net,
it will have a complexity of the form CMLP .
The algorithmic complexity for M2 is of the form CM2 = LCM1 , where L is the number of labels
and CM1 is the cost of evaluating the gradients of each conditional bound Jy (x), which is the same
as for M1. The stacked generative semi-supervised model has an algorithmic complexity of the
form CM1 + CM2 . But with the advantage that the cost CM2 is calculated in a low-dimensional space
(formed by the latent variables of the model M1 that provides the embeddings).
These complexities make this approach extremely appealing, since they are no more expensive than
alternative approaches based on auto-encoder or neural models, which have the lowest computational complexity amongst existing competitive approaches. In addition, our models are fully probabilistic, allowing for a wide range of inferential queries, which is not possible with many alternative
approaches for semi-supervised learning.
5
Table 1: Benchmark results of semi-supervised classification on MNIST with few labels.
N
100
600
1000
3000
4
NN
25.81
11.44
10.7
6.04
CNN
22.98
7.68
6.45
3.35
TSVM
16.81
6.16
5.38
3.45
CAE
13.47
6.3
4.77
3.22
MTC
12.03
5.13
3.64
2.57
AtlasRBF
8.10 (? 0.95)
?
3.68 (? 0.12)
?
M1+TSVM
11.82 (? 0.25)
5.72 (? 0.049)
4.24 (? 0.07)
3.49 (? 0.04)
M2
11.97 (? 1.71)
4.94 (? 0.13)
3.60 (? 0.56)
3.92 (? 0.63)
M1+M2
3.33 (? 0.14)
2.59 (? 0.05)
2.40 (? 0.02)
2.18 (? 0.04)
Experimental Results
Open source code, with which the most important results and figures can be reproduced, is available at http://github.com/dpkingma/nips14-ssl. For the latest experimental results,
please see http://arxiv.org/abs/1406.5298.
4.1
Benchmark Classification
We test performance on the standard MNIST digit classification benchmark. The data set for semisupervised learning is created by splitting the 50,000 training points between a labelled and unlabelled set, and varying the size of the labelled from 100 to 3000. We ensure that all classes are
balanced when doing this, i.e. each class has the same number of labelled points. We create a number of data sets using randomised sampling to confidence bounds for the mean performance under
repeated draws of data sets.
For model M1 we used a 50-dimensional latent variable z. The MLPs that form part of the generative
and inference models were constructed with two hidden layers, each with 600 hidden units, using
softplus log(1+ex ) activation functions. On top, a transductive SVM (TSVM) was learned on values
of z inferred with q? (z|x). For model M2 we also used 50-dimensional z. In each experiment, the
MLPs were constructed with one hidden layer, each with 500 hidden units and softplus activation
functions. In case of SVHN and NORB, we found it helpful to pre-process the data with PCA.
This makes the model one level deeper, and still optimizes a lower bound on the likelihood of the
unprocessed data.
Table 1 shows classification results. We compare to a broad range of existing solutions in semisupervised learning, in particular to classification using nearest neighbours (NN), support vector
machines on the labelled set (SVM), the transductive SVM (TSVM), and contractive auto-encoders
(CAE). Some of the best results currently are obtained by the manifold tangent classifier (MTC)
(Rifai et al., 2011) and the AtlasRBF method (Pitelis et al., 2014). Unlike the other models in this
comparison, our models are fully probabilistic but have a cost in the same order as these alternatives.
Results: The latent-feature discriminative model (M1) performs better than other models based
on simple embeddings of the data, demonstrating the effectiveness of the latent space in providing
robust features that allow for easier classification. By combining these features with a classification
mechanism directly in the same model, as in the conditional generative model (M2), we are able to
get similar results without a separate TSVM classifier.
However, by far the best results were obtained using the stack of models M1 and M2. This combined model provides accurate test-set predictions across all conditions, and easily outperforms the
previously best methods. We also tested this deep generative model for supervised learning with
all available labels, and obtain a test-set performance of 0.96%, which is among the best published
results for this permutation-invariant MNIST classification task.
4.2
Conditional Generation
The conditional generative model can be used to explore the underlying structure of the data, which
we demonstrate through two forms of analogical reasoning. Firstly, we demonstrate style and content separation by fixing the class label y, and then varying the latent variables z over a range of
values. Figure 1 shows three MNIST classes in which, using a trained model with two latent variables, and the 2D latent variable varied over a range from -5 to 5. In all cases, we see that nearby
regions of latent space correspond to similar writing styles, independent of the class; the left region
represents upright writing styles, while the right-side represents slanted styles.
As a second approach, we use a test image and pass it through the inference network to infer a
value of the latent variables corresponding to that image. We then fix the latent variables z to this
6
(a) Handwriting styles for MNIST obtained by fixing the class label and varying the 2D latent variable z
(b) MNIST analogies
(c) SVHN analogies
Figure 1: (a) Visualisation of handwriting styles learned by the model with 2D z-space. (b,c)
Analogical reasoning with generative semi-supervised models using a high-dimensional z-space.
The leftmost columns show images from the test set. The other columns show analogical fantasies
of x by the generative model, where the latent variable z of each row is set to the value inferred from
the test-set image on the left by the inference network. Each column corresponds to a class label y.
Table 2: Semi-supervised classification on
the SVHN dataset with 1000 labels.
KNN
77.93
(? 0.08)
TSVM
66.55
(? 0.10)
M1+KNN
65.63
(? 0.15)
M1+TSVM
54.33
(? 0.11)
Table 3: Semi-supervised classification on
the NORB dataset with 1000 labels.
M1+M2
36.02
(? 0.10)
KNN
78.71
(? 0.02)
TSVM
26.00
(? 0.06)
M1+KNN
65.39
(? 0.09)
M1+TSVM
18.79
(? 0.05)
value, vary the class label y, and simulate images from the generative model corresponding to that
combination of z and y. This again demonstrate the disentanglement of style from class. Figure 1
shows these analogical fantasies for the MNIST and SVHN datasets (Netzer et al., 2011). The
SVHN data set is a far more complex data set than MNIST, but the model is able to fix the style of
house number and vary the digit that appears in that style well. These generations represent the best
current performance in simulation from generative models on these data sets.
The model used in this way also provides an alternative model to the stochastic feed-forward networks (SFNN) described by Tang and Salakhutdinov (2013). The performance of our model significantly improves on SFNN, since instead of an inefficient Monte Carlo EM algorithm relying on
importance sampling, we are able to perform efficient joint inference that is easy to scale.
4.3
Image Classification
We demonstrate the performance of image classification on the SVHN, and NORB image data sets.
Since no comparative results in the semi-supervised setting exists, we perform nearest-neighbour
and TSVM classification with RBF kernels and compare performance on features generated by
our latent-feature discriminative model to the original features. The results are presented in tables 2
and 3, and we again demonstrate the effectiveness of our approach for semi-supervised classification.
7
4.4
Optimization details
The parameters were initialized by sampling randomly from N (0, 0.0012 I), except for the bias parameters which were initialized as 0. The objectives were optimized using minibatch gradient ascent
until convergence, using a variant of RMSProp with momentum and initialization bias correction, a
constant learning rate of 0.0003, first moment decay (momentum) of 0.1, and second moment decay
of 0.001. For MNIST experiments, minibatches for training were generated by treating normalised
pixel intensities of the images as Bernoulli probabilities and sampling binary images from this distribution. In the M2 model, a weight decay was used corresponding to a prior of (?, ?) ? N (0, I).
5
Discussion and Conclusion
The approximate inference methods introduced here can be easily extended to the model?s parameters, harnessing the full power of variational learning. Such an extension also provides a principled
ground for performing model selection. Efficient model selection is particularly important when the
amount of available data is not large, such as in semi-supervised learning.
For image classification tasks, one area of interest is to combine such methods with convolutional
neural networks that form the gold-standard for current supervised classification methods. Since all
the components of our model are parametrised by neural networks we can readily exploit convolutional or more general locally-connected architectures ? and forms a promising avenue for future
exploration.
A limitation of the models we have presented is that they scale linearly in the number of classes
in the data sets. Having to re-evaluate the generative likelihood for each class during training is an
expensive operation. Potential reduction of the number of evaluations could be achieved by using a
truncation of the posterior mass. For instance we could combine our method with the truncation algorithm suggested by Pal et al. (2005), or by using mechanisms such as error-correcting output codes
(Dietterich and Bakiri, 1995). The extension of our model to multi-label classification problems that
is essential for image-tagging is also possible, but requires similar approximations to reduce the
number of likelihood-evaluations per class.
We have developed new models for semi-supervised learning that allow us to improve the quality of
prediction by exploiting information in the data density using generative models. We have developed
an efficient variational optimisation algorithm for approximate Bayesian inference in these models
and demonstrated that they are amongst the most competitive models currently available for semisupervised learning. We hope that these results stimulate the development of even more powerful
semi-supervised classification methods based on generative models, of which there remains much
scope.
Acknowledgements. We are grateful for feedback from the reviewers. We would also like to
thank the SURFFoundation for the use of the Dutch national e-infrastructure for a significant part of
the experiments.
8
References
Adams, R. P. and Ghahramani, Z. (2009). Archipelago: nonparametric Bayesian semi-supervised learning. In
Proceedings of the International Conference on Machine Learning (ICML).
Blum, A., Lafferty, J., Rwebangira, M. R., and Reddy, R. (2004). Semi-supervised learning using randomized
mincuts. In Proceedings of the International Conference on Machine Learning (ICML).
Dayan, P. (2000). Helmholtz machines and wake-sleep learning. Handbook of Brain Theory and Neural
Network. MIT Press, Cambridge, MA, 44(0).
Dietterich, T. G. and Bakiri, G. (1995). Solving multiclass learning problems via error-correcting output codes.
arXiv preprint cs/9501101.
Duchi, J., Hazan, E., and Singer, Y. (2010). Adaptive subgradient methods for online learning and stochastic
optimization. Journal of Machine Learning Research, 12:2121?2159.
Fergus, R., Weiss, Y., and Torralba, A. (2009). Semi-supervised learning in gigantic image collections. In
Advances in Neural Information Processing Systems (NIPS).
Joachims, T. (1999). Transductive inference for text classification using support vector machines. In Proceeding
of the International Conference on Machine Learning (ICML), volume 99, pages 200?209.
Kemp, C., Griffiths, T. L., Stromsten, S., and Tenenbaum, J. B. (2003). Semi-supervised learning with trees. In
Advances in Neural Information Processing Systems (NIPS).
Kingma, D. P. and Welling, M. (2014). Auto-encoding variational Bayes. In Proceedings of the International
Conference on Learning Representations (ICLR).
Li, P., Ying, Y., and Campbell, C. (2009). A variational approach to semi-supervised clustering. In Proceedings
of the European Symposium on Artificial Neural Networks (ESANN), pages 11 ? 16.
Liang, P. (2005). Semi-supervised learning for natural language. PhD thesis, Massachusetts Institute of Technology.
Liu, Y. and Kirchhoff, K. (2013). Graph-based semi-supervised learning for phone and segment classification.
In Proceedings of Interspeech.
Netzer, Y., Wang, T., Coates, A., Bissacco, A., Wu, B., and Ng, A. Y. (2011). Reading digits in natural images
with unsupervised feature learning. In NIPS workshop on deep learning and unsupervised feature learning.
Pal, C., Sutton, C., and McCallum, A. (2005). Fast inference and learning with sparse belief propagation. In
Advances in Neural Information Processing Systems (NIPS).
Pitelis, N., Russell, C., and Agapito, L. (2014). Semi-supervised learning using an unsupervised atlas. In
Proceddings of the European Conference on Machine Learning (ECML), volume LNCS 8725, pages 565 ?
580.
Ranzato, M. and Szummer, M. (2008). Semi-supervised learning of compact document representations with
deep networks. In Proceedings of the 25th International Conference on Machine Learning (ICML), pages
792?799.
Rezende, D. J., Mohamed, S., and Wierstra, D. (2014). Stochastic backpropagation and approximate inference
in deep generative models. In Proceedings of the International Conference on Machine Learning (ICML),
volume 32 of JMLR W&CP.
Rifai, S., Dauphin, Y., Vincent, P., Bengio, Y., and Muller, X. (2011). The manifold tangent classifier. In
Advances in Neural Information Processing Systems (NIPS), pages 2294?2302.
Rosenberg, C., Hebert, M., and Schneiderman, H. (2005). Semi-supervised self-training of object detection models. In Proceedings of the Seventh IEEE Workshops on Application of Computer Vision
(WACV/MOTION?05).
Shi, M. and Zhang, B. (2011). Semi-supervised learning improves gene expression-based prediction of cancer
recurrence. Bioinformatics, 27(21):3017?3023.
Stuhlm?uller, A., Taylor, J., and Goodman, N. (2013). Learning stochastic inverses. In Advances in neural
information processing systems, pages 3048?3056.
Tang, Y. and Salakhutdinov, R. (2013). Learning stochastic feedforward neural networks. In Advances in
Neural Information Processing Systems (NIPS), pages 530?538.
Wang, Y., Haffari, G., Wang, S., and Mori, G. (2009). A rate distortion approach for semi-supervised conditional random fields. In Advances in Neural Information Processing Systems (NIPS), pages 2008?2016.
Weston, J., Ratle, F., Mobahi, H., and Collobert, R. (2012). Deep learning via semi-supervised embedding. In
Neural Networks: Tricks of the Trade, pages 639?655. Springer.
Zhu, X. (2006). Semi-supervised learning literature survey. Technical report, Computer Science, University of
Wisconsin-Madison.
Zhu, X., Ghahramani, Z., Lafferty, J., et al. (2003). Semi-supervised learning using Gaussian fields and harmonic functions. In Proceddings of the International Conference on Machine Learning (ICML), volume 3,
pages 912?919.
9
| 5352 |@word cnn:1 version:1 open:2 termination:1 cm2:3 simulation:1 contrastive:1 sgd:1 reduction:1 moment:3 liu:2 configuration:1 contains:1 bootstrapped:1 document:1 outperforms:1 existing:4 current:2 com:2 z2:5 activation:2 diederik:1 parsing:1 slanted:1 readily:1 atlas:2 treating:1 update:2 alone:2 generative:40 mccallum:1 bissacco:1 core:1 nips14:1 provides:6 infrastructure:1 node:1 location:1 org:2 firstly:1 zhang:2 wierstra:1 along:1 constructed:3 direct:1 become:1 symposium:1 combine:4 introduce:2 tagging:1 multi:1 brain:1 ratle:1 salakhutdinov:2 relying:1 spherical:1 peu:1 increasing:1 underlying:1 mass:1 lowest:1 pel:1 fantasy:2 deepmind:1 developed:2 unified:1 finding:1 transformation:5 unobserved:2 every:1 classifier:14 control:1 unit:2 omit:1 appear:1 yn:1 gigantic:1 generalised:1 before:1 local:1 limit:1 sgvb:1 sutton:1 encoding:1 approximately:1 initialization:1 limited:1 range:5 contractive:1 practical:2 testing:1 backpropagation:2 digit:5 procedure:1 lncs:1 area:1 empirical:1 significantly:1 inferential:1 confidence:1 pre:1 regular:1 griffith:1 sfnn:2 get:1 cannot:1 close:1 undesirable:1 selection:2 impossible:1 writing:3 map:1 deterministic:1 shi:2 tangentprop:1 missing:4 straightforward:1 latest:1 demonstrated:1 reviewer:1 survey:1 splitting:1 correcting:2 m2:16 estimator:2 rule:1 embedding:4 updated:1 exact:1 trick:1 element:1 helmholtz:1 expensive:3 recognition:1 particularly:1 predicts:1 observed:1 preprint:1 wang:4 solved:1 region:2 ensures:1 connected:1 ranzato:2 russell:1 trade:1 balanced:1 principled:1 rmsprop:2 complexity:8 asked:1 ideally:1 trained:2 grateful:1 solving:1 segment:1 predictive:1 purely:1 upon:3 preconditioning:1 easily:3 joint:3 kirchhoff:2 cae:2 cat:3 represented:1 train:3 univ:1 stacked:2 fast:2 effective:1 describe:2 monte:2 kp:1 query:1 artificial:1 hyper:1 harnessing:1 whose:5 heuristic:1 distortion:1 encoder:3 knn:4 unseen:1 transductive:6 jointly:1 shakir:2 reproduced:1 online:1 sequence:1 advantage:1 net:1 propose:1 product:1 combining:1 gold:1 analogical:4 description:1 scalability:1 exploiting:2 convergence:1 extending:1 comparative:1 adam:2 object:1 derive:1 develop:3 fixing:2 nearest:2 eq:6 esann:1 c:1 stochastic:10 subsequently:1 exploration:1 mtc:3 require:1 fix:2 disentanglement:1 extension:3 pl:1 correction:1 gradientbased:1 ground:1 algorithmic:4 scope:1 vary:2 torralba:1 label:27 currently:3 sensitive:1 create:1 uller:2 minimization:1 brought:1 mit:1 hope:1 gaussian:11 aim:2 avoid:1 varying:3 rosenberg:2 conjunction:1 rezende:6 derived:1 joachim:2 improvement:1 modelling:1 likelihood:9 bernoulli:3 indicates:1 helpful:1 inference:33 dayan:2 nn:2 entire:2 typically:1 hidden:5 visualisation:1 pixel:1 overall:1 classification:31 among:1 dauphin:1 stateof:1 development:1 art:1 ssl:1 marginal:3 equal:1 construct:5 field:2 having:1 ng:1 sampling:4 represents:2 broad:1 kd2:1 unsupervised:5 icml:6 future:1 report:1 few:2 modern:2 randomly:1 neighbour:2 national:1 ab:2 detection:1 interest:2 highly:2 intra:1 evaluation:3 mixture:4 nl:1 parametrised:1 immense:1 chain:1 accurate:4 netzer:2 tree:2 taylor:1 initialized:2 abundant:1 re:1 instance:2 column:3 cost:6 deviation:1 subset:3 successful:1 seventh:1 pal:2 motivating:1 encoders:2 answer:1 dependency:1 combined:4 confident:1 referring:1 density:5 adaptively:1 international:7 randomized:1 probabilistic:6 connecting:1 again:2 thesis:1 choose:3 resort:1 inefficient:2 style:11 li:2 potential:1 factorised:1 collobert:1 doing:1 hazan:1 reached:1 competitive:4 bayes:2 tsvm:11 vem:1 contribution:1 mlps:5 formed:4 convolutional:2 efficiently:2 correspond:1 bayesian:3 raw:1 vincent:1 marginally:1 carlo:2 published:1 whenever:1 energy:1 mohamed:2 associated:1 handwriting:2 dataset:3 popular:2 massachusetts:1 knowledge:1 dimensionality:1 improves:2 campbell:1 appears:1 feed:2 higher:1 supervised:52 danilo:1 follow:1 improved:1 wei:1 done:1 though:1 parameterised:2 until:2 proceddings:2 nonlinear:1 propagation:1 google:2 cm1:3 minibatch:2 quality:1 stimulate:1 semisupervised:3 building:1 dietterich:2 agapito:1 true:2 remedy:1 atlasrbf:2 inductive:1 analytically:2 symmetric:1 during:4 self:2 interspeech:1 please:2 recurrence:1 generalized:1 leftmost:1 demonstrate:7 duchi:2 performs:1 svhn:6 cp:1 motion:1 reasoning:2 image:16 variational:24 wise:1 harmonic:1 recently:1 multinomial:3 qp:1 volume:4 extend:1 m1:22 approximates:1 relating:1 significant:3 refer:1 cambridge:1 rd:1 language:2 specification:1 gt:2 pu:1 add:1 posterior:12 own:1 recent:4 optimizes:1 phone:1 binary:1 yi:7 muller:1 captured:1 minimum:1 additional:3 seen:1 semi:48 full:1 infer:1 technical:1 unlabelled:10 faster:1 jy:1 laplacian:1 ensuring:1 prediction:7 scalable:6 regression:1 variant:1 optimisation:7 expectation:3 vision:1 arxiv:3 dutch:1 kernel:2 represent:1 achieved:1 addition:2 wake:1 source:1 goodman:1 unlike:2 ascent:1 lafferty:2 effectiveness:2 near:1 noting:1 feedforward:1 bengio:1 embeddings:5 easy:1 variety:1 variate:1 zi:8 architecture:1 reduce:1 rifai:3 idea:1 avenue:1 multiclass:1 unprocessed:1 expression:1 pca:1 penalty:1 speech:1 deep:16 clear:1 amount:2 nonparametric:1 locally:1 tenenbaum:1 svms:3 simplest:1 stromsten:1 http:3 coates:1 revisit:1 estimated:2 per:2 summarised:1 discrete:1 group:1 thereafter:1 demonstrating:1 blum:2 imputation:1 rwebangira:1 graph:8 subgradient:1 utilising:1 schneiderman:1 inverse:1 powerful:1 wu:1 separation:1 recognise:1 draw:1 decision:1 layer:5 bound:12 followed:1 sleep:1 nearby:1 simulate:2 extremely:1 performing:3 separable:1 developing:1 combination:1 poor:1 conjugate:1 inflexible:1 describes:1 em:2 across:1 appealing:1 pitelis:3 making:1 explained:1 invariant:2 handling:1 mori:1 equation:3 reddy:1 previously:3 randomised:1 remains:1 mechanism:2 singer:1 tractable:1 end:3 available:7 operation:1 rewritten:1 spectral:1 alternative:4 eigen:1 original:1 top:1 clustering:3 dirichlet:1 ensure:1 archipelago:1 madison:1 exploit:4 ghahramani:3 bakiri:2 objective:11 question:2 parametric:2 strategy:1 diagonal:1 amongst:5 gradient:9 iclr:1 separate:4 reinforce:1 thank:1 manifold:6 considers:1 kemp:2 code:3 index:1 providing:2 ying:1 liang:2 perform:5 allowing:2 observation:8 markov:1 datasets:3 benchmark:4 ecml:1 extended:2 ever:1 variability:1 y1:1 perturbation:1 stack:1 varied:1 intensity:1 inferred:3 introduced:2 pair:1 specified:1 kl:2 z1:7 optimized:1 learned:2 kingma:7 nip:7 address:1 able:4 suggested:1 haffari:1 reading:1 max:2 including:1 belief:1 power:2 suitable:2 difficulty:2 natural:3 treated:2 hybrid:1 zhu:4 scheme:1 improve:3 github:1 technology:1 created:1 categorical:1 auto:6 danilor:1 genomics:1 faced:1 prior:4 text:1 acknowledgement:1 tangent:3 literature:1 adagrad:2 relative:1 wisconsin:1 lacking:2 loss:3 fully:2 bear:1 permutation:1 generation:3 parameterise:1 limitation:1 wacv:1 analogy:3 integrate:1 propagates:1 principle:2 share:1 row:1 prone:1 cancer:1 repeat:1 truncation:2 hebert:1 side:1 allow:8 deeper:1 bias:2 normalised:1 wide:2 institute:1 sparse:1 boundary:1 dimension:1 xn:2 evaluating:1 calculated:1 rich:1 feedback:1 forward:2 made:1 qualitatively:1 commonly:1 adaptive:1 collection:1 far:3 employing:1 welling:7 approximate:8 compact:1 dpkingma:1 gene:1 global:1 handbook:1 generalising:1 norb:3 fergus:3 xi:9 discriminative:8 search:1 latent:36 continuous:2 table:5 promising:1 learn:4 robust:2 obtaining:2 contributes:1 complex:1 european:2 diag:2 uva:1 linearly:1 repeated:2 x1:1 referred:1 en:2 fashion:1 momentum:2 explicit:1 wish:1 lie:1 house:1 jmlr:1 extractor:1 learns:1 tang:2 mobahi:1 explored:2 decay:3 svm:5 fusion:1 essential:2 intractable:1 mnist:9 exists:1 workshop:2 importance:2 phd:1 margin:2 gap:1 easier:1 specialised:1 explore:1 amsterdam:1 springer:1 corresponds:1 minibatches:1 ma:1 weston:2 conditional:7 rbf:2 labelled:13 content:2 generalisation:1 upright:1 except:1 total:1 mincuts:1 pas:1 experimental:3 caes:1 support:2 softplus:2 szummer:2 stuhlm:2 latter:1 bioinformatics:1 evaluate:1 tested:1 ex:1 |
4,808 | 5,353 | Two-Stream Convolutional Networks
for Action Recognition in Videos
Karen Simonyan
Andrew Zisserman
Visual Geometry Group, University of Oxford
{karen,az}@robots.ox.ac.uk
Abstract
We investigate architectures of discriminatively trained deep Convolutional Networks (ConvNets) for action recognition in video. The challenge is to capture
the complementary information on appearance from still frames and motion between frames. We also aim to generalise the best performing hand-crafted features
within a data-driven learning framework.
Our contribution is three-fold. First, we propose a two-stream ConvNet architecture which incorporates spatial and temporal networks. Second, we demonstrate
that a ConvNet trained on multi-frame dense optical flow is able to achieve very
good performance in spite of limited training data. Finally, we show that multitask learning, applied to two different action classification datasets, can be used to
increase the amount of training data and improve the performance on both. Our
architecture is trained and evaluated on the standard video actions benchmarks of
UCF-101 and HMDB-51, where it is competitive with the state of the art. It also
exceeds by a large margin previous attempts to use deep nets for video classification.
1
Introduction
Recognition of human actions in videos is a challenging task which has received a significant amount
of attention in the research community [11, 14, 17, 26]. Compared to still image classification, the
temporal component of videos provides an additional (and important) clue for recognition, as a
number of actions can be reliably recognised based on the motion information. Additionally, video
provides natural data augmentation (jittering) for single image (video frame) classification.
In this work, we aim at extending deep Convolutional Networks (ConvNets) [19], a state-of-theart still image representation [15], to action recognition in video data. This task has recently been
addressed in [14] by using stacked video frames as input to the network, but the results were significantly worse than those of the best hand-crafted shallow representations [20, 26]. We investigate
a different architecture based on two separate recognition streams (spatial and temporal), which
are then combined by late fusion. The spatial stream performs action recognition from still video
frames, whilst the temporal stream is trained to recognise action from motion in the form of dense
optical flow. Both streams are implemented as ConvNets. Decoupling the spatial and temporal nets
also allows us to exploit the availability of large amounts of annotated image data by pre-training
the spatial net on the ImageNet challenge dataset [1]. Our proposed architecture is related to the
two-streams hypothesis [9], according to which the human visual cortex contains two pathways: the
ventral stream (which performs object recognition) and the dorsal stream (which recognises motion);
though we do not investigate this connection any further here.
The rest of the paper is organised as follows. In Sect. 1.1 we review the related work on action
recognition using both shallow and deep architectures. In Sect. 2 we introduce the two-stream
architecture and specify the Spatial ConvNet. Sect. 3 introduces the Temporal ConvNet and in
particular how it generalizes the previous architectures reviewed in Sect. 1.1. A mult-task learning
framework is developed in Sect. 4 in order to allow effortless combination of training data over
multiple datasets. Implementation details are given in Sect. 5, and the performance is evaluated
in Sect. 6 and compared to the state of the art. Our experiments on two challenging datasets (UCF101 [24] and HMDB-51 [16]) show that the two recognition streams are complementary, and our
1
deep architecture significantly outperforms that of [14] and is competitive with the state of the art
shallow representations [20, 21, 26] in spite of being trained on relatively small datasets.
1.1
Related work
Video recognition research has been largely driven by the advances in image recognition methods,
which were often adapted and extended to deal with video data. A large family of video action
recognition methods is based on shallow high-dimensional encodings of local spatio-temporal features. For instance, the algorithm of [17] consists in detecting sparse spatio-temporal interest points,
which are then described using local spatio-temporal features: Histogram of Oriented Gradients
(HOG) [7] and Histogram of Optical Flow (HOF). The features are then encoded into the Bag Of
Features (BoF) representation, which is pooled over several spatio-temporal grids (similarly to spatial pyramid pooling) and combined with an SVM classifier. In a later work [28], it was shown that
dense sampling of local features outperforms sparse interest points.
Instead of computing local video features over spatio-temporal cuboids, state-of-the-art shallow
video representations [20, 21, 26] make use of dense point trajectories. The approach, first introduced in [29], consists in adjusting local descriptor support regions, so that they follow dense
trajectories, computed using optical flow. The best performance in the trajectory-based pipeline
was achieved by the Motion Boundary Histogram (MBH) [8], which is a gradient-based feature,
separately computed on the horizontal and vertical components of optical flow. A combination of
several features was shown to further boost the accuracy. Recent improvements of trajectory-based
hand-crafted representations include compensation of global (camera) motion [10, 16, 26], and the
use of the Fisher vector encoding [22] (in [26]) or its deeper variant [23] (in [21]).
There has also been a number of attempts to develop a deep architecture for video recognition. In
the majority of these works, the input to the network is a stack of consecutive video frames, so the
model is expected to implicitly learn spatio-temporal motion-dependent features in the first layers,
which can be a difficult task. In [11], an HMAX architecture for video recognition was proposed
with pre-defined spatio-temporal filters in the first layer. Later, it was combined [16] with a spatial
HMAX model, thus forming spatial (ventral-like) and temporal (dorsal-like) recognition streams.
Unlike our work, however, the streams were implemented as hand-crafted and rather shallow (3layer) HMAX models. In [4, 18, 25], a convolutional RBM and ISA were used for unsupervised
learning of spatio-temporal features, which were then plugged into a discriminative model for action
classification. Discriminative end-to-end learning of video ConvNets has been addressed in [12]
and, more recently, in [14], who compared several ConvNet architectures for action recognition.
Training was carried out on a very large Sports-1M dataset, comprising 1.1M YouTube videos of
sports activities. Interestingly, [14] found that a network, operating on individual video frames,
performs similarly to the networks, whose input is a stack of frames. This might indicate that
the learnt spatio-temporal features do not capture the motion well. The learnt representation, finetuned on the UCF-101 dataset, turned out to be 20% less accurate than hand-crafted state-of-the-art
trajectory-based representation [20, 27].
Our temporal stream ConvNet operates on multiple-frame dense optical flow, which is typically
computed in an energy minimisation framework by solving for a displacement field (typically at
multiple image scales). We used a popular method of [2], which formulates the energy based on
constancy assumptions for intensity and its gradient, as well as smoothness of the displacement field.
Recently, [30] proposed an image patch matching scheme, which is reminiscent of deep ConvNets,
but does not incorporate learning.
2
Two-stream architecture for video recognition
Video can naturally be decomposed into spatial and temporal components. The spatial part, in the
form of individual frame appearance, carries information about scenes and objects depicted in the
video. The temporal part, in the form of motion across the frames, conveys the movement of the
observer (the camera) and the objects. We devise our video recognition architecture accordingly,
dividing it into two streams, as shown in Fig. 1. Each stream is implemented using a deep ConvNet,
softmax scores of which are combined by late fusion. We consider two fusion methods: averaging
and training a multi-class linear SVM [6] on stacked L2 -normalised softmax scores as features.
Spatial stream ConvNet operates on individual video frames, effectively performing action recognition from still images. The static appearance by itself is a useful clue, since some actions are
2
Spatial stream ConvNet
single frame
conv1
conv2
conv3
conv4
conv5
full6
full7
7x7x96
stride 2
norm.
pool 2x2
5x5x256
stride 2
norm.
pool 2x2
3x3x512
stride 1
3x3x512
stride 1
3x3x512
stride 1
pool 2x2
4096
dropout
2048
dropout
softmax
class
score
fusion
Temporal stream ConvNet
input
video
multi-frame
optical flow
conv1
conv2
conv3
conv4
conv5
full6
full7
7x7x96
stride 2
norm.
pool 2x2
5x5x256
stride 2
pool 2x2
3x3x512
stride 1
3x3x512
stride 1
3x3x512
stride 1
pool 2x2
4096
dropout
2048
dropout
softmax
Figure 1: Two-stream architecture for video classification.
strongly associated with particular objects. In fact, as will be shown in Sect. 6, action classification
from still frames (the spatial recognition stream) is fairly competitive on its own. Since a spatial
ConvNet is essentially an image classification architecture, we can build upon the recent advances
in large-scale image recognition methods [15], and pre-train the network on a large image classification dataset, such as the ImageNet challenge dataset. The details are presented in Sect. 5. Next, we
describe the temporal stream ConvNet, which exploits motion and significantly improves accuracy.
3
Optical flow ConvNets
In this section, we describe a ConvNet model, which forms the temporal recognition stream of our
architecture (Sect. 2). Unlike the ConvNet models, reviewed in Sect. 1.1, the input to our model is
formed by stacking optical flow displacement fields between several consecutive frames. Such input
explicitly describes the motion between video frames, which makes the recognition easier, as the
network does not need to estimate motion implicitly. We consider several variations of the optical
flow-based input, which we describe below.
(a)
(b)
(c)
(d)
(e)
Figure 2: Optical flow. (a),(b): a pair of consecutive video frames with the area around a moving hand outlined with a cyan rectangle. (c): a close-up of dense optical flow in the outlined area;
(d): horizontal component dx of the displacement vector field (higher intensity corresponds to positive values, lower intensity to negative values). (e): vertical component dy . Note how (d) and (e)
highlight the moving hand and bow. The input to a ConvNet contains multiple flows (Sect. 3.1).
3.1
ConvNet input configurations
Optical flow stacking. A dense optical flow can be seen as a set of displacement vector fields dt
between the pairs of consecutive frames t and t + 1. By dt (u, v) we denote the displacement vector
at the point (u, v) in frame t, which moves the point to the corresponding point in the following
frame t + 1. The horizontal and vertical components of the vector field, dxt and dyt , can be seen
as image channels (shown in Fig. 2), well suited to recognition using a convolutional network. To
represent the motion across a sequence of frames, we stack the flow channels dx,y
of L consecutive
t
frames to form a total of 2L input channels. More formally, let w and h be the width and height
of a video; a ConvNet input volume I? ? Rw?h?2L for an arbitrary frame ? is then constructed as
follows:
I? (u, v, 2k ? 1) = dx?+k?1 (u, v),
I? (u, v, 2k) =
dy? +k?1 (u, v),
(1)
u = [1; w], v = [1; h], k = [1; L].
For an arbitrary point (u, v), the channels I? (u, v, c), c = [1; 2L] encode the motion at that point
over a sequence of L frames (as illustrated in Fig. 3-left).
Trajectory stacking. An alternative motion representation, inspired by the trajectory-based descriptors [29], replaces the optical flow, sampled at the same locations across several frames, with
3
the flow, sampled along the motion trajectories. In this case, the input volume I? , corresponding to
a frame ? , takes the following form:
I? (u, v, 2k ? 1) = dx?+k?1 (pk ),
I? (u, v, 2k) = dy? +k?1 (pk ),
(2)
u = [1; w], v = [1; h], k = [1; L].
where pk is the k-th point along the trajectory, which starts at the location (u, v) in the frame ? and
is defined by the following recurrence relation:
p1 = (u, v);
pk = pk?1 + d? +k?2 (pk?1 ), k > 1.
Compared to the input volume representation (1), where the channels I? (u, v, c) store the displacement vectors at the locations (u, v), the input volume (2) stores the vectors sampled at the locations
pk along the trajectory (as illustrated in Fig. 3-right).
input volume channels
at point
input volume channels
at point
Figure 3: ConvNet input derivation from the multi-frame optical flow. Left: optical flow stacking (1) samples the displacement vectors d at the same location in multiple frames. Right: trajectory
stacking (2) samples the vectors along the trajectory. The frames and the corresponding displacement vectors are shown with the same colour.
Bi-directional optical flow. Optical flow representations (1) and (2) deal with the forward optical
flow, i.e. the displacement field dt of the frame t specifies the location of its pixels in the following
frame t + 1. It is natural to consider an extension to a bi-directional optical flow, which can be
obtained by computing an additional set of displacement fields in the opposite direction. We then
construct an input volume I? by stacking L/2 forward flows between frames ? and ? +L/2 and L/2
backward flows between frames ? ? L/2 and ? . The input I? thus has the same number of channels
(2L) as before. The flows can be represented using either of the two methods (1) and (2).
Mean flow subtraction. It is generally beneficial to perform zero-centering of the network input,
as it allows the model to better exploit the rectification non-linearities. In our case, the displacement
vector field components can take on both positive and negative values, and are naturally centered in
the sense that across a large variety of motions, the movement in one direction is as probable as the
movement in the opposite one. However, given a pair of frames, the optical flow between them can
be dominated by a particular displacement, e.g. caused by the camera movement. The importance
of camera motion compensation has been previously highlighted in [10, 26], where a global motion
component was estimated and subtracted from the dense flow. In our case, we consider a simpler
approach: from each displacement field d we subtract its mean vector.
Architecture. Above we have described different ways of combining multiple optical flow displacement fields into a single volume I? ? Rw?h?2L . Considering that a ConvNet requires a fixed-size
input, we sample a 224 ? 224 ? 2L sub-volume from I? and pass it to the net as input. The hidden layers configuration remains largely the same as that used in the spatial net, and is illustrated
in Fig. 1. Testing is similar to the spatial ConvNet, and is described in detail in Sect. 5.
3.2
Relation of the temporal ConvNet architecture to previous representations
In this section, we put our temporal ConvNet architecture in the context of prior art, drawing connections to the video representations, reviewed in Sect. 1.1. Methods based on feature encodings [17, 29] typically combine several spatio-temporal local features. Such features are computed
from the optical flow and are thus generalised by our temporal ConvNet. Indeed, the HOF and MBH
local descriptors are based on the histograms of orientations of optical flow or its gradient, which
can be obtained from the displacement field input (1) using a single convolutional layer (containing
4
orientation-sensitive filters), followed by the rectification and pooling layers. The kinematic features
of [10] (divergence, curl and shear) are also computed from the optical flow gradient, and, again, can
be captured by our convolutional model. Finally, the trajectory feature [29] is computed by stacking
the displacement vectors along the trajectory, which corresponds to the trajectory stacking (2). In the
supplementary material we visualise the convolutional filters, learnt in the first layer of the temporal
network. This provides further evidence that our representation generalises hand-crafted features.
As far as the deep networks are concerned, a two-stream video classification architecture of [16]
contains two HMAX models which are hand-crafted and less deep than our discriminatively trained
ConvNets, which can be seen as a learnable generalisation of HMAX. The convolutional models
of [12, 14] do not decouple spatial and temporal recognition streams, and rely on the motionsensitive convolutional filters, learnt from the data. In our case, motion is explicitly represented
using the optical flow displacement field, computed based on the assumptions of constancy of the
intensity and smoothness of the flow. Incorporating such assumptions into a ConvNet framework
might be able to boost the performance of end-to-end ConvNet-based methods, and is an interesting
direction for future research.
4
Multi-task learning
Unlike the spatial stream ConvNet, which can be pre-trained on a large still image classification
dataset (such as ImageNet), the temporal ConvNet needs to be trained on video data ? and the
available datasets for video action classification are still rather small. In our experiments (Sect. 6),
training is performed on the UCF-101 and HMDB-51 datasets, which have only: 9.5K and 3.7K
videos respectively. To decrease over-fitting, one could consider combining the two datasets into
one; this, however, is not straightforward due to the intersection between the sets of classes. One
option (which we evaluate later) is to only add the images from the classes, which do not appear in
the original dataset. This, however, requires manual search for such classes and limits the amount
of additional training data.
A more principled way of combining several datasets is based on multi-task learning [5]. Its aim
is to learn a (video) representation, which is applicable not only to the task in question (such as
HMDB-51 classification), but also to other tasks (e.g. UCF-101 classification). Additional tasks act
as a regulariser, and allow for the exploitation of additional training data. In our case, a ConvNet
architecture is modified so that it has two softmax classification layers on top of the last fullyconnected layer: one softmax layer computes HMDB-51 classification scores, the other one ? the
UCF-101 scores. Each of the layers is equipped with its own loss function, which operates only on
the videos, coming from the respective dataset. The overall training loss is computed as the sum of
the individual tasks? losses, and the network weight derivatives can be found by back-propagation.
5
Implementation details
ConvNets configuration. The layer configuration of our spatial and temporal ConvNets is schematically shown in Fig. 1. It corresponds to CNN-M-2048 architecture of [3] and is similar to the
network of [31]. All hidden weight layers use the rectification (ReLU) activation function; maxpooling is performed over 3 ? 3 spatial windows with stride 2; local response normalisation uses the
same settings as [15]. The only difference between spatial and temporal ConvNet configurations is
that we removed the second normalisation layer from the latter to reduce memory consumption.
Training. The training procedure can be seen as an adaptation of that of [15] to video frames, and
is generally the same for both spatial and temporal nets. The network weights are learnt using the
mini-batch stochastic gradient descent with momentum (set to 0.9). At each iteration, a mini-batch
of 256 samples is constructed by sampling 256 training videos (uniformly across the classes), from
each of which a single frame is randomly selected. In spatial net training, a 224 ? 224 sub-image is
randomly cropped from the selected frame; it then undergoes random horizontal flipping and RGB
jittering. The videos are rescaled beforehand, so that the smallest side of the frame equals 256. We
note that unlike [15], the sub-image is sampled from the whole frame, not just its 256 ? 256 center.
In the temporal net training, we compute an optical flow volume I for the selected training frame as
described in Sect. 3. From that volume, a fixed-size 224 ? 224 ? 2L input is randomly cropped and
flipped. The learning rate is initially set to 10?2 , and then decreased according to a fixed schedule,
which is kept the same for all training sets. Namely, when training a ConvNet from scratch, the rate
is changed to 10?3 after 50K iterations, then to 10?4 after 70K iterations, and training is stopped
5
after 80K iterations. In the fine-tuning scenario, the rate is changed to 10?3 after 14K iterations, and
training stopped after 20K iterations.
Testing. At test time, given a video, we sample a fixed number of frames (25 in our experiments)
with equal temporal spacing between them. From each of the frames we then obtain 10 ConvNet
inputs [15] by cropping and flipping four corners and the center of the frame. The class scores for the
whole video are then obtained by averaging the scores across the sampled frames and crops therein.
Pre-training on ImageNet ILSVRC-2012. When pre-training the spatial ConvNet, we use the
same training and test data augmentation as described above (cropping, flipping, RGB jittering).
This yields 13.5% top-5 error on ILSVRC-2012 validation set, which compares favourably to 16.0%
reported in [31] for a similar network. We believe that the main reason for the improvement is
sampling of ConvNet inputs from the whole image, rather than just its center.
Multi-GPU training. Our implementation is derived from the publicly available Caffe toolbox [13],
but contains a number of significant modifications, including parallel training on multiple GPUs
installed in a single system. We exploit the data parallelism, and split each SGD batch across several
GPUs. Training a single temporal ConvNet takes 1 day on a system with 4 NVIDIA Titan cards,
which constitutes a 3.2 times speed-up over single-GPU training.
Optical flow is computed using the off-the-shelf GPU implementation of [2] from the OpenCV
toolbox. In spite of the fast computation time (0.06s for a pair of frames), it would still introduce
a bottleneck if done on-the-fly, so we pre-computed the flow before training. To avoid storing
the displacement fields as floats, the horizontal and vertical components of the flow were linearly
rescaled to a [0, 255] range and compressed using JPEG (after decompression, the flow is rescaled
back to its original range). This reduced the flow size for the UCF-101 dataset from 1.5TB to 27GB.
6
Evaluation
Datasets and evaluation protocol. The evaluation is performed on UCF-101 [24] and
HMDB-51 [16] action recognition benchmarks, which are among the largest available annotated
video datasets1 . UCF-101 contains 13K videos (180 frames/video on average), annotated into 101
action classes; HMDB-51 includes 6.8K videos of 51 actions. The evaluation protocol is the same
for both datasets: the organisers provide three splits into training and test data, and the performance
is measured by the mean classification accuracy across the splits. Each UCF-101 split contains 9.5K
training videos; an HMDB-51 split contains 3.7K training videos. We begin by comparing different
architectures on the first split of the UCF-101 dataset. For comparison with the state of the art, we
follow the standard evaluation protocol and report the average accuracy over three splits on both
UCF-101 and HMDB-51.
Spatial ConvNets. First, we measure the performance of the spatial stream ConvNet. Three scenarios are considered: (i) training from scratch on UCF-101, (ii) pre-training on ILSVRC-2012
followed by fine-tuning on UCF-101, (iii) keeping the pre-trained network fixed and only training
the last (classification) layer. For each of the settings, we experiment with setting the dropout regularisation ratio to 0.5 or to 0.9. From the results, presented in Table 1a, it is clear that training the
ConvNet solely on the UCF-101 dataset leads to over-fitting (even with high dropout), and is inferior
to pre-training on a large ILSVRC-2012 dataset. Interestingly, fine-tuning the whole network gives
only marginal improvement over training the last layer only. In the latter setting, higher dropout
over-regularises learning and leads to worse accuracy. In the following experiments we opted for
training the last layer on top of a pre-trained ConvNet.
Temporal ConvNets. Having evaluated spatial ConvNet variants, we now turn to the temporal
ConvNet architectures, and assess the effect of the input configurations, described in Sect. 3.1. In
particular, we measure the effect of: using multiple (L = {5, 10}) stacked optical flows; trajectory
stacking; mean displacement subtraction; using the bi-directional optical flow. The architectures
are trained on the UCF-101 dataset from scratch, so we used an aggressive dropout ratio of 0.9 to
help improve generalisation. The results are shown in Table 1b. First, we can conclude that stacking
multiple (L > 1) displacement fields in the input is highly beneficial, as it provides the network with
long-term motion information, which is more discriminative than the flow between a pair of frames
1
Very recently, [14] released the Sports-1M dataset of 1.1M automatically annotated YouTube sports videos.
Processing the dataset of such scale is very challenging, and we plan to address it in future work.
6
Table 1: Individual ConvNets accuracy on UCF-101 (split 1).
(a) Spatial ConvNet.
Dropout ratio
Training setting
0.5
0.9
From scratch
42.5% 52.3%
Pre-trained + fine-tuning 70.8% 72.8%
Pre-trained + last layer
72.7% 59.9%
(b) Temporal ConvNet.
Mean subtraction
off
on
Single-frame optical flow (L = 1)
73.9%
Optical flow stacking (1) (L = 5)
80.4%
Optical flow stacking (1) (L = 10)
79.9% 81.0%
Trajectory stacking (2)(L = 10)
79.6% 80.2%
Optical flow stacking (1)(L = 10), bi-dir.
81.2%
Input configuration
(L = 1 setting). Increasing the number of input flows from 5 to 10 leads to a smaller improvement,
so we kept L fixed to 10 in the following experiments. Second, we find that mean subtraction is
helpful, as it reduces the effect of global motion between the frames. We use it in the following
experiments as default. The difference between different stacking techniques is marginal; it turns
out that optical flow stacking performs better than trajectory stacking, and using the bi-directional
optical flow is only slightly better than a uni-directional forward flow. Finally, we note that temporal
ConvNets significantly outperform the spatial ConvNets (Table 1a), which confirms the importance
of motion information for action recognition.
We also implemented the ?slow fusion? architecture of [14], which amounts to applying a ConvNet
to a stack of RGB frames (11 frames in our case). When trained from scratch on UCF-101, it
achieved 56.4% accuracy, which is better than a single-frame architecture trained from scratch
(52.3%), but is still far off the network trained from scratch on optical flow. This shows that while
multi-frame information is important, it is also important to present it to a ConvNet in an appropriate
manner.
Multi-task learning of temporal ConvNets. Training temporal ConvNets on UCF-101 is challenging due to the small size of the training set. An even bigger challenge is to train the ConvNet on
HMDB-51, where each training split is 2.6 times smaller than that of UCF-101. Here we evaluate
different options for increasing the effective training set size of HMDB-51: (i) fine-tuning a temporal
network pre-trained on UCF-101; (ii) adding 78 classes from UCF-101, which are manually selected
so that there is no intersection between these classes and the native HMDB-51 classes; (iii) using the
multi-task formulation (Sect. 4) to learn a video representation, shared between the UCF-101 and
HMDB-51 classification tasks. The results are reported in Table 2. As expected, it is beneficial to
Table 2: Temporal ConvNet accuracy on HMDB-51 (split 1 with additional training data).
Training setting
Training on HMDB-51 without additional data
Fine-tuning a ConvNet, pre-trained on UCF-101
Training on HMDB-51 with classes added from UCF-101
Multi-task learning on HMDB-51 and UCF-101
Accuracy
46.6%
49.0%
52.8%
55.4%
utilise full (all splits combined) UCF-101 data for training (either explicitly by borrowing images, or
implicitly by pre-training). Multi-task learning performs the best, as it allows the training procedure
to exploit all available training data.
We have also experimented with multi-task learning on the UCF-101 dataset, by training a network
to classify both the full HMDB-51 data (all splits combined) and the UCF-101 data (a single split).
On the first split of UCF-101, the accuracy was measured to be 81.5%, which improves on 81.0%
achieved using the same settings, but without the additional HMDB classification task (Table 1b).
Two-stream ConvNets. Here we evaluate the complete two-stream model, which combines the
two recognition streams. One way of combining the networks would be to train a joint stack of
fully-connected layers on top of full6 or full7 layers of the two nets. This, however, was not feasible
in our case due to over-fitting. We therefore fused the softmax scores using either averaging or
a linear SVM. From Table 3 we conclude that: (i) temporal and spatial recognition streams are
complementary, as their fusion significantly improves on both (6% over temporal and 14% over
spatial nets); (ii) SVM-based fusion of softmax scores outperforms fusion by averaging; (iii) using
bi-directional flow is not beneficial in the case of ConvNet fusion; (iv) temporal ConvNet, trained
using multi-task learning, performs the best both alone and when fused with a spatial net.
Comparison with the state of the art. We conclude the experimental evaluation with the comparison against the state of the art on three splits of UCF-101 and HMDB-51. For that we used a
7
Table 3: Two-stream ConvNet accuracy on UCF-101 (split 1).
Spatial ConvNet
Pre-trained + last layer
Pre-trained + last layer
Pre-trained + last layer
Pre-trained + last layer
Temporal ConvNet
bi-directional
uni-directional
uni-directional, multi-task
uni-directional, multi-task
Fusion Method
averaging
averaging
averaging
SVM
Accuracy
85.6%
85.9%
86.2%
87.0%
spatial net, pre-trained on ILSVRC, with the last layer trained on UCF or HMDB. The temporal
net was trained on UCF and HMDB using multi-task learning, and the input was computed using
uni-directional optical flow stacking with mean subtraction. The softmax scores of the two nets were
combined using averaging or SVM. As can be seen from Table 4, both our spatial and temporal nets
alone outperform the deep architectures of [14, 16] by a large margin. The combination of the two
nets further improves the results (in line with the single-split experiments above), and is comparable
to the very recent state-of-the-art hand-crafted models [20, 21, 26].
Table 4: Mean accuracy (over three splits) on UCF-101 and HMDB-51.
Method
Improved dense trajectories (IDT) [26, 27]
IDT with higher-dimensional encodings [20]
IDT with stacked Fisher encoding [21] (based on Deep Fisher Net [23])
Spatio-temporal HMAX network [11, 16]
?Slow fusion? spatio-temporal ConvNet [14]
Spatial stream ConvNet
Temporal stream ConvNet
Two-stream model (fusion by averaging)
Two-stream model (fusion by SVM)
7
UCF-101
85.9%
87.9%
65.4%
73.0%
83.7%
86.9%
88.0%
HMDB-51
57.2%
61.1%
66.8%
22.8%
40.5%
54.6%
58.0%
59.4%
Conclusions and directions for improvement
We proposed a deep video classification model with competitive performance, which incorporates
separate spatial and temporal recognition streams based on ConvNets. Currently it appears that
training a temporal ConvNet on optical flow (as here) is significantly better than training on raw
stacked frames [14]. The latter is probably too challenging, and might require architectural changes
(for example, a combination with the deep matching approach of [30]). Despite using optical flow
as input, our temporal model does not require significant hand-crafting, since the flow is computed
using a method based on the generic assumptions of constancy and smoothness.
As we have shown, extra training data is beneficial for our temporal ConvNet, so we are planning to
train it on large video datasets, such as the recently released collection of [14]. This, however, poses
a significant challenge on its own due to the gigantic amount of training data (multiple TBs).
There still remain some essential ingredients of the state-of-the-art shallow representation [26],
which are missed in our current architecture. The most prominent one is local feature pooling
over spatio-temporal tubes, centered at the trajectories. Even though the input (2) captures the optical flow along the trajectories, the spatial pooling in our network does not take the trajectories into
account. Another potential area of improvement is explicit handling of camera motion, which in our
case is compensated by mean displacement subtraction.
Acknowledgements
This work was supported by ERC grant VisRec no. 228180. We gratefully acknowledge the support
of NVIDIA Corporation with the donation of the GPUs used for this research.
References
[1] A. Berg, J. Deng, and L. Fei-Fei. Large scale visual recognition challenge (ILSVRC), 2010. URL
http://www.image-net.org/challenges/LSVRC/2010/.
[2] T. Brox, A. Bruhn, N. Papenberg, and J. Weickert. High accuracy optical flow estimation based on a
theory for warping. In Proc. ECCV, pages 25?36, 2004.
[3] K. Chatfield, K. Simonyan, A. Vedaldi, and A. Zisserman. Return of the devil in the details: Delving deep
into convolutional nets. In Proc. BMVC., 2014.
[4] B. Chen, J. A. Ting, B. Marlin, and N. de Freitas. Deep learning of invariant spatio-temporal features
from video. In NIPS Deep Learning and Unsupervised Feature Learning Workshop, 2010.
8
[5] R. Collobert and J. Weston. A unified architecture for natural language processing: deep neural networks
with multitask learning. In Proc. ICML, pages 160?167, 2008.
[6] K. Crammer and Y. Singer. On the algorithmic implementation of multiclass kernel-based vector machines. JMLR, 2:265?292, 2001.
[7] N. Dalal and B Triggs. Histogram of Oriented Gradients for Human Detection. In Proc. CVPR, volume 2,
pages 886?893, 2005.
[8] N. Dalal, B. Triggs, and C. Schmid. Human detection using oriented histograms of flow and appearance.
In Proc. ECCV, pages 428?441, 2006.
[9] M. A. Goodale and A. D. Milner. Separate visual pathways for perception and action. Trends in Neurosciences, 15(1):20?25, 1992.
[10] M. Jain, H. Jegou, and P. Bouthemy. Better exploiting motion for better action recognition. In Proc.
CVPR, pages 2555?2562, 2013.
[11] H. Jhuang, T. Serre, L. Wolf, and T. Poggio. A biologically inspired system for action recognition. In
Proc. ICCV, pages 1?8, 2007.
[12] S. Ji, W. Xu, M. Yang, and K. Yu. 3D convolutional neural networks for human action recognition. IEEE
PAMI, 35(1):221?231, 2013.
[13] Y. Jia. Caffe: An open source convolutional architecture for fast feature embedding. http://caffe.
berkeleyvision.org/, 2013.
[14] A. Karpathy, G. Toderici, S. Shetty, T. Leung, R. Sukthankar, and L. Fei-Fei. Large-scale video classication with convolutional neural networks. In Proc. CVPR, 2014.
[15] A. Krizhevsky, I. Sutskever, and G. E. Hinton. ImageNet classification with deep convolutional neural
networks. In NIPS, pages 1106?1114, 2012.
[16] H. Kuehne, H. Jhuang, E. Garrote, T. Poggio, and T. Serre. HMDB: A large video database for human
motion recognition. In Proc. ICCV, pages 2556?2563, 2011.
[17] I. Laptev, M. Marsza?ek, C. Schmid, and B. Rozenfeld. Learning realistic human actions from movies. In
Proc. CVPR, 2008.
[18] Q. V. Le, W. Y. Zou, S. Y. Yeung, and A. Y. Ng. Learning hierarchical invariant spatio-temporal features
for action recognition with independent subspace analysis. In Proc. CVPR, pages 3361?3368, 2011.
[19] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel. Backpropagation applied to handwritten zip code recognition. Neural Computation, 1(4):541?551, 1989.
[20] X. Peng, L. Wang, X. Wang, and Y. Qiao. Bag of visual words and fusion methods for action recognition:
Comprehensive study and good practice. CoRR, abs/1405.4506, 2014.
[21] X. Peng, C. Zou, Y. Qiao, and Q. Peng. Action recognition with stacked fisher vectors. In Proc. ECCV,
pages 581?595, 2014.
[22] F. Perronnin, J. S?anchez, and T. Mensink. Improving the Fisher kernel for large-scale image classification.
In Proc. ECCV, 2010.
[23] K. Simonyan, A. Vedaldi, and A. Zisserman. Deep Fisher networks for large-scale image classification.
In NIPS, 2013.
[24] K. Soomro, A. R. Zamir, and M. Shah. UCF101: A dataset of 101 human actions classes from videos in
the wild. CoRR, abs/1212.0402, 2012.
[25] G. W. Taylor, R. Fergus, Y. LeCun, and C. Bregler. Convolutional learning of spatio-temporal features.
In Proc. ECCV, pages 140?153, 2010.
[26] H. Wang and C. Schmid. Action recognition with improved trajectories. In Proc. ICCV, pages 3551?3558,
2013.
[27] H. Wang and C. Schmid. LEAR-INRIA submission for the THUMOS workshop. In ICCV Workshop on
Action Recognition with a Large Number of Classes, 2013.
[28] H. Wang, M. M. Ullah, A. Kl?aser, I. Laptev, and C. Schmid. Evaluation of local spatio-temporal features
for action recognition. In Proc. BMVC., pages 1?11, 2009.
[29] H. Wang, A. Kl?aser, C. Schmid, and C.-L. Liu. Action recognition by dense trajectories. In Proc. CVPR,
pages 3169?3176, 2011.
[30] P. Weinzaepfel, J. Revaud, Z. Harchaoui, and C. Schmid. DeepFlow: Large displacement optical flow
with deep matching. In Proc. ICCV, pages 1385?1392, 2013.
[31] M. D. Zeiler and R. Fergus. Visualizing and understanding convolutional networks. CoRR, abs/1311.2901,
2013.
9
| 5353 |@word multitask:2 exploitation:1 cnn:1 dalal:2 norm:3 triggs:2 open:1 confirms:1 rgb:3 sgd:1 carry:1 configuration:7 contains:7 score:10 liu:1 interestingly:2 outperforms:3 freitas:1 ullah:1 current:1 comparing:1 activation:1 dx:4 reminiscent:1 gpu:3 realistic:1 mbh:2 alone:2 selected:4 accordingly:1 provides:4 detecting:1 location:6 org:2 simpler:1 height:1 along:6 constructed:2 consists:2 pathway:2 combine:2 fitting:3 fullyconnected:1 wild:1 manner:1 introduce:2 peng:3 indeed:1 expected:2 p1:1 planning:1 multi:17 jegou:1 inspired:2 decomposed:1 automatically:1 toderici:1 equipped:1 considering:1 window:1 increasing:2 begin:1 linearity:1 developed:1 whilst:1 unified:1 marlin:1 corporation:1 temporal:61 act:1 classifier:1 uk:1 grant:1 appear:1 gigantic:1 rozenfeld:1 positive:2 before:2 generalised:1 local:10 limit:1 installed:1 despite:1 encoding:5 oxford:1 solely:1 pami:1 might:3 inria:1 therein:1 challenging:5 limited:1 bi:7 range:2 camera:5 lecun:2 testing:2 practice:1 backpropagation:1 procedure:2 displacement:23 area:3 significantly:6 mult:1 ucf101:2 matching:3 pre:21 vedaldi:2 word:1 spite:3 close:1 put:1 effortless:1 context:1 applying:1 sukthankar:1 www:1 center:3 compensated:1 straightforward:1 attention:1 conv4:2 embedding:1 variation:1 milner:1 us:1 hypothesis:1 trend:1 recognition:44 finetuned:1 submission:1 native:1 bouthemy:1 database:1 constancy:3 fly:1 wang:6 capture:3 zamir:1 region:1 revaud:1 connected:1 sect:18 movement:4 decrease:1 removed:1 rescaled:3 principled:1 goodale:1 jittering:3 trained:26 solving:1 laptev:2 upon:1 joint:1 represented:2 derivation:1 stacked:6 train:4 jain:1 fast:2 describe:3 effective:1 caffe:3 whose:1 encoded:1 supplementary:1 cvpr:6 drawing:1 compressed:1 simonyan:3 highlighted:1 itself:1 sequence:2 net:19 propose:1 coming:1 adaptation:1 turned:1 combining:4 bow:1 achieve:1 az:1 exploiting:1 sutskever:1 extending:1 cropping:2 object:4 help:1 donation:1 andrew:1 ac:1 pose:1 develop:1 measured:2 received:1 conv5:2 dividing:1 implemented:4 indicate:1 direction:4 annotated:4 filter:4 stochastic:1 centered:2 human:8 material:1 organiser:1 require:2 probable:1 bregler:1 extension:1 around:1 considered:1 algorithmic:1 opencv:1 ventral:2 consecutive:5 smallest:1 released:2 estimation:1 proc:18 applicable:1 bag:2 currently:1 marsza:1 jackel:1 sensitive:1 hubbard:1 largest:1 aim:3 modified:1 rather:3 avoid:1 shelf:1 minimisation:1 encode:1 derived:1 improvement:6 opted:1 sense:1 helpful:1 dependent:1 perronnin:1 leung:1 typically:3 initially:1 hidden:2 relation:2 borrowing:1 comprising:1 pixel:1 overall:1 classification:24 orientation:2 among:1 plan:1 spatial:39 art:11 softmax:9 brox:1 fairly:1 marginal:2 field:15 construct:1 equal:2 having:1 sampling:3 manually:1 ng:1 flipped:1 yu:1 unsupervised:2 constitutes:1 theart:1 bruhn:1 icml:1 future:2 report:1 idt:3 oriented:3 randomly:3 divergence:1 comprehensive:1 individual:5 papenberg:1 geometry:1 attempt:2 ab:3 detection:2 interest:2 normalisation:2 investigate:3 kinematic:1 highly:1 evaluation:7 henderson:1 introduces:1 accurate:1 beforehand:1 deepflow:1 poggio:2 respective:1 iv:1 plugged:1 taylor:1 stopped:2 instance:1 classify:1 jpeg:1 formulates:1 stacking:18 hof:2 krizhevsky:1 too:1 reported:2 learnt:5 dir:1 combined:7 off:3 pool:6 fused:2 augmentation:2 again:1 tube:1 containing:1 worse:2 corner:1 ek:1 derivative:1 return:1 aggressive:1 account:1 potential:1 de:1 stride:11 pooled:1 availability:1 includes:1 titan:1 explicitly:3 caused:1 stream:38 collobert:1 later:3 performed:3 observer:1 competitive:4 start:1 option:2 parallel:1 jia:1 contribution:1 ass:1 formed:1 publicly:1 accuracy:14 convolutional:17 descriptor:3 largely:2 who:1 yield:1 directional:11 raw:1 handwritten:1 trajectory:24 manual:1 centering:1 bof:1 against:1 energy:2 conveys:1 naturally:2 associated:1 rbm:1 static:1 sampled:5 dataset:17 adjusting:1 popular:1 kuehne:1 improves:4 schedule:1 back:2 appears:1 higher:3 dt:3 day:1 follow:2 zisserman:3 specify:1 response:1 improved:2 formulation:1 evaluated:3 ox:1 though:2 strongly:1 done:1 just:2 bmvc:2 mensink:1 convnets:18 aser:2 hand:11 horizontal:5 favourably:1 propagation:1 undergoes:1 weinzaepfel:1 believe:1 effect:3 serre:2 illustrated:3 deal:2 visualizing:1 width:1 recurrence:1 inferior:1 berkeleyvision:1 prominent:1 recognised:1 complete:1 demonstrate:1 performs:6 motion:26 image:21 recently:5 shear:1 ji:1 volume:12 visualise:1 significant:4 smoothness:3 curl:1 tuning:6 grid:1 outlined:2 similarly:2 decompression:1 erc:1 gratefully:1 language:1 moving:2 robot:1 ucf:35 cortex:1 operating:1 maxpooling:1 add:1 own:3 recent:3 driven:2 scenario:2 store:2 nvidia:2 devise:1 seen:5 captured:1 additional:8 zip:1 deng:1 subtraction:6 ii:3 multiple:10 full:2 isa:1 reduces:1 harchaoui:1 exceeds:1 generalises:1 full6:3 long:1 bigger:1 variant:2 crop:1 essentially:1 yeung:1 histogram:6 represent:1 iteration:6 kernel:2 pyramid:1 achieved:3 schematically:1 cropped:2 separately:1 fine:6 addressed:2 decreased:1 spacing:1 float:1 source:1 extra:1 rest:1 unlike:4 probably:1 pooling:4 incorporates:2 flow:63 yang:1 split:18 iii:3 concerned:1 variety:1 relu:1 architecture:32 opposite:2 reduce:1 multiclass:1 bottleneck:1 chatfield:1 colour:1 gb:1 url:1 soomro:1 karen:2 action:34 deep:21 useful:1 generally:2 clear:1 karpathy:1 amount:6 rw:2 reduced:1 http:2 specifies:1 outperform:2 estimated:1 neuroscience:1 group:1 four:1 kept:2 rectangle:1 backward:1 sum:1 family:1 architectural:1 patch:1 recognise:1 missed:1 garrote:1 dy:3 comparable:1 dropout:9 layer:25 cyan:1 followed:2 weickert:1 fold:1 replaces:1 activity:1 adapted:1 fei:4 scene:1 x2:6 dominated:1 speed:1 performing:2 optical:44 relatively:1 gpus:3 according:2 combination:4 thumos:1 across:8 describes:1 beneficial:5 smaller:2 slightly:1 remain:1 shallow:7 modification:1 biologically:1 invariant:2 iccv:5 pipeline:1 rectification:3 previously:1 remains:1 turn:2 singer:1 end:4 generalizes:1 available:4 denker:1 hierarchical:1 appropriate:1 generic:1 subtracted:1 alternative:1 batch:3 shetty:1 shah:1 original:2 top:4 include:1 zeiler:1 exploit:5 ting:1 build:1 crafting:1 move:1 warping:1 question:1 recognises:1 flipping:3 added:1 gradient:7 subspace:1 convnet:56 separate:3 card:1 majority:1 consumption:1 reason:1 code:1 mini:2 ratio:3 difficult:1 hog:1 negative:2 implementation:5 reliably:1 regulariser:1 conv2:2 perform:1 vertical:4 anchez:1 datasets:11 benchmark:2 acknowledge:1 howard:1 compensation:2 descent:1 extended:1 hinton:1 frame:57 stack:5 arbitrary:2 community:1 intensity:4 introduced:1 pair:5 namely:1 toolbox:2 kl:2 connection:2 imagenet:5 qiao:2 boser:1 datasets1:1 boost:2 nip:3 address:1 able:2 below:1 parallelism:1 perception:1 challenge:7 tb:2 including:1 memory:1 video:58 natural:3 rely:1 scheme:1 improve:2 movie:1 carried:1 schmid:7 full7:3 review:1 prior:1 l2:1 acknowledgement:1 understanding:1 regularisation:1 loss:3 fully:1 discriminatively:2 highlight:1 dxt:1 interesting:1 organised:1 ingredient:1 validation:1 conv1:2 storing:1 classication:1 eccv:5 changed:2 jhuang:2 supported:1 last:10 keeping:1 side:1 allow:2 deeper:1 generalise:1 normalised:1 conv3:2 sparse:2 boundary:1 default:1 computes:1 forward:3 collection:1 clue:2 far:2 uni:5 implicitly:3 cuboid:1 global:3 visrec:1 conclude:3 spatio:17 discriminative:3 fergus:2 search:1 reviewed:3 additionally:1 table:11 learn:3 channel:8 delving:1 decoupling:1 improving:1 zou:2 protocol:3 pk:7 dense:11 main:1 linearly:1 whole:4 complementary:3 xu:1 crafted:8 fig:6 slow:2 sub:3 momentum:1 explicit:1 jmlr:1 late:2 hmax:6 learnable:1 experimented:1 svm:7 evidence:1 fusion:14 incorporating:1 essential:1 workshop:3 adding:1 effectively:1 importance:2 corr:3 margin:2 chen:1 easier:1 subtract:1 suited:1 depicted:1 intersection:2 appearance:4 forming:1 hmdb:25 visual:5 sport:4 regularises:1 corresponds:3 utilise:1 wolf:1 weston:1 lear:1 shared:1 fisher:6 feasible:1 change:1 youtube:2 lsvrc:1 generalisation:2 operates:3 uniformly:1 averaging:9 decouple:1 total:1 dyt:1 pas:1 experimental:1 formally:1 ilsvrc:6 berg:1 support:2 latter:3 crammer:1 dorsal:2 devil:1 incorporate:1 evaluate:3 scratch:7 handling:1 |
4,809 | 5,354 | Rounding-based Moves for Metric Labeling
M. Pawan Kumar
Ecole Centrale Paris & INRIA Saclay
[email protected]
Abstract
Metric labeling is a special case of energy minimization for pairwise Markov random fields. The energy function consists of arbitrary unary potentials, and pairwise potentials that are proportional to a given metric distance function over the
label set. Popular methods for solving metric labeling include (i) move-making
algorithms, which iteratively solve a minimum st-cut problem; and (ii) the linear
programming (LP) relaxation based approach. In order to convert the fractional
solution of the LP relaxation to an integer solution, several randomized rounding procedures have been developed in the literature. We consider a large class
of parallel rounding procedures, and design move-making algorithms that closely
mimic them. We prove that the multiplicative bound of a move-making algorithm
exactly matches the approximation factor of the corresponding rounding procedure for any arbitrary distance function. Our analysis includes all known results
for move-making algorithms as special cases.
1
Introduction
A Markov random field (MRF) is a graph whose vertices are random variables, and whose edges
specify a neighborhood over the random variables. Each random variable can be assigned a value
from a set of labels, resulting in a labeling of the MRF. The putative labelings of an MRF are
quantitatively distinguished from each other by an energy function, which is the sum of potential
functions that depend on the cliques of the graph. An important optimization problem associate with
the MRF framework is energy minimization, that is, finding a labeling with the minimum energy.
Metric labeling is a special case of energy minimization, which models several useful low-level
vision tasks [3, 4, 18]. It is characterized by a finite, discrete label set and a metric distance function
over the labels. The energy function in metric labeling consists of arbitrary unary potentials and
pairwise potentials that are proportional to the distance between the labels assigned to them. The
problem is known to be NP-hard [20]. Two popular approaches for metric labeling are: (i) movemaking algorithms [4, 8, 14, 15, 21], which iteratively improve the labeling by solving a minimum
st-cut problem; and (ii) linear programming (LP) relaxation [5, 13, 17, 22], which is obtained by
dropping the integral constraints in the corresponding integer programming formulation. Movemaking algorithms are very efficient due to the availability of fast minimum st-cut solvers [2] and
are very popular in the computer vision community. In contrast, the LP relaxation is significantly
slower, despite the development of specialized solvers [7, 9, 11, 12, 16, 19, 22, 23, 24, 25]. However,
when used in conjunction with randomized rounding algorithms, the LP relaxation provides the best
known polynomial-time theoretical guarantees for metric labeling [1, 5, 10].
At first sight, the difference between move-making algorithms and the LP relaxation appears to be
the standard accuracy vs. speed trade-off. However, for some special cases of distance functions,
it has been shown that appropriately designed move-making algorithms can match the theoretical
guarantees of the LP relaxation [14, 15, 20]. In this paper, we extend this result for a large class
of randomized rounding procedures, which we call parallel rounding. In particular we prove that
for any arbitrary (semi-)metric distance function, there exist move-making algorithms that match
the theoretical guarantees provided by parallel rounding. The proofs, the various corollaries of our
1
theorems (which cover all previously known guarantees) and our experimental results are deferred
to the accompanying technical report.
2
Preliminaries
Metric Labeling. The problem of metric labeling is defined over an undirected graph G =
(X, E). The vertices X = {X1 , X2 , ? ? ? , Xn } are random variables, and the edges E specify a
neighborhood relationship over the random variables. Each random variable can be assigned a value
from the label set L = {l1 , l2 , ? ? ? , lh }. We assume that we are also provided with a metric distance
function d : L ? L ? R+ over the labels.
We refer to an assignment of values to all the random variables as a labeling. In other words, a
labeling is a vector x ? Ln , which specifies the label xa assigned to each random variable Xa . The
hn different labelings are quantitatively distinguished from each other by an energy function Q(x),
which is defined as follows:
X
X
Q(x) =
wab d(xa , xb ).
?a (xa ) +
Xa ?X
(Xa ,Xb )?E
Here, the unary potentials ?a (?) are arbitrary, and the edge weights wab are non-negative. Metric
labeling requires us to find a labeling with the minimum energy. It is known to be NP-hard.
Multiplicative Bound. As metric labeling plays a central role in low-level vision, several approximate algorithms have been proposed in the literature. A common theoretical measure of accuracy
for an approximate algorithm is the multiplicative bound. In this work, we are interested in the
multiplicative bound of an algorithm with respect to a distance function. Formally, given a distance
function d, the multiplicative bound of an algorithm is said to be B if the following condition is
satisfied for all possible values of unary potentials ?a (?) and non-negative edge weights wab :
X
X
X
X
wab d(x?a , x?b ). (1)
wab d(?
xa , x
?b ) ?
?a (x?a ) + B
?a (?
xa ) +
Xa ?X
Xa ?X
(Xa ,Xb )?E
(Xa ,Xb )?E
? is the labeling estimated by the algorithm for the given values of unary potentials and edge
Here, x
weights, and x? is an optimal labeling. Multiplicative bounds are greater than or equal to 1, and are
invariant to reparameterizations of the unary potentials. A multiplicative bound B is said to be tight
if the above inequality holds as an equality for some value of unary potentials and edge weights.
Linear Programming Relaxation. An overcomplete representation of a labeling can be specified
using the following variables: (i) unary variables ya (i) ? {0, 1} for all Xa ? X and li ? L such
that ya (i) = 1 if and only if Xa is assigned the label li ; and (ii) pairwise variables yab (i, j) ? {0, 1}
for all (Xa , Xb ) ? E and li , lj ? L such that yab (i, j) = 1 if and only if Xa and Xb are assigned
labels li and lj respectively. This allows us to formulate metric labeling as follows:
X X
X
X
min
?a (li )ya (i) +
wab d(li , lj )yab (i, j),
y
s.t.
Xa ?X li ?L
(Xa ,Xb )?E li ,lj ?L
X
ya (i) = 1, ?Xa ? X,
X
yab (i, j) = ya (i), ?(Xa , Xb ) ? E, li ? L,
X
yab (i, j) = yb (j), ?(Xa , Xb ) ? E, lj ? L,
li ?L
lj ?L
li ?L
ya (i) ? {0, 1}, yab (i, j) ? {0, 1}, ?Xa ? X, (Xa , Xb ) ? E, li , lj ? L.
By relaxing the final set of constraints such that the optimization variables can take any value between 0 and 1 inclusive, we obtain a linear program (LP). The computational complexity of solving
the LP relaxation is polynomial in the size of the problem.
Rounding Procedure. In order to prove theoretical guarantees of the LP relaxation, it is common
to use a rounding procedure that can covert a feasible fractional solution y of the LP relaxation to
? of the integer linear program. Several rounding procedures have been
a feasible integer solution y
2
proposed in the literature. In this work, we focus on the randomized parallel rounding procedures
proposed in [5, 10]. These procedures have the property that, given a fractional solution y, the
probability of assigning a label li ? L to a random variable Xa ? X is equal to ya (i), that is,
Pr(?
ya (i) = 1) = ya (i).
(2)
We will describe the various rounding procedures in detail in sections 3-5. For now, we would like
to note that our reason for focusing on the parallel rounding of [5, 10] is that they provide the best
known polynomial-time theoretical guarantees for metric labeling. Specifically, we are interested in
their approximation factor, which is defined next.
Approximation Factor. Given a distance function d, the approximation factor for a rounding procedure is said to be F if the following condition is satisfied for all feasible fractional solutions y:
?
?
X
X
d(li , lj )yab (i, j).
(3)
d(li , lj )?
ya (i)?
yb (j)? ? F
E?
li ,lj ?L
li ,lj ?L
? refers to the integer solution, and the expectation is taken with respect to the randomized
Here, y
rounding procedure applied to the feasible solution y.
Given a rounding procedure with an approximation factor of F , an optimal fractional solution y? of
? that satisfies the following condition:
the LP relaxation can be rounded to a labeling y
?
?
X
X
X X
wab d(li , lj )?
ya (i)?
yb (j)?
?a (li )?
ya (i) +
E?
Xa ?X li ?L
?
X X
Xa ?X li ?L
?a (li )ya? (i)
(Xa ,Xb )?E li ,lj ?L
+F
X
X
?
wab d(li , lj )yab
(i, j).
(Xa ,Xb )?E li ,lj ?L
The above inequality follows directly from properties (2) and (3). Similar to multiplicative bounds,
approximation factors are always greater than or equal to 1, and are invariant to reparameterizations
of the unary potentials. An approximation factor F is said to be tight if the above inequality holds
as an equality for some value of unary potentials and edge weights.
Submodular Energy Function. We will use the following important fact throughout this paper.
Given an energy function defined using arbitrary unary potentials, non-negative edge weights and a
submodular distance function, an optimal labeling can be computed in polynomial time by solving
an equivalent minimum st-cut problem [6]. Recall that a submodular distance function d? over a
label set L = {l1 , l2 , ? ? ? , lh } satisfies the following properties: (i) d? (li , lj ) ? 0 for all li , lj ? L,
and d? (li , lj ) = 0 if and only if i = j; and (ii) d? (li , lj ) + d? (li+1 , lj+1 ) ? d? (li , lj+1 ) + d? (li+1 , lj )
for all li , lj ? L\{lh } (where \ refers to set difference).
3
Complete Rounding and Complete Move
We start with a simple rounding scheme, which we call complete rounding. While complete rounding is not very accurate, it would help illustrate the flavor of our results. We will subsequently
consider its generalizations, which have been useful in obtaining the best-known approximation
factors for various special cases of metric labeling.
The complete rounding procedure consists of a single stage where we use the set of all unary variables to obtain a labeling (as opposed to other rounding procedures discussed subsequently). Algorithm 1 describes its main steps. Intuitively, it treats the value of the unary variable ya (i) as the
probability of assigning the label li to the random variable Xa . It obtains a labeling by sampling
from all the distributions ya = [ya (i), ?li ? L] simultaneously using the same random number.
It can be shown that using a different random number to sample the distributions ya and yb of
two neighboring random variables (Xa , Xb ) ? E results in an infinite approximation factor. For
example, let y a (i) = y b (i) = 1/h for all li ? L, where h is the number of labels. The pairwise
variables yab that minimize the energy function are y ab (i, i) = 1/h and y ab (i, j) = 0 when i 6= j.
For the above feasible solution of the LP relaxation, the RHS of inequality (3) is 0 for any finite F ,
while the LHS of inequality (3) is strictly greater than 0 if h > 1. However, we will shortly show that
using the same random number r for all random variables provides a finite approximation factor.
3
Algorithm 1 The complete rounding procedure.
input A feasible solution y of the LP relaxation.
1: Pick a real number r uniformly from [0, 1].
2: for all Xa ? X do
P
3:
Define Ya (0) = 0 and Ya (i) = ij=1 ya (j) for all li ? L.
4:
Assign the label li ? L to the random variable Xa if Ya (i ? 1) < r ? Ya (i).
5: end for
We now turn our attention to designing a move-making algorithm whose multiplicative bound
matches the approximation factor of the complete rounding procedure. To this end, we modify
the range expansion algorithm proposed in [15] for truncated convex pairwise potentials to a general
(semi-)metric distance function. Our method, which we refer to as the complete move-making algorithm, considers all putative labels of all random variables, and provides an approximate solution
in a single iteration. Algorithm 2 describes its two main steps. First, it computes a submodular
overestimation of the given distance function by solving the following optimization problem:
d=
argmin t
(4)
d?
s.t.
d? (li , lj ) ? td(li , lj ), ?li , lj ? L,
d? (li , lj ) ? d(li , lj ), ?li , lj ? L,
d? (li , lj ) + d? (li+1 , lj+1 ) ? d? (li , lj+1 ) + d? (li+1 , lj ), ?li , lj ? L\{lh }.
The above problem minimizes the maximum ratio of the estimated distance to the original distance
over all pairs of labels, that is, maxi6=j d? (li , lj )/d(li , lj ). We will refer to the optimal value of
problem (4) as the submodular distortion of the distance function d. Second, it replaces the original
distance function by the submodular overestimation and computes an approximate solution to the
original metric labeling problem by solving a single minimum st-cut problem. Note that, unlike
the range expansion algorithm [15] that uses the readily available submodular overestimation of
a truncated convex distance (namely, the corresponding convex distance function), our approach
estimates the submodular overestimation via the LP (4). Since the LP (4) can be solved for any
arbitrary distance function, it makes complete move-making more generally applicable.
Algorithm 2 The complete move-making algorithm.
input Unary potentials ?a (?), edge weights wab , distance function d.
1: Compute a submodular overestimation of d by solving problem (4).
2: Using the approach of [6], solve the following problem via an equivalent minimum st-cut problem:
X
X
? = argmin
wab d(xa , xb ).
x
?a (xa ) +
x?Ln
Xa ?X
(Xa ,Xb )?E
The following theorem establishes the theoretical guarantees of the complete move-making algorithm and the complete rounding procedure.
Theorem 1. The tight multiplicative bound of the complete move-making algorithm is equal to the
submodular distortion of the distance function. Furthermore, the tight approximation factor of the
complete rounding procedure is also equal to the submodular distortion of the distance function.
In terms of computational complexities, complete move-making is significantly faster than solving
the LP relaxation. Specifically, given an MRF with n random variables and m edges, and a label
set with h labels, the LP relaxation requires at least O(m3 h3 log(m2 h3 )) time, since it consists
of O(mh2 ) optimization variables and O(mh) constraints. In contrast, complete move-making
requires O(nmh3 log(m)) time, since the graph constructed using the method of [6] consists of
O(nh) nodes and O(mh2 ) arcs. Note that complete move-making also requires us to solve the
linear program (4). However, since problem (4) is independent of the unary potentials and the edge
weights, it only needs to be solved once beforehand in order to compute the approximate solution
for any metric labeling problem defined using the distance function d.
4
4
Interval Rounding and Interval Moves
Theorem 1 implies that the approximation factor of the complete rounding procedure is very large
for distance functions that are highly non-submodular. For example, consider the truncated linear
distance function defined as follows over a label set L = {l1 , l2 , ? ? ? , lh }:
d(li , lj ) = min{|i ? j|, M }.
Here, M is a user specified parameter that determines the maximum distance. The tightest submodular overestimation of the above distance function is the linear distance function, that is,
d(li , lj ) = |i ? j|. This implies that the submodular distortion of the truncated linear metric is
(h ? 1)/M , and therefore, the approximation factor for the complete rounding procedure is also
(h ? 1)/M . In order to avoid this large approximation factor, Chekuri et al. [5] proposed an interval
rounding procedure, which captures the intuition that it is beneficial to assign similar labels to as
many random variables as possible.
Algorithm 3 provides a description of interval rounding. The rounding procedure chooses an interval
of at most q consecutive labels (step 2). It generates a random number r (step 3), and uses it to
attempt to assign labels to previously unlabeled random variables from the selected interval (steps
4-7). It can be shown that the overall procedure converges in a polynomial number of iterations with
a probability of 1 [5]. Note that if we fix q = h and z = 1, interval rounding becomes equivalent
to complete rounding. However, the analyses in [5, 10] shows that other values of q provide better
approximation factors for various special cases.
Algorithm 3 The interval rounding procedure.
input A feasible solution y of the LP relaxation.
1: repeat
2:
Pick an integer z uniformly from [?q + 2, h]. Define an interval of labels I = {ls , ? ? ? , le },
where s = max{z, 1} is the start index and e = min{z + q ? 1, h} is the end index.
3:
Pick a real number r uniformly from [0, 1].
4:
for all Unlabeled random variables
a do
PX
s+i?1
5:
Define Ya (0) = 0 and Ya (i) = j=s ya (j) for all i ? {1, ? ? ? , e ? s + 1}.
6:
Assign the label ls+i?1 ? I to the Xa if Ya (i ? 1) < r ? Ya (i).
7:
end for
8: until All random variables have been assigned a label.
Our goal is to design a move-making algorithm whose multiplicative bound matches the approximation factor of interval rounding for any choice of q. To this end, we propose the interval move-making
algorithm that generalizes the range expansion algorithm [15], originally proposed for truncated convex distances, to arbitrary distance functions. Algorithm 4 provides its main steps. The central idea
? by allowing each random variable Xa to either retain
of the method is to improve a given labeling x
its current label x
?a or to choose a new label from an interval of consecutive labels. In more detail, let
I = {ls , ? ? ? , le } ? L be an interval of labels of length at most q (step 4). For
S the sake of simplicity,
let us assume that x
?a ?
/ I for any random variable Xa . We define Ia = I {?
xa } (step 5). For each
pair of neighboring random variables (Xa , Xb ) ? E, we compute a submodular distance function
dx?a ,?xb : Ia ? Ib ? R+ by solving the following linear program (step 6):
dx?a ,?xb =
s.t.
argmin t
(5)
d?
d? (li , lj ) ? td(li , lj ), ?li ? Ia , lj ? Ib ,
d? (li , lj ) ? d(li , lj ), ?li ? Ia , lj ? Ib ,
d? (li , lj ) + d? (li+1 , lj+1 ) ? d? (li , lj+1 ) + d? (li+1 , lj ), ?li , lj ? I\{le },
d? (li , le ) + d? (li+1 , x
?b ) ? d? (li , x
?b ) + d? (li+1 , le ), ?li ? I\{le },
?
?
?
d (le , lj ) + d (?
xa , lj+1 ) ? d (le , lj+1 ) + d? (?
xa , lj ), ?lj ? I\{le },
?
?
?
d (le , le ) + d(?
xa , x
?b ) ? d (le , x
?b ) + d (?
xa , le ).
Similar to problem (4), the above problem minimizes the maximum ratio of the estimated distance
to the original distance. However, instead of introducing constraints for all pairs of labels, it is only
5
considers pairs of labels li and lj where li ? Ia and lj ? Ib . Furthermore, it does not modify the
distance between the current labels x
?a and x
?b (as can be seen in the last constraint of problem (5)).
Given the submodular distance functions dx?a ,?xb , we can compute a new labeling x by solving the
following optimization problem via minimum st-cut using the method of [6] (step 7):
X
X
wab dx?a ,?xb (xa , xb )
x=
argmin
?a (xa ) +
x
s.t.
Xa ?X
(Xa ,Xb )?E
xa ? Ia , ?Xa ? X.
(6)
? , then we update our
If the energy of the new labeling x is less than that of the current labeling x
labeling to x (steps 8-10). Otherwise, we retain the current estimate of the labeling and consider
another interval. The algorithm converges when the energy does not decrease for any interval of
length at most q. Note that, once again, the main difference between interval move-making and the
range expansion algorithm is the use of an appropriate optimization problem, namely the LP (5), to
obtain a submodular overestimation of the given distance function. This allows us to use interval
move-making for the general metric labeling problem, instead of focusing on only truncated convex
models.
Algorithm 4 The interval move-making algorithm.
input Unary potentials ?a (?), edge weights wab , distance function d, initial labeling x0 .
? = x0 .
1: Set current labeling to initial labeling, that is, x
2: repeat
3:
for all z ? [?q + 2, h] do
4:
Define an interval of labels I = {ls , ? ? ? , le }, where s = max{z, 1} is the start index and
e = min{z + qS? 1, h} is the end index.
5:
Define Ia = I {?
xa } for all random variables Xa ? X.
6:
Obtain submodular overestimates dx?a ,?xb for each pair of neighboring random variables
(Xa , Xb ) ? E by solving problem (5).
7:
Obtain a new labeling x by solving problem (6).
? then
8:
if Energy of x is less than energy of x
? = x.
9:
Update x
10:
end if
11:
end for
12: until Energy cannot be decreased further.
The following theorem establishes the theoretical guarantees of the interval move-making algorithm
and the interval rounding procedure.
Theorem 2. The tight multiplicative bound of the interval move-making algorithm is equal to the
tight approximation factor of the interval rounding procedure.
An interval move-making algorithm that uses an interval length of q runs for at most O(h/q) iterations. This follows from a simple modification of the result by Gupta and Tardos [8] (specifically,
theorem 3.7). Hence, the total time complexity of interval move-making is O(nmhq 2 log(m)),
since each iteration solves a minimum st-cut problem of a graph with O(nq) nodes and O(mq 2 )
arcs. In other words, interval move-making is at most as computationally complex as complete
move-making, which in turn is significantly less complex than solving the LP relaxation. Note that
problem (5), which is required for interval move-making, is independent of the unary potentials
and the edge weights. Hence, it only needs to be solved once beforehand for all pairs of labels
(?
xa , x
?b ) ? L ? L in order to obtain a solution for any metric labeling problem defined using the
distance function d.
5
Hierarchical Rounding and Hierarchical Moves
We now consider the most general form of parallel rounding that has been proposed in the literature,
namely the hierarchical rounding procedure [10]. The rounding relies on a hierarchical clustering
of the labels. Formally, we denote a hierarchical clustering of m levels for the label set L by C =
{C(i), i = 1, ? ? ? , m}. At each level i, the clustering C(i) = {C(i, j) ? L, j = 1, ? ? ? , hi } is
6
mutually exclusive and collectively exhaustive, that is,
[
C(i, j) = L, C(i, j) ? C(i, j ? ) = ?, ?j 6= j ? .
j
Furthermore, for each cluster C(i, j) at the level i > 2, there exists a unique cluster C(i ? 1, j ? ) in
the level i ? 1 such that C(i, j) ? C(i ? 1, j ? ). We call the cluster C(i ? 1, j ? ) the parent of the
cluster C(i, j) and define p(i, j) = j ? . Similarly, we call C(i, j) a child of C(i ? 1, j ? ). Without
loss of generality, we assume that there exists a single cluster at level 1 that contains all the labels,
and that each cluster at level m contains a single label.
Algorithm 5 The hierarchical rounding procedure.
input A feasible solution y of the LP relaxation.
1: Define fa1 = 1 for all Xa ? X.
2: for all i ? {2, ? ? ? , m} do
3:
for all Xa ? X do
4:
Define zai (j) for all j ? {1, ? ? ? , hi } as follows:
P
k,lk ?C(i,j) ya (k)
zai (j) =
0
5:
6:
7:
8:
9:
10:
11:
12:
13:
14:
if p(i, j) = fai?1 ,
otherwise.
Define yai (j) for all j ? {1, ? ? ? , hi } as follows:
z i (j)
yai (j) = Phi a
i ?
j ? =1 za (j )
end for
Using a rounding procedure (complete or interval) on yi = [yai (j), ?Xa ? X, j ?
? i.
{1, ? ? ? , hi }], obtain an integer solution y
for all Xa ? X do
Let ka ? {1, ? ? ? , hi } such that y?i (ka ) = 1. Define fai = ka .
end for
end for
for all Xa ? X do
Let lk be the unique label present in the cluster C(m, fam ). Assign lk to Xa .
end for
Algorithm 5 describes the hierarchical rounding procedure. Given a clustering C, it proceeds in a
top-down fashion through the hierarchy while assigning each random variable to a cluster in the
current level. Let fai be the index of the cluster assigned to the random variable Xa in the level
i. In the first step, the rounding procedure assigns all the random variables to the unique cluster
C(1, 1) (step 1). At each step i, it assigns each random variable to a unique cluster in the level i
i
by computing a conditional probability distribution as follows. The conditional
Pprobability ya (j)
of assigning the random variable Xa to the cluster C(i, j) is proportional to lk ?C(i,j) ya (k) if
p(i, j) = fai?1 (steps 3-6). The conditional probability yai (j) = 0 if p(i, j) 6= fai?1 , that is, a
random variable cannot be assigned to a cluster C(i, j) if it wasn?t assigned to its parent in the
previous step. Using a rounding procedure (complete or interval) for yi , we obtain an assignment
of random variables to the clusters at level i (step 7). Once such an assignment is obtained, the
values fai are computed for all random variables Xa (steps 8-10). At the end of step m, hierarchical
rounding would have assigned each random variable to a unique cluster in the level m. Since each
cluster at level m consists of a single label, this provides us with a labeling of the MRF (steps 12-14).
Our goal is to design a move-making algorithm whose multiplicative bound matches the approximation factor of the hierarchical rounding procedure for any choice of hierarchical clustering C. To
this end, we propose the hierarchical move-making algorithm, which extends the hierarchical graph
cuts approach for hierarchically well-separated tree (HST) metrics proposed in [14]. Algorithm 6
provides its main steps. In contrast to hierarchical rounding, the move-making algorithm traverses
the hierarchy in a bottom-up fashion while computing a labeling for each cluster in the current level.
Let xi,j be the labeling corresponding to the cluster C(i, j). At the first step, when considering the
level m of the clustering, all the random variables are assigned the same label. Specifically, xm,j
a
7
Algorithm 6 The hierarchical move-making algorithm.
input Unary potentials ?a (?), edge weights wab , distance function d.
1: for all j ? {1, ? ? ? , h} do
2:
Let lk be the unique label is the cluster C(m, j). Define xm,j
= lk for all Xa ? X.
a
3: end for
4: for all i ? {2, ? ? ? , m} do
5:
for all j ? {1, ? ? ? , hm?i+1 } do
?
6:
Define Lm?i+1,j
= {xm?i+2,j
, p(m ? i + 2, j ? ) = j, j ? ? {1, ? ? ? , hm?i+2 }}.
a
a
7:
Using a move-making algorithm (complete or interval), compute the labeling xm?i+1,j
under the constraint xm?i+1,j
? Lm?i+1,j
.
a
a
8:
end for
9: end for
10: The final solution is x1,1 .
is equal to the unique label contained in the cluster C(m, j) (steps 1-3). At step i, it computes the
labeling xm?i+1,j for each cluster C(m ? i + 1, j) by using the labelings computed in the previous
step. Specifically, it restricts the label assigned to a random variable Xa in the labeling xm?i+1,j
to the subset of labels that were assigned to it by the labelings corresponding to the children of
C(m ? i + 1, j) (step 6). Under this restriction, the labeling xm?i+1,j is computed by approximately minimizing the energy using a move-making algorithm (step 7). Implicit in our description
is the assumption that that we will use a move-making algorithm (complete or interval) in step 7 of
Algorithm 6 whose multiplicative bound matches the approximation factor of the rounding procedure (complete or interval) used in step 7 of Algorithm 5. Note that, unlike the hierarchical graph
cuts approach [14], the hierarchical move-making algorithm can be used for any arbitrary clustering
and not just the one specified by an HST metric.
The following theorem establishes the theoretical guarantees of the hierarchical move-making algorithm and the hierarchical rounding procedure.
Theorem 3. The tight multiplicative bound of the hierarchical move-making algorithm is equal to
the tight approximation factor of the hierarchical rounding procedure.
Note that hierarchical move-making solves a series of problems defined on a smaller label set. Since
the complexity of complete and interval move-making is superlinear in the number of labels, it can
be verified that the hierarchical move-making algorithm is at most as computationally complex as
the complete move-making algorithm (corresponding to the case when the clustering consists of
only one cluster that contains all the labels). Hence, hierarchical move-making is significantly faster
than solving the LP relaxation.
6
Discussion
For any general distance function that can be used to specify the (semi-)metric labeling problem, we
proved that the approximation factor of a large family of parallel rounding procedures is matched by
the multiplicative bound of move-making algorithms. This generalizes previously known results on
the guarantees of move-making algorithms in two ways: (i) in contrast to previous results [14, 15, 20]
that focused on special cases of distance functions, our results are applicable to arbitrary semi-metric
distance functions; and (ii) the guarantees provided by our theorems are tight. Our experiments
(described in the technical report) confirm that the rounding-based move-making algorithms provide
similar accuracy to the LP relaxation, while being significantly faster due to the use of efficient
minimum st-cut solvers.
Several natural questions arise. What is the exact characterization of the rounding procedures for
which it is possible to design matching move-making algorithms? Can we design rounding-based
move-making algorithms for other combinatorial optimization problems? Answering these questions will not only expand our theoretical understanding, but also result in the development of efficient and accurate algorithms.
Acknowledgements. This work is funded by the European Research Council under the European Community?s Seventh Framework Programme (FP7/2007-2013)/ERC Grant agreement number 259112.
8
References
[1] A. Archer, J. Fakcharoenphol, C. Harrelson, R. Krauthgamer, K. Talvar, and E. Tardos. Approximate classification via earthmover metrics. In SODA, 2004.
[2] Y. Boykov and V. Kolmogorov. An experimental comparison of min-cut/max-flow algorithms
for energy minimization in vision. PAMI, 2004.
[3] Y. Boykov, O. Veksler, and R. Zabih. Markov random fields with efficient approximations. In
CVPR, 1998.
[4] Y. Boykov, O. Veksler, and R. Zabih. Fast approximate energy minimization via graph cuts. In
ICCV, 1999.
[5] C. Chekuri, S. Khanna, J. Naor, and L. Zosin. Approximation algorithms for the metric labeling
problem via a new linear programming formulation. In SODA, 2001.
[6] B. Flach and D. Schlesinger. Transforming an arbitrary minsum problem into a binary one.
Technical report, TU Dresden, 2006.
[7] A. Globerson and T. Jaakkola. Fixing max-product: Convergent message passing algorithms
for MAP LP-relaxations. In NIPS, 2007.
[8] A. Gupta and E. Tardos. A constant factor approximation algorithm for a class of classification
problems. In STOC, 2000.
[9] T. Hazan and A. Shashua. Convergent message-passing algorithms for inference over general
graphs with convex free energy. In UAI, 2008.
[10] J. Kleinberg and E. Tardos. Approximation algorithms for classification problems with pairwise relationships: Metric labeling and Markov random fields. In STOC, 1999.
[11] V. Kolmogorov. Convergent tree-reweighted message passing for energy minimization. PAMI,
2006.
[12] N. Komodakis, N. Paragios, and G. Tziritas. MRF optimization via dual decomposition:
Message-passing revisited. In ICCV, 2007.
[13] A. Koster, C. van Hoesel, and A. Kolen. The partial constraint satisfaction problem: Facets
and lifting theorems. Operations Research Letters, 1998.
[14] M. P. Kumar and D. Koller. MAP estimation of semi-metric MRFs via hierarchical graph cuts.
In UAI, 2009.
[15] M. P. Kumar and P. Torr. Improved moves for truncated convex models. In NIPS, 2008.
[16] P. Ravikumar, A. Agarwal, and M. Wainwright. Message-passing for graph-structured linear
programs: Proximal projections, convergence and rounding schemes. In ICML, 2008.
[17] M. Schlesinger. Syntactic analysis of two-dimensional visual signals in noisy conditions.
Kibernetika, 1976.
[18] R. Szeliski, R. Zabih, D. Scharstein, O. Veksler, V. Kolmogorov, A. Agarwala, M. Tappen, and
C. Rother. A comparative study of energy minimization methods for Markov random fields
with smoothness-based priors. PAMI, 2008.
[19] D. Tarlow, D. Batra, P. Kohli, and V. Kolmogorov. Dynamic tree block coordinate ascent. In
ICML, 2011.
[20] O. Veksler. Efficient graph-based energy minimization methods in computer vision. PhD thesis,
Cornell University, 1999.
[21] O. Veksler. Graph cut based optimization for MRFs with truncated convex priors. In CVPR,
2007.
[22] M. Wainwright, T. Jaakkola, and A. Willsky. MAP estimation via agreement on trees: Message
passing and linear programming. Transactions on Information Theory, 2005.
[23] Y. Weiss, C. Yanover, and T. Meltzer. MAP estimation, linear programming and belief propagation with convex free energies. In UAI, 2007.
[24] T. Werner. A linear programming approach to max-sum problem: A review. PAMI, 2007.
[25] T. Werner. Revisting the linear programming relaxation approach to Gibbs energy minimization and weighted constraint satisfaction. PAMI, 2010.
9
| 5354 |@word kohli:1 polynomial:5 flach:1 decomposition:1 pick:3 initial:2 contains:3 series:1 ecole:1 current:7 ka:3 assigning:4 dx:5 readily:1 designed:1 update:2 v:1 selected:1 nq:1 tarlow:1 characterization:1 provides:7 node:2 revisited:1 traverse:1 constructed:1 consists:7 prove:3 naor:1 x0:2 pairwise:7 td:2 solver:3 considering:1 becomes:1 provided:3 matched:1 what:1 argmin:4 minimizes:2 developed:1 finding:1 guarantee:11 exactly:1 grant:1 overestimate:1 treat:1 modify:2 despite:1 approximately:1 pami:5 inria:1 dresden:1 relaxing:1 range:4 unique:7 globerson:1 block:1 procedure:39 significantly:5 matching:1 projection:1 word:2 refers:2 cannot:2 unlabeled:2 superlinear:1 restriction:1 equivalent:3 map:4 attention:1 l:4 convex:9 focused:1 formulate:1 minsum:1 simplicity:1 assigns:2 fam:1 m2:1 q:1 mq:1 coordinate:1 tardos:4 hierarchy:2 play:1 user:1 exact:1 programming:9 us:3 designing:1 agreement:2 associate:1 tappen:1 cut:15 bottom:1 role:1 solved:3 capture:1 trade:1 decrease:1 intuition:1 transforming:1 complexity:4 overestimation:7 reparameterizations:2 dynamic:1 depend:1 solving:14 tight:9 mh:1 various:4 kolmogorov:4 separated:1 fast:2 describe:1 labeling:51 neighborhood:2 exhaustive:1 whose:6 solve:3 cvpr:2 distortion:4 otherwise:2 zosin:1 syntactic:1 noisy:1 final:2 propose:2 product:1 fr:1 neighboring:3 tu:1 zai:2 description:2 parent:2 cluster:22 convergence:1 comparative:1 maxi6:1 converges:2 help:1 illustrate:1 fixing:1 ij:1 h3:2 solves:2 hst:2 tziritas:1 implies:2 closely:1 subsequently:2 yai:4 assign:5 fix:1 generalization:1 preliminary:1 strictly:1 accompanying:1 hold:2 lm:2 consecutive:2 estimation:3 applicable:2 label:49 combinatorial:1 council:1 establishes:3 weighted:1 minimization:9 always:1 sight:1 avoid:1 cornell:1 jaakkola:2 conjunction:1 corollary:1 focus:1 contrast:4 inference:1 mrfs:2 unary:18 lj:56 koller:1 expand:1 archer:1 labelings:4 interested:2 overall:1 classification:3 dual:1 agarwala:1 development:2 special:7 field:5 equal:8 once:4 sampling:1 icml:2 mimic:1 np:2 report:3 quantitatively:2 simultaneously:1 pawan:2 ab:2 attempt:1 message:6 highly:1 deferred:1 xb:24 accurate:2 beforehand:2 edge:14 integral:1 partial:1 lh:6 tree:4 overcomplete:1 theoretical:10 schlesinger:2 facet:1 cover:1 assignment:3 werner:2 introducing:1 vertex:2 subset:1 veksler:5 rounding:59 seventh:1 proximal:1 chooses:1 st:9 randomized:5 retain:2 off:1 rounded:1 thesis:1 again:1 central:2 satisfied:2 opposed:1 hn:1 choose:1 li:71 potential:19 kolen:1 includes:1 availability:1 multiplicative:16 hazan:1 shashua:1 start:3 parallel:7 minimize:1 accuracy:3 wab:13 za:1 fa1:1 energy:26 proof:1 proved:1 popular:3 recall:1 fractional:5 appears:1 focusing:2 originally:1 specify:3 improved:1 wei:1 yb:4 formulation:2 generality:1 furthermore:3 xa:66 stage:1 implicit:1 chekuri:2 until:2 just:1 mh2:2 propagation:1 fai:6 khanna:1 equality:2 assigned:14 hence:3 iteratively:2 reweighted:1 komodakis:1 complete:28 covert:1 l1:3 boykov:3 common:2 specialized:1 nh:1 extend:1 discussed:1 refer:3 gibbs:1 smoothness:1 similarly:1 erc:1 submodular:18 funded:1 inequality:5 binary:1 yi:2 seen:1 minimum:11 greater:3 signal:1 semi:5 ii:5 technical:3 match:7 faster:3 characterized:1 ravikumar:1 mrf:7 vision:5 metric:32 expectation:1 iteration:4 agarwal:1 interval:34 decreased:1 appropriately:1 unlike:2 ascent:1 undirected:1 flow:1 integer:7 call:4 meltzer:1 idea:1 wasn:1 passing:6 useful:2 generally:1 yab:9 zabih:3 specifies:1 exist:1 restricts:1 estimated:3 discrete:1 dropping:1 verified:1 graph:13 relaxation:23 convert:1 sum:2 run:1 koster:1 letter:1 soda:2 extends:1 throughout:1 family:1 ecp:1 putative:2 bound:16 hi:5 convergent:3 replaces:1 constraint:8 inclusive:1 x2:1 sake:1 generates:1 kleinberg:1 speed:1 min:5 kumar:4 px:1 structured:1 centrale:1 describes:3 beneficial:1 smaller:1 lp:25 making:48 modification:1 intuitively:1 invariant:2 pr:1 iccv:2 taken:1 ln:2 computationally:2 mutually:1 previously:3 turn:2 fp7:1 end:17 available:1 tightest:1 generalizes:2 operation:1 hierarchical:24 appropriate:1 distinguished:2 shortly:1 slower:1 original:4 top:1 clustering:8 include:1 krauthgamer:1 move:53 question:2 exclusive:1 said:4 distance:44 considers:2 reason:1 willsky:1 rother:1 length:3 index:5 relationship:2 ratio:2 minimizing:1 stoc:2 negative:3 design:5 allowing:1 markov:5 arc:2 finite:3 truncated:8 arbitrary:11 community:2 pair:6 paris:1 specified:3 namely:3 required:1 nip:2 proceeds:1 xm:8 saclay:1 program:5 max:5 belief:1 wainwright:2 ia:7 satisfaction:2 natural:1 yanover:1 scheme:2 improve:2 lk:6 hoesel:1 hm:2 prior:2 literature:4 l2:3 understanding:1 acknowledgement:1 review:1 loss:1 proportional:3 repeat:2 last:1 free:2 szeliski:1 kibernetika:1 van:1 xn:1 computes:3 programme:1 transaction:1 scharstein:1 approximate:7 obtains:1 clique:1 confirm:1 uai:3 xi:1 obtaining:1 expansion:4 complex:3 european:2 main:5 hierarchically:1 rh:1 arise:1 child:2 x1:2 fashion:2 fakcharoenphol:1 paragios:1 answering:1 ib:4 theorem:11 down:1 pprobability:1 gupta:2 exists:2 lifting:1 phd:1 flavor:1 visual:1 contained:1 phi:1 collectively:1 satisfies:2 determines:1 relies:1 conditional:3 goal:2 feasible:8 hard:2 specifically:5 infinite:1 uniformly:3 torr:1 total:1 batra:1 experimental:2 ya:30 m3:1 formally:2 |
4,810 | 5,355 | Stochastic Gradient Descent, Weighted Sampling, and
the Randomized Kaczmarz algorithm
Nathan Srebro
Toyota Technological Institute at Chicago
and Dept. of Computer Science, Technion
[email protected]
Deanna Needell
Department of Mathematical Sciences
Claremont McKenna College
Claremont CA 91711
[email protected]
Rachel Ward
Department of Mathematics
Univ. of Texas, Austin
[email protected]
Abstract
We improve a recent guarantee of Bach and Moulines on the linear convergence
of SGD for smooth and strongly convex objectives, reducing a quadratic dependence on the strong convexity to a linear dependence. Furthermore, we show how
reweighting the sampling distribution (i.e. importance sampling) is necessary in
order to further improve convergence, and obtain a linear dependence on average
smoothness, dominating previous results, and more broadly discus how importance sampling for SGD can improve convergence also in other scenarios. Our
results are based on a connection between SGD and the randomized Kaczmarz algorithm, which allows us to transfer ideas between the separate bodies of literature
studying each of the two methods.
1
Introduction
This paper concerns two algorithms which until now have remained somewhat disjoint in the literature: the randomized Kaczmarz algorithm for solving linear systems and the stochastic gradient
descent (SGD) method for optimizing a convex objective using unbiased gradient estimates. The
connection enables us to make contributions by borrowing from each body of literature to the other.
In particular, it helps us highlight the role of weighted sampling for SGD and obtain a tighter guarantee on the linear convergence regime of SGD.
Our starting point is a recent analysis on convergence of the SGD iterates. Considering a stochastic
objective F (x) = Ei [fi (x)], classical analyses of SGD show a polynomial rate on the suboptimality
of the objective value F (xk ) ? F (x? ). Bach and Moulines [1] showed that if F (x) is ?-strongly
convex, fi (x) are Li -smooth (i.e. their gradients are Li -Lipschitz), and x? is a minimizer of (almost)
all fi (x) (i.e. Pi (?fi (x? ) = 0) = 1), then Ekxk ? x? k goes to zero exponentially, rather then
polynomially, in k. That is, reaching a desired accuracy of Ekxk ? x? k2 ? ? requires a number of
steps that scales only logarithmically in 1/?. Bach and Moulines?s bound on the required number of
iterations further depends on the average squared conditioning number E[(Li /?)2 ].
In a seemingly independent line of research, the Kaczmarz method was proposed as an iterative
method for solving overdetermined systems of linear equations [7]. The simplicity of the method
makes it popular in applications ranging from computer tomography to digital signal processing [5,
1
9, 6]. Recently, Strohmer and Vershynin [19] proposed a variant of the Kaczmarz method which
selects rows with probability proportional to their squared norm, and showed that using this selection
strategy, a desired accuracy of ? can be reached in the noiseless setting in a number of steps that
scales with log(1/?) and only linearly in the condition number. As we discuss in Section 5, the
randomized Kaczmarz algorithm is in fact a special case of stochastic gradient descent.
Inspired by the above analysis, we prove improved convergence results for generic SGD, as well as
for SGD with gradient estimates chosen based on a weighted sampling distribution, highlighting the
role of importance sampling in SGD:
We first show that without perturbing the sampling distribution, we can obtain a linear dependence
on the uniform conditioning (sup Li /?), but it is not possible to obtain a linear dependence on
the average conditioning E[Li ]/?. This is a quadratic improvement over [1] in regimes where the
components have similar Lipschitz constants (Theorem 2.1 in Section 2).
We then show that with weighted sampling we can obtain a linear dependence on the average conditioning E[Li ]/?, dominating the quadratic dependence of [1] (Corollary 3.1 in Section 3).
In Section 4, we show how also for smooth but not-strongly-convex objectives, importance sampling
can improve a dependence on a uniform bound over smoothness, (sup Li ), to a dependence on the
average smoothness E[Li ]?such an improvement is not possible without importance sampling.
For non-smooth objectives, we show that importance sampling can eliminate a dependence on the
variance in the Lipschitz constants of the components.
Finally, in Section 5, we turn to the Kaczmarz algorithm, and show we can improve known guarantees in this context as well.
2
SGD for Strongly Convex Smooth Optimization
We consider the problem of minimizing a strongly convex function of the form F (x) = Ei?D fi (x)
where fi : H ? R are smooth functionals over H = Rd endowed with the standard Euclidean
norm k?k2 , or over a Hilbert space H with the norm k?k2 . Here i is drawn from some source
distribution D over an arbitrary probability space. Throughout this manuscript, unless explicitly
specified otherwise, expectations will be with respect to indices drawn from the source distribution
D. We denote the unique minimum x? = arg min F (x) and denote by ? 2 the ?residual? quantity at
the minimum, ? 2 = Ek?fi (x? )k22 .
Assumptions Our bounds will be based on the following assumptions and quantities: First, F has
strong convexity parameter ?; that is, hx ? y, ?F (x) ? ?F (y)i ? ?kx ? yk22 for all vectors x
and y. Second, each fi is continuously differentiable and the gradient function ?fi has Lipschitz
constant Li ; that is, k?fi (x) ? ?fi (y)k2 ? Li kx ? yk2 for all vectors x and y. We denote sup L
the supremum of the support of Li , i.e. the smallest L such that Li ? L a.s., and similarly denote
inf L the infimum. We denote the average Lipschitz constant as L = ELi .
An unbiased gradient estimate for F (x) can be obtained by drawing i ? D and using ?fi (x) as the
estimate. The SGD updates with (fixed) step size ? based on these gradient estimates are given by:
xk+1 ? xk ? ??fik (xk )
(2.1)
x? k22
where {ik } are drawn i.i.d. from D. We are interested in the distance kxk ?
from the unique minimum, and denote the initial distance by ?0 = kx0 ? x? k22 .
of the iterates
Bach and Moulines [1, Theorem 1] considered this setting1 and established that
EL2
?2
i
k = 2 log(?0 /?)
+ 2
(2.2)
2
?
? ?
SGD iterations of the form (2.1), with an appropriate step-size, are sufficient to ensure
Ekxk ? x? k22 ? ?, where the expectation is over the random sampling. As long as ? 2 = 0, i.e. the
1
Bach and Moulines?s results are somewhat more general. Their Lipschitz requirement is a bit weaker and
more complicated, but in terms of Li yields (2.2). They also study the use of polynomial decaying step-sizes,
but these do not lead to improved runtime if the target accuracy is known ahead of time.
2
same minimizer x? minimizes all components fi (x) (though of course it need not be a unique minimizer of any of them); this yields linear convergence to x? , with a graceful degradation as ? 2 > 0.
However, in the linear convergence regime, the number of required iterations scales with the expected squared conditioning EL2i /?2 . In this paper, we reduce this quadratic dependence to a linear
dependence. We begin with a guarantee ensuring linear dependence on sup L/?:
Theorem 2.1 Let each fi be convex where ?fi has Lipschitz constant Li , with Li ? sup L a.s.,
and let F (x) = Efi (x) be ?-strongly convex. Set ? 2 = Ek?fi (x? )k22 , where x? = argminx F (x).
Suppose that ? ? 1/?. Then the SGD iterates given by (2.1) satisfy:
h
ik
?? 2
.
Ekxk ? x? k22 ? 1 ? 2??(1 ? ? sup L)
kx0 ? x? k22 +
(2.3)
? 1 ? ? sup L
That is, for any desired ?, using a step-size of
?=
??
2?? sup L + 2? 2
sup L
?2
ensures that after k = 2 log(?0 /?)
+ 2
?
? ?
(2.4)
SGD iterations, Ekxk ? x? k22 ? ?, where ?0 = kx0 ? x? k22 and where both expectations are with
respect to the sampling of {ik }.
Proof sketch: The crux of the improvement over [1] is a tighter recursive equation. Instead of:
kxk+1 ? x? k22 ? 1 ? 2?? + 2? 2 L2ik kxk ? x? k22 + 2? 2 ? 2 ,
we use the co-coercivity Lemma (Lemma A.1 in the supplemental material) to obtain:
kxk+1 ? x? k22 ? 1 ? 2?? + 2? 2 ?Lik kxk ? x? k22 + 2? 2 ? 2 .
The significant difference is that one of the factors of Lik , an upper bound on the second derivative
(where ik is the random index selected in the kth iteration) in the third term inside the parenthesis,
is replaced by ?, a lower bound on the second derivative of F . A complete proof can be found in the
supplemental material.
Comparison to [1] Our bound (2.4) improves a quadratic dependence on ?2 to a linear dependence and replaces the dependence on the average squared smoothness EL2i with a linear dependence
on the smoothness bound sup L. When all Lipschitz constants Li are of similar magnitude, this is a
quadratic improvement in the number of required iterations. However, when different components
fi have widely different scaling, i.e. Li are highly variable, the supremum might be significantly
larger then the average square conditioning.
Tightness Considering the above, one might hope to obtain a linear dependence on the average
smoothness L. However, as the following example shows, this is not possible. Consider a uniform
source distribution over N + 1 quadratics, with the first quadratic f1 being N (x[1] ? b)2 and all
others being x[2]2 , and b = ?1. Any method must examine f1 in order to recover x to within
error less then one, but by uniformly sampling indices i, this takes N iterations in expectation.
2
?1)
?1)
We can calculate sup L = L1 = 2N , L = 2(2N
, EL2i = 4(N +N
, and ? = 1. Both
N
N
2
2
sup L/? = ELi /? = O(N ) scale correctly with the expected number of iterations, while error
reduction in O(L/?) = O(1) iterations is not possible for this example.
We therefore see that the choice between EL2i and sup L is unavoidable. In the next Section, we
will show how we can obtain a linear dependence on the average smoothness L, using importance
sampling, i.e. by sampling from a modified distribution.
3
Importance Sampling
For a weight function w(i) which assigns a non-negative weight w(i) ? 0 to each index i, the
weighted distribution D(w) is defined as the distribution such that
PD(w) (I) ? Ei ?D [1I (i)w(i)] ,
3
where I is an event (subset of indices) and 1I (?) its indicator function. For a discrete distribution
D with probability mass function p(i) this corresponds to weighting the probabilities to obtain a
new probability mass function, which we write as p(w) (i) ? w(i)p(i). Similarly, for a continuous
distribution, this corresponds to multiplying the density by w(i) and renormalizing. Importance
sampling has appeared in both the Kaczmarz method [19] and in coordinate-descent methods [14,
15], where the weights are proportional to some power of the Lipschitz constants (of the gradient
coordinates). Here we analyze this type of sampling in the context of SGD.
One way to construct D(w) is through rejection sampling: sample i ? D, and accept with probability
w(i)/W , for some W ? supi w(i). Otherwise, reject and continue to re-sample until a suggestion
i is accepted. The accepted samples are then distributed according to D(w) .
We use E(w) [?] = Ei?D(w) [?] to denote expectation where indices are sampled from the weighted
distribution D(w) . An important property of such an expectation is that for any quantity X(i):
h
i
1
E(w) w(i)
X(i) = E [w(i)] ? E [X(i)] ,
(3.1)
where recall that the expectationsh on the r.h.s.
i are with respect to i ? D. In particular, when
1
E[w(i)] = 1, we have that E(w) w(i)
X(i) = EX(i). In fact, we will consider only weights
s.t. E[w(i)] = 1, and refer to such weights as normalized.
Reweighted SGD For any normalized weight function w(i), we can write:
(w)
fi
(x) =
1
fi (x) and
w(i)
(w)
F (x) = E(w) [fi
(x)].
(3.2)
This is an equivalent, and equally valid, stochastic representation of the objective F (x), and we can
just as well base SGD on this representation. In this case, at each iteration we sample i ? D(w)
(w)
1
and then use ?fi (x) = w(i)
?fi (x) as an unbiased gradient estimate. SGD iterates based on the
representation (3.2), which we will refer to as w-weighted SGD, are then given by
xk+1 ? xk ?
?
?fik (xk )
w(ik )
(3.3)
where {ik } are drawn i.i.d. from D(w) .
The important observation here is that all SGD guarantees are equally valid for the w-weighted
updates (3.3)?the objective is the same objective F (x), the sub-optimality is the same, and the
minimizer x? is the same. We do need, however, to calculate the relevant quantities controlling SGD
(w)
convergence with respect to the modified components fi and the weighted distribution D(w) .
Strongly Convex Smooth Optimization using Weighted SGD We now return to the analysis of
strongly convex smooth optimization and investigate how re-weighting can yield a better guarantee.
(w)
(w)
(w)
1
Li .
The Lipschitz constant Li of each component fi is now scaled, and we have Li = w(i)
The supremum is then given by:
(w)
sup L(w) = sup Li
= sup
i
i
Li
.
w(i)
(3.4)
It is easy to verify that (3.4) is minimized by the weights
w(i) =
Li
,
L
so that
sup L(w) = sup
i
Li
= L.
Li /L
(3.5)
Before applying Theorem 2.1, we must also calculate:
(w)
2
?(w)
= E(w) [k?fi
(x? )k22 ] = E[
1
L
L 2
k?fi (x? )k22 ] = E[ k?fi (x? )k22 ] ?
? .
w(i)
Li
inf L
4
(3.6)
Now, applying Theorem 2.1 to the w-weighted SGD iterates (3.3) with weights (3.5), we have that,
with an appropriate stepsize,
2
sup L
L
?(w)
?2
L
(w)
k = 2 log(?0 /?)
= 2 log(?0 /?)
(3.7)
+ 2
+
? 2
?
? ?
?
inf L ? ?
iterations are sufficient for E(w) kxk ? x? k22 ? ?, where x? , ? and ?0 are exactly as in Theorem 2.1.
If ? 2 = 0, i.e. we are in the ?realizable? situation, with true linear convergence, then we also have
2
?(w)
= 0. In this case, we already obtain the desired guarantee: linear convergence with a linear
dependence on the average conditioning L/?, strictly improving over the best known results [1].
However, when ? 2 > 0 we get a dissatisfying scaling of the second term, by a factor of L/inf L.
Fortunately, we can easily overcome this factor. To do so, consider sampling from a distribution
which is a mixture of the original source distribution and its re-weighting:
1 1 Li
w(i) = + ? .
(3.8)
2 2 L
We refer to this as partially biased sampling. Instead of an even mixture as in (3.9), we could also
use a mixture with any other constant proportion, i.e. w(i) = ? + (1 ? ?)Li /L for 0 < ? < 1.
Using these weights, we have
1
1
2
sup L(w) = sup 1 1 Li Li ? 2L and ?(w)
= E[ 1 1 Li k?fi (x? )k22 ] ? 2? 2 . (3.9)
i 2 + 2 ?
2 + 2 ? L
L
Corollary 3.1 Let each fi be convex where ?fi has Lipschitz constant Li and let F (x) =
Ei?D [fi (x)], where F (x) is ?-strongly convex. Set ? 2 = Ek?fi (x? )k22 , where x? =
argminx F (x). For any desired ?, using a stepsize of
L
?2
??
ensures
that
after
k
=
4
log(?
/?)
+
(3.10)
?=
0
?
?2 ?
4(??L + ? 2 )
iterations of w-weighted SGD (3.3) with weights specified by (3.8), E(w) kxk ? x? k22 ? ?, where
?0 = kx0 ? x? k22 and L = ELi .
This result follows by substituting (3.9) into Theorem 2.1. We now obtain the desired linear scaling
on L/?, without introducing any additional factor to the residual term, except for a constant factor.
We thus obtain a result which dominates Bach and Moulines (up to a factor of 2) and substantially
improves upon it (with a linear rather than quadratic dependence on the conditioning). Such ?partially biased weights? are not only an analysis trick, but might indeed improve actual performance
over either no weighting or the ?fully biased? weights (3.5), as demonstrated in Figure 1.
Implementing Importance Sampling In settings where linear systems need to be solved repeatedly, or when the Lipschitz constants are easily computed from the data, it is straightforward to
sample by the weighted distribution. However, when we only have sampling access to the source
distribution D (or the implied distribution over gradient estimates), importance sampling might be
difficult. In light of the above results, one could use rejection sampling to simulate sampling from
D(w) . For the weights (3.5), this can be done by accepting samples with probability proportional
to Li / sup L. The overall probability of accepting a sample is then L/ sup L, introducing an additional factor of sup L/L. This yields a sample complexity with a linear dependence on sup L, as
in Theorem 2.1, but a reduction in the number of actual gradient calculations and updates. In even
less favorable situations, if Lipschitz constants cannot be bounded for individual components, even
importance sampling might not be possible.
4
Importance Sampling for SGD in Other Scenarios
In the previous Section, we considered SGD for smooth and strongly convex objectives, and were
particularly interested in the regime where the residual ? 2 is low, and the linear convergence term is
dominant. Weighted SGD is useful also in other scenarios, and we now briefly survey them, as well
as relate them to our main scenario of interest.
5
Error || xk ? x* ||2
Error || xk ? x* ||2
?=0
? = 0.2
?=1
1
0
10
?1
10
0
10
10
10
0
1000
2000
Iteration k
3000
0
10
?1
?1
10
?=0
? = 0.2
?=1
1
10
Error || xk ? x* ||2
?=0
? = 0.2
?=1
1
10
0
1000
2000 3000
Iteration k
4000
5000
0
1000
2000 3000
Iteration k
4000
5000
Figure 1: Performance of SGD with weights w(i) = ? + (1 ? ?) LLi on synthetic overdetermined least
squares problems of the form (5.1) (? = 1 is unweighted, ? = 0 is fully weighted). Left: ai are standard
spherical Gaussian, bi = hai , x0 i + N (0, 0.12 ). Center: ai is spherical Gaussian with variance i, bi =
hai , x0 i + N (0, 202 ). Right: ai is spherical Gaussian with variance i, bi = hai , x0 i + N (0, 0.12 ). In all
cases, matrix A with rows ai is 1000 ? 100 and the corresponding least squares problem is strongly convex;
the stepsize was chosen as in (3.10).
2
2
?=0
?=.4
?=1
0
10
?2
10
Error F(xk) ? F(x*)
k
0
10
Error F(xk) ? F(x*)
10
?=0
?=.4
?=1
*
Error F(x ) ? F(x )
10
?=0
? = .4
?=1
1
10
0
0
2000
4000
Iteration k
6000
8000
0
5000
10000
Iteration k
15000
10
0
5000
10000
Iteration k
15000
Figure 2: Performance of SGD with weights w(i) = ? + (1 ? ?) LLi on synthetic underdetermined least
squares problems of the form (5.1) (? = 1 is unweighted, ? = 0 is fully weighted). We consider 3 cases. Left:
ai are standard spherical Gaussian, bi = hai , x0 i+N (0, 0.12 ). Center: ai is spherical Gaussian with variance
i, bi = hai , x0 i + N (0, 202 ). Right: ai is spherical Gaussian with variance i, bi = hai , x0 i + N (0, 0.12 ).
In all cases, matrix A with rows ai is 50 ? 100 and so the corresponding least squares problem is not strongly
convex; the step-size was chosen as in (3.10).
Smooth, Not Strongly Convex When each component fi is convex, non-negative, and has an
Li -Lipschitz gradient, but the objective F (x) is not necessarily strongly convex, then after
(sup L)kx? k22 F (x? ) + ?
k=O
?
(4.1)
?
?
iterations of SGD with an appropriately chosen step-size we will have F (xk ) ? F (x? ) + ?, where
xk is an appropriate averaging of the k iterates [18]. The relevant quantity here determining the iteration complexity is again sup L. Furthermore, the dependence on the supremum is unavoidable and
cannot be replaced with the average Lipschitz constant L [3, 18]: if we sample gradients according
to the source distribution D, we must have a linear dependence on sup L.
The only quantity in the bound (4.1) that changes with a re-weighting is sup L?all other quantities
(kx? k22 , F (x? ), and the sub-optimality ?) are invariant to re-weightings. We can therefore replace
the dependence on sup L with a dependence on sup L(w) by using a weighted SGD as in (3.3). As we
already calculated, the optimal weights are given by (3.5), and using them we have sup L(w) = L.
In this case, there is no need for partially biased sampling, and we obtain that
Lkx? k22 F (x? ) + ?
k=O
?
(4.2)
?
?
iterations of weighed SGD updates (3.3) using the weights (3.5) suffice. Empirical evidence suggests
that this is not a theoretical artifact; full weighted sampling indeed exhibits better convergence rates
compared to partially biased sampling in the non-strongly convex setting (see Figure 2), in contrast
6
to the strongly convex regime (see Figure 1). We again see that using importance sampling allows us
to reduce the dependence on sup L, which is unavoidable without biased sampling, to a dependence
on L. An interesting question for further consideration is to what extent importance sampling can
also help with stochastic optimization procedures such as SAG [8] and SDCA [17] which achieve
faster convergence on finite data sets. Indeed, weighted sampling was shown empirically to achieve
faster convergence rates for SAG [16], but theoretical guarantees remain open.
Non-Smooth Objectives We now turn to non-smooth objectives, where the components fi might
not be smooth, but each component is Gi -Lipschitz. Roughly speaking, Gi is a bound on the first
derivative (the subgradients) of fi , while Li is a bound on the second derivatives of fi . Here,
the performance of SGD (actually stochastic subgradient decent) depends on the second moment
G2 = E[G2i ] [12]. The precise iteration complexity depends on whether the objective is strongly
convex or whether x? is bounded, but in either case depends linearly on G2 .
Using weighted SGD, we get linear dependence on
2
2
i
h
Gi
Gi
(w)
G2(w) = E(w) (Gi )2 = E(w)
=
E
w(i)2
w(i)
(w)
(4.3)
(w)
where Gi = Gi /w(i) is the Lipschitz constant of the scaled fi . This is minimized by the
2
weights w(i) = Gi /G, where G = EGi , yielding G2(w) = G . Using importance sampling, we
2
therefore reduce the dependence on G2 to a dependence on G . Its helpful to recall that G2 =
2
G + Var[Gi ]. What we save is thus exactly the variance of the Lipschitz constants Gi .
Parallel work we recently became aware of [22] shows a similar improvement for a non-smooth
composite objective. Rather than relying on a specialized analysis as in [22], here we show this
follows from SGD analysis applied to different gradient estimates.
Non-Realizable Regime Returning to the smooth and strongly convex setting of Sections 2 and 3,
let us consider more carefully the residual term ? 2 = Ek?fi (x? )k22 . This quantity depends on the
weighting, and in Section 3, we avoided increasing it, introducing partial biasing for this purpose.
However, if this is the dominant term, we might want to choose weights to minimize this term. The
optimal weights here would be proportional to k?fi (x? )k2 , which is not known in general.
An alternative approach is to bound k?fi (x? )k2 ? Gi and so ? 2 ? G2 . Taking this bound, we are
back to the same quantity as in the non-smooth case, and the optimal weights are proportional to Gi .
Note that this differs from using weights proportional to Li , which optimize the linear-convergence
term as studied in Section 3.
To understand how weighting according to Gi and Li are different, consider a generalized linear
objective fi (x) = ?i (hzi , xi), where ?i is a scalar function with bounded |?0i | , |?00i |. We have
that Gi ? kzi k2 while Li ? kzi k22 . Weighting according to (3.5), versus weighting with w(i) =
Gi /G, thus corresponds to weighting according to kzi k22 versus kzi k2 , and are rather different. E.g.,
weighting by Li ? kzi k22 yields G2(w) = G2 : the same sub-optimal dependence as if no weighting
at all were used. A good solution could be to weight by a mixture of Gi and Li , as in the partial
weighting scheme of Section 3.
5
The least squares case and the Randomized Kaczmarz Method
A special case of interest is the least squares problem, where
n
F (x) =
1
1X
(hai , xi ? bi )2 = kAx ? bk22
2 i=1
2
(5.1)
with b ? Cn , A an n ? d matrix with rows ai , and x? = argminx 21 kAx ? bk22 is the least-squares
solution. We can also write (5.1) as a stochastic objective, where the source distribution D is uniform
over {1, 2, . . . , n} and fi = n2 (hai , xi ? bi )2 . In this setting, ? 2 = kAx? ? bk22 is the residual error
7
at the least squares solution x? , which can also be interpreted as noise variance in a linear regression
model.
The randomized Kaczmarz method introduced for solving the least squares problem (5.1) in the
case where A is an overdetermined full-rank matrix, begins with an arbitrary estimate x0 , and in the
kth iteration selects a row i at random from the matrix A and iterates by:
xk+1 = xk + c ?
bi ? hai , xk i
ai ,
kai k22
(5.2)
where c = 1 in the standard method. This is almost an SGD update with step-size ? = c/n, except
for the scaling by kai k22 .
Strohmer and Vershynin [19] provided the first non-asymptotic convergence rates, showing that
drawing rows proportionally to kai k22 leads to provable exponential convergence in expectation [19].
With such a weighting, (5.2) is exactly weighted SGD, as in (3.3), with the fully biased weights
(3.5).
The reduction of the quadratic dependence on the conditioning to a linear dependence in Theorem
2.1, and the use of biased sampling, was inspired by the analysis of [19]. Indeed, applying Theorem
2.1 to the weighted SGD iterates with weights as in (3.5) and a stepsize of ? = 1 yields precisely the
guarantee of [19]. Furthermore, understanding the randomized Kaczmarz method as SGD, allows
us to obtain the following improvements:
Partially Biased Sampling. Using partially biased sampling weights (3.8) yields a better dependence on the residual over the fully biased sampling weights (3.5) considered by [19].
Using Step-sizes. The randomized Kaczmarz method with weighted sampling exhibits exponential
convergence, but only to within a radius, or convergence horizon, of the least-squares solution [19,
10]. This is because a step-size of ? = 1 is used, and so the second term in (2.3) does not vanish.
It has been shown [21, 2, 20, 4, 11] that changing the step size can allow for convergence inside of
this convergence horizon, but only asymptotically. Our results allow for finite-iteration guarantees
with arbitrary step-sizes and can be immediately applied to this setting.
Uniform Row Selection. Strohmer and Vershynin?s variant of the randomized Kaczmarz method
calls for weighted row sampling, and thus requires pre-computing all the row norms. Although
certainly possible in some applications, in other cases this might be better avoided. Understanding
the randomized Kaczmarz as SGD allows us to apply Theorem 2.1 also with uniform weights (i.e. to
the unbiased SGD), and obtain a randomized Kaczmarz using uniform sampling, which converges
to the least-squares solution and enjoys finite-iteration guarantees.
6
Conclusion
We consider this paper as making three main contributions. First, we improve the dependence on
the conditioning for smooth and strongly convex SGD from quadratic to linear. Second, we investigate SGD and importance sampling and show how it can yield improvements not possible without
reweighting. Lastly, we make connections between SGD and the randomized Kaczmarz method.
This connection along with our new results show that the choice in step-size of the Kaczmarz method
offers a tradeoff between convergence rate and horizon and also allows for a convergence bound
when the rows are sampled uniformly.
For simplicity, we only considered SGD with fixed step-size ?, which is appropriate when the target
accuracy in known in advance. Our analysis can be adapted also to decaying step-sizes.
Our discussion of importance sampling is limited to a static reweighting of the sampling distribution.
A more sophisticated approach would be to update the sampling distribution dynamically as the
method progresses, and as we gain more information about the relative importance of components
(e.g. about k?fi (x? )k). Such dynamic sampling is sometimes attempted heuristically, and obtaining
a rigorous framework for this would be desirable.
8
References
[1] F. Bach and E. Moulines. Non-asymptotic analysis of stochastic approximation algorithms for machine
learning. Advances in Neural Information Processing Systems (NIPS), 2011.
[2] Y. Censor, P. P. B. Eggermont, and D. Gordon. Strong underrelaxation in Kaczmarz?s method for inconsistent systems. Numerische Mathematik, 41(1):83?92, 1983.
[3] R. Foygel and N. Srebro. Concentration-based guarantees for low-rank matrix reconstruction. 24th Ann.
Conf. Learning Theory (COLT), 2011.
[4] M. Hanke and W. Niethammer. On the acceleration of Kaczmarz?s method for inconsistent linear systems.
Linear Algebra and its Applications, 130:83?98, 1990.
[5] G. T. Herman. Fundamentals of computerized tomography: image reconstruction from projections.
Springer, 2009.
[6] G. N Hounsfield. Computerized transverse axial scanning (tomography): Part 1. description of system.
British Journal of Radiology, 46(552):1016?1022, 1973.
[7] S. Kaczmarz. Angen?aherte aufl?osung von systemen linearer gleichungen. Bull. Int. Acad. Polon. Sci. Lett.
Ser. A, pages 335?357, 1937.
[8] N. Le Roux, M. W. Schmidt, and F. Bach. A stochastic gradient method with an exponential convergence
rate for finite training sets. Advances in Neural Information Processing Systems (NIPS), pages 2672?2680,
2012.
[9] F. Natterer. The mathematics of computerized tomography, volume 32 of Classics in Applied Mathematics. Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA, 2001. ISBN 0-89871-4931. doi: 10.1137/1.9780898719284. URL http://dx.doi.org/10.1137/1.9780898719284.
Reprint of the 1986 original.
[10] D. Needell.
Randomized Kaczmarz solver for noisy linear systems.
(2):395?403, 2010.
ISSN 0006-3835.
doi:
10.1007/s10543-010-0265-5.
http://dx.doi.org/10.1007/s10543-010-0265-5.
BIT,
50
URL
[11] D. Needell and R. Ward. Two-subspace projection method for coherent overdetermined linear systems.
Journal of Fourier Analysis and Applications, 19(2):256?269, 2013.
[12] Arkadi Nemirovski. Efficient methods in convex programming. 2005.
[13] Y. Nesterov. Introductory Lectures on Convex Optimization. Kluwer, 2004.
[14] Y. Nesterov. Efficiency of coordinate descent methods on huge-scale optimization problems. SIAM J.
Optimiz., 22(2):341?362, 2012.
[15] P. Richt?arik and M. Tak?ac? . Iteration complexity of randomized block-coordinate descent methods for
minimizing a composite function. Math. Program., pages 1?38, 2012.
[16] M. Schmidt, N. Roux, and F. Bach. Minimizing finite sums with the stochastic average gradient. arXiv
preprint arXiv:1309.2388, 2013.
[17] S. Shalev-Shwartz and T. Zhang. Stochastic dual coordinate ascent methods for regularized loss. J. Mach.
Learn. Res., 14(1):567?599, 2013.
[18] N. Srebro, K. Sridharan, and A. Tewari. Smoothness, low noise and fast rates. In Advances in Neural
Information Processing Systems, 2010.
[19] T. Strohmer and R. Vershynin. A randomized Kaczmarz algorithm with exponential convergence. J.
Fourier Anal. Appl., 15(2):262?278, 2009. ISSN 1069-5869. doi: 10.1007/s00041-008-9030-4. URL
http://dx.doi.org/10.1007/s00041-008-9030-4.
[20] K. Tanabe. Projection method for solving a singular system of linear equations and its applications.
Numerische Mathematik, 17(3):203?214, 1971.
[21] T. M. Whitney and R. K. Meany. Two algorithms related to the method of steepest descent. SIAM Journal
on Numerical Analysis, 4(1):109?118, 1967.
[22] P. Zhao and T. Zhang. Stochastic optimization with importance sampling. Submitted, 2014.
9
| 5355 |@word briefly:1 polynomial:2 norm:4 proportion:1 open:1 heuristically:1 sgd:47 moment:1 reduction:3 initial:1 kx0:4 dx:3 must:3 numerical:1 chicago:1 rward:1 enables:1 update:6 selected:1 xk:17 dissatisfying:1 steepest:1 accepting:2 iterates:8 math:2 org:3 zhang:2 mathematical:1 along:1 ik:6 prove:1 introductory:1 inside:2 x0:7 indeed:4 expected:2 roughly:1 examine:1 moulines:7 inspired:2 relying:1 spherical:6 actual:2 considering:2 increasing:1 solver:1 begin:2 provided:1 bounded:3 suffice:1 mass:2 what:2 interpreted:1 minimizes:1 substantially:1 supplemental:2 guarantee:12 runtime:1 exactly:3 sag:2 k2:8 scaled:2 returning:1 ser:1 before:1 el2:1 acad:1 mach:1 might:8 studied:1 dynamically:1 suggests:1 appl:1 co:1 limited:1 nemirovski:1 bi:9 unique:3 recursive:1 block:1 kaczmarz:22 differs:1 procedure:1 sdca:1 empirical:1 significantly:1 reject:1 composite:2 projection:3 pre:1 get:2 cannot:2 selection:2 context:2 applying:3 weighed:1 equivalent:1 optimize:1 demonstrated:1 center:2 go:1 straightforward:1 starting:1 convex:25 survey:1 numerische:2 simplicity:2 roux:2 needell:3 assigns:1 immediately:1 fik:2 classic:1 coordinate:5 target:2 suppose:1 controlling:1 programming:1 overdetermined:4 trick:1 logarithmically:1 pa:1 particularly:1 role:2 preprint:1 solved:1 calculate:3 ensures:2 richt:1 technological:1 pd:1 convexity:2 complexity:4 nesterov:2 dynamic:1 solving:4 algebra:1 upon:1 efficiency:1 easily:2 univ:1 fast:1 doi:6 shalev:1 widely:1 dominating:2 larger:1 kai:3 drawing:2 otherwise:2 tightness:1 ward:2 gi:16 radiology:1 noisy:1 seemingly:1 differentiable:1 niethammer:1 isbn:1 reconstruction:2 relevant:2 achieve:2 description:1 convergence:26 requirement:1 renormalizing:1 converges:1 help:2 ac:1 axial:1 progress:1 strong:3 cmc:1 radius:1 stochastic:13 material:2 implementing:1 hx:1 crux:1 f1:2 tighter:2 underdetermined:1 strictly:1 considered:4 substituting:1 smallest:1 purpose:1 favorable:1 utexas:1 weighted:24 hope:1 gaussian:6 arik:1 modified:2 rather:4 reaching:1 corollary:2 improvement:7 rank:2 contrast:1 rigorous:1 industrial:1 realizable:2 helpful:1 censor:1 eliminate:1 accept:1 borrowing:1 tak:1 selects:2 interested:2 arg:1 overall:1 colt:1 dual:1 special:2 construct:1 aware:1 sampling:51 minimized:2 others:1 gordon:1 individual:1 replaced:2 argminx:3 interest:2 huge:1 highly:1 investigate:2 certainly:1 mixture:4 yielding:1 light:1 strohmer:4 partial:2 necessary:1 unless:1 euclidean:1 desired:6 re:6 theoretical:2 whitney:1 bull:1 introducing:3 subset:1 uniform:7 technion:1 scanning:1 synthetic:2 vershynin:4 density:1 fundamental:1 randomized:15 siam:3 continuously:1 squared:4 again:2 unavoidable:3 von:1 choose:1 g2i:1 hzi:1 conf:1 ek:4 derivative:4 zhao:1 return:1 li:40 int:1 satisfy:1 explicitly:1 depends:5 analyze:1 sup:33 reached:1 decaying:2 recover:1 complicated:1 parallel:1 hanke:1 arkadi:1 contribution:2 minimize:1 square:12 accuracy:4 became:1 variance:7 yield:8 lli:2 computerized:3 multiplying:1 submitted:1 proof:2 static:1 sampled:2 gain:1 popular:1 recall:2 improves:2 hilbert:1 carefully:1 actually:1 back:1 sophisticated:1 manuscript:1 improved:2 done:1 though:1 strongly:19 furthermore:3 just:1 lastly:1 until:2 sketch:1 ei:5 reweighting:3 infimum:1 artifact:1 k22:31 normalized:2 unbiased:4 verify:1 true:1 reweighted:1 suboptimality:1 generalized:1 complete:1 l1:1 ranging:1 image:1 consideration:1 fi:43 recently:2 specialized:1 empirically:1 perturbing:1 conditioning:10 exponentially:1 volume:1 kluwer:1 significant:1 refer:3 bk22:3 ai:10 smoothness:8 rd:1 mathematics:4 similarly:2 access:1 yk2:1 lkx:1 base:1 dominant:2 recent:2 showed:2 optimizing:1 inf:4 scenario:4 continue:1 minimum:3 fortunately:1 somewhat:2 additional:2 signal:1 lik:2 full:2 desirable:1 smooth:17 faster:2 calculation:1 bach:9 long:1 offer:1 equally:2 parenthesis:1 ensuring:1 kax:3 variant:2 regression:1 noiseless:1 expectation:7 supi:1 arxiv:2 iteration:26 sometimes:1 want:1 singular:1 source:7 appropriately:1 biased:11 ascent:1 inconsistent:2 sridharan:1 call:1 yk22:1 easy:1 decent:1 reduce:3 idea:1 cn:1 tradeoff:1 texas:1 whether:2 url:3 speaking:1 repeatedly:1 useful:1 tewari:1 proportionally:1 tomography:4 http:3 disjoint:1 correctly:1 broadly:1 discrete:1 write:3 drawn:4 changing:1 asymptotically:1 subgradient:1 sum:1 eli:3 rachel:1 almost:2 throughout:1 scaling:4 bit:2 bound:13 quadratic:11 replaces:1 adapted:1 ahead:1 precisely:1 optimiz:1 nathan:1 simulate:1 fourier:2 min:1 optimality:2 subgradients:1 graceful:1 department:2 according:5 remain:1 aufl:1 making:1 invariant:1 equation:3 mathematik:2 foygel:1 discus:2 turn:2 studying:1 endowed:1 efi:1 apply:1 generic:1 appropriate:4 stepsize:4 save:1 alternative:1 schmidt:2 original:2 ensure:1 eggermont:1 classical:1 society:1 implied:1 objective:17 already:2 quantity:9 question:1 strategy:1 concentration:1 dependence:36 osung:1 hai:9 exhibit:2 gradient:18 kth:2 subspace:1 distance:2 separate:1 sci:1 extent:1 provable:1 issn:2 index:6 minimizing:3 difficult:1 relate:1 negative:2 anal:1 upper:1 observation:1 finite:5 descent:7 situation:2 precise:1 arbitrary:3 ttic:1 transverse:1 introduced:1 required:3 specified:2 connection:4 coherent:1 established:1 nip:2 deanna:1 regime:6 appeared:1 biasing:1 herman:1 program:1 power:1 event:1 regularized:1 indicator:1 residual:6 egi:1 scheme:1 improve:7 reprint:1 philadelphia:1 literature:3 understanding:2 nati:1 determining:1 asymptotic:2 relative:1 fully:5 lecture:1 highlight:1 loss:1 suggestion:1 interesting:1 proportional:6 srebro:3 var:1 versus:2 digital:1 sufficient:2 pi:1 austin:1 row:10 course:1 enjoys:1 weaker:1 understand:1 allow:2 institute:1 taking:1 distributed:1 overcome:1 calculated:1 lett:1 valid:2 unweighted:2 tanabe:1 avoided:2 polynomially:1 kzi:5 functionals:1 supremum:4 xi:3 shwartz:1 continuous:1 iterative:1 learn:1 transfer:1 ca:1 obtaining:1 improving:1 necessarily:1 coercivity:1 main:2 linearly:2 noise:2 n2:1 body:2 ekxk:5 sub:3 exponential:4 vanish:1 toyota:1 third:1 weighting:15 theorem:11 remained:1 british:1 showing:1 mckenna:1 concern:1 dominates:1 evidence:1 importance:20 magnitude:1 kx:4 horizon:3 rejection:2 highlighting:1 kxk:7 partially:6 g2:9 scalar:1 springer:1 corresponds:3 minimizer:4 ann:1 acceleration:1 lipschitz:18 replace:1 change:1 except:2 reducing:1 uniformly:2 averaging:1 degradation:1 lemma:2 accepted:2 attempted:1 college:1 support:1 dept:1 ex:1 |
4,811 | 5,356 | An Accelerated Proximal Coordinate Gradient Method
Qihang Lin
University of Iowa
Iowa City, IA, USA
[email protected]
Zhaosong Lu
Simon Fraser University
Burnaby, BC, Canada
[email protected]
Lin Xiao
Microsoft Research
Redmond, WA, USA
[email protected]
Abstract
We develop an accelerated randomized proximal coordinate gradient (APCG)
method, for solving a broad class of composite convex optimization problems.
In particular, our method achieves faster linear convergence rates for minimizing
strongly convex functions than existing randomized proximal coordinate gradient methods. We show how to apply the APCG method to solve the dual of the
regularized empirical risk minimization (ERM) problem, and devise efficient implementations that avoid full-dimensional vector operations. For ill-conditioned
ERM problems, our method obtains improved convergence rates than the state-ofthe-art stochastic dual coordinate ascent (SDCA) method.
1
Introduction
Coordinate descent methods have received extensive attention in recent years due to their potential
for solving large-scale optimization problems arising from machine learning and other applications.
In this paper, we develop an accelerated proximal coordinate gradient (APCG) method for solving
convex optimization problems with the following form:
def
F (x) = f (x) + ?(x) ,
(1)
minimize
x?RN
where f is differentiable on dom (?), and ? has a block separable structure. More specifically,
?(x) =
n
X
?i (xi ),
(2)
i=1
where each xi denotes a sub-vector of x with cardinality Ni , and each ?i : RNi ? R ? {+?}
is a closed convex function. We assume the collection {xi : i = 1, . . . , n} form a partition of
the components of x ? RN . In addition to the capability of modeling nonsmooth regularization
terms such as ?(x) = ?kxk1 , this model also includes optimization problems with block separable
constraints. More precisely, each block constraint xi ? Ci , where Ci is a closed convex set, can be
modeled by an indicator function defined as ?i (xi ) = 0 if xi ? Ci and ? otherwise.
At each iteration, coordinate descent methods choose one block of coordinates xi to sufficiently
reduce the objective value while keeping other blocks fixed. One common and simple approach
for choosing such a block is the cyclic scheme. The global and local convergence properties of the
cyclic coordinate descent method have been studied in, for example, [21, 11, 16, 2, 5]. Recently,
randomized strategies for choosing the block to update became more popular. In addition to its theoretical benefits [13, 14, 19], numerous experiments have demonstrated that randomized coordinate
descent methods are very powerful for solving large-scale machine learning problems [3, 6, 18, 19].
Inspired by the success of accelerated full gradient methods (e.g., [12, 1, 22]), several recent work
applied Nesterov?s acceleration schemes to speed up randomized coordinate descent methods. In
particular, Nesterov [13] developed an accelerated randomized coordinate gradient method for minimizing unconstrained smooth convex functions, which corresponds to the case of ?(x) ? 0 in (1).
1
Lu and Xiao [10] gave a sharper convergence analysis of Nesterov?s method, and Lee and Sidford [8] developed extensions with weighted random sampling schemes. More recently, Fercoq
and Richt?arik [4] proposed an APPROX (Accelerated, Parallel and PROXimal) coordinate descent
method for solving the more general problem (1) and obtained accelerated sublinear convergence
rates, but their method cannot exploit the strong convexity to obtain accelerated linear rates.
In this paper, we develop a general APCG method that achieves accelerated linear convergence
rates when the objective function is strongly convex. Without the strong convexity assumption, our
method recovers the APPROX method [4]. Moreover, we show how to apply the APCG method to
solve the dual of the regularized empirical risk minimization (ERM) problem, and devise efficient
implementations that avoid full-dimensional vector operations. For ill-conditioned ERM problems,
our method obtains faster convergence rates than the state-of-the-art stochastic dual coordinate ascent (SDCA) method [19], and the improved iteration complexity matches the accelerated SDCA
method [20]. We present numerical experiments to illustrate the advantage of our method.
1.1
Notations and assumptions
For any partition of x ? RN into {xi ? RNi : i = 1, . . . , n}, there is an N ? N permutation
matrix U partitioned as U = [U1 ? ? ? Un ], where Ui ? RN ?Ni , such that
n
X
Ui xi ,
and xi = UiT x, i = 1, . . . , n.
x=
i=1
For any x ? RN , the partial gradient of f with respect to xi is defined as
?i f (x) = UiT ?f (x), i = 1, . . . , n.
We associate each subspace RNi , for i = 1, . . . , n, with the standard Euclidean norm, denoted
by k ? k. We make the following assumptions which are standard in the literature on coordinate
descent methods (e.g., [13, 14]).
Assumption 1. The gradient of function f is block-wise Lipschitz continuous with constants Li , i.e.,
k?i f (x + Ui hi ) ? ?i f (x)k ? Li khi k, ? hi ? RNi , i = 1, . . . , n, x ? RN .
For convenience, we define the following norm in the whole space RN :
X
1/2
n
kxkL =
Li kxi k2
, ? x ? RN .
(3)
i=1
Assumption 2. There exists ? ? 0 such that for all y ? RN and x ? dom (?),
?
f (y) ? f (x) + h?f (x), y ? xi + ky ? xk2L .
2
The convexity parameter of f with respect to the norm k ? kL is the largest ? such that the above
inequality holds. Every convex function satisfies Assumption 2 with ? = 0. If ? > 0, the function f
is called strongly convex.
We note that an immediate consequence of Assumption 1 is
Li
f (x + Ui hi ) ? f (x) + h?i f (x), hi i + khi k2 , ? hi ? RNi ,
2
This together with Assumption 2 implies ? ? 1.
2
i = 1, . . . , n,
x ? RN . (4)
The APCG method
In this section we describe the general APCG method, and several variants that are more suitable
for implementation under different assumptions. These algorithms extend Nesterov?s accelerated
gradient methods [12, Section 2.2] to the composite and coordinate descent setting.
We first explain the notations used in our algorithms. The algorithms proceed in iterations, with k
being the iteration counter. Lower case letters x, y, z represent vectors in the full space RN , and
x(k) , y (k) and z (k) are their values at the kth iteration. Each block coordinate is indicated with a
(k)
subscript, for example, xi represents the value of the ith block of the vector x(k) . The Greek letters
?, ?, ? are scalars, and ?k , ?k and ?k represent their values at iteration k.
2
Algorithm 1: the APCG method
(0)
Input: x ? dom (?) and convexity parameter ? ? 0.
Initialize: set z (0) = x(0) and choose 0 < ?0 ? [?, 1].
Iterate: repeat for k = 0, 1, 2, . . .
1. Compute ?k ? (0, n1 ] from the equation
n2 ?k2 = (1 ? ?k ) ?k + ?k ?,
and set
?k ?
.
?k+1
?k ?k z (k) + ?k+1 x(k) .
?k+1 = (1 ? ?k )?k + ?k ?,
2. Compute y (k) as
y (k) =
1
?k ?k + ?k+1
(5)
?k =
(6)
(7)
3. Choose ik ? {1, . . . , n} uniformly at random and compute
o
n n?
2
k
z (k+1) = arg min
x?(1??k )z (k) ??k y (k)
L +h?ik f (y (k) ), xik i+?ik (xik ) .
2
x?RN
4. Set
x(k+1) = y (k) + n?k (z (k+1) ? z (k) ) +
? (k)
(z ? y (k) ).
n
(8)
The general APCG method is given as Algorithm 1. At each iteration k, it chooses a random
coordinate ik ? {1, . . . , n} and generates y (k) , x(k+1) and z (k+1) . One can observe that x(k+1) and
z (k+1) depend on the realization of the random variable
?k = {i0 , i1 , . . . , ik },
while y (k) is independent of ik and only depends on ?k?1 . To better understand this method, we
make the following observations. For convenience, we define
n n?
o
2
k
z?(k+1) = arg min
x ? (1 ? ?k )z (k) ? ?k y (k)
L + h?f (y (k) ), x ? y (k) i + ?(x) , (9)
2
x?RN
which is a full-dimensional update version of Step 3. One can observe that z (k+1) is updated as
( (k+1)
z?i
if i = ik ,
(k+1)
zi
=
(10)
(k)
(k)
(1 ? ?k )zi + ?k yi
if i 6= ik .
Notice that from (5), (6), (7) and (8) we have
x(k+1) = y (k) + n?k z (k+1) ? (1 ? ?k )z (k) ? ?k y (k) ,
which together with (10) yields
?
? y (k) + n? z (k+1) ? z (k) +
k
(k+1)
i
i
i
xi
=
? y (k)
i
?
n
(k)
zi
(k)
? yi
(k+1)
That is, in Step 4, we only need to update the block coordinates xik
if i = ik ,
if i 6= ik .
(11)
(k)
and set the rest to be yi .
We now state a theorem concerning the expected rate of convergence of the APCG method, whose
proof can be found in the full report [9].
Theorem 1. Suppose Assumptions 1 and 2 hold. Let F ? be the optimal value of problem (1), and
{x(k) } be the sequence generated by the APCG method. Then, for any k ? 0, there holds:
(
2 )
? k
?
?0
2n
(k)
?
1?
E?k?1 [F (x )] ? F ? min
F (x(0) ) ? F ? + R02 ,
,
?
n
2n + k ?0
2
where
def
?
R0 = min
kx(0) ? x? kL ,
?
?
x ?X
and X is the set of optimal solutions of problem (1).
3
(12)
Our result in Theorem 1 improves upon the convergence rates of the proximal coordinate gradient
methods in [14, 10], which have convergence rates on the order of
n
o
k
n
O min 1 ? n? , n+k
.
For n = 1, our result matches exactly that of the accelerated full gradient method in [12, Section 2.2].
2.1
Two special cases
Here we give two simplified versions of the APCG method, for the special cases of ? = 0 and
? > 0, respectively. Algorithm 2 shows the simplified version for ? = 0, which can be applied to
problems without strong convexity, or if the convexity parameter ? is unknown.
Algorithm 2: APCG with ? = 0
(0)
Input: x ? dom (?).
Initialize: set z (0) = x(0) and choose ?0 ? (0, n1 ].
Iterate: repeat for k = 0, 1, 2, . . .
1. Compute y (k) = (1 ? ?k )x(k) + ?k z (k) .
2. Choose ik ? {1, . . . , n}nuniformly at random and compute
o
n?k Lik
(k+1)
x ? z (k)
2 + h?i f (y (k) ), x ? y (k) i + ?i (x) .
zik
= arg minx?RN
k
k
ik
ik
2
(k+1)
and set zi
(k)
= zi
for all i 6= ik .
3. Set x(k+1) = y (k) + n?k (z (k+1) ? z (k) ).
p
?k4 + 4?k2 ? ?k2 .
4. Compute ?k+1 = 12
According to Theorem 1, Algorithm 2 has an accelerated sublinear convergence rate, that is
2
1
2n
F (x(0) ) ? F ? + R02 .
E?k?1 [F (x(k) )] ? F ? ?
2n + kn?0
2
With the choice of ?0 = 1/n, Algorithm 2 reduces to the APPROX method [4] with single block
update at each iteration (i.e., ? = 1 in their Algorithm 1).
For the strongly convex case with ? > ?
0, we can initialize Algorithm 1 with the parameter ?0 = ?,
which implies ?k = ? and ?k = ?k = ?/n for all k ? 0. This results in Algorithm 3.
Algorithm 3: APCG with ?0 = ? > 0
Input: x
(0)
? dom (?) and convexity parameter ? > 0.
Initialize: set z (0) = x(0) and and ? =
Iterate: repeat for k = 0, 1, 2, . . .
1. Compute y (k) =
?
?
n .
x(k) +?z (k)
.
1+?
2. Choose ik ? {1, . . . , n} uniformly at random and compute
o
n
2
(k)
(k)
??y (k)
L +h?ik f (y (k) ), xik ?yik i+?ik (xik ) .
z (k+1) = arg min n?
2 x?(1??)z
x?RN
3. Set x(k+1) = y (k) + n?(z (k+1) ? z (k) ) + n?2 (z (k) ? y (k) ).
As a direct corollary of Theorem 1, Algorithm 3 enjoys an accelerated linear convergence rate:
? k
?
?
F (x(0) ) ? F ? + R02 .
E?k?1 [F (x(k) )] ? F ? ? 1 ?
n
2
To the best of our knowledge, this is the first time such an accelerated rate is obtained for solving
the general problem (1) (with strong convexity) using coordinate descent type of methods.
4
2.2
Efficient implementation
The APCG methods we presented so far all need to perform full-dimensional vector operations at
each iteration. For example, y (k) is updated as a convex combination of x(k) and z (k) , and this
can be very costly since in general they can be dense vectors. Moreover, for the strongly convex case (Algorithms 1 and 3), all blocks of z (k+1) need to be updated at each iteration, although
only the ik -th block needs to compute the partial gradient and perform a proximal mapping. These
full-dimensional vector updates cost O(N ) operations per iteration and may cause the overall computational cost of APCG to be even higher than the full gradient methods (see discussions in [13]).
In order to avoid full-dimensional vector operations, Lee and Sidford [8] proposed a change of
variables scheme for accelerated coordinated descent methods for unconstrained smooth minimization. Fercoq and Richt?arik [4] devised a similar scheme for efficient implementation in the ? = 0
case for composite minimization. Here we show that such a scheme can also be developed for the
case of ? > 0 in the composite optimization setting. For simplicity, we only present an equivalent
implementation of the simplified APCG method described in Algorithm 3.
Algorithm 4: Efficient implementation of APCG with ?0 = ? > 0
Input: x
(0)
? dom (?) and convexity parameter ? > 0.
?
?
(0)
Initialize: set ? = n and ? = 1??
= 0 and v (0) = x(0) .
1+? , and initialize u
Iterate: repeat for k = 0, 1, 2, . . .
1. Choose ik ? {1, . . . , n} uniformly at random and compute
o
n
n?Lik
(k)
(k)
(k)
?ik = arg min
k?k2 + h?ik f (?k+1 u(k) +v (k) ), ?i + ?ik (??k+1 uik +vik +?) .
2
??R
Ni
k
2. Let u(k+1) = u(k) and v (k+1) = v (k) , and update
(k+1)
uik
(k)
= uik ?
1?n? (k)
? ,
2?k+1 ik
(k+1)
vik
(k)
= vik +
1+n? (k)
2 ?ik .
(13)
Output: x(k+1) = ?k+1 u(k+1) + v (k+1)
The following Proposition is proved in the full report [9].
Proposition 1. The iterates of Algorithm 3 and Algorithm 4 satisfy the following relationships:
x(k) = ?k u(k) + v (k) ,
y (k) = ?k+1 u(k) + v (k) ,
z (k) = ??k u(k) + v (k) .
(14)
We note that in Algorithm 4, only a single block coordinate of the vectors u(k) and v (k) are updated
at each iteration, which cost O(Ni ). However, computing the partial gradient ?ik f (?k+1 u(k) +v (k) )
may still cost O(N ) in general. In the next section, we show how to further exploit structure in many
ERM problems to completely avoid full-dimensional vector operations.
3
Application to regularized empirical risk minimization (ERM)
Let A1 , . . . , An be vectors in Rd , ?1 , . . . , ?n be a sequence of convex functions defined on R, and g
be a convex function on Rd . Regularized ERM aims to solve the following problem:
n
minimize P (w),
w?Rd
with
1X
P (w) =
?i (ATi w) + ?g(w),
n i=1
where ? > 0 is a regularization parameter. For example, given a label bi ? {?1} for each vector Ai ,
for i = 1, . . . , n, we obtain the linear SVM problem by setting ?i (z) = max{0, 1?bi z} and g(w) =
(1/2)kwk22 . Regularized logistic regression is obtained by setting ?i (z) = log(1+exp(?bi z)). This
formulation also includes regression problems. For example, ridge regression is obtained by setting
(1/2)?i (z) = (z ? bi )2 and g(w) = (1/2)kwk22 , and we get Lasso if g(w) = kwk1 .
5
Let ??i be the convex conjugate of ?i , that is, ??i (u) = maxz?R (zu ? ?i (z)). The dual of the
regularized ERM problem is (see, e.g., [19])
n
1X
1
?
?
maximize
D(x),
with
D(x)
=
Ax
,
??
(?x
)
?
?g
i
i
x?Rn
n i=1
?n
def
where A = [A1 , . . . , An ]. This is equivalent to minimize F (x) = ?D(x), that is,
n
X
1
def 1
?
?
F
(x)
=
minimize
Ax
.
?
(?x
)
+
?g
i
x?Rn
n i=1 i
?n
1
Ax and
The structure of F (x) above matches the formulation in (1) and (2) with f (x) = ?g ? ?n
?i (xi ) = n1 ??i (?xi ), and we can apply the APCG method to minimize F (x). In order to exploit
the fast linear convergence rate, we make the following assumption.
Assumption 3. Each function ?i is 1/? smooth, and the function g has unit convexity parameter 1.
Here we slightly abuse the notation by overloading ?, which also appeared in Algorithm 1. But
in this section it solely represents the (inverse) smoothness parameter of ?i . Assumption 3 implies
that each ??i has strong convexity parameter ? (with respect to the local Euclidean norm) and g ?
is differentiable and ?g ? has Lipschitz constant 1. In the following, we split the function F (x) =
f (x) + ?(x) by relocating the strong convexity term as follows:
n
1
?
1 X ?
?
?
f (x) = ?g
(15)
Ax +
kxk2 ,
?(x) =
? (?xi ) ? kxi k2 .
?n
2n
n i=1
2
As a result, the function f is strongly convex and each ?i is still convex. Now we can apply the
APCG method to minimize F (x) = ?D(x), and obtain the following guarantee.
Theorem 2. Suppose Assumption 3 holds and kAi k ? R for all i = 1, . . . , n. In order to obtain an
expected dual optimality gap E[D? ? D(x(k) )] ? ? by using the APCG method, it suffices to have
q
2
k ? n + nR
log(C/?).
(16)
??
where D? = maxx?Rn D(x) and the constant C = D? ? D(x(0) ) + (?/(2n))kx(0) ? x? k2 .
2
2
?
R +??n
ik
Proof. The function f (x) in (15) has coordinate Lipschitz constants Li = kA
?n2 + n ?
?n2
?
and convexity parameter n with respect to the unweighted Euclidean norm. The strong convexity
parameter of f (x) with respect to the norm k ? kL defined in(3) is
. 2
+??n
= R2??n
? = n? R ?n
2
+??n .
?
? k
?
?
According to Theorem 1, we have E[D? ?D(x(0) )] ? 1 ? n
C ? exp ? n k C. Therefore
it suffices to have the number of iterations k to be larger than
q
q
q
R2 +??n
nR2
2 + nR2 log(C/?) ?
?n log(C/?) = n
log(C/?).
log(C/?)
=
n
+
n
?
??n
??
??
This finishes the proof.
Several state-of-the-art algorithms for ERM, including SDCA [19], SAG [15, 17] and SVRG [7, 23]
obtain the iteration complexity
2
O n+ R
log(1/?)
.
(17)
??
We note that our result in (16) can be much better for ill-conditioned problems, i.e., when the condi2
tion number R
?? is larger than n. This is also confirmed by our numerical experiments in Section 4.
The complexity bound in (17) for the aforementioned work is for minimizing the primal objective
P (x) or the duality gap P (x) ? D(x), but our result in Theorem 2 is in terms of the dual optimality.
In the full report [9], we show that the same guarantee on accelerated primal-dual convergence can be
obtained by our method with an extra primal gradient step, without affecting the overall complexity.
The experiments in Section 4 illustrate superior performance of our algorithm on reducing the primal
objective value, even without performing the extra step.
6
We note that Shalev-Shwartz and Zhang
an accelerated SDCA method
[20]qrecently
developed
n
which achieves the same complexity O n + ?? log(1/?) as our method. Their method calls
the SDCA method in a full-dimensional accelerated gradient method in an inner-outer iteration procedure. In contrast, our APCG method is a straightforward single loop coordinate gradient method.
3.1
Implementation details
Here we show how to exploit the structure of the regularized ERM problem to efficiently compute
the coordinate gradient ?ik f (y (k) ), and totally avoid full-dimensional updates in Algorithm 4. We
focus on the special case g(w) = 21 kwk2 and show how to compute ?ik f (y (k) ). According to (15),
?ik f (y (k) ) =
1
T
(k)
)
?n2 Ai (Ay
(k)
+ n? yik .
Since we do not form y (k) in Algorithm 4, we update Ay (k) by storing and updating two vectors
in Rd : p(k) = Au(k) and q (k) = Av (k) . The resulting method is detailed in Algorithm 5.
Algorithm 5: APCG for solving dual ERM
Input: x
(0)
? dom (?) and convexity parameter ? > 0.
?
?
(0)
Initialize: set ? = n and ? = 1??
= 0, v (0) = x(0) , p(0) = 0 and q (0) = Ax(0) .
1+? , and let u
Iterate: repeat for k = 0, 1, 2, . . .
1. Choose ik ? {1, . . . , n} uniformly at random, compute the coordinate gradient
(k)
(k)
(k)
?ik = ?n1 2 ?k+1 ATik p(k) + ATik q (k) + n? ?k+1 uik + vik .
2. Compute coordinate increment
o
n
?(kAik k2 +??n)
(k)
(k)
(k)
(k)
k?k2 + h?ik , ?i + n1 ??ik (?k+1 uik ? vik ? ?) .
?ik = arg min
2?n
3. Let u
??R
(k+1)
Ni
k
= u(k) and v (k+1) = v (k) , and update
(k+1)
uik
(k)
= uik ?
p(k+1) = p(k) ?
1?n? (k)
? ,
2?k+1 ik
(k+1)
vik
(k)
1?n?
A ? ,
2?k+1 ik ik
Output: approximate primal and dual solutions
1
w(k+1) = ?n
?k+2 p(k+1) + q (k+1) ,
(k)
= vik +
q (k+1) = q (k) +
1+n? (k)
2 ?ik ,
(k)
1+n?
2 Aik ?ik .
(18)
x(k+1) = ?k+1 u(k+1) + v (k+1) .
Each iteration of Algorithm 5 only involves the two inner products ATik p(k) , ATik q (k) in computing
(k)
?ik and the two vector additions in (18). They all cost O(d) rather than O(n). When the Ai ?s are
sparse (the case of most large-scale problems), these operations can be carried out very efficiently.
Basically, each iteration of Algorithm 5 only costs twice as much as that of SDCA [6, 19].
4
Experiments
In our experiments, we solve ERM problems with smoothed hinge loss for binary classification.
That is, we pre-multiply each feature vector Ai by its label bi ? {?1} and use the loss function
?
if a ? 1,
? 0
?
1
?
a
?
if a ? 1 ? ?,
?(a) =
2
? 1 (1 ? a)
2
otherwise.
2?
The conjugate function of ? is ?? (b) = b + ?2 b2 if b ? [?1, 0] and ? otherwise. Therefore we have
?xi
?
1 ?
if xi ? [0, 1]
n
? (?xi ) ? kxi k2 =
?i (xi ) =
?
otherwise.
n
2
The dataset used in our experiments are summarized in Table 1.
7
?
?5
10
rcv1
100
100
10?3
10?3
10?3
10?6
10?6
10?6
AFG
SDCA
APCG
10?12
10?15
0
20
40
60
80
10?12
100
10?15
0
20
40
60
80
10?12
100
10?15
0
100
10?3
10?3
10?3
10?6
10?6
10?6
20
40
60
80
100
10?9
0
20
40
60
80
AFG
SDCA
APCG
10?9
100
100
10?9
0
100
100
100
10?1
10?2
10?3
10?4
10?5
10?6
0
10?1
10?2
10?3
10?4
10?5
10?6
0
10?1
10?2
10?3
10?4
10?5
10?6
0
20
40
60
80
100
100
AFG
SDCA
APCG
10?1
10?8
AFG
SDCA
APCG
10?9
100
10?9
0
10?7
news20
100
10?9
10?6
covertype
10?2
20
40
60
80
100
100
100
10?1
10?1
20
40
60
80
100
20
40
60
80
100
20
40
60
80
100
20
40
60
80
100
10?2
10?2
10?3
10?3
10?3
10?4
0
20
40
60
80
100
10?4
0
10?4
20
40
60
80
100
10?5
0
Figure 1: Comparing the APCG method with SDCA and the accelerated full gradient method (AFG)
with adaptive line search. In each plot, the vertical axis is the primal objective gap P (w(k) )?P ? , and
the horizontal axis is the number of passes through the entire dataset. The three columns correspond
to the three datasets, and each row corresponds to a particular value of the regularization parameter ?.
In our experiments, we compare the APCG method with SDCA and the accelerated full gradient
method (AFG) [12] with an additional line search procedure to improve efficiency. When the regularization parameter ? is not too small (around 10?4 ), then APCG performs similarly as SDCA as
predicted by our complexity results, and they both outperform AFG by a substantial margin.
Figure 1 shows the results in the ill-conditioned setting, with ? varying form 10?5 to 10?8 . Here
we see that APCG has superior performance in reducing the primal objective value compared with
SDCA and AFG, even though our theory only gives complexity for solving the dual ERM problem.
AFG eventually catches up for cases with very large condition number (see the plots for ? = 10?8 ).
datasets
rcv1
covtype
news20
number of samples n
20,242
581,012
19,996
number of features d
47,236
54
1,355,191
sparsity
0.16%
22%
0.04%
Table 1: Characteristics of three binary classification datasets (available from the LIBSVM web
page: http://www.csie.ntu.edu.tw/?cjlin/libsvmtools/datasets).
8
References
[1] A. Beck and M. Teboulle. A fast iterative shrinkage-threshold algorithm for linear inverse
problems. SIAM Journal on Imaging Sciences, 2(1):183?202, 2009.
[2] A. Beck and L. Tetruashvili. On the convergence of block coordinate descent type methods.
SIAM Journal on Optimization, 13(4):2037?2060, 2013.
[3] K.-W. Chang, C.-J. Hsieh, and C.-J. Lin. Coordinate descent method for large-scale l2 -loss
linear support vector machines. Journal of Machine Learning Research, 9:1369?1398, 2008.
[4] O. Fercoq and P. Richt?arik. Accelerated, parallel and proximal coordinate descent. Manuscript,
arXiv:1312.5799, 2013.
[5] M. Hong, X. Wang, M. Razaviyayn, and Z. Q. Luo. Iteration complexity analysis of block
coordinate descent methods. Manuscript, arXiv:1310.6957, 2013.
[6] C.-J. Hsieh, K.-W. Chang, C.-J. Lin, S.-S. Keerthi, and S. Sundararajan. A dual coordinate
descent method for large-scale linear svm. In Proceedings of the 25th International Conference
on Machine Learning (ICML), pages 408?415, 2008.
[7] R. Johnson and T. Zhang. Accelerating stochastic gradient descent using predictive variance
reduction. In Advances in Neural Information Processing Systems 26, pages 315?323. 2013.
[8] Y. T. Lee and A. Sidford. Efficient accelerated coordinate descent methods and faster algorithms for solving linear systems. arXiv:1305.1922.
[9] Q. Lin, Z. Lu, and L. Xiao. An accelerated proximal coordinate gradient method and its
application to regularized empirical risk minimization. Technical Report MSR-TR-2014-94,
Microsoft Research, 2014. (arXiv:1407.1296).
[10] Z. Lu and L. Xiao. On the complexity analysis of randomized block-coordinate descent methods. Accepted by Mathematical Programming, Series A, 2014. (arXiv:1305.4723).
[11] Z. Q. Luo and P. Tseng. On the convergence of the coordinate descent method for convex differentiable minimization. Journal of Optimization Theory & Applications, 72(1):7?35, 2002.
[12] Y. Nesterov. Introductory Lectures on Convex Optimization: A Basic Course. Kluwer, Boston,
2004.
[13] Y. Nesterov. Efficiency of coordinate descent methods on huge-scale optimization problems.
SIAM Journal on Optimization, 22(2):341?362, 2012.
[14] P. Richt?arik and M. Tak?ac? . Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function. Mathematical Programming, 144(1):1?38, 2014.
[15] N. Le Roux, M. Schmidt, and F. Bach. A stochastic gradient method with an exponential
convergence rate for finite training sets. In Advances in Neural Information Processing Systems
25, pages 2672?2680. 2012.
[16] A. Saha and A. Tewari. On the non-asymptotic convergence of cyclic coordinate descent
methods. SIAM Jorunal on Optimization, 23:576?601, 2013.
[17] M. Schmidt, N. Le Roux, and F. Bach. Minimizing finite sums with the stochastic average
gradient. Technical Report HAL 00860051, INRIA, Paris, France, 2013.
[18] S. Shalev-Shwartz and A. Tewari. Stochastic methods for ?1 regularized loss minimization. In
Proceedings of the 26th International Conference on Machine Learning (ICML), pages 929?
936, Montreal, Canada, 2009.
[19] S. Shalev-Shwartz and T. Zhang. Stochastic dual coordinate ascent methods for regularized
loss minimization. Journal of Machine Learning Research, 14:567?599, 2013.
[20] S. Shalev-Shwartz and T. Zhang. Accelerated proximal stochastic dual coordinate ascent for
regularized loss minimization. Proceedings of the 31st International Conference on Machine
Learning (ICML), JMLR W&CP, 32(1):64?72, 2014.
[21] P. Tseng. Convergence of a block coordinate descent method for nondifferentiable minimization. Journal of Optimization Theory and Applications, 140:513?535, 2001.
[22] P. Tseng. On accelerated proximal gradient methods for convex-concave optimization. Unpublished manuscript, 2008.
[23] L. Xiao and T. Zhang. A proximal stochastic gradient method with progressive variance reduction. Technical Report MSR-TR-2014-38, Microsoft Research, 2014. (arXiv:1403.4699).
9
| 5356 |@word msr:2 version:3 norm:6 hsieh:2 tr:2 reduction:2 cyclic:3 series:1 bc:1 ati:1 existing:1 ka:1 com:1 comparing:1 luo:2 numerical:2 partition:2 plot:2 update:9 zik:1 ith:1 iterates:1 zhang:5 mathematical:2 direct:1 ik:40 introductory:1 news20:2 expected:2 inspired:1 cardinality:1 totally:1 moreover:2 notation:3 developed:4 guarantee:2 every:1 concave:1 sag:1 exactly:1 k2:11 unit:1 local:2 consequence:1 subscript:1 solely:1 abuse:1 inria:1 twice:1 au:1 studied:1 bi:5 block:20 procedure:2 sdca:15 empirical:4 maxx:1 composite:5 pre:1 get:1 cannot:1 uiowa:1 convenience:2 risk:4 www:1 equivalent:2 maxz:1 demonstrated:1 straightforward:1 attention:1 convex:20 simplicity:1 roux:2 coordinate:42 increment:1 updated:4 suppose:2 aik:1 programming:2 associate:1 updating:1 kxk1:1 csie:1 wang:1 richt:4 counter:1 substantial:1 convexity:15 complexity:10 ui:4 nesterov:6 dom:7 depend:1 solving:9 predictive:1 upon:1 efficiency:2 completely:1 afg:9 fast:2 describe:1 choosing:2 shalev:4 whose:1 kai:1 solve:4 larger:2 otherwise:4 advantage:1 differentiable:3 sequence:2 product:1 loop:1 realization:1 ky:1 r02:3 convergence:19 illustrate:2 develop:3 ac:1 montreal:1 received:1 strong:7 predicted:1 involves:1 implies:3 greek:1 stochastic:9 libsvmtools:1 suffices:2 ntu:1 proposition:2 extension:1 hold:4 sufficiently:1 around:1 exp:2 mapping:1 achieves:3 label:2 largest:1 city:1 weighted:1 minimization:11 arik:4 aim:1 rather:1 avoid:5 shrinkage:1 varying:1 corollary:1 ax:5 focus:1 contrast:1 i0:1 burnaby:1 entire:1 tak:1 france:1 i1:1 arg:6 dual:14 ill:4 overall:2 denoted:1 aforementioned:1 classification:2 art:3 special:3 initialize:7 sampling:1 represents:2 broad:1 progressive:1 icml:3 nonsmooth:1 report:6 saha:1 beck:2 keerthi:1 microsoft:4 n1:5 huge:1 multiply:1 zhaosong:2 primal:7 partial:3 condi2:1 euclidean:3 theoretical:1 column:1 modeling:1 teboulle:1 sidford:3 cost:6 nr2:2 johnson:1 too:1 kn:1 proximal:12 kxi:3 chooses:1 st:1 international:3 randomized:8 siam:4 lee:3 together:2 choose:8 li:5 potential:1 b2:1 summarized:1 includes:2 coordinated:1 satisfy:1 depends:1 tion:1 closed:2 capability:1 parallel:2 simon:1 minimize:6 ni:5 became:1 variance:2 characteristic:1 efficiently:2 yield:1 ofthe:1 correspond:1 basically:1 lu:4 confirmed:1 explain:1 proof:3 recovers:1 proved:1 dataset:2 popular:1 knowledge:1 improves:1 manuscript:3 higher:1 improved:2 formulation:2 though:1 strongly:6 horizontal:1 web:1 logistic:1 indicated:1 hal:1 usa:2 regularization:4 hong:1 ay:2 ridge:1 performs:1 cp:1 wise:1 recently:2 common:1 superior:2 extend:1 kluwer:1 kwk2:1 sundararajan:1 ai:4 smoothness:1 approx:3 unconstrained:2 rd:4 similarly:1 recent:2 inequality:1 binary:2 success:1 kwk1:1 yi:3 devise:2 additional:1 r0:1 maximize:1 lik:2 full:18 reduces:1 smooth:3 technical:3 match:3 faster:3 bach:2 lin:7 devised:1 concerning:1 fraser:1 a1:2 variant:1 regression:3 basic:1 arxiv:6 iteration:19 represent:2 addition:3 affecting:1 extra:2 rest:1 ascent:4 pass:1 kwk22:2 call:1 split:1 iterate:5 finish:1 gave:1 zi:5 lasso:1 reduce:1 inner:2 vik:7 accelerating:1 proceed:1 cause:1 yik:2 tewari:2 detailed:1 http:1 outperform:1 qihang:2 notice:1 relocating:1 arising:1 per:1 threshold:1 k4:1 libsvm:1 imaging:1 year:1 sum:1 inverse:2 letter:2 powerful:1 sfu:1 def:4 hi:5 bound:1 covertype:1 constraint:2 precisely:1 generates:1 u1:1 speed:1 fercoq:3 min:8 optimality:2 performing:1 separable:2 rcv1:2 according:3 combination:1 conjugate:2 slightly:1 partitioned:1 tw:1 erm:13 equation:1 eventually:1 cjlin:1 available:1 operation:7 apply:4 observe:2 schmidt:2 tetruashvili:1 denotes:1 hinge:1 exploit:4 objective:6 strategy:1 costly:1 nr:1 gradient:27 kth:1 subspace:1 minx:1 outer:1 nondifferentiable:1 tseng:3 modeled:1 relationship:1 minimizing:5 sharper:1 xik:5 atik:4 implementation:8 unknown:1 perform:2 av:1 observation:1 vertical:1 datasets:4 finite:2 descent:23 immediate:1 rn:18 smoothed:1 canada:2 unpublished:1 paris:1 kl:3 extensive:1 redmond:1 appeared:1 sparsity:1 max:1 including:1 ia:1 suitable:1 regularized:11 indicator:1 scheme:6 improve:1 numerous:1 axis:2 carried:1 catch:1 literature:1 l2:1 asymptotic:1 loss:6 lecture:1 permutation:1 sublinear:2 iowa:2 rni:5 xiao:6 storing:1 row:1 course:1 repeat:5 keeping:1 svrg:1 enjoys:1 understand:1 sparse:1 benefit:1 apcg:31 uit:2 unweighted:1 collection:1 adaptive:1 simplified:3 far:1 approximate:1 obtains:2 global:1 xi:21 shwartz:4 un:1 continuous:1 search:2 iterative:1 table:2 ca:1 dense:1 whole:1 n2:4 razaviyayn:1 uik:7 sub:1 khi:2 exponential:1 kxk2:1 jmlr:1 theorem:8 zu:1 r2:2 svm:2 covtype:1 exists:1 overloading:1 ci:3 conditioned:4 kx:2 margin:1 gap:3 boston:1 scalar:1 chang:2 corresponds:2 satisfies:1 acceleration:1 lipschitz:3 change:1 specifically:1 uniformly:4 reducing:2 called:1 duality:1 accepted:1 support:1 accelerated:26 |
4,812 | 5,357 | Inference by Learning: Speeding-up Graphical
Model Optimization via a Coarse-to-Fine Cascade of
Pruning Classifiers
Bruno Conejo?
GPS Division, California Institute of Technology, Pasadena, CA, USA
Universite Paris-Est, Ecole des Ponts ParisTech, Marne-la-Vallee, France
[email protected]
Nikos Komodakis
Universite Paris-Est, Ecole des Ponts ParisTech, Marne-la-Vallee, France
[email protected]
Sebastien Leprince & Jean Philippe Avouac
GPS Division, California Institute of Technology, Pasadena, CA, USA
[email protected] [email protected]
Abstract
We propose a general and versatile framework that significantly speeds-up graphical model optimization while maintaining an excellent solution accuracy. The
proposed approach, refereed as Inference by Learning or in short as IbyL, relies
on a multi-scale pruning scheme that progressively reduces the solution space by
use of a coarse-to-fine cascade of learnt classifiers. We thoroughly experiment
with classic computer vision related MRF problems, where our novel framework
constantly yields a significant time speed-up (with respect to the most efficient
inference methods) and obtains a more accurate solution than directly optimizing
the MRF. We make our code available on-line [4].
1
Introduction
Graphical models in computer vision Optimization of undirected graphical models such as
Markov Random Fields, MRF, or Conditional Random Fields, CRF, is of fundamental importance in
computer vision. Currently, a wide spectrum of problems including stereo matching [25, 13], optical flow estimation [27, 16], image segmentation [23, 14], image completion and denoising [10], or,
object recognition [8, 2] rely on finding the mode of the distribution associated to the random field,
i.e., the Maximum A Posteriori (MAP) solution. The MAP estimation, often referred as the labeling
problem, is posed as an energy minimization task. While this task is NP-Hard, strong optimum solutions or even the optimal solutions can be obtained [3]. Over the past 20 years, tremendous progress
has been made in term of computational cost, and, many different techniques have been developed
such as move making approaches [3, 19, 22, 21, 28], and message passing methods [9, 32, 18, 20].
A review of their effectiveness has been published in [31, 12]. Nevertheless, the ever increasing
dimensionality of the problems and the need for larger solution space greatly challenge these tech?
This work was supported by USGS through the Measurements of surface ruptures produced by continental
earthquakes from optical imagery and LiDAR project (USGS Award G13AP00037), the Terrestrial Hazard
Observation and Reporting Center of Caltech, and the Moore foundation through the Advanced Earth Surface
Observation Project (AESOP Grant 2808).
1
niques as even the best ones have a highly super-linear computational cost and memory requirement
relatively to the dimensionality of the problem.
Our goal in this work is to develop a general MRF optimization framework that can provide a
significant speed-up for such methods while maintaining the accuracy of the estimated solutions.
Our strategy for accomplishing this goal will be to gradually reduce (by a significant amount) the
size of the discrete state space via exploiting the fact that an optimal labeling is typically far from
being random. Indeed, most MRF optimization problems favor solutions that are piecewise smooth.
In fact, this spatial structure of the MAP solution has already been exploited in prior work to reduce
the dimensionality of the solution space.
Related work A first set of methods of this type, referred here for short as the super-pixel approach
[30], defines a grouping heuristic to merge many random variables together in super-pixels. The
grouping heuristic can be energy-aware if it is based on the energy to minimize as in [15], or, energyagnostic otherwise as in [7, 30]. All random variables belonging to the same super-pixel are forced
to take the same label. This restricts the solution space and results in an optimization speed-up as
a smaller number of variables needs to be optimized. The super-pixel approach has been applied
with segmentation, stereo and object recognition [15]. However, if the grouping heuristic merges
variables that should have a different label in the MAP solution, only an approximate labeling is
computed. In practice, defining general yet efficient grouping heuristics is difficult. This represents
the key limitation of super-pixel approaches.
One way to overcome this limitation is to mimic the multi-scale scheme used in continuous optimization by building a coarse to fine representation of the graphical model. Similarly to the superpixel approach, such a multi-scale method, relies again on a grouping of variables for building the
required coarse to fine representation [17, 24, 26]. However, contrary to the super-pixel approach,
if the grouping merges variables that should have a different label in the MAP solution, there always exists a scale at which these variables are not grouped. This property thus ensures that the
MAP solution can still be recovered. Nevertheless, in order to manage a significant speed-up of
the optimization, the multi-scale approach also needs to progressively reduce the number of labels
per random variable (i.e., the solution space). Typically, this is achieved by use, for instance, of a
heuristic that keeps only a small fixed number of labels around the optimal label of each node found
at the current scale, while pruning all other labels, which are therefore not considered thereafter [5].
This strategy, however, may not be optimal or even valid for all types of problems. Furthermore,
such a pruning heuristic is totally inappropriate (and can thus lead to errors) for nodes located along
discontinuity boundaries of an optimal solution, where such boundaries are always expected to exist
in practice. An alternative strategy followed by some other methods relies on selecting a subset of
the MRF nodes at each scale (based on some criterion) and then fixing their labels according to the
optimal solution estimated at the current scale (essentially, such methods contract the entire label
set of a node to a single label). However, such a fixing strategy may be too aggressive and can also
easily lead to eliminating good labels.
Proposed approach Our method simultaneously makes use of the following two strategies for
speeding-up the MRF optimization process:
(i) it solves the problem through a multi-scale approach that gradually refines the MAP estimation based on a coarse-to-fine representation of the graphical model,
(ii) and, at the same time, it progressively reduces the label space of each variable by cleverly
utilizing the information computed during the above coarse-to-fine process.
To achieve that, we propose to significantly revisit the way that the pruning of the solution space
takes place. More specifically:
(i) we make use of and incorporate into the above process a fine-grained pruning scheme that
allows an arbitrary subset of labels to be discarded, where this subset can be different for
each node,
(ii) additionally, and most importantly, instead of trying to manually come up with some criteria
for deciding what labels to prune or keep, we introduce the idea of relying entirely on
a sequence of trained classifiers for taking such decisions, where different classifiers per
scale are used.
2
We name such an approach Inference by Learning, and show that it is particularly efficient and effective in reducing the label space while omitting very few correct labels. Furthermore, we demonstrate
that the training of these classifiers can be done based on features that are not application specific
but depend solely on the energy function, which thus makes our approach generic and applicable
to any MRF problem. The end result of this process is to obtain both an important speed-up and
a significant decrease in memory consumption as the solution space is progressively reduced. Furthermore, as each scale refines the MAP estimation, a further speed-up is obtained as a result of a
warm-start initialization that can be used when transitioning between different scales.
Before proceeding, it is worth also noting that there exists a body of prior work [29] that focuses on
fixing the labels of a subset of nodes of the graphical model by searching for a partial labeling with
the so-called persistency property (which means that this labeling is provably guaranteed to be part
of an optimal solution). However, finding such a set of persistent variables is typically very time
consuming. Furthermore, in many cases only a limited number of these variables can be detected.
As a result, the focus of these works is entirely different from ours, since the main motivation in our
case is how to obtain a significant speed-up for the optimization.
Hereafter, we assume without loss of generality that the graphical model is a discrete pairwise
CRF/MRF. However, one can straightforwardly apply our approach to higher order models.
Outline of the paper We briefly review the optimization problem related to a discrete pairwise
MRF and introduce the necessary notations in section 2. We describe our general multi-scale pruning
framework in section 3. We explain how classifiers are trained in section 4. Experimental results
and their analysis are presented in 5. Finally, we conclude the paper in section 6.
2
Notation and preliminaries
To represent a discrete MRF model M, we use the following notation
M = V, E, L, {?i }i?V , {?ij }(i,j)?E .
(1)
Here V and E represent respectively the nodes and edges of a graph, and L represents a discrete
label set. Furthermore, for every i ? V and (i, j) ? E, the functions ?i : L ? R and ?ij : L2 ? R
represent
and pairwise costs (that are also known connectively as MRF potentials
respectively unary
? = {?i }i?V , {?ij }(i,j)?E ). A solution x = (xi )i?V of this model consists of one variable per
vertex i, taking values in the label set L, and the total cost (energy) E(x|M) of such a solution is
E(x|M) =
X
X
?i (xi ) +
i?V
?ij (xi , xj ) .
(i,j)?E
The goal of MAP estimation is to find a solution that has minimum energy, i.e., computes
xMAP = arg min E(x|M) .
x?L|V|
The above minimization takes place over the full solution space of model M, which is L|V| . Here
we will also make use of a pruned solution space S(M, A), which is defined based on a binary
function A : V ? L ? {0, 1} (referred to as the pruning matrix hereafter) that specifies the status
(active or pruned) of a label for a given vertex, i.e.,
1
if label l is active at vertex i
A(i, l) =
(2)
0
if label l is pruned at vertex i
During optimization, active labels are retained while pruned labels are discarded. Based on a given
A, the corresponding pruned solution space of model M is defined as
n
o
S(M, A) = x ? L|V| | (?i), A(i, xi ) = 1 .
3
Multiscale Inference by Learning
In this section we describe the overall structure of our MAP estimation framework, beginning by
explaining how to construct the coarse-to-fine representation of the input graphical model.
3
3.1
Model coarsening
Given a model M (defined as in (1)), we wish to create a ?coarser? version of this model M0 =
V 0 , E 0 , L, {?0i }i?V 0 , {?0ij }(i,j)?E 0 . Intuitively, we want to partition the nodes of M into groups,
and treat each group as a single node of the coarser model M0 (the implicit assumption is that nodes
of M that are grouped together are assigned the same label). To that end, we will make use of a
grouping function g : V ? N . The nodes and edges of the coarser model are then defined as follows
V 0 = {i0 | ?i ? V, i0 = g(i)} ,
0
0
0
(3)
0
0
0
0
E = {(i , j ) | ?(i, j) ? E, i = g(i), j = g(j), i 6= j } .
(4)
Furthermore, the unary and pairwise potentials of M0 are given by
(?i0 ? V 0 ),
Figure 1: Black circles:
V, Black lines: E, Red
squares: V 0 , Blue lines:
E 0.
(?(i0 , j 0 ) ? E 0 ),
?0i0 (l)
?0i0 j 0 (l0 , l1 ) =
P
= P i?V|i0 =g(i)
+ (i,j)?E|i0 =g(i)=g(j)
X
(i,j)?E|i0 =g(i),j 0 =g(j)
?i (l)
, (5)
?ij (l, l)
?ij (l0 , l1 ) . (6)
With a slight abuse of notation, we will hereafter use g(M) to denote the coarser model resulting
from M when using the grouping function g, i.e., we define g(M) = M0 . Also, given a solution x0
of M0 , we can ?upsample? it into a solution x of M by setting xi = x0g(i) for each i ? V. We will
use the following notation in this case: g ?1 (x0 ) = x. We provide a toy example in supplementary
materials.
3.2
Coarse-to-fine optimization and label pruning
To estimate the MAP of an input model M, we first construct a series of N +1 progressively coarser
models (M(s) )0?s?N by use of a sequence of N grouping functions (g (s) )0?s<N , where
M(0) = M and (?s), M(s+1) = g (s) (M(s) ) .
This provides a multiscale (coarse-to-fine) representation of the original model., where the elements
of the resulting models are denoted as follows:
(s)
(s)
M(s) = V (s) , E (s) , L, {?i }i?V (s) , {?ij }(i,j)?E (s)
In our framework, MAP estimation proceeds from the coarsest to the finest scale (i.e., from model
M(N ) to M(0) ). During this process, a pruning matrix A(s) is computed at each scale s, which is
used for defining a restricted solution space S(M(s) , A(s) ). The elements of the matrix A(N ) at the
coarsest scale are all set equal to 1 (i.e., no label pruning is used in this case), whereas in all other
scales A(s) is computed by use of a trained classifier f (s) .
More specifically, at any given scale s, the following steps take place:
i. We approximately minimize (via any existing MRF optimization method) the energy of the
model M(s) over the restricted solution space S(M(s) , A(s) ), i.e., we compute
x(s) ? arg minx?S(M(s) ,A(s) ) E(x|M(s) ) .
ii. Given the estimated solution x(s) , a feature map z (s) : V (s) ? L ? RK is computed at
the current scale, and a trained classifier f (s) : RK ? {0, 1} uses this feature map z (s) to
construct the pruning matrix A(s?1) for the next scale as follows
(?i ? V (s?1) , ?l ? L), A(s?1) (i, l) = f (s) (z (s) (g (s?1) (i), l)) .
iii. Solution x(s) is ?upsampled? into x(s?1) = [g (s?1) ]?1 (x(s) ) and used as the initialization for the optimization at the next scale s ? 1. Note that, due to (5) and (6), it holds
E(x(s?1) |M(s?1) ) = E(x(s) |M(s) ). Therefore, this initialization ensures that energy will
continually decrease if the same is true for the optimization applied per scale. Furthermore,
it can allow for a warm-starting strategy when transitioning between scales.
The pseudocode of the resulting algorithm appears in Algo. 1.
4
Algorithm 1: Inference by learning framework
Data: Model M, grouping functions (g (s) )0?s<N , classifiers (f (s) )0<s?N
Result: x(0)
Compute the coarse to fine sequence of MRFs:
M(0) ? M
for s = [0 . . . N ? 1] do
M(s+1) ? g (s) (M(s) )
Optimize the coarse to fine sequence of MRFs over pruned solution spaces:
(?i ? V (N ) , ?l ? L), A(N ) (i, l) ? 1
Initialize x(N )
for s = [N...0] do
Update x(s) by iterative minimization: x(s) ? arg minx?S(M(s) ,A(s) ) E(x|M(s) )
if s 6= 0 then
Compute feature map z (s)
Update pruning matrix for next finer scale: A(s?1) (i, l) = f (s) (z (s) (g (s?1) (i), l))
Upsample x(s) for initializing solution x(s?1) at next scale: x(s?1) ? [g (s?1) ]?1 (x(s) )
4
Features and classifier for label pruning
For each scale s, we explain how the set of features comprising the feature map z (s) is computed
and how we train (off-line) the classifier f (s) . This is a crucial step for our approach. Indeed, if the
classifier wrongly prunes labels that belong to the MAP solution, then, only an approximate labeling
might be found at the finest scale. Moreover, keeping too many active labels will result in a poor
speed-up for MAP estimation.
4.1
Features
The feature map z (s) : V (s) ? L ? RK is formed by stacking K individual real-valued features
defined on V (s) ? L. We propose to compute features that are not application specific but depend
solely on the energy function and the current solution x(s) . This makes our approach generic and
applicable to any MRF problem. However, as we establish a general framework, specific application
features can be straightforwardly added in future work.
Presence of strong discontinuity This binary feature, PSD(s) , accounts for the existence of dis(s)
(s)
continuity in solution x(s) when a strong link (i.e., ?ij (xi , xj ) > ?) exists between neighbors.
Its definition follows for any vertex i ? V (s) and any label l ? L :
(s)
(s)
1
?(i, j) ? E (s) | ?ij (xi , xj ) > ?
PSD(s) (i, l) =
(7)
0
otherwise
Local energy variation This feature represents the local variation of the energy around the current
solution x(s) . It accounts for both the unary and pairwise terms associated to a vertex and a label.
As in [11], we remove the local energy of the current solution as it leads to a higher discriminative
power. The local energy variation feature, LEV(s) , is defined for any i ? V (s) and l ? L as follows:
(s)
LEV
(s)
(i, l) =
(s)
(s)
NV (i)
(s)
(s)
?i (l) ? ?i (xi )
(s)
(s)
(s)
(s)
X
?ij (l, xj ) ? ?ij (xi , xj )
j:(i,j)?E (s)
NE (i)
+
(s)
(s)
(8)
(s)
with NV (i) = card{i0 ? V (s?1) : g (s?1) (i0 ) = i} and NE (i) = card{(i0 , j 0 ) ? E (s?1) :
g (s?1) (i0 ) = i, g (s?1) (j 0 ) = j}.
Unary ?coarsening? This feature, UC(s) , aims to estimate an approximation of the coarsening
induced in the MRF unary terms when going from model M(s?1) to model M(s) , i.e., as a result of
5
applying the grouping function g (s?1) . It is defined for any i ? V (s) and l ? L as follows
(s?1)
UC(s) (i, l) =
|?i0
X
i0 ?V (s?1) |g (s?1) (i0 )=i
(s)
(l) ?
?i (l)
(s)
NV (i)
(s)
|
(9)
NV (i)
Feature normalization The features are by design insensitive to any additive term applied on all
the unary and pairwise terms. However, we still need to apply a normalization to the LEV(s) and
UC(s) features to make them insensitive to any positive global scaling factor applied on both the
unary and pairwise terms (such scaling variations are commonly used in computer vision). Hence,
we simply divide group of features, LEV(s) and UC(s) by their respective mean value.
4.2
Classifier
To train the classifiers, we are given as input a set of MRF instances (all of the same class, e.g.,
stereo-matching) along with the ground truth MAP solutions. We extract a subset of MRFs for offline learning and a subset for on-line testing. For each MRF instance in the training set, we apply
the algorithm 1 without any pruning (i.e., A(s) ? 1) and, at each scale, we keep track of the features
(s)
z (s) and also compute the binary function XMAP : V (s) ? L ? {0, 1} defined as follows:
1, if l is the ground truth label for node i
(0)
(?i ? V, ?l ? L), XMAP (i, l) =
0, otherwise
_
(s)
(s?1)
(?s > 0)(?i ? V (s) , ?l ? L), XMAP (i, l) =
XMAP (i0 , l) ,
i0 ?V (s?1) :g (s) (i0 )=i
(s)
W
where denotes the binary OR operator. The values 0 and 1 in XMAP define respectively the two
classes c0 and c1 when training the classifier f (s) , where c0 means that the label can be pruned and
c1 that the label should not be pruned.
To treat separately the nodes that are on the border of a strong discontinuity, we split the feature map
(s)
(s)
(s)
(s)
z (s) into two groups z0 and z1 , where z0 contains only features where PSD(s) = 0 and z1
contains only features where PSD(s) = 1 (strong discontinuity). For each group, we train a standard
linear C-SVM classifier with l2 -norm regularization (regularization parameter was set to C = 10).
The linear classifiers give good enough accuracy during training while also being fast to evaluate at
test time
During training (and for each group), we also introduce weights to balance the different number of
elements in each class (c0 is much larger than c1 ), and to also strongly penalize misclassification in
c1 (as such misclassification can have a more drastic impact on the accuracy of MAP estimation). To
card(c0 )
accomplish that, we set the weight for class c0 to 1, and the weight for class c1 to ? card(c
, where
1)
card(?) counts the number of training samples in each class. Parameter ? is a positive scalar (common to both groups) used for tuning the penalization of misclassification in c1 (it will be referred
to as the pruning aggressiveness factor hereafter as it affects the amount of labels that get pruned).
During on-line testing, depending on the value of the PSD feature, f (s) applies the linear classifier
(s)
(s)
learned on group z0 if PSD(s) = 0, or the linear classifier learned on group z1 if PSD(s) = 1.
5
Experimental results
We evaluate our framework on pairwise MRFs from stereo-matching, image restoration, and, optical
flow estimation problems. The corresponding MRF graphs consist of regular 4-connected grids in
this case. At each scale, the grouping function merges together vertices of 2 ? 2 subgrids. We leave
more advanced grouping functions [15] for future work. As MRF optimization subroutine, we use
the Fast-PD algorithm [21]. We make our code available on-line [4].
Experimental setup For the stereo matching problem, we estimate the disparity map from images
IR and IL where each label encodes a potential disparity d (discretized at quarter of a pixel precision), with MRF potentials ?p (d) = ||IL (yp , xp )?IR (yp , xp ?d)||1 and ?pq (d0 , d1 ) = wpq |d0 ?d1 |,
with the weight wpq varying based on the image gradient (parameters are adjusted for each sequence). We train the classifier on the well-known Tsukuba stereo-pair (61 labels), and use all other
6
(a) Speed-up
(b) Active label ratio
(c) Energy ratio
(d) Label agreement
Figure 2: Performance of our Inference by Learning framework: (Top row) stereo matching, (Middle row)
optical flow, (Bottom row) image restoration. For stereo matching, the Average Middlebury curve represents
the average value of the statistic for the entire Middlebury dataset [6] (2001, 2003, 2005 and 2006) (37 stereopairs).
stereo-pairs of [6] (2001, 2003, 2005 and 2006) for testing. For image restoration, we estimate the
pixel intensity of a noisy and incomplete image I with MRF potentials ?p (l) = ||I(yp , xp ) ? l||22
and ?(l0 , l1 ) = 25 min(||l0 ? l1 ||22 , 200). We train the classifier on the Penguin image stereo-pair
(256 labels), and use House (256 labels) for testing (dataset [31]). For the optical flow estimation,
we estimate a subpixel-accurate 2D displacement field between two frames by extending the stereo
matching formulation to 2D. Using the dataset of [1], we train the classifier on Army (1116 labels),
and test on RubberWhale (625 labels) and Dimetrodon (483 labels). For all experiments, we use 5
scales and set in (7) ? = 5w
?pq with w
?pq being the mean value of edge weights.
Evaluations We evaluate three optimization strategies: the direct optimization (i.e., optimizing
the full MRF at the finest scale), the multi-scale optimization (? = 0, i.e., our framework without
any pruning), and our Inference by Learning optimization, where we experiment with different error
ratios ? that range between 0.001 and 1.
We assess the performance by computing the energy ratio, i.e., the ratio between the current energy
and the energy computed by the direct optimization, the best label agreement, i.e., the proportion
of labels that coincides with the labels of the lowest computed energy, the speed-up factor, i.e., the
ratio of computation time between the direct optimization and the current optimization strategy, and,
the active label ratio, i.e., the percentage of active labels at the finest scale.
Results and discussion For all problems, we present in Fig. 2 the performance of our Inference
by Learning approach for all tested aggressiveness factors and show in Fig. 3 estimated results for
? = 0.01. We present additional results in the supplementary material.
For every problem and aggressiveness factors until ? = 0.1, our pruning-based optimization obtains
a lower energy (column (c) of Fig. 2) in less computation time, achieving a speed-up factor (column
(a) of Fig. 2) close to 5 for Stereo-matching, above 10 for Optical-flow and up to 3 for image
restoration. (note that these speed-up factors are with respect to an algorithm, FastPD, that was the
most efficient one in recent comparisons [12]). The percentage of active labels (Fig. 2 column (b))
strongly correlates with the speed-up factor. The best labeling agreement (Fig. 2 column (d)) is
never worse than 97% (except for the image restoration problems because of the in-painted area)
7
Tsukuba
Venus
Teddy
Army
Dimet.
House
(a)
(b)
(c)
(d)
(e)
(f)
Figure 3: Results of our Inference by Learning framework for ? = 0.1. Each row is a different MRF problem.
(a) original image, (b) ground truth, (c) solution of the pruning framework, (d,e,f) percentage of active labels
per vertex for scale 0, 1 and 2 (black 0%, white 100%).
and is always above 99% for ? 6 0.1. As expected, less pruning happens near label discontinuities
as illustrated in column (d,e,f) of Fig. 3 justifying the use of a dedicated linear classifier. Moreover,
large homogeneously labeled regions are pruned earlier in the coarse to fine scale.
6
Conclusion and future work
Our Inference by Learning approach consistently speeds-up the graphical model optimization by
a significant amount while maintaining an excellent accuracy of the labeling estimation. On most
experiments, it even obtains a lower energy than direct optimization.
In future work, we plan to experiment with problems that require general pairwise potentials where
message-passing techniques can be more effective than graph-cut based methods but are at the same
time much slower. Our framework is guaranteed to provide an even more dramatic speedup in this
case since the computational complexity of message-passing methods is quadratic with respect to
the number of labels while being linear for graph-cut based methods used in our experiments. We
also intend to explore the use of application specific features, learn the grouping functions used in
the coarse-to-fine scheme, jointly train the cascade of classifiers, and apply our framework to high
order graphical models.
References
[1] S. Baker, S. Roth, D. Scharstein, M.J. Black, J. P. Lewis, and R. Szeliski. A database and evaluation
methodology for optical flow. In ICCV 2007., 2007.
[2] Martin Bergtholdt, J?org Kappes, Stefan Schmidt, and Christoph Schn?orr. A study of parts-based object
class detection using complete graphs. IJCV, 2010.
[3] Y. Boykov, O. Veksler, and R. Zabih. Fast approximate energy minimization via graph cuts. PAMI, 2001.
[4] B. Conejo. http://imagine.enpc.fr/?conejob/ibyl/.
8
[5] B. Conejo, S. Leprince, F. Ayoub, and J. P. Avouac. Fast global stereo matching via energy pyramid
minimization. ISPRS Ann. Photogramm. Remote Sens. Spatial Inf. Sci., 2014.
[6] Middlebury Stereo Datasets.
[7] Pedro F. Felzenszwalb and Daniel P. Huttenlocher. Efficient graph-based image segmentation. IJCV,
2004.
[8] Pedro F. Felzenszwalb and Daniel P. Huttenlocher. Pictorial structures for object recognition. IJCV, 2005.
[9] P.F. Felzenszwalb and D.P. Huttenlocher. Efficient belief propagation for early vision. In CVPR, 2004.
[10] W.T. Freeman and E.C. Pasztor. Learning low-level vision. In ICCV, 1999.
[11] Xiaoyan Hu and P. Mordohai. A quantitative evaluation of confidence measures for stereo vision. PAMI.,
2012.
[12] J.H. Kappes, B. Andres, F.A. Hamprecht, C. Schnorr, S. Nowozin, D. Batra, Sungwoong Kim, B.X.
Kausler, J. Lellmann, N. Komodakis, and C. Rother. A comparative study of modern inference techniques
for discrete energy minimization problems. In CVPR, 2013.
[13] Junhwan Kim, V. Kolmogorov, and R. Zabih. Visual correspondence using energy minimization and
mutual information. In ICCV, 2003.
[14] S. Kim, C. Yoo, S. Nowozin, and P. Kohli. Image segmentation using higher-order correlation clustering,
2014.
[15] Taesup Kim, S. Nowozin, P. Kohli, and C.D. Yoo. Variable grouping for energy minimization. In CVPR,
2011.
[16] T. Kohlberger, C. Schnorr, A. Bruhn, and J. Weickert. Domain decomposition for variational optical-flow
computation. IEEE Transactions on Information Theory/Image Processing, 2005.
[17] Pushmeet Kohli, Victor S. Lempitsky, and Carsten Rother. Uncertainty driven multi-scale optimization.
In DAGM-Symposium, 2010.
[18] V. Kolmogorov. Convergent tree-reweighted message passing for energy minimization. PAMI, 2006.
[19] V. Kolmogorov and R. Zabin. What energy functions can be minimized via graph cuts? PAMI, 2004.
[20] N. Komodakis, N. Paragios, and G. Tziritas. Mrf optimization via dual decomposition: Message-passing
revisited. In CVPR, 2007.
[21] N. Komodakis, G. Tziritas, and N. Paragios. Fast, approximately optimal solutions for single and dynamic
mrfs. In CVPR, 2007.
[22] M. Pawan Kumar and Daphne Koller. Map estimation of semi-metric mrfs via hierarchical graph cuts. In
UAI, 2009.
[23] M.P. Kumar, P.H.S. Ton, and A. Zisserman. Obj cut. In CVPR, 2005.
[24] H. Lombaert, Yiyong Sun, L. Grady, and Chenyang Xu. A multilevel banded graph cuts method for fast
image segmentation. In ICCV 2005., 2005.
[25] T. Meltzer, C. Yanover, and Y. Weiss. Globally optimal solutions for energy minimization in stereo vision
using reweighted belief propagation. In ICCV, 2005.
[26] P. Perez and F. Heitz. Restriction of a markov random field on a graph and multiresolution statistical
image modeling. IEEE Transactions on Information Theory/Image Processing, 1996.
[27] S. Roth and M.J. Black. Fields of experts: a framework for learning image priors. In CVPR, 2005.
[28] C. Rother, V. Kolmogorov, V. Lempitsky, and M. Szummer. Optimizing binary mrfs via extended roof
duality. In CVPR, 2007.
[29] Alexander Shekhovtsov. Maximum persistency in energy minimization. In CVPR, 2014.
[30] Jianbo Shi and Jitendra Malik. Normalized cuts and image segmentation. PAMI., 2000.
[31] R. Szeliski, R. Zabih, D. Scharstein, O. Veksler, V. Kolmogorov, Aseem Agarwala, M. Tappen, and
C. Rother. A comparative study of energy minimization methods for markov random fields with
smoothness-based priors. PAMI, 2008.
[32] M.J. Wainwright, T.S. Jaakkola, and A.S. Willsky. Map estimation via agreement on trees: messagepassing and linear programming. IEEE Transactions on Information Theory/Image Processing, 2005.
9
| 5357 |@word kohli:3 version:1 briefly:1 eliminating:1 norm:1 middle:1 proportion:1 c0:5 hu:1 decomposition:2 dramatic:1 versatile:1 series:1 contains:2 selecting:1 hereafter:4 disparity:2 ecole:2 ours:1 daniel:2 past:1 existing:1 recovered:1 enpc:2 current:8 yet:1 finest:4 refines:2 additive:1 partition:1 remove:1 progressively:5 update:2 beginning:1 short:2 persistency:2 coarse:13 provides:1 node:13 revisited:1 org:1 daphne:1 along:2 direct:4 symposium:1 persistent:1 consists:1 ijcv:3 introduce:3 x0:2 pairwise:9 expected:2 indeed:2 multi:8 discretized:1 freeman:1 relying:1 globally:1 inappropriate:1 increasing:1 totally:1 project:2 notation:5 moreover:2 baker:1 lowest:1 what:2 developed:1 finding:2 quantitative:1 every:2 classifier:24 jianbo:1 grant:1 continually:1 before:1 positive:2 local:4 treat:2 middlebury:3 painted:1 lev:4 ponts:2 solely:2 merge:1 abuse:1 black:5 approximately:2 might:1 initialization:3 pami:6 christoph:1 limited:1 range:1 earthquake:1 testing:4 practice:2 displacement:1 area:1 cascade:3 significantly:2 matching:9 confidence:1 regular:1 upsampled:1 get:1 close:1 operator:1 wrongly:1 applying:1 optimize:1 restriction:1 map:25 center:1 roth:2 shi:1 starting:1 utilizing:1 importantly:1 d1:2 classic:1 searching:1 variation:4 imagine:1 programming:1 gps:3 us:1 superpixel:1 agreement:4 element:3 recognition:3 particularly:1 located:1 tappen:1 cut:8 coarser:5 labeled:1 database:1 bottom:1 huttenlocher:3 initializing:1 region:1 ensures:2 connected:1 kappes:2 sun:1 remote:1 decrease:2 pd:1 complexity:1 dynamic:1 trained:4 depend:2 algo:1 division:2 easily:1 kolmogorov:5 train:7 forced:1 fast:6 effective:2 describe:2 detected:1 labeling:8 jean:1 heuristic:6 posed:1 larger:2 supplementary:2 valued:1 cvpr:9 otherwise:3 favor:1 statistic:1 aseem:1 jointly:1 noisy:1 sequence:5 sen:1 propose:3 fr:2 achieve:1 multiresolution:1 exploiting:1 optimum:1 requirement:1 extending:1 comparative:2 leave:1 object:4 depending:1 develop:1 completion:1 fixing:3 ij:12 progress:1 solves:1 strong:5 tziritas:2 come:1 correct:1 aggressiveness:3 material:2 require:1 multilevel:1 preliminary:1 adjusted:1 hold:1 around:2 considered:1 ground:3 deciding:1 m0:5 early:1 earth:1 estimation:14 applicable:2 label:54 currently:1 grouped:2 create:1 minimization:12 stefan:1 always:3 super:7 aim:1 varying:1 jaakkola:1 l0:4 focus:2 subpixel:1 consistently:1 greatly:1 tech:1 kim:4 posteriori:1 inference:12 mrfs:7 unary:7 i0:19 typically:3 entire:2 dagm:1 pasadena:2 koller:1 going:1 france:2 comprising:1 subroutine:1 provably:1 pixel:8 agarwala:1 arg:3 overall:1 dual:1 denoted:1 plan:1 spatial:2 initialize:1 uc:4 mutual:1 field:7 aware:1 construct:3 never:1 equal:1 manually:1 represents:4 bruhn:1 mimic:1 future:4 np:1 minimized:1 piecewise:1 few:1 penguin:1 modern:1 simultaneously:1 individual:1 roof:1 pictorial:1 bergtholdt:1 pawan:1 psd:7 detection:1 message:5 highly:1 evaluation:3 hamprecht:1 perez:1 accurate:2 edge:3 partial:1 necessary:1 respective:1 tree:2 incomplete:1 divide:1 circle:1 instance:3 column:5 earlier:1 modeling:1 restoration:5 cost:4 stacking:1 vertex:8 subset:6 veksler:2 too:2 straightforwardly:2 learnt:1 accomplish:1 thoroughly:1 fundamental:1 contract:1 off:1 together:3 imagery:1 again:1 manage:1 worse:1 expert:1 yp:3 toy:1 aggressive:1 potential:6 account:2 de:2 orr:1 jitendra:1 red:1 start:1 grady:1 ass:1 minimize:2 square:1 accuracy:5 accomplishing:1 formed:1 ir:2 il:2 yield:1 shekhovtsov:1 andres:1 produced:1 worth:1 finer:1 published:1 explain:2 banded:1 definition:1 energy:30 universite:2 associated:2 dataset:3 dimensionality:3 segmentation:6 appears:1 higher:3 methodology:1 zisserman:1 wei:1 formulation:1 done:1 strongly:2 generality:1 furthermore:7 implicit:1 until:1 correlation:1 multiscale:2 propagation:2 continuity:1 defines:1 mode:1 building:2 usa:2 name:1 omitting:1 true:1 normalized:1 hence:1 assigned:1 regularization:2 moore:1 illustrated:1 white:1 reweighted:2 komodakis:5 during:6 coincides:1 criterion:2 trying:1 outline:1 crf:2 demonstrate:1 complete:1 l1:4 dedicated:1 image:21 variational:1 novel:1 boykov:1 common:1 pseudocode:1 quarter:1 insensitive:2 belong:1 slight:1 significant:7 measurement:1 smoothness:1 tuning:1 grid:1 similarly:1 bruno:1 refereed:1 pq:3 surface:2 recent:1 xiaoyan:1 optimizing:3 inf:1 driven:1 taesup:1 binary:5 exploited:1 caltech:4 victor:1 minimum:1 additional:1 nikos:2 prune:2 ii:3 semi:1 full:2 reduces:2 d0:2 smooth:1 hazard:1 sungwoong:1 justifying:1 award:1 impact:1 mrf:24 vision:8 essentially:1 metric:1 represent:3 normalization:2 pyramid:1 achieved:1 c1:6 penalize:1 whereas:1 want:1 fine:14 separately:1 crucial:1 nv:4 induced:1 undirected:1 contrary:1 flow:7 effectiveness:1 coarsening:3 obj:1 near:1 noting:1 presence:1 iii:1 split:1 enough:1 meltzer:1 xj:5 affect:1 reduce:3 idea:1 venus:1 stereo:16 passing:5 amount:3 zabih:3 reduced:1 http:1 specifies:1 exist:1 restricts:1 percentage:3 revisit:1 estimated:4 per:5 track:1 blue:1 discrete:6 group:9 key:1 thereafter:1 nevertheless:2 achieving:1 graph:11 usgs:2 year:1 uncertainty:1 reporting:1 place:3 decision:1 scaling:2 entirely:2 followed:1 guaranteed:2 weickert:1 tsukuba:2 correspondence:1 quadratic:1 convergent:1 encodes:1 speed:15 min:2 pruned:10 wpq:2 coarsest:2 optical:8 kumar:2 relatively:1 martin:1 speedup:1 according:1 poor:1 belonging:1 cleverly:1 smaller:1 mordohai:1 making:1 happens:1 intuitively:1 gradually:2 restricted:2 iccv:5 count:1 drastic:1 end:2 available:2 apply:4 hierarchical:1 generic:2 homogeneously:1 alternative:1 schmidt:1 slower:1 existence:1 original:2 denotes:1 top:1 clustering:1 graphical:11 maintaining:3 establish:1 move:1 intend:1 already:1 marne:2 added:1 malik:1 strategy:8 minx:2 gradient:1 continental:1 link:1 card:5 sci:1 consumption:1 terrestrial:1 willsky:1 rother:4 code:2 retained:1 ratio:7 balance:1 dimet:1 difficult:1 setup:1 zabin:1 design:1 sebastien:1 observation:2 markov:3 discarded:2 datasets:1 pasztor:1 teddy:1 philippe:1 defining:2 extended:1 ever:1 frame:1 arbitrary:1 intensity:1 pair:3 paris:2 required:1 optimized:1 z1:3 schn:1 california:2 merges:3 learned:2 tremendous:1 discontinuity:5 proceeds:1 challenge:1 including:1 memory:2 belief:2 wainwright:1 power:1 misclassification:3 kausler:1 rely:1 warm:2 advanced:2 yanover:1 scheme:4 technology:2 ne:2 rupture:1 extract:1 speeding:2 review:2 prior:4 l2:2 loss:1 limitation:2 penalization:1 foundation:1 xp:3 nowozin:3 row:4 supported:1 keeping:1 dis:1 offline:1 allow:1 institute:2 wide:1 explaining:1 taking:2 neighbor:1 szeliski:2 felzenszwalb:3 overcome:1 boundary:2 curve:1 valid:1 heitz:1 computes:1 made:1 commonly:1 far:1 pushmeet:1 correlate:1 transaction:3 scharstein:2 pruning:20 obtains:3 approximate:3 status:1 keep:3 global:2 active:9 uai:1 conclude:1 consuming:1 xi:9 discriminative:1 spectrum:1 continuous:1 iterative:1 additionally:1 learn:1 schnorr:2 ca:2 messagepassing:1 excellent:2 domain:1 main:1 motivation:1 border:1 lellmann:1 rubberwhale:1 body:1 xu:1 fig:7 referred:4 xmap:6 precision:1 paragios:2 wish:1 house:2 grained:1 rk:3 z0:3 transitioning:2 specific:4 ton:1 svm:1 grouping:15 exists:3 consist:1 niques:1 importance:1 simply:1 army:2 explore:1 visual:1 upsample:2 scalar:1 applies:1 pedro:2 truth:3 relies:3 constantly:1 lewis:1 conditional:1 lempitsky:2 goal:3 carsten:1 ann:1 paristech:2 hard:1 lidar:1 specifically:2 except:1 reducing:1 isprs:1 denoising:1 called:1 total:1 batra:1 duality:1 experimental:3 la:2 est:2 szummer:1 alexander:1 incorporate:1 evaluate:3 yoo:2 tested:1 |
4,813 | 5,358 | Probabilistic low-rank matrix completion on finite
alphabets
?
Eric
Moulines
Institut Mines-T?el?ecom
T?el?ecom ParisTech
CNRS LTCI
Olga Klopp
CREST et MODAL?X
Universit?e Paris Ouest
Jean Lafond
Institut Mines-T?el?ecom
T?el?ecom ParisTech
CNRS LTCI
[email protected]
[email protected]
[email protected]
Joseph Salmon
Institut Mines-T?el?ecom
T?el?ecom ParisTech
CNRS LTCI
[email protected]
Abstract
The task of reconstructing a matrix given a sample of observed entries is known
as the matrix completion problem. It arises in a wide range of problems, including recommender systems, collaborative filtering, dimensionality reduction,
image processing, quantum physics or multi-class classification to name a few.
Most works have focused on recovering an unknown real-valued low-rank matrix from randomly sub-sampling its entries. Here, we investigate the case where
the observations take a finite number of values, corresponding for examples to
ratings in recommender systems or labels in multi-class classification. We also
consider a general sampling scheme (not necessarily uniform) over the matrix
entries. The performance of a nuclear-norm penalized estimator is analyzed theoretically. More precisely, we derive bounds for the Kullback-Leibler divergence
between the true and estimated distributions. In practice, we have also proposed
an efficient algorithm based on lifted coordinate gradient descent in order to tackle
potentially high dimensional settings.
1
Introduction
Matrix completion has attracted a lot of contributions over the past decade. It consists in recovering
the entries of a potentially high dimensional matrix, based on their random and partial observations.
In the classical noisy matrix completion problem, the entries are assumed to be real valued and observed in presence of additive (homoscedastic) noise. In this paper, it is assumed that the entries take
values in a finite alphabet that can model categorical data. Such a problem arises in analysis of voting patterns, recovery of incomplete survey data (typical survey responses are true/false, yes/no or
do not know, agree/disagree/indifferent), quantum state tomography [13] (binary outcomes), recommender systems [18, 2] (for instance in common movie rating datasets, e.g., MovieLens or Neflix,
ratings range from 1 to 5) among many others. It is customary in this framework that rows represent
individuals while columns represent items e.g., movies, survey responses, etc. Of course, the observations are typically incomplete, in the sense that a significant proportion of the entries are missing.
Then, a crucial question to be answered is whether it is possible to predict the missing entries from
these partial observations.
1
Since the problem of matrix completion is ill-posed in general, it is necessary to impose a lowdimensional structure on the matrix, one particularly popular example being a low rank constraint.
The classical noisy matrix completion problem (real valued observations and additive noise), can be
solved provided that the unknown matrix is low rank, either exactly or approximately; see [7, 15, 17,
20, 5, 16] and the references therein. Most commonly used methods amount to solve a least square
program under a rank constraint or a convex relaxation of a rank constraint provided by the nuclear
(or trace norm) [10].
The problem of probabilistic low rank matrix completion over a finite alphabet has received much
less attention; see [22, 8, 6] among others. To the best of our knowledge, only the binary case
(also referred to as the 1-bit matrix completion problem) has been covered in depth. In [8], the
authors proposed to model the entries as Bernoulli random variables whose success rate depend
upon the matrix to be recovered through a convex link function (logistic and probit functions being
natural examples). The estimated matrix is then obtained as a solution of a maximization of the
log-likelihood of the observations under an explicit low-rank constraint. Moreover, the sampling
model proposed in [8] assumes that the entries are sampled uniformly at random. Unfortunately,
this condition is not totally realistic in recommender system applications: in such a context some
users are more active than others and some popular items are rated more frequently. Theoretically,
an important issue is that the method from [8] requires the knowledge of an upper bound on the
nuclear norm or on the rank of the unknown matrix.
Variations on the 1-bit matrix completion was further considered in [6] where a max-norm (though
the name is similar, this is different from the sup-norm) constrained minimization is considered. The
method of [6] allows more general non-uniform samplings but still requires an upper bound on the
max-norm of the unknown matrix.
In the present paper we consider a penalized maximum log-likelihood method, in which the loglikelihood of the observations is penalized by the nuclear norm (i.e., we focus on the Lagrangian
version rather than on the constrained one). We first establish an upper bound of the KullbackLeibler divergence between the true and the estimated distribution under general sampling distributions; see Section 2 for details. One should note that our method only requires the knowledge of
an upper bound on the maximum absolute value of the probabilities, and improves upon previous
results found in the literature.
Last but not least, we propose an efficient implementation of our statistical procedure, which is
adapted from the lifted coordinate descent algorithm recently introduced in [9, 14]. Unlike other
methods, this iterative algorithm is designed to solve the convex optimization and not (possibly nonconvex) approximated formulation as in [21]. It also has the benefit that it does not need to perform
full/partial SVD (Singular Value Decomposition) at every iteration; see Section 3 for details.
Notation
Define m1 ? m2 := min(m1 , m2 ) and m1 ? m2 := max(m1 , m2 ). We equip the set of m1 ? m2
matrices with real entries (denoted Rm1 ?m2 ) with the scalar product hX|X 0 i := tr(X > X 0 ). For a
given matrix X ? Rm1 ?m2 we write kXk? := maxi,j |Xi,j | and, for q ? 1, we denote its Schatten
q-norm by
!1/q
mX
1 ?m2
q
kXk?,q :=
?i (X)
,
i=1
where ?i (X) are the singular values of X ordered in decreasing order (see [1] for more details on
such norms). The operator norm of X is given by kXk?,? := ?1 (X). Consider two vectors of
j
p ? 1 matrices (X j )p?1
and (X 0j )p?1
j=1 such that for any (k, l) ? [m1 ] ? [m2 ] we have Xk,l ? 0,
Pp?1 j=1j
Pp?1 0j
0j
Xk,l ? 0, 1 ? j=1 Xk,l ? 0 and 1 ? j=1 Xk,l ? 0. Their square Hellinger distance is
d2H (X, X 0 ) :=
1
m1 m2
?
v
?v
?2 ?
u
2 u
p?1 q
p?1
p?1
q
X ?X
X
X
u
u
j
0j
j
0j ? ?
Xk,l
+?t1 ?
Xk,l
Xk,l
? Xk,l
? t1 ?
?
?
k?[m1 ] j=1
l?[m2 ]
j=1
2
j=1
and their Kullback-Leibler divergence is
?
Pp?1 j ?
j
p?1
p?1
X
X X
1
?
X
1
j=1 Xk,l
k,l
j
j
?
KL (X, X 0 ) :=
Xk,l
) log
Xk,l
log 0j + (1 ?
Pp?1 0j ? .
m1 m2
Xk,l
1 ? j=1 Xk,l
j=1
k?[m1 ] j=1
l?[m2 ]
Given an integer p > 1, a function f : Rp?1 ? Rp?1 is called a p-link function if for any x ? Rp?1
Pp?1
it satisfies f j (x) ? 0 for j ? [p ? 1] and 1 ? j=1 f j (x) ? 0. For any collection of p ? 1 matrices
j
j
j p?1
(X j )p?1
j=1 , f (X) denotes the vector of matrices (f (X) )j=1 such that f (X)k,l = f (Xk,l ) for any
(k, l) ? [m1 ] ? [m2 ] and j ? [p ? 1].
2
Main results
Let p denote the cardinality of our finite alphabet, that is the number of classes of the logistic model
(e.g., ratings have p possible values or surveys p possible answers). For a vector of p ? 1 matrices
m1 ?m2
and an index ? ? [m1 ] ? [m2 ], we denote by X? the vector (X?j )p?1
X = (X j )p?1
j=1 of R
j=1 .
We consider an i.i.d. sequence (?i )1?i?n over [m1 ] ? [m2 ], with a probability distribution function
? that controls the way the matrix entries are revealed. It is customary to consider the simple
uniform sampling distribution over the set [m1 ] ? [m2 ], though more general sampling schemes
could be considered as well. We observe n independent random elements (Yi )1?i?n ? [p]n . The
observations (Y1 , . . . , Yn ) are assumed to be independent and to follow a multinomial distribution
with success probabilities given by
? ?1 , . . . , X
? ?p?1 ) j ? [p ? 1] and
P(Yi = j) = f (X
i
i
j
P(Yi = p) = 1 ?
p?1
X
P(Yi = j)
j=1
?
? j p?1
where {f j }p?1
j=1 is a p-link function and X = (X )j=1 is the vector of true (unknown) parameters
? i instead of X
? ? . Let us denote by ?Y
we aim at recovering. For ease of notation, we often write X
i
the (normalized) negative log-likelihood of the observations:
?
?
??
p?1
p?1
n
X
1 X ?X
1{Yi =j} log f j (Xi ) + 1{Yi =p} log ?1 ?
f j (Xi )?? ,
(1)
?Y (X) = ?
n i=1 j=1
j=1
For any ? > 0 our proposed estimator is the following:
?=
X
arg min
X?(Rm1 ?m2 )p?1
maxj?[p?1] kX j k? ??
??Y (X) ,
where ??Y (X) = ?Y (X) + ?
p?1
X
kX j k?,1 ,
(2)
j=1
with ? > 0 being a regularization parameter controlling the rank of the estimator. In the rest of the
paper we assume that the negative log-likelihood ?Y is convex (this is the case for the multinomial
logit function, see for instance [3]).
? in the binomial setting
In this section we present two results controlling the estimation error of X
(i.e., when p = 2). Before doing so, let us introduce some additional notation and assumptions. The
?
score function (defined as the gradient of the negative log-likelihood) taken at the true parameter X,
? := ? ?Y (X).
? We also need the following constants depending on the link function
is denoted by ?
f and ? > 0:
M? = sup 2| log(f (x))| ,
|x|??
|f 0 (x)|
|f 0 (x)|
L? = max sup
, sup
|x|?? f (x) |x|?? 1 ? f (x)
K? = inf
|x|??
f 0 (x)2
.
8f (x)(1 ? f (x))
3
!
,
In our framework, we allow for a general distribution for observing the coefficients. However, we
need to control deviations of the sampling mechanism from the uniform distribution and therefore
we consider the following assumptions.
H1. There exists a constant ? ? 1 such that for all indexes (k, l) ? [m1 ] ? [m2 ]
min(?k,l ) ? 1/(?m1 m2 ) .
k,l
with ?k,l := ?(?1 = (k, l)).
Pm2
Pm1
?k,l ) for any l ? [m2 ] (resp. k ? [m1 ]) the
?k,l (resp. Rk := l=1
Let us define Cl := k=1
probability of sampling a coefficient in column l (resp. in row k).
H2. There exists a constant ? ? 1 such that
max(Rk , Cl ) ? ?/(m1 ? m2 ) ,
k,l
Assumption H1 ensures that each coefficient has a non-zero probability of being sampled whereas
H2 requires that no column nor row is sampled with too high probability (see also [11, 16] for more
details on this condition).
We define the sequence of matrices (Ei )ni=1 associated to the revealed coefficient (?i )ni=1 by
0 m2
1
Ei := eki (e0li )> where (ki , li ) = ?i and with (ek )m
k=1 (resp. (el )l=1 ) being the canonical bam2
m1
sis of R (resp. R ). Furthermore, if (?i )1?i?n is a Rademacher sequence independent from
(?i )ni=1 and (Yi )1?i?n we define
n
1X
?R :=
?i Ei .
n i=1
We can now state our first result. For completeness, the proofs can be found in the supplementary
material.
? ?,? and kXk
? ? ? ?. Then, with probability at least
Theorem 1. Assume H1 holds, ? ? 2k?k
1 ? 2/d the Kullback-Leibler divergence between the true and estimated distribution is bounded by
!
p
log(d)
?2
2
? 2
2
?
?
?
KL f (X), f (X) ? 8 max
m1 m2 rank(X) ? + c L? (Ek?R k?,? ) , ?eM?
,
K?
n
where c? is a universal constant.
? ?,? is stochastic and that its expectation Ek?R k?,? is unknown. However, thanks to
Note that k?k
Assumption H2 these quantities can be controlled.
To ease notation let us also define m := m1 ? m2 , M := m1 ? m2 and d := m1 + m2 .
? ? ? ?. Assume in addition that n ?
Theorem 2. Assume H 1 and H 2p
hold and that kXk
2m log(d)/(9?). Taking ? = 6L? 2? log(d)/(mn), then with probability at least 1 ? 3/d the
folllowing holds
!
p
? ? Xk
? 2
? log(d)
kX
??2 L2? M rank(X)
log(d)
?,2
?
?
? KL f (X), f (X) ? max c?
, 8?eM?
,
K?
m1 m2
K?
n
n
where c? is a universal constant.
Remark. Let us compare the rate of convergence of Theorem 2 with those obtained in previous
? is estimated by minimizing the negative
works on 1-bit matrix completion. In [8], the parameter X
?
log-likelihood under the constraints kXk? ? ? and kXk?,1 ? ? rm1 m2 for some r > 0. Under
? ? r, they could prove that
the assumption that rank(X)
r
? ? Xk
? 2
kX
rd
?,2
? C?
,
m1 m2
n
where C? is a constant depending on ? (see [8, Theorem 1]). This rate of convergence is slower
than the rate of convergence given by Theorem 2. [6] studied a max-norm constrained maximum
likelihood estimate and obtained a rate of convergence similar to [8].
4
3
Numerical Experiments
Implementation For numerical experiments, data were simulated according to a multinomial
logit distribution. In this setting, an observation Yk,l associated to row k and column l is distributed
p?1
1
as P(Yk,l = j) = f j (Xk,l
, . . . , Xk,l
) where
?
??1
p?1
X
f j (x1 , . . . , xp?1 ) = exp(xj ) ?1 +
exp(xj )? , for j ? [p ? 1] .
(3)
j=1
With this choice, ?Y is convex and problem (2) can be solved using convex optimization algorithms. Moreover, following the advice of [8] we considered the unconstrained version of problem
(2) (i.e., with no constraint on kXk? ), which reduces significantly the computation burden and has
no significant impact on the solution in practice. To solve this problem, we have extended to the
multinomial case the coordinate gradient descent algorithm introduced by [9]. This type of algorithm has the advantage, say over the Soft-Impute [19] or the SVT [4] algorithm, that it does not
require the computation of a full SVD at each step of the main loop of an iterative (proximal) algorithm (bare in mind that the proximal operator associated to the nuclear norm is the soft-thresholding
operator of the singular values). The proposed version only computes the largest singular vectors
and singular values. This potentially decreases the computation by a factor close to the value of the
upper bound on the rank commonly used (see the aforementioned paper for more details).
Let us present the algorithm. Any vector of p ? 1 matrices X = (X j )p?1
j=1 is identified as an element
of the tensor product space Rm1 ?m2 ? Rp?1 and denoted by:
X=
p?1
X
X j ? ej ,
(4)
j=1
p?1
where again (ej )p?1
and ? stands for the tensor product. The set of
j=1 is the canonical basis on R
normalized rank-one matrices is denoted by
M := M ? Rm1 ?m2 |M = uv > | kuk = kvk = 1, u ? Rm1 , v ? Rm2 .
Define ? the linear space of real-valued functions on M with finite support, i.e., ?(M
P ) = 0 except
for a finite number of M ? M. This space is equipped with the `1 -norm k?k1 = M ?M |?(M )|.
Define by ?+ the positive orthant, i.e., the cone of functions ? ? ? such that ?(M ) ? 0 for all
p?1
M ? M. Any tensor X can be associated with a vector ? = (?1 , . . . , ?p?1 ) ? ?+
, i.e.,
X=
p?1 X
X
?j (M )M ? ej .
(5)
j=1 M ?M
Such representations are not unique, and among them, the one associated to the SVD plays a key
role, as we will see below. For a given X represented by (4) and for any j ? {1, . . . , p ? 1}, denote
j
j
by {?kj }nk=1 the (non-zero) singular values of the matrix X j and {ujk ,vkj }nk=1 the associated singular
vectors. Then, X may be expressed as
j
X=
p?1 X
n
X
?kj ujk (vkj )> ? ej .
(6)
j=1 k=1
Defining ?j the function ?j (M ) = ?kj if M = ujk (vkj )> , k ? [nj ] and ?j (M ) = 0 otherwise, one
obtains a representation of the type given in Eq. (5).
Conversely, for any ? = (?1 , . . . , ?p?1 ) ? ?p?1 , define the map
W : ? ? W? :=
p?1
X
W?j ? ej
with W?j :=
X
?j (M )M
M ?M
j=1
and the auxiliary objective function
? ?Y (?) = ?
?
p?1 X
X
j=1 M ?M
5
?j (M ) + ?Y (W? ) .
(7)
The map ? 7? W? is a continuous linear map from (?p?1 , k ? k1 ) to Rm1 ?m2 ? Rp?1 , where
Pp?1 P
p?1
k?k1 = j=1 M ?M |?j (M )|. In addition, for all ? ? ?+
p?1
X
kW?j k?,1 ? k?k1 ,
j=1
Pp?1
j
j=1 kW? k?,1
and one obtains k?k1 =
when ? is the representation associated to the SVD decomposition. An important consequence, outlined in [9, Proposition 3.1], is that the minimization of (7)
is actually equivalent to the minimization of (2); see [9, Theorem 3.2].
The proposed coordinate gradient descent algorithm updates at each step the nonnegative finite support function ?. For ? ? ? we denote by supp(?) the support of ? and for M ? M, by ?M ? ? the
Dirac function on M satisfying ?M (M ) = 1 and ?M (M 0 ) = 0 if M 0 6= M . In our experiments we
have set to zero the initial ?0 .
Algorithm 1: Multinomial lifted coordinate gradient descent
Data: Observations: Y , tuning parameter ?
initial parameter: ?0 ? ?p?1
+ ; tolerance: ; maximum number of iterations: K
Result: ? ? ?p?1
+
Initialization: ? ? ?0 , k ? 0
while k ? K do
for j = 0 to p ? 1 do
Compute top singular vectors pair of (?? ?Y (W? ))j : uj , vj
Let g = ? + minj=1,...,p?1 h? ?Y | uj (v j )> i
if g ? ?/2 then
? ? ? + (b0 ?u0 (v0 )> , . . . , bp?1 ?up?1 (vp?1 )> )
(?0 , . . . , ?p?1 ) =
arg min
?
Y
(b0 ,...,bp?1 )?Rp?1
+
? ? ? + (?0 ?u0 (v0 )> , . . . , ?p?1 ?up?1 (vp?1 )> )
k ?k+1
else
Let gmax = maxj?[p?1] maxuj (vj )> ?supp(?j ) |? + h? ?Y | uj (v j )> i|
if gmax ? then
break
else
? ?Y (?0 )
??
arg min
?
? 0 ??p?1
,supp(? 0j )?supp(? j ),j?[p?1]
+
k ?k+1
A major interest of Algorithm 1 is that it requires to store the value of the parameter entries only
for the indexes which are actually observed. Since in practice the number of observations is much
smaller than the total number of coefficients m1 m2 , this algorithm is both memory and computationally efficient. Moreover, using an SVD algorithm such as Arnoldi iterations to compute the top
singular values and vector pairs (see [12, Section 10.5] for instance) allows us to take full advantage
of gradient sparse structure. Algorithm 1 was implemented in C and Table 1 gives a rough idea
of the execution time for the case of two classes on a 3.07Ghz w3550 Xeon CPU (RAM 1.66 Go,
Cache 8Mo).
Simulated experiments To evaluate our procedure we have performed simulations for matrices with p = 2 or 5. For each class matrix X j we sampled uniformly five unitary vector pairs
(ujk , vkj )5k=1 . We have then generated matrices of rank equals to 5, such that
5
X
?
X j = ? m1 m2
?k ujk (vkj )> ,
k=1
?
with (?1 , . . . , ?5 ) = (2, 1, 0.5, 0.25, 0.1) and ? is a scaling factor. The m1 m2 factor, guarantees
that E[kX j k? ] does not depend on the sizes of the problem m1 and m2 .
6
Parameter Size
Observations
Execution Time (s.)
103 ? 103
105
4.5
3 ? 103 ? 3 ? 103
105
52
104 ? 104
107
730
Table 1: Execution time of the proposed algorithm for the binary case.
We then sampled the entries uniformly and the observations according to a logit distribution given
by Eq. (3). We have then considered and compared the two following estimators both computed
using Algorithm 1:
? the logit version of our method (with the link function given by Eq. (3))
? N ), that consists in using the Gaussian
? the Gaussian completion method (denoted by X
log-likelihood instead of the multinomial in (2), i.e., using a classical squared Frobenius
norm (the implementation being adapted mutatis mutandis). Moreover an estimation of the
standard deviation is obtained by the classical analysis of the residue.
Contrary to the logit version, the Gaussian matrix completion does not directly recover the probabilities of observing a rating. However, we can estimate this probability by the following quantity:
?
if j = 1 ,
?
?0
?N
j?0.5?X
N
?
k,l
P(Xk,l = j) = FN (0,1) (pj+1 ) ? FN (0,1) (pj ) with pj =
if 0 < j < p
?
?
?
?
1
if j = p ,
where FN (0,1) is the cdf of a zero-mean standard Gaussian random variable.
As we see on Figure 1, the logistic estimator outperforms the Gaussian for both cases p = 2 and
p = 5 in terms of the Kullback-Leibler divergence. This was expected because the Gaussian model
allows uniquely symmetric distributions with the same variance for all the ratings, which is not the
case for logistic distributions. The choice of the ? parameter has been set for both methods by
performing 5-fold cross-validation on a geometric grid of size 0.8 log(n).
Table 2 and Table 3 summarize the results obtained for a 900 ? 1350 matrix respectively for p = 2
and p = 5. For both the binomial case p = 2 and the multinomial case p = 5, the logistic model
slightly outperforms the Gaussian model. This is partly due to the fact that in the multinomial case,
some ratings can have a multi-modal distribution. In such a case, the Gaussian model is unable
to predict these ratings, because its distribution is necessarily centered around a single value and
is not flexible enough. For instance consider the case of a rating distribution with high probability
of seeing 1 or 5, low probability of getting 2, 3 and 4, where we observed both 1?s and 5?s. The
estimator based on a Gaussian model will tend to center its distribution around 2.5 and therefore
misses the bimodal shape of the distribution.
Observations
Gaussian prediction error
Logistic prediction error
10 ? 103
0.49
0.42
50 ? 103
0.34
0.30
100 ? 103
0.29
0.27
500 ? 103
0.26
0.24
Table 2: Prediction errors for a binomial (2 classes) underlying model, for a 900 ? 1350 matrix.
Observations
Gaussian prediction error
Logistic prediction error
10 ? 103
0.78
0.75
50 ? 103
0.76
0.54
100 ? 103
0.73
0.47
500 ? 103
0.69
0.43
Table 3: Prediction Error for a multinomial (5 classes) distribution against a 900 ? 1350 matrix.
Real dataset We have also run the same estimators on the MovieLens 100k dataset. In the case
of real data we cannot calculate the Kullback-Leibler divergence since no ground truth is available.
Therefore, to compare the prediction errors, we randomly selected 20% of the entries as a test set,
and the remaining entries were split between a training set (80%) and a validation set (20%).
7
Normalized KL divergence for logistic (plain), Gaussian (dashed)
Normalized KL divergence for logistic (plain), Gaussian (dashed)
0.20
0.20
size: 100x150
size: 300x450
size: 900x1350
0.15
Mean KL divergence
Mean KL divergence
size: 100x150
size: 300x450
size: 900x1350
0.10
0.10
0.05
0.05
0.00
0.15
100000
200000
300000
400000
0.00
500000
100000
Number of observations
200000
300000
400000
500000
Number of observations
Figure 1: Kullback-Leibler divergence between the estimated and the true model for different
matrices sizes and sampling fraction, normalized by number of classes. Right figure: binomial
and Gaussian models ; left figure: multinomial with five classes and Gaussian model. Results are
averaged over five samples.
For this dataset, ratings range from 1 to 5. To consider the benefit of a binomial model, we have
tested each rating against the others (e.g., ratings 5 are set to 0 and all others are set to 1). Interestingly we see that the Gaussian prediction error is significantly better when choosing labels ?1, 1
instead of labels 0, 1. This is another motivation for not using the Gaussian version: the sensibility to
the alphabet choice seems to be crucial for the Gaussian version, whereas the binomial/multinomial
ones are insensitive to it. These results are summarized in table 4.
Rating
Gaussian prediction error (labels ?1 and 1)
Gaussian prediction error (labels 0 and 1)
Logistic prediction error
1
0.06
0.12
0.06
2
0.12
0.20
0.11
3
0.28
0.39
0.27
4
0.35
0.46
0.34
5
0.19
0.30
0.20
Table 4: Binomial prediction error when performing one versus the others procedure on the MovieLens 100k dataset.
4
Conclusion and future work
We have proposed a new nuclear norm penalized maximum log-likelihood estimator and have provided strong theoretical guarantees on its estimation accuracy in the binary case. Compared to
previous works on 1-bit matrix completion, our method has some important advantages. First, it
works under quite mild assumptions on the sampling distribution. Second, it requires only an upper bound on the maximal absolute value of the unknown matrix. Finally, the rates of convergence
given by Theorem 2 are faster than the rates of convergence obtained in [8] and [6]. In future work,
we could consider the extension to more general data fitting terms, and to possibly generalize the
results to tensor formulations, or to penalize directly the nuclear norm of the matrix probabilities
themselves.
Acknowledgments
Jean Lafond is grateful for fundings from the Direction G?en?erale de l?Armement (DGA) and to the
labex LMH through the grant no ANR-11-LABX-0056-LMH in the framework of the ?Programme
des Investissements d?Avenir?. Joseph Salmon acknowledges Chair Machine Learning for Big Data
for partial financial support. The authors would also like to thank Alexandre Gramfort for helpful
discussions.
8
References
[1] R. Bhatia. Matrix analysis, volume 169 of Graduate Texts in Mathematics. Springer-Verlag,
New York, 1997.
[2] J. Bobadilla, F. Ortega, A. Hernando, and A. Guti?errez.
Knowledge-Based Systems, 46(0):109 ? 132, 2013.
Recommender systems survey.
[3] S. Boyd and L. Vandenberghe. Convex optimization. Cambridge University Press, Cambridge,
2004.
[4] J-F. Cai, E. J. Cand`es, and Z. Shen. A singular value thresholding algorithm for matrix completion. SIAM Journal on Optimization, 20(4):1956?1982, 2010.
[5] T. T. Cai and W-X. Zhou. Matrix completion via max-norm constrained optimization. CoRR,
abs/1303.0341, 2013.
[6] T. T. Cai and W-X. Zhou. A max-norm constrained minimization approach to 1-bit matrix
completion. J. Mach. Learn. Res., 14:3619?3647, 2013.
[7] E. J. Cand`es and Y. Plan. Matrix completion with noise. Proceedings of the IEEE, 98(6):925?
936, 2010.
[8] M. A. Davenport, Y. Plan, E. van den Berg, and M. Wootters. 1-bit matrix completion. CoRR,
abs/1209.3672, 2012.
[9] M. Dud??k, Z. Harchaoui, and J. Malick. Lifted coordinate descent for learning with trace-norm
regularization. In AISTATS, 2012.
[10] M. Fazel. Matrix rank minimization with applications. PhD thesis, Stanford University, 2002.
[11] R. Foygel, R. Salakhutdinov, O. Shamir, and N. Srebro. Learning with the weighted trace-norm
under arbitrary sampling distributions. In NIPS, pages 2133?2141, 2011.
[12] G. H. Golub and C. F. van Loan. Matrix computations. Johns Hopkins University Press,
Baltimore, MD, fourth edition, 2013.
[13] D. Gross. Recovering low-rank matrices from few coefficients in any basis. Information
Theory, IEEE Transactions on, 57(3):1548?1566, 2011.
[14] Z. Harchaoui, A. Juditsky, and A. Nemirovski. Conditional gradient algorithms for normregularized smooth convex optimization. Mathematical Programming, pages 1?38, 2014.
[15] R. H. Keshavan, A. Montanari, and S. Oh. Matrix completion from noisy entries. J. Mach.
Learn. Res., 11:2057?2078, 2010.
[16] O. Klopp. Noisy low-rank matrix completion with general sampling distribution. Bernoulli,
2(1):282?303, 02 2014.
[17] V. Koltchinskii, A. B. Tsybakov, and K. Lounici. Nuclear-norm penalization and optimal rates
for noisy low-rank matrix completion. Ann. Statist., 39(5):2302?2329, 2011.
[18] Y. Koren, R. Bell, and C. Volinsky. Matrix factorization techniques for recommender systems.
Computer, 42(8):30?37, 2009.
[19] R. Mazumder, T. Hastie, and R. Tibshirani. Spectral regularization algorithms for learning
large incomplete matrices. J. Mach. Learn. Res., 11:2287?2322, 2010.
[20] S. Negahban and M. J. Wainwright. Restricted strong convexity and weighted matrix completion: optimal bounds with noise. J. Mach. Learn. Res., 13:1665?1697, 2012.
[21] B. Recht and C. R?e. Parallel stochastic gradient algorithms for large-scale matrix completion.
Mathematical Programming Computation, 5(2):201?226, 2013.
[22] A. Todeschini, F. Caron, and M. Chavent. Probabilistic low-rank matrix completion with
adaptive spectral regularization algorithms. In NIPS, pages 845?853, 2013.
9
| 5358 |@word mild:1 version:7 norm:21 proportion:1 logit:5 seems:1 simulation:1 decomposition:2 tr:1 reduction:1 initial:2 score:1 interestingly:1 past:1 outperforms:2 recovered:1 si:1 attracted:1 john:1 fn:3 additive:2 realistic:1 numerical:2 shape:1 designed:1 update:1 juditsky:1 selected:1 item:2 xk:19 completeness:1 math:1 five:3 mathematical:2 consists:2 prove:1 fitting:1 introduce:1 hellinger:1 theoretically:2 expected:1 themselves:1 frequently:1 nor:1 multi:3 cand:2 moulines:2 salakhutdinov:1 decreasing:1 cpu:1 equipped:1 cardinality:1 totally:1 cache:1 provided:3 moreover:4 notation:4 bounded:1 underlying:1 nj:1 sensibility:1 guarantee:2 every:1 voting:1 tackle:1 exactly:1 universit:1 control:2 grant:1 yn:1 arnoldi:1 t1:2 before:1 positive:1 svt:1 consequence:1 mach:4 approximately:1 therein:1 studied:1 initialization:1 koltchinskii:1 conversely:1 ease:2 factorization:1 nemirovski:1 range:3 graduate:1 averaged:1 fazel:1 unique:1 acknowledgment:1 practice:3 procedure:3 universal:2 bell:1 significantly:2 boyd:1 seeing:1 cannot:1 close:1 operator:3 context:1 equivalent:1 map:3 lagrangian:1 missing:2 center:1 go:1 attention:1 convex:8 focused:1 survey:5 shen:1 recovery:1 m2:38 estimator:8 nuclear:8 vandenberghe:1 oh:1 financial:1 coordinate:6 variation:1 resp:5 controlling:2 play:1 shamir:1 user:1 programming:2 element:2 approximated:1 particularly:1 satisfying:1 observed:4 role:1 solved:2 calculate:1 ensures:1 decrease:1 yk:2 gross:1 convexity:1 mine:3 depend:2 grateful:1 upon:2 eric:1 basis:2 represented:1 alphabet:5 bhatia:1 outcome:1 choosing:1 jean:3 whose:1 posed:1 valued:4 solve:3 loglikelihood:1 supplementary:1 say:1 otherwise:1 anr:1 stanford:1 noisy:5 eki:1 sequence:3 advantage:3 cai:3 propose:1 lowdimensional:1 product:3 maximal:1 fr:4 loop:1 erale:1 frobenius:1 dirac:1 getting:1 convergence:6 rademacher:1 derive:1 depending:2 completion:24 ouest:1 b0:2 received:1 eq:3 strong:2 recovering:4 auxiliary:1 implemented:1 avenir:1 direction:1 stochastic:2 centered:1 material:1 require:1 hx:1 proposition:1 extension:1 hold:3 around:2 considered:5 ground:1 exp:2 predict:2 mo:1 major:1 homoscedastic:1 estimation:3 label:5 mutandis:1 largest:1 weighted:2 minimization:5 rough:1 gaussian:20 aim:1 rather:1 zhou:2 ej:5 lifted:4 focus:1 rank:21 bernoulli:2 likelihood:9 sense:1 helpful:1 el:7 cnrs:4 typically:1 issue:1 classification:2 among:3 ill:1 denoted:5 arg:3 aforementioned:1 flexible:1 plan:2 constrained:5 gramfort:1 malick:1 equal:1 sampling:13 kw:2 future:2 others:6 few:2 randomly:2 divergence:11 individual:1 maxj:2 ab:2 ltci:3 interest:1 investigate:1 indifferent:1 golub:1 analyzed:1 kvk:1 pm2:1 partial:4 necessary:1 gmax:2 institut:3 incomplete:3 re:4 theoretical:1 instance:4 column:4 soft:2 xeon:1 maximization:1 deviation:2 entry:17 uniform:4 too:1 kullbackleibler:1 answer:1 proximal:2 thanks:1 recht:1 siam:1 negahban:1 probabilistic:3 physic:1 hopkins:1 again:1 squared:1 thesis:1 possibly:2 davenport:1 ek:3 li:1 supp:4 de:2 summarized:1 coefficient:6 dga:1 performed:1 h1:3 lot:1 break:1 doing:1 sup:4 observing:2 recover:1 parallel:1 collaborative:1 contribution:1 square:2 ni:3 accuracy:1 variance:1 yes:1 pm1:1 vp:2 generalize:1 minj:1 against:2 volinsky:1 pp:7 associated:7 proof:1 sampled:5 dataset:4 popular:2 knowledge:4 dimensionality:1 improves:1 actually:2 alexandre:1 follow:1 modal:2 response:2 formulation:2 lounici:1 though:2 furthermore:1 ei:3 keshavan:1 logistic:10 name:2 normalized:5 true:7 regularization:4 dud:1 symmetric:1 leibler:6 impute:1 uniquely:1 ortega:1 image:1 salmon:3 recently:1 funding:1 guti:1 common:1 multinomial:11 insensitive:1 volume:1 m1:30 significant:2 cambridge:2 caron:1 rd:1 unconstrained:1 uv:1 lafond:3 outlined:1 tuning:1 grid:1 mathematics:1 v0:2 etc:1 inf:1 store:1 verlag:1 nonconvex:1 binary:4 success:2 yi:7 additional:1 impose:1 dashed:2 u0:2 full:3 harchaoui:2 reduces:1 smooth:1 faster:1 cross:1 controlled:1 impact:1 prediction:12 expectation:1 iteration:3 represent:2 bimodal:1 penalize:1 whereas:2 addition:2 residue:1 baltimore:1 else:2 singular:10 crucial:2 rest:1 unlike:1 tend:1 contrary:1 integer:1 unitary:1 presence:1 revealed:2 split:1 enough:1 quite:1 xj:2 ujk:5 hastie:1 identified:1 idea:1 whether:1 york:1 remark:1 wootters:1 covered:1 amount:1 tsybakov:1 statist:1 tomography:1 chavent:1 canonical:2 estimated:6 tibshirani:1 write:2 key:1 pj:3 kuk:1 ram:1 relaxation:1 fraction:1 cone:1 run:1 fourth:1 scaling:1 bit:6 bound:8 ki:1 koren:1 fold:1 nonnegative:1 adapted:2 precisely:1 constraint:6 bp:2 answered:1 min:5 chair:1 performing:2 according:2 vkj:5 smaller:1 slightly:1 reconstructing:1 em:2 joseph:3 errez:1 den:1 restricted:1 taken:1 ecom:6 computationally:1 agree:1 foygel:1 mechanism:1 know:1 mind:1 available:1 observe:1 spectral:2 slower:1 customary:2 rp:6 assumes:1 denotes:1 binomial:7 top:2 remaining:1 todeschini:1 k1:5 uj:3 establish:1 klopp:3 classical:4 tensor:4 objective:1 question:1 quantity:2 md:1 gradient:8 mx:1 distance:1 link:5 unable:1 schatten:1 simulated:2 lmh:2 thank:1 equip:1 index:3 minimizing:1 unfortunately:1 potentially:3 trace:3 negative:4 implementation:3 unknown:7 perform:1 recommender:6 disagree:1 observation:18 upper:6 datasets:1 finite:8 descent:6 orthant:1 defining:1 extended:1 y1:1 arbitrary:1 rating:13 introduced:2 pair:3 paris:1 kl:7 nip:2 below:1 pattern:1 summarize:1 program:1 including:1 max:10 memory:1 wainwright:1 natural:1 mn:1 scheme:2 movie:2 rated:1 acknowledges:1 categorical:1 bare:1 kj:3 text:1 literature:1 l2:1 geometric:1 probit:1 x150:2 filtering:1 srebro:1 versus:1 validation:2 h2:3 labex:1 penalization:1 xp:1 thresholding:2 row:4 course:1 penalized:4 last:1 allow:1 wide:1 taking:1 rm2:1 absolute:2 sparse:1 benefit:2 distributed:1 tolerance:1 depth:1 ghz:1 stand:1 plain:2 van:2 quantum:2 computes:1 author:2 commonly:2 collection:1 adaptive:1 programme:1 transaction:1 crest:1 obtains:2 kullback:6 active:1 assumed:3 xi:3 continuous:1 iterative:2 decade:1 table:8 learn:4 mazumder:1 necessarily:2 cl:2 vj:2 aistats:1 main:2 montanari:1 motivation:1 noise:4 big:1 edition:1 x1:1 advice:1 telecom:3 referred:1 en:1 sub:1 explicit:1 rk:2 theorem:7 normregularized:1 maxi:1 exists:2 burden:1 false:1 corr:2 phd:1 execution:3 kx:5 nk:2 kxk:8 ordered:1 expressed:1 mutatis:1 scalar:1 d2h:1 springer:1 truth:1 satisfies:1 cdf:1 labx:1 conditional:1 ann:1 rm1:8 paristech:6 loan:1 typical:1 movielens:3 uniformly:3 except:1 olga:2 miss:1 called:1 total:1 partly:1 svd:5 e:2 berg:1 support:4 arises:2 evaluate:1 tested:1 |
4,814 | 5,359 | Controlling privacy in recommender systems
Tommi Jaakkola
CSAIL, MIT
[email protected]
Yu Xin
CSAIL, MIT
[email protected]
Abstract
Recommender systems involve an inherent trade-off between accuracy of recommendations and the extent to which users are willing to release information about
their preferences. In this paper, we explore a two-tiered notion of privacy where
there is a small set of ?public? users who are willing to share their preferences
openly, and a large set of ?private? users who require privacy guarantees. We
show theoretically and demonstrate empirically that a moderate number of public
users with no access to private user information already suffices for reasonable
accuracy. Moreover, we introduce a new privacy concept for gleaning relational
information from private users while maintaining a first order deniability. We
demonstrate gains from controlled access to private user preferences.
1
Introduction
Recommender systems exploit fragmented information available from each user. In a realistic system there?s also considerable ?churn?, i.e., users/items entering or leaving the system. The core
problem of transferring the collective experience of many users to an individual user can be understood in terms of matrix completion ([13, 14]). Given a sparsely populated matrix of preferences,
where rows and columns of the matrix correspond to users and items, respectively, the goal is to
predict values for the missing entries.
Matrix completion problems can be solved as convex regularization problems, using trace norm
as a convex surrogate to rank. A number of algorithms are available for solving large-scale tracenorm regularization problems. Such algorithms typically operate by iteratively building the matrix
from rank-1 components (e.g., [7, 17]). Under reasonable assumptions (e.g., boundedness, noise,
restricted strong convexity), the resulting empirical estimators have been shown to converge to the
underlying matrix with high probability ([12, 8, 2]). Consistency guarantees have mostly involved
matrices of fixed dimension, i.e., generalization to new users is not considered. In this paper, we
reformulate the regularization problem in a manner that depends only on the item (as opposed to
user) features, and characterize the error for out-of-sample users.
The completion accuracy depends directly on the amount of information that each user is willing to share with the system ([1]). It may be possible in some cases to side-step this statistical
trade-off by building Peer-to-Peer networks with homomorphic encryption that is computationally
challenging([3, 11]). We aim to address the statistical question directly.
The statistical trade-off between accuracy and privacy further depends on the notion of privacy we
adopt. A commonly used privacy concept is Differential Privacy (DP) ([6]), first introduced to
protect information leaked from database queries. In a recommender context, users may agree to a
trusted party to hold and aggregate their data, and perform computations on their behalf. Privacy
guarantees are then sought for any results published beyond the trusted party (including back to
the users). In this setting, differential privacy can be achieved through obfuscation (adding noise)
without a significant loss of accuracy ([10]).
1
In contrast to [10], we view the system as an untrusted entity, and assume that users wish to guard
their own data. We depart from differential privacy and separate computations that can be done
locally (privately) by individual users and computations that must be performed by the system (e.g.,
aggregation). For example, in terms of low rank matrices, only the item features have to be solved by
the system. The corresponding user features can be obtained locally by the users and subsequently
used for ranking.
From this perspective, we divide the set of users into two pools, the set of public users who openly
share their preferences, and the larger set of private users who require explicit privacy guarantees.
We show theoretically and demonstrate empirically that a moderate number of public users suffice
for accurate estimation of item features. The remaining private users can make use of these item
features without any release of information. Moreover, we propose a new 2nd order privacy concept
which uses limited (2nd order) information from the private users as well, and illustrate how recommendations can be further improved while maintaining marginal deniability of private information.
2
Problem formulation and summary of results
Recommender setup without privacy Consider a recommendation problem with n users and
? ? Rn?m . If only a few
m items. The underlying complete rating matrix to be recovered is X
? can be assumed to have low rank. As such, it is also
latent factors affect user preferences, X
recoverable from a small number of observed entries. We assume that entries are observed with
noise. Specifically,
?i,j + i,j , (i, j) ? ?
Yi,j = X
(1)
where ? denotes the set of observed entries. Noise is assumed to be i.i.d and follows a zeromean sub-Gaussian distribution with parameter kk?2 = ?. Following [16], we refer to the noise
distribution as Sub(? 2 ).
To bias our estimated rating matrix X to have low rank, we use convex relaxation of rank P
in the form
of trace norm. The trace-norm is the sum of singular values of the matrix or kXk? = i ?i (X).
The basic estimation problem, without any privacy considerations, is then given by
1 X
?
min
(Yi,j ? Xi,j )2 + ?
kXk?
(2)
m?n
N
mn
X?R
(i,j)??
where ?
? is a regularization parameter and N = |?| is the total number of observed ratings. The
factor mn ensures that the regularization does not grow with either dimension.
The above formulation requires the server to explicitly
? obtain predictions for each user, i.e., solve
for X. We can instead write X = U V T and ? = (1/ mn)V V T , and solve for ? only. If the server
then communicates the resulting low rank ? (or just V ) to each user, the users can reconstruct the
relevant part of U locally, and reproduce X as it pertains to them. To this end, let ?i = {j : (i, j) ?
?} be the set of observed entries for user i, and let Yi,?i be a column vector of user i?s ratings. Then
we can show that Eq.(2) is equivalent to solving
min+
??S
n
X
T
Yi,?
(?0 I + ??i ,?i )Yi,?i +
i
?
nm k?k?
(3)
i=1
?
?
where S + is the set of positive semi-definite m ? m matrices and ?0 = ?N/ nm. By solving ?,
c
we can predict ratings for unobserved items (index set ?i for user i) by
? i,?c = ??c ,? (?0 I + ?? ,? )?1 Yi,?
X
i
i
i
i
i
i
(4)
Note that we have yet to address any privacy concerns. The solution to Eq.(3) still requires access
to full ratings Yi,?i for each user.
Recommender setup with privacy Our privacy setup assumes an untrusted server. Any user
interested in guarding their data must therefore keep and process their data locally, releasing information to the server only in a controlled manner. We will initially divide users into two broad
2
categories, public and private. Public users are willing to share all their data with the server while
private users are unwilling to share any. This strict division is removed later when we permit private
users to release, in a controlled manner, limited information pertaining to their ratings (2nd order
information) so as to improve recommendations.
Any data made available to the server enables the server to model the collective experience of users,
for example, to solve Eq.(3). We will initially consider the setting where Eq.(3) is solved on the
basis of public users only. We use an EM type algorithm for training. In the E-step, the current ?
is sent to public users to complete their rating vectors and send back to the server. In the M-step,
? (or V? ) can be subsequently
? is then updated based on these full rating vectors. The resulting ?
shared with the private users, enabling the private users (their devices) to locally rank candidate
? is then improved by asking
items without any release of private information. The estimation of ?
private users to share 2nd order relational information about their ratings without any release of
marginal selections/ratings.
Note that we do not consider privacy beyond ratings. In other words, we omit any subsequent release
of information due to users exploring items recommended to them.
Summary of contributions We outline here our major contributions towards characterizing the
role of public users and the additional controlled release of information from private users.
p
?T X/
? ?nm can be estimated in a consistent, accurate manner on the basis
1) We show that ?
?= X
? ??
of public users alone. In particular, we express the error k?
?kF as a function of the total number
of observations. Moreover, if the underlying public user ratings can be thought of as i.i.d. samples,
we also bound k?
? ? ?? kF in terms of the number of public users. Here ?? is the true limiting
estimate. See section 3.1 for details.
? i,?c for private users relates to the accuracy of
2) We show how the accuracy of predicted ratings X
i
? (primarily from public users). Since the ratings for user i may not be related to the
estimating ?
? lies in, we can only characterize the accuracy when sufficient overlap exists. We
subspace that ?
? i,?c ? X
?i,?c k depends on this overlap, accuracy of ?,
? and
quantify this overlap, and show how kX
i
i
the observation noise. See section 3.2 for details.
3) Having established the accuracy of predictions based on public users alone, we go one step further
and introduce a new privacy mechanism and algorithms for gleaning additional relational (2nd order)
information from private users. This 2nd order information is readily used by the server to estimate
? The privacy concept constructively maintains first order (marginal) deniability for private users.
?.
We demonstrate empirically the gains from the additional 2nd order information. See section 4.
3
3.1
Analysis
?
Statistical Consistency of ?
?T U
? be a solution to Eq.(2). We can write X
? = U
? V? T , where U
? = I?m with 0/1 diagonal.
Let X
p
1
T
?
?
?
?
? To this end,
Since ? = ?
X X we can first analyze errors in X and then relate them to ?.
mn
we follow the restricted strong convexity (RSC) analysis[12]. However, their result depends on
the inverse of the minimum number of ratings of all users and items. In practice (see below), the
number of ratings decays exponentially across sorted users, making such a result loose. We provide
a modified analysis that depends only on the total number of observations N .
?i,? belongs to a fixed r dimensional
Throughout the analysis, we assume that each row vector X
?i,j | ?
subspace. We also assume that both noiseless and noisy entries are bounded,
i.e. |Yi,j |, |X
P
2
?, ?(i, j). For brevity, we use kY ? Xk? to denote the empirical loss (i,j)?? (Yi,j ? Xi,j )2 . The
restricted strong convexity property (RSC) assumes that there exists a constant ? > 0 such that
? ?
? 2F ? 1 kX
? ? Xk
? 2?
kX ? Xk
mn
N
3
(5)
? ?X
? in a certain subset. RSC provides the step from approximating observations to apfor X
proximating the full underlying matrix. It is satisfied with high probability provided that N =
(m + n) log(m + n)).
?=P
?S Q
?T , and let row(X) and col(X) denote the row and column spaces of
Assume the SVD of X
X. We define the following two sets,
A(P, Q)
B(P, Q)
?, col(X) ? Q}
?
:= {X, row(X) ? P
?? , col(X) ? Q
?? }
:= {X, row(X) ? P
(6)
Let ?A (X) and ?B (X) be the projection of X onto sets A and B, respectively, and ?A = I ? ?A ,
? ?X
? be the difference between the estimated and the underlying rating
?B = I ? ?B . Let ? = X
matrices. Our first lemma demonstrates that ? lies primarily in a restricted subspace and the second
one guarantees that the noise remains bounded.
Lemma 3.1. Assume i,j for (i, j) ? ? are i.i.d. sub-gaussian with ? = ki,j k?1 . Then with
2 2?
e
probability 1 ? N 4ch
, k?B (?)k? ? k?B (?)k? + 2c ?N ? mn log2 N . Here h > 0 is an absolute
constant associated with the sub-gaussian noise.
pn
2 2?
2
mn log N
N p mn
? N , then c ?
If ? = ?0 c? log
= c? log
N?
?0
N = b log N
N where we leave the deN
pendence on n explicit. Let D(b, n, N ) denote the set of difference matrices that satisfy lemma 3.1
above. By combining the lemma and the RSC property, we obtain the following theorem.
Theorem 3.2. Assume RSC for the set D(b, n, N ) with parameter ? > 0 where b =
? N,
?0 c? log
N
then we have
?=
where h, c > 0 are constants.
? 1 k?kF
mn
?
2c?( ?1?
?
+
2r log
?N
? ) N
?
c? m
?0 .
Let
e
with probability at least 1? N 4ch
The bound in the theorem consists of two terms, pertaining to the noise and the regularization. In
contrast to [12], the terms only relate to the total number of observations N .
? First, we map the accuracy of X
? to that of ?
? using a
We now turn our focus on the accuracy of ?.
perturbation bound for polar decomposition (see [9]).
?
? ? Xk
? F ? ?, then k?
? ??
Lemma 3.3. If ? 1 kX
?kF ? 2?
mn
? As a
This completes our analysis in terms of recovering ?
? for a fixed size underlying matrix X.
final step, we turn to the question of how the estimation error changes as the number of users or n
?T ?
?i be the underlying rating vector for user i and define ?n = 1 Pn X
grows. Let X
i=1 i Xi . Then
mn
1
1
n
?
n
?
?
?
? = (? ) 2 . If ? is the limit of ? , then ? = (? ) 2 . We bound the distance between ?
? and ?? .
?i are i.i.d samples from a distribution with support only in a subspace
Theorem 3.4. Assume X
?i k ? ??m. Let ?1 and ?r be the smallest and largest
of dimension r and bounded norm kX
?
eigenvalues of ? . Then, for large enough n, with probability at least 1 ? nr2 ,
s
?
?r log n
log n
?
k?
? ? ? kF ? 2 r?
+ o(
)
(7)
?1 n
n
Combining the two theorems and using triangle inequality, we obtain a high probability bound on
? ? ?? kF . Assuming the number of ratings for each user is larger than ?m, then N > ?nm and
k?
?
the bound grows in the rate of ?(log n/ n) with ? being a constant that depends on ?. For large
enough ?, the required n to achieve a certain error bound is small. Therefore a few public users with
large number of ratings could be enough to obtain a good estimate of ?? .
3.2
Prediction accuracy
? i,?c for all users as defined in
We are finally ready to characterize the error in the predicted ratings X
i
? ? ?? k obtained on the basis of our results
Eq.(4). For brevity, we use ? to represent the bound on k?
above. We also use x? and x?c as shorthands for Xi,?i and Xi,?ci with the idea that x? typically
refers to a new private user.
4
The key issue for us here is that the partial rating vector x? may be of limited use. For example,
if the number of observed ratings is less than rank r, then we would be unable to identify a rating
vector in the r dimensional subspace even without noise. We seek to control this in our analysis by
assuming that the observations have enough signal to be useful. Let SVD of ?? be Q? S ? (Q? )T ,
and ?1 be its minimum eigenvalue. We constrain the index set of observations ? such that it belongs
to the set
m
2
2
? T
D(?) = ? ? {1, . . . , m}, s.t.kxkF ? ? kx? kF , ?x ? row((Q ) )
|?|
The parameter ? depends on how the low dimensional sub-space is aligned with the coordinate axes.
We are only interested in characterizing prediction errors for observations that are in D(?). This is
quite different from the RSC property. Our main result is then
Theorem 3.5. Suppose k? ? ?? kF ? ? ?1 , ? ? D(?). For any ?
x ? row((Q? )T ), our
observation x? = ?
x? + ? where ? ? Sub(? 2 ) is the noise vector. The predicted ratings over
the remaining entries p
are given by x
??c = ??c ,? (?0 I + ??,? )?1 x? . Then, with probability at least
1 ? exp(?c2 min(c41 , |?|c21 )),
r
1
?
m
k?
xk F
2c1 ?|?| 4
0
c
c
?
?
kx? ? ?
x? kF ? 2 ? + ?( ?
+ 1)(
+
)
|?|
?1
?0
where c1 , c2 > 0 are constants.
?
All the proofs are provided in the supplementary material. The term proportional
to k?
xkF / ?1 is
?
1
due to the estimation error of ?? , while the term proportional to 2c1 ?|?| 4 / ?0 comes from the
noise in the observed ratings.
4
Controlled privacy for private users
Our theoretical results already demonstrate that a relatively small number of public users with many
ratings suffices for a reasonable performance guarantee for both public and private users. Empirical
results (next section) support this claim. However, since public users enjoy no privacy guarantees,
we would like to limit the required number of such users by requesting private users to contribute in
a limited manner while maintaining specific notions of privacy.
Definition 4.1. : Privacy preserving mechanism. Let M : Rm?1 ? Rm?1 be a random mechanism that takes a rating vector r as input and outputs M(r) of the same dimension with j th element
M(r)j . We say that M(r) is element-wise privacy preserving if Pr(M(r)j = z) = p(z) for
j = 1, ..., m, and some fixed distribution p.
For example, a privacy preserving mechanism M(r) is element-wise private if its coordinates follow the same marginal distribution such as uniform. Note that such a mechanism can still release
information about how different ratings interact (co-vary) which is necessary for estimation.
Discrete values. Assume that each element in r and M(r) belongs to a discrete set S with |S| = K.
A natural privacy constraint is to insist that the marginal distribution of M(r)j is uniform, i.e.,
Pr(M(r)j = z) = 1/K, for z ? S. A trivial mechanism that satisfies the privacy constraint is to
select each value uniformly at random from S. In this case, the returned rating vector contributes
nothing to the server model. Our goal is to design a mechanism that preserves useful 2nd order
information.
We assume that a small number of public user profiles are available, from which we can learn
an initial model parameterized by (?, V ), where ? is the item mean vector and V is a low rank
component of ?. The server provides each private user the pair (?, V ) and asks, once, for a response
M(r). Here r is the user?s full rating vector, completed (privately) with the help of the server model
(?, V ).
The mechanism M(r) is assumed to be element-wise privacy preserving, thus releasing nothing
about a single element in isolation. What information should it carry? If each user i provided their
Pn
1
1
full rating vector ri , the server could estimate ? according to ?nm
( i=1 (ri ??)(ri ??)T ) 2 . Thus,
5
if M(r) preserves the second order statistics to the extent possible, the server could still obtain an
accurate estimate of ?.
Consider a particular user and their completed rating vector r. Let P(x) = Pr(M(r) = x). We
select distribution P(x) by solving the following optimization problem geared towards preserving
interactions between the ratings under the uniform marginal constraint.
min
P
Ex?P k(x ? ?)(x ? ?)T ? (r ? ?)(r ? ?)T k2F
s.t.
P(xi = s) = 1/K, ?i, ?s ? S.
(8)
where K = |S|. The exact solution is difficult to obtain because the number of distinct assignments
of x is K m . Instead, we consider an approximate solution. Let x1 , ..., xK ? Rm?1 be K different
vectors such that, for each i, {x1i , x2i , ..., xK
i } forms a permutation of S. If we choose x with
Pr(x = xj ) = 1/K, then the marginal distribution of each element is uniform therefore maintaining
element-wise privacy. It remains to optimize the set x1 , ..., xK to capture interactions between
ratings.
We use a greedy coordinate descent algorithm to optimize x1 , ..., xK . For each coordinate i, we
randomly select a pair xp and xq , and switch xpi and xqi if the objective function in (8) is reduced.
The process is repeated a few times before we move on to the next coordinate. The algorithm can
be implemented efficiently because each operation deals only with a single coordinate.
Finally, according to the mechanism, the private user selects one of xj , j = 1, . . . , K, uniformly
at random and sends the discrete vector back to the server. Since the resulting rating vectors from
private users are noisy, the server decreases their weight relative to the information from public users
as part of the overall M-step for estimating ?.
Continuous values. Assuming the rating values are continuous and unbounded, we require instead
that the returned rating vectors follow the marginal distributions with a given mean and variance.
Specifically, E[M(r)i ] = 0 and Var[M(r)i ] = v where v is a constant that remains to be determined. Note that, again, any specific element of the returned vector will not, in isolation, carry any
information specific to the element.
As before, we search for the distribution P so as to minimize the L2 error of the second order
statistics under marginal constraints. For simplicity, denote r0 = r ? ? where r is the true completed
rating vector, and ui = M(r)i . The objective is given by
min
P,v
Eu?P kuuT ? r0 r0T k2F
s.t.
E[ui ] = 0, Var[ui ] = v, ?i.
(9)
Note that the formulation does not directly constrain that P has identical marginals, only that the
means and variances agree. However, the optimal solution does, as shown next.
Pm
Theorem 4.2. Let zi = sign(ri0 ) and h = ( i=1 |ri0 |)/m. The minimizing distribution of (9) is
given by Pr(u = zh) = Pr(u = ?zh) = 1/2.
We leave the proof in the supplementary material. A few remarks are in order. The mechanism in this
case is a two component mixture distribution, placing a probability mass on sign(r0 )h (vectorized)
and ?sign(r0 )h with equal probability. As a result, the server, knowing the algorithm that private
users follow, can reconstruct sign(r0 ) up to an overall randomly chosen sign. Note also that the
value of h is computed from user?s private rating vector and therefore releases some additional
information about r0 = r ? ? albeit weakly. To remove this information altogether, we could use
the same h for all users and estimate it based on public users.
The privacy constraints will clearly have a negative impact on the prediction accuracy in comparison
to having direct access to all the ratings. However, the goal is to improve accuracy beyond the public
users alone by obtaining limited information from private users. While improvements are possible,
the limited information surfaces in several ways. First, since private users provide no first order
information, the estimation of mean rating values cannot be improved beyond public users. Second,
the sampling method we use to enforce element-wise privacy adds noise to the aggregate second
order information from which V is constructed. Finally, the server can run the M-step with respect to
the private users? information only once, whereas the original EM algorithm could entertain different
completions for user ratings iteratively. Nevertheless, as illustrated in the next section, the algorithm
can still achieve a good accuracy, improving with each additional private user.
6
5
Experiments
We perform experiments on the Movielens 10M dataset which contains 10 million ratings from
69878 users on 10677 movies. The test set contains 10 ratings for each user. We begin by demonstrating that indeed a few public users suffice for making accurate recommendations. Figure 1 left
shows the test performance of both weighted (see [12]) and unweighted (uniform) trace norm regularization as we add users. Here users with most ratings are added first.
0.96
1.5
Uniform
Weighted
0.95
Most ratings
Random
1.4
0.94
1.3
0.92
Test RMSE
Test RMSE
0.93
0.91
0.9
1.2
1.1
0.89
1
0.88
0.9
0.87
0.86
0
0.2
0.4
0.6
Percentage of Users
0.8
1
0.8
200
400
600
800 1000 1200 1400 1600 1800 2000
Number of ratings (k)
Figure 1: Left: Test RMSE as a function of the percentage of public users; Right: Performance
changes with different rating numbers.
With only 1% of public users added, the test RMSE of unweighted trace norm regularization is
0.876 which is already close to the optimal prediction error. Note that the loss of weighted trace
norm regularization actually starts to go up when the number of users increases. The reason is that
the weighted trace norm penalizes less for users with few ratings. As a result, the resulting low
dimensional subspace used for prediction is influenced more by users with few ratings.
The statistical convergence bound in section 3.1 involves both terms that decrease as a function of
the number of ratings N and the number of public users n. The two factors are usually coupled. It
is interesting to see how they impact performance individually. Given a number of total ratings, we
compare two different methods of selecting public users. In the first method, users with most ratings
are selected first, whereas the second method selects users uniformly at random. As a result, if we
equalize the total number of ratings from each method, the second method selects a lot more users.
Figure 1 Right shows that the second method achieves better performance. An interpretation, based
on the theory, is that the right side of error bound (7) decreases as the number of users increases.
We also show how performance improves based on controlled access to private user preferences.
First, we take the top 100 users with the most ratings as the public users, and learn the initial
prediction model from their ratings. To highlight possible performance gains, private users with
more ratings are selected first. The results remain close if we select private users uniformly.
The rating values are from 0.5 to 5 with totally 10 different discrete values. Following the privacy
mechanism for discrete values, each private user generates ten different candidate vectors and returns
one of them uniformly at random. In the M-step, the weight for each private user is set to 1/2
compared to 1 for public users. During training, after processing w = 20 private users, we update
parameters (?, V ), re-complete the rating vectors of public users, making predictions for next batch
of private users more accurate. The privacy mechanism for continuous values is also tested under
the same setup. We denote the two privacy mechanism as PMD and PMC, respectively.
We compare five different scenarios. First, we use a standard DP method that adds Laplace noise to
the completed rating vector. Let the DP parameter be , because the maximum difference between
rating values is 4.5, the noise follows Lap(0, 4.5/). As before, we give a smaller weight to the
noisy rating vectors and this is determined by cross validation. Second, [5] proposed a notion of
?local privacy? in which differential privacy is guaranteed for each user separately. An optimal
strategy for d-dimensional multinomial distribution in this case reduces effective samples from n to
n2 /d where is the DP parameter. In our case the dimension corresponds to the number of items
7
0.92
0.915
0.91
Test RMSE
0.905
0.9
0.895
0.89
0.885
0.88
0.875
0.87
0
PMC
PMD
Lap eps=1
Lap eps=5
SSLP eps=5
Exact 2nd order
Full EM
50
100
150
200
250
Number of ??private?? users
300
350
400
Figure 2: Test RMSE as a function of private user numbers. PMC: the privacy mechanism for
continuous values; PMD: the privacy mechanism for discrete values; Lap eps=1: DP with Laplace
noise, = 1; Lap eps=5: same as before except = 5; SSLP eps=5: sampling strategy described
in [4] with DP parameter = 5; Exact 2nd order: with exact second order statistics from private
users (not a valid privacy mechanism); Full EM: EM without any privacy protection.
making estimation challenging under DP constraints. We also compare to this method and denote it
as SSLP (sampling strategy for local privacy).
In addition, to understand how our approximation to second order statistics affects the performance,
we also compare to the case that r0 a is given to the server directly where a = {?1, 1} with probability 1/2. In this way, the server can obtain the exact second order statistics using r0 r0T . Note that
this is not a valid privacy preserving mechanism. Finally, we compare to the case that the algorithm
can access private user rating vectors multiple times and update the parameters iteratively. Again,
this is not a valid mechanism but illustrates how much could be gained.
Figure 2 shows the performance as a function of the number of private users. The standard Laplace
noise method performs reasonably well when = 5, however the corresponding privacy guarantee
is very weak. SSLP improves the accuracy mildly.
In contrast, with the privacy mechanism we defined in section 4 the test RMSE decreases significantly as more private users are added. If we use the exact second order information without the
sampling method, the final test RMSE would be reduced by 0.07 compared to PMD. Lastly, full
EM without privacy protection reduces the test RMSE by another 0.08. These performance gaps are
expected because there is an inherent trade-off between accuracy and privacy.
6
Conclusion
Our contributions in this paper are three-fold. First, we provide explicit guarantees for estimating
item features in matrix completion problems. Second, we show how the resulting estimates, if shared
with new users, can be used to predict their ratings depending on the degree of overlap between
their private ratings and the relevant item subspace. The empirical results demonstrate that only a
small number of public users with large number of ratings suffices for a good performance. Third,
we introduce a new privacy mechanism for releasing 2nd order information needed for estimating
item features while maintaining 1st order deniability. The experiments show that this mechanism
indeed performs well in comparison to other mechanisms. We believe that allowing different levels
of privacy is an exciting research topic. An extension of our work would be applying the privacy
mechanism to the learning of graphical models in which 2nd or higher order information plays an
important role.
7
Acknowledgement
The work was partially supported by Google Research Award and funding from Qualcomm Inc.
8
References
[1] M?ario S Alvim, Miguel E Andr?es, Konstantinos Chatzikokolakis, Pierpaolo Degano, and
Catuscia Palamidessi. Differential privacy: on the trade-off between utility and information
leakage. In Formal Aspects of Security and Trust, pages 39?54. Springer, 2012.
[2] E. Candes and Y. Plan. Matrix completion with noise. In Proceedings of the IEEE, 2010.
[3] J. Canny. Collaborative filtering with privacy via factor analysis. In SIGIR, 2002.
[4] John Duchi, Martin J Wainwright, and Michael Jordan. Local privacy and minimax bounds:
Sharp rates for probability estimation. In Advances in Neural Information Processing Systems,
pages 1529?1537, 2013.
[5] John C Duchi, Michael I Jordan, and Martin J Wainwright. Privacy aware learning. In NIPS,
pages 1439?1447, 2012.
[6] C. Dwork. Differential privacy: A survey of results. In Theory and Applications of Models of
Computation, 2008.
[7] M. Jaggi and M. Sulovsk. A simple algorithm for nuclear norm regularized problems. In
ICML, 2010.
[8] R. Keshavan, A. Montanari, and Sewoong Oh. Matrix completion from noisy entries. JMLR,
2010.
[9] R. Mathias. Perturbation bounds for the polar decomposition. BIT Numerical Mathematics,
1997.
[10] F. McSherry and I. Mironov. Differentially private recommender systems: Building privacy
into the netflix prize contenders. In SIGKDD, 2009.
[11] B. N. Miller, J. A. Konstan, and J. Riedl. Pocketlens: Toward a personal recommender system.
ACM Trans. Inf. Syst., 2004.
[12] S. Negahban and M. J. Wainwright. Restricted strong convexity and weighted matrix completion: optimal bounds with noise. JMLR, 2012.
[13] R. Salakhutdinov and N. Srebro. Collaborative filtering in a non-uniform world: Learning with
the weighted trace norm. In NIPS, 2010.
[14] N. Srebro, J. Rennie, and T. Jaakkola. Maximum margin matrix factorization. In NIPS, 2004.
[15] J. A. Tropp. User-friendly tail bounds for sums of random matrices. Found. Comput. Math.,
2012.
[16] R. Vershynin.
arXiv:1011.3027.
Introduction to the non-asymptotic analysis of random matrices.
[17] Y. Xin and T. Jaakkola. Primal-dual methods for sparse constrained matrix completion. In
AISTATS, 2012.
9
| 5359 |@word private:48 norm:10 nd:12 willing:4 seek:1 guarding:1 decomposition:2 asks:1 boundedness:1 carry:2 initial:2 contains:2 selecting:1 recovered:1 current:1 protection:2 yet:1 must:2 readily:1 john:2 realistic:1 subsequent:1 numerical:1 enables:1 remove:1 update:2 alone:3 greedy:1 selected:2 device:1 item:16 xk:9 prize:1 core:1 yuxin:1 provides:2 math:1 contribute:1 preference:7 five:1 unbounded:1 guard:1 c2:2 direct:1 differential:6 constructed:1 consists:1 shorthand:1 privacy:57 manner:5 introduce:3 theoretically:2 expected:1 indeed:2 salakhutdinov:1 insist:1 totally:1 provided:3 estimating:4 moreover:3 underlying:7 suffice:2 bounded:3 mass:1 begin:1 what:1 unobserved:1 guarantee:9 friendly:1 demonstrates:1 rm:3 control:1 omit:1 enjoy:1 positive:1 before:4 understood:1 local:3 limit:2 challenging:2 co:1 limited:6 factorization:1 c21:1 practice:1 definite:1 empirical:4 thought:1 significantly:1 projection:1 word:1 refers:1 onto:1 cannot:1 selection:1 close:2 context:1 applying:1 optimize:2 equivalent:1 map:1 missing:1 send:1 go:2 convex:3 sigir:1 survey:1 unwilling:1 simplicity:1 mironov:1 estimator:1 nuclear:1 oh:1 notion:4 coordinate:6 laplace:3 updated:1 limiting:1 controlling:1 suppose:1 play:1 user:131 exact:6 us:1 element:11 sparsely:1 database:1 observed:7 role:2 solved:3 capture:1 ensures:1 eu:1 trade:5 removed:1 decrease:4 equalize:1 convexity:4 ui:3 personal:1 weakly:1 solving:4 division:1 basis:3 untrusted:2 triangle:1 distinct:1 effective:1 pertaining:2 query:1 aggregate:2 peer:2 quite:1 larger:2 solve:3 supplementary:2 say:1 rennie:1 reconstruct:2 statistic:5 qualcomm:1 noisy:4 final:2 eigenvalue:2 propose:1 interaction:2 canny:1 relevant:2 combining:2 aligned:1 achieve:2 ky:1 differentially:1 convergence:1 leave:2 encryption:1 help:1 illustrate:1 depending:1 completion:9 miguel:1 eq:6 strong:4 recovering:1 predicted:3 implemented:1 come:1 involves:1 quantify:1 tommi:2 subsequently:2 deniability:4 public:32 material:2 require:3 suffices:3 generalization:1 exploring:1 extension:1 hold:1 considered:1 exp:1 proximating:1 predict:3 claim:1 major:1 sought:1 adopt:1 smallest:1 vary:1 achieves:1 estimation:9 polar:2 individually:1 largest:1 trusted:2 weighted:6 mit:4 gleaning:2 clearly:1 gaussian:3 aim:1 modified:1 pn:3 jaakkola:3 release:9 focus:1 ax:1 improvement:1 rank:10 contrast:3 sigkdd:1 typically:2 transferring:1 initially:2 reproduce:1 interested:2 selects:3 issue:1 overall:2 dual:1 plan:1 constrained:1 marginal:9 equal:1 once:2 aware:1 having:2 sampling:4 identical:1 placing:1 broad:1 yu:1 k2f:2 icml:1 inherent:2 few:7 primarily:2 randomly:2 preserve:2 individual:2 dwork:1 mixture:1 primal:1 mcsherry:1 pendence:1 accurate:5 partial:1 necessary:1 experience:2 entertain:1 r0t:2 divide:2 penalizes:1 re:1 theoretical:1 rsc:6 column:3 asking:1 kxkf:1 assignment:1 nr2:1 entry:8 subset:1 uniform:7 characterize:3 sulovsk:1 contender:1 vershynin:1 st:1 negahban:1 xpi:1 csail:3 off:5 pool:1 michael:2 again:2 nm:5 satisfied:1 opposed:1 homomorphic:1 choose:1 return:1 syst:1 inc:1 satisfy:1 explicitly:1 ranking:1 depends:8 performed:1 view:1 later:1 lot:1 analyze:1 start:1 aggregation:1 maintains:1 netflix:1 candes:1 rmse:9 contribution:3 minimize:1 collaborative:2 accuracy:18 variance:2 who:4 efficiently:1 miller:1 correspond:1 identify:1 weak:1 churn:1 published:1 influenced:1 definition:1 involved:1 associated:1 proof:2 gain:3 dataset:1 improves:2 pmd:4 actually:1 back:3 higher:1 follow:4 response:1 improved:3 formulation:3 done:1 zeromean:1 just:1 lastly:1 tropp:1 trust:1 keshavan:1 google:1 grows:2 believe:1 building:3 concept:4 true:2 regularization:9 entering:1 iteratively:3 illustrated:1 deal:1 leaked:1 during:1 xqi:1 outline:1 complete:3 demonstrate:6 performs:2 duchi:2 wise:5 consideration:1 funding:1 multinomial:1 empirically:3 exponentially:1 million:1 tail:1 interpretation:1 marginals:1 eps:6 significant:1 refer:1 consistency:2 populated:1 pm:1 tracenorm:1 mathematics:1 access:6 geared:1 surface:1 add:3 jaggi:1 own:1 perspective:1 moderate:2 belongs:3 inf:1 scenario:1 certain:2 server:20 inequality:1 yi:9 preserving:6 minimum:2 additional:5 r0:8 converge:1 recommended:1 signal:1 semi:1 recoverable:1 full:8 relates:1 multiple:1 reduces:2 cross:1 award:1 controlled:6 impact:2 prediction:9 basic:1 noiseless:1 arxiv:1 represent:1 achieved:1 c1:3 whereas:2 addition:1 separately:1 completes:1 singular:1 leaving:1 grow:1 sends:1 operate:1 releasing:3 strict:1 sent:1 jordan:2 enough:4 switch:1 affect:2 isolation:2 xj:2 zi:1 idea:1 knowing:1 requesting:1 konstantinos:1 utility:1 returned:3 remark:1 useful:2 involve:1 amount:1 locally:5 ten:1 category:1 reduced:2 percentage:2 andr:1 sign:5 estimated:3 write:2 discrete:6 express:1 ario:1 key:1 nevertheless:1 demonstrating:1 relaxation:1 sum:2 run:1 inverse:1 parameterized:1 throughout:1 reasonable:3 bit:1 bound:14 guaranteed:1 c41:1 fold:1 constraint:6 constrain:2 ri:3 generates:1 aspect:1 min:5 relatively:1 martin:2 ri0:2 according:2 riedl:1 across:1 remain:1 em:6 smaller:1 making:4 den:1 openly:2 restricted:5 pr:6 computationally:1 agree:2 remains:3 turn:2 loose:1 mechanism:23 apfor:1 needed:1 end:2 available:4 operation:1 permit:1 enforce:1 batch:1 altogether:1 xkf:1 original:1 denotes:1 remaining:2 assumes:2 top:1 completed:4 graphical:1 log2:1 maintaining:5 exploit:1 approximating:1 leakage:1 objective:2 move:1 already:3 question:2 depart:1 added:3 strategy:3 diagonal:1 surrogate:1 behalf:1 dp:7 subspace:7 distance:1 separate:1 unable:1 entity:1 topic:1 extent:2 trivial:1 reason:1 toward:1 assuming:3 index:2 tiered:1 reformulate:1 minimizing:1 pmc:3 setup:4 mostly:1 difficult:1 relate:2 trace:8 negative:1 constructively:1 design:1 collective:2 perform:2 allowing:1 recommender:8 observation:9 enabling:1 descent:1 relational:3 rn:1 perturbation:2 sharp:1 rating:68 introduced:1 pair:2 required:2 security:1 protect:1 established:1 nip:3 trans:1 address:2 beyond:4 below:1 usually:1 including:1 wainwright:3 overlap:4 natural:1 regularized:1 mn:11 minimax:1 improve:2 movie:1 x2i:1 ready:1 coupled:1 xq:1 l2:1 acknowledgement:1 kf:9 zh:2 relative:1 asymptotic:1 loss:3 permutation:1 highlight:1 interesting:1 proportional:2 filtering:2 srebro:2 var:2 validation:1 degree:1 sufficient:1 consistent:1 xp:1 vectorized:1 exciting:1 sewoong:1 share:6 row:8 summary:2 supported:1 side:2 bias:1 understand:1 formal:1 characterizing:2 absolute:1 sparse:1 fragmented:1 dimension:5 valid:3 world:1 unweighted:2 commonly:1 made:1 party:2 approximate:1 keep:1 assumed:3 xi:6 continuous:4 latent:1 search:1 learn:2 reasonably:1 obtaining:1 contributes:1 improving:1 interact:1 aistats:1 main:1 montanari:1 privately:2 noise:19 profile:1 nothing:2 repeated:1 x1:3 sub:6 obfuscation:1 wish:1 explicit:3 konstan:1 col:3 candidate:2 lie:2 x1i:1 comput:1 communicates:1 third:1 jmlr:2 theorem:7 specific:3 decay:1 concern:1 exists:2 albeit:1 adding:1 gained:1 ci:1 illustrates:1 kx:7 margin:1 mildly:1 gap:1 lap:5 explore:1 kxk:2 partially:1 recommendation:5 springer:1 ch:2 corresponds:1 satisfies:1 acm:1 goal:3 sorted:1 towards:2 shared:2 considerable:1 change:2 specifically:2 determined:2 uniformly:5 movielens:1 except:1 lemma:5 total:6 mathias:1 svd:2 xin:2 e:1 select:4 support:2 pertains:1 brevity:2 tested:1 ex:1 |
4,815 | 536 | Tangent Prop - A formalism for specifying
selected invariances in an adaptive network
Patrice Simard
AT&T Bell Laboratories
101 Crawford Corner Rd
Holmdel, NJ 07733
Yann Le Cun
AT&T Bell Laboratories
101 Crawford Corner Rd
Holmdel, NJ 07733
Bernard Victorri
Universite de Caen
Caen 14032 Cedex
France
John Denker
AT&T Bell Laboratories
101 Crawford Corner Rd
Holmdel, NJ 07733
Abstract
In many machine learning applications, one has access, not only to training
data, but also to some high-level a priori knowledge about the desired behavior of the system. For example, it is known in advance that the output
of a character recognizer should be invariant with respect to small spatial distortions of the input images (translations, rotations, scale changes,
etcetera).
We have implemented a scheme that allows a network to learn the derivative of its outputs with respect to distortion operators of our choosing.
This not only reduces the learning time and the amount of training data,
but also provides a powerful language for specifying what generalizations
we wish the network to perform.
1
INTRODUCTION
In machine learning, one very often knows more about the function to be learned
than just the training data. An interesting case is when certain directional derivatives of the desired function are known at certain points. For example, an image
895
896
Simard, Victorri, Le Cun, and Denker
Figure 1: Top: Small rotations of an original digital image of the digit "3" (center).
Middle: Representation of the effect of the rotation in the input vector space space
(assuming there are only 3 pixels). Bottom: Images obtained by moving along the
tangent to the transformation curve for the same original digital image (middle).
recognition system might need to be invariant with respect to small distortions of
the input image such as translations, rotations, scalings, etc.; a speech recognition
system n.ight need to be invariant to time distortions or pitch shifts. In other
words, the derivative of the system's output should be equal to zero when the input
is transformed in certain ways.
Given a large amount of training data and unlimited training time, the system
could learn these invariances from the data alone, but this is often infeasible. The
limitation on data can be overcome by training the system with additional data
obtained by distorting (translating, rotating, etc.) the original patterns (Baird,
1990). The top of Fig. 1 shows artificial data generated by rotating a digital image of
the digit "3" (with the original in the center). This procedure, called the "distortion
model" , has two drawbacks. First, the user must choose the magnitude of distortion
and how many instances should be generated. Second, and more importantly, the
distorted data is highly correlated with the original data. This makes traditional
learning algorithms such as back propagation very inefficient. The distorted data
carries only a very small incremental amount of information, since the distorted
patterns are not very different from the original ones. It may not be possible to
adjust the learning system so that learning the invariances proceeds at a reasonable
rate while learning the original points is non-divergent.
The key idea in this paper is that it is possible to directly learn the effect on
the output of distorting the input, independently from learning the undistorted
Tangent Prop-A formalism for specifying selected invariances in an adaptive network
F(x)
F(x)
x1
x2
x3
x4
x
x1
x2
x3
x4
x
Figure 2: Learning a given function (solid line) from a limited set of example (Xl
to X4). The fitted curves are shown in dotted line. Top: The only constraint is that
the fitted curve goes through the examples. Bottom: The fitted curves not only
goes through each examples but also its derivatives evaluated at the examples agree
with the derivatives of the given function.
patterns. When a pattern P is transformed (e.g. rotated) with a transformation
s that depends on one parameter a (e.g. the angle of the rotation), the set of all
the transformed patterns S(P) = {sea, P) Va} is a one dimensional curve in the
vector space of the inputs (see Fig. 1). In certain cases, such as rotations of digital
images, this curve must be made continuous using smoothing techniques, as will be
shown below. When the set of transformations is parameterized by n parameters
ai (rotation, translation, scaling, etc.), S(P) is a manifold of at most n dimensions.
The patterns in S(P) that are obtained through small transformations of P, i.e.
the part of S( P) that is close to P, can be approximated by a plane tangent to
the manifold S(P) at point P. Small transformations of P can be obtained by
adding to P a linear combination of vectors that span the tangent plane (tangent
vectors). The images at the bottom of Fig. 1 were obtained by that procedure.
More importantly, the tangent vectors can be used to specify high order constraints
on the function to be learned, as explained below.
To illustrate the method, consider the problem of learning a single-valued function
F from a limited set of examples. Fig. 2 (left) represents a simple case where the
desired function F (solid line) is to be approximated by a function G (dotted line)
from four examples {(Xi, F(Xi))}i=1,2,3,4. As exemplified in the picture, the fitted
function G largely disagrees with the desired function F between the examples. If
the functions F and G are assumed to be differentiable (which is generally the case),
the approximation G can be greatly improved by requiring that G's derivatives
evaluated at the points {xd are equal to the derivatives of F at the same points
(Fig. 2 right). This result can be extended to multidimensional inputs. In this case,
we can impose the equality of the derivatives of F and G in certain directions, not
necessarily in all directions of the input space.
Such constraints find immediate use in traditional learning problems. It is often the
case that a priori knowledge is available on how the desired function varies with
897
898
Simard, Victorri, Le Cun, and Denker
pattern P
rotated by ex
pattern P
tangent
vector
--
Figure 3: How to compute a tangent vector for a given transformation (in this case
a rotation).
respect to some transformations of the input. It is straightforward to derive the
corresponding constraint on the directional derivatives of the fitted function G in
the directions of the transformations (previously named tangent vectors). Typical
examples can be found in pattern recognition where the desired classification function is known to be invariant with respect to some transformation of the input such
as translation, rotation, scaling, etc., in other words, the directional derivatives of
the classification function in the directions of these transformations is zero.
2
IMPLEMENTATION
The implementation can be divided into two parts. The first part consists in computing the tangent vectors. This part is independent from the learning algorithm
used subsequently. The second part consists in modifying the learning algorithm
(for instance backprop) to incorporate the information about the tangent vectors.
Part I: Let x be an input pattern and s be a transformation operator acting
on the input space and depending on a parameter a. If s is a rotation operator
for instance, then s( a, x) denotes the input x rotated by the angle a. We will
require that the transformation operator s be differentiable with respect to a and
x, and that s(O, x) = x. The tangent vector is by definition 8s(a, x)/8a. It can be
approximated by a finite difference, as shown in Fig. 3. In the figure, the input space
is a 16 by 16 pixel image and the patterns are images of handwritten digits. The
transformations considered are rotations of the digit images. The tangent vector
is obtained in two steps. First the image is rotated by an infinitesimal amount a.
This is done by computing the rotated coordinates of each pixel and interpolating
the gray level values at the new coordinates. This operation can be advantageously
combined with some smoothing using a convolution. A convolution with a Gaussian
provides an efficient interpolation scheme in O(nm) multiply-adds, where nand m
are the (gaussian) kernel and image sizes respectively. The next step is to subtract
(pixel by pixel) the rotated image from the original image and to divide the result
Tangent Prop-A formalism for specifying selected invariances in an adaptive network
by the scalar 0 (see Fig. 3). If Ie types of transformations are considered, there
will be Ie different tangent vectors per pattern. For most algorithms, these do not
require any storage space since they can be generated as needed from the original
pattern at negligible cost.
Part IT: Tangent prop is an extension of the backpropagation algorithm, allowing
it to learn directional derivatives. Other algorithms such as radial basis functions
can be extended in a similar fashion.
To implement our idea, we will modify the usual weight-update rule:
~w =
oE
-7] ow
is replaced with
~w
=
0
-7] ow (E
+ J.tEr)
(1)
where 7] is the learning rate, E the usual objective function, Er an additional objective function (a regularizer) that measures the discrepancy between the actual and
desired directional derivatives in the directions of some selected transformations,
and J.t is a weighting coefficient.
=
Let x be an input pattern, y G(x) be the input-output function of the network.
The regularizer Er is of the form
Er(x)
:e
etrainingset
where Er(x) is
(2)
Here, Ki(x) is the desired directional derivative of G in the direction induced by
transformation Si applied to pattern x. The second term in the norm symbol is the
actual directional derivative, which can be rewritten as
=G'{x). OSi(O, x)
0=0
00
0=0
where G'(x) is the Jacobian of G for pattern x, and OSi(O, x)Joo is the tangent
vector associated to transformation Si as described in Part I. Multiplying the tangent
vector by the Jacobian involves one forward propagation through a "linearized"
version of the network. In the special case where local invariance with respect to
the Si'S is desired, Ki(x) is simply set to o.
Composition of transformations: The theory of Lie groups (Gilmore, 1974)
ensures that compositions of local (small) transformations Si correspond to linear
combinations of the corresponding tangent vectors (the local transformations Si
have a structure of Lie algebra). Consequently, if Er{x) = 0 is verified, the network
derivative in the direction of a linear combination of the tangent vectors is equal
to the same linear combination of the desired derivatives. In other words if the
network is successfully trained to be locally invariant with respect to, say, horizontal
translation and vertical translations, it will be invariant with respect to compositions
thereof.
We have derived and implemented an efficient algorithm, "tangent prop" , for performing the weight update (Eq. 1). It is analogous to ordinary backpropagation,
899
900
Simard, Victorri, Le Cun, and Denker
W l+ 1
W'+l
Iti
Iti
e: l
,
b'.-l
j3J-1
,
x?'-I
Network
e;-I
Jacobian nework
Figure 4: forward propagated variables (a, x, a, e), and backward propagated variables (b, y, p, t/J) in the regular network (roman symbols) and the Jacobian (linearized) network (greek symbols)
but in addition to propagating neuron activations, it also propagates the tangent
vectors. The equations can be easily derived from Fig. 4.
Forward propagation:
a~
?
=~
wL x '.-l
L...J I, ,
x~
= u(aD
(3)
i
Tangent forward propagation:
,_
ai -
~
,
~'-1
L...J wW"i
i
e! = u'(a~)a~
(4)
Tangent gradient backpropagation:
~ w'+1.I.l+1
(3i1 -- L...J
Iti ?lit
(5)
It
Gradient backpropagation:
~ w1+ 1yl+1
bi' -- L...J
Iti It
(6)
It
Weight update:
8[E(W, Up) + I'Er (W, Up, Tp)] _ 1-1 , + ~'-l.I.'
- Xi Yi I'\oi ?Ii
8w??'
I,
(7)
Tangent Prop--A formalism for specifying selected invariances in an adaptive network
60
50
%Erroron
the test set
20
10
160
320
Training set size
Figure 5: Generalization performance curve as a function of the training set size for
the tangent prop and the backprop algorithms
The regularization parameter jJ is tremendously important, because it determines
the tradeoff between minimizing the usual objective function and minimizing the
directional derivative error.
3
RESULTS
Two experiments illustrate the advantages of tangent prop. The first experiment
is a classification task, using a small (linearly separable) set of 480 binarized handwritten digit . The training sets consist of 10, 20, 40, 80, 160 or 320 patterns, and
the training set contains the remaining 160 patterns. The patterns are smoothed
using a gaussian kernel with standard deviation of one half pixel. For each of the
training set patterns, the tangent vectors for horizontal and vertical translation
are computed. The network has two hidden layers with locally connected shared
weights, and one output layer with 10 units (5194 connections, 1060 free parameters) (Le Cun, 1989). The generalization performance as a function of the training
set size for traditional backprop and tangent prop are compared in Fig. 5. We have
conducted additional experiments in which we implemented not only translations
but also rotations, expansions and hyperbolic deformations. This set of 6 generators is a basis for all linear transformations of coordinates for two dimensional
images. It is straightforward to implement other generators including gray-Ievelshifting, "smooth" segmentation, local continuous coordinate transformations and
independent image segment transformations.
The next experiment is designed to show that in applications where data is highly
901
902
Simard, Victorri, Le Cun, and Denker
A-. NMSE VI.
Av"ge NMSE VI 1ge
0.15
.15
0.1
.1
o
o
oL-~~==~~=;~==+=~~~
1000 2000 3000 4000 5000 6000 7000 8000 0000 10000
0
.....
1000 2000 3000 4000 5000 6000 7000 8000 0000 10000
.....
15
"
-
o
15
"
-
0
-0.5
-.5
-1
-1
+--_+_-_--+_-_+_-_-_
-1
-0.5
0
0.5
1.5
-1 .5
-1 .5
Distortion model
+--_+_-_--+--_+_-_-__t
.5
1.5
- .5
o
-1.5
-1.5
-1
Tangent prop
Figure 6: Comparison of the distortion model (left column) and tangent prop (right
column). The top row gives the learning curves (error versus number of sweeps
through the training set). The bottom row gives the final input-output function of
the network; the dashed line is the result for unadorned back prop.
Tangent Prop-A formalism for specifying selected invariances in an adaptive network
correlated, tangent prop yields a large speed advantage. Since the distortion model
implies adding lots of highly correlated data, the advantage of tangent prop over
the distortion model becomes clear.
The task is to approximate a function that has plateaus at three locations. We want
to enforce local invariance near each of the training points (Fig. 6, bottom). The
network has one input unit, 20 hidden units and one output unit. Two strategies are
possible: either generate a small set of training point covering each of the plateaus
(open squares on Fig. 6 bottom), or generate one training point for each plateau
(closed squares), and enforce local invariance around them (by setting the desired
derivative to 0). The training set of the former method is used as a measure the
performance for both methods. All parameters were adjusted for approximately
optimal performance in all cases. The learning curves for both models are shown in
Fig. 6 (top). Each sweep through the training set for tangent prop is a little faster
since it requires only 6 forward propagations, while it requires 9 in the distortion
model. As can be seen, stable performance is achieved after 1300 sweeps for the
tangent prop, versus 8000 for the distortion model. The overall speedup is therefore
about 10.
Tangent prop in this example can take advantage of a very large regularization term.
The distortion model is at a disadvantage because the only parameter that effectively controls the amount of regularization is the magnitude of the distortions, and
this cannot be increased to large values because the right answer is only invariant
under small distortions.
4
CONCLUSIONS
When a priori information about invariances exists, this information must be made
available to the adaptive system. There are several ways of doing this, including the
distortion model and tangent prop. The latter may be much more efficient in some
applications, and it permits separate control of the emphasis and learning rate for
the invariances, relative to the original training data points. Training a system to
have zero derivatives in some directions is a powerful tool to express invariances to
transformations of our choosing. Tests of this procedure on large-scale applications
(handwritten zipcode recognition) are in progress.
References
Baird, H. S. (1990). Document Image Defect Models. In IAPR 1990 Workshop on
Sytactic and Structural Pattern Recognition, pages 38-46, Murray Hill, NJ.
Gilmore, R. (1974). Lie Groups, Lie Algebras and some of their Applications. Wiley,
New York.
Le Cun, Y. (1989) . Generalization and Network Design Strategies. In Pfeifer, R.,
Schreter, Z., Fogelman, F., and Steels, L., editors, Connectionism in Perspective, Zurich, Switzerland. Elsevier. an extended version was published as a
technical report of the University of Toronto.
903
| 536 |@word version:2 middle:2 norm:1 open:1 linearized:2 solid:2 carry:1 contains:1 document:1 si:5 activation:1 must:3 john:1 designed:1 update:3 alone:1 half:1 selected:6 plane:2 provides:2 location:1 toronto:1 along:1 j3j:1 consists:2 behavior:1 ol:1 actual:2 little:1 becomes:1 what:1 nework:1 transformation:24 nj:4 multidimensional:1 binarized:1 xd:1 control:2 unit:4 negligible:1 local:6 modify:1 interpolation:1 approximately:1 might:1 emphasis:1 specifying:6 limited:2 bi:1 implement:2 x3:2 backpropagation:4 digit:5 procedure:3 bell:3 hyperbolic:1 word:3 radial:1 regular:1 cannot:1 close:1 operator:4 storage:1 center:2 go:2 straightforward:2 independently:1 rule:1 importantly:2 coordinate:4 analogous:1 user:1 recognition:5 approximated:3 bottom:6 ensures:1 connected:1 oe:1 trained:1 segment:1 algebra:2 iapr:1 basis:2 easily:1 caen:2 regularizer:2 artificial:1 choosing:2 valued:1 distortion:16 say:1 patrice:1 final:1 zipcode:1 advantage:4 differentiable:2 osi:2 sea:1 incremental:1 rotated:6 illustrate:2 derive:1 depending:1 propagating:1 undistorted:1 progress:1 eq:1 implemented:3 involves:1 implies:1 direction:8 greek:1 switzerland:1 drawback:1 modifying:1 subsequently:1 translating:1 etcetera:1 backprop:3 require:2 generalization:4 connectionism:1 adjusted:1 extension:1 around:1 considered:2 recognizer:1 wl:1 successfully:1 tool:1 gaussian:3 derived:2 greatly:1 tremendously:1 elsevier:1 nand:1 hidden:2 transformed:3 france:1 i1:1 pixel:6 overall:1 classification:3 fogelman:1 priori:3 spatial:1 smoothing:2 special:1 equal:3 x4:3 represents:1 lit:1 discrepancy:1 report:1 roman:1 replaced:1 highly:3 multiply:1 adjust:1 divide:1 rotating:2 desired:11 deformation:1 fitted:5 instance:3 formalism:5 column:2 increased:1 tp:1 disadvantage:1 ordinary:1 cost:1 deviation:1 conducted:1 answer:1 varies:1 combined:1 ie:2 yl:1 w1:1 nm:1 choose:1 corner:3 simard:5 derivative:19 inefficient:1 de:1 coefficient:1 baird:2 depends:1 ad:1 vi:2 lot:1 closed:1 doing:1 oi:1 square:2 largely:1 correspond:1 yield:1 directional:8 handwritten:3 multiplying:1 published:1 plateau:3 definition:1 infinitesimal:1 thereof:1 universite:1 associated:1 propagated:2 knowledge:2 segmentation:1 back:2 specify:1 improved:1 evaluated:2 done:1 just:1 horizontal:2 propagation:5 gray:2 effect:2 requiring:1 gilmore:2 former:1 regularization:3 equality:1 laboratory:3 covering:1 hill:1 image:19 rotation:12 composition:3 ai:2 rd:3 language:1 joo:1 moving:1 access:1 stable:1 etc:4 add:1 perspective:1 certain:5 yi:1 seen:1 additional:3 impose:1 ight:1 dashed:1 ii:1 reduces:1 smooth:1 technical:1 faster:1 divided:1 va:1 pitch:1 kernel:2 achieved:1 addition:1 want:1 victorri:5 cedex:1 induced:1 structural:1 near:1 ter:1 idea:2 tradeoff:1 shift:1 distorting:2 advantageously:1 speech:1 york:1 jj:1 generally:1 clear:1 amount:5 locally:2 generate:2 dotted:2 per:1 express:1 group:2 key:1 four:1 verified:1 backward:1 defect:1 angle:2 parameterized:1 powerful:2 distorted:3 named:1 reasonable:1 yann:1 unadorned:1 holmdel:3 scaling:3 ki:2 layer:2 constraint:4 x2:2 unlimited:1 schreter:1 speed:1 span:1 performing:1 separable:1 speedup:1 combination:4 character:1 cun:7 explained:1 invariant:7 equation:1 agree:1 previously:1 zurich:1 needed:1 know:1 ge:2 available:2 operation:1 rewritten:1 permit:1 denker:5 enforce:2 original:10 top:5 denotes:1 remaining:1 murray:1 sweep:3 objective:3 strategy:2 usual:3 traditional:3 ow:2 gradient:2 separate:1 manifold:2 assuming:1 minimizing:2 steel:1 implementation:2 design:1 perform:1 allowing:1 vertical:2 convolution:2 neuron:1 av:1 iti:4 finite:1 immediate:1 extended:3 ww:1 smoothed:1 connection:1 learned:2 proceeds:1 below:2 pattern:21 exemplified:1 including:2 scheme:2 picture:1 crawford:3 disagrees:1 tangent:39 relative:1 interesting:1 limitation:1 versus:2 generator:2 digital:4 propagates:1 editor:1 translation:8 row:2 free:1 infeasible:1 curve:9 overcome:1 dimension:1 forward:5 made:2 adaptive:6 approximate:1 assumed:1 xi:3 continuous:2 learn:4 correlated:3 expansion:1 necessarily:1 interpolating:1 linearly:1 x1:2 nmse:2 fig:12 fashion:1 wiley:1 wish:1 xl:1 lie:4 weighting:1 jacobian:4 pfeifer:1 er:6 symbol:3 divergent:1 consist:1 exists:1 workshop:1 adding:2 effectively:1 magnitude:2 subtract:1 simply:1 scalar:1 determines:1 prop:19 consequently:1 shared:1 change:1 typical:1 acting:1 called:1 bernard:1 invariance:13 latter:1 incorporate:1 ex:1 |
4,816 | 5,360 | Content-based recommendations
with Poisson factorization
Laurent Charlin
Department of Computer Science
Columbia University
New York, NY 10027
[email protected]
Prem Gopalan
Department of Computer Science
Princeton University
Princeton, NJ 08540
[email protected]
David M. Blei
Departments of Statistics & Computer Science
Columbia University
New York, NY 10027
[email protected]
Abstract
We develop collaborative topic Poisson factorization (CTPF), a generative model
of articles and reader preferences. CTPF can be used to build recommender systems by learning from reader histories and content to recommend personalized
articles of interest. In detail, CTPF models both reader behavior and article texts
with Poisson distributions, connecting the latent topics that represent the texts
with the latent preferences that represent the readers. This provides better recommendations than competing methods and gives an interpretable latent space for
understanding patterns of readership. Further, we exploit stochastic variational
inference to model massive real-world datasets. For example, we can fit CPTF
to the full arXiv usage dataset, which contains over 43 million ratings and 42
million word counts, within a day. We demonstrate empirically that our model
outperforms several baselines, including the previous state-of-the art approach.
1
Introduction
In this paper we develop a probabilistic model of articles and reader behavior data. Our model is
called collaborative topic Poisson factorization (CTPF). It identifies the latent topics that underlie the articles, represents readers in terms of their preferences for those topics, and captures how
documents about one topic might be interesting to the enthusiasts of another.
As a recommendation system, CTPF performs well in the face of massive, sparse, and long-tailed
data. Such data is typical because most readers read or rate only a few articles, while a few readers
may read thousands of articles. Further, CTPF provides a natural mechanism to solve the ?cold start?
problem, the problem of recommending previously unread articles to existing readers. Finally, CTPF
provides a new exploratory window into the structure of the collection. It organizes the articles
according to their topics and identifies important articles both in terms of those important to their
topic and those that have transcended disciplinary boundaries.
We illustrate the model with an example. Consider the classic paper ?Maximum likelihood from
incomplete data via the EM algorithm? [5]. This paper, published in the Journal of the Royal Statistical Society (B) in 1977, introduced the expectation-maximization (EM) algorithm. The EM
algorithm is a general method for finding maximum likelihood estimates in models with hidden
random variables. As many readers will know, EM has had an enormous impact on many fields,
1
including computer vision, natural language processing, and machine learning. This original paper
has been cited over 37,000 times.
Figure 1 illustrates the CTPF representation of the EM paper. (This model was fit to the shared
libraries of scientists on the Mendeley website; the number of readers is 80,000 and the number of
articles is 261,000.) In the figure, the horizontal axes contains topics, latent themes that pervade
the collection [2]. Consider the black bars in the left figure. These represent the topics that the
EM paper is about. (These were inferred from the abstract of the paper.) Specifically, it is about
probabilistic modeling and statistical algorithms. Now consider the red bars on the right, which are
summed with the black bars. These represent the preferences of the readers who have the EM paper
in their libraries. CTPF has uncovered the interdisciplinary impact of the EM paper. It is popular
with readers interested in many fields outside of those the paper discusses, including computer vision
and statistical network analysis.
The CTPF representation has advantages. For forming recommendations, it naturally interpolates
between using the text of the article (the black bars) and the inferred representation from user behavior data (the red bars). On one extreme, it recommends rarely or never read articles based mainly on
their text; this is the cold start problem. On the other extreme, it recommends widely-read articles
based mainly on their readership. In this setting, it can make good inferences about the red bars.
Further, in contrast to traditional matrix factorization algorithms, we combine the space of preferences and articles via interpretable topics. CTPF thus offers reasons for making recommendations,
readable descriptions of reader preferences, and an interpretable organization of the collection. For
example, CTPF can recognize the EM paper is among the most important statistics papers that has
had an interdisciplinary impact.
In more detail, CTPF draws on ideas from two existing models: collaborative topic regression [20]
and Poisson factorization [9]. Poisson factorization is a form of probabilistic matrix factorization [17] that replaces the usual Gaussian likelihood and real-valued representations with a Poisson
likelihood and non-negative representations. Compared to Gaussian factorization, Poisson factorization enjoys more efficient inference and better handling of sparse data. However, PF is a basic
recommendation model. It cannot handle the cold start problem or easily give topic-based representations of readers and articles.
Collaborative topic regression is a model of text and reader data that is based on the same intuitions
as we described above. (Wang and Blei [20] also use the EM paper as an example.) However, in its
implementation, collaborative topic regression is a non-conjugate model that is complex to fit, difficult to work with on sparse data, and difficult to scale without stochastic optimization. Further, it is
based on a Gaussian likelihood of reader behavior. Collaborative Poisson factorization, because it is
based on Poisson and gamma variables, enjoys an easier-to-implement and more efficient inference
algorithm and a better fit to sparse real-world data. As we show below, it scales more easily and
provides significantly better recommendations than collaborative topic regression.
2
The collaborative topic Poisson factorization model
In this section we describe the collaborative topic Poisson factorization model (CTPF), and discuss
its statistical properties. We are given data about users (readers) and documents (articles), where
each user has read or placed in his library a set of documents. The rating rud equals one if user u
consulted document d, can be greater than zero if the user rated the document and is zero otherwise.
Most of the values of the matrix y are typically zero, due to sparsity of user behavior data.
Background: Poisson factorization. CTPF builds on Poisson matrix factorization [9]. In collaborative filtering, Poisson factorization (PF) is a probabilistic model of users and items. It associates
each user with a latent vector of preferences, each item with a latent vector of attributes, and constrains both sets of vectors to be sparse and non-negative. Each cell of the observed matrix is
assumed drawn from a Poisson distribution, whose rate is a linear combination of the corresponding
user and item attributes. Poisson factorization has also been used as a topic model [3], and developed
as an alternative text model to latent Dirichlet allocation (LDA). In both applications Poisson factorization has been shown to outperform competing methods [3, 9]. PF is also more easily applicable
to real-life preference datasets than the popular Gaussian matrix factorization [9].
2
40
40
!
30
30
20
algorithm, efficient,
optimal, clustering,
optimization, show
probability, prior,
bayesian, likelihood,
inference, maximum
20
network, connected,
modules, nodes,
links, topology
!
!
10
image, object,
matching, tracking"
motion,segmentation
10
!
!
!
!
!
!!!
!
!
!
!
!
0
0
!
!
!
!
!
!
!
!
! ! !
!!!
!
!!!!!!!!!!!!!!!!!!! !!! ! !!!!!!!!!!!!!!!! !!!!! !!!!!!!!!! !!!!!!!!!!!!!!!!!!!!!!!! !!!!!!!!!!!!!!!! !!!!!!!!!!!!!!! !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !! !!
Topic
Topic
Figure 1: We visualized the inferred topic intensities ? (the black bars) and the topic offsets (the
red bars) of an article in the Mendeley [13] dataset. The plots are for the statistics article titled
?Maximum likelihood from incomplete data via the EM algorithm?. The black bars represent the
topics that the EM paper is about. These include probabilistic modeling and statistical algorithms.
The red bars represent the preferences of the readers who have the EM paper in their libraries. It
is popular with readers interested in many fields outside of those the paper discusses, including
computer vision and statistical network analysis.
Collaborative topic Poisson factorization. CTPF is a latent variable model of user ratings and
document content. CTPF uses Poisson factorization to model both types of data. Rather than modeling them as independent factorization problems, we connect the two latent factorizations using a
correction term [20] which we?ll describe below.
Suppose we have data containing D documents and U users. CTPF assumes a collection of K
unormalized topics ?1:K . Each topic ?k is a collection of word intensities on a vocabulary of size
V . Each component ?vk of the unnormalized topics is drawn from a Gamma distribution. Given
the topics, CTPF assumes that a document d is generated with a vector of K latent topic intensities
?d , and represents users with a vector of K latent topic preferences ?u . Additionally, the model
associates each document with K latent topic offsets d that capture the document?s deviation from
the topic intensities. These deviations occur when the content of a document is insufficient to explain
its ratings. For example, these variables can capture that a machine learning article is interesting to
a biologist, because other biologists read it.
We now define a generative process for the observed word counts in documents and observed user
ratings of documents under CTPF:
1. Document model:
(a) Draw topics ?vk ? Gamma(a, b)
(b) Draw document topic intensities ?dk ? Gamma(c, d)
(c) Draw word count wdv ? Poisson(?dT ?v ).
2. Recommendation model:
(a) Draw user preferences ?uk ? Gamma(e, f )
(b) Draw document topic offsets dk ? Gamma(g, h)
(c) Draw rud ? Poisson(?uT (?d + d )).
CTPF specifies that the conditional probability that a user u rated document d with rating rud is
drawn from a Poisson distribution with rate parameter ?uT (?d + d ). The form of the factorization
couples the user preferences for the document topic intensities ?d and the document topic offsets d .
This allows the user preferences to be interpreted as affinity to latent topics.
CTPF has two main advantages over previous work (e.g., [20]), both of which contribute to its
superior empirical performance (see Section 5). First, CTPF is a conditionally conjugate model
when augmented with auxiliary variables. This allows CTPF to conveniently use standard variational
inference with closed-form updates (see Section 3). Second, CTPF is built on Poisson factorization;
it can take advantage of the natural sparsity of user consumption of documents and can analyze
massive real-world data. This follows from the likelihood of the observed data under the model [9].
3
We analyze user preferences and document content with CTPF via its posterior distribution over
latent variables p(?1:K , ?1:D , 1:D , ?1:U |w, r). By estimating this distribution over the latent structure, we can characterize user preferences and document readership in many useful ways. Figure 1
gives an example.
Recommending old and new documents. Once the posterior is fit, we use CTPF to recommend
in-matrix documents and out-matrix or cold-start documents to users. We define in-matrix documents as those that have been rated by at least one user in the recommendation system. All other
documents are new to the system. A cold-start recommendation of a new document is based entirely
on its content. For predicting both in-matrix and out-matrix documents, we rank each user?s unread
documents by their posterior expected Poisson parameters,
scoreud = E[?uT (?d + d )|w, r].
(1)
The intuition behind the CTPF posterior is that when there is no reader data, we depend on the
topics to make recommendations. When there is both reader data and article content, this gives
information about the topic offsets. We emphasize that under CTPF the in-matrix recommendations
and cold-start recommendations are not disjoint tasks. There is a continuum between these tasks.
For example, the model can provide better predictions for articles with few ratings by leveraging its
latent topic intensities ?d .
3
Approximate posterior inference
Given a set of observed document ratings r and their word counts w, our goal is to infer the topics
?1:K , the user preferences ?1:U , the document topic intensities ?1:D , the document topic offsets 1:D .
With estimates of these quantities, we can recommend in-matrix and out-matrix documents to users.
Computing the exact posterior distribution p(?1:K , ?1:D , 1:D , ?1:U |w, r) is intractable; we use variational inference [15]. We first develop a coordinate ascent algorithm?a batch algorithm that iterates over only the non-zero document-word counts and the non-zero user-document ratings. We then
present a more scalable stochastic variational inference algorithm.
In variational inference we first define a parameterized family of distributions over the hidden variables. We then fit the parameters to find a distribution that minimizes the KL divergence to the
posterior. The model is conditionally conjugate if the complete conditional of each latent variable
is in the exponential family and is in the same family as its prior. (The complete conditional is the
conditional distribution of a latent variable given the observations and the other latent variables in
the model [8].) For the class of conditionally conjugate models, we can perform this optimization
with a coordinate ascent algorithm and closed form updates.
Auxiliary variables. To facilitate inference, we first augment CTPF with auxiliary variables. Following Ref. [6] and Ref. [9],
P we add K latent variables zdv,k ? Poisson(?dk ?vk ), which are
integers such that wdv =
k zdv,k . Similarly, for each observed rating rud , we add K latent
a
b
variables yud,k
? Poisson(?uk ?dk ) and K latent variables yud,k
? Poisson(?uk dk ) such that
P a
b
rud = k yud,k + yud,k . A sum of independent Poisson random variables is itself a Poisson with
rate equal to the sum of the rates. Thus, these new latent variables preserve the marginal distribution
of the observations, wdv and rud . Further, when the observed counts are 0, these auxiliary variables
are not random. Consequently, our inference procedure need only consider the auxiliary variables
for non-zero observations.
CTPF with the auxiliary variables is conditionally conjugate; its complete conditionals are shown
in Table 1. The complete conditionals of the Gamma variables ?vk , ?dk , dk , and ?uk are Gamma
distributions with shape and rate parameters as shown in Table 1. For the auxiliary Poisson variables,
observe that zdv is a K-dimensional latent vector of Poisson counts, which when conditioned on
their observed sum wdv , is distributed as a multinomial [14, 4]. A similar reasoning underlies the
conditional for yud which is a 2K-dimensional latent vector of Poisson counts. With our complete
conditionals in place, we now derive the coordinate ascent algorithm for the expanded set of latent
variables.
4
Latent Variable
Type
?dk
?vk
?uk
dk
zdv
Gamma
Gamma
Gamma
Gamma
Mult
yud
Mult
Complete conditional
P
P a
P
P
c + v zdv,k + u yud,k
, d + v ?vk + u ?uk
P
P
a + d zdv,k , b + d ?dk
P a
P b
P
e + d yud,k
+ d yud,k
, f + d (?dk + dk )
P b
P
g + u yud,k , h + u ?uk
log
( ?dk + log ?vk
log ?uk + log ?dk if k < K,
log ?uk + log dk if K ? k < 2K
Variational parameters
shp ?rte
??dk
, ?dk
shp ?rte
?
?vk
, ?vk
shp
rte
??uk
, ??uk
shp rte
?dk , ?dk
?dv
?ud
Table 1: CTPF: latent variables, complete conditionals and variational parameters.
Variational family. We define the mean-field variational family q(?, ?, ?, , z, y) over the latent
variables where we consider these variables to be independent and each governed by its own distribution,
Y
Y
Y
Y
Y
q(?, ?, , ?, z, y) =
q(?vk )
q(?dk )q(dk )
q(?uk )
q(yud,k )
q(zdv,k ).
(2)
v,k
d,k
u,k
ud,k
dv,k
The variational factors for topic components ?vk , topic intensities ?dk , user preferences ?uk are
all Gamma distributions?the same as their conditional distributions?with freely set shape and
rate variational parameters. For example, the variational distribution for the topic intensities ?dk
shp ?rte
is Gamma(?dk ; ??dk
, ?dk ). We denote shape with the superscript ?shp? and rate with the superscript
?rte?. The variational factor for zdv is a multinomial Mult(wdv , ?dv ) where the variational parameter
a
b
?dv is a point on the K-simplex. The variational factor for yud = (yud
, yud
) is also a multinomial
Mult(rud , ?ud ) but here ?ud is a point in the 2K-simplex.
Optimal coordinate updates. In coordinate ascent we iteratively optimize each variational parameter while holding the others fixed. Under the conditionally conjugate augmented CTPF, we
can optimize each coordinate in closed form by setting the variational parameter equal to the expected natural parameter (under q) of the complete conditional. For a given random variable, this
expected conditional parameter is the expectation of a function of the other random variables and
observations. (For details, see [9, 10]). We now describe two of these updates; the other updates
are similarly derived.
The update for the variational shape and rate parameters of topic intensities ?dk is
??dk
= hc +
P
v
wdv ?dv,k +
P
u rud ?ud,k , d
+
shp
??vk
v ??rte
vk
P
+
shp
??uk
rte i.
u ??uk
P
(3)
The Gamma update in Equation 3 derives from the expected natural parameter (under q) of the
complete conditional for ?dk in Table 1. In the shape parameter for topic intensities for document d,
a
we use that Eq [zdv,k ] = wdv ?dv,k for the word indexed by v and Eq [yud,k
] = rud ?ud,k for the user
indexed by u. In the rate parameter, we use that the expectation of a Gamma variable is the shape
divided by the rate.
The update for the multinomial ?dv is
?dv
?
shp
shp
rte
rte
exp{?(??dk
) ? log ??dk
+ ?(??vk
) ? log ??vk
},
(4)
where ?(?) is the digamma function (the first derivative of the log ? function). This update comes
shp
rte
from the expectation of the log of a Gamma variable, for example, Eq [log ?dk ] = ?(??dk
) ? log ??dk
.
Coordinate ascent algorithm. The CTPF coordinate ascent algorithm is illustrated in Figure 2.
Similar to the algorithm of [9], our algorithm is efficient on sparse matrices. In steps 1 and 2, we
need to only update variational multinomials for the non-zero word counts wdv and the non-zero
ratings rud . In step 3, the sums over the expected zdv,k and the expected yud,k need only to consider
non-zero observations. This efficiency comes from the likelihood of the full matrix depending only
on the non-zero observations [9].
5
Initialize the topics ?1:K and topic intensities ?1:D using LDA [2] as described in Section 3.
Repeat until convergence:
1. For each word count wdv > 0, set ?dv to the expected conditional parameter of zdv .
2. For each rating rud > 0, set ?ud to the expected conditional parameter of yud .
3. For each document d and each k, update the block of variational topic intensities ??dk to
their expected conditional parameters using Equation 3. Perform similar block updates
for ??vk , ??uk and ?dk , in sequence.
Figure 2: The CTPF coordinate ascent algorithm. The expected conditional parameters of the latent
variables are computed from Table 1.
Stochastic algorithm. The CTPF coordinate ascent algorithm is efficient: it only iterates over the
non-zero observations in the observed matrices. The algorithm computes approximate posteriors for
datasets with ten million observations within hours (see Section 5). To fit to larger datasets, within
hours, we develop an algorithm that subsamples a document and estimates variational parameters
using stochastic variational inference [10]. The stochastic algorithm is also useful in settings where
new items continually arrive in a stream. The CTPF SVI algorithm is described in the Appendix.
Computational efficiency. The SVI algorithm is more efficient than the batch algorithm. The
batch algorithm has a per-iteration computational complexity of O((W + R)K) where R and W
are the total number of non-zero observations in the document-user and document-word matrices,
respectively. For the SVI algorithm, this is O((wd + rd )K) where rd is the number of users rating
the sampled document d and wd is the number of unique words in it. (We assume that a single
document is sampled in each iteration.) In Figure 2, the sums involving the multinomial parameters
can be tracked for efficient memory usage. The bound on memory usage is O((D + V + U )K).
Hyperparameters, initialization and stopping criteria: Following [9], we fix each Gamma shape
and rate hyperparameter at 0.3. We initialize the variational parameters for ?uk and dk to the prior
on the corresponding latent variables and add small uniform noise. We initialize ??vk and ??dk using
estimates of their normalized counterparts from LDA [2] fitted to the document-word matrix w. For
the SVI algorithm described in the Appendix, we set learning rate parameters ?0 = 1024, ? = 0.5
and use a mini-batch size of 1024. In both algorithms, we declare convergence when the change in
expected predictive likelihood is less than 0.001%.
4
Related work
Several research efforts propose joint models of item covariates and user activity. Singh and Gordon [19] present a framework for simultaneously factorizing related matrices, using generalized link
functions and coupled latent spaces. Hong et al. [11] propose Co-factorization machines for modeling user activity on twitter with tweet features, including content. They study several design choices
for sharing latent spaces. While CTPF is roughly an instance of these frameworks, we focus on the
task of recommending articles to readers.
Agarwal and Chen [1] propose fLDA, a latent factor model which combines document features
through their empirical LDA [2] topic intensities and other covariates, to predict user preferences.
The coupling of matrix decomposition and topic modeling through shared latent variables is also
considered in [18, 22]. Like fLDA, both papers tie latent spaces without corrective terms. Wang
and Blei [20] have shown the importance of using corrective terms through the collaborative topic
regression (CTR) model which uses a latent topic offset to adjust a document?s topic proportions.
CTR has been shown to outperform a variant of fLDA [20]. Our proposed model CTPF uses the
CTR approach to sharing latent spaces.
CTR [20] combines topic modeling using LDA [2] with Gaussian matrix factorization for one-class
collaborative filtering [12]. Like CTPF, the underlying MF algorithm has a per-iteration complexity that is linear in the number of non-zero observations. Unlike CTPF, CTR is not conditionally
6
Mean precision
CTPF (Section 2)
Decoupled PF (Section 5)
mendeley.in
Content Only
Ratings Only [9]
mendeley.out
Collaborative Topic Regression [20]
arxiv.in
arxiv.out
0.4%
0.4%
1.0%
2.0%
1.5%
1.0%
0.5%
0.3%
0.2%
0.5%
0.1%
10 30 50 70
100
10 30 50 70
mendeley.in
4%
2%
50
70
100
100
10 30 50 70
100
Collaborative Topic Regression [20]
arxiv.in
arxiv.out
4%
5%
4%
3%
2%
1%
30
0.1%
10 30 50 70
mendeley.out
6%
10
0.2%
Decoupled PF (Section
5)
Content
Only
Ratings Only [9]
Number
of recommendations
CTPF (Section 2)
Mean recall
100
0.3%
2.0%
1.5%
1.0%
0.5%
3%
2%
1%
0%
10
30
50
70
100
10
30
50
70
100
10
30
50
70
100
Number of recommendations
Figure 3: The CTPF coordinate ascent algorithm outperforms CTR and other competing algorithms on both
in-matrix and out-matrix predictions. Each panel shows the in-matrix or out-matrix recommendation task on
the Mendeley data set or the 1-year arXiv data set. Note that the Ratings-only model cannot make out-matrix
predictions. The mean precision and mean recall are computed from a random sample of 10,000 users.
conjugate, and the inference algorithm depends on numerical optimization of topic intensities. Further, CTR requires setting confidence parameters that govern uncertainty around a class of observed
ratings. As we show in Section 5, CTPF scales more easily and provides significantly better recommendations than CTR.
.
5
Empirical results
We use the predictive approach to evaluating model fitness [7], comparing the predictive accuracy
of the CTPF coordinate ascent algorithm in Figure 2 to collaborative topic regression (CTR) [21].
We also compare to variants of CTPF to demonstrate that coupling the latent spaces using corrective
terms is essential for good predictive performance, and that CTPF predicts significantly better than
its variants and CTR. Finally, we explore large real-world data sets revealing the interaction patterns
between readers and articles.
Data sets. We study the CTPF algorithm of Figure 2 on two data sets. The Mendeley data set [13]
of scientific articles is a binary matrix of 80,000 users and 260,000 articles with 5 million observations. Each cell corresponds to the presence or absence of an article in a scientist?s online library.
The arXiv data set is a matrix of 120,297 users and 825,707 articles, with 43 million observations.
Each observation indicates whether or not a user has consulted an article (or its abstract). This data
was collected from the access logs of registered users on the http://arXiv.org paper repository. The
articles and the usage data spans a timeline of 10 years (2003-2012). In our experiments on predictive performance, we use a subset of the data set, with 64,978 users 636,622 papers and 7.6 million
clicks, which spans one year of usage data (2012). We treat the user clicks as implicit feedback and
specifically as binary data. For each article in the above data sets, we remove stop words and use
tf-idf to choose the top 10,000 distinct words (14,000 for arXiv) as the vocabulary. We implemented
the batch and stochastic algorithms for CTPF in 4500 lines of C++ code.1
Competing methods. We study the predictive performance of the following models. With the
exception of the Poisson factorization [9], which does not model content, the topics and topic intensities (or proportions) in all CTPF models are initialized using LDA [2], and fit using batch
variational inference. We set K = 100 in all of our experiments.
? CTPF: CTPF is our proposed model (Section 2) with latent user preferences tied to a single
vector ?u , and interpreted as affinity to latent topics ?.
1
Our source code is available from: https://github.com/premgopalan/collabtm
7
Topic: "Statistical Inference Algorithms"
Topic: ?Information Retrieval?
A) Articles about the topic; readers in the field
On the ergodicity properties of adaptive MCMC algorithms
Particle filtering within adaptive Metropolis Hastings sampling
An Adaptive Sequential Monte Carlo Sampler
A) Articles about the topic; readers in the field
The anatomy of a large-scale hypertextual Web search engine
Authoritative sources in a hyperlinked environment
A translation approach to portable ontology specifications
B) Articles outside the topic; readers in the field
A comparative review of dimension reduction methods in ABC
Computational methods for Bayesian model choice
The Proof of Innocence
B) Articles outside the topic; readers in the field
How to choose a good scientific problem.
Practical Guide to Support Vector Classification
Maximum likelihood from incomplete data via the EM?
C) Articles about this field; readers outside the field
Introduction to Monte Carlo Methods
An introduction to Monte Carlo simulation of statistical...
The No-U-Turn Sampler: Adaptively setting path lengths...
C) Articles about this field; readers outside the field
Data clustering: a review
Defrosting the digital library: bibliographic tools?
Top 10 algorithms in data mining
Figure 4: The top articles by the expected weight ?dk from a component discovered by our stochastic variational inference in the arXiv data set (Left) and Mendeley (Right). Using the expected topic proportions ?dk
and the expected topic offsets dk , we identified subclasses of articles: A) corresponds to the top articles by
topic proportions in the field of ?Statistical inference algorithms? for arXiv and ?Ontologies and applications?
for Mendeley; B) corresponds to the top articles with low topic proportions in this field, but a large ?dk + dk ,
demonstrating the outside interests of readers of that field (e.g., very popular papers often appear such as ?The
Proof of Innocence? which describes a rigorous way to ?fight your traffic tickets?). C) corresponds to the top
articles with high topic proportions in this field but that also draw significant interest from outside readers.
? Decoupled Poisson Factorization: This model is similar to CTPF but decouples the user
latent preferences into distinct components pu and qu , each of dimension K. We have,
wdv ? Poisson(?dT ?v ); rud ? Poisson(pTu ?d + quT d ).
(5)
The user preference parameters for content and ratings can vary freely. The qu are independent of topics and offer greater modeling flexibility, but they are less interpretable than
the ?u in CTPF. Decoupling the factorizations has been proposed by Porteous et al. [16].
? Content Only: We use the CTPF model without the document topic offsets d . This resembles the idea developed in [1] but using Poisson generating distributions.
? Ratings Only [9]: We use Poisson factorization to the observed ratings. This model can
only make in-matrix predictions.
? CTR [20]: A full optimization of this model does not scale to the size of our data sets
despite running for several days. Accordingly, we fix the topics and document topic proportions to their LDA values. This procedure is shown to perform almost as well as jointly
optimizing the full model in [20]. We follow the authors? experimental settings. Specifically, for hyperparameter selection we started with the values of hyperparameters suggested
by the authors and explored various values of the learning rate as well as the variance of
the prior over the correction factor (?v in [20]). Training convergence was assessed using
the model?s complete log-likelihood on the training observations. (CTR does not use a
validation set.)
Evaluation. Prior to training models, we randomly select 20% of ratings and 1% of documents in
each data set to be used as a held-out test set. Additionally, we set aside 1% of the training ratings as
a validation set (20% for arXiv) and use it to determine convergence. We used the CTPF settings described in Section 3 across both data sets. During testing, we generate the top M recommendations
for each user as those items with the highest predictive score under each method. Figure 3 shows the
mean precision and mean recall at varying number of recommendations for each method and data
set. We see that CTPF outperforms CTR and the Ratings-only model on all data sets. CTPF outperforms the Decoupled PF model and the Content-only model on all data sets except on cold-start
predictions on the arXiv data set, where it performs equally well. The Decoupled PF model lacks
CTPF?s interpretable latent space. The Content-only model performs poorly on most tasks; it lacks a
corrective term on topics to account for user ratings. In Figure 4, we explored the Mendeley and the
arXiv data sets using CTPF. We fit the Mendeley data set using the coordinate ascent algorithm, and
the full arXiv data set using the stochastic algorithm from Section 3. Using the expected document
topic intensities ?dk and the expected document topic offsets dk , we identified interpretable topics
and subclasses of articles that reveal the interaction patterns between readers and articles.
8
References
[1] D. Agarwal and B. Chen. fLDA: Matrix factorization through latent Dirichlet allocation. In Proceedings
of the third ACM international conference on web search and data mining, pages 91?100. ACM, 2010.
[2] D. Blei, A. Ng, and M. Jordan. latent Dirichlet allocation. Journal of Machine Learning Research, 3:
993?1022, January 2003.
[3] J. Canny. GaP: A factor model for discrete data. In Proceedings of the 27th Annual International ACM
SIGIR Conference on Research and Development in Information Retrieval, 2004.
[4] A. Cemgil. Bayesian inference for nonnegative matrix factorization models. Computational Intelligence
and Neuroscience, 2009.
[5] A. Dempster, N. Laird, and D. Rubin. Maximum likelihood from incomplete data via the EM algorithm.
Journal of the Royal Statistical Society, Series B, 39:1?38, 1977.
[6] D. B Dunson and A. H. Herring. Bayesian latent variable models for mixed discrete outcomes. Biostatistics, 6(1):11?25, 2005.
[7] S. Geisser and W.F. Eddy. A predictive approach to model selection. Journal of the American Statistical
Association, pages 153?160, 1979.
[8] Z. Ghahramani and M. Beal. Variational inference for Bayesian mixtures of factor analysers. In Neural
Information Processing Systems, volume 12, 2000.
[9] P. Gopalan, J.M. Hofman, and D. Blei. Scalable recommendation with Poisson factorization. arXiv
preprint arXiv:1311.1704, 2013.
[10] M. Hoffman, D. Blei, C. Wang, and J. Paisley. Stochastic variational inference. Journal of Machine
Learning Research, 14:1303?1347, 2013.
[11] L. Hong, A. S. Doumith, and B.D. Davison. Co-factorization machines: Modeling user interests and
predicting individual decisions in Twitter. In Proceedings of the sixth ACM international conference on
web search and data mining, pages 557?566. ACM, 2013.
[12] Y. Hu, Y. Koren, and C. Volinsky. Collaborative filtering for implicit feedback datasets. In Eighth IEEE
International Conference on Data Mining., pages 263?272. IEEE, 2008.
[13] K. Jack, J. Hammerton, D. Harvey, J. J. Hoyt, J. Reichelt, and V. Henning. Mendeley?s reply to the
datatel challenge. Procedia Computer Science, 1(2):1?3, 2010. URL http://www.mendeley.com/
research/sei-whale/.
[14] N. Johnson, A. Kemp, and S. Kotz. Univariate Discrete Distributions. John Wiley & Sons, 2005.
[15] M. Jordan, Z. Ghahramani, T. Jaakkola, and L. Saul. Introduction to variational methods for graphical
models. Machine Learning, 37:183?233, 1999.
[16] I. Porteous, A. U. Asuncion, and M. Welling. Bayesian matrix factorization with side information and
Dirichlet process mixtures. In Maria Fox and David Poole, editors, In the conference of the Association
for the Advancement of Artificial Intelligence. AAAI Press, 2010.
[17] R. Salakhutdinov and A. Mnih. Bayesian probabilistic matrix factorization using Markov chain Monte
Carlo. In Proceedings of the 25th international conference on machine learning, pages 880?887, 2008.
[18] H. Shan and A. Banerjee. Generalized probabilistic matrix factorizations for collaborative filtering. In
Data Mining (ICDM), 2010 IEEE 10th International Conference on, pages 1025?1030. IEEE, 2010.
[19] A. P. Singh and G. J. Gordon. Relational learning via collective matrix factorization. In Proceedings of the
14th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 650?658.
ACM, 2008.
[20] C. Wang and D. Blei. Collaborative topic modeling for recommending scientific articles. In Knowledge
Discovery and Data Mining, 2011.
[21] C. Wang, J. Paisley, and D. Blei. Online variational inference for the hierarchical Dirichlet process. In
Artificial Intelligence and Statistics, 2011.
[22] X. Zhang and L. Carin. Joint modeling of a matrix with associated text via latent binary features. In
Advances in Neural Information Processing Systems, pages 1556?1564, 2012.
9
| 5360 |@word repository:1 proportion:7 hu:1 simulation:1 decomposition:1 reduction:1 contains:2 uncovered:1 score:1 bibliographic:1 series:1 document:51 outperforms:4 existing:2 wd:2 comparing:1 com:2 herring:1 john:1 numerical:1 shape:7 remove:1 plot:1 interpretable:6 update:12 aside:1 generative:2 intelligence:3 website:1 item:6 advancement:1 accordingly:1 blei:9 davison:1 provides:5 iterates:2 node:1 contribute:1 preference:21 org:1 zhang:1 combine:3 expected:16 roughly:1 ontology:2 behavior:5 salakhutdinov:1 window:1 pf:7 estimating:1 underlying:1 panel:1 biostatistics:1 interpreted:2 minimizes:1 developed:2 finding:1 nj:1 subclass:2 tie:1 decouples:1 uk:17 underlie:1 appear:1 continually:1 declare:1 scientist:2 treat:1 cemgil:1 despite:1 laurent:1 path:1 might:1 black:5 initialization:1 resembles:1 co:2 factorization:38 pervade:1 unique:1 practical:1 testing:1 pgopalan:1 block:2 implement:1 rte:11 svi:4 cold:7 procedure:2 empirical:3 significantly:3 mult:4 matching:1 revealing:1 word:14 confidence:1 cannot:2 selection:2 optimize:2 www:1 sigir:1 his:1 classic:1 handle:1 exploratory:1 coordinate:13 suppose:1 massive:3 user:46 exact:1 us:3 associate:2 predicts:1 observed:11 module:1 preprint:1 wang:5 capture:3 thousand:1 connected:1 highest:1 intuition:2 environment:1 govern:1 complexity:2 constrains:1 covariates:2 dempster:1 depend:1 unread:2 singh:2 hofman:1 predictive:8 efficiency:2 easily:4 joint:2 various:1 corrective:4 distinct:2 describe:3 monte:4 artificial:2 analyser:1 outside:8 outcome:1 whose:1 widely:1 solve:1 valued:1 larger:1 otherwise:1 statistic:4 jointly:1 itself:1 laird:1 superscript:2 online:2 beal:1 subsamples:1 advantage:3 sequence:1 hyperlinked:1 propose:3 interaction:2 canny:1 flexibility:1 poorly:1 description:1 convergence:4 comparative:1 generating:1 object:1 illustrate:1 develop:4 derive:1 depending:1 coupling:2 ticket:1 eq:3 auxiliary:7 c:2 implemented:1 come:2 anatomy:1 attribute:2 stochastic:10 disciplinary:1 fix:2 unormalized:1 correction:2 around:1 considered:1 exp:1 predict:1 continuum:1 vary:1 applicable:1 sei:1 tf:1 tool:1 hoffman:1 gaussian:5 rather:1 varying:1 jaakkola:1 ax:1 derived:1 focus:1 vk:17 maria:1 rank:1 likelihood:13 mainly:2 indicates:1 contrast:1 digamma:1 rigorous:1 baseline:1 sigkdd:1 inference:22 twitter:2 stopping:1 typically:1 fight:1 hidden:2 interested:2 zdv:11 among:1 classification:1 augment:1 development:1 art:1 summed:1 initialize:3 biologist:2 marginal:1 field:16 equal:3 never:1 once:1 ng:1 sampling:1 whale:1 represents:2 geisser:1 carin:1 simplex:2 others:1 recommend:3 gordon:2 few:3 randomly:1 gamma:18 recognize:1 divergence:1 preserve:1 simultaneously:1 fitness:1 individual:1 organization:1 interest:4 mining:7 mnih:1 evaluation:1 adjust:1 mixture:2 extreme:2 behind:1 held:1 chain:1 decoupled:5 fox:1 indexed:2 incomplete:4 old:1 initialized:1 fitted:1 instance:1 modeling:10 maximization:1 deviation:2 subset:1 uniform:1 johnson:1 characterize:1 connect:1 adaptively:1 cited:1 international:7 interdisciplinary:2 probabilistic:7 hoyt:1 connecting:1 ctr:13 aaai:1 containing:1 choose:2 american:1 derivative:1 account:1 depends:1 stream:1 closed:3 analyze:2 traffic:1 red:5 start:7 asuncion:1 collaborative:19 accuracy:1 variance:1 who:2 bayesian:7 carlo:4 published:1 history:1 explain:1 sharing:2 shp:11 sixth:1 volinsky:1 ptu:1 naturally:1 proof:2 associated:1 couple:1 sampled:2 stop:1 dataset:2 popular:4 recall:3 knowledge:2 ut:3 segmentation:1 eddy:1 innocence:2 dt:2 day:2 follow:1 charlin:1 ergodicity:1 implicit:2 reply:1 until:1 hastings:1 horizontal:1 web:3 banerjee:1 lack:2 lda:7 reveal:1 scientific:3 usage:5 facilitate:1 normalized:1 counterpart:1 read:6 iteratively:1 illustrated:1 conditionally:6 ll:1 during:1 unnormalized:1 criterion:1 generalized:2 hong:2 complete:10 demonstrate:2 performs:3 motion:1 reasoning:1 image:1 variational:29 jack:1 superior:1 multinomial:6 empirically:1 tracked:1 volume:1 million:6 association:2 significant:1 paisley:2 rd:2 similarly:2 particle:1 language:1 had:2 access:1 specification:1 add:3 pu:1 posterior:8 own:1 optimizing:1 harvey:1 binary:3 life:1 greater:2 freely:2 determine:1 ud:7 full:5 infer:1 offer:2 long:1 retrieval:2 divided:1 icdm:1 equally:1 impact:3 prediction:5 involving:1 variant:3 regression:8 basic:1 vision:3 expectation:4 poisson:40 scalable:2 arxiv:17 underlies:1 represent:6 iteration:3 agarwal:2 cell:2 background:1 conditionals:4 source:2 unlike:1 ascent:11 henning:1 leveraging:1 jordan:2 integer:1 presence:1 recommends:2 fit:9 competing:4 topology:1 click:2 identified:2 idea:2 whether:1 rud:12 url:1 effort:1 titled:1 interpolates:1 york:2 useful:2 gopalan:2 ten:1 visualized:1 http:3 specifies:1 outperform:2 generate:1 neuroscience:1 disjoint:1 per:2 discrete:3 hyperparameter:2 demonstrating:1 enormous:1 drawn:3 tweet:1 sum:5 year:3 parameterized:1 uncertainty:1 place:1 family:5 reader:33 arrive:1 almost:1 kotz:1 draw:8 decision:1 appendix:2 entirely:1 bound:1 shan:1 koren:1 replaces:1 hypertextual:1 annual:1 activity:2 nonnegative:1 occur:1 idf:1 your:1 personalized:1 ctpf:61 span:2 expanded:1 department:3 according:1 combination:1 conjugate:7 describes:1 across:1 em:15 son:1 metropolis:1 qu:2 making:1 dv:9 equation:2 previously:1 discus:3 count:10 mechanism:1 turn:1 know:1 available:1 observe:1 hierarchical:1 alternative:1 batch:6 original:1 assumes:2 dirichlet:5 clustering:2 include:1 top:7 porteous:2 running:1 graphical:1 readable:1 exploit:1 ghahramani:2 build:2 society:2 quantity:1 usual:1 traditional:1 affinity:2 link:2 consumption:1 topic:86 collected:1 portable:1 kemp:1 reason:1 code:2 length:1 insufficient:1 mini:1 difficult:2 dunson:1 holding:1 negative:2 implementation:1 design:1 collective:1 perform:3 recommender:1 observation:14 datasets:5 markov:1 january:1 relational:1 discovered:1 consulted:2 intensity:18 inferred:3 rating:24 david:3 introduced:1 kl:1 engine:1 registered:1 hour:2 timeline:1 bar:10 suggested:1 below:2 pattern:3 poole:1 eighth:1 sparsity:2 challenge:1 built:1 including:5 royal:2 memory:2 natural:5 predicting:2 github:1 rated:3 library:6 flda:4 identifies:2 started:1 coupled:1 columbia:4 text:7 prior:5 understanding:1 review:2 discovery:2 mixed:1 interesting:2 filtering:5 allocation:3 digital:1 validation:2 authoritative:1 article:45 rubin:1 editor:1 doumith:1 translation:1 placed:1 repeat:1 enjoys:2 guide:1 side:1 saul:1 face:1 sparse:6 distributed:1 boundary:1 feedback:2 vocabulary:2 world:4 evaluating:1 dimension:2 computes:1 author:2 collection:5 adaptive:3 welling:1 approximate:2 emphasize:1 assumed:1 recommending:4 factorizing:1 search:3 latent:48 tailed:1 table:5 additionally:2 decoupling:1 hc:1 complex:1 main:1 noise:1 hyperparameters:2 ref:2 augmented:2 qut:1 ny:2 wiley:1 precision:3 theme:1 exponential:1 governed:1 tied:1 third:1 offset:10 dk:45 explored:2 derives:1 intractable:1 essential:1 sequential:1 importance:1 illustrates:1 conditioned:1 chen:2 easier:1 mf:1 gap:1 explore:1 univariate:1 forming:1 conveniently:1 tracking:1 recommendation:20 corresponds:4 abc:1 acm:7 conditional:14 procedia:1 goal:1 consequently:1 shared:2 absence:1 content:15 change:1 typical:1 specifically:3 except:1 sampler:2 called:1 total:1 experimental:1 organizes:1 rarely:1 exception:1 select:1 support:1 assessed:1 prem:1 mcmc:1 princeton:3 handling:1 |
4,817 | 5,361 | Minimax-optimal Inference from Partial Rankings
Bruce Hajek
UIUC
[email protected]
Sewoong Oh
UIUC
[email protected]
Jiaming Xu
UIUC
[email protected]
Abstract
This paper studies the problem of rank aggregation under the Plackett-Luce model.
The goal is to infer a global ranking and related scores of the items, based on partial rankings provided by multiple users over multiple subsets of items. A question
of particular interest is how to optimally assign items to users for ranking and how
many item assignments are needed to achieve a target estimation error. Without
any assumptions on how the items are assigned to users, we derive an oracle lower
bound and the Cram?er-Rao lower bound of the estimation error. We prove an upper bound on the estimation error achieved by the maximum likelihood estimator,
and show that both the upper bound and the Cram?er-Rao lower bound inversely depend on the spectral gap of the Laplacian of an appropriately defined comparison
graph. Since random comparison graphs are known to have large spectral gaps,
this suggests the use of random assignments when we have the control. Precisely,
the matching oracle lower bound and the upper bound on the estimation error imply that the maximum likelihood estimator together with a random assignment is
minimax-optimal up to a logarithmic factor. We further analyze a popular rankbreaking scheme that decompose partial rankings into pairwise comparisons. We
show that even if one applies the mismatched maximum likelihood estimator that
assumes independence (on pairwise comparisons that are now dependent due to
rank-breaking), minimax optimal performance is still achieved up to a logarithmic
factor.
1
Introduction
Given a set of individual preferences from multiple decision makers or judges, we address the problem of computing a consensus ranking that best represents the preference of the population collectively. This problem, known as rank aggregation, has received much attention across various
disciplines including statistics, psychology, sociology, and computer science, and has found numerous applications including elections, sports, information retrieval, transportation, and marketing
[1, 2, 3, 4]. While consistency of various rank aggregation algorithms has been studied when a
growing number of sampled partial preferences is observed over a fixed number of items [5, 6],
little is known in the high-dimensional setting where the number of items and number of observed
partial rankings scale simultaneously, which arises in many modern datasets. Inference becomes
even more challenging when each individual provides limited information. For example, in the well
known Netflix challenge dataset, 480,189 users submitted ratings on 17,770 movies, but on average
a user rated only 209 movies. To pursue a rigorous study in the high-dimensional setting, we assume
that users provide partial rankings over subsets of items generated according to the popular PlackettLuce (PL) model [7] from some hidden preference vector over all the items and are interested in
estimating the preference vector (see Definition 1).
Intuitively, inference becomes harder when few users are available, or each user is assigned few
items to rank, meaning fewer observations. The first goal of this paper is to quantify the number of
item assignments needed to achieve a target estimation error. Secondly, in many practical scenarios
such as crowdsourcing, the systems have the control over the item assignment. For such systems, a
1
natural question of interest is how to optimally assign the items for a given budget on the total number of item assignments. Thirdly, a common approach in practice to deal with partial rankings is to
break them into pairwise comparisons and apply the state-of-the-art rank aggregation methods specialized for pairwise comparisons [8, 9]. It is of both theoretical and practical interest to understand
how much the performance degrades when rank breaking schemes are used.
Notation. For any set S, let |S| denote its cardinality. Let sn1 = {s1 , . . . , sn } denote a set with
n elements. For any positive integer N , let [N ] = {1, . . . , N }. We use standard big O notations,
e.g., for any sequences {an } and {bn }, an = ?(bn ) if there is an absolute constant C > 0 such that
1/C ? an /bn ? C. For a partial ranking ? over S, i.e., ? is a mapping from [|S|] to S, let ? ?1
denote the inverse mapping. All logarithms are natural unless the base is explicitly specified. We
say a sequence of events {An } holds with high probability if P[An ] ? 1 ? c1 n?c2 for two positive
constants c1 , c2 .
1.1
Problem setup
We describe our model in the context of recommender systems, but it is applicable to other systems
with partial rankings. Consider a recommender system with m users indexed by [m] and n items
indexed by [n]. For each item i ? [n], there is a hidden parameter ?i? measuring the underlying
preference. Each user j, independent of everyone else, randomly generates a partial ranking ?j
over a subset of items Sj ? [n] according to the PL model with the underlying preference vector
?? = (?1? , . . . , ?n? ).
Definition 1 (PL model). A partial ranking ? : [|S|] ? S is generated from {?i? , i ? S} under
the PL model in two steps: (1) independently assign each item i ? S an unobserved value Xi ,
?
exponentially distributed with mean e??i ; (2) select ? so that X?(1) ? X?(2) ? ? ? ? ? X?(|S|) .
The PL model can be equivalently described in the following sequential manner.
To generate a
P
?
?i?0
partial ranking ?, first select ?(1) in S randomly from the distribution e?i /
e
; secondly,
i0 ?S
P
?i?0
?i?
select ?(2) in S \ {?(1)} with the probability distribution e /
; continue the
i0 ?S\{?(1)} e
process in the same fashion until all the items in S are assigned. The PL model is a special case of
the following class of models.
Definition 2 (Thurstone model, or random utility model (RUM) ). A partial ranking ? : [|S|] ? S
is generated from {?i? , i ? S} under the Thurstone model for a given CDF F in two steps: (1)
independently assign each item i ? S an unobserved utility Ui , with CDF F (c ? ?i? ); (2) select ? so
that U?(1) ? U?(2) ? ? ? ? ? U?(|S|) .
To recover the PL model from the Thurstone model, take F to be the CDF for the standard Gumbel
?c
distribution: F (c) = e?(e ) . Equivalently, take F to be the CDF of ? log(X) such that X has the
exponential distribution with mean one. For this choice of F, the utility Ui having CDF F (c ? ?i? ),
?
is equivalent to Ui = ? log(Xi ) such that Xi is exponentially distributed with mean e??i . The
corresponding partial permutation ? is such that X?(1) ? X?(2) ? ? ? ? ? X?(|S|) , or equivalently,
U?(1) ? U?(2) ? ? ? ? ? U?(|S|) . (Note the opposite ordering of X?s and U ?s.)
Given the observation of all partial rankings {?j }j?[m] over the subsets {Sj }j?[m] of items, the
task is to infer the underlying preference vector ?? . For the PL model, and more generally for the
Thurstone model, we see that ?? and ?? + a1 for any a ? R are statistically indistinguishable,
where 1 is an all-ones vector. Indeed, under our model, the preference vector ?? is the equivalence
class [?? ] = {? : P
?a ? R, ? = ?? + a1}. To get a unique representation of the equivalence
n
class, we assumeP i=1 ?i? = 0. Then the space of all possible preference vectors is given by
n
? = {? ? Rn : i=1 ?i = 0}. Moreover, if ?i? ? ?i?0 becomes arbitrarily large for all i0 6= i, then
with high probability item i is ranked higher than any other item i0 and there is no way to estimate
?i to any accuracy. Therefore, we further put the constraint that ?? ? [?b, b]n for some b ? R
and define ?b = ? ? [?b, b]n . The parameter b characterizes the dynamic range of the underlying
preference. In this paper, we assume b is a fixed constant. As observed in [10], if b were scaled with
n, then it would be easy to rank items with high preference versus items with low preference and
one can focus on ranking items with close preference.
2
We denote the number of items
Pm assigned to user j by kj := |Sj | and the average number of assigned
1
items per use by k = m j=1 kj ; parameter k may scale with n in this paper. We consider two
scenarios for generating the subsets {Sj }m
j=1 : the random item assignment case where the Sj ?s are
chosen independently and uniformly at random from all possible subsets of [n] with sizes given by
the kj ?s, and the deterministic item assignment case where the Sj ?s are chosen deterministically.
Our main results depend on the structure of a weighted undirected graph G defined as follows.
Definition 3 (Comparison graph G). Each item i ? [n] corresponds to a vertex i ? [n]. For any pair
of vertices i, i0 , there is a P
weighted edge between them if there exists a user who ranks both items i
and i0 ; the weight equals j:i,i0 ?Sj kj1?1 .
P
Let A denote the weighted adjacency matrix of G. Let di = j Aij , so di is the number of users
who rank item i, and without loss of generality assume d1 ? d2 ? ? ? ? ? dn . Let D denote the n ? n
diagonal matrix formed by {di , i ? [n]} and define the graph Laplacian L as L = D ? A. Observe
that L is positive semi-definite and the smallest eigenvalue of L is zero with the corresponding
eigenvector given by the normalized all-one vector. Let 0 = ?1 ? ?2 ? ? ? ? ? ?n denote the
eigenvalues of L in ascending order.
Summary
Pn 1 of main results. Theorem 1 gives a lower bound for the estimation error that scales as
i=2 di . The lower bound is derived based on a genie-argument and holds for both the PL model
andPthe more general Thurstone model. Theorem 2 shows that the Cram?er-Rao lower bound scales
n
as i=2 ?1i . Theorem 3 gives an upper bound for the squared error of the maximum likelihood (ML)
log n
?
estimator that scales as (?mk
2 . Under the full rank breaking scheme that decomposes a k-way
2 ? ?n )
n
k
comparison into 2 pairwise comparisons, Theorem 4 gives an upper bound that scales as mk?log
.
2
2
If the comparison graph is an expander graph, i.e., ?2 ? ?n and mk = ?(n
log
n),
our
lower
and
P
P
upper bounds match up to a log n factor. This follows from the fact that i ?i = i di = mk,
and for expanders mk = ?(n?2 ). Since the Erd?os-R?enyi random graph is an expander graph with
high probability for average degree larger than log n, when the system is allowed to choose the
item assignment, we propose a random assignment scheme under which the items for each user are
chosen independently and uniformly at random. It follows from Theorem 1 that mk = ?(n) is
necessary for any item assignment scheme to reliably infer the underlying preference vector, while
our upper bounds imply that mk = ?(n log n) is sufficient with the random assignment scheme and
can be achieved by either the ML estimator or the full rank breaking or the independence-preserving
breaking that decompose a k-way comparison into bk/2c non-intersecting pairwise comparisons,
proving that rank breaking schemes are also nearly optimal.
1.2
Related Work
There is a vast literature on rank aggregation, and here we can only hope to cover a fraction of them
we see most relevant. In this paper, we study a statistical learning approach, assuming the observed
ranking data is generated from a probabilistic model. Various probabilistic models on permutations
have been studied in the ranking literature (see, e.g., [11, 12]). A nonparametric approach to modeling distributions over rankings using sparse representations has been studied in [13]. Most of the
parametric models fall into one of the following three categories: noisy comparison model, distance
based model, and random utility model. The noisy comparison model assumes that there is an underlying true ranking over n items, and each user independently gives a pairwise comparison which
agrees with the true ranking with probability p > 1/2. It is shown in [14] that O(n log n) pairwise
comparisons, when chosen adaptively, are sufficient for accurately estimating the true ranking.
The Mallows model is a distance-based model, which randomly generates a full ranking ? over n
?
items from some underlying true ranking ? ? with probability proportional to e??d(?,? ) , where ? is
a fixed spread parameter and d(?, ?) can be any permutation distance such as the Kemeny distance.
It is shown in [14] that the true ranking ? ? can be estimated accurately given O(log n) independent
full rankings generated under the Mallows model with the Kemeny distance.
In this paper, we study a special case of random utility models (RUMs) known as the Plackett-Luce
(PL) model. It is shown in [7] that the likelihood function under the PL model is concave and the
ML estimator can be efficiently found using a minorization-maximization (MM) algorithm which is
3
a variation of the general EM algorithm. We give an upper bound on the error achieved by such
an ML estimator, and prove that this is matched by a lower bound. The lower bound is derived by
comparing to an oracle estimator which observes the random utilities of RUM directly. The BradleyTerry (BT) model is the special case of the PL model where we only observe pairwise comparisons.
For the BT model, [10] proposes RankCentrality algorithm based on the stationary distribution of a
random walk over a suitably defined comparison graph and shows ?(npoly(log n)) randomly chosen
pairwise comparisons are sufficient to accurately estimate the underlying parameters; one corollary
of our result is a matching performance guarantee for the ML estimator under the BT model. More
recently, [15] analyzed various algorithms including RankCentrality and the ML estimator under a
general, not necessarily uniform, sampling scheme.
In a PL model with priors, MAP inference becomes computationally challenging. Instead, an efficient message-passing algorithm is proposed in [16] to approximate the MAP estimate. For a more
general family of random utility models, Soufiani et al. in [17, 18] give a sufficient condition under
which the likelihood function is concave, and propose a Monte-Carlo EM algorithm to compute the
ML estimator for general RUMs. More recently in [8, 9], the generalized method of moments together with the rank-breaking is applied to estimate the parameters of the PL model and the random
utility model when the data consists of full rankings.
2
Main results
In this section, we present our theoretical findings and numerical experiments.
2.1
Oracle lower bound
In this section, we derive an oracle lower bound for any estimator of ?? . The lower bound is constructed by considering an oracle who reveals all the hidden scores in the PL model as side information and holds for the general Thurstone models.
Theorem 1. Suppose ?1m are generated from the Thurstone model for some CDF F. For any estib
mator ?,
n
X
1
1
(n ? 1)2
1
? 2
b
inf sup E[||? ? ? ||2 ] ?
,
?
2
2
di
mk
2I(?) + b2 (d2?
2I(?) + b2 (d2?
?b ? ? ??b
1 +d2 ) i=2
1 +d2 )
R (?0 (x))2
where ? is the probability density function of F , i.e., ? = F 0 and I(?) =
?(x) dx; the second
inequality follows from the Jensen?s inequality. For the PL model, which is a special case of the
Thurstone models with F being the standard Gumbel distribution, I(?) = 1.
Pn
Theorem 1 shows that the oracle lower bound scales as i=2 d1i . We remark that the summation
begins with 1/d2 . This makes some sense, in view of the fact that the parameters ?i? need to sum
to zero. For example, if d1 is a moderate value and all the other di ?s are very large, then with the
hidden scores as side information, we may be able to accurately estimate ?i? for i 6= 1 and therefore
accurately estimate ?1? . The oracle lower bound also depends on the dynamic range b and is tight for
b = 0, because a trivial estimator that always outputs the all-zero vector achieves the lower bound.
Comparison to previous work Theorem 1 implies that mk = ?(n) is necessary for any item
assignment scheme to reliably infer ?? , i.e., ensuring E[||?b? ?? ||22 ] = o(n). It provides the first converse result on inferring the parameter vector under the general Thurstone models to our knowledge.
For the Bradley-Terry model, which is a special case of the PL model where all the partial rankings
reduce to the pairwise comparisons, i.e., k = 2, it is shown in [10] that m = ?(n) is necessary
for the random item assignment scheme to achieve the reliable inference based on the informationtheoretic argument. In contrast, our converse result is derived based on the Bayesian Cram?e-Rao
lower bound [19], applies to the general models with any item assignment, and is considerably
tighter if di ?s are of different orders.
2.2
Cram?er-Rao lower bound
In this section, we derive the Cram?er-Rao lower bound for any unbiased estimator of ?? .
4
Theorem 2. Let kmax = maxj?[m] kj and U denote the set of all unbiased estimators of ?? , i.e.,
b ? = ?] = ?, ?? ? ?b . If b > 0, then
?b ? U if and only if E[?|?
!?1 n
!?1
kX
kX
max
max
X 1
1
1
1
1
(n ? 1)2
inf sup E[k?b ? ?? k22 ] ? 1 ?
? 1?
,
b
kmax
`
?
kmax
`
mk
? ? ??b
??U
i=2 i
`=1
`=1
where the second inequality follows from the Jensen?s inequality.
Pn
The Cram?er-Rao lower bound scales as i=2 ?1i . When G is disconnected, i.e., all the items can be
partitioned into two groups such that no user ever compares an item in one group with an item in
the other group, ?2 = 0 and the Cram?er-Rao lower bound is infinity, which is valid (and of course
tight) because there is no basis for gauging any item in one connected component with respect to any
item in the other connected component and the accurate inference is impossible for any estimator.
Although the Cram?er-Rao lower bound only holds for any unbiased estimator, we suspect that a
lower bound with the same scaling holds for any estimator, but we do not have a proof.
2.3
ML upper bound
In this section, we study the ML estimator based on the partial rankings. The ML estimator of ?? is
defined as ?bML ? arg max???b L(?), where L(?) is the log likelihood function given by
L(?) =
log P? [?1m ]
j ?1
m kX
X
??j (`) ? log exp(??j (`) ) + ? ? ? + exp(??j (kj ) ) .
=
(1)
j=1 `=1
As observed in [7], L(?) is concave in ? and thus the ML estimator can be efficiently computed
either via the gradient descent method or the EM type algorithms.
The following theorem gives an upper bound on the error rates inversely dependent on ?2 . Intuitively, by the well-known Cheeger?s inequality, if the spectral gap ?2 becomes larger, then there are
more edges across any bi-partition of G, meaning more pairwise comparisons are available between
any bi-partition of movies, and therefore ?? can be estimated more accurately.
Theorem 3. Assume ?n ? C log n for a sufficiently large constant C in the case with k > 2. Then
with high probability,
(
?
4(1 + e2b )2 ??1
2? m log n If k = 2,
?
b
k?ML ? ? k2 ?
8e4b 2mk
? log n
If k > 2.
? ?16e2b ? log n
2
n
We compare
the above upper bound with the Cram?er-RaoP
lower bound given by Theorem 2. Notice
Pn
n
1
that i=1 ?i = mk and ?1 = 0. Therefore, mk
?
2
i=2 ?i and the upper bound is always
?2
larger than the Cram?er-Rao lower bound. When the comparison graph G is an expander and mk =
?(n log n), by the well-known Cheeger?s inequality, ?2 ? ?n = ?(log n) , the upper bound is only
larger than the Cram?er-Rao lower bound by a logarithmic factor. In particular, with the random item
assignment scheme, we show that ?2 , ?n ? mk
n if mk ? C log n and as a corollary of Theorem 3,
?
b
mk = ?(n log n) is sufficient to ensure k?ML ??? k2 = o( n), proving the random item assignment
scheme with the ML estimation is minimax-optimal up to a log n factor.
Corollary 1. Suppose S1m are chosen independently and uniformly at random among all possible
subsets of [n]. Then there exists a positive constant C > 0 such that if m ? Cn log n when k = 2
and mk ? Ce2b log n when k > 2, then with high probability
?
q
? 4(1 + e2b )2 n2 log n ,
if k = 2,
m
q
k?bML ? ?? k2 ?
2
log n
?
32e4b 2n mk
,
if k > 2.
Comparison to previous work Theorem 3 provides the first finite-sample error rates for inferring
the parameter vector under the PL model to our knowledge. For the Bradley-Terry model, which
is a special case of the PL model with k = 2, [10] derived the similar performance guarantee by
analyzing the rank centrality algorithm and the ML estimator. More recently, [15] extended the
results to the non-uniform sampling scheme of item pairs, but the performance guarantees obtained
when specialized
to the uniform sampling scheme require at least m = ?(n4 log n) to ensure k?b ?
?
?
? k2 = o( n), while our results only require m = ?(n log n).
5
2.4
Rank breaking upper bound
In this section, we study two rank-breaking schemes which decompose partial rankings into pairwise
comparisons.
Definition 4. Given a partial ranking ? over the subset S ? [n] of size k, the independencepreserving breaking scheme (IB) breaks ? into bk/2c non-intersecting pairwise comparisons of form
bk/2c
{it , i0t , yt }t=1 such that {is , i0s } ? {it , i0t } = ? for any s 6= t and yt = 1 if ? ?1 (it ) < ? ?1 (i0t ) and
bk/2c
0 otherwise. The random IB chooses {it , i0t }t=1 uniformly at random among all possibilities.
If ? is generated under the PL model, then the IB breaks ? into independent pairwise comparisons
generated under the PL model. Hence, we can first break partial rankings ?1m into independent pairwise comparisons using the random IB and then apply the ML estimator on the generated pairwise
comparisons with the constraint that ? ? ?b , denoted by ?bIB . Under the random assignment scheme,
?
as a corollary of Theorem 3, mk = ?(n log n) is sufficient to ensure k?bIB ? ?? k2 = o( n), proving
the random item assignment scheme with the random IB is minimax-optimal up to a log n factor in
view of the oracle lower bound in Theorem 1.
Corollary 2. Suppose S1m are chosen independently and uniformly at random among all possible
subsets of [n] with size k. There exists a positive constant C > 0 such that if mk ? Cn log n, then
with high probability,
r
2n2 log n
?
2b 2
b
k?IB ? ? k2 ? 4(1 + e )
.
mk
Definition 5. Given a partial ranking ? over the subset S ? [n] of size k, the full breaking scheme
(k2)
such that yt = 1 if
(FB) breaks ? into all k2 possible pairwise comparisons of form {it , i0t , yt }t=1
? ?1 (it ) < ? ?1 (i0t ) and 0 otherwise.
If ? is generated under the PL model, then the FB breaks ? into pairwise comparisons which are not
independently generated under the PL model. We pretend the pairwise comparisons induced from
the full breaking are all independent and maximize the weighted log likelihood function given by
m
X
X
1
L(?) =
?i I{??1 (i)<??1 (i0 )} + ?i0 I{??1 (i)>??1 (i0 )} ? log e?i + e?i0
j
j
j
j
2(kj ? 1) 0
j=1
i,i ?Sj
(2)
with the constraint that ? ? ?b . Let ?bFB denote the maximizer. Notice that we put the weight kj1?1
to adjust the contributions of the pairwise comparisons generated from the partial rankings over
subsets with different sizes.
?
Theorem 4. With high probability, k?bFB ? ?? k2 ? 2(1 + e2b )2 mk?2log n . Furthermore, suppose S1m
are chosen independently and uniformly at random among all possible subsets of [n]. There exists
a positive constant
C > 0 such that if mk ? Cn log n, then with high probability, k?bFB ? ?? k2 ?
q
4(1 + e2b )2
n2 log n
mk .
Theorem 4 shows that the error rates of ?bFB inversely depend on ?2 . When the comparison graph G
is an expander, i.e., ?2 ? ?n , the upper bound is only larger than the Cram?er-Rao lower bound by a
logarithmic factor. The similar observation holds for the ML estimator as shown in Theorem 3. With
the random item assignment scheme, Theorem 4 imply that the FB only need mk = ?(n log n)
to achieve the reliable inference, which is optimal up to a log n factor in view of the oracle lower
bound in Theorem 1.
Comparison to previous work The rank breaking schemes considered in [8, 9] breaks the full
rankings according to rank positions while our schemes break the partial rankings according to the
item indices. The results in [8, 9] establish the consistency of the generalized method of moments
under the rank breaking schemes when the data consists of full rankings. In contrast, Corollary 2 and
Theorem 4 apply to the more general setting with partial rankings and provide the finite-sample error
rates, proving the optimality of the random IB and FB with the random item assignment scheme.
6
2.5
Numerical experiments
Suppose there are n = 1024 items and ?? is uniformly distributed over [?b, b]. We first generate
d full rankings over 1024 items according to the PL model with parameter ?? . Then for each fixed
k ? {512, 256, . . . , 2}, we break every full ranking ? into n/k partial rankings over subsets of
n/k
size k as follows: Let {Sj }j=1 denote a partition of [n] generated uniformly at random such that
n/k
Sj ? Sj 0 = ? for j 6= j 0 and |Sj | = k for all j; generate {?j }j=1 such that ?j is the partial ranking
over set Sj consistent with ?. In this way, in total we get m = dn/k k-way comparisons which
are all independently generated from the PL model. We apply the minorization-maximization (MM)
algorithm proposed in [7] to compute the ML estimator ?bML based on the k-way comparisons and
the estimator ?bFB based on the pairwise comparisons induced by the
error is
FB. The estimation
mk b
? 2
measured by the rescaled mean square error (MSE) defined by log2 n2 k? ? ? k2 .
We run the simulation with b = 2 and d = 16, 64. The results are depicted in Fig. 1. We also plot
?1
Pk
as per Theorem 2. The oracle lower
the Cram?er-Rao (CR) limit given by log2 1 ? k1 l=1 1l
bound in Theorem 1 implies that the rescaled MSE is at least 0. We can see that the rescaled MSE of
the ML estimator ?bML is close to the CR limit and approaches the oracle lower bound as k becomes
large, suggesting the ML estimator is minimax-optimal. Furthermore, the rescaled MSE of ?bFB under
FB is approximately twice larger than the CR limit, suggesting that the FB is minimax-optimal up
to a constant factor.
3
FB (d=16)
FB (d=64)
d=16
d=64
CR Limit
Rescaled MSE
2.5
2
1.5
1
0.5
0
1
2
3
4
5
6
7
8
9
10
log2(k)
Figure 1: The error rate based on nd/k k-way comparisons with and without full breaking.
Finally, we point out that when d = 16 and log2 (k) = 1, the MSE returned by the MM algorithm
is infinity. Such singularity occurs for the following reason. Suppose we consider a directed comparison graph with nodes corresponding to items such that for each (i, j), there is a directed edge
(i ? j) if item i is ever ranked higher than j. If the graph is not strongly connected, i.e., if there
exists a partition of the items into two groups A and B such that items in A are always ranked higher
than items in B, then if all {?i : i ? A} are increased by a positive constant a, and all {?i : i ? B}
are decreased by another positive constant a0 such that all {?i , i ? [n]} still sum up to zero, the log
likelihood (1) must increase; thus, the log likelihood has no maximizer over the parameter space
?, and the MSE returned by the MM algorithm will diverge. Theoretically, if b is a constant and
d exceeds the order of log n, the directed comparison graph will be strongly connected with high
probability and so such singularity does not occur in our numerical experiments when d ? 64. In
practice we can deal with this singularity issue in three ways: 1) find the strongly connected components and then run MM in each component to come up with an estimator of ?? restricted to each
component; 2) introduce a proper prior on the parameters and use Bayesian inference to come up
with an estimator (see [16]); 3) add to the log likelihood objective function a regularization term
based on k?k2 and solve the regularized ML using the gradient descent algorithms (see [10]).
7
3
Proofs
We sketch the proof of our two upper bounds given by Theorem 3 and Theorem 4. The proofs of
other results can be found in the supplementary file. We introduce some additional notations used
in the proof. For a vector x, let kxk2 denote the usual l2 norm. Let 1 denote the all-one vector
and 0 denote the all-zero vector with the appropriate dimension. Let S n denote the set of n ? n
symmetric matrices with real-valued entries. For X ? S n , let ?1P
(X) ? ?2 (X) ? ? ? ? ? ?n (X)
n
denote its eigenvalues sorted in increasing order. Let Tr(X) = i=1 ?i (X) denote its trace and
kXk = max{??1 (X), ?n (X)} denote its spectral norm. For two matrices X, Y ? S n , we write
X ? Y if Y ?X is positive semi-definite, i.e., ?1 (Y ?X) ? 0. Recall that L(?) is the log likelihood
function. Let ?L(?) denote its gradient and H(?) ? S n denote its Hessian matrix.
3.1
Proof of Theorem 3
The main idea of the proof is inspired from the proof of [10, Theorem 4]. We first introduce several
key auxiliary results used in the proof. Observe that E?? [?L(?? )] = 0. The following lemma upper
bounds the deviation of ?L(?? ) from its mean.
Lemma 1. With probability at least 1 ?
2e2
n ,
?
k?L(? )k2 ?
p
2mk log n
(3)
Observed that ?H(?) is positive semi-definite with the smallest eigenvalue equal to zero. The
following lemma lower bounds its second smallest eigenvalue.
Lemma 2. Fix any ? ? ?b . Then
(
e2b
?2 If k = 2,
(1+e2b )2
(4)
?2 (?H(?)) ?
?
1
2b
?2 ? 16e
?n log n If k > 2,
4e4b
where the inequality holds with probability at least 1 ? n?1 in the case with k > 2.
Proof of Theorem 3. Define ? = ?bML ? ?? . It follows from the definition that ? is orthogonal to
the all-one vector. By the definition of the ML estimator, L(?bML ) ? L(?? ) and thus
L(??ML ) ? L(?? ) ? h?L(?? ), ?i ? ?h?L(?? ), ?i ? ?k?L(?? )k2 k?k2 ,
(5)
where the last inequality holds due to the Cauchy-Schwartz inequality. By the Taylor expansion,
there exists a ? = a?bML + (1 ? a)?? for some a ? [0, 1] such that
1
1
L(??ML ) ? L(?? ) ? h?L(?? ), ?i = ?> H(?)? ? ? ?2 (?H(?))k?k22 ,
(6)
2
2
where the last inequality holds because the Hessian matrix ?H(?) is positive semi-definite with
H(?)1 = 0 and ?> 1 = 0. Combining (5) and (6),
k?k2 ? 2k?L(?? )k2 /?2 (?H(?)).
(7)
Note that ? ? ?b by definition. The theorem follows by Lemma 1 and Lemma 2.
3.2
Proof of Theorem 4
It follows from the definition of L(?) given by (2) that
X
X
1
exp(?i? )
?
,
?i L(? ) =
I{??1 (i)<??1 (i0 )} ?
j
j
kj ? 1 0
exp(?i? ) + exp(?i?0 )
0
j:i?Sj
(8)
i ?Sj :i 6=i
which is a sum of di independent ?
random variables with mean zero and bounded by 1. By Ho?
effding?s inequality,
|?
L(?
)|
?
di log n with probability at least 1 ? 2n?2 . By union bound,
i
?
?
k?L(? )k2 ? mk log n with probability at least 1 ? 2n?1 . The Hessian matrix is given by
m
X
X
1
exp(?i + ?i0 )
H(?) = ?
(ei ? ei0 )(ei ? ei0 )>
2.
2(kj ? 1) 0
[exp(?i ) + exp(?i0 )]
j=1
i,i ?Sj
exp(?i +?i0 )
If |?i | ? b, ?i ? [n], [exp(?
2 ?
i )+exp(?i0 )]
and the theorem follows from (7).
e2b
.
(1+e2b )2
8
It follows that ?H(?) ?
e2b
L
(1+e2b )2
for ? ? ?b
References
[1] M. E. Ben-Akiva and S. R. Lerman, Discrete choice analysis: theory and application to travel
demand. MIT press, 1985, vol. 9.
[2] P. M. Guadagni and J. D. Little, ?A logit model of brand choice calibrated on scanner data,?
Marketing science, vol. 2, no. 3, pp. 203?238, 1983.
[3] D. McFadden, ?Econometric models for probabilistic choice among products,? Journal of Business, vol. 53, no. 3, pp. S13?S29, 1980.
[4] P. Sham and D. Curtis, ?An extended transmission/disequilibrium test (TDT) for multi-allele
marker loci,? Annals of human genetics, vol. 59, no. 3, pp. 323?336, 1995.
[5] G. Simons and Y. Yao, ?Asymptotics when the number of parameters tends to infinity in the
Bradley-Terry model for paired comparisons,? The Annals of Statistics, vol. 27, no. 3, pp.
1041?1060, 1999.
[6] J. C. Duchi, L. Mackey, and M. I. Jordan, ?On the consistency of ranking algorithms,? in
Proceedings of the ICML Conference, Haifa, Israel, June 2010.
[7] D. R. Hunter, ?MM algorithms for generalized Bradley-Terry models,? The Annals of Statistics, vol. 32, no. 1, pp. 384?406, 02 2004.
[8] H. A. Soufiani, W. Chen, D. C. Parkes, and L. Xia, ?Generalized method-of-moments for rank
aggregation,? in Advances in Neural Information Processing Systems 26, 2013, pp. 2706?2714.
[9] H. Azari Soufiani, D. Parkes, and L. Xia, ?Computing parametric ranking models via rankbreaking,? in Proceedings of the International Conference on Machine Learning, 2014.
[10] S. Negahban, S. Oh, and D. Shah, ?Rank centrality: Ranking from pair-wise comparisons,?
arXiv:1209.1688, 2012.
[11] T. Qin, X. Geng, and T. yan Liu, ?A new probabilistic model for rank aggregation,? in Advances
in Neural Information Processing Systems 23, 2010, pp. 1948?1956.
[12] J. A. Lozano and E. Irurozki, ?Probabilistic modeling on rankings,? Available at http:// www.
sc.ehu.es/ ccwbayes/ members/ ekhine/ tutorial ranking/ info.html, 2012.
[13] S. Jagabathula and D. Shah, ?Inferring rankings under constrained sensing.? in NIPS, vol. 2008,
2008.
[14] M. Braverman and E. Mossel, ?Sorting from noisy information,? arXiv:0910.1191, 2009.
[15] A. Rajkumar and S. Agarwal, ?A statistical convergence perspective of algorithms for rank
aggregation from pairwise data,? in Proceedings of the International Conference on Machine
Learning, 2014.
[16] J. Guiver and E. Snelson, ?Bayesian inference for Plackett-Luce ranking models,? in Proceedings of the 26th Annual International Conference on Machine Learning, New York, NY, USA,
2009, pp. 377?384.
[17] A. S. Hossein, D. C. Parkes, and L. Xia, ?Random utility theory for social choice,? in Proceeedings of the 25th Annual Conference on Neural Information Processing Systems, 2012.
[18] H. A. Soufiani, D. C. Parkes, and L. Xia, ?Preference elicitation for general random utility
models,? arXiv preprint arXiv:1309.6864, 2013.
[19] R. D. Gill and B. Y. Levit, ?Applications of the van Trees inequality: a Bayesian Cram?er-Rao
bound,? Bernoulli, vol. 1, no. 1-2, pp. 59?79, 03 1995.
9
| 5361 |@word norm:2 nd:1 suitably:1 logit:1 d2:6 simulation:1 bn:3 tr:1 harder:1 moment:3 liu:1 score:3 e2b:11 bradley:4 comparing:1 dx:1 must:1 numerical:3 partition:4 plot:1 mackey:1 stationary:1 fewer:1 item:60 parkes:4 provides:3 node:1 preference:16 minorization:2 dn:2 c2:2 constructed:1 prove:2 consists:2 introduce:3 manner:1 theoretically:1 pairwise:23 indeed:1 uiuc:3 growing:1 multi:1 inspired:1 election:1 little:2 cardinality:1 considering:1 becomes:6 provided:1 estimating:2 notation:3 underlying:8 moreover:1 matched:1 npoly:1 increasing:1 israel:1 begin:1 bounded:1 pursue:1 eigenvector:1 unobserved:2 finding:1 guarantee:3 every:1 concave:3 scaled:1 k2:18 schwartz:1 control:2 converse:2 positive:11 tends:1 limit:4 analyzing:1 approximately:1 twice:1 studied:3 equivalence:2 suggests:1 challenging:2 limited:1 range:2 statistically:1 bi:2 directed:3 practical:2 unique:1 mallow:2 practice:2 union:1 definite:4 asymptotics:1 yan:1 matching:2 cram:15 get:2 close:2 put:2 context:1 kmax:3 impossible:1 www:1 equivalent:1 deterministic:1 map:2 transportation:1 yt:4 attention:1 independently:10 bfb:6 guiver:1 estimator:31 oh:2 population:1 proving:4 thurstone:9 variation:1 annals:3 target:2 suppose:6 user:16 element:1 rajkumar:1 observed:6 preprint:1 soufiani:4 connected:5 azari:1 ordering:1 rescaled:5 observes:1 cheeger:2 ui:3 rankcentrality:2 dynamic:2 depend:3 tight:2 basis:1 various:4 enyi:1 describe:1 monte:1 sc:1 larger:6 solve:1 supplementary:1 say:1 valued:1 otherwise:2 estib:1 statistic:3 noisy:3 sequence:2 eigenvalue:5 propose:2 product:1 qin:1 relevant:1 combining:1 achieve:4 convergence:1 transmission:1 generating:1 ben:1 derive:3 measured:1 received:1 auxiliary:1 judge:1 implies:2 quantify:1 come:2 bib:2 allele:1 human:1 adjacency:1 require:2 assign:4 fix:1 decompose:3 tighter:1 secondly:2 summation:1 singularity:3 pl:25 hold:9 mm:6 sufficiently:1 considered:1 scanner:1 exp:11 mapping:2 achieves:1 smallest:3 estimation:8 travel:1 applicable:1 maker:1 agrees:1 weighted:4 hope:1 mit:1 always:3 pn:4 cr:4 corollary:6 derived:4 focus:1 june:1 rank:25 likelihood:12 bernoulli:1 contrast:2 rigorous:1 sense:1 inference:9 plackett:3 dependent:2 i0:16 bt:3 a0:1 hidden:4 interested:1 issue:1 arg:1 among:5 html:1 denoted:1 hossein:1 proposes:1 art:1 special:6 constrained:1 equal:2 having:1 sampling:3 represents:1 icml:1 nearly:1 gauging:1 geng:1 few:2 modern:1 randomly:4 simultaneously:1 individual:2 maxj:1 tdt:1 interest:3 message:1 possibility:1 braverman:1 adjust:1 analyzed:1 d1i:1 accurate:1 edge:3 partial:26 necessary:3 orthogonal:1 unless:1 indexed:2 tree:1 taylor:1 logarithm:1 walk:1 haifa:1 plackettluce:1 theoretical:2 sociology:1 mk:29 increased:1 modeling:2 rao:14 cover:1 measuring:1 assignment:21 maximization:2 deviation:1 vertex:2 subset:13 entry:1 uniform:3 optimally:2 considerably:1 chooses:1 adaptively:1 calibrated:1 density:1 international:3 negahban:1 rankbreaking:2 probabilistic:5 discipline:1 diverge:1 together:2 yao:1 intersecting:2 squared:1 choose:1 suggesting:2 s13:1 b2:2 explicitly:1 ranking:50 depends:1 break:9 view:3 analyze:1 characterizes:1 sup:2 netflix:1 aggregation:8 recover:1 simon:1 bruce:1 levit:1 contribution:1 formed:1 square:1 accuracy:1 mator:1 who:3 efficiently:2 bayesian:4 accurately:6 hunter:1 carlo:1 sn1:1 submitted:1 definition:10 pp:9 e2:1 proof:11 di:10 sampled:1 dataset:1 popular:2 recall:1 knowledge:2 genie:1 hajek:2 higher:3 erd:1 strongly:3 generality:1 furthermore:2 marketing:2 until:1 sketch:1 ei:2 o:1 maximizer:2 marker:1 bml:7 usa:1 k22:2 normalized:1 true:5 unbiased:3 lozano:1 hence:1 assigned:5 regularization:1 symmetric:1 deal:2 indistinguishable:1 generalized:4 duchi:1 meaning:2 wise:1 snelson:1 recently:3 common:1 specialized:2 exponentially:2 thirdly:1 swoh:1 consistency:3 pm:1 illinois:3 base:1 add:1 perspective:1 inf:2 moderate:1 scenario:2 inequality:12 continue:1 arbitrarily:1 assumep:1 preserving:1 additional:1 gill:1 maximize:1 semi:4 multiple:3 full:12 sham:1 infer:4 exceeds:1 match:1 retrieval:1 a1:2 laplacian:2 ensuring:1 paired:1 arxiv:4 agarwal:1 jiaming:1 achieved:4 c1:2 decreased:1 else:1 appropriately:1 file:1 suspect:1 induced:2 expander:4 undirected:1 member:1 s29:1 jordan:1 integer:1 easy:1 independence:2 psychology:1 opposite:1 reduce:1 idea:1 cn:3 luce:3 utility:10 returned:2 passing:1 hessian:3 york:1 remark:1 generally:1 nonparametric:1 category:1 generate:3 http:1 tutorial:1 notice:2 estimated:2 disequilibrium:1 per:2 write:1 discrete:1 vol:8 group:4 key:1 econometric:1 vast:1 graph:15 fraction:1 sum:3 run:2 inverse:1 family:1 decision:1 scaling:1 bound:50 oracle:12 annual:2 occur:1 precisely:1 constraint:3 infinity:3 generates:2 argument:2 optimality:1 according:5 disconnected:1 across:2 em:3 partitioned:1 n4:1 s1:1 ei0:2 intuitively:2 restricted:1 jagabathula:1 computationally:1 needed:2 i0t:6 locus:1 ascending:1 available:3 apply:4 observe:3 spectral:4 appropriate:1 centrality:2 ho:1 shah:2 assumes:2 ensure:3 log2:4 pretend:1 kj1:2 k1:1 establish:1 objective:1 question:2 occurs:1 degrades:1 parametric:2 usual:1 diagonal:1 kemeny:2 gradient:3 distance:5 cauchy:1 consensus:1 trivial:1 reason:1 assuming:1 index:1 equivalently:3 setup:1 proceeedings:1 info:1 trace:1 reliably:2 proper:1 upper:17 recommender:2 observation:3 datasets:1 finite:2 descent:2 extended:2 ever:2 rn:1 rating:1 bk:4 pair:3 specified:1 s1m:3 nip:1 address:1 able:1 elicitation:1 challenge:1 including:3 reliable:2 max:4 everyone:1 terry:4 event:1 natural:2 ranked:3 regularized:1 business:1 minimax:7 scheme:24 movie:3 rated:1 mossel:1 inversely:3 imply:3 numerous:1 sn:1 kj:8 prior:2 literature:2 l2:1 loss:1 permutation:3 mcfadden:1 proportional:1 versus:1 degree:1 sufficient:6 consistent:1 sewoong:1 course:1 summary:1 genetics:1 last:2 aij:1 side:2 understand:1 mismatched:1 fall:1 absolute:1 sparse:1 distributed:3 van:1 xia:4 dimension:1 rum:4 valid:1 fb:9 social:1 sj:16 approximate:1 informationtheoretic:1 ml:24 global:1 reveals:1 xi:3 decomposes:1 curtis:1 expansion:1 mse:7 necessarily:1 pk:1 main:4 spread:1 big:1 expanders:1 n2:4 allowed:1 xu:1 fig:1 fashion:1 ny:1 inferring:3 position:1 deterministically:1 exponential:1 kxk2:1 breaking:15 ib:7 theorem:32 er:14 jensen:2 sensing:1 exists:6 sequential:1 budget:1 kx:3 gumbel:2 gap:3 demand:1 chen:1 sorting:1 depicted:1 logarithmic:4 kxk:1 sport:1 applies:2 collectively:1 corresponds:1 cdf:6 goal:2 sorted:1 uniformly:8 lemma:6 total:2 e:1 lerman:1 brand:1 akiva:1 select:4 arises:1 d1:2 crowdsourcing:1 |
4,818 | 5,362 | Efficient Optimization for Average Precision SVM
Pritish Mohapatra
IIIT Hyderabad
[email protected]
C.V. Jawahar
IIIT Hyderabad
[email protected]
M. Pawan Kumar
Ecole Centrale Paris & INRIA Saclay
[email protected]
Abstract
The accuracy of information retrieval systems is often measured using average
precision (AP). Given a set of positive (relevant) and negative (non-relevant) samples, the parameters of a retrieval system can be estimated using the AP - SVM
framework, which minimizes a regularized convex upper bound on the empirical AP loss. However, the high computational complexity of loss-augmented inference, which is required for learning an AP - SVM, prohibits its use with large
training datasets. To alleviate this deficiency, we propose three complementary
approaches. The first approach guarantees an asymptotic decrease in the computational complexity of loss-augmented inference by exploiting the problem structure. The second approach takes advantage of the fact that we do not require a
full ranking during loss-augmented inference. This helps us to avoid the expensive step of sorting the negative samples according to their individual scores. The
third approach approximates the AP loss over all samples by the AP loss over difficult samples (for example, those that are incorrectly classified by a binary SVM),
while ensuring the correct classification of the remaining samples. Using the PAS CAL VOC action classification and object detection datasets, we show that our
approaches provide significant speed-ups during training without degrading the
test accuracy of AP - SVM.
1
Introduction
Information retrieval systems require us to rank a set of samples according to their relevance to a
query. The parameters of a retrieval system can be estimated by minimizing the prediction risk on a
training dataset, which consists of positive and negative samples. Here, positive samples are those
that are relevant to a query, and negative samples are those that are not relevant to the query. Several
risk minimization frameworks have been proposed in the literature, including structured support
vector machines (SSVM) [15, 16], neural networks [14], decision forests [11] and boosting [13]. In
this work, we focus on SSVMs for clarity while noting the methods we develop are also applicable
to other learning frameworks.
The SSVM framework provides a linear prediction rule to obtain a structured output for a structured
input. Specifically, the score of a putative output is the dot product of the parameters of an SSVM
with the joint feature vector of the input and the output. The prediction requires us to maximize
the score over all possible outputs for an input. During training, the parameters of an SSVM are
estimated by minimizing a regularized convex upper bound on a user-specified loss function. The
loss function measures the prediction risk, and should be chosen according to the evaluation criterion
for the system. While in theory the SSVM framework can be employed in conjunction with any loss
function, in practice its feasibility depends on the computational efficiency of the corresponding lossaugmented inference. In other words, given the current estimate of the parameters, it is important to
be able to efficiently maximize the sum of the score and the loss function over all possible outputs.
1
A common measure of accuracy for information retrieval is average precision (AP), which is used
in several standard challenges such as the PASCAL VOC object detection, image classification and
action classification tasks [7], and the TREC Web Track corpora. The popularity of AP inspired Yue
et al. [19] to propose the AP - SVM framework, which is a special case of SSVM. The input of AP SVM is a set of samples, the output is a ranking and the loss function is one minus the AP of the
ranking. In order to learn the parameters of an AP - SVM, Yue et al. [19] developed an optimal greedy
algorithm for loss-augmented inference. Their algorithm consists of two stages. First, it sorts the
positive samples P and the negative samples N separately in descending order of their individual
scores. The individual score of a sample is equal to the dot product of the parameters with the feature
vector of the sample. Second, starting from the negative sample with the highest score, it iteratively
finds the optimal interleaving rank for each of the |N | negative samples. The interleaving rank for a
negative sample is the index of the highest ranked positive sample ranked below it. which requires
at most O(|P|) time per iteration. The overall algorithm is described in detail in the next section.
Note that, typically |N | |P|, that is, the negative samples significantly outnumber the positive
samples.
While the AP - SVM has been successfully applied for ranking using high-order information in mid to
large size datasets [5], many methods continue to use the simpler binary SVM framework for large
datasets. Unlike AP - SVM, a binary SVM optimizes the surrogate 0-1 loss. Its main advantage is
the efficiency of the corresponding loss-augmented inference algorithm, which has a complexity of
O(|P| + |N |). However, this gain in training efficiency often comes at the cost of a loss in testing
accuracy, which is especially significant when training with weakly supervised datasets [1].
In order to facilitate the use of AP - SVM, we present three complementary approaches to speed-up
its learning. Our first approach exploits an interesting structure in the problem corresponding to the
computation of the rank of the j-th negative sample. Specifically, we show that when j > |P|, the
rank of the j-th negative sample is obtained by maximizing a discrete unimodal function. Here, a
discrete function defined over points {1, ? ? ? , p} is said to be unimodal if it is non-decreasing from
{1, ? ? ? , k} and non-increasing from {k, ? ? ? , p} for some k ? {1, ? ? ? , p}. Since the mode of a
discrete unimodal function can be computed efficiently using binary search, it reduces the computational complexity of computing the rank of the j-th negative sample from O(|P|) to O(log(|P|)). To
the best of our knowledge, ours is the first work to improve the speed of loss-augmented inference
for AP - SVM by taking advantage of the special structure of the problem. Unlike [2] which proposes
an efficient method for a similar framework of structured output ranking, our method optimizes the
APloss.
Our second approach relies on the fact that in many cases we do not need to explicitly compute the
optimal interleaving rank for all the negative samples. Specifically, we only need to compute the
interleaving rank for the set of negative samples that would have an interleaving rank of less than
|P| + 1. We identify this set using a binary search over the list of negative samples. While training,
after the initial few training iterations the size of this set rapidly reduces, allowing us to significantly
reduce the training time in practice.
Our third approach uses the intuition that the 0-1 loss and the AP loss differ only when some of the
samples are difficult to classify (that is, some positive samples that can be confused as negatives and
vice versa). In other words, when the 0-1 loss over the training dataset is 0, then the AP loss is also
0. Thus, instead of optimizing the AP loss over all the samples, we adopt a two-stage approximate
strategy. In the first stage, we identify a subset of difficult samples (specifically, those that are
incorrectly classified by a binary SVM). In the second stage, we optimize the AP loss over the subset
of difficult samples, while ensuring the correct classification of the remaining easy samples. Using
the PASCAL VOC action classification and object detection datasets, we empirically demonstrate that
each of our approaches greatly reduces the training time of AP - SVM while not decreasing the testing
accuracy.
2
The AP-SVM Framework
We provide a brief overview of the AP - SVM framework, highlighting only those aspects that are
necessary for the understanding of this paper. For a detailed description, we refer the reader to [19].
Input and Output. The input of an AP - SVM is a set of n samples, which we denote by X =
{xi , i = 1, ? ? ? , n}. Each sample can either belong to the positive class (that is, the sample is
2
relevant) or the negative class (that is, the sample is not relevant). The indices for the positive and
negative samples are denoted by P and N respectively. In other words, if i ? P and j ? N then xi
belongs to positive class and xj belongs to the negative class.
The desired output is a ranking matrix R of size n ? n, such that (i) Rij = 1 if xi is ranked higher
than xj ; (ii) Rij = ?1 if xi is ranked lower than xj ; and (iii) Rij = 0 if xi and xj are assigned
the same rank. During training, the ground-truth ranking matrix R? is defined as: (i) R?ij = 1 and
R?ji = ?1 for all i ? P and j ? N ; (ii) R?ii0 = 0 and R?jj 0 = 0 for all i, i0 ? P and j, j 0 ? N .
Joint Feature Vector. For a sample xi , let ?(xi ) denote its feature vector. The joint feature vector
of the input X and an output R is specified as
1 XX
?(X, R) =
Rij (?(xi ) ? ?(xj )).
(1)
|P||N |
i?P j?N
In other words, the joint feature vector is the scaled sum of the difference between the features of all
pairs of samples, where one sample is positive and the other is negative.
Parameters and Prediction. The parameter vector of AP - SVM is denoted by w, and is of the same
size as the joint feature vector. Given the parameters w, the ranking of an input X is predicted by
maximizing the score, that is,
R = argmax w> ?(X, R).
(2)
R
Yue et al. [19] showed that the above optimization can be performed efficiently by sorting the samples xk in descending order of their individual scores, that is, sk = w> ?(xk ).
Parameter Estimation. Given the input X and the ground-truth ranking matrix R? , we estimate
the AP - SVM parameters by optimizing a regularized upper bound on the empirical AP loss. The
AP loss of an output R is defined as 1 ? AP(R? , R), where AP(?, ?) corresponds to the AP of the
ranking R with respect to the true ranking R? . Specifically, the parameters are obtained by solving
the following convex optimization problem:
min
w
s.t.
1
||w||2 + C?,
2
w> ?(X, R? ) ? w> ?(X, R) ? ?(R? , R) ? ?, ?R
(3)
The computational complexity of solving the above problem depends on the complexity of the corresponding loss-augmented inference, that is,
? = argmax w> ?(X, R) + ?(R? , R).
R
(4)
R
For a given set of parameters w, the above problem requires us to find the most violated ranking,
that is, the ranking that maximizes the sum of the score and the AP loss. To be more precise, what
? and the AP loss ?(R? , R)
? corresponding to the most
we require is the joint feature vector ?(X, R)
violated ranking. Yue et al. [19] provided an optimal greedy algorithm for problem (4), which is
summarized in Algorithm 1. It consists of two stages. First, it sorts the positive and the negative
samples separately in descending order of their scores (steps 1-2). This takes O(|P| log(|P|) +
|N | log(|N |)) time. Second, starting with the highest scoring negative sample, it iteratively finds
the interleaving rank of each negative sample xj . This involves maximizing the quantity ?j (i),
defined in equation (5), over all i ? {1, ? ? ? , |P|} (steps 3-7), which takes O(|P||N |) time.
3
Efficient Optimization for AP-SVM
In this section, we propose three methods to speed up the training procedure of AP - SVM. The first
two methods are exact. Specifically, they reduce the time taken to perform loss-augmented inference
while ensuring the computation of the same most violated ranking as Algorithm 1. The third method
provides a framework for a sensible trade-off between training efficiency and test accuracy.
3.1 Efficient Search for Loss-Augmented Inference
In order to find the most violated ranking, the greedy algorithm of Yue et al. [19] iteratively computes the optimal interleaving rank optj ? {1, ? ? ? , |P| + 1} for each negative sample xj (step 5
3
Algorithm 1 The optimal greedy algorithm for loss-augmented inference for training AP - SVM.
input Training samples X containing positive samples P and negative samples N , parameters w.
p
1: Sort the positive samples in descending order of the scores si = w> ?(xi ), i ? {1, . . . , |P|}.
>
2: Sort the negative samples in descending order of the scores sn
j = w ?(xj ), j ? {1, . . . , |N |}.
3: Set j = 1.
4: repeat
5:
Compute the interleaving rank optj = argmaxi?{1,??? ,|P|} ?j (i), where
|P|
X
2(spk ? snj )
1
j
j?1
?
?
.
?j (i) =
|P| j + k j + k ? 1
|P||N |
(5)
k=i
The j-th negative sample is ranked between the (optj ? 1)-th and the optj -th positive sample.
6:
Set j ? j + 1.
7: until j > |N |.
of Algorithm 1). The interleaving rank optj specifies that the negative sample xj must be ranked
between the (optj ? 1)-th and the optj -th positive sample. The computation of the optimal interleaving rank for a particular negative sample requires us to maximize the discrete function ?j (i) over
the domain i ? {1, ? ? ? , |P|}. Yue et al. [19] use a simple linear algorithm for this step, which takes
O(|P|) time. In contrast, we propose a more efficient algorithm to maximize ?j (?), which exploits
the special structure of this discrete function.
Before we describe our efficient algorithm in detail, we require the definition of a unimodal function.
A discrete function f : {1, ? ? ? , p} ? R is said to be unimodal if and only if there exists a k ?
{1, ? ? ? , p} such that
f (i) ? f (i + 1), ?i ? {1, ? ? ? , k ? 1},
f (i ? 1) ? f (i), ?i ? {k + 1, ? ? ? , p}.
(6)
In other words, a unimodal discrete function is monotonically non-decreasing in the interval [1, k]
and monotonically non-increasing in the interval [k, p]. The maximization of a unimodal discrete
function over its domain {1, ? ? ? , p} simply requires us to find the index k that satisfies the above
properties. The maximization can be performed efficiently, in O(log(p)) time, using binary search.
We are now ready to state the main result that allows us to compute the optimal interleaving rank of
a negative sample efficiently.
Proposition 1. The discrete function ?j (i), defined in equation (5), is unimodal in the domain
{1, ? ? ? , p}, where p = min{|P|, j}.
The proof of the above proposition is provided in Appendix A (supplementary material).
Algorithm 2 Efficient search for the optimal interleaving rank of a negative sample.
input {?j (i), i = 1, ? ? ? , |P|}.
1: p = min{|P|, j}.
2: Compute an interleaving rank i1 as
ii = argmax ?j (i).
(7)
i?{1,??? ,p}
3: Compute an interleaving rank i2 as
i2 =
argmax
?j (i).
(8)
i?{p+1,??? ,|P|}
4: Compute the optimal interleaving rank optj as
optj =
i1
i2
if ?j (i1 ) ? ?j (i2 ),
otherwise.
4
(9)
Using the above proposition, the discrete function ?j (i) can be optimized over the domain
{1, ? ? ? , |P|} efficiently as described in Algorithm 2. Briefly, our efficient search algorithm finds
an interleaving ranking i1 over the domain {1, ? ? ? , p}, where p is set to min{|P|, j} in order to
ensure that the function ?j (?) is unimodal (step 2 of Algorithm 2). Since i1 can be computed using binary search, the computational complexity of this step is O(log(p)). Furthermore, we find an
interleaving ranking i2 over the domain {p + 1, ? ? ? , |P|} (step 3 of Algorithm 2). Since i2 needs
to be computed using linear search, the computational complexity of this step is O(|P| ? p) when
p < |P| and 0 otherwise. The optimal interleaving ranking optj of the negative sample xj can then
be computed by comparing the values of ?j (i1 ) and ?j (i2 ) (step 4 of Algorithm 2).
Note that, in a typical training dataset, the negative samples significantly outnumber the positive
samples, that is, |N | |P|. For all the negative samples xj where j ? |P|, p will be equal to |P|.
Hence, the maximization of ?j (?) can be performed efficiently over the entire domain {1, ? ? ? , |P|}
using binary search in O(log(|P|)) as opposed to the O(|P|) time suggested in [19].
3.2
Selective Ranking for Loss-Augmented Inference
While the efficient search algorithm described in the previous subsection allows us to find the optimal interleaving rank for a particular negative sample, the overall loss-augmented inference would
still remain computationally inefficient when the number of negative samples is large (as is typically the case). This is due to the following two reasons. First, loss-augmented inference spends a
considerable amount of time sorting the negative samples according to their individual scores (step
2 of Algorithm 1). Second, if we were to apply our efficient search algorithm to every negative
sample, the total computational complexity of the second stage of loss-augmented inference (step
3-7 of Algorithm 1) will still be O(|P|2 + (|N | ? |P|) log(|P|)).
In order to overcome the above computational issues, we exploit two key properties of lossaugmented inference in AP - SVM. First, if a negative sample xj has the optimal interleaving rank
optj = |P| + 1, then all the negative samples that have lower score than xj would also have the
same optimal interleaving rank (that is, optk = optj = |P| + 1 for all k > j). This property follows
directly from the analysis of Yue et al. [19] who showed that, for k < j, optk ? optj and for any
negative sample xj , optj ? [1, |P| + 1]. We refer the reader to [19] for a detailed proof. Second, we
? but the
note that the desired output of loss-augmented inference is not the most violated ranking R,
? and the AP loss AP(R? , R).
? From the definition of the joint feature
joint feature vector ?(X, R)
vector and the AP loss, it follows that they do not depend on the relative ranking of the negative
samples that share the same optimal interleaving rank. Specifically, both the joint feature vector
and the AP loss only depend on the number of negatives that are ranked higher and lower than each
positive sample.
The above two observations suggest the following alternate strategy to Algorithm 1. Instead of explicitly computing the optimal interleaving rank for each negative
sample (which can be computationally expensive), we
compute it only for negative samples that are expected
to have optimal interleaving rank less than |P| + 1. Algorithm 3 outlines the procedure we propose in detail. We
first find the score s? such that every negative sample xj
with score snj < s? has optj = |P| + 1. We do a binary
search over the list of scores of negative samples to find s?
(step 4 of algorithm 3). We do not need to sort the scores
of all the negative samples, as we use the quick select algorithm to find the k-th highest score wherever required. Figure 1: A row corresponds to the interleaving ranks of the negative samIf the output of the loss-augmented inference is such that
ples after a training iteration. Here,
a large number of negative samples have optimal interthere are 4703 negative samples, and
leaving rank as |P| + 1, then this alternate strategy would
131 training iterations. The interleavresult in a significant speed-up during training. In our
ing ranks are represented using a heat
experiments, we found that in later iterations of the opmap where the deepest red represents
timization, this is indeed the case in practice. Figure 1
interleaving rank of |P| + 1. (The figshows how the number of negative samples with optimal
ure is best viewed in colour.)
interleaving rank equal to |P| + 1, rapidly increases after
5
Algorithm 3 The selective ranking algorithm for loss-augmented inference in AP - SVM.
input S x , S x? , |P|, |N |
1: Sort the positive samples in descending order of their scores S x .
2: Do binarysearch over S x? to find s?.
3: Set Nl = j ? N |sn
?
j <s
4: Sort Nl in descending order of the scores.
5: for all j ? Nl do
6:
Compute optj using Algorithm 2.
7: end for
8: Set Nr = N ? Nl .
9: for all j ? Nr do
10:
Set optj = |P| + 1.
11: end for
output optj , ?j ? N
a few training iterations for a typical experiment. A large number of negative samples have optimal
interleaving rank equal to |P| + 1, while the negative samples that have other values of optimal
interleaving rank decrease considerably.
It would be worth taking note that here, even though we take advantage of the fact that a long
sequence of negative samples at the end of the list take the same optimal interleaving rank, such
sequences also occur at other locations throughout the list. This can be leveraged for further speedup by computing the interleaving rank for only the boundary samples of such sequences and setting
all the intermediate samples to the same interleaving rank as the boundary samples. We can use a
method similar to the one presented in this section to search for such sequences by using the quick
select algorithm to compute the interleaving rank for any particular negative sample on the list.
3.3
Efficient Approximation of AP-SVM
The previous two subsections provide exact algorithms for loss-augmented inference that reduce
the time require for training an AP - SVM. However, despite these improvements, AP - SVM might
be slower to learn compared to simpler frameworks such as the binary SVM, which optimizes the
surrogate 0-1 loss. The disadvantage of using the binary SVM is that, in general, the 0-1 loss is a
poor approximation for the AP loss. However, the quality of the approximation is not uniformly
poor for all samples, but depends heavily on their separability. Specifically, when the 0-1 loss of a
set of samples is 0 (that is, they are linearly separable by a binary SVM), their AP loss is also 0. This
observation inspires us to approximate the AP loss over the entire set of training samples using the
AP loss over the subset of difficult samples. In this work, we define the subset of difficult samples
as those that are incorrectly classified by a simple binary SVM.
Formally, given the complete input X and the ground-truth ranking matrix R? , we represent individual samples as xi and their class as yi . In other words, yi = 1 if i ? P and yi = ?1 if i ? N .
In order to approximate the AP - SVM, we adopt a two stage strategy. In the first stage, we learn
a binary SVM by minimizing the regularized convex upper bound on the 0-1 loss over the entire
training set. Since the loss-augmented inference for 0-1 loss is very fast, the parameters w0 of the
binary SVM can be estimated efficiently. We use the binary SVM to define the set of easy samples as
Xe = {xi , yi w0> ?i (x) ? 1}. In other words, a positive sample is easy if it is assigned a score that
is greater than 1 by the binary SVM. Similarly, a negative sample is easy if it is assigned a score that
is less than -1 by the binary SVM. The remaining difficult samples are denoted by Xd = X ? Xe
and the corresponding ground-truth ranking matrix by R?d . In the second stage, we approximate the
AP loss over the entire set of samples X by the AP loss over the difficult samples Xd while ensuring
that the samples Xe are correctly classified. In order to accomplish this, we solve the following
optimization problem:
min
w
s.t.
1
||w||2 + C?
2
w> ?(Xd , R?d ) ? w> ?(Xd , Rd ) ? ?(R?d , Rd ) ? ?, ?Rd ,
yi w> ?(xi ) > 1, ?xi ? Xe .
6
(10)
In practice, we can choose to retain only the top k% of Xe ranked in descending order of their score
and push the remaining samples into the difficult set Xd . This gives the AP - SVM more flexibility to
update the parameters at the cost of some additional computation.
4
Experiments
We demonstrate the efficacy of our methods, described in the previous section, on the challenging
problems of action classification and object detection.
4.1 Action Classification
Dataset. We use the PASCAL VOC 2011 [7] action classification dataset for our experiments. This
dataset consists of 4846 images, which include 10 different action classes. The dataset is divided
into two parts: 3347 ?trainval? person bounding boxes and 3363 ?test? person bounding boxes. We
use the ?trainval? bounding boxes for training since their ground-truth action classes are known.
We evaluate the accuracy of the different instances of SSVM on the ?test? bounding boxes using the
PASCAL evaluation server.
Features. We use the standard poselet [12] activation features to define the sample feature for
each person bounding box. The feature vector consists of 2400 action poselet activations and 4
object detection scores. We refer the reader to [12] for details regarding the feature vector.
Methods. We present results on five different methods. First, the standard binary SVM, which
optimizes the 0-1 loss. Second, the standard AP - SVM, which uses the inefficient loss-augmented
inference described in Algorithm 1. Third, AP - SVM - SEARCH, which uses efficient search to compute the optimal interleaving rank for each negative sample using Algorithm 2. Fourth, AP - SVM SELECT, which uses the selective ranking strategy outlined in Algorithm 3. Fifth, AP - SVM - APPX ,
which employs the approximate AP - SVM framework described in subsection 3.3. Note that, AP SVM , AP - SVM - SEARCH and AP - SVM - SELECT are guaranteed to provide the same set of parameters
since both efficient search and selective ranking are exact methods. The hyperparameters of all five
methods are fixed using 5-fold cross-validation on the ?trainval? set.
Results. Table 1 shows the AP for the rankings obtained by the five methods for ?test? set. Note that
AP - SVM (and therefore, AP - SVM - SEARCH and AP - SVM - SELECT) consistently outperforms binary
SVM by optimizing a more appropriate loss function during training. The approximate AP - SVM APPX provides comparable results to the exact AP - SVM formulations by optimizing the AP loss over
difficult samples, while ensuring the correct classification of easy samples. The time required to
compute the most violated rankings for each of the five methods in shown in Table 2. Note that
all three methods described in this paper result in substantial improvement in training time. The
overall time required for loss-augmented inference is reduced by a factor of 5 ? 10 compared to the
original AP - SVM approach. It can also be observed that though each loss-augmented inference step
for binary SVM is significantly more efficient than for AP - SVM (Table 3), in some cases we observe
that we required more cutting plane iterations for binary SVM to converge. As a result, in some cases
training binary SVM is slower than training AP - SVM with our proposed speed-ups.
Object class
Jumping
Phoning
Playing instrument
Reading
Riding bike
Running
Taking photo
Using computer
Walking
Riding horse
Binary SVM
52.580
32.090
35.210
27.410
72.240
73.090
21.880
30.620
54.400
79.820
AP - SVM
55.230
32.630
41.180
26.600
81.060
76.850
25.980
32.050
57.090
83.290
AP - SVM - APPX
k=25%
54.660
31.380
40.510
27.100
80.660
75.720
25.360
32.460
57.380
83.650
k=50%
55.640
30.660
38.650
25.530
79.950
74.670
23.680
32.810
57.430
83.560
k=75%
54.570
29.610
37.260
24.980
78.660
72.550
22.860
32.840
55.790
82.390
Table 1: Test AP for the different action classes of PASCAL VOC 2011 action dataset. For AP - SVM APPX , we report test results for 3 different values of k, which is the percentage of samples that are
included in the easy set among all the samples that the binary SVM had classified with margin > 1.
7
Binary SVM
0.1068
AP - SVM
AP - SVM - SEARCH
AP - SVM - SELECT
AP - SVM - APPX ( K =50)
ALL
0.5660
0.0671
0.0404
0.2341
0.0251
Table 2: Computation time (in seconds) for computing the most violated ranking when using the
different methods. The reported time is averaged over the training for all the action classes.
Binary SVM
0.637
AP - SVM
AP - SVM - SEARCH
AP - SVM - SELECT
AP - SVM - APPX ( K =50)
ALL
13.192
1.565
0.942
8.217
0.689
Table 3: Computation time (in milli-seconds) for computing the most violated ranking per iteration
when using the different methods. The reported time is averaged over all training iterations and
over all the action classes.
4.2 Object Detection
Dataset. We use the PASCAL VOC 2007 [6] object detection dataset, which consists of a total of
9963 images. The dataset is divided into a ?trainval? set of 5011 images and a ?test? set of 4952
images. All the images are labelled to indicate the presence or absence of the instances of 20
different object categories. In addition, we are also provided with tight bounding boxes around the
object instances, which we ignore during training and testing. Instead, we treat the location of the
objects as a latent variable. In order to reduce the latent variable space, we use the selective-search
algorithm [17] in its fast mode, which generates an average of 2000 candidate windows per image.
Features. For each of the candidate windows, we use a feature representation that is extracted
from a trained Convolutional Neural Network (CNN). Specifically, we pass the image as input to the
CNN and use the activation vector of the penultimate layer of the CNN as the feature vector. Inspired
by the work of Girshick et al. [9], we use the CNN that is trained on the ImageNet dataset [4], by
rescaling each candidate window to a fixed size of 224 ? 224. The length of the resulting feature
vector is 4096.
Methods. We train latent AP - SVMs [1] as object detectors for 20 object categories. In our experiments, we determine the value of the hyperparameters using 5-fold cross-validation. During testing,
we evaluate each candidate window generated by selective search, and use non-maxima suppression
to prune highly overlapping detections.
Results. This experiment places high computational demands due to the size of the dataset (5011
?trainval? images), as well as the size of the latent space (2000 candidate windows per image). We
compare the computational efficiency of the loss-augmented inference algorithm proposed in [19]
and the exact methods proposed by us. The total time taken for loss-augmented inference during
training, averaged over the all the 20 classes, is 0.3302 sec for our exact methods (SEARCH+SELECT)
which is significantly better than the 6.237 sec taken by the algorithm used in [19].
5
Discussion
We proposed three complementary approaches to improve the efficiency of learning AP - SVM. The
first two approaches exploit the problem structure to speed-up the computation of the most violated
ranking using exact loss-augmented inference. The third approach provides an accurate approximation of AP - SVM, which facilitates the trade-off of test accuracy and training time.
As mentioned in the introduction, our approaches can also be used in conjunction with other learning
frameworks, such as the popular deep convolutional neural networks. A combination of methods
proposed in this paper and the speed-ups proposed in [10] may prove to be effective in such a
framework. The efficacy of optimizing AP efficiently using other frameworks needs to be empirically
evaluated. Another computational bottleneck of all SSVM frameworks is the computation of the joint
feature vector. An interesting direction of future research would be to combine our approaches with
those of sparse feature coding [3, 8, 18] to improve the speed to AP - SVM learning further.
6
Acknowledgement
This work is partially funded by the European Research Council under the European Community?s
Seventh Framework Programme (FP7/2007-2013)/ERC Grant agreement number 259112. Pritish is
supported by the TCS Research Scholar Program.
8
References
[1] A. Behl, C. V. Jawahar, and M. P. Kumar. Optimizing average precision using weakly supervised data. In CVPR, 2014.
[2] M. Blaschko, A. Mittal, and E. Rahtu. An O(n log n) cutting plane algorithm for structured
output ranking. In GCPR, 2014.
[3] X. Boix, G. Roig, C. Leistner, and L. Van Gool. Nested sparse quantization for efficient feature
coding. In ECCV. 2012.
[4] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. ImageNet: A Large-Scale
Hierarchical Image Database. In CVPR, 2009.
[5] P. Dokania, A. Behl, C. V. Jawahar, and M. P. Kumar. Learning to rank using high-order
information. In ECCV, 2014.
[6] M. Everingham, L. Van Gool, C. Williams, J. Winn, and A. Zisserman. The PASCAL Visual Object Classes Challenge 2007 (VOC2007) Results.
http://www.pascalnetwork.org/challenges/VOC/voc2007/workshop/index.html.
[7] M. Everingham, L. Van Gool, C. Williams, J. Winn, and A. Zisserman. The PASCAL visual
object classes (VOC) challenge. IJCV, 2010.
[8] T. Ge, Q. Ke, and J. Sun. Sparse-coded features for image retrieval. In BMVC, 2013.
[9] R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich feature hierarchies for accurate object
detection and semantic segmentation. In CVPR, 2014.
[10] M. Jaderberg, A. Vedaldi, and A. Zisserman. Speeding up convolutional neural networks with
low rank expansions. In BMVC, 2014.
[11] D. Kim. Minimizing structural risk on decision tree classification. In Multi-Objective Machine
Learning, Springer. 2006.
[12] S. Maji, L. Bourdev, and J. Malik. Action recognition from a distributed representation of pose
and appearance. In CVPR, 2011.
[13] C. Shen, H. Li, and N. Barnes. Totally corrective boosting for regularized risk minimization.
arXiv preprint arXiv:1008.5188, 2010.
[14] C. Szegedy, A. Toshev, and D. Erhan. Deep neural networks for object detection. In NIPS,
2013.
[15] B. Taskar, C. Guestrin, and D. Koller. Max-margin Markov networks. In NIPS, 2003.
[16] I. Tsochantaridis, T. Hofmann, Y. Altun, and T. Joachims. Support vector machine learning for
interdependent and structured output spaces. In ICML, 2004.
[17] J. Uijlings, K. van de Sande, T. Gevers, and A. Smeulders. Selective search for object recognition. IJCV, 2013.
[18] J. Yang, K. Yu, and T. Huang. Efficient highly over-complete sparse coding using a mixture
model. In ECCV. 2010.
[19] Y. Yue, T. Finley, F. Radlinski, and T. Joachims. A support vector method for optimizing
average precision. In SIGIR, 2007.
9
| 5362 |@word cnn:4 briefly:1 everingham:2 minus:1 initial:1 score:26 efficacy:2 trainval:5 ecole:1 ours:1 outperforms:1 current:1 comparing:1 si:1 activation:3 must:1 hofmann:1 update:1 greedy:4 plane:2 xk:2 provides:4 boosting:2 location:2 org:1 simpler:2 five:4 consists:6 prove:1 ijcv:2 combine:1 indeed:1 expected:1 multi:1 inspired:2 voc:8 decreasing:3 window:5 optj:18 increasing:2 totally:1 confused:1 xx:1 provided:3 blaschko:1 maximizes:1 bike:1 what:1 minimizes:1 prohibits:1 degrading:1 developed:1 spends:1 guarantee:1 every:2 xd:5 scaled:1 grant:1 positive:20 before:1 treat:1 despite:1 ure:1 ap:89 inria:1 might:1 challenging:1 averaged:3 testing:4 practice:4 procedure:2 empirical:2 significantly:5 vedaldi:1 ups:3 word:7 pritish:3 suggest:1 altun:1 tsochantaridis:1 cal:1 risk:5 descending:8 optimize:1 www:1 quick:2 maximizing:3 williams:2 starting:2 convex:4 sigir:1 ke:1 shen:1 ii0:1 rule:1 roig:1 hierarchy:1 heavily:1 user:1 exact:7 us:4 agreement:1 pa:1 expensive:2 recognition:2 walking:1 database:1 hyderabad:2 observed:1 taskar:1 preprint:1 rij:4 sun:1 decrease:2 highest:4 trade:2 substantial:1 intuition:1 mentioned:1 complexity:9 optk:2 weakly:2 solving:2 depend:2 tight:1 trained:2 efficiency:6 spk:1 joint:10 milli:1 represented:1 corrective:1 maji:1 train:1 heat:1 fast:2 describe:1 effective:1 argmaxi:1 query:3 horse:1 supplementary:1 solve:1 cvpr:4 otherwise:2 advantage:4 sequence:4 propose:5 product:2 fr:1 relevant:6 rapidly:2 flexibility:1 description:1 exploiting:1 darrell:1 object:18 help:1 develop:1 ac:2 pose:1 bourdev:1 measured:1 ij:1 predicted:1 involves:1 come:1 indicate:1 differ:1 direction:1 correct:3 material:1 require:5 scholar:1 leistner:1 alleviate:1 proposition:3 around:1 ground:5 adopt:2 estimation:1 applicable:1 jawahar:4 council:1 vice:1 mittal:1 successfully:1 minimization:2 avoid:1 conjunction:2 focus:1 joachim:2 improvement:2 consistently:1 rank:40 greatly:1 contrast:1 suppression:1 kim:1 inference:27 i0:1 typically:2 entire:4 koller:1 selective:7 i1:6 overall:3 classification:11 issue:1 among:1 denoted:3 pascal:8 html:1 proposes:1 special:3 equal:4 represents:1 yu:1 icml:1 future:1 report:1 few:2 employ:1 individual:6 argmax:4 pawan:2 detection:10 highly:2 evaluation:2 mixture:1 nl:4 accurate:2 necessary:1 jumping:1 ples:1 tree:1 desired:2 girshick:2 instance:3 classify:1 disadvantage:1 maximization:3 cost:2 subset:4 seventh:1 inspires:1 reported:2 accomplish:1 considerably:1 person:3 retain:1 off:2 dong:1 containing:1 opposed:1 leveraged:1 choose:1 huang:1 inefficient:2 rescaling:1 li:3 szegedy:1 de:1 summarized:1 sec:2 coding:3 explicitly:2 ranking:33 depends:3 performed:3 later:1 red:1 sort:7 gevers:1 smeulders:1 accuracy:8 convolutional:3 who:1 efficiently:9 identify:2 worth:1 classified:5 detector:1 definition:2 proof:2 outnumber:2 gain:1 dataset:13 popular:1 knowledge:1 subsection:3 segmentation:1 higher:2 supervised:2 zisserman:3 bmvc:2 formulation:1 evaluated:1 though:2 box:6 furthermore:1 stage:9 until:1 web:1 overlapping:1 mode:2 quality:1 riding:2 facilitate:1 pascalnetwork:1 true:1 hence:1 assigned:3 iteratively:3 i2:7 semantic:1 during:9 criterion:1 outline:1 complete:2 demonstrate:2 image:12 common:1 empirically:2 overview:1 ji:1 belong:1 approximates:1 significant:3 refer:3 versa:1 rd:3 outlined:1 similarly:1 erc:1 had:1 dot:2 funded:1 showed:2 optimizing:7 optimizes:4 belongs:2 poselet:2 server:1 sande:1 binary:29 continue:1 xe:5 yi:5 iiit:4 scoring:1 guestrin:1 greater:1 additional:1 employed:1 prune:1 deng:1 converge:1 maximize:4 determine:1 monotonically:2 ii:3 full:1 unimodal:9 reduces:3 ing:1 cross:2 long:1 retrieval:6 phoning:1 divided:2 coded:1 feasibility:1 ensuring:5 prediction:5 lossaugmented:2 iteration:9 represent:1 arxiv:2 addition:1 separately:2 interval:2 winn:2 leaving:1 ssvms:1 unlike:2 snj:2 yue:8 facilitates:1 structural:1 noting:1 presence:1 intermediate:1 iii:1 easy:6 yang:1 xj:15 appx:6 reduce:4 regarding:1 bottleneck:1 colour:1 dokania:1 jj:1 action:14 deep:2 ssvm:8 detailed:2 amount:1 mid:1 svms:1 category:2 reduced:1 http:1 specifies:1 percentage:1 estimated:4 track:1 popularity:1 per:4 correctly:1 discrete:10 key:1 clarity:1 sum:3 fourth:1 place:1 throughout:1 reader:3 ecp:1 putative:1 decision:2 appendix:1 comparable:1 bound:4 layer:1 guaranteed:1 fold:2 barnes:1 occur:1 deficiency:1 fei:2 generates:1 aspect:1 speed:9 toshev:1 min:5 kumar:4 separable:1 speedup:1 structured:6 according:4 alternate:2 centrale:1 poor:2 combination:1 remain:1 separability:1 wherever:1 taken:3 mohapatra:2 computationally:2 equation:2 ge:1 fp7:1 instrument:1 end:3 photo:1 apply:1 observe:1 hierarchical:1 appropriate:1 slower:2 original:1 top:1 remaining:4 ensure:1 include:1 running:1 exploit:4 especially:1 malik:2 objective:1 quantity:1 strategy:5 nr:2 surrogate:2 said:2 penultimate:1 sensible:1 w0:2 reason:1 length:1 index:4 minimizing:4 timization:1 difficult:10 negative:58 perform:1 allowing:1 upper:4 observation:2 datasets:6 markov:1 incorrectly:3 precise:1 trec:1 rahtu:1 community:1 pair:1 paris:1 required:5 specified:2 optimized:1 imagenet:2 nip:2 able:1 suggested:1 below:1 reading:1 challenge:4 saclay:1 program:1 including:1 max:1 gool:3 ranked:8 regularized:5 improve:3 voc2007:2 brief:1 ready:1 finley:1 sn:2 speeding:1 literature:1 understanding:1 deepest:1 acknowledgement:1 interdependent:1 asymptotic:1 relative:1 loss:62 interesting:2 validation:2 playing:1 share:1 row:1 eccv:3 repeat:1 supported:1 taking:3 fifth:1 sparse:4 van:4 distributed:1 overcome:1 boundary:2 rich:1 computes:1 programme:1 erhan:1 approximate:6 ignore:1 cutting:2 jaderberg:1 corpus:1 xi:13 search:25 latent:4 sk:1 table:6 learn:3 forest:1 expansion:1 european:2 uijlings:1 domain:7 main:2 linearly:1 bounding:6 hyperparameters:2 complementary:3 augmented:25 boix:1 precision:5 candidate:5 third:5 interleaving:34 donahue:1 list:5 svm:79 exists:1 socher:1 quantization:1 workshop:1 push:1 margin:2 demand:1 sorting:3 tc:1 simply:1 appearance:1 visual:2 highlighting:1 partially:1 springer:1 corresponds:2 truth:5 satisfies:1 relies:1 extracted:1 nested:1 viewed:1 labelled:1 absence:1 considerable:1 included:1 specifically:9 typical:2 uniformly:1 total:3 pas:1 select:8 formally:1 support:3 radlinski:1 relevance:1 violated:9 evaluate:2 |
4,819 | 5,363 | Ranking via Robust Binary Classification
Hyokun Yun
Amazon
Seattle, WA 98109
[email protected]
Parameswaran Raman, S. V. N. Vishwanathan
Department of Computer Science
University of California
Santa Cruz, CA 95064
{params,vishy}@ucsc.edu
Abstract
We propose RoBiRank, a ranking algorithm that is motivated by observing a close
connection between evaluation metrics for learning to rank and loss functions
for robust classification. It shows competitive performance on standard benchmark datasets against a number of other representative algorithms in the literature.
We also discuss extensions of RoBiRank to large scale problems where explicit
feature vectors and scores are not given. We show that RoBiRank can be efficiently parallelized across a large number of machines; for a task that requires
386, 133 ? 49, 824, 519 pairwise interactions between items to be ranked, RoBiRank finds solutions that are of dramatically higher quality than that can be found
by a state-of-the-art competitor algorithm, given the same amount of wall-clock
time for computation.
1
Introduction
Learning to rank is a problem of ordering a set of items according to their relevances to a given
context [8]. While a number of approaches have been proposed in the literature, in this paper we
provide a new perspective by showing a close connection between ranking and a seemingly unrelated
topic in machine learning, namely, robust binary classification.
In robust classification [13], we are asked to learn a classifier in the presence of outliers. Standard
models for classification such as Support Vector Machines (SVMs) and logistic regression do not
perform well in this setting, since the convexity of their loss functions does not let them give up
their performance on any of the data points [16]; for a classification model to be robust to outliers,
it has to be capable of sacrificing its performance on some of the data points. We observe that
this requirement is very similar to what standard metrics for ranking try to evaluate. Discounted
Cumulative Gain (DCG) [17] and its normalized version NDCG, popular metrics for learning to
rank, strongly emphasize the performance of a ranking algorithm at the top of the list; therefore, a
good ranking algorithm in terms of these metrics has to be able to give up its performance at the
bottom of the list if that can improve its performance at the top.
In fact, we will show that DCG and NDCG can indeed be written as a natural generalization of robust
loss functions for binary classification. Based on this observation we formulate RoBiRank, a novel
model for ranking, which maximizes the lower bound of (N)DCG. Although the non-convexity
seems unavoidable for the bound to be tight [9], our bound is based on the class of robust loss
functions that are found to be empirically easier to optimize [10]. Indeed, our experimental results
suggest that RoBiRank reliably converges to a solution that is competitive as compared to other
representative algorithms even though its objective function is non-convex.
While standard deterministic optimization algorithms such as L-BFGS [19] can be used to estimate
parameters of RoBiRank, to apply the model to large-scale datasets a more efficient parameter estimation algorithm is necessary. This is of particular interest in the context of latent collaborative
1
retrieval [24]; unlike standard ranking task, here the number of items to rank is very large and explicit feature vectors and scores are not given.
Therefore, we develop an efficient parallel stochastic optimization algorithm for this problem. It has
two very attractive characteristics: First, the time complexity of each stochastic update is independent of the size of the dataset. Also, when the algorithm is distributed across multiple number of
machines, no interaction between machines is required during most part of the execution; therefore,
the algorithm enjoys near linear scaling. This is a significant advantage over serial algorithms, since
it is very easy to deploy a large number of machines nowadays thanks to the popularity of cloud
computing services, e.g. Amazon Web Services.
We apply our algorithm to latent collaborative retrieval task on Million Song Dataset [3] which consists of 1,129,318 users, 386,133 songs, and 49,824,519 records; for this task, a ranking algorithm
has to optimize an objective function that consists of 386, 133 ? 49, 824, 519 number of pairwise
interactions. With the same amount of wall-clock time given to each algorithm, RoBiRank leverages
parallel computing to outperform the state-of-the-art with a 100% lift on the evaluation metric.
2
Robust Binary Classification
Suppose we are given training data which consists of n data points (x1 , y1 ), (x2 , y2 ), . . . , (xn , yn ),
where each xi ? Rd is a d-dimensional feature vector and yi ? {?1, +1} is a label associated
with it. A linear model attempts to learn a d-dimensional parameter ?, and for a given feature
vector x it predicts label +1 if hx, ?i ? 0 and ?1 otherwise. Here h?, ?i denotes the Euclidean dot
product between
Pn two vectors. The quality of ? can be measured by the number of mistakes it makes:
L(?) := i=1 I(yi ? hxi , ?i < 0). The indicator function I(? < 0) is called the 0-1 loss function,
because it has a value of 1 if the decision rule makes a mistake, and 0 otherwise. Unfortunately,
since the 0-1 loss is a discrete function its minimization is difficult [11]. The most popular solution
to this problem in machine learning is to upper bound the 0-1 loss by an easy to optimize function
[2]. For example, logistic regression uses the logistic loss function ?0 (t) := log2 (1 + 2?t ), to come
up with a continuous and convex objective function
L(?) :=
n
X
i=1
?0 (yi ? hxi , ?i),
(1)
which upper bounds L(?). It is clear that for each i, ?0 (yi ? hxi , ?i) is a convex function in ?;
therefore, L(?), a sum of convex functions, is also a convex function which is relatively easier to
optimize [6]. Support Vector Machines (SVMs) on the other hand can be recovered by using the
hinge loss to upper bound the 0-1 loss.
However, convex upper bounds such as L(?) are known to be sensitive to outliers [16]. The basic
intuition here is that when yi ? hxi , ?i is a very large negative number for some data point i, ?(yi ?
hxi , ?i) is also very large, and therefore the optimal solution of (1) will try to decrease the loss on
such outliers at the expense of its performance on ?normal? data points.
In order to construct robust loss functions, consider the following two transformation functions:
?1 (t) := log2 (t + 1), ?2 (t) := 1 ?
1
,
log2 (t + 2)
(2)
which, in turn, can be used to define the following loss functions:
?1 (t) := ?1 (?0 (t)), ?2 (t) := ?2 (?0 (t)).
(3)
One can see that ?1 (t) ? ? as t ? ??, but at a much slower rate than ?0 (t) does; its derivative
?10 (t) ? 0 as t ? ??. Therefore, ?1 (?) does not grow as rapidly as ?0 (t) on hard-to-classify data
points. Such loss functions are called Type-I robust loss functions by Ding [10], who also showed
that they enjoy statistical robustness properties. ?2 (t) behaves even better: ?2 (t) converges to a
constant as t ? ??, and therefore ?gives up? on hard to classify data points. Such loss functions
are called Type-II loss functions, and they also enjoy statistical robustness properties [10].
In terms of computation, of course, ?1 (?) and ?2 (?) are not convex, and therefore the objective
function based on such loss functions is more difficult to optimize. However, it has been observed
2
in Ding [10] that models based on optimization of Type-I functions are often empirically much
more successful than those which optimize Type-II functions. Furthermore, the solutions of Type-I
optimization are more stable to the choice of parameter initialization. Intuitively, this is because
Type-II functions asymptote to a constant, reducing the gradient to almost zero in a large fraction
of the parameter space; therefore, it is difficult for a gradient-based algorithm to determine which
direction to pursue. See Ding [10] for more details.
3
Ranking Model via Robust Binary Classification
Let X = {x1 , x2 , . . . , xn } be a set of contexts, and Y = {y1 , y2 , . . . , ym } be a set of items to
be ranked. For example, in movie recommender systems X is the set of users and Y is the set of
movies. In some problem settings, only a subset of Y is relevant to a given context x ? X ; e.g.
in document retrieval systems, only a subset of documents is relevant to a query. Therefore, we
define Yx ? Y to be a set of items relevant to context x. Observed data can be described by a set
W := {Wxy }x?X ,y?Yx where Wxy is a real-valued score given to item y in context x.
We adopt a standard problem setting used in the literature of learning to rank. For each context x
and an item y ? Yx , we aim to learn a scoring function f (x, y) : X ? Yx ? R that induces a
ranking on the item set Yx ; the higher the score, the more important the associated item is in the
given context. To learn such a function, we first extract joint features of x and y, which will be
denoted by ?(x, y). Then, we parametrize f (?, ?) using a parameter ?, which yields the linear model
f? (x, y) := h?(x, y), ?i, where, as before, h?, ?i denotes the Euclidean dot product between two
vectors. ? induces a ranking on the set of items Yx ; we define rank? (x, y) to be the rank of item y
in a given context x induced by ?. Observe that rank? (x, y) can also be written as a sum of 0-1 loss
functions (see e.g. Usunier et al. [23]):
X
rank? (x, y) =
I (f? (x, y) ? f? (x, y 0 ) < 0) .
(4)
y 0 ?Yx ,y 0 6=y
3.1
Basic Model
If an item y is very relevant in context x, a good parameter ? should position y at the top of the list;
in other words, rank? (x, y) has to be small, which motivates the following objective function [7]:
X
X
L(?) :=
cx
v(Wxy ) ? rank? (x, y),
(5)
x?X
y?Yx
where cx is an weighting factor for each context x, and v(?) : R+ ? R+ quantifies the relevance
level of y on x. Note that {cx } and v(Wxy ) can be chosen to reflect the metric the model is going to
be evaluated on (this will be discussed in Section 3.2). Note that (5) can be rewritten using (4) as a
sum of indicator functions. Following the strategy in Section 2, one can form an upper bound of (5)
by bounding each 0-1 loss function by a logistic loss function:
X X
X
L(?) :=
cx
v (Wxy ) ?
?0 (f? (x, y) ? f? (x, y 0 )) .
(6)
x?X
y?Yx
y 0 ?Yx ,y 0 6=y
Just like (1), (6) is convex in ? and hence easy to minimize.
3.2
DCG
Although (6) enjoys convexity, it may not be a good objective function for ranking. This is because
in most applications of learning to rank, it is more important to do well at the top of the list than at
the bottom, as users typically pay attention only to the top few items. Therefore, it is desirable to
give up performance on the lower part of the list in order to gain quality at the top. This intuition is
similar to that of robust classification in Section 2; a stronger connection will be shown below.
Discounted Cumulative Gain (DCG) [17] is one of the most popular metrics for ranking. For each
context x ? X , it is defined as:
X
v (Wxy )
DCG(?) := cx
,
(7)
log2 (rank? (x, y) + 2)
y?Yx
3
where v(t) = 2t ? 1 and cx = 1. Since 1/ log(t + 2) decreases quickly and then asymptotes to
a constant as t increases, this metric emphasizes the quality of the ranking at the top of the list.
Normalized DCG (NDCG) simply normalizes the metric to bound it between 0 and 1 by calculating
the maximum achievable DCG value mx and dividing by it [17].
3.3
RoBiRank
Now we formulate RoBiRank, which optimizes the lower bound of metrics for ranking in form (7).
Observe that max? DCG(?) can be rewritten as
X
X
1
min
cx
v (Wxy ) ? 1 ?
.
(8)
?
log2 (rank? (x, y) + 2)
x?X
y?Yx
Using (4) and the definition of the transformation function ?2 (?) in (2), we can rewrite the objective
function in (8) as:
?
?
X
X
X
L2 (?) :=
cx
v (Wxy ) ? ?2 ?
I (f? (x, y) ? f? (x, y 0 ) < 0)? .
(9)
x?X
y?Yx
y 0 ?Yx ,y 0 6=y
Since ?2 (?) is a monotonically increasing function, we can bound (9) with a continuous function by
bounding each indicator function using the logistic loss:
?
?
X
X
X
L2 (?) :=
cx
v (Wxy ) ? ?2 ?
?0 (f? (x, y) ? f? (x, y 0 ))? .
(10)
x?X
y?Yx
y 0 ?Yx ,y 0 6=y
This is reminiscent of the basic model in (6); as we applied the transformation ?2 (?) on the logistic
loss ?0 (?) to construct the robust loss ?2 (?) in (3), we are again applying the same transformation
on (6) to construct a loss function that respects the DCG metric used in ranking. In fact, (10) can be
seen as a generalization of robust binary classification by applying the transformation on a group of
logistic losses instead of a single loss. In both robust classification and ranking, the transformation
?2 (?) enables models to give up on part of the problem to achieve better overall performance.
As we discussed in Section 2, however, transformation of logistic loss using ?2 (?) results in Type-II
loss function, which is very difficult to optimize. Hence, instead of ?2 (?) we use an alternative transformation ?1 (?), which generates Type-I loss function, to define the objective function of RoBiRank:
?
?
X
X
X
L1 (?) :=
cx
v (Wxy ) ? ?1 ?
?0 (f? (x, y) ? f? (x, y 0 ))? .
(11)
x?X
y?Yx
y 0 ?Yx ,y 0 6=y
Since ?1 (t) ? ?2 (t) for every t > 0, we have L1 (?) ? L2 (?) ? L2 (?) for every ?. Note
that L1 (?) is continuous and twice differentiable. Therefore, standard gradient-based optimization
techniques can be applied to minimize it. As is standard, a regularizer on ? can be added to avoid
overfitting; for simplicity, we use the `2 -norm in our experiments.
3.4
Standard Learning to Rank Experiments
We conducted experiments to check the performance of RoBiRank (11) in a standard learning to
rank setting, with a small number of labels to rank. We pitch RoBiRank against the following
algorithms: RankSVM [15], the ranking algorithm of Le and Smola [14] (called LSRank in the sequel), InfNormPush [22], IRPush [1], and 8 standard ranking algorithms implemented in RankLib1
namely MART, RankNet, RankBoost, AdaRank, CoordAscent, LambdaMART, ListNet and RandomForests. We use three sources of datasets: LETOR 3.0 [8] , LETOR 4.02 and YAHOO LTRC
[20], which are standard benchmarks for ranking (see Table 2 for summary statistics). Each dataset
consists of five folds; we consider the first fold, and use the training, validation, and test splits provided. We train with different values of regularization parameter, and select one with the best NDCG
1
2
http://sourceforge.net/p/lemur/wiki/RankLib
http://research.microsoft.com/en-us/um/beijing/projects/letor/letor4dataset.aspx
4
on the validation dataset. The performance of the model with this parameter on the test dataset is
reported. We used implementation of the L-BFGS algorithm provided by the Toolkit for Advanced
Optimization (TAO)3 for estimating the parameter of RoBiRank. For the other algorithms, we either
implemented them using our framework or used the implementations provided by the authors.
TD 2004
1
TD 2004
1
RoBiRank
RankSVM
LSRank
InfNormPush
IRPush
0.8
NDCG@k
NDCG@k
0.8
0.6
0.4
RoBiRank
MART
RankNet
RankBoost
AdaRank
CoordAscent
LambdaMART
ListNet
RandomForests
0.6
5
10
15
0.4
20
k
5
10
15
20
k
Figure 1: Comparison of RoBiRank with a number of competing algorithms.
We use values of NDCG at different levels of truncation as our evaluation metric [17]; see Figure 1.
RoBiRank outperforms its competitors on most of the datasets; due to space constraints, we only
present plots for the TD 2004 dataset here and other plots can be found in Appendix B. The performance of RankSVM seems insensitive to the level of truncation for NDCG. On the other hand,
RoBiRank, which uses non-convex loss function to concentrate its performance at the top of the
ranked list, performs much better especially at low truncation levels. It is also interesting to note
that the NDCG@k curve of LSRank is similar to that of RoBiRank, but RoBiRank consistently outperforms at each level. RoBiRank dominates Inf-Push and IR-Push at all levels. When compared to
standard algorithms, Figure 1 (right), again RoBiRank outperforms especially at the top of the list.
Overall, RoBiRank outperforms IRPush and InfNormPush on all datasets except TD 2003 and
OHSUMED where IRPush seems to fare better at the top of the list. Compared to the 8 standard
algorithms, again RobiRank either outperforms or performs comparably to the best algorithm except
on two datasets (TD 2003 and HP 2003), where MART and Random Forests overtake RobiRank at
few values of NDCG. We present a summary of the NDCG values obtained by each algorithm in
Table 2 in the appendix. On the MSLR30K dataset, some of the additional algorithms like InfNormPush and IRPush did not complete within the time period available; indicated by dashes in the table.
4
Latent Collaborative Retrieval
For each context x and an item y ? Y, the standard problem setting of learning to rank requires
training data to contain feature vector ?(x, y) and score Wxy assigned on the x, y pair. When the
number of contexts |X | or the number of items |Y| is large, it might be difficult to define ?(x, y)
and measure Wxy for all x, y pairs. Therefore, in most learning to rank problems we define the set
of relevant items Yx ? Y to be much smaller than Y for each context x, and then collect data only
for Yx . Nonetheless, this may not be realistic in all situations; in a movie recommender system, for
example, for each user every movie is somewhat relevant.
On the other hand, implicit user feedback data is much more abundant. For example, a lot of users
on Netflix would simply watch movie streams on the system but do not leave an explicit rating. By
the action of watching a movie, however, they implicitly express their preference. Such data consist
only of positive feedback, unlike traditional learning to rank datasets which have score Wxy between
each context-item pair x, y. Again, we may not be able to extract feature vectors for each x, y pair.
In such a situation, we can attempt to learn the score function f (x, y) without a feature vector ?(x, y)
by embedding each context and item in an Euclidean latent space; specifically, we redefine the score
function to be: f (x, y) := hUx , Vy i, where Ux ? Rd is the embedding of the context x and Vy ? Rd
3
http://www.mcs.anl.gov/research/projects/tao/index.html
5
is that of the item y. Then, we can learn these embeddings by a ranking model. This approach was
introduced in Weston et al. [24], and was called latent collaborative retrieval.
Now we specialize RoBiRank model for this task. Let us define ? to be the set of context-item pairs
(x, y) which was observed in the dataset. Let v(Wxy ) = 1 if (x, y) ? ?, and 0 otherwise; this is a
natural choice since the score information is not available. For simplicity, we set cx = 1 for every
x. Now RoBiRank (11) specializes to:
?
?
X
X
L1 (U, V ) =
?1 ?
?0 (f (Ux , Vy ) ? f (Ux , Vy0 ))? .
(12)
y 0 6=y
(x,y)??
Note that now the summation inside the parenthesis of (12) is over all items Y instead of a smaller
set Yx , therefore we omit specifying the range of y 0 from now on. To avoid overfitting, a regularizer
is added to (12); for simplicity we use the Frobenius norm of U and V in our experiments.
4.1
Stochastic Optimization
When the size of the data |?| or the number of items |Y| is large, however, methods that require
exact evaluation of the function value and its gradient will become very slow since the evaluation
takes O (|?| ? |Y|) computation. In this case, stochastic optimization methods are desirable [4];
in this subsection, we will develop a stochastic gradient descent algorithm whose complexity is
independent of |?| and |Y|.
For simplicity, let ? be a concatenation of all parameters {Ux }x?X , {Vy }y?Y . The gradient
?? L1 (U, V ) of (12) is
?
?
X
X
?? ? 1 ?
?0 (f (Ux , Vy ) ? f (Ux , Vy0 ))? .
(x,y)??
y 0 6=y
Finding an unbiased estimator of the gradient whose computation is independent of |?| is not difficult; if we sample a pair (x, y) uniformly from ?, then it is easy to see that the following estimator
?
?
X
|?| ? ?? ?1 ?
?0 (f (Ux , Vy ) ? f (Ux , Vy0 ))?
(13)
y 0 6=y
is unbiased. This still involves a summation over Y, however, so it requires O(|Y|) calculation.
Since ?1 (?) is a nonlinear function it seems unlikely that an unbiased stochastic gradient which
randomizes over Y can be found; nonetheless, to achieve convergence guarantees of the stochastic
gradient descent algorithm, unbiasedness of the estimator is necessary [18].
We attack this problem by linearizing the objective function by parameter expansion. Note the
following property of ?1 (?) [5]:
?1 (t) = log2 (t + 1) ? ? log2 ? +
? ? (t + 1) ? 1
.
log 2
(14)
1
This holds for any ? > 0, and the bound is tight when ? = t+1
. Now introducing an auxiliary
parameter ?xy for each (x, y) ? ? and applying this bound, we obtain an upper bound of (12) as
P
0
?xy
X
y 0 6=y ?0 (f (Ux , Vy ) ? f (Ux , Vy )) + 1 ? 1
L(U, V, ?) :=
? log2 ?xy +
. (15)
log 2
(x,y)??
Now we propose an iterative algorithm in which, each iteration consists of (U, V )-step and ?-step;
in the (U, V )-step we minimize (15) in (U, V ) and in the ?-step we minimize in ?. Pseudo-code can
be found in Algorithm 1 in Appendix C.
(U, V )-step The partial derivative
of (15) in terms of U and V can be calculated as:
P
P
0
?U,V L(U, V, ?) := log1 2 (x,y)?? ?xy
y 0 6=y ?U,V ?0 (f (Ux , Vy ) ? f (Ux , Vy )) . Now it is easy
to see that the following stochastic procedure unbiasedly estimates the above gradient:
6
RoBiRank 4
RoBiRank 16
RoBiRank 32
RoBiRank 1
0.1
0
0
0.5
1
1.5
2
2.5
3
0.2
Mean Precision@10
0.2
0.2
Weston et al. (2012)
RoBiRank 1
RoBiRank 4
RoBiRank 16
RoBiRank 32
0.3
Mean Precision@1
Mean Precision@1
0.3
0.1
0
Weston et al. (2012)
RoBiRank 1
RoBiRank 4
RoBiRank 16
RoBiRank 32
0.15
0.1
5 ? 10?2
0
0.2
number of machines ? seconds elapsed ?106
0.4
0.6
seconds elapsed
0.8
0
1
?105
0
0.2
0.4
0.6
seconds elapsed
0.8
1
?105
Figure 2: Left: Scaling of RoBiRank on Million Song Dataset. Center, Right: Comparison of
RoBiRank and Weston et al. [24] when the same amount of wall-clock computation time is given.
? Sample (x, y) uniformly from ?
? Sample y 0 uniformly from Y \ {y}
? Estimate the gradient by
|?| ? (|Y| ? 1) ? ?xy
? ?U,V ?0 (f (Ux , Vy ) ? f (Ux , Vy0 )).
log 2
(16)
Therefore a stochastic gradient descent algorithm based on (16) will converge to a local minimum
of the objective function (15) with probability one [21]. Note that the time complexity of calculating
(16) is independent of |?| and |Y|. Also, it is a function of only Ux and Vy ; the gradient is zero in
terms of other variables.
?-step When U and V are fixed, minimization of ?xy variable is independent of each other and
a simple analytic solution exists: ?xy = P 0 ?0 (f (Ux ,V1y )?f (Ux ,V 0 ))+1 . This of course requires
y
y 6=y
O(|Y|) work. In principle, we can avoid summation over Y by taking stochastic gradient in terms of
?xy as we did for U and V . However, since the exact solution is simple to compute and also because
most of the computation time is spent on (U, V )-step, we found this update rule to be efficient.
Parallelization The linearization trick in (15) not only enables us to construct an efficient stochastic gradient algorithm, but also makes possible to efficiently parallelize the algorithm across multiple
number of machines. Due to lack of space, details are relegated to Appendix D.
4.2
Experiments
In this subsection we use the Million Song Dataset (MSD) [3], which consists of 1,129,318 users
(|X |), 386,133 songs (|Y|), and 49,824,519 records (|?|) of a user x playing a song y in the training
dataset. The objective is to predict the songs from the test dataset that a user is going to listen to4 .
Since explicit ratings are not given, NDCG is not applicable for this task; we use precision at 1 and
10 [17] as our evaluation metric. In our first experiment we study the scaling behavior of RoBiRank
as a function of number of machines. RoBiRank p denotes the parallel version of RoBiRank which
is distributed across p machines. In Figure 2 (left) we plot mean Precision@1 as a function of the
number of machines ? the number of seconds elapsed; this is a proxy for CPU time. If an algorithm
linearly scales across multiple processors, then all lines in the figure should overlap with each other.
As can be seen RoBiRank exhibits near ideal speed up when going from 4 to 32 machines5 .
In our next experiment we compare RoBiRank with a state of the art algorithm from Weston et al.
[24], which optimizes a similar objective function (17). We compare how fast the quality of the
solution improves as a function of wall clock time. Since the authors of Weston et al. [24] do not
make available their code, we implemented their algorithm within our framework using the same
data structures and libraries used by our method. Furthermore, for a fair comparison, we used the
same initialization for U and V and performed an identical grid-search over the step size parameter.
4
the original data also provides the number of times a song was played by a user, but we ignored this in our
experiment.
5
The graph for RoBiRank 1 is hard to see because it was run for only 100,000 CPU-seconds.
7
It can be seen from Figure 2 (center, right) that on a single machine the algorithm of Weston et al.
[24] is very competitive and outperforms RoBiRank. The reason for this might be the introduction
of the additional ? variables in RoBiRank, which slows down convergence. However, RoBiRank
training can be distributed across processors, while it is not clear how to parallelize the algorithm
of Weston et al. [24]. Consequently, RoBiRank 32 which uses 32 machines for its computation can
produce a significantly better model within the same wall clock time window.
5
Related Work
In terms of modeling, viewing ranking problems as generalization of binary classification problems
is not a new idea; for example, RankSVM defines the objective function as a sum of hinge losses,
similarly to our basic model (6) in Section 3.1. However, it does not directly optimize the ranking
metric such as NDCG; the objective function and the metric are not immediately related to each
other. In this respect, our approach is closer to that of Le and Smola [14] which constructs a convex
upper bound on the ranking metric and Chapelle et al. [9] which improves the bound by introducing
non-convexity. The objective function of Chapelle et al. [9] is also motivated by ramp loss, which
is used for robust classification; nonetheless, to our knowledge the direct connection between the
ranking metrics in form (7) (DCG, NDCG) and the robust loss (3) is our novel contribution. Also,
our objective function is designed to specifically bound the ranking metric, while Chapelle et al. [9]
proposes a general recipe to improve existing convex bounds.
Stochastic optimization of the objective function for latent collaborative retrieval has been also explored in Weston et al. [24]. They attempt to minimize
?
?
X
X
? ?1 +
I(f (Ux , Vy ) ? f (Ux , Vy0 ) < 0)? ,
(17)
(x,y)??
y 0 6=y
Pt
where ?(t) = k=1 k1 . This is similar to our objective function (15); ?(?) and ?2 (?) are asymptotically equivalent. However, we argue that our formulation (15) has two major advantages. First, it is
a continuous and differentiable function, therefore gradient-based algorithms such as L-BFGS and
stochastic gradient descent have convergence guarantees. On the other hand, the objective function
of Weston et al. [24] is not even continuous, since their formulation is based on a function ?(?)
that is defined for only natural numbers. Also, through the linearization trick in (15) we are able
to obtain an unbiased stochastic gradient, which is necessary for the convergence guarantee, and to
parallelize the algorithm across multiple machines as discussed in Appendix D. It is unclear how
these techniques can be adapted for the objective function of Weston et al. [24].
6
Conclusion
In this paper, we developed RoBiRank, a novel model on ranking, based on insights and techniques
from robust binary classification. Then, we proposed a scalable and parallelizable stochastic optimization algorithm that can be applied to latent collaborative retrieval task which large-scale data
without feature vectors and explicit scores have to take care of. Experimental results on both learning
to rank datasets and latent collaborative retrieval dataset suggest the advantage of our approach.
As a final note, the experiments in Section 4.2 are arguably unfair towards WSABIE. For instance,
one could envisage using clever engineering tricks to derive a parallel variant of WSABIE (e.g.,
by averaging gradients from various machines), which might outperform RoBiRank on the MSD
dataset. While performance on a specific dataset might be better, we would lose global convergence
guarantees. Therefore, rather than obsess over the performance of a specific algorithm on a specific
dataset, via this paper we hope to draw the attention of the community to the need for developing
principled parallel algorithms for this important problem.
Acknowledgments We thank anonymous reviewers for their constructive comments, and Texas
Advanced Computing Center for infrastructure and support for experiments. This material is partially based upon work supported by the National Science Foundation under grant No. IIS-1117705.
8
References
[1] S. Agarwal. The infinite push: A new support vector ranking algorithm that directly optimizes
accuracy at the absolute top of the list. In SDM, pages 839?850. SIAM, 2011.
[2] P. L. Bartlett, M. I. Jordan, and J. D. McAuliffe. Convexity, classification, and risk bounds.
Journal of the American Statistical Association, 101(473):138?156, 2006.
[3] T. Bertin-Mahieux, D. P. Ellis, B. Whitman, and P. Lamere. The million song dataset. In Proceedings of the 12th International Conference on Music Information Retrieval (ISMIR 2011),
2011.
[4] L. Bottou and O. Bousquet. The tradeoffs of large-scale learning. Optimization for Machine
Learning, page 351, 2011.
[5] G. Bouchard. Efficient bounds for the softmax function, applications to inference in hybrid
models. 2007.
[6] S. Boyd and L. Vandenberghe. Convex Optimization. Cambridge University Press, Cambridge,
England, 2004.
[7] D. Buffoni, P. Gallinari, N. Usunier, and C. Calauz`enes. Learning scoring functions with
order-preserving losses and standardized supervision. In Proceedings of the 28th International
Conference on Machine Learning (ICML-11), pages 825?832, 2011.
[8] O. Chapelle and Y. Chang. Yahoo! learning to rank challenge overview. Journal of Machine
Learning Research-Proceedings Track, 14:1?24, 2011.
[9] O. Chapelle, C. B. Do, C. H. Teo, Q. V. Le, and A. J. Smola. Tighter bounds for structured
estimation. In Advances in neural information processing systems, pages 281?288, 2008.
[10] N. Ding. Statistical Machine Learning in T-Exponential Family of Distributions. PhD thesis,
PhD thesis, Purdue University, West Lafayette, Indiana, USA, 2013.
[11] V. Feldman, V. Guruswami, P. Raghavendra, and Y. Wu. Agnostic learning of monomials by
halfspaces is hard. SIAM Journal on Computing, 41(6):1558?1590, 2012.
[12] R. Gemulla, E. Nijkamp, P. J. Haas, and Y. Sismanis. Large-scale matrix factorization with
distributed stochastic gradient descent. In Conference on Knowledge Discovery and Data
Mining, pages 69?77, 2011.
[13] P. J. Huber. Robust Statistics. John Wiley and Sons, New York, 1981.
[14] Q. V. Le and A. J. Smola. Direct optimization of ranking measures. Technical Report
0704.3359, arXiv, April 2007. http://arxiv.org/abs/0704.3359.
[15] C.-P. Lee and C.-J. Lin. Large-scale linear ranksvm. Neural Computation, 2013. To Appear.
[16] P. Long and R. Servedio. Random classification noise defeats all convex potential boosters.
Machine Learning Journal, 78(3):287?304, 2010.
[17] C. D. Manning, P. Raghavan, and H. Sch?utze. Introduction to Information Retrieval. Cambridge University Press, 2008. URL http://nlp.stanford.edu/IR-book/.
[18] A. Nemirovski, A. Juditsky, G. Lan, and A. Shapiro. Robust stochastic approximation approach
to stochastic programming. SIAM Journal on Optimization, 19(4):1574?1609, 2009.
[19] J. Nocedal and S. J. Wright. Numerical Optimization. Springer Series in Operations Research.
Springer, 2nd edition, 2006.
[20] T. Qin, T.-Y. Liu, J. Xu, and H. Li. Letor: A benchmark collection for research on learning to
rank for information retrieval. Information Retrieval, 13(4):346?374, 2010.
[21] H. E. Robbins and S. Monro. A stochastic approximation method. Annals of Mathematical
Statistics, 22:400?407, 1951.
[22] C. Rudin. The p-norm push: A simple convex ranking algorithm that concentrates at the top
of the list. The Journal of Machine Learning Research, 10:2233?2271, 2009.
[23] N. Usunier, D. Buffoni, and P. Gallinari. Ranking with ordered weighted pairwise classification. In Proceedings of the International Conference on Machine Learning, 2009.
[24] J. Weston, C. Wang, R. Weiss, and A. Berenzweig. Latent collaborative retrieval. arXiv
preprint arXiv:1206.4603, 2012.
9
| 5363 |@word version:2 achievable:1 norm:3 seems:4 stronger:1 nd:1 liu:1 series:1 score:10 document:2 outperforms:6 existing:1 recovered:1 com:2 written:2 reminiscent:1 john:1 cruz:1 numerical:1 realistic:1 enables:2 analytic:1 asymptote:2 plot:3 designed:1 update:2 juditsky:1 rudin:1 item:22 record:2 infrastructure:1 provides:1 wxy:14 preference:1 attack:1 org:1 five:1 unbiasedly:1 mahieux:1 mathematical:1 ucsc:1 direct:2 become:1 consists:6 specialize:1 redefine:1 inside:1 pairwise:3 huber:1 indeed:2 behavior:1 discounted:2 td:5 gov:1 cpu:2 ohsumed:1 window:1 increasing:1 provided:3 project:2 unrelated:1 estimating:1 maximizes:1 agnostic:1 what:1 pursue:1 developed:1 finding:1 transformation:8 indiana:1 guarantee:4 pseudo:1 every:4 um:1 classifier:1 gallinari:2 grant:1 omit:1 yn:1 enjoy:2 arguably:1 mcauliffe:1 before:1 service:2 lemur:1 positive:1 local:1 engineering:1 mistake:2 sismanis:1 randomizes:1 parallelize:3 ndcg:14 might:4 twice:1 initialization:2 collect:1 specifying:1 factorization:1 nemirovski:1 range:1 lafayette:1 acknowledgment:1 procedure:1 significantly:1 boyd:1 word:1 suggest:2 close:2 clever:1 lamere:1 context:19 applying:3 risk:1 optimize:8 www:1 deterministic:1 equivalent:1 center:3 reviewer:1 attention:2 convex:14 formulate:2 amazon:3 simplicity:4 immediately:1 rule:2 estimator:3 insight:1 vandenberghe:1 embedding:2 annals:1 pt:1 deploy:1 suppose:1 user:10 exact:2 programming:1 us:3 trick:3 predicts:1 bottom:2 cloud:1 observed:3 preprint:1 ding:4 wang:1 ordering:1 decrease:2 halfspaces:1 principled:1 intuition:2 convexity:5 complexity:3 asked:1 tight:2 rewrite:1 upon:1 whitman:1 joint:1 various:1 regularizer:2 train:1 fast:1 query:1 lift:1 whose:2 stanford:1 valued:1 ramp:1 otherwise:3 statistic:3 envisage:1 final:1 seemingly:1 advantage:3 differentiable:2 sdm:1 net:1 propose:2 interaction:3 product:2 qin:1 relevant:6 rapidly:1 achieve:2 ranklib:1 frobenius:1 sourceforge:1 recipe:1 seattle:1 convergence:5 requirement:1 letor:4 produce:1 converges:2 leave:1 spent:1 derive:1 develop:2 measured:1 dividing:1 implemented:3 auxiliary:1 involves:1 come:1 direction:1 concentrate:2 stochastic:19 raghavan:1 viewing:1 material:1 require:1 hx:1 generalization:3 wall:5 anonymous:1 tighter:1 summation:3 extension:1 hold:1 wright:1 normal:1 ranksvm:5 predict:1 major:1 adopt:1 utze:1 estimation:2 applicable:1 lose:1 label:3 sensitive:1 teo:1 robbins:1 weighted:1 minimization:2 hope:1 rankboost:2 aim:1 rather:1 pn:1 avoid:3 berenzweig:1 consistently:1 rank:23 check:1 parameswaran:1 inference:1 dcg:11 typically:1 unlikely:1 relegated:1 going:3 tao:2 overall:2 classification:18 html:1 denoted:1 yahoo:2 proposes:1 art:3 softmax:1 construct:5 identical:1 icml:1 report:1 few:2 national:1 microsoft:1 attempt:3 ab:1 interest:1 mining:1 evaluation:6 nowadays:1 capable:1 partial:1 necessary:3 xy:8 closer:1 euclidean:3 abundant:1 sacrificing:1 instance:1 classify:2 modeling:1 elli:1 introducing:2 subset:2 monomials:1 successful:1 conducted:1 reported:1 params:1 unbiasedness:1 thanks:1 overtake:1 international:3 siam:3 sequel:1 lee:1 ym:1 quickly:1 again:4 reflect:1 unavoidable:1 thesis:2 watching:1 booster:1 american:1 derivative:2 book:1 li:1 potential:1 bfgs:3 gemulla:1 ranking:32 stream:1 performed:1 try:2 lot:1 observing:1 competitive:3 netflix:1 parallel:5 bouchard:1 nijkamp:1 monro:1 collaborative:8 minimize:5 contribution:1 ir:2 accuracy:1 characteristic:1 efficiently:2 who:1 yield:1 raghavendra:1 emphasizes:1 comparably:1 mc:1 processor:2 parallelizable:1 definition:1 against:2 competitor:2 nonetheless:3 servedio:1 vishy:1 associated:2 gain:3 calauz:1 dataset:17 popular:3 subsection:2 listen:1 improves:2 knowledge:2 higher:2 listnet:2 wei:1 april:1 formulation:2 evaluated:1 though:1 strongly:1 furthermore:2 just:1 smola:4 implicit:1 clock:5 hand:4 web:1 nonlinear:1 lack:1 defines:1 logistic:8 quality:5 indicated:1 usa:1 normalized:2 y2:2 contain:1 unbiased:4 hence:2 regularization:1 assigned:1 attractive:1 during:1 linearizing:1 yun:1 complete:1 performs:2 l1:5 novel:3 behaves:1 empirically:2 overview:1 insensitive:1 defeat:1 million:4 discussed:3 fare:1 association:1 significant:1 cambridge:3 feldman:1 rd:3 grid:1 hp:1 similarly:1 dot:2 hxi:5 toolkit:1 stable:1 chapelle:5 supervision:1 showed:1 perspective:1 optimizes:3 inf:1 binary:8 yi:6 scoring:2 seen:3 minimum:1 additional:2 somewhat:1 care:1 preserving:1 parallelized:1 determine:1 converge:1 period:1 monotonically:1 ii:5 multiple:4 desirable:2 technical:1 adarank:2 england:1 calculation:1 long:1 retrieval:13 lin:1 msd:2 serial:1 parenthesis:1 pitch:1 scalable:1 regression:2 basic:4 variant:1 metric:18 arxiv:4 iteration:1 agarwal:1 buffoni:2 grow:1 source:1 sch:1 parallelization:1 unlike:2 comment:1 induced:1 jordan:1 near:2 presence:1 leverage:1 ideal:1 split:1 easy:5 embeddings:1 competing:1 idea:1 tradeoff:1 texas:1 motivated:2 bartlett:1 guruswami:1 url:1 song:9 york:1 action:1 ranknet:2 dramatically:1 ignored:1 santa:1 clear:2 amount:3 induces:2 svms:2 http:5 wiki:1 outperform:2 shapiro:1 vy:13 popularity:1 track:1 discrete:1 express:1 group:1 lan:1 nocedal:1 graph:1 asymptotically:1 fraction:1 sum:4 beijing:1 run:1 almost:1 family:1 ismir:1 wu:1 draw:1 raman:1 decision:1 appendix:5 scaling:3 ob:1 bound:21 pay:1 dash:1 played:1 fold:2 adapted:1 vishwanathan:1 constraint:1 x2:2 bousquet:1 generates:1 speed:1 min:1 relatively:1 department:1 developing:1 according:1 structured:1 manning:1 across:7 smaller:2 wsabie:2 son:1 outlier:4 intuitively:1 ene:1 discus:1 turn:1 usunier:3 parametrize:1 available:3 rewritten:2 operation:1 apply:2 observe:3 alternative:1 robustness:2 slower:1 original:1 top:12 denotes:3 standardized:1 nlp:1 log2:8 hinge:2 yx:21 calculating:2 music:1 appear:1 k1:1 especially:2 objective:20 added:2 strategy:1 traditional:1 unclear:1 exhibit:1 gradient:20 mx:1 thank:1 concatenation:1 topic:1 haas:1 argue:1 reason:1 code:2 index:1 difficult:6 unfortunately:1 expense:1 negative:1 slows:1 implementation:2 reliably:1 motivates:1 perform:1 upper:7 recommender:2 observation:1 datasets:8 benchmark:3 purdue:1 descent:5 situation:2 y1:2 community:1 rating:2 introduced:1 namely:2 required:1 pair:6 connection:4 california:1 elapsed:4 able:3 below:1 challenge:1 max:1 overlap:1 ranked:3 natural:3 hybrid:1 indicator:3 advanced:2 improve:2 movie:6 library:1 specializes:1 log1:1 extract:2 literature:3 l2:4 discovery:1 loss:35 interesting:1 bertin:1 validation:2 foundation:1 proxy:1 principle:1 playing:1 normalizes:1 course:2 summary:2 supported:1 truncation:3 enjoys:2 taking:1 absolute:1 distributed:4 curve:1 feedback:2 xn:2 calculated:1 cumulative:2 author:2 collection:1 emphasize:1 implicitly:1 global:1 overfitting:2 xi:1 anl:1 continuous:5 latent:9 iterative:1 quantifies:1 search:1 table:3 learn:6 robust:20 ca:1 forest:1 expansion:1 bottou:1 did:2 linearly:1 bounding:2 noise:1 edition:1 fair:1 x1:2 xu:1 representative:2 west:1 en:1 slow:1 v1y:1 wiley:1 precision:5 position:1 explicit:5 exponential:1 unfair:1 weighting:1 down:1 specific:3 showing:1 list:11 explored:1 dominates:1 consist:1 exists:1 phd:2 execution:1 linearization:2 push:4 aspx:1 easier:2 cx:11 simply:2 ordered:1 ux:19 partially:1 watch:1 chang:1 springer:2 mart:3 weston:12 consequently:1 towards:1 hard:4 specifically:2 except:2 reducing:1 uniformly:3 averaging:1 infinite:1 called:5 experimental:2 select:1 support:4 lambdamart:2 relevance:2 constructive:1 evaluate:1 |
4,820 | 5,364 | Tight Bounds for Influence in Diffusion Networks and
Application to Bond Percolation and Epidemiology
R?emi Lemonnier1,2
Kevin Scaman1
Nicolas Vayatis1
1
2
CMLA ? ENS Cachan, CNRS, France, 1000mercis, Paris, France
{lemonnier, scaman, vayatis}@cmla.ens-cachan.fr
Abstract
In this paper, we derive theoretical bounds for the long-term influence of a node
in an Independent Cascade Model (ICM). We relate these bounds to the spectral
radius of a particular matrix and show that the behavior is sub-critical when this
spectral radius is lower than 1. More specifically,
? we point out that, in general
networks, the sub-critical regime behaves in O( n) where n is the size of the
network, and that this upper bound is met for star-shaped networks. We apply our
results to epidemiology and percolation on arbitrary networks, and derive a bound
for the critical value beyond which a giant connected component arises. Finally,
we show empirically the tightness of our bounds for a large family of networks.
1
Introduction
The emergence of social graphs of the World Wide Web has had a considerable effect on propagation of ideas or information. For advertisers, these new diffusion networks have become a favored
vector for viral marketing operations, that consist of advertisements that people are likely to share
by themselves with their social circle, thus creating a propagation dynamics somewhat similar to
the spreading of a virus in epidemiology ([1]). Of particular interest is the problem of influence
maximization, which consists of selecting the top-k nodes of the network to infect at time t = 0
in order to maximize in expectation the final number of infected nodes at the end of the epidemic.
This problem was first formulated by Domingues and Richardson in [2] and later expressed in [3]
as an NP-hard discrete optimization problem under the Independent Cascade (IC) framework, a
widely-used probabilistic model for information propagation.
From an algorithmic point of view, influence maximization has been fairly well studied. Assuming
the transmission probability of all edges are known, Kempe, Kleinberg and Tardos ([3]) derived
a greedy algorithm based on Monte-Carlo simulations that was shown to approximate the optimal
solution up to a factor 1 ? 1e , building on classical results of optimization theory. Since then, various
techniques were proposed in order to significantly improve the scalability of this algorithm ([4, 5, 6,
7]), and also to provide an estimate of the transmission probabilities from real data ([8, 9]). Recently,
a series of papers ([10, 11, 12]) introduced continuous-time diffusion networks in which infection
spreads during a time period T at varying rates across the different edges. While these models
provide a more accurate representation of real-world networks for finite T , they are equivalent to the
IC model when T ? ?. In this paper, will focus on such long-term behavior of the contagion.
From a theoretical point of view, little is known about the influence maximization problem under the
IC model framework. The most celebrated result established by Newman ([13]) proves the equivalence between bond percolation and the Susceptible-Infected-Removed (SIR) model in epidemiology
([14]) that can be identified to a special case of IC model where transmission probability are equal
amongst all infectious edges.
In this paper, we propose new bounds on the influence of any set of nodes. Moreover, we prove the
existence of an epidemic threshold for a key quantity defined by the spectral radius of a given hazard
1
matrix.
? Under this threshold, the influence of any given set of nodes in a network of size n will be
O( n), while the influence of a randomly chosen set of nodes will be O(1). We provide empirical
evidence that these bounds are sharp for a family of graphs and sets of initial influencers and can
therefore be used as what is to our knowledge the first closed-form formulas for influence estimation.
We show that these results generalize bounds obtained on the SIR model by Draief, Ganesh and
Massouli?e ([15]) and are closely related to recent results on percolation on finite inhomogeneous
random graphs ([16]).
The rest of the paper is organized as follows. In Sec. 2, we recall the definition of Information
Cascades Model and introduce useful notations. In Sec. 3, we derive theoretical bounds for the
influence. In Sec. 4, we show that our results also apply to the fields of percolation and epidemiology
and generalize existing results in these fields. In Sec. 5, we illustrate our results by applying them
on simple networks and retrieving well-known results. In Sec. 6, we perform experiments in order
to show that our bounds are sharp for a family of graphs and sets of initial nodes.
2
2.1
Information Cascades Model
Influence in random networks and infection dynamics
Let G = (V, E) be a directed network of n nodes and A ? V be a set of n0 nodes that are initially
contagious (e.g. aware of a piece of information, infected by a disease or adopting a product). In
the sequel, we will refer to A as the influencers. The behavior of the cascade is modeled using a
probabilistic framework. The influencer nodes spread the contagion through the network by means
of transmission through the edges of the network. More specifically, each contagious node can infect
its neighbors with a certain probability. The influence of A, denoted as ?(A), is the expected number
of nodes reached by the contagion originating from A, i.e.
X
?(A) =
P(v is infected by the contagion |A).
(1)
v?V
We consider three infection dynamics that we will show in the next section to be equivalent regarding
the total number of infected nodes at the end of the epidemic.
Discrete-Time Information Cascades [DT IC(P)] At time t = 0, only the influencers are infected. Given a matrix P = (pij )ij ? [0, 1]n?n , each node i that receives the contagion at time t
may transmit it at time t + 1 along its outgoing edge (i, j) ? E with probability pij . Node i cannot
make any attempt to infect its neighbors in subsequent rounds. The process terminates when no
more infections are possible.
Continuous-Time Information Cascades [CT IC(F, T )] At time t = 0, only the influencers
are infected. Given a matrix F = (fij )ij of non-negative integrable functions, each node i that
receives the contagion at time t may transmit it at time s > t along its outgoing edge (i, j) ? E
with stochastic rate of occurrence fij (s ? t). The process terminates at a given deterministic time
T > 0. This model is much richer than Discrete-time IC, but we will focus here on its behavior
when T = ?.
Random Networks [RN (P)] Given a matrix P = (pij )ij ? [0, 1]n?n , each edge (i, j) ? E is
removed independently of the others with probability 1 ? pij . A node i ? V is said to be infected if
i is linked to at least one element of A in the spanning subgraph G 0 = (V, E 0 ) where E 0 ? E is the
set of non-removed edges.
For any v ? V, we will designate by influence of v the influence of the set containing only v,
i.e. ?({v}). We will show in Section 4.2 that, if P is symmetric and G undirected, these three
infection processes are equivalent to bond percolation and the influence of a node v is also equal
to the expected size of the connected component containing v in G 0 . This will make our results
applicable to percolation in arbitrary networks. Following the percolation literature, we will denote
as sub-critical a cascade whose influence is not proportional to the size of the network n.
2
2.2
The hazard matrix
In order to linearize the influence problem and derive upper bounds, we introduce the concept of
hazard matrix, which describes the behavior of the information cascade. As we will see in the
following, in the case of Continuous-time Information Cascades, this matrix gives, for each edge of
the network, the integral of the instantaneous rate of transmission (known as hazard function). The
spectral radius of this matrix will play a key role in the influence of the cascade.
Definition. For a given graph G = (V, E) and edge transmission probabilities pij , let H be the
n ? n matrix, denoted as the hazard matrix, whose coefficients are
? ln(1 ? pij ) if (i, j) ? E
Hij =
.
(2)
0
otherwise
Next lemma shows the equivalence between the three definitions of the previous section.
Lemma 1. For a given graph G = (V, E), set of influencers A, and transmission probabilities matrix P, the distribution of the set of infected nodes is equal under
R ?the infection dynamics
DT IC(P), CT IC(F, ?) and RN (P), provided that for any (i, j) ? E, 0 fij (t)dt = Hij .
Definition. For a given set of influencers A ? V, we will denote as H(A) the hazard matrix except
for zeros along the columns whose indices are in A:
H(A)ij = 1{j ?A}
Hij .
/
(3)
We recall that for any square matrix M , its spectral radius ?(M ) is defined by ?(M ) = maxi (|?i |)
where ?1 , ..., ?n are the (possibly repeated) eigenvalues of matrix M . We will also use that, when
>
>
M is a real square matrix with positive entries, ?( M +M
) = supX XX >MXX .
2
Remark. When the pij are small, the hazard matrix is very close to the transmission matrix P.
This implies that, for low pij values, the spectral radius of H will be very close to that of P. More
specifically, a simple calculation holds
?(P) ? ?(H) ?
? ln(1 ? kPk? )
?(P),
kPk?
(4)
for x ? 1? implies that the
where kPk? = maxi,j pij . The relatively slow increase of ? ln(1?x)
x
behavior of ?(P) and ?(H) will be of the same order of magnitude even for high (but lower than 1)
values of kPk? .
3
Upper bounds for the influence of a set of nodes
Given A ? V the set of influencer nodes and |A| = n0 < n, we derive here two upper bounds
for the influence of A. The first bound (Proposition 1) applies to any set of influencers A such
that |A| = n0 . Intuitively, this result correspond to a best-case scenario (or a worst-case scenario,
depending on the viewpoint), since we can target any set of nodes so as to maximize the resulting
contagion.
>
Proposition 1. Define ?c (A) = ?( H(A)+H(A)
). Then, for any A such that |A| = n0 < n, denoting
2
by ?(A) the expected number of nodes reached by the cascade starting from A:
?(A) ? n0 + ?1 (n ? n0 ),
where ?1 is the smallest solution in [0, 1] of the following equation:
?c (A)n0
?1 ? 1 + exp ??c (A)?1 ?
= 0.
?1 (n ? n0 )
Corollary 1. Under the same assumptions:
3
(5)
(6)
s
? if ?c (A) < 1,
? if ?c (A) ? 1,
?(A) ? n0 +
?c (A) p
n0 (n ? n0 ),
1 ? ?c (A)
2?c (A)
!
.
?(A) ? n ? (n ? n0 ) exp ??c (A) ? p
4n/n0 ? 3 ? 1
?
In particular, when ?c (A) < 1, ?(A) = O( n) and the regime is sub-critical.
The second result (Proposition 2) applies in the case where A is drawn from a uniform distribution
over the ensemble of sets of n0 nodes chosen amongst n (denoted as Pn0 (V)). This result corresponds to the average-case scenario in a setting where the initial influencer nodes are not known and
drawn independently of the transmissions over each edge.
>
Proposition 2. Define ?c = ?( H+H
). Assume the set of influencers A is drawn from a uniform
2
distribution over Pn0 (V). Then, denoting by ?uniform the expected number of nodes reached by the
cascade starting from A:
?uniform ? n0 + ?2 (n ? n0 ),
where ?2 is the unique solution in [0, 1] of the following equation:
?c n 0
?2 ? 1 + exp ??c ?2 ?
= 0.
n ? n0
(7)
(8)
Corollary 2. Under the same assumptions:
? if ?c < 1,
? if ?c ? 1,
n0
,
1 ? ?c
?c
?uniform ? n ? (n ? n0 ) exp ?
.
1 ? nn0
?uniform ?
In particular, when ?c < 1, ?uniform = O(1) and the regime is sub-critical.
?
The difference in the sub-critical regime between O( n) and O(1) for the worst and average case
influence is an important feature of our results, and is verified in our experiments (see Sec. 6). Intuitively, when the network is inhomogeneous and contains highly central nodes (e.g. scale-free networks), there will be a significant difference between specifically targeting the most central nodes
and random targeting (which will most probably target a peripheral node).
4
Application to epidemiology and percolation
Building on the celebrated equivalences between the fields of percolation, epidemiology and influence maximization, we show that our results generalize existing results in these fields.
4.1
Susceptible-Infected-Removed (SIR) model in epidemiology
We show here that Proposition 1 further improves results on the SIR model in epidemiology. This
widely used model was introduced by Kermac and McKendrick ([14]) in order to model the propagation of a disease in a given population. In this setting, nodes represent individuals, that can be
in one of three possible states, susceptible (S), infected (I) or removed (R). At t = 0, a subset A of
n0 nodes is infected and the epidemic spreads according to the following evolution. Each infected
node transmits the infection along its outgoing edge (i, j) ? E at stochastic rate of occurrence ? and
is removed from the graph at stochastic rate of occurrence ?. The process ends for a given T > 0.
It is straightforward that, if the removed events are not observed, this infection process is equivalent
to CT IC(F, T ) where for any (i, j) ? E,fij (t) = ? exp(??t). The hazard matrix H is therefore
equal to ?? A where A = 1{(i,j)?E} ij is the adjacency matrix of the underlying network. Note
4
that, by Lemma 1, our results can be used in order to model the total number of infected nodes in a
setting where infection and recovery rates of a given node exhibit a non-exponential behavior. For
instance, incubation periods for different individuals generally follow a log-normal distribution [17],
which indicates that continuous-time IC with a log-normal rate of removal might be well-suited to
model some kind of infections.
It was recently shown by Draief, Ganesh and Massouli?e ([15]) that, in the case of undirected networks, and if ??(A) < ?,
?
nn0
?(A) ?
.
(9)
?
1 ? ? ?(A)
?
This result shows, that, when ?(H) = ?? ?(A) < 1, the influence of set of nodes A is O( n).
We show in the next lemma that this result is a direct consequence of Corollary 1: the condition
?c (A) < 1 is weaker than ?(H) < 1 and, under these conditions, the bound of Corollary 1 is tighter.
Lemma 2. For any symmetric adjacency matrix A, initial set of influencers A such that |A| = n0 <
?
n, ? > 0 and ? < ?(A)
, we have simultaneously ?c (A) ? ?? ?(A) and
s
?
nn0
?c (A) p
,
(10)
n0 (n ? n0 ) ?
n0 +
?
1 ? ?c (A)
1 ? ? ?(A)
where the condition ? <
?
?(A)
imposes that the regime is sub-critical.
Moreover, these new bounds capture with more accuracy the behavior of the influence in extreme
cases. In the limit ? ? 0, the difference between the two?
bounds is significant, because Proposition
1 yields ?(A) ? n0 whereas (9) only ensures ?(A) ? nn0 . When n = n0 , Proposition 1 also
0
ensures that ?(A) = n0 whereas (9) yields ?(A) ? 1? ?n?(A)
. Secondly, Proposition 1 gives also
?
bounds in the case ??(A) ? ?. Finally, Proposition 1 applies to more general cases that the classical
homogeneous SIR model, and allows infection and recovery rates to vary across individuals.
4.2
Bond percolation
Given a finite undirected graph G = (V, E), bond percolation theory describes the behavior of
connected clusters of the spanning subgraph of G obtained by retaining a subset E 0 ? E of edges
of G according to a given distribution.When these removals occur independently along each edge
with same probability 1 ? p, this process is called homogeneous percolation and is fairly well known
(see e.g [18]). The inhomogeneous case, where the independent edge removal probabilities 1 ? pij
vary across the edges, is more intricate and has been the subject of recent studies. In particular,
results on critical probabilities and size of the giant component have been obtained by Bollobas,
Janson and Riordan in [16]. However, these bounds hold for a particular class of asymptotic graphs
(inhomogeneous random graphs) when n ? ?. In the next lemma, we show that our results can be
used in order to obtain bounds that hold in expectation for any fixed graph.
Lemma 3. Let P = (pij )ij ? [0, 1]n?n be a symmetric matrix. Let G 0 = (V, E 0 ) be the undirected
subgraph of G such that each edge {i, j} ? E is removed independently with probability 1 ? pij . Let
Gd = (V, Ed ) be the directed graph such that (i, j) ? Ed ?? {i, j} ? E. Then, for any v ? V,
the expected size of the connected component containing v in G 0 is equal to the influence of v in Gd
under the infection process DT IC(P).
We now derive an upper bound for C1 (G 0 ), the size of the largest connected component of the
spanning subgraph G 0 = (V, E 0 ). In the following, we will denote by E[C1 (G 0 )] the expected value
of this random variable, given P = (pij )ij .
Proposition 3. Let G = (V, E) be an undirected network where each edge {i, j} ? E has an independent probability 1 ? pij of being removed. The expected size of the largest connected component
of the resulting subgraph G 0 is upper bounded by:
?
(11)
E[C1 (G 0 )] ? n ?3 ,
where ?3 is the unique solution in [0, 1] of the following equation:
n?1
n
?3 ? 1 +
exp ?
?(H)?3 = 0.
n
n?1
5
(12)
Moreover, the resulting network has a probability of being connected upper bounded by:
P(G 0 is connected) ? ?3 .
(13)
In the case ?(H) < 1, we can further simplify our bounds in the same way than for Propositions 1
and 2.
q
n
Corollary 3. In the case ?(H) < 1, E[C1 (G 0 )] ? 1??(H)
.
Whereas our results hold for any n ? N, classical results in percolation theory study the asymptotic
behavior of sequences of graphs when n ? ?. In order to further compare our results, we therefore
consider sequences of spanning subgraphs (G 0 n )n ?N , obtained by removing each edge of graphs
of n nodes (Gn )n ?N with probability 1 ? pnij . A previous result ([16], Corollary 3.2 of section
5) states that, for particular sequences known as inhomogeneous random graphs and under a given
sub-criticality condition, C1 (G 0 n ) = o(n) asymptotically almost surely (a.a.s.), i.e with probability
going to 1 as n ? ?. Using Proposition 3, we get for our part the following result:
Corollary 4. Assume the sequence Hn = ? ln(1 ? pnij ) ij
is such that
n ?N
n
lim sup ?(H ) < 1.
(14)
n??
Then, for any > 0, we have asymptotically almost surely when n ? ?,
C1 (Gn0 ) = o(n1/2+ ).
(15)
This result is to our knowledge the first to bound the expected size of the largest connected component in general arbitrary networks.
5
Application to particular networks
In order to illustrate our theoretical results, we now apply our bounds to three specific networks and
compare them to existing results, showing that our bounds are always of the same order than these
specific results. We consider three particular networks: 1) star-shaped networks, 2) Erd?os-R?enyi
networks and 3) random graphs with an expected degree distribution. In order to simplify these
problems and exploit existing theorems, we will consider in this section that pij = p is fixed for
each edge {i, j} ? E. Infection dynamics thus only depend on p, the set of influencers A, and the
structure of the underlying network.
5.1
Star-shaped networks
For a star shaped network centered around a given node v1 , and A = {v1 }, the exact influence is
computable and writes ?({v1 }) = 1 + p(n ? 1). As H(A)ij = ? ln(1 ? p)1{i=1,j6=1} , the spectral
radius is given by
H(A) + H(A)>
? ln(1 ? p) ?
?
=
n ? 1.
(16)
2
2
Therefore, Proposition 1 states that ?({v1 }) ? 1 + (n ? 1)?1 where ?1 is the solution of equation
?
1
ln(1 ? p)
1 ? ?1 = exp
?1 n ? 1 + ?
.
(17)
2
?1 n ? 1
1
1
It is worth mentionning that, when p = ?n?1
, ?1 = ?n?1
is solution of (17) and therefore the
?
bound is ?({v1 }) ? 1 + n ? 1 which is tight. Note that, in the case of star-shaped networks, the
influence does not present a critical behavior and is always linear with respect to the total number of
nodes n.
5.2
Erd?os-R?enyi networks
For Erd?os-R?enyi networks G(n, p) (i.e. an undirected network with n nodes where each couple of
nodes (i, j) ? V 2 belongs to E independently of the others with probability p), the exact influence
6
of a set of nodes is not known. However, percolation theory characterizes the limit behavior of the
giant connected component when n ? ?. In the simplest case of Erd?os-R?enyi networks G(n, nc )
the following result holds:
Lemma 4. (taken from [16]) For a given sequence of Erd?os-R?enyi networks G(n, nc ), we have:
? if c < 1, C1 (G(n, nc )) ?
3
(1?c)2
log(n) a.a.s.
? if c > 1, C1 (G(n, nc )) = (1 + o(1))?n a.a.s. where ? ? 1 + exp(??c) = 0.
As previously stated, our results hold for any given graph, and not only asymptotically. However,
we get an asymptotic behavior consistent with the aforementioned result. Indeed, using notations of
n
section 4.2, Hij
= ? ln(1 ? nc )1{i6=j} and ?(Hn ) = ?(n ? 1) ln(1 ? nc ). Using Proposition 3, and
noting that ?3 = (1 + o(1))?, we get that, for any > 0:
? if c < 1, C1 (G(n, nc )) = o(n1/2+ ) a.a.s.
? if c > 1, C1 (G(n, nc )) ? (1 + o(1))?n1+ a.a.s., where ? ? 1 + exp(??c) = 0.
5.3
Random graphs with given expected degree distribution
In this section, we apply our bounds to random graphs whose expected degree distribution is fixed
(see e.g [19], section 13.2.2). More specifically, let w = (wi )i?{1,...,n} be the expected degree of
each node of the network. For a fixed w, let G(w) be a random graph whose edges are selected
independently and randomly with probability
1{i6=j} wi wj
.
(18)
qij = P
k wk
For these graphs, results on the volume of connected components (i.e the expected sum of degrees
of the nodes in these components) were derived in [20] but our work gives to our knowledge the first
result on the size of the giant component. Note that Erd?os-R?enyi G(n, p) networks are a special case
of (18) where wi = np for any i ? V.
In order to further compare our results, we note that these graphs are also very similar to the widely
used configuration model where node degrees are fixed to a sequence w, the main difference being
that the occupation probabilities pij are in this P
case not independent
anymore. For configuration
P
models, a giant component exists if and only if i wi2 > 2 i wi ([21, 22]). In theP
case of graphs
P
with given expected degree distribution, we retrieve the key role played by the ratio i wi2 / i wi
>
) < 1 where
in our criterion of non-existence of the giant component given by ?( H+H
2
P
w2
H + H>
?
? ?((qij )ij ) ? Pi i .
(19)
2
i wi
The left-hand approximation is particularly good when the qij are small. This is for instance the case
?
as soon as there exists ? < 1 such that, for any i ? V, wi = o(n
P ). The right-hand side
P is based
P on
2
the fact that the spectral radius of the matrix (qij + 1{i=j} wi / k wk )ij is given by i wi2 / i wi .
6
Experimental results
In this section, we show that the bounds given in Sec. 3 are tight (i.e. very close to empirical results in
particular graphs), and are good approximations of the influence on a large set of random networks.
Fig. 1a compares experimental simulations of the influence to the bound derived in proposition 1.
The considered networks have n = 1000 nodes and are of 6 types (see e.g [19] for further details on
these different networks): 1) Erd?os-R?enyi networks, 2) Preferential attachment networks, 3) Smallworld networks, 4) Geometric random networks ([23]), 5) 2D regular grids and 6) totally connected
networks with fixed weight b ? [0, 1] except for the ingoing and outgoing edges of the influencer
node A = {v1 } having weight a ? [0, 1]. Except for totally connected networks, edge probabilities
are set to the same value p for each edge (this parameter was used to tune the spectral radius ?c (A)).
All points of the plots are averages over 100 simulations. The results show that the bound in proposition 1 is tight (see totally connected networks in Fig. 1a) and close to the real influence for a large
7
class of random
? networks. In particular, the tightness of the bound around ?c (A) = 1 validates the
behavior in n of the worst-case influence in the sub-critical regime. Similarly, Fig. 1b compares
1000
900
900
800
800
700
700
uniform
600
influence (?
influence (?(A))
)
1000
500
400
totally connected
erdos renyi
preferential attachment
small World
geometric random
2D grid
upper bound
300
200
100
0
0
2
4
6
8
spectral radius of the Hazard matrix (?c(A))
(a) Fixed set of influencers
600
500
400
totally connected
erdos renyi
preferential attachment
small World
geometric random
2D grid
upper bound
300
200
100
0
10
0
2
4
6
8
spectral radius of the Hazard matrix (?c)
10
(b) Uniformly distributed set of influencers
Figure 1: Empirical influence on random networks of various types. The solid lines are the upper
bounds in propositions 1 (for Fig. 1a) and 2 (for Fig. 1b).
experimental simulations of the influence to the bound derived in proposition 2 in the case of random
initial influencers. While this bound is not as tight as the previous one, the behavior of the bound
agrees with experimental simulations, and proves a relatively good approximation of the influence
under a random set of initial influencers. It is worth mentioning that the bound is tight for the subcritical regime and shows that corollary 2 is a good approximation of ?uniform when ?c < 1. In
order to verify the criticality of ?c (A) = 1, we compared the behavior of ?(A)?w.r.t the size of the
network n. When ?c (A) < 1 (see Fig. 2a in which ?c (A) = 0.5), ?(A) = O( n), and the bound
is tight. On the contrary, when ?c (A) > 1 (see Fig. 2b in which ?c (A) = 1.5), ?(A) = O(n), and
?(A) is linear w.r.t. n for most random networks.
totally connected
erdos renyi
preferential attachment
small World
geometric random
2D grid
upper bound
influence (?(A))
25
20
400
15
300
200
10
100
5
0
totally connected
erdos renyi
preferential attachment
small World
geometric random
2D grid
upper bound
500
influence (?(A))
30
0
200
400
600
size of the network (n)
800
1000
(a) Sub-critical regime: ?c (A) = 0.5
0
0
200
400
600
size of the network (n)
800
1000
(b) Super-critical regime: ?c (A) = 1.5
Figure 2: Influence w.r.t. the size of the network in the sub-critical and super-critical regime. The
solid line is the upper bound in proposition 1. Note the square-root versus linear behavior.
7
Conclusion
In this paper, we derived the first upper bounds for the influence of a given set of nodes in any
finite graph under the Independent Cascade Model (ICM) framework, and relate them to the spectral
radius of a given hazard matrix. We show that these bounds can also be used to generalize previous
results in the fields of epidemiology and percolation. Finally, we provide empirical evidence that
these bounds are close to the best possible for general graphs.
Acknowledgments
This research is part of the SODATECH project funded by the French Government within the program of ?Investments for the Future ? Big Data?.
8
References
[1] Justin Kirby and Paul Marsden. Connected marketing: the viral, buzz and word of mouth revolution.
Elsevier, 2006.
[2] Pedro Domingos and Matt Richardson. Mining the network value of customers. In Proceedings of the
seventh ACM SIGKDD international conference on Knowledge discovery and data mining, pages 57?66.
ACM, 2001.
? Tardos. Maximizing the spread of influence through a social
[3] David Kempe, Jon Kleinberg, and Eva
network. In Proceedings of the Ninth ACM SIGKDD International Conference on Knowledge Discovery
and Data Mining, KDD ?03, pages 137?146, New York, NY, USA, 2003. ACM.
[4] Wei Chen, Yajun Wang, and Siyu Yang. Efficient influence maximization in social networks. In Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining, pages
199?208. ACM, 2009.
[5] Wei Chen, Chi Wang, and Yajun Wang. Scalable influence maximization for prevalent viral marketing
in large-scale social networks. In Proceedings of the 16th ACM SIGKDD international conference on
Knowledge discovery and data mining, pages 1029?1038. ACM, 2010.
[6] Amit Goyal, Wei Lu, and Laks VS Lakshmanan. Celf++: optimizing the greedy algorithm for influence
maximization in social networks. In Proceedings of the 20th international conference companion on
World wide web, pages 47?48. ACM, 2011.
[7] Kouzou Ohara, Kazumi Saito, Masahiro Kimura, and Hiroshi Motoda. Predictive simulation framework
of stochastic diffusion model for identifying top-k influential nodes. In Asian Conference on Machine
Learning, pages 149?164, 2013.
[8] Manuel Gomez Rodriguez, Jure Leskovec, and Andreas Krause. Inferring networks of diffusion and
influence. In Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery
and data mining, pages 1019?1028. ACM, 2010.
[9] Seth A. Myers and Jure Leskovec. On the convexity of latent social network inference. In NIPS, pages
1741?1749, 2010.
[10] Manuel Gomez-Rodriguez, David Balduzzi, and Bernhard Sch?olkopf. Uncovering the temporal dynamics
of diffusion networks. In ICML, pages 561?568, 2011.
[11] Manuel G Rodriguez and Bernhard Sch?olkopf. Influence maximization in continuous time diffusion
networks. In Proceedings of the 29th International Conference on Machine Learning (ICML-12), pages
313?320, 2012.
[12] Nan Du, Le Song, Manuel Gomez-Rodriguez, and Hongyuan Zha. Scalable influence estimation in
continuous-time diffusion networks. In NIPS, pages 3147?3155, 2013.
[13] Mark EJ Newman. Spread of epidemic disease on networks. Physical review E, 66(1):016128, 2002.
[14] William O Kermack and Anderson G McKendrick. Contributions to the mathematical theory of epidemics. ii. the problem of endemicity. Proceedings of the Royal society of London. Series A, 138(834):55?
83, 1932.
[15] Moez Draief, Ayalvadi Ganesh, and Laurent Massouli?e. Thresholds for virus spread on networks. In Proceedings of the 1st international conference on Performance evaluation methodolgies and tools, page 51.
ACM, 2006.
[16] B?ela Bollob?as, Svante Janson, and Oliver Riordan. The phase transition in inhomogeneous random
graphs. Random Structures & Algorithms, 31(1):3?122, 2007.
[17] Kenrad E Nelson. Epidemiology of infectious disease: general principles. Infectious Disease Epidemiology Theory and Practice. Gaithersburg, MD: Aspen Publishers, pages 17?48, 2007.
[18] Svante Janson, Tomasz Luczak, and Andrzej Rucinski. Random graphs, volume 45. John Wiley & Sons,
2011.
[19] Mark Newman. Networks: An Introduction. Oxford University Press, Inc., New York, NY, USA, 2010.
[20] Fan Chung and Linyuan Lu. Connected components in random graphs with given expected degree sequences. Annals of combinatorics, 6(2):125?145, 2002.
[21] Michael Molloy and Bruce Reed. A critical point for random graphs with a given degree sequence.
Random structures & algorithms, 6(2-3):161?180, 1995.
[22] Michael Molloy and Bruce Reed. The size of the giant component of a random graph with a given degree
sequence. Combinatorics probability and computing, 7(3):295?305, 1998.
[23] Mathew Penrose. Random geometric graphs, volume 5. Oxford University Press Oxford, 2003.
9
| 5364 |@word pnij:2 motoda:1 simulation:6 lakshmanan:1 solid:2 initial:6 celebrated:2 series:2 contains:1 selecting:1 configuration:2 denoting:2 janson:3 existing:4 yajun:2 virus:2 manuel:4 john:1 subsequent:1 kdd:1 plot:1 n0:27 v:1 greedy:2 selected:1 node:48 mathematical:1 along:5 direct:1 become:1 retrieving:1 qij:4 consists:1 prove:1 mckendrick:2 introduce:2 intricate:1 indeed:1 expected:15 themselves:1 behavior:17 chi:1 kazumi:1 little:1 totally:7 provided:1 xx:1 moreover:3 notation:2 underlying:2 bounded:2 project:1 what:1 kind:1 giant:7 kimura:1 temporal:1 draief:3 positive:1 limit:2 consequence:1 oxford:3 laurent:1 might:1 studied:1 equivalence:3 mentioning:1 directed:2 unique:2 acknowledgment:1 linyuan:1 investment:1 goyal:1 practice:1 writes:1 moez:1 saito:1 empirical:4 cascade:14 significantly:1 word:1 regular:1 get:3 cannot:1 close:5 targeting:2 influence:48 applying:1 equivalent:4 deterministic:1 customer:1 bollobas:1 maximizing:1 straightforward:1 starting:2 independently:6 recovery:2 identifying:1 subgraphs:1 retrieve:1 population:1 nn0:4 siyu:1 tardos:2 transmit:2 target:2 play:1 annals:1 exact:2 cmla:2 homogeneous:2 domingo:1 element:1 particularly:1 observed:1 role:2 wang:3 capture:1 worst:3 wj:1 ensures:2 connected:20 eva:1 removed:9 disease:5 convexity:1 dynamic:6 depend:1 tight:7 predictive:1 seth:1 various:2 enyi:7 london:1 monte:1 hiroshi:1 newman:3 kevin:1 whose:5 richer:1 widely:3 tightness:2 otherwise:1 epidemic:6 richardson:2 emergence:1 validates:1 final:1 sequence:9 eigenvalue:1 myers:1 propose:1 product:1 fr:1 scaman:1 subgraph:5 infectious:3 scalability:1 olkopf:2 cluster:1 transmission:9 derive:6 illustrate:2 linearize:1 depending:1 ij:11 implies:2 met:1 radius:12 closely:1 inhomogeneous:6 fij:4 stochastic:4 centered:1 adjacency:2 government:1 proposition:19 tighter:1 secondly:1 designate:1 hold:6 around:2 considered:1 ic:12 normal:2 exp:9 algorithmic:1 vary:2 smallest:1 estimation:2 applicable:1 bond:5 spreading:1 percolation:16 ohara:1 largest:3 agrees:1 tool:1 always:2 super:2 ej:1 varying:1 corollary:8 derived:5 focus:2 prevalent:1 indicates:1 sigkdd:5 elsevier:1 inference:1 cnrs:1 kermack:1 initially:1 originating:1 going:1 france:2 uncovering:1 aforementioned:1 denoted:3 favored:1 retaining:1 special:2 fairly:2 kempe:2 equal:5 field:5 aware:1 shaped:5 having:1 icml:2 jon:1 future:1 np:2 others:2 simplify:2 endemicity:1 randomly:2 simultaneously:1 individual:3 asian:1 phase:1 n1:3 william:1 attempt:1 interest:1 highly:1 mining:6 evaluation:1 gn0:1 extreme:1 accurate:1 oliver:1 edge:24 integral:1 preferential:5 circle:1 theoretical:4 leskovec:2 instance:2 column:1 gn:1 infected:14 maximization:8 entry:1 subset:2 uniform:9 seventh:1 supx:1 gd:2 st:1 pn0:2 epidemiology:12 international:8 sequel:1 probabilistic:2 influencers:14 michael:2 central:2 containing:3 hn:2 possibly:1 creating:1 chung:1 star:5 sec:7 wk:2 coefficient:1 inc:1 combinatorics:2 gaithersburg:1 piece:1 later:1 root:1 view:2 closed:1 linked:1 sup:1 reached:3 characterizes:1 zha:1 tomasz:1 bruce:2 contribution:1 square:3 accuracy:1 ensemble:1 correspond:1 yield:2 generalize:4 lu:2 carlo:1 worth:2 buzz:1 j6:1 kpk:4 infection:13 lemonnier:1 definition:4 ed:2 transmits:1 couple:1 recall:2 knowledge:8 lim:1 improves:1 organized:1 ingoing:1 dt:4 follow:1 wei:3 erd:7 anderson:1 marketing:3 hand:2 receives:2 web:2 ganesh:3 o:7 propagation:4 rodriguez:4 french:1 building:2 effect:1 matt:1 concept:1 verify:1 usa:2 evolution:1 symmetric:3 round:1 during:1 criterion:1 instantaneous:1 recently:2 viral:3 behaves:1 empirically:1 physical:1 volume:3 refer:1 significant:2 grid:5 i6:2 similarly:1 had:1 funded:1 influencer:4 recent:2 optimizing:1 belongs:1 scenario:3 certain:1 integrable:1 somewhat:1 surely:2 maximize:2 advertiser:1 period:2 ii:1 calculation:1 aspen:1 long:2 hazard:11 marsden:1 scalable:2 expectation:2 represent:1 adopting:1 c1:10 vayatis:1 whereas:3 krause:1 publisher:1 sch:2 w2:1 rest:1 probably:1 smallworld:1 subject:1 undirected:6 contrary:1 noting:1 yang:1 identified:1 andreas:1 idea:1 regarding:1 computable:1 song:1 york:2 remark:1 useful:1 generally:1 tune:1 simplest:1 discrete:3 key:3 threshold:3 drawn:3 verified:1 diffusion:8 v1:6 graph:32 asymptotically:3 subcritical:1 sum:1 massouli:3 family:3 almost:2 cachan:2 bound:46 ct:3 nan:1 played:1 gomez:3 fan:1 mathew:1 occur:1 kleinberg:2 emi:1 relatively:2 influential:1 according:2 peripheral:1 across:3 terminates:2 describes:2 son:1 wi:9 kirby:1 intuitively:2 taken:1 ln:9 equation:4 previously:1 end:3 operation:1 molloy:2 apply:4 spectral:12 occurrence:3 anymore:1 existence:2 top:2 andrzej:1 laks:1 exploit:1 balduzzi:1 prof:2 amit:1 classical:3 society:1 quantity:1 md:1 said:1 exhibit:1 amongst:2 riordan:2 nelson:1 spanning:4 assuming:1 modeled:1 index:1 reed:2 ratio:1 bollob:1 nc:8 susceptible:3 relate:2 hij:4 negative:1 stated:1 perform:1 upper:14 finite:4 criticality:2 rn:2 ninth:1 arbitrary:3 sharp:2 introduced:2 david:2 paris:1 established:1 nip:2 jure:2 beyond:1 justin:1 masahiro:1 wi2:3 regime:10 program:1 royal:1 mouth:1 critical:16 event:1 ela:1 improve:1 contagion:7 attachment:5 review:1 literature:1 geometric:6 removal:3 discovery:5 asymptotic:3 sir:5 occupation:1 proportional:1 versus:1 degree:10 pij:16 consistent:1 imposes:1 principle:1 viewpoint:1 share:1 pi:1 free:1 soon:1 side:1 weaker:1 wide:2 neighbor:2 distributed:1 celf:1 world:7 transition:1 social:7 approximate:1 erdos:4 bernhard:2 hongyuan:1 svante:2 thep:1 continuous:6 latent:1 nicolas:1 du:1 spread:6 main:1 big:1 paul:1 repeated:1 icm:2 fig:7 en:2 slow:1 ny:2 wiley:1 sub:11 inferring:1 exponential:1 advertisement:1 renyi:4 infect:3 formula:1 removing:1 theorem:1 companion:1 specific:2 contagious:2 showing:1 revolution:1 maxi:2 evidence:2 consist:1 exists:2 magnitude:1 chen:2 suited:1 likely:1 penrose:1 expressed:1 luczak:1 applies:3 pedro:1 corresponds:1 acm:12 formulated:1 considerable:1 hard:1 specifically:5 except:3 uniformly:1 lemma:8 total:3 called:1 experimental:4 people:1 mark:2 arises:1 outgoing:4 |
4,821 | 5,365 | Shaping Social Activity by Incentivizing Users
Mehrdad Farajtabar?
Nan Du?
Manuel Gomez-Rodriguez?
?
Isabel Valera
Hongyuan Zha?
Le Song?
?
?
Georgia Institute of Technology
MPI for Software Systems
Univ. Carlos III in Madrid?
{mehrdad,dunan}@gatech.edu
[email protected]
{zha,lsong}@cc.gatech.edu
[email protected]
Abstract
Events in an online social network can be categorized roughly into endogenous
events, where users just respond to the actions of their neighbors within the network, or exogenous events, where users take actions due to drives external to the
network. How much external drive should be provided to each user, such that the
network activity can be steered towards a target state? In this paper, we model
social events using multivariate Hawkes processes, which can capture both endogenous and exogenous event intensities, and derive a time dependent linear relation between the intensity of exogenous events and the overall network activity.
Exploiting this connection, we develop a convex optimization framework for determining the required level of external drive in order for the network to reach a
desired activity level. We experimented with event data gathered from Twitter,
and show that our method can steer the activity of the network more accurately
than alternatives.
1
Introduction
Online social platforms routinely track and record a large volume of event data, which may correspond to the usage of a service (e.g., url shortening service, bit.ly). These events can be categorized
roughly into endogenous events, where users just respond to the actions of their neighbors within
the network, or exogenous events, where users take actions due to drives external to the network.
For instance, a user?s tweets may contain links provided by bit.ly, either due to his forwarding of a
link from his friends, or due to his own initiative to use the service to create a new link.
Can we model and exploit these data to steer the online community to a desired activity level?
Specifically, can we drive the overall usage of a service to a certain level (e.g., at least twice per
day per user) by incentivizing a small number of users to take more initiatives? What if the goal is
to make the usage level of a service more homogeneous across users? What about maximizing the
overall service usage for a target group of users? Furthermore, these activity shaping problems need
to be addressed by taking into account budget constraints, since incentives are usually provided in
the form of monetary or credit rewards.
Activity shaping problems are significantly more challenging than traditional influence maximization problems, which aim to identify a set of users, who, when convinced to adopt a product, shall
influence others in the network and trigger a large cascade of adoptions [1, 2]. First, in influence
maximization, the state of each user is often assumed to be binary, either adopting a product or
not [1, 3, 4, 5]. However, such assumption does not capture the recurrent nature of product usage,
where the frequency of the usage matters. Second, while influence maximization methods identify
a set of users to provide incentives, they do not typically provide a quantitative prescription on how
much incentive should be provided to each user. Third, activity shaping concerns a larger variety of
target states, such as minimum activity and homogeneity of activity, not just activity maximization.
In this paper, we will address the activity shaping problems using multivariate Hawkes processes [6],
which can model both endogenous and exogenous recurrent social events, and were shown to be a
good fit for such data in a number of recent works (e.g., [7, 8, 9, 10, 11, 12]). More importantly,
1
we will go beyond model fitting, and derive a novel predictive formula for the overall network activity given the intensity of exogenous events in individual users, using a connection between the
processes and branching processes [13, 14, 15, 16]. Based on this relation, we propose a convex
optimization framework to address a diverse range of activity shaping problems given budget constraints. Compared to previous methods for influence maximization, our framework can provide
more fine-grained control of network activity, not only steering the network to a desired steady-state
activity level but also do so in a time-sensitive fashion. For example, our framework allows us to
answer complex time-sensitive queries, such as, which users should be incentivized, and by how
much, to steer a set of users to use a product twice per week after one month?
In addition to the novel framework, we also develop an efficient gradient based optimization algorithm, where the matrix exponential needed for gradient computation is approximated using the
truncated Taylor series expansion [17]. This algorithm allows us to validate our framework in a
variety of activity shaping tasks and scale up to networks with tens of thousands of nodes. We also
conducted experiments on a network of 60,000 Twitter users and more than 7,500,000 uses of a popular url shortening services. Using held-out data, we show that our algorithm can shape the network
behavior much more accurately than alternatives.
2
Modeling Endogenous-Exogenous Recurrent Social Events
We model the events generated by m users in a social network as a m-dimensional counting process
N (t) = (N1 (t), N2 (t), . . . , Nm (t))" , where Ni (t) records the total number of events generated by
user i up to time t. Furthermore, we represent each event as a tuple (ui , ti ), where ui is the user identity and ti is the event timing. Let the history of the process up to time t be Ht := {(ui , ti ) | ti ! t},
and Ht? be the history until just before time t. Then the increment of the process, dN (t), in an infinitesimal window [t, t + dt] is parametrized by the intensity ?(t) = (?1 (t), . . . , ?m (t))" " 0, i.e.,
E[dN (t)|Ht? ] = ?(t) dt.
(1)
Intuitively, the larger the intensity ?(t), the greater the likelihood of observing an event in the time
window [t, t + dt]. For instance, a Poisson process in [0, ?) can be viewed as a special counting
process with a constant intensity function ?, independent of time and history. To model the presence
of both endogenous and exogenous events, we will decompose the intensity into two terms
?(t)
!"#$
overall event intensity
=
?(0) (t)
! "# $
exogenous event intensity
+
?? (t)
! "# $
,
(2)
endogenous event intensity
where the exogenous event intensity models drive outside the network, and the endogenous event
intensity models interactions within the network. We assume that hosts of social platforms can
potentially drive up or down the exogenous events intensity by providing incentives to users; while
endogenous events are generated due to users? own interests or under the influence of network peers,
and the hosts do not interfere with them directly. The key questions in the activity shaping context
are how to model the endogenous event intensity which are realistic to recurrent social interactions,
and how to link the exogenous event intensity to the endogenous event intensity. We assume that the
exogenous event intensity is independent of the history and time, i.e., ?(0) (t) = ?(0) .
2.1 Multivariate Hawkes Process
Recurrent endogenous events often exhibit the characteristics of self-excitation, where a user tends
to repeat what he has been doing recently, and mutual-excitation, where a user simply follows what
his neighbors are doing due to peer pressure. These social phenomena have been made analogy to
the occurrence of earthquake [18] and the spread of epidemics [19], and can be well-captured by
multivariate Hawkes processes [6] as shown in a number of recent works (e.g., [7, 8, 9, 10, 11, 12]).
More specifically, a multivariate Hawkes process is a counting process who has a particular form
of intensity. We assume that the strength of influence between users is parameterized by a sparse
nonnegative influence matrix A = (auu! )u,u! ?[m] , where auu! > 0 means user u% directly excites
user u. We also allow A to have nonnegative diagonals to model self-excitation of a user. Then, the
intensity of the u-th dimension is
& t
%
%
?
?u (t) =
auui g(t ? ti ) =
auu!
g(t ? s) dNu! (s),
(3)
i:ti <t
u! ?[m]
0
'?
where g(s) is a nonnegative kernel function such that g(s) = 0 for s ? 0 and 0 g(s) ds <
?; the second equality is obtained by grouping events according to users and use the fact that
2
1
2
1
3
t1
3
1
2
5
4
2
1
3
5
1
5
1
6
1
5
t2
2
3
t3
3
6
5
5
6
3
4
2
2 3 1
1 2 4
t
(a) An example social network
(b) Branching structure of events
Figure 1: In Panel (a), each directed edge indicates that the target node follows, and can be influenced
by, the source node. The activity in this network is modeled using Hawkes processes, which result in
branching structure of events shown in Panel (b). Each exogenous event is the root node of a branch
(e.g., top left most red circle at t1 ), and it occurs due to a user?s own initiative; and each event can
trigger one or more endogenous events (blue square at t2 ). The new endogenous events can create
the next generation of endogenous events (green triangles at t3 ), and so forth. The social network
will constrain the branching structure of events, since an event produced by a user (e.g., user 1) can
only trigger endogenous events in the same user or one or more of her followers (e.g., user 2 or 3).
't
(
?
g(t ? s) dNu! (s) =
ui =u! ,ti <t g(t ? ti ). Intuitively, ?u (t) models the propagation of peer
0
influence over the network ? each event (ui , ti ) occurred in the neighbor of a user will boost her
intensity by a certain amount which itself decays over time. Thus, the more frequent the events
occur in the user?s neighbor, the more likely she will be persuaded to generate a new event.
For simplicity, we will focus on an exponential kernel, g(t ? ti ) = exp(??(t ? ti )) in the reminder
of the paper. However, multivariate Hawkes processes and the branching processed explained in
next section is independent of the kernel choice and can be extended to other kernels such as powerlaw, Rayleigh or any other long tailed distribution over nonnegative real domain. Furthermore, we
can rewrite equation (3) in vectorial format
& t
?
? (t) =
G(t ? s) dN (s),
(4)
0
by defining a m ? m time-varying matrix G(t) = (auu! g(t))u,u! ?[m] . Note that, for multivariate
Hawkes processes, the intensity, ?(t), itself is a random quantity, which depends on the history Ht .
We denote the expectation of the intensity with respect to history as
?(t) := EHt? [?(t)]
(5)
2.2 Connection to Branching Processes
A branching process is a Markov process that models a population in which each individual in
generation k produces some random number of individuals in generation k + 1, according some
distribution [20]. In this section, we will conceptually assign both exogenous events and endogenous
events in the multivariate Hawkes process to levels (or generations), and associate these events with
a branching structure which records the information on which event triggers which other events (see
Figure 1 for an example). Note that this genealogy of events should be interpreted in probabilistic
terms and may not be observed in actual data. Such connection has been discussed in Hawkes?
original paper on one dimensional Hawkes processes [21], and it has recently been revisited in the
context of multivariate Hawkes processes by [11]. The branching structure will play a crucial role in
deriving a novel link between the intensity of the exogenous events and the overall network activity.
More specifically, we assign all exogenous events to the zero-th generation, and record the number
of such events as N (0) (t). These exogenous events will trigger the first generation of endogenous
events whose number will be recorded as N (1) (t). Next these first generation of endogenous events
will further trigger a second generation of endogenous events N (2) (t), and so on. Then the total
number of events in the network is the sum of the numbers of events from all generations
N (t) = N (0) (t) + N (1) (t) + N (2) (t) + . . .
(k?1)
Ht
.
(6)
Furthermore, denote all events in generation k ? 1 as
Then, independently for each event
(k?1)
(ui , ti ) ? Ht
in generation k ? 1, it triggers a Poisson process in its neighbor u independently
with intensity auui g(t?ti ). Due to the superposition theorem of independent Poisson processes [22],
3
(k)
the intensity, ?u (t), of events at node u and generation k is simply the sum of conditional intensities
(
(k)
of the Poisson processes triggered by all its neighbors, i.e., ?u (t) = (ui ,ti )?H(k?1) auui g(t ?
t
't
(
(k?1)
ti ) =
(s). Concatenate the intensity for all u ? [m], and use the
u! ?[m] 0 g(t ? s) dNu!
time-varying matrix G(t) (4), we have
& t
?(k) (t) =
G(t ? s) dN (k?1) (s),
(7)
0
(k)
(k)
(?1 (t), . . . , ?m (t))"
where ? (t) =
is the intensity for counting process N (k) (t) at k-th generation. Again, due to the superposition of independent Poisson processes, we can decompose the
intensity of N (t) into a sum of conditional intensities from different generation
(k)
?(t) = ?(0) (t) + ?(1) (t) + ?(2) (t) + . . .
(8)
Next, based on the above decomposition, we will develop a closed form relation between the expected intensity ?(t) = EHt? [?(t)] and the intensity, ?(0) (t), of the exogenous events. This relation will form the basis of our activity shaping framework.
3
Linking Exogenous Event Intensity to Overall Network Activity
Our strategy is to first link the expected intensity ?(k) (t) := EHt? [?(k) (t)] of events at the k-th
generation with ?(0) (t), and then derive a close form for the infinite series sum
?(t) = ?(0) (t) + ?(1) (t) + ?(2) (t) + . . .
(9)
Define a series of auto-convolution matrices, one for each generation, with G (t) = I and
& t
G(!k) (t) =
G(t ? s) G(!k?1) (s) ds = G(t) # G(!k?1) (t)
(10)
(!0)
0
Then the expected intensity of events at the k-th generation is related to exogenous intensity ?(0) by
Lemma 1 ?(k) (t) = G(!k) (t) ?(0) .
Next, by summing together all auto-convolution matrices,
?(t) := I + G(!1) (t) + G(!2) (t) + . . .
we obtain a linear relation between the expected intensity of the network and the intensity of the
exogenous events, i.e., ?(t) = ?(t)?(0) . The entries in the matrix ?(t) roughly encode the ?influence? between pairs of users. More precisely, the entry ?uv (t) is the expected intensity of events
at node u due to a unit level of exogenous intensity at(node v. We can also derive several other
useful quantities from ?(t). For example, ??v (t) := u ?uv (t) can be thought of as the overall
influence user v has on all users. Surprisingly, for exponential kernel, the infinite sum of matrices
results in a closed form using matrix exponentials. First, let )? denote the Laplace transform of a
function, and we have the following intermediate results on the Laplace transform of G(!k) (t).
) (!k) (z) =
Lemma 2 G
'?
0
G(!k) (t) dt =
1
z
?
Ak
(z+?)k
With Lemma 2, we are in a position
* to prove our main theorem below:
+
Theorem 3 ?(t) = ?(t)?(0) = e(A??I)t + ?(A ? ?I)?1 (e(A??I)t ? I) ?(0) .
Theorem 3 provides us a linear relation between exogenous event intensity and the expected overall
intensity at any point in time but not just stationary intensity. The significance of this result is that
it allows us later to design a diverse range of convex programs to determine the intensity of the
exogenous event in order to achieve a target intensity.
In fact, we can recover the previous results in the stationary case as a special case of our general
result. More specifically, a multivariate Hawkes process is stationary if the spectral radius
,& ?
-.
& ?
/
A
? :=
G(t) dt =
g(t) dt
auu!
=
(11)
!
?
u,u ?[m]
0
0
is strictly smaller than 1 [6]. In this case, the expected intensity is ? = (I ? ?)?1 ?(0) independent
of the time. We can obtain this relation from theorem 3 if we let t ? ?.
?1
Corollary 4 ? = (I ? ?) ?(0) = limt?? ?(t) ?(0) .
Refer to Appendix A for all the proofs.
4
4
Convex Activity Shaping Framework
Given the linear relation between exogenous event intensity and expected overall event intensity, we
now propose a convex optimization framework for a variety of activity shaping tasks. In all tasks
discussed below, we will optimize the exogenous event intensity ?(0) such that the expected overall
event intensity ?(t) is maximized with respect to some concave utility U (?) in ?(t), i.e.,
maximize?(t),?(0) U (?(t))
(12)
subject to
?(t) = ?(t)?(0) , c" ?(0) ! C, ?(0) " 0
where c = (c1 , . . . , cm )" " 0 is the cost per unit event for each user and C is the total budget.
Additional regularization can also be added to ?(0) either to restrict the number of incentivized
users (with $0 norm '?(0) '0 ), or to promote a sparse solution (with $1 norm '?(0) '1 , or to obtain a
smooth solution (with $2 regularization '?(0) '2 ). We next discuss several instances of the general
framework which achieve different goals (their constraints remain the same and hence omitted).
Capped Activity Maximization. In real networks, there is an upper bound (or a cap) on the activity
each user can generate due to limited attention of a user. For example, a Twitter user typically posts
a limited number of shortened urls or retweets a limited number of tweets [23]. Suppose we know
the upper bound, ?u , on a user?s activity, i.e., how much activity each user is willing to generate.
Then we can perform the following capped activity maximization task
(
maximize?(t),?(0)
(13)
u?[m] min {?u (t), ?u }
Minimax Activity Shaping. Suppose our goal is instead maintaining the activity of each user in the
network above a certain minimum level, or, alternatively make the user with the minimum activity
as active as possible. Then, we can perform the following minimax activity shaping task
maximize?(t),?(0) minu ?u (t)
(14)
Least-Squares Activity Shaping. Sometimes we want to achieve a pre-specified target activity
levels, v, for users. For example, we may like to divide users into groups and desire a different level
of activity in each group. Inspired by these examples, we can perform the following least-squares
activity shaping task
maximize?(t),?(0) ?'B?(t) ? v'22
(15)
where B encodes potentially additional constraints (e.g., group partitions). Besides Euclidean distance, the family of Bregman divergences can be used to measure the difference between B?(t)
and v here. That is, given a function f (?) : Rm (? R convex in its argument, we can use
D(B?(t)'v) := f (B?(t)) ? f (v) ? )?f (v), B?(t) ? v+ as our objective function.
Activity Homogenization. Many other concave utility functions can be used. For example, we may
want to steer users activities to a more homogeneous profile. If we measure homogeneity of activity
with Shannon entropy, then we can perform the following activity homogenization task
(
maximize?(t),?(0) ? u?[m] ?u (t) ln ?u (t)
(16)
5
Scalable Algorithm
All the activity shaping problems defined above require an efficient evaluation of the instantaneous
average intensity ?(t) at time t, which entails computing matrix exponentials to obtain ?(t). In
small or medium networks, we can rely on well-known numerical methods to compute matrix exponentials [24]. However, in large networks, the explicit computation of ?(t) becomes intractable.
Fortunately, we can exploit the following key property of our convex activity shaping framework:
the instantaneous average intensity only depends on ?(t) through matrix-vector product operations.
In particular, we start by using Theorem 3* to rewrite the +multiplication of ?(t) and a vector v
as ?(t)v = e(A??I)t v + ?(A ? ?I)?1 e(A??I)t v ? v . We then get a tractable solution by
first computing e(A??I)t v *efficiently, subtracting
v from it, and solving a sparse linear system of
+
equations, (A ? ?I)x = e(A??I)t v ? v , efficiently. The steps are illustrated in Algorithm 1.
Next, we elaborate on two very efficient algorithms for computing the product of matrix exponential
with a vector and for solving a sparse linear system of equations.
For the computation of the product of matrix exponential with a vector, we rely on the iterative
algorithm by Al-Mohy et al. [17], which combines a scaling and squaring method with a truncated
Taylor series approximation to the matrix exponential. For solving the sparse linear system of equa5
Algorithm 1: Average Instantaneous Intensity
Algorithm 2: PGD for Activity Shaping
input : A, ?, t, v
output: ?(t)v
v1 = e(A??I)t v
v2 = v2 ? v;
v3 = (A ? ?I)?1 v2
return v1 + ?v3 ;
Initialize ?(0) ;
repeat
1- Project ?(0) into ?(0) " 0, c! ?(0) ! C;
2- Evaluate the gradient g(?(0) ) at ?(0) ;
3- Update ?(0) using the gradient g(?(0) );
until convergence;
tion, we use the well-known GMRES method [25], which is an Arnoldi process for constructing
an l2 -orthogonal basis of Krylov subspaces. The method solves the linear system by iteratively
minimizing the norm of the residual vector over a Krylov subspace.
Perhaps surprisingly, we will now show that it is possible to compute the gradient of the objective functions of all our activity shaping problems using the algorithm developed above for computing the average instantaneous intensity. We only need to define the vector v appropriately
for each problem, as follows: (i) Activity maximization: g(?(0) ) = ?(t)" v, where v is defined such that vj = 1 if ?j > ?j , and vj = 0, otherwise. (ii) Minimax activity shaping:
g(?(0) ) = ?(t)" e, where e is defined such that ej = 1 *if ?j = ?min , and
+ ej = 0, otherwise. (iii)
Least-squares activity shaping: g(?(0) ) = 2?(t)" B " B?(t)?(0) ? v . (iv) Activity homogenization: g(?(0) ) = ?(t)" ln (?(t)?(0) ) + ?(t)" 1, where ln(?) on a vector is the element-wise
natural logarithm. Since the activity maximization and the minimax activity shaping tasks require
only one evaluation of ?(t) times a vector, Algorithm 1 can be used directly. However, computing
the gradient for least-squares activity shaping and activity homogenization is slightly more involved
and it requires to be careful with the order in which we perform the operations (Refer to Appendix B
for details). Equipped with an efficient way to compute of gradients, we solve the corresponding
convex optimization problem for each activity shaping problem by applying projected gradient descent (PGD) [26] with the appropriate gradient1 . Algorithm 2 summarizes the key steps.
6
Experimental Evaluation
We evaluate our framework using both simulated and real world held-out data, and show that our
approach significantly outperforms several baselines. The appendix contains additional experiments.
Dataset description and network inference. We use data gathered from Twitter as reported in [27],
which comprises of all public tweets posted by 60,000 users during a 8-month period, from January
2009 to September 2009. For every user, we record the times she uses any of six popular url shortening services (refer to Appendix C for details). We evaluate the performance of our framework on
a subset of 2,241 active users, linked by 4,901 edges, which we call 2K dataset, and we evaluate its
scalability on the overall 60,000 users, linked by ? 200,000 edges, which we call 60K dataset. The
2K dataset accounts for 691,020 url shortened service uses while the 60K dataset accounts for ?7.5
million uses. Finally, we treat each service as independent cascades of events.
In the experiments, we estimated the nonnegative influence matrix A and the exogenous intensity
?(0) using maximum log-likelihood, as in previous work [8, 9, 12]. We used a temporal resolution
of one minute and selected the bandwidth ? = 0.1 by cross validation. Loosely speaking, ? = 0.1
corresponds to loosing 70% of the initial influence after 10 minutes, which may be explained by the
rapid rate at which each user? news feed gets updated.
Evaluation schemes. We focus on three tasks: capped activity maximization, minimax activity
shaping, and least square activity shaping. We set the total budget to C = 0.5, which corresponds
to supporting a total extra activity equal to 0.5 actions per unit time, and assume all users entail the
same cost. In the capped activity maximization, we set the upper limit of each user?s intensity, ?,
by adding a nonnegative random vector to their inferred initial intensity. In the least-squares activity
shaping, we set B = I and aim to create three user groups: less-active, moderate, and super-active.
We use three different evaluation schemes, with an increasing resemblance to a real world scenario:
Theoretical objective: We compute the expected overall (theoretical) intensity by applying Theorem 3 on the optimal exogenous event intensities to each of the three activity shaping tasks, as well
as the learned A and ?. We then compute and report the value of the objective functions.
1
For nondifferential objectives, subgradient algorithms can be used instead.
6
K
G
0
?4
4
3
0.4
0.2
0
2
0 1 2 3 4 5 6 7 8 9
logarithm of time
?4
1.2
0 1 2 3 4 5 6 7 8 9
logarithm of time
0.4
0.2
0
D
1.4
0.6
GR
1.6
0.8
LS
LSGRD
OP
1.2
0 1 2 3 4 5 6 7 8 9
logarithm of time
PROP
PR
1.4
LSASH
H
1.6
1.8 x 10
AS
LSGRD
LS
PROP
rank correlation
LSASH
Euclidean distance
?4
D
5
*
0.6
GR
GRD
LP
LP
I
2
MINMU
MU
4
UNI
6
MI
N
MMASH
UN
GRD
AS
H
LP
MM
MINMU
rank correlation
x 10
UNI
minimum activity
minimum activity
0.5
0 1 2 3 4 5 6 7 8 9
logarithm of time
0
0 1 2 3 4 5 6 7 8 9
logarithm of time
Euclidean distance
*
1
PR
0.6
?4
1.8 x 10
PRK
0.65
x 10
MMASH
DEG
0.7
0 1 2 3 4 5 6 7 8 9
logarithm of time
6
WEI
DE
0.6
XMU
U
0.65
CAM
WE
I
0.7
0.75
M
PRK
XM
DEG
CA
WEI
rank correlation
XMU
sum of users? activity
sum of users? activity
CAM
0.75
(a) Theoretical objective
(b) Simulated objective
(c) Held-out data
Figure 2: Row 1: Capped activity maximization. Row 2: Minimax activity shaping. Row 3: Leastsquares activity shaping. * means statistical significant at level of 0.01 with paired t-test between
our method and the second best
Simulated objective: We simulate 50 cascades with Ogata?s thinning algorithm [28], using the optimal exogenous event intensities to each of the three activity shaping tasks, and the learned A and ?.
We then estimate empirically the overall event intensity based on the simulated cascades, by computing a running average over non-overlapping time windows, and report the value of the objective
functions based on this estimated overall intensity. Appendix D provides a comparison between the
simulated and the theoretical objective.
Held-out data: The most interesting evaluation scheme would entail carrying out real interventions
in a social platform. However, since this is very challenging to do, instead, in this evaluation scheme,
we use held-out data to simulate such process, proceeding as follows. We first partition the 8-month
data into 50 five-day long contiguous intervals. Then, we use one interval for training and the
remaining 49 intervals for testing. Suppose interval 1 is used for training, the procedure is as follows:
(0)
1. We estimate A1 , ?1 and ?1 using the events from interval 1. Then, we fix A1 and ?1 ,
(0)
and estimate ?i for all other intervals, i = 2, . . . , 49.
(0)
2. Given A1 and ?1 , we find the optimal exogenous event intensities, ?opt , for each of the
three activity shaping task, by solving the associated convex program. We then sort the
(0)
(0)
estimated ?i (i = 2, . . . , 49) according to their similarity to ?opt , using the Euclidean
(0)
(0)
distance '?opt ? ?i '2 .
3. We estimate the overall event intensity for each of the 49 intervals (i = 2, . . . , 49), as in the
?simulated objective? evaluation scheme, and sort these intervals according to the value of
their corresponding objective function.
4. Last, we compute and report the rank correlation score between the two orderings obtained
in step 2 and 3.2 The larger the rank correlation, the better the method.
We repeat this procedure 50 times, choosing each different interval for training once, and compute
and report the average rank correlations. More details can be found in the appendix.
2
rank correlation = number of pairs with consistent ordering / total number of pairs.
7
Capped activity maximization (CAM). We compare to a number of alternatives. XMU: heuristic
based on ?(t) without optimization; DEG and WEI: heuristics based on the degree of the user;
PRANK: heuristic based on page rank (refer to Appendix C for further details). The first row of
Figure 2 summarizes the results for the three different evaluation schemes. We find that our method
(CAM) consistently outperforms the alternatives. For the theoretical objective, CAM is 11 % better
than the second best, DEG. The difference in overall users? intensity from DEG is about 0.8 which,
roughly speaking, leads to at least an increase of about 0.8 ? 60 ? 24 ? 30 = 34, 560 in the overall
number of events in a month. In terms of simulated objective and held-out data, the results are
similar and provide empirical evidence that, compared to other heuristics, degree is an appropriate
surrogate for influence, while, based on the poor performance of XMU, it seems that high activity
does not necessarily entail being influential. To elaborate on the interpretability of the real-world
experiment on held-out data, consider for example the difference in rank correlation between CAM
and DEG, which is almost 0.1. Then, roughly speaking, this means that incentivizing users based
on our approach accommodates with the ordering of real activity patterns in 0.1 ? 50?49
= 122.5
2
more pairs of realizations.
Minimax activity shaping (MMASH). We compare to a number of alternatives. UNI: heuristic
based on equal allocation; MINMU: heuristic based on ?(t) without optimization; LP: linear programming based heuristic; GRD: a greedy approach to leverage the activity (see Appendix C for
more details). The second row of Figure 2 summarizes the results for the three different evaluation
schemes. We find that our method (MMASH) consistently outperforms the alternatives. For the theoretical objective, it is about 2? better than the second best, LP. Importantly, the difference between
MMASH and LP is not trifling and the least active user carries out 2?10?4 ?60?24?30 = 4.3 more
actions in average over a month. As one may have expected, GRD and LP are the best among the
heuristics. The poor performance of MINMU, which is directly related to the objective of MMASH,
may be because it assigns the budget to a low active user, regardless of their influence. However,
our method, by cleverly distributing the budget to the users whom actions trigger many other users?
actions (like those ones with low activity), it benefits from the budget most. In terms of simulated
objective and held-out data, the algorithms? performance become more similar.
Least-squares activity shaping (LSASH). We compare to two alternatives. PROP: Assigning the
budget proportionally to the desired activity; LSGRD: greedily allocating budget according the difference between current and desired activity (refer to Appendix C for more details). The third row of
Figure 2 summarizes the results for the three different evaluation schemes. We find that our method
(LSASH) consistently outperforms the alternatives. Perhaps surprisingly, PROP, despite its simplicity, seems to perform slightly better than LSGRD. This is may be due to the way it allocates the
budget to users, e.g., it does not aim to strictly fulfill users? target activity but benefit more users by
assigning budget proportionally. Refer to Appendix E for additional experiments.
Sparsity and Activity Shaping. In some applications there is a limitation on the number of users we
can incentivize. In our proposed framework, we can handle this requirement by including a sparsity
constraint on the optimization problem. In order to maintain the convexity of the optimization
problem, we consider a l1 regularization term, where a regularization parameter ? provides the
trade-off between sparsity and the activity shaping goal. Refer to Appendix F for more details and
experimental results for different values of ?.
Scalability. The most computationally demanding part of the proposed algorithm is the evaluation of
matrix exponentials, which we scale up by utilizing techniques from matrix algebra, such as GMRES
and Al-Mohy methods. As a result, we are able to run our methods in a reasonable amount of time
on the 60K dataset, specifically, in comparison with a naive implementation of matrix exponential
evaluations. Refer to Appendix G for detailed experimental results on scalability.
Appendix H discusses the limitations of our framework and future work.
Acknowledgement. This project was supported in part by NSF IIS1116886, NSF/NIH BIGDATA
1R01GM108341, NSF CAREER IIS1350983 and Raytheon Faculty Fellowship to Le Song. Isabel Valera acknowledge the support of Plan Regional-Programas I+D of Comunidad de Madrid
(AGES-CM S2010/BMD-2422), Ministerio de Ciencia e Innovaci?on of Spain (project DEIPRO
TEC2009-14504-C02-00 and program Consolider-Ingenio 2010 CSD2008-00010 COMONSENS).
8
References
? Tardos. Maximizing the spread of influence through a social
[1] David Kempe, Jon Kleinberg, and Eva
network. In KDD, pages 137?146. ACM, 2003.
[2] Matthew Richardson and Pedro Domingos. Mining knowledge-sharing sites for viral marketing. In KDD,
pages 61?70. ACM, 2002.
[3] Wei Chen, Yajun Wang, and Siyu Yang. Efficient influence maximization in social networks. In KDD,
pages 199?208. ACM, 2009.
[4] Manuel G. Rodriguez and Bernard Sch?olkopf. Influence maximization in continuous time diffusion
networks. In ICML, 2012.
[5] Nan Du, Le Song, Manuel Gomez Rodriguez, and Hongyuan Zha. Scalable influence estimation in
continuous-time diffusion networks. In NIPS 26, 2013.
[6] Thomas .J. Liniger. Multivariate Hawkes Processes. PhD thesis, SWISS FEDERAL INSTITUTE OF
TECHNOLOGY ZURICH, 2009.
[7] Charles Blundell, Jeff Beck, and Katherine A Heller. Modelling reciprocating relationships with Hawkes
processes. In NIPS, 2012.
[8] Ke Zhou, Hongyuan Zha, and Le Song. Learning social infectivity in sparse low-rank networks using
multi-dimensional Hawkes processes. In AISTATS, 2013.
[9] Ke Zhou, Hongyuan Zha, and Le Song. Learning triggering kernels for multi-dimensional Hawkes processes. In ICML, 2013.
[10] Tomoharu Iwata, Amar Shah, and Zoubin Ghahramani. Discovering latent influence in online social
activities via shared cascade poisson processes. In KDD, pages 266?274. ACM, 2013.
[11] Scott W Linderman and Ryan P Adams. Discovering latent network structure in point process data. arXiv
preprint arXiv:1402.0914, 2014.
[12] Isabel Valera, Manuel Gomez-Rodriguez, and Krishna Gummadi. Modeling adoption of competing products and conventions in social media. arXiv preprint arXiv:1406.0516, 2014.
[13] Ian Dobson, Benjamin A Carreras, and David E Newman. A branching process approximation to cascading load-dependent system failure. In System Sciences, 2004. Proceedings of the 37th Annual Hawaii
International Conference on, pages 10?pp. IEEE, 2004.
[14] Jakob Gulddahl Rasmussen. Bayesian inference for Hawkes processes. Methodology and Computing in
Applied Probability, 15(3):623?642, 2013.
[15] Alejandro Veen and Frederic P Schoenberg. Estimation of space?time branching process models in seismology using an em?type algorithm. JASA, 103(482):614?624, 2008.
[16] Jiancang Zhuang, Yosihiko Ogata, and David Vere-Jones. Stochastic declustering of space-time earthquake occurrences. JASA, 97(458):369?380, 2002.
[17] Awad H Al-Mohy and Nicholas J Higham. Computing the action of the matrix exponential, with an
application to exponential integrators. SIAM journal on scientific computing, 33(2):488?511, 2011.
[18] David Marsan and Olivier Lengline. Extending earthquakes? reach through cascading. Science,
319(5866):1076?1079, 2008.
[19] Shuang-Hong Yang and Hongyuan Zha. Mixture of mutually exciting processes for viral diffusion. In
ICML, pages 1-9, 2013.
[20] Theodore E Harris. The theory of branching processes. Courier Dover Publications, 2002.
[21] Alan G Hawkes. Spectra of some self-exciting and mutually exciting point processes. Biometrika,
58(1):83?90, 1971.
[22] John Frank Charles Kingman. Poisson processes, volume 3. Oxford university press, 1992.
[23] Manuel Gomez-Rodriguez, Krishna Gummadi, and Bernhard Schoelkopf. Quantifying Information Overload in Social Media and its Impact on Social Contagions. In ICWSM, 2014.
[24] Gene H Golub and Charles F Van Loan. Matrix computations, volume 3. JHU Press, 2012.
[25] Youcef Saad and Martin H Schultz. Gmres: A generalized minimal residual algorithm for solving nonsymmetric linear systems. SIAM Journal on scientific and statistical computing, 7(3):856?869, 1986.
[26] Stephen Boyd and Lieven Vandenberghe. Convex Optimization. Cambridge University Press, Cambridge,
England, 2004.
[27] Meeyoung Cha, Hamed Haddadi, Fabricio Benevenuto, and P Krishna Gummadi. Measuring User Influence in Twitter: The Million Follower Fallacy. In ICWSM, 2010.
[28] Yosihiko Ogata. On lewis? simulation method for point processes. Information Theory, IEEE Transactions
on, 27(1):23?31, 1981.
9
| 5365 |@word faculty:1 norm:3 seems:2 auu:5 consolider:1 cha:1 willing:1 simulation:1 decomposition:1 pressure:1 carry:1 initial:2 series:4 contains:1 score:1 outperforms:4 yajun:1 current:1 manuel:5 assigning:2 follower:2 vere:1 tec2009:1 john:1 realistic:1 concatenate:1 partition:2 numerical:1 shape:1 ministerio:1 kdd:4 prk:2 update:1 s2010:1 stationary:3 greedy:1 selected:1 discovering:2 ivalera:1 dover:1 record:5 provides:3 node:7 revisited:1 org:1 five:1 dn:4 become:1 initiative:3 prove:1 fitting:1 combine:1 expected:11 rapid:1 behavior:1 roughly:5 multi:2 integrator:1 inspired:1 actual:1 iis1350983:1 window:3 equipped:1 increasing:1 becomes:1 provided:4 project:3 spain:1 panel:2 medium:3 what:4 cm:2 interpreted:1 developed:1 deipro:1 temporal:1 quantitative:1 every:1 ti:15 concave:2 biometrika:1 rm:1 control:1 unit:3 ly:2 intervention:1 arnoldi:1 before:1 service:10 manuelgr:1 timing:1 t1:2 tends:1 treat:1 limit:1 infectivity:1 despite:1 shortened:2 ak:1 oxford:1 twice:2 theodore:1 forwarding:1 challenging:2 limited:3 range:2 adoption:2 bmd:1 directed:1 earthquake:3 testing:1 swiss:1 procedure:2 veen:1 empirical:1 jhu:1 significantly:2 cascade:5 thought:1 courier:1 pre:1 boyd:1 zoubin:1 get:2 close:1 context:2 influence:21 applying:2 optimize:1 maximizing:2 go:1 attention:1 regardless:1 independently:2 convex:10 l:2 resolution:1 ke:2 simplicity:2 assigns:1 powerlaw:1 utilizing:1 importantly:2 deriving:1 cascading:2 vandenberghe:1 his:4 population:1 handle:1 yosihiko:2 increment:1 laplace:2 updated:1 tardos:1 target:7 trigger:8 play:1 user:75 suppose:3 programming:1 homogeneous:2 us:4 olivier:1 domingo:1 associate:1 element:1 approximated:1 marsan:1 observed:1 role:1 preprint:2 wang:1 capture:2 thousand:1 eva:1 news:1 schoelkopf:1 ordering:3 trade:1 benjamin:1 mu:1 ui:7 convexity:1 reward:1 ciencia:1 cam:6 carrying:1 rewrite:2 solving:5 algebra:1 predictive:1 basis:2 triangle:1 isabel:3 routinely:1 univ:1 query:1 newman:1 outside:1 choosing:1 peer:3 whose:1 heuristic:8 larger:3 solve:1 otherwise:2 epidemic:1 richardson:1 amar:1 transform:2 itself:2 online:4 triggered:1 propose:2 subtracting:1 interaction:2 product:8 frequent:1 monetary:1 realization:1 achieve:3 forth:1 description:1 validate:1 scalability:3 olkopf:1 exploiting:1 convergence:1 requirement:1 extending:1 produce:1 adam:1 seismology:1 derive:4 develop:3 friend:1 recurrent:5 excites:1 op:1 lsong:1 solves:1 convention:1 radius:1 stochastic:1 public:1 require:2 assign:2 fix:1 decompose:2 opt:3 leastsquares:1 ryan:1 strictly:2 genealogy:1 mm:1 credit:1 exp:1 minu:1 week:1 matthew:1 adopt:1 omitted:1 estimation:2 superposition:2 sensitive:2 create:3 federal:1 aim:3 super:1 fulfill:1 zhou:2 ej:2 varying:2 gatech:2 publication:1 corollary:1 encode:1 focus:2 she:2 consistently:3 rank:10 likelihood:2 indicates:1 iis1116886:1 modelling:1 greedily:1 baseline:1 inference:2 twitter:5 dependent:2 squaring:1 typically:2 her:2 relation:8 overall:18 among:1 ingenio:1 gmres:3 platform:3 special:2 initialize:1 mutual:1 kempe:1 equal:2 once:1 plan:1 jones:1 icml:3 jon:1 promote:1 future:1 others:1 t2:2 report:4 divergence:1 homogeneity:2 individual:3 beck:1 comunidad:1 n1:1 maintain:1 interest:1 mining:1 evaluation:13 golub:1 mixture:1 r01gm108341:1 held:8 allocating:1 bregman:1 tuple:1 edge:3 allocates:1 orthogonal:1 iv:1 taylor:2 divide:1 euclidean:4 circle:1 desired:5 logarithm:7 loosely:1 theoretical:6 minimal:1 instance:3 modeling:2 steer:4 contiguous:1 measuring:1 maximization:15 cost:2 entry:2 subset:1 innovaci:1 shuang:1 conducted:1 gr:2 reported:1 answer:1 international:1 siam:2 probabilistic:1 off:1 together:1 again:1 thesis:1 nm:1 recorded:1 hawaii:1 external:4 steered:1 kingman:1 return:1 account:3 de:3 matter:1 depends:2 tion:1 later:1 root:1 endogenous:20 exogenous:31 observing:1 doing:2 red:1 zha:6 closed:2 carlos:1 recover:1 start:1 linked:2 sort:2 awad:1 square:8 ni:1 who:2 characteristic:1 maximized:1 gathered:2 correspond:1 identify:2 t3:2 efficiently:2 conceptually:1 bayesian:1 accurately:2 produced:1 cc:1 drive:7 history:6 hamed:1 reach:2 influenced:1 sharing:1 infinitesimal:1 failure:1 frequency:1 involved:1 pp:1 proof:1 mi:1 associated:1 dataset:6 popular:2 reminder:1 cap:1 knowledge:1 shaping:36 uc3m:1 thinning:1 feed:1 dt:6 day:2 methodology:1 wei:4 tomoharu:1 furthermore:4 just:5 marketing:1 until:2 d:2 correlation:8 overlapping:1 propagation:1 rodriguez:5 interfere:1 perhaps:2 resemblance:1 scientific:2 usage:6 dnu:3 contain:1 equality:1 regularization:4 hence:1 iteratively:1 illustrated:1 during:1 branching:12 self:3 hawkes:19 mpi:2 steady:1 excitation:3 hong:1 generalized:1 tsc:1 l1:1 wise:1 instantaneous:4 novel:3 recently:2 charles:3 nih:1 viral:2 empirically:1 volume:3 million:2 discussed:2 he:1 occurred:1 linking:1 reciprocating:1 nonsymmetric:1 homogenization:4 refer:8 significant:1 lieven:1 cambridge:2 uv:2 entail:4 similarity:1 grd:4 alejandro:1 carreras:1 multivariate:11 own:3 recent:2 moderate:1 scenario:1 certain:3 binary:1 captured:1 minimum:5 greater:1 additional:4 steering:1 fortunately:1 krishna:3 determine:1 maximize:5 v3:2 period:1 ii:1 branch:1 stephen:1 smooth:1 alan:1 england:1 cross:1 long:2 prescription:1 host:2 post:1 gummadi:3 paired:1 a1:3 impact:1 scalable:2 expectation:1 poisson:7 arxiv:4 represent:1 adopting:1 kernel:6 limt:1 sometimes:1 c1:1 addition:1 want:2 fine:1 fellowship:1 addressed:1 interval:9 source:1 crucial:1 appropriately:1 extra:1 sch:1 regional:1 saad:1 subject:1 call:2 counting:4 presence:1 intermediate:1 iii:2 leverage:1 yang:2 variety:3 fit:1 restrict:1 bandwidth:1 triggering:1 competing:1 dunan:1 blundell:1 six:1 utility:2 url:5 distributing:1 song:5 speaking:3 action:9 useful:1 proportionally:2 detailed:1 amount:2 shortening:3 ten:1 processed:1 generate:3 nsf:3 estimated:3 track:1 per:5 blue:1 diverse:2 dobson:1 schoenberg:1 shall:1 incentive:4 group:5 key:3 ht:6 diffusion:3 incentivize:1 v1:2 subgradient:1 tweet:3 sum:7 run:1 parameterized:1 respond:2 farajtabar:1 family:1 almost:1 reasonable:1 c02:1 appendix:13 scaling:1 summarizes:4 bit:2 bound:2 nan:2 gomez:4 nonnegative:6 activity:84 annual:1 strength:1 occur:1 vectorial:1 constraint:5 precisely:1 constrain:1 software:1 encodes:1 kleinberg:1 simulate:2 argument:1 min:2 format:1 martin:1 influential:1 according:5 poor:2 cleverly:1 across:1 smaller:1 remain:1 slightly:2 em:1 lp:7 intuitively:2 explained:2 pr:2 ln:3 equation:3 computationally:1 zurich:1 mutually:2 discus:2 needed:1 know:1 tractable:1 operation:2 linderman:1 v2:3 spectral:1 appropriate:2 nicholas:1 occurrence:2 alternative:8 shah:1 original:1 thomas:1 top:1 running:1 remaining:1 maintaining:1 sw:1 liniger:1 exploit:2 ghahramani:1 objective:17 question:1 quantity:2 occurs:1 added:1 strategy:1 mehrdad:2 traditional:1 diagonal:1 surrogate:1 exhibit:1 gradient:8 september:1 subspace:2 distance:4 link:6 incentivized:2 simulated:8 accommodates:1 parametrized:1 whom:1 besides:1 modeled:1 relationship:1 providing:1 minimizing:1 katherine:1 potentially:2 eht:3 frank:1 siyu:1 design:1 implementation:1 perform:6 upper:3 convolution:2 markov:1 acknowledge:1 descent:1 truncated:2 january:1 defining:1 extended:1 supporting:1 prank:1 jakob:1 community:1 intensity:65 inferred:1 david:4 pair:4 required:1 specified:1 comonsens:1 connection:4 learned:2 boost:1 nip:2 address:2 beyond:1 capped:6 krylov:2 usually:1 below:2 xm:1 pattern:1 able:1 scott:1 sparsity:3 program:3 green:1 interpretability:1 including:1 event:85 demanding:1 natural:1 rely:2 valera:3 residual:2 minimax:7 scheme:8 zhuang:1 technology:2 contagion:1 auto:2 naive:1 csd2008:1 heller:1 l2:1 acknowledgement:1 multiplication:1 determining:1 generation:17 interesting:1 allocation:1 limitation:2 analogy:1 age:1 validation:1 degree:2 jasa:2 consistent:1 exciting:3 row:6 convinced:1 repeat:3 surprisingly:3 last:1 supported:1 rasmussen:1 allow:1 institute:2 neighbor:7 taking:1 sparse:6 benefit:2 van:1 dimension:1 world:3 made:1 mohy:3 projected:1 schultz:1 social:20 transaction:1 uni:3 bernhard:1 gene:1 deg:6 active:6 hongyuan:5 summing:1 assumed:1 alternatively:1 spectrum:1 un:1 iterative:1 continuous:2 latent:2 tailed:1 fallacy:1 nature:1 ca:1 career:1 du:2 expansion:1 complex:1 posted:1 constructing:1 domain:1 vj:2 necessarily:1 aistats:1 significance:1 spread:2 main:1 profile:1 n2:1 categorized:2 site:1 madrid:2 elaborate:2 georgia:1 fashion:1 retweets:1 position:1 comprises:1 explicit:1 exponential:13 third:2 programas:1 grained:1 ian:1 ogata:3 incentivizing:3 formula:1 down:1 theorem:7 minute:2 load:1 experimented:1 decay:1 concern:1 grouping:1 intractable:1 evidence:1 frederic:1 adding:1 higham:1 phd:1 budget:11 chen:1 entropy:1 rayleigh:1 simply:2 likely:1 desire:1 pedro:1 corresponds:2 iwata:1 lewis:1 acm:4 harris:1 prop:4 conditional:2 goal:4 month:5 identity:1 viewed:1 careful:1 towards:1 loosing:1 jeff:1 shared:1 quantifying:1 loan:1 specifically:5 infinite:2 pgd:2 raytheon:1 lemma:3 total:6 bernard:1 e:1 experimental:3 shannon:1 support:1 icwsm:2 overload:1 bigdata:1 evaluate:4 phenomenon:1 |
4,822 | 5,366 | Learning Time-Varying Coverage Functions
Nan Du? , Yingyu Liang? , Maria-Florina Balcan , Le Song?
?
College of Computing, Georgia Institute of Technology
?
Department of Computer Science, Princeton University
School of Computer Science, Carnegie Mellon University
[email protected],[email protected]
[email protected],[email protected]
Abstract
Coverage functions are an important class of discrete functions that capture the
law of diminishing returns arising naturally from applications in social network
analysis, machine learning, and algorithmic game theory. In this paper, we propose a new problem of learning time-varying coverage functions, and develop a
novel parametrization of these functions using random features. Based on the connection between time-varying coverage functions and counting processes, we also
propose an efficient parameter learning algorithm based on likelihood maximization, and provide a sample complexity analysis. We applied our algorithm to the
influence function estimation problem in information diffusion in social networks,
and show that with few assumptions about the diffusion processes, our algorithm
is able to estimate influence significantly more accurately than existing approaches
on both synthetic and real world data.
1
Introduction
Coverage functions are a special class of the more general submodular functions which play important role in combinatorial optimization with many interesting applications in social network analysis [1], machine learning [2], economics and algorithmic game theory [3], etc. A particularly
important example of coverage functions in practice is the influence function of users in information
diffusion modeling [1] ? news spreads across social networks by word-of-mouth and a set of influential sources can collectively trigger a large number of follow-ups. Another example of coverage
functions is the valuation functions of customers in economics and game theory [3] ? customers are
thought to have certain requirements and the items being bundled and offered fulfill certain subsets
of these demands.
Theoretically, it is usually assumed that users? influence or customers? valuation are known in advance as an oracle. In practice, however, these functions must be learned. For example, given past
traces of information spreading in social networks, a social platform host would like to estimate
how many follow-ups a set of users can trigger. Or, given past data of customer reactions to different bundles, a retailer would like to estimate how likely customer would respond to new packages of
goods. Learning such combinatorial functions has attracted many recent research efforts from both
theoretical and practical sides (e.g., [4, 5, 6, 7, 8]), many of which show that coverage functions can
be learned from just polynomial number of samples.
However, the prior work has widely ignored an important dynamic aspect of the coverage functions.
For instance, information spreading is a dynamic process in social networks, and the number of
follow-ups of a fixed set of sources can increase as observation time increases. A bundle of items
or features offered to customers may trigger a sequence of customer actions over time. These real
world problems inspire and motivate us to consider a novel time-varying coverage function, f (S, t),
which is a coverage function of the set S when we fix a time t, and a continuous monotonic function
of time t when we fix a set S. While learning time-varying combinatorial structures has been ex1
plored in graphical model setting (e.g., [9, 10]), as far as we are aware of, learning of time-varying
coverage function has not been addressed in the literature. Furthermore, we are interested in estimating the entire function of t, rather than just treating the time t as a discrete index and learning
the function value at a small number of discrete points. From this perspective, our formulation is the
generalization of the most recent work [8] with even less assumptions about the data used to learn
the model.
Generally, we assume that the historical data are provided in pairs of a set and a collection of timestamps when caused events by the set occur. Hence, such a collection of temporal events associated
with a particular set Si can be modeled principally by a counting process Ni (t), t > 0 which is a
stochastic process with values that are positive, integer, and increasing along time [11]. For instance,
in the information diffusion setting of online social networks, given a set of earlier adopters of some
new product, Ni (t) models the time sequence of all triggered events of the followers, where each
jump in the process records the timing tij of an action. In the economics and game theory setting, the
counting process Ni (t) records the number of actions a customer has taken over time given a particular bundled offer. This essentially raises an interesting question of how to estimate the time-varying
coverage function from the angle of counting processes. We thus propose a novel formulation which
builds a connection between the two by modeling the cumulative intensity function of a counting
process as a time-varying coverage function. The key idea is to parametrize the intensity function
as a weighted combination of random kernel functions. We then develop an efficient learning algorithm TC OVERAGE L EARNER to estimate the parameters of the function using maximum likelihood
approach. We show that our algorithm can provably learn the time-varying coverage function using
only polynomial number of samples. Finally, we validate TC OVERAGE L EARNER on both influence
estimation and maximization problems by using cascade data from information diffusion. We show
that our method performs significantly better than alternatives with little prior knowledge about the
dynamics of the actual underlying diffusion processes.
2
Time-Varying Coverage Function
We will first give a formal definition of the time-varying coverage function, and then explain its
additional properties in details.
Definition. Let U be a (potentially uncountable) domain. We endow U with some ?-algebra A and
denote a probability distribution on U by P. A coverage function is a combinatorial function over a
finite set V of items, defined as
[
f (S) := Z ? P
Us , for all S ? 2V ,
(1)
s?S
where Us ? U is the subset of domain U covered by item s ? V, and Z is the additional normalization constant. For time-varying coverage functions, we let the size of the subset Us to grow
monotonically over time, that is
Us (t) ? Us (? ),
for all t 6 ? and s ? V,
which results in a combinatorial temporal function
[
f (S, t) = Z ? P
Us (t) ,
s?S
for all S ? 2V .
(2)
(3)
In this paper, we assume that f (S, t) is smooth and continuous, and its first order derivative with
respect to time, f 0 (S, t), is also smooth and continuous.
Representation. We now show that a time-varying coverage function, f (S, t), can be represented
as an expectation over random functions based on multidimensional step basis functions. Since
Us (t) is varying over time, we can associate each u ? U with a |V|-dimensional vector ?u of change
points. In particular, the s-th coordinate of ?u records the time that source node s covers u. Let ?
to be a random variable obtained by sampling u according to P and setting ? = ?u . Note that given
all ?u we can compute f (S, t); now we claim that the distribution of ? is sufficient.
We first introduce some notations. Based on ?u we define a |V|-dimensional step function ru (t) :
|V|
R+ 7? {0, 1} , where the s-th dimension of ru (t) is 1 if u is covered by the set Us (t) at time t, and
0 otherwise. To emphasize the dependence of the function ru (t) on ?u , we will also write ru (t) as
ru (t|?u ). We denote the indicator vector of a set S by ?S ? S
{0, 1}|V| where the s-th dimension of
?S is 1 if s ? S, and 0 otherwise. Then u ? U is covered by s?S Us (t) at time t if ?>
S ru (t) > 1.
2
Lemma 1. There exists a distribution Q(? ) over the vector of change points ? , such that the timevarying coverage function can be represented as
f (S, t) = Z ? E? ?Q(? ) ?(?>
(4)
S r(t|? ))
where ?(x) := min {x, 1}, and r(t|? ) is a multidimensional step function parameterized by ? .
S
Proof. Let US := s?S Us (t). By definition (3), we have the following integral representation
Z
Z
>
f (S, t) = Z ?
I {u ? US } dP(u) = Z ?
?(?>
S ru (t)) dP(u) = Z ? Eu?P(u) ?(?S ru (t)) .
U
U
We can define the setR of u having the same ? as U? := {u ? U | ?u = ? } and define a distribution
over ? as dQ(? ) := U? dP(u). Then the integral representation of f (S, t) can be rewritten as
>
Z ? Eu?P(u) ?(?>
S ru (t)) = Z ? E? ?Q(? ) ?(?S r(t|? )) ,
which proves the lemma.
3
Model for Observations
In general, we assume that the input data are provided in the form of pairs, (Si , Ni (t)), where Si is
a set, and Ni (t) is a counting process in which each jump of Ni (t) records the timing of an event.
We first give a brief overview of a counting process [11] and then motivate our model in details.
Counting Process. Formally, a counting process {N (t), t > 0} is any nonnegative, integer-valued
stochastic process such that N (t0 ) 6 N (t) whenever t0 6 t and N (0) = 0. The most common
use of a counting process is to count the number of occurrences of temporal events happening along
time, so the index set is usually taken to be the nonnegative real numbers R+ . A counting process
is a submartingale: E[N (t) | Ht0 ] > N (t0 ) for all t > t0 where Ht0 denotes the history up to time t0 .
By Doob-Meyer theorem [11], N (t) has the unique decomposition:
N (t) = ?(t) + M (t)
(5)
where ?(t) is a nondecreasing predictable process called the compensator (or cumulative intensity),
and M (t) is a mean zero martingale. Since E[dM (t) | Ht? ] = 0, where dM (t) is the increment of
M (t) over a small time interval [t, t + dt), and Ht? is the history until just before time t,
E[dN (t) | Ht? ] = d?(t) := a(t) dt
(6)
where a(t) is called the intensity of a counting process.
Model formulation. We assume that the cumulative intensity of the counting process is modeled
by a time-varying coverage function, i.e., the observation pair (Si , Ni (t)) is generated by
Ni (t) = f (Si , t) + Mi (t)
(7)
in the time window [0, T ] for some T > 0, and df (S, t) = a(S, t)dt. In other words, the timevarying coverage function controls the propensity of occurring events over time. Specifically, for a
fixed set Si , as time t increases, the cumulative number of events observed grows accordingly for
that f (Si , t) is a continuous monotonic function over time; for a given time t, as the set Si changes
to another set Sj , the amount of coverage over domain U may change and hence can result in a
different cumulative intensity. This abstract model can be mapped to real world applications. In
the information diffusion context, for a fixed set of sources Si , as time t increases, the number of
influenced nodes in the social network tends to increase; for a given time t, if we change the sources
to Sj , the number of influenced nodes may be different depending on how influential the sources
are. In the economics and game theory context, for a fixed bundle of offers Si , as time t increases, it
is more likely that the merchant will observe the customers? actions in response to the offers; even
at the same time t, different bundles of offers, Si and Sj , may have very different ability to drive the
customers? actions.
Compared to a regression model yi = g(Si ) + i with i.i.d. input data (Si , yi ), our model outputs
a special random function over time, that is, a counting process Ni (t) with the noise being a zero
mean martingale Mi (t). In contrast to functional regression models, our model exploits much more
interesting structures of the problem. For instance, the random function representation in the last
section can be used to parametrize the model. Such special structure of the counting process allows
us to estimate the parameter of our model using maximum likelihood approach efficiently, and the
martingale noise enables us to use exponential concentration inequality in analyzing our algorithm.
3
4
Parametrization
Based on the following two mild assumptions, we will show how to parametrize the intensity function as a weighted combination of random kernel functions, learn the parameters by maximum
likelihood estimation, and eventually derive a sample complexity.
(A1) a(S, t) is smooth and boundedR on [0, T ]: 0 < amin 6 a 6 amax < ?, and a
? := d2 a/dt2
is absolutely continuous with a
?(t)dt < ?.
(A2) There is a known distribution Q0 (? ) and a constant C with Q0 (? )/C 6 Q(? ) 6 CQ0 (? ).
Kernel Smoothing To facilitate our finite dimensional parameterization, we first convolve the
intensity function with K(t) = k(t/?)/? where ? is the bandwidth parameter and k is a kernel
?
2
function (such as the Gaussian RBF kernel k(t) = e?t /2 / 2?) with
Z
Z
Z
2
0 6 k(t) 6 ?max ,
k(t) dt = 1,
t k(t) dt = 0, and ?k := t2 k(t) dt < ?. (8)
The convolution results in a smoothed intensity aK (S, t) = K(t) ? (df (S, t)/dt) = d(K(t) ?
?(S, t))/dt. By the property of convolution and exchanging derivative with integral, we have that
aK (S, t) = d(Z ? E? ?Q(? ) [K(t) ? ?(?>
by definition of f (?)
S r(t|? )])/dt
>
= Z ? E? ?Q(? ) d(K(t) ? ?(?S r(t|? ))/dt exchange derivative and integral
= Z ? E? ?Q(? ) [K(t) ? ?(t ? t(S, r)]
by property of convolution and function ?(?)
= Z ? E? ?Q(? ) [K(t ? t(S, ? ))]
by definition of ?(?)
?(?>
S r(t|? ))
jumps from 0 to 1. If we choose small enough
where t(S, ? ) is the time when function
kernel bandwidth, aK only incurs a small bias from a. But the smoothed intensity still results in
infinite number of parameters, due to the unknown distribution Q(? ). To address this problem, we
design the following random approximation with finite number of parameters.
Random Function Approximation The key idea is to sample a collection of W random change
points ? from a known distribution Q0 (? ) which can be different from Q(? ). If Q0 (? ) is not very
far way from Q(? ), the random approximation will be close to aK , and thus close to a. More
specifically, we will denote the space of weighted combination of W random kernel function by
(
)
W
X
Z
i.i.d.
K
wi K(t ? t(S, ?i )) : w > 0, 6 kwk1 6 ZC , {?i } ? Q0 (? ). (9)
A = aw (S, t) =
C
i=1
? 2 /(?)2 ), then with probability > 1 ? ?, there exists an e
Lemma 2. If W = O(Z
a ? A such that
RT
ES Et (a(S, t) ? e
a(S, t))2 := ES?P(S) 0 (a(S, t) ? e
a(S, t))2 dt/T = O(2 + ? 4 ).
?
The lemma then suggests to set the kernel bandwidth ? = O( ) to get O(2 ) approximation error.
5
Learning Algorithm
We develop a learning algorithm, referred to as TC OVERAGE L EARNER, to estimate the parameters
of aK
w (S, t) by maximizing the joint likelihood of all observed events based on convex optimization
techniques as follows.
Maximum Likelihood Estimation Instead of directly estimating the time-varying coverage function, which is the cumulative intensity function of the counting process, we turn to estimate
the intensity function a(S, t) = ??(S, t)/?t. Given m i.i.d. counting processes, Dm :=
{(S1 , N1 (t)), . . . , (Sm , Nm (t))} up to observation time T , the log-likelihood of the dataset is [11]
(Z
)
Z T
m
T
X
m
(10)
`(D |a) =
{log a(Si , t)} dNi (t) ?
a(Si , t) dt .
i=1
0
0
Maximizing the log-likelihood with respect to the intensity function a(S, t) then gives us the estimation b
a(S, t). The W -term random kernel function approximation reduces a function optimization
problem to a finite dimensional optimization problem, while incurring only small bias in the estimated function.
4
Algorithm 1 TC OVERAGE L EARNER
INPUT : {(Si , Ni (t))} , i = 1, . . . , m;
Sample W random features ?1 , . . . , ?W from Q0 (? );
Compute {t(Si , ?w )} , {gi } , {k(tij )} , i ? {1, . . . , m} , w = 1, . . . , W, tij < T ;
Initialize w0 ? ? = {w > 0, kwk1 6 1};
Apply projected quasi-newton algorithm [12] to solve 11;
PW
OUTPUT : aK
w (S, t) =
i=1 wi K(t ? t(S, ?i ))
Convex Optimization. By plugging the parametrization aK
w (S, t) (9) into the log-likelihood (10),
we formulate the optimization problem as :
?
?
m ?
?
X
X
min
w> gi ?
log w> k(tij )
subject to w > 0, kwk1 6 1,
(11)
w
?
?
i=1
tij <T
where we define
Z
T
K (t ? t(Si , ?k )) dt and
gik =
kl (tij ) = K(tij ? t(Si , ?l )),
(12)
0
tij when the j-th event occurs in the i-th counting process. By treating the normalization constant
Z as a free variable which will be tuned by cross validation later, we simply require that kwk1 6 1.
By applying the Gaussian RBF kernel, we can derive a closed form of gik and the gradient O` as
?
?
m ?
X
X
T ? t(Si , ?k )
k(tij ) ?
t(Si , ?k )
1
?
? erfc
, O` =
gi ?
.
erfc ? ?
gik =
?
2
w> k(tij ) ?
2h
2h
i=1
tij <T
(13)
A pleasing feature of this formulation is that it is convex in the argument w, allowing us to apply
various convex optimization techniques to solve the problem efficiently. Specifically, we first draw
W random features ?1 , . . . , ?W from Q0 (? ). Then, we precompute the jumping time t(Si , ?w )
m
W
for every source set {Si }i=1 on each random feature {?w }w=1 . Because in general |Si | << n,
this computation costs O(mW ). Based on the achieved m-by-W jumping-time matrix, we preprom
cess the feature vectors {gi }i=1 and k(tij ), i ? {1, . . . , m} , tij < T , which costs O(mW ) and
O(mLW ) where L is the maximum number of events caused by a particular source set before time
T . Finally, we apply the projected quasi-newton algorithm [12] to find the weight w that minimizes
the negative log-likelihood of observing the given event data. Because the evaluation of the objective
function and the gradient, which costs O(mLW ), is much more expensive than the projection onto
the convex constraint set, and L << n, the worst case computation complexity is thus O(mnW ).
Algorithm 1 summarizes the above steps in the end.
Sample Strategy. One important constitution of our parametrization is to sample W random change
points ? from a known distribution Q0 (? ). Because given a set Si , we can only observe the jumping
time of the events in each counting process without knowing the identity of the covered items (which
is a key difference from [8]), the best thing we can do is to sample from these historical data.
Specifically, let the number of counting processes that a single item s ? V is involved to induce
be Ns , and the collection of all the jumping timestamps before time T be Js . Then, for the s-th
entry of ? , with probability |Js |/nNs , we uniformly draw a sample from Js ; and with probability
1 ? |Js |/nNs , we assign a time much greater than T to indicate that the item will never be covered
until infinity. Given the very limited information, although this Q0 (? ) might be quite different from
Q(? ), by drawing sufficiently large number of samples and adjusting the weights, we expect it still
can lead to good results, as illustrated in our experiments later.
6
Sample Complexity
Suppose we use W random features and m training examples to compute an ` -MLE solution b
a, i.e.,
`(Dm |b
a) > max
`(Dm |a0 ) ? ` .
0
a ?A
The goal is to analyze how well the function fb induced by b
a approximates the true function f . This
sections describes the intuition and the complete proof is provided in the appendix.
5
A natural choice for connecting the error between f and fb with the log-likelihood cost used in MLE
is the Hellinger distance [22]. So it suffices to prove an upper bound on the Hellinger distance
h(a, b
a) between b
a and the true intensity a, for which we need to show a high probability bound on
b 2 (a, a0 ) between the two. Here, h and H
b are defined as
the (total) empirical Hellinger distance H
h
i
p
p
2
1
h2 (a, a0 ) := ES Et
a(S, t) ? a0 (S, t) ,
2
m Z T hp
i2
X
p
b 2 (a, a0 ) := 1
a(Si , t) ? a0 (Si , t) dt.
H
2 i=1 0
The key for the analysis is to show that the empirical Hellinger distance can be bounded by a martingale plus some other additive error terms, which we then bound respectively. This martingale is
defined based on our hypotheses and the martingales Mi associated with the counting process Ni :
!
Z t
m Z t
X
X
M (t|g) :=
g(t)d
Mi (t) =
g(t)dMi (t)
0
n
where g ? G = ga0 =
1
2
log
a+a
2a
0
i
i=1
0
o
: a0 ? A . More precisely, we have the following lemma.
Lemma 3. Suppose b
a is an ` -MLE. Then
m 0
b 2 (b
H
a, a) 6 16M (T ; gba ) + 4 `(Dm |a) ? max
`(D
|a
)
+ 4` .
0
a ?A
The right hand side has three terms: the martingale (estimation error), the likelihood gap between
the truth and the best one in our hypothesis class (approximation error), and the optimization error.
We then focus on bounding the martingale and the likelihood gap.
To bound the martingale, we first introduce a notion called (d, d0 )-covering dimension measuring
the complexity of the hypothesis class, generalizing that in [25]. Based on this notion, we prove
a uniform convergence inequality, combining the ideas in classic works on MLE [25] and counting process [13]. Compared to the classic uniform inequality, our result is more general, and the
complexity notion has more clear geometric interpretation and are thus easier to verify. For the likelihood gap, recall that by Lemma 2, there exists an good approximation a
? ? A. The likelihood gap
is then bounded by that between a and a
?, which is small since a and a
? are close.
Combining the two leads to a bound on the Hellinger distance based on bounded dimension of the
hypothesis class. We then show that the dimension of our specific hypothesis class is at most the
b 2 (b
number of random features W , and convert H
a, a) to the desired `2 error bound on f and fb.
5/4
ZT 5/2
ZT
2
? ZT [W + ` ] . Then
?
Theorem 4. Suppose W = O Z
+ amin
and m = O
W
with probability > 1 ? ? over the random sample of {?i }i=1 , we have that for any 0 6 t 6 T ,
h
i2
ES fb(S, t) ? f (S, t) 6 .
The theorem shows that the number of random functions needed to achieve error is roughly
O(?5/2 ), and the sample size is O(?7/2 ). They also depend on amin , which means with more
random functions and data, we can deal with intensities with more extreme values. Finally, they
increase with the time T , i.e., it is more difficult to learn the function values at later time points.
7
Experiments
We evaluate TC OVERAGE L EARNER on both synthetic and real world information diffusion data.
We show that our method can be more robust to model misspecification than other state-of-the-art
alternatives by learning a temporal coverage function all at once.
7.1 Competitors
Because our input data only include pairs of a source set and the temporal information of its trigm
gered events {(Si , Ni (t))}i=1 with unknown identity, we first choose the general kernel ridge regression model as the major baseline, which directly estimates the influence value of a source set
6
15
TCoverageLearner
Kernel Ridge Regression
CIC
DIC
5
5
TCoverageLearner
Kernel Ridge Regression
CIC
DIC
4
10
MAE
10
MAE
20
MAE
10
30
TCoverageLearner
Kernel Ridge Regression
CIC
DIC
MAE
15
5
TCoverageLearner
Kernel Ridge Regression
CIC
DIC
3
2
1
0
1
2
3
4
5 6
Time
7
8
9
10
0
1
2
3
4
5 6
Time
7
8
9
10
0
1
2
3
4
5 6
Time
7
8
9
10
0
1
2
3
4
5 6
Time
7
8
9
10
(a) Weibull (CIC)
(b) Exponential (CIC)
(c) DIC
(d) LT
Figure 1: MAE of the estimated influence on test data along time with the true diffusion model being
continuous-time independent cascade with pairwise Weibull (a) and Exponential (b) transmission
functions, (c) discrete-time independent cascade model and (d) linear-threshold cascade model.
?S by f (?S ) = k(?S )(K + ?I)?1 y where k(?S ) = K(?Si , ?S ), and K is the kernel matrix. We discretize the time into several steps and fit a separate model to each of them. Between
two consecutive time steps, the predictions are simply interpolated. In addition, to further demonstrate the robustness of TC OVERAGE L EARNER, we compare it to the two-stage methods which
must know the identity of the nodes involved in an information diffusion process to first learn
a specific diffusion model based on which they can then estimate the influence. We give them
such an advantage and study three well-known diffusion models : (I) Continuous-time Independent
Cascade model(CIC)[14, 15]; (II) Discrete-time Independent Cascade model(DIC)[1]; and (III)
Linear-Threshold cascade model(LT)[1].
7.2
Influence Estimation on Synthetic Data
We generate Kronecker synthetic networks ([0.9 0.5;0.5 0.3]) which mimic real world information
diffusion patterns [16]. For CIC, we use both Weibull distribution (Wbl) and Exponential distribution (Exp) for the pairwise transmission function associated with each edge, and randomly set their
parameters to capture the heterogeneous temporal dynamics. Then, we use NETRATE [14] to learn
the model by assuming an exponential pairwise transmission function. For DIC, we choose the pairwise infection probability uniformly from 0 to 1 and fit the model by [17]. For LT, we assign the edge
weight wuv between u and v as 1/dv , where dv is the degree of node v following [1]. Finally, 1,024
source sets are sampled with power-law distributed cardinality (with exponent 2.5), each of which
induces eight independent cascades(or counting processes), and the test data contains another 128
independently sampled source sets with the ground truth influence estimated from 10,000 simulated
cascades up to time T = 10. Figure 1 shows the MAE(Mean Absolute Error) between the estimated
influence value and the true value up to the observation window T = 10. The average influence
is 16.02, 36.93, 9.7 and 8.3. We use 8,192 random features and two-fold cross validation on the
train data to tune the normalization Z, which has the best
? value 1130, 1160, 1020, and 1090, respectively. We choose the RBF kernel bandwidth h = 1/ 2? so that the magnitude of the smoothed
approximate function still equals to 1 (or it can be tuned by cross-validation as well), which matches
the original indicator function. For the kernel ridge regression, the RBF kernel bandwidth and the
regularization ? are all chosen by the same two-fold cross validation. For CIC and DIC, we learn
the respective model up to time T for once.
Figure 1 verifies that even though the underlying diffusion models can be dramatically different,
the prediction performance of TC OVERAGE L EARNER is robust to the model changes and consistently outperforms the nontrivial baseline significantly. In addition, even if CIC and DIC are
provided with extra information, in Figure 1(a), because the ground-truth is continuous-time diffusion model with Weibull functions, they do not have good performance. CIC assumes the right
model but the wrong family of transmission functions. In Figure 1(b), we expect CIC should have
the best performance for that it assumes the correct diffusion model and transmission functions.
Yet, TC OVERAGE L EARNER still has comparable performance with even less information. In Figure 1(c), although DIC has assumed the correct model, it is hard to determine the correct step size to
discretize the time line, and since we only learn the model once up to time T (instead of at each time
point), it is harder to fit the whole process. In Figure1(d), both CIC and DIC have the wrong model,
so we have similar trend as Figure synthetic(a). Moreover, for kernel ridge regression, we have to
first partition the timeline with arbitrary step size, fit the model to each of time, and interpolate the
value between neighboring time legs. Not only will the errors from each stage be accumulated to
the error of the final prediction, but also we cannot rely on this method to predict the influence of a
source set beyond the observation window T .
7
5
80
2
6
4
10
influence
10
100
10
8
time(s)
15
0
3
10
TCoverageLearner
Kernel Ridge Regression
CIC
DIC
20
Average MAE
Average MAE
25
1
10
2
3
4
5
6
Groups of Memes
7
0
0
128 256 512 1024 2048 4096 8192
# Random features
10
60
40
2
1
TCoverageLearner
Kernel Ridge Regression
CIC
DIC
128 256 512 1024 2048 4096 8192
# random features
20
1
2
3
4
5 6
Time
7
8
9
10
(a) Average MAE
(b) Features? Effect
(c) Runtime
(d) Influence maximization
Figure 2: (a) Average MAE from time 1 to 10 on seven groups of real cascade data; (b) Improved
estimation with increasing number of random features; (c) Runtime in log-log scale; (d) Maximized
influence of selected sources on the held-out testing data along time.
Overall, compared to the kernel ridge regression, TC OVERAGE L EARNER only needs to be trained
once given all the event data up to time T in a compact and principle way, and then can be used to infer the influence of any given source set at any particular time much more efficiently and accurately.
In contrast to the two-stage methods, TC OVERAGE L EARNER is able to address the more general
setting with much less assumption and information but still can produce consistently competitive
performance.
7.3
Influence Estimation on Real Data
MemeTracker is a real-world dataset [18] to study information diffusion. The temporal flow of information was traced using quotes which are short textual phrases spreading through the websites.
We have selected seven typical groups of cascades with the representative keywords like ?apple and
jobs?, ?tsunami earthquake?, etc., among the top active 1,000 sites. Each set of cascades is split into
60%-train and 40%-test. Because we often can observe cascades only from single seed node, we
rarely have cascades produced from multiple sources simultaneously. However, because our model
can capture the correlation among multiple sources, we challenge TC OVERAGE L EARNER with sets
of randomly chosen multiple source nodes on the independent hold-out data. Although the generation of sets of multiple source nodes is simulated, the respective influence is calculated from the real
test data as follows : Given a source set S, for each node u ? S, let C(u) denote the set of cascades
generated from u on the testing data. We uniformly sample cascades from C(u). The average length
of all sampled cascades is treated as the true influence of S. We draw 128 source sets and report
the average MAE along time in Figure 2(a). Again, we can observe that TC OVERAGE L EARNER
has consistent and robust estimation performance across all testing groups. Figure 2(b) verifies that
the prediction can be improved as more random features are exploited, because the representational
power of TC OVERAGE L EARNER increases to better approximate the unknown true coverage function. Figure 2(c) indicates that the runtime of TC OVERAGE L EARNER is able to scale linearly with
large number of random features. Finally, Figure 2(d) shows the application of the learned coverage
function to the influence maximization problem along time, which seeks to find a set of source nodes
that maximize the expected number of infected nodes by time T . The classic greedy algorithm[19]
is applied to solve the problem, and the influence is calculated and averaged over the seven held-out
test data. It shows that TC OVERAGE L EARNER is very competitive to the two-stage methods with
much less assumption. Because the greedy algorithm mainly depends on the relative rank of the
selected sources, although the estimated influence value can be different, the selected set of sources
could be similar, so the performance gap is not large.
8
Conclusions
We propose a new problem of learning temporal coverage functions with a novel parametrization
connected with counting processes and develop an efficient algorithm which is guaranteed to learn
such a combinatorial function from only polynomial number of training samples. Empirical study
also verifies our method outperforms existing methods consistently and significantly.
Acknowledgments This work was supported in part by NSF grants CCF-0953192, CCF-1451177,
CCF-1101283, and CCF-1422910, ONR grant N00014-09-1-0751, AFOSR grant FA9550-09-10538, Raytheon Faculty Fellowship, NSF IIS1116886, NSF/NIH BIGDATA 1R01GM108341, NSF
CAREER IIS1350983 and Facebook Graduate Fellowship 2014-2015.
8
References
? Tardos. Maximizing the spread of influence through a social
[1] David Kempe, Jon Kleinberg, and Eva
network. In SIGKDD 2003, pages 137?146. ACM, 2003.
[2] C. Guestrin, A. Krause, and A. Singh. Near-optimal sensor placements in gaussian processes. In International Conference on Machine Learning ICML?05, 2005.
[3] Benny Lehmann, Daniel Lehmann, and Noam Nisan. Combinatorial auctions with decreasing marginal
utilities. In EC ?01, pages 18?28, 2001.
[4] Maria-Florina Balcan and Nicholas JA Harvey. Learning submodular functions. In Proceedings of the
43rd annual ACM symposium on Theory of computing, pages 793?802. ACM, 2011.
[5] A. Badanidiyuru, S. Dobzinski, H. Fu, R. D. Kleinberg, N. Nisan, and T. Roughgarden. Sketching valuation functions. In Annual ACM-SIAM Symposium on Discrete Algorithms, 2012.
[6] Vitaly Feldman and Pravesh Kothari. Learning coverage functions. arXiv preprint arXiv:1304.2079,
2013.
[7] Vitaly Feldman and Jan Vondrak. Optimal bounds on approximation of submodular and xos functions by
juntas. In FOCS, 2013.
[8] Nan Du, Yingyu Liang, Nina Balcan, and Le Song. Influence function learning in information diffusion
networks. In ICML 2014, 2014.
[9] L. Song, M. Kolar, and E. P. Xing. Time-varying dynamic bayesian networks. In Neural Information
Processing Systems, pages 1732?1740, 2009.
[10] M. Kolar, L. Song, A. Ahmed, and E. P. Xing. Estimating time-varying networks. Ann. Appl. Statist.,
4(1):94?123, 2010.
[11] Odd Aalen, Oernulf Borgan, and H?akon K Gjessing. Survival and event history analysis: a process point
of view. Springer, 2008.
[12] M. P. Friedlander K. Murphy M. Schmidt, E. van den Berg. Optimizing costly functions with simple
constraints: A limited-memory projected quasi-newton algorithm. In AISTATS 2009.
[13] Sara van de Geer. Exponential inequalities for martingales, with application to maximum likelihood
estimation for counting processes. The Annals of Statistics, pages 1779?1801, 1995.
[14] Manuel Gomez Rodriguez, David Balduzzi, and Bernhard Sch?olkopf. Uncovering the temporal dynamics
of diffusion networks. arXiv preprint arXiv:1105.0697, 2011.
[15] Nan Du, Le Song, Hongyuhan Zha, and Manuel Gomez Rodriguez. Scalable influence estimation in
continuous time diffusion networks. In NIPS 2013, 2013.
[16] Jure Leskovec, Deepayan Chakrabarti, Jon Kleinberg, Christos Faloutsos, and Zoubin Ghahramani. Kronecker graphs: An approach to modeling networks. 11(Feb):985?1042, 2010.
[17] Praneeth Netrapalli and Sujay Sanghavi. Learning the graph of epidemic cascades.
RICS/PERFORMANCE, pages 211?222. ACM, 2012.
In SIGMET-
[18] Jure Leskovec, Lars Backstrom, and Jon Kleinberg. Meme-tracking and the dynamics of the news cycle.
In SIGKDD2009, pages 497?506. ACM, 2009.
[19] G. Nemhauser, L. Wolsey, and M. Fisher. An analysis of the approximations for maximizing submodular
set functions. Mathematical Programming, 14:265?294, 1978.
[20] L. Wasserman. All of Nonparametric Statistics. Springer, 2006.
[21] Ali Rahimi and Benjamin Recht. Weighted sums of random kitchen sinks: Replacing minimization with
randomization in learning. In Neural Information Processing Systems, 2009.
[22] Sara van de Geer. Hellinger-consistency of certain nonparametric maximum likelihood estimators. The
Annals of Statistics, pages 14?44, 1993.
[23] G.R. Shorack and J.A. Wellner. Empirical Processes with Applications to Statistics. Wiley, New York,
1986.
[24] Wing Hung Wong and Xiaotong Shen. Probability inequalities for likelihood ratios and convergence rates
of sieve mles. The Annals of Statistics, pages 339?362, 1995.
[25] L. Birg?e and P. Massart. Minimum Contrast Estimators on Sieves: Exponential Bounds and Rates of
Convergence. Bernoulli, 4(3), 1998.
[26] Kenneth S Alexander. Rates of growth and sample moduli for weighted empirical processes indexed by
sets. Probability Theory and Related Fields, 75(3):379?423, 1987.
9
| 5366 |@word mild:1 faculty:1 pw:1 polynomial:3 d2:1 seek:1 decomposition:1 incurs:1 harder:1 memetracker:1 contains:1 daniel:1 tuned:2 past:2 existing:2 reaction:1 outperforms:2 manuel:2 si:29 yet:1 follower:1 must:2 attracted:1 timestamps:2 additive:1 partition:1 enables:1 treating:2 greedy:2 selected:4 website:1 item:7 parameterization:1 accordingly:1 parametrization:5 short:1 record:4 fa9550:1 junta:1 node:11 mathematical:1 along:6 dn:1 symposium:2 chakrabarti:1 focs:1 prove:2 yingyu:2 hellinger:6 introduce:2 pairwise:4 theoretically:1 expected:1 roughly:1 decreasing:1 little:1 actual:1 window:3 cardinality:1 increasing:2 iis1350983:1 provided:4 estimating:3 underlying:2 notation:1 bounded:3 moreover:1 minimizes:1 weibull:4 temporal:9 every:1 multidimensional:2 growth:1 runtime:3 wrong:2 control:1 grant:3 positive:1 before:3 timing:2 wuv:1 tends:1 ak:7 analyzing:1 might:1 plus:1 suggests:1 sara:2 appl:1 limited:2 graduate:1 averaged:1 practical:1 unique:1 earthquake:1 testing:3 acknowledgment:1 practice:2 jan:1 yingyul:1 empirical:5 significantly:4 thought:1 cascade:18 ups:3 word:2 projection:1 induce:1 zoubin:1 get:1 onto:1 close:3 cannot:1 context:2 influence:26 applying:1 wong:1 customer:10 maximizing:4 economics:4 independently:1 convex:5 formulate:1 shen:1 wasserman:1 estimator:2 amax:1 classic:3 notion:3 coordinate:1 increment:1 tardos:1 annals:3 play:1 trigger:3 user:3 suppose:3 programming:1 hypothesis:5 associate:1 trend:1 expensive:1 particularly:1 observed:2 role:1 preprint:2 capture:3 worst:1 eva:1 news:2 connected:1 cycle:1 eu:2 gjessing:1 benny:1 borgan:1 intuition:1 predictable:1 meme:2 complexity:6 benjamin:1 dt2:1 dynamic:7 motivate:2 raise:1 depend:1 trained:1 algebra:1 singh:1 badanidiyuru:1 ali:1 basis:1 sink:1 joint:1 represented:2 various:1 train:2 quite:1 widely:1 valued:1 solve:3 drawing:1 otherwise:2 epidemic:1 ability:1 statistic:5 gi:4 nondecreasing:1 final:1 online:1 sequence:2 triggered:1 advantage:1 propose:4 product:1 neighboring:1 combining:2 achieve:1 representational:1 amin:3 validate:1 olkopf:1 convergence:3 requirement:1 transmission:5 produce:1 depending:1 develop:4 derive:2 lsong:1 keywords:1 odd:1 school:1 job:1 netrapalli:1 coverage:30 c:2 indicate:1 correct:3 stochastic:2 xos:1 lars:1 exchange:1 require:1 ja:1 assign:2 fix:2 generalization:1 suffices:1 randomization:1 cic:15 hold:1 sufficiently:1 ground:2 exp:1 seed:1 algorithmic:2 predict:1 claim:1 dobzinski:1 major:1 consecutive:1 a2:1 estimation:12 pravesh:1 combinatorial:7 spreading:3 quote:1 propensity:1 akon:1 weighted:5 minimization:1 sensor:1 gaussian:3 fulfill:1 rather:1 varying:18 gatech:2 timevarying:2 endow:1 focus:1 bundled:2 maria:2 consistently:3 rank:1 likelihood:18 indicates:1 mainly:1 iis1116886:1 contrast:3 sigkdd:1 bernoulli:1 baseline:2 accumulated:1 entire:1 submartingale:1 a0:7 diminishing:1 doob:1 quasi:3 interested:1 provably:1 overall:1 among:2 uncovering:1 exponent:1 platform:1 special:3 kempe:1 smoothing:1 marginal:1 field:1 initialize:1 art:1 dmi:1 aware:1 having:1 never:1 sampling:1 equal:1 once:4 icml:2 jon:3 mimic:1 t2:1 report:1 sanghavi:1 few:1 randomly:2 simultaneously:1 interpolate:1 murphy:1 kitchen:1 n1:1 pleasing:1 evaluation:1 extreme:1 r01gm108341:1 held:2 bundle:4 integral:4 edge:2 fu:1 jumping:4 respective:2 indexed:1 desired:1 theoretical:1 leskovec:2 instance:3 modeling:3 earlier:1 cover:1 infected:1 measuring:1 exchanging:1 maximization:4 phrase:1 cost:4 subset:3 entry:1 uniform:2 ga0:1 aw:1 synthetic:5 nns:2 mles:1 recht:1 international:1 siam:1 connecting:1 sketching:1 again:1 nm:1 choose:4 earner:15 derivative:3 wing:1 return:1 de:2 caused:2 depends:1 nisan:2 later:3 view:1 closed:1 observing:1 analyze:1 competitive:2 xing:2 zha:1 ni:12 efficiently:3 maximized:1 bayesian:1 accurately:2 produced:1 cc:1 drive:1 apple:1 history:3 explain:1 influenced:2 whenever:1 infection:1 facebook:1 definition:5 competitor:1 ninamf:1 involved:2 dm:6 naturally:1 associated:3 proof:2 mi:4 sampled:3 dataset:2 adjusting:1 recall:1 knowledge:1 dt:15 follow:3 figure1:1 response:1 inspire:1 improved:2 formulation:4 though:1 furthermore:1 just:3 stage:4 until:2 correlation:1 hand:1 replacing:1 rodriguez:2 grows:1 modulus:1 facilitate:1 effect:1 verify:1 true:6 ccf:4 hence:2 regularization:1 sieve:2 q0:9 i2:2 illustrated:1 deal:1 ex1:1 game:5 covering:1 complete:1 ridge:10 demonstrate:1 performs:1 balcan:3 auction:1 novel:4 nih:1 common:1 functional:1 overview:1 interpretation:1 approximates:1 mae:11 mellon:1 feldman:2 rd:1 sujay:1 consistency:1 hp:1 submodular:4 etc:2 feb:1 j:4 recent:2 perspective:1 optimizing:1 certain:3 constitution:1 harvey:1 n00014:1 retailer:1 inequality:5 onr:1 kwk1:4 yi:2 exploited:1 guestrin:1 minimum:1 additional:2 greater:1 determine:1 maximize:1 monotonically:1 ii:1 multiple:4 reduces:1 d0:1 infer:1 smooth:3 rahimi:1 match:1 ahmed:1 offer:4 cross:4 host:1 mle:4 dic:13 a1:1 plugging:1 prediction:4 scalable:1 regression:12 florina:2 overage:15 essentially:1 cmu:1 expectation:1 df:2 heterogeneous:1 arxiv:4 kernel:23 normalization:3 achieved:1 addition:2 fellowship:2 krause:1 addressed:1 interval:1 grow:1 source:24 sch:1 extra:1 massart:1 subject:1 induced:1 wbl:1 thing:1 vitaly:2 flow:1 integer:2 mw:2 counting:25 near:1 iii:1 enough:1 split:1 fit:4 bandwidth:5 idea:3 praneeth:1 knowing:1 dunan:1 t0:5 utility:1 wellner:1 effort:1 song:5 york:1 action:5 ignored:1 generally:1 tij:13 covered:5 clear:1 tune:1 dramatically:1 amount:1 nonparametric:2 statist:1 induces:1 generate:1 nsf:4 estimated:5 arising:1 carnegie:1 discrete:6 write:1 group:4 key:4 threshold:2 traced:1 ce:1 diffusion:20 ht:3 kenneth:1 ht0:2 graph:2 sum:1 convert:1 package:1 angle:1 parameterized:1 respond:1 lehmann:2 family:1 draw:3 summarizes:1 appendix:1 comparable:1 bound:8 nan:3 guaranteed:1 gomez:2 fold:2 oracle:1 nonnegative:2 nontrivial:1 annual:2 occur:1 placement:1 constraint:2 infinity:1 precisely:1 kronecker:2 roughgarden:1 interpolated:1 aspect:1 kleinberg:4 argument:1 min:2 vondrak:1 xiaotong:1 department:1 influential:2 according:1 combination:3 precompute:1 across:2 describes:1 wi:2 backstrom:1 s1:1 leg:1 dv:2 den:1 principally:1 tsunami:1 taken:2 turn:1 count:1 eventually:1 needed:1 know:1 end:1 parametrize:3 rewritten:1 incurring:1 apply:3 observe:4 eight:1 birg:1 nicholas:1 occurrence:1 alternative:2 robustness:1 schmidt:1 faloutsos:1 original:1 uncountable:1 denotes:1 convolve:1 include:1 assumes:2 graphical:1 top:1 newton:3 exploit:1 balduzzi:1 build:1 prof:1 erfc:2 ghahramani:1 objective:1 question:1 occurs:1 strategy:1 concentration:1 dependence:1 compensator:1 rt:1 costly:1 gradient:2 dp:3 nemhauser:1 distance:5 separate:1 mapped:1 simulated:2 w0:1 seven:3 valuation:3 nina:1 assuming:1 ru:9 length:1 index:2 modeled:2 deepayan:1 ratio:1 kolar:2 liang:2 difficult:1 potentially:1 trace:1 negative:1 noam:1 design:1 zt:3 unknown:3 allowing:1 upper:1 discretize:2 observation:6 convolution:3 kothari:1 sm:1 finite:4 merchant:1 misspecification:1 smoothed:3 arbitrary:1 intensity:15 david:2 pair:4 kl:1 connection:2 learned:3 textual:1 timeline:1 nip:1 address:2 able:3 beyond:1 jure:2 usually:2 pattern:1 challenge:1 max:3 memory:1 mouth:1 power:2 event:15 natural:1 rely:1 treated:1 indicator:2 gba:1 technology:1 brief:1 prior:2 literature:1 geometric:1 friedlander:1 relative:1 law:2 afosr:1 expect:2 interesting:3 generation:1 wolsey:1 validation:4 h2:1 degree:1 offered:2 sufficient:1 consistent:1 dq:1 principle:1 supported:1 last:1 free:1 zc:1 side:2 formal:1 bias:2 institute:1 absolute:1 dni:1 distributed:1 van:3 dimension:5 calculated:2 world:6 cumulative:6 fb:4 collection:4 jump:3 projected:3 historical:2 far:2 ec:1 social:10 sj:3 approximate:2 emphasize:1 compact:1 bernhard:1 active:1 assumed:2 continuous:9 learn:9 robust:3 adopter:1 career:1 du:3 domain:3 aistats:1 spread:2 linearly:1 bounding:1 noise:2 netrate:1 whole:1 verifies:3 site:1 referred:1 representative:1 georgia:1 martingale:10 wiley:1 n:1 christos:1 meyer:1 exponential:7 theorem:3 specific:2 gik:3 survival:1 exists:3 magnitude:1 occurring:1 demand:1 gap:5 easier:1 tc:15 generalizing:1 lt:3 simply:2 likely:2 happening:1 tracking:1 collectively:1 monotonic:2 springer:2 truth:3 acm:6 identity:3 goal:1 ann:1 rbf:4 fisher:1 change:8 hard:1 specifically:4 infinite:1 uniformly:3 typical:1 raytheon:1 lemma:7 called:3 total:1 geer:2 e:4 rarely:1 formally:1 college:1 aalen:1 berg:1 alexander:1 absolutely:1 bigdata:1 evaluate:1 princeton:2 hung:1 |
4,823 | 5,367 | Online and Stochastic Gradient Methods for
Non-decomposable Loss Functions
Purushottam Kar?
Harikrishna Narasimhan?
Prateek Jain?
Microsoft Research, INDIA
?
Indian Institute of Science, Bangalore, INDIA
{t-purkar,prajain}@microsoft.com, [email protected]
?
Abstract
Modern applications in sensitive domains such as biometrics and medicine frequently require the use of non-decomposable loss functions such as precision@k,
F-measure etc. Compared to point loss functions such as hinge-loss, these offer much more fine grained control over prediction, but at the same time present
novel challenges in terms of algorithm design and analysis. In this work we initiate
a study of online learning techniques for such non-decomposable loss functions
with an aim to enable incremental learning as well as design scalable solvers for
batch problems. To this end, we propose an online learning framework for such
loss functions. Our model enjoys several nice properties, chief amongst them being the existence of efficient online learning algorithms with sublinear regret and
online to batch conversion bounds. Our model is a provable extension of existing
online learning models for point loss functions. We instantiate two popular losses,
Prec@k and pAUC, in our model and prove sublinear regret bounds for both of
them. Our proofs require a novel structural lemma over ranked lists which may
be of independent interest. We then develop scalable stochastic gradient descent
solvers for non-decomposable loss functions. We show that for a large family
of loss functions satisfying a certain uniform convergence property (that includes
Prec@k , pAUC, and F-measure), our methods provably converge to the empirical
risk minimizer. Such uniform convergence results were not known for these losses
and we establish these using novel proof techniques. We then use extensive experimentation on real life and benchmark datasets to establish that our method can be
orders of magnitude faster than a recently proposed cutting plane method.
1
Introduction
Modern learning applications frequently require a level of fine-grained control over prediction performance that is not offered by traditional ?per-point? performance measures such as hinge loss.
Examples include datasets with mild to severe label imbalance such as spam classification wherein
positive instances (spam emails) constitute a tiny fraction of the available data, and learning tasks
such as those in medical diagnosis which make it imperative for learning algorithms to be sensitive to class imbalances. Other popular examples include ranking tasks where precision in the top
ranked results is valued more than overall precision/recall characteristics. The performance measures of choice in these situations are those that evaluate algorithms over the entire dataset in a
holistic manner. Consequently, these measures are frequently non-decomposable over data points.
More specifically, for these measures, the loss on a set of points cannot be expressed as the sum of
losses on individual data points (unlike hinge loss, for example). Popular examples of such measures
include F-measure, Precision@k, (partial) area under the ROC curve etc.
Despite their success in these domains, non-decomposable loss functions are not nearly as well
understood as their decomposable counterparts. The study of point loss functions has led to a deep
1
understanding about their behavior in batch and online settings and tight characterizations of their
generalization abilities. The same cannot be said for most non-decomposable losses. For instance, in
the popular online learning model, it is difficult to even instantiate a non-decomposable loss function
as defining the per-step penalty itself becomes a challenge.
1.1
Our Contributions
Our first main contribution is a framework for online learning with non-decomposable loss functions.
The main hurdle in this task is a proper definition of instantaneous penalties for non-decomposable
losses. Instead of resorting to canonical definitions, we set up our framework in a principled way
that fulfills the objectives of an online model. Our framework has a very desirable characteristic that
allows it to recover existing online learning models when instantiated with point loss functions. Our
framework also admits online-to-batch conversion bounds.
We then propose an efficient Follow-the-Regularized-Leader [1] algorithm within our framework.
We show that for loss functions
that satisfy a generic ?stability? condition, our algorithm is able
to offer vanishing O ?1T regret. Next, we instantiate within our framework, convex surrogates
for two popular performances measures namely, Precision at k (Prec@k ) and partial area under the
ROC curve (pAUC) [2] and show, via a stability analysis, that we do indeed achieve sublinear
regret bounds for these loss functions. Our stability proofs involve a structural lemma on sorted
lists of inner products which proves Lipschitz continuity properties for measures on such lists (see
Lemma 2) and might be useful for analyzing non-decomposable loss functions in general.
A key property of online learning methods is their applicability in designing solvers for offline/batch problems. With this goal in mind, we design a stochastic gradient-based solver for
non-decomposable loss functions. Our methods apply to a wide family of loss functions (including
Prec@k , pAUC and F-measure) that were introduced in [3] and have been widely adopted [4, 5, 6]
in the literature. We design several variants of our method and show that our methods provably converge to the empirical risk minimizer of the learning instance at hand. Our proofs involve uniform
convergence-style results which were not known for the loss functions we study and require novel
techniques, in particular the structural lemma mentioned above.
Finally, we conduct extensive experiments on real life and benchmark datasets with pAUC and
Prec@k as performance measures. We compare our methods to state-of-the-art methods that are
based on cutting plane techniques [7]. The results establish that our methods can be significantly
faster, all the while offering comparable or higher accuracy values. For example, on a KDD 2008
challenge dataset, our method was able to achieve a pAUC value of 64.8% within 30ms whereas it
took the cutting plane method more than 1.2 seconds to achieve a comparable performance.
1.2
Related Work
Non-decomposable loss functions such as Prec@k , (partial) AUC, F-measure etc, owing to their
demonstrated ability to give better performance in situations with label imbalance etc, have generated significant interest within the learning community. From their role in early works as indicators
of performance on imbalanced datasets [8], their importance has risen to a point where they have
become the learning objectives themselves. Due to their complexity, methods that try to indirectly
optimize these measures are very common e.g. [9], [10] and [11] who study the F-measure. However, such methods frequently seek to learn a complex probabilistic model, a task arguably harder
than the one at hand itself. On the other hand are algorithms that perform optimization directly via
structured losses. Starting from the seminal work of [3], this method has received a lot of interest
for measures such as the F-measure [3], average precision [4], pAUC [7] and various ranking losses
[5, 6]. These formulations typically use cutting plane methods to design dual solvers.
We note that the learning and game theory communities are also interested in non-additive notions
of regret and utility. In particular [12] provides a generic framework for online learning with nonadditive notions of regret with a focus on showing regret bounds for mixed strategies in a variety of
problems. However, even polynomial time implementation of their strategies is difficult in general.
Our focus, on the other hand, is on developing efficient online algorithms that can be used to solve
large scale batch problems. Moreover, it is not clear how (if at all) can the loss functions considered
here (such as Prec@k ) be instantiated in their framework.
2
Recently, online learning for AUC maximization has received some attention [13, 14]. Although
AUC is not a point loss function, it still decomposes over pairs of points in a dataset, a fact that [13]
and [14] crucially use. The loss functions in this paper do not exhibit any such decomposability.
2
Problem Formulation
Let x1:t := {x1 , . . . , xt }, xi ? Rd and y1:t := {y1 , . . . , yt }, yi ? {?1, 1} be the observed data
points and true binary labels. We will use yb1:t := {b
y1 , . . . , ybt }, ybi ? R to denote the predictions of a
learning algorithm. We shall, for sake of simplicity, restrict ourselves to linear predictors ybi = w> xi
for parameter vectors w ? Rd . A performance measure P : {?1, 1}t ? Rt ? R+ shall be used to
evaluate the the predictions of the learning algorithm against the true labels. Our focus shall be on
non-decomposable performance measures such as Prec@k , partial AUC etc.
Since these measures are typically non-convex, convex surrogate loss functions are used instead (we
will use the terms loss function and performance measure interchangeably). A popular technique for
constructing such loss functions is the structural SVM formulation [3] given below. For simplicity,
we shall drop mention of the training points and use the notation `P (w) := `P (x1:T , y1:T , w).
`P (w) =
max
? ?{?1,+1}T
y
T
X
(?
yi ? yi )x>
y, y).
i w ? P(?
(1)
i=1
Precision@k. The Prec@k measure ranks the data points in order of the predicted scores ybi and then
returns the number of true positives in the top ranked positions. This is valuable in situations where
there are very few positives. To formalize this, for any predictor w and set of points x1:t , define
S(x, w) := {j : w> x > w> xj } to be the set of points which w ranks above x. Then define
1, if |S(x, w)| < d?te,
T?,t (x, w) =
(2)
0, otherwise.
i.e. T?,t (x, w) is non-zero iff x is in the top-? fraction of the set. Then we define1
X
Prec@k (w) :=
I [yj = 1] .
j:Tk,t (xj ,w)=1
The structural surrogate for this measure is then calculated as 2
t
t
X
X
`Prec@k (w) =
max t
(?
yi ? yi )xTi w ?
yi y?i .
? ?{?1,+1}
y
P
i=1
yi +1)=2kt
i (?
(3)
i=1
Partial AUC. This measures the area under the ROC curve with the false positive rate restricted to
the range [0, ?]. This is in contrast to AUC that considers the entire range [0, 1] of false positive
rates. pAUC is useful in medical applications such as cancer detection where a small false positive
rate is desirable. Let us extend notation to use the indicator T?
top ? fraction of
?,t (x, w) to select the
j : yj < 0, w> x > w> xj ? d?t? e where
the negatively labeled points i.e. T?
(x,
w)
=
1
iff
?,t
t? is the number of negatives. Then we define
X X
>
>
pAUC(w) =
T?
(4)
?,t (xj , w) ? I[xi w ? xj w].
i:yi >0 j:yj <0
Let ? : R ? R+ be any convex, monotone, Lipschitz, classification surrogate. Then we can obtain
convex surrogates for pAUC(w) by replacing the indicator functions above with ?(?).
X X
>
>
`pAUC (w) =
T?
(5)
?,t (xj , w) ? ?(xi w ? xj w),
i:yi >0 j:yj <0
It can be shown [7, Theorem 4] that the structural surrogate for pAUC is equivalent to (5) with
?(c) = max(0, 1 ? c), the hinge loss function. In the next section we will develop an online
learning framework for non-decomposable performance measures and instantiate loss functions such
as `Prec@k and `pAUC in our framework. Then in Section 4, we will develop stochastic gradient
methods for non-decomposable loss functions and prove error bounds for the same. There we will
focus on a much larger family of loss functions including Prec@k , pAUC and F-measure.
1
2
An equivalent definition considers k to be the number of top ranked points instead.
[3] uses a slightly modified, but equivalent, definition that considers labels to be Boolean.
3
3
Online Learning with Non-decomposable Loss Functions
We now present our online learning framework for non-decomposable loss functions. Traditional
online learning takes place in several rounds, in each of which the player proposes some wt ? W
while the adversary responds with a penalty function Lt : W ? R and a loss Lt (wt ) is incurred.
PT
PT
The goal is to minimize the regret i.e.
t=1 Lt (w). For point loss
t=1 Lt (wt ) ? arg minw?W
functions, the instantaneous penalty Lt (?) is encoded using a data point (xt , yt ) ? Rd ? {?1, 1}
as Lt (w) = `P (xt , yt , w). However, for (surrogates of) non-decomposable loss functions such as
`pAUC and `Prec@k the definition of instantaneous penalty itself is not clear and remains a challenge.
To guide us in this process we turn to some properties of standard online learning frameworks. For
point losses, we note that the best solution in hindsight is also the batch optimal solution. This is
PT
equivalent to the condition arg minw?W t=1 Lt (w) = arg minw?W `P (x1:T , y1:T , w) for nondecomposable losses. Also, since the batch optimal solution is agnostic to the ordering of points,
PT
we should expect t=1 Lt (w) to be invariant to permutations within the stream. By pruning away
several naive definitions of Lt using these requirements, we arrive at the following definition:
Lt (w) = `P (x1:t , y1:t , w) ? `P (x1:(t?1) , y1:(t?1) , w).
(6)
It turns out that the above is a very natural penalty function as it measures the amount of ?extra?
penalty incurred due to the inclusion of xt into the set of points. It can be readily verified that
PT
generalizes
t=1 Lt (w) = `P (x1:T , y1:T , w) as required. Also, this penalty function seamlessly
Pt
online learning frameworks since for point losses, we have `P (x1:t , y1:t , w) = i=1 `P (xi , yi , w)
and thus Lt (w) = `P (xt , yt , w). We note that our framework also recovers the model for online
AUC maximization used in [13] and [14]. The notion of regret corresponding to this penalty is
R(T ) =
T
1X
1
Lt (wt ) ? arg min `P (x1:T , y1:T , w).
T t=1
w?W T
We note that Lt , being the difference of two loss functions, is non-convex in general and thus, standard online convex programming regret bounds cannot be applied in our framework. Interestingly, as
we show below, by exploiting structural properties of our penalty function, we can still get efficient
low-regret learning algorithms, as well as online-to-batch conversion bounds in our framework.
3.1
Low Regret Online Learning
We propose an efficient Follow-the-Regularized-Leader (FTRL) style algorithm in our framework.
Let w1 = arg minw?W kwk22 and consider the following update:
t
X
?
?
Lt (w) + kwk22 = arg min `P (x1:t , y1:t , w) + kwk22
w?W
w?W
2
2
t=1
wt+1 = arg min
(FTRL)
We would like to stress that despite the non-convexity of Lt , the FTRL objective is strongly convex
if `P is convex and thus the update can be implemented efficiently by solving a regularized batch
problem on x1:t . We now present our regret bound analysis for the FTRL update given above.
Theorem 1. Let `P (?, w) be a convex loss function and W ? Rd be a convex set. Assume w.l.o.g.
kxt k2 ? 1, ?t. Also, for the penalty function Lt in (6), let |Lt (w) ? Lt (w0 )| ? Gt ? kw ? w0 k2 ,
for all t and all w, w0 ? W, for some Gt > 0. Suppose we use the update step given in ((FTRL)) to
obtain wt+1 , 0 ? t ? T ? 1. Then for all w? , we have
q P
T
T
2 t=1 G2t
1X
1
Lt (wt ) ? `P (x1:T , y1:T , w? ) + kw? k2
.
T t=1
T
T
See Appendix A for a proof. The above result requires the penalty function Lt to be Lipschitz
continuous i.e. be ?stable? w.r.t. w. Establishing this for point losses such as hinge loss is relatively
straightforward. However, the same becomes non-trivial for non-decomposable loss functions as
4
Lt is now the difference of two loss functions, both of which involve ? (t) data points. A naive
argument would thus, only be able to show Gt ? O(t) which would yield vacuous regret bounds.
Instead, we now show that for the surrogate loss functions for Prec@k and pAUC, this Lipschitz
continuity property does indeed hold. Our proofs crucially use a structural lemma given below that
shows that sorted lists of inner products are Lipschitz at each fixed position.
Lemma 2 (Structural Lemma). Let x1 , . . . , xt be t points with kxi k2 ? 1 ?t. Let w, w0 ? W be any
two vectors. Let zi = hw, xi i ? ci and zi0 = hw0 , xi i ? ci , where ci ? R are constants independent
of w, w0 . Also, let {i1 , . . . , it } and {j1 , . . . , jt } be ordering of indices such that zi1 ? zi2 ? ? ? ? ?
zit and zj0 1 ? zj0 2 ? ? ? ? ? zj0 t . Then for any 1-Lipschitz increasing function g : R ? R (i.e.
|g(u) ? g(v)| ? |u ? v| and u ? v ? g(u) ? g(v)), we have, ?k |g(zik ) ? g(zj0 k )| ? 3kw ? w0 k2 .
See Appendix B for a proof. Using this lemma
qwecan show that the Lipschitz constant for `Prec@k
1
is bounded by Gt ? 8 which gives us a O
regret bound for Prec@k (see Appendix C for
T
the proof). In Appendix D, we show that the same technique can be used to prove a stability result
for the structural SVM surrogate of the Precision-Recall Break Even Point (PRBEP) performance
measure [3] as well. The case of pAUC is handled similarly. However, since pAUC discriminates
between positives and negatives, our previous analysis cannot be applied directly. Nevertheless, we
can obtain the following regret bound for pAUC (a proof will appear in the full version of the paper).
Theorem 3. Let T+ and T? resp. be the number of positive and negative points in the stream and
let wt+1 , 0 ? t ? T ? 1 be obtained using the FTRL algorithm ((FTRL)). Then we have
s
!
T
X
1
1
1
1
Lt (wt ) ? min
`pAUC (x1:T , y1:T , w) + O
+
.
w?W ?T+ T?
?T+ T? t=1
T+
T?
Notice that the above regret bound depends on both T+ and T? and the regret becomes large even
if one of them is small. This is actually quite intuitive because if, say T+ = 1 and T? = T ? 1,
an adversary may wish to provide the lone positive point in the last round. Naturally the algorithm,
having only seen negatives till now, would not be able to perform well and would incur a large error.
3.2
Online-to-batch Conversion
To present our bounds we generalize our framework slightly: we now consider the stream of T
points to be composed of T /s batches Z1 , . . . , ZT /s of size s each. Thus, the instantaneous penalty
is now defined as Lt (w) = `P (Z1 , . . . , Zt , w) ? `P (Z1 , . . . , Zt?1 , w) for t = 1 . . . T /s and the
PT /s
regret becomes R(T, s) = T1 t=1 Lt (wt ) ? arg minw?W T1 `P (x1:T , y1:T , w). Let RP denote
the population risk for the (normalized) performance measure P. Then we have:
Theorem 4. Suppose the sequence of points (xt , yt ) is generated i.i.d. and let w1 , w2 , . . . , wT /s
be an ensemble of models generated by an online learning algorithm upon receiving these T /s
batches. Suppose the online learning algorithm has a guaranteed regret bound R(T, s). Then for
PT /s
w = T1/s t=1 wt , any w? ? W, ? (0, 0.5] and ? > 0, with probability at least 1 ? ?,
!
r
s
ln(1/?)
??(s2 )
?
?
RP (w) ? (1 + )RP (w ) + R(T, s) + e
+O
.
T
p
?
? T ) and = 4 1/T gives us, with probability at least 1 ? ?,
In particular, setting s = O(
!
r
?
4 ln(1/?)
?
?
RP (w) ? RP (w ) + R(T, T ) + O
.
T
We conclude by noting that for Prec@k and pAUC, R(T,
4
?
T) ? O
p
4
1/
T
(see Appendix E).
Stochastic Gradient Methods for Non-decomposable Losses
The online learning algorithms discussed in the previous section present attractive guarantees in the
sequential prediction model but are required to solve batch problems at each stage. This rapidly
5
Algorithm 1 1PMB: Single-Pass with Mini-batches
Algorithm 2 2PMB: Two-Passes with Mini-batches
Input: Step length scale ?, Buffer B of size s
Input: Step length scale ?, Buffers B+ , B? of size s
Output: A good predictor w ? W
Output: A good predictor w ? W
1: w0 ? 0, B ? ?, e ? 0
Pass 1: B+ ? ?
+
2: while stream not exhausted do
1: Collect random sample of pos. x+
1 , . . . , xs in B+
3:
Collect s data points (xe1 , y1e ), . . . , (xes , yse ) in
Pass 2: w0 ? 0, B? ? ?, e ? 0
buffer B
2: while stream of negative points not exhausted do
e?
4:
Set step length ?e ? ??e
3:
Collect s negative points xe?
1 , . . . , xs in B?
?
e
4:
Set step length ?e ? ?e
5:
we+1 ? ?W [we + ?e ?w `P (xe1:s , y1:s
, we )]
+
//?W projects onto the set W
5:
we+1 ? ?W we + ?e ?w `P (xe?
1:s , x1:s , we )
6:
Flush buffer B
6:
Flush buffer B?
7:
e?e+1
//start a new epoch 7:
e?e+1
//start a new epoch
8: end while
8: end while
P
P
9: return w = 1e ei=1 wi
9: return w = 1e ei=1 wi
becomes infeasible for large scale data. To remedy this, we now present memory efficient stochastic
gradient descent methods for batch learning with non-decomposable loss functions. The motivation
for our approach comes from mini-batch methods used to make learning methods for point loss
functions amenable to distributed computing environments [15, 16], we exploit these techniques to
offer scalable algorithms for non-decomposable loss functions.
Single-pass Method with Mini-batches. The method assumes access to a limited memory buffer
and takes a pass over the data stream. The stream is partitioned into epochs. In each epoch, the
method accumulates points in the stream, uses them to form gradient estimates and takes descent
steps. The buffer is flushed after each epoch. Algorithm 1 describes the 1PMB method. Gradient
computations can be done using Danskin?s theorem (see Appendix H).
Two-pass Method with Mini-batches. The previous algorithm is unable to exploit relationships
between data points across epochs which may help improve performance for loss functions such as
pAUC. To remedy this, we observe that several real life learning scenarios exhibit mild to severe
label imbalance (see Table 2 in Appendix H) which makes it possible to store all or a large fraction
of points of the rare label. Our two pass method exploits this by utilizing two passes over the data:
the first pass collects all (or a random subset of) points of the rare label using some stream sampling
technique [13]. The second pass then goes over the stream, restricted to the non-rare label points,
and performs gradient updates. See Algorithm 2 for details of the 2PMB method.
4.1
Error Bounds
Given a set of n labeled data points (xi , yi ), i = 1 . . . n and a performance measure P, our goal is to
approximate the empirical risk minimizer w? = arg min `P (x1:n , y1:n , w) as closely as possible.
w?W
In this section we shall show that our methods 1PMB and 2PMB provably converge to the empirical
risk minimizer. We first introduce the notion of uniform convergence for a performance measure.
Definition 5. We say that a loss function ` demonstrates
uniform convergence with respect to a set of
?1, . . . , x
? s chosen
predictors W if for some ?(s, ?) = poly 1s , log 1? , when given a set of s points x
randomly from an arbitrary set of n points {(x1 , y1 ), . . . , (xn , yn )} then w.p. at least 1 ? ?, we have
sup |`P (x1:n , y1:n , w) ? `P (?
x1:s , y?1:s , w)| ? ?(s, ?).
w?W
Such uniform convergence results are fairly common for decomposable loss functions such as the
squared loss, logistic loss etc. However, the same is not true for non-decomposable loss functions
barring a few exceptions [17, 10]. To bridge this gap, below we show that a large family of surrogate
loss functions for popular non decomposable performance measures does indeed exhibit uniform
convergence. Our proofs require novel techniques and do not follow from traditional proof progressions. However, we first show how we can use these results to arrive at an error bound.
Theorem 6. Suppose the loss function ` is convex and demonstrates ?(s, ?)-uniform convergence.
Also suppose we have an arbitrary set of n points which are randomly ordered, then the predictor
6
CP
PSG
1PMB
2PMB
0.3
0.2
0.1
0
1
2
3
4
Training time (secs)
0.6
0.4
CP
PSG
1PMB
2PMB
0.2
5
0
(a) PPI
0.2
0.4
0.6
Training time (secs)
0.6
0.4
CP
PSG
1PMB
2PMB
0.2
0.8
0
(b) KDDCup08
Average pAUC in [0, 0.1]
0.4
Average pAUC in [0, 0.1]
0.5
Average pAUC in [0, 0.1]
Average pAUC in [0, 0.1]
0.6
0.6
0.5
0.4
CP
PSG
1PMB
2PMB
0.3
0.2
0.1
0.5
1
1.5
Training time (secs)
0
(c) IJCNN
0.1
0.2
0.3
Training time (secs)
(d) Letter
Figure 1: Comparison of stochastic gradient methods with the cutting plane (CP) and projected
subgradient (PSG) methods on partial AUC maximization tasks. The epoch lengths/buffer sizes for
1PMB and 2PMB were set to 500.
0.1
0
2
0.4
0.3
CP
1PMB
2PMB
0.2
0.1
4
6
8
10
Training time (secs)
0
(a) PPI
10
20
30
Training time (secs)
0.6
0.4
CP
1PMB
2PMB
0.2
0
(b) KDDCup08
Average Prec@k
CP
1PMB
2PMB
Average Prec@k
0.2
Average Prec@k
Average Prec@k
0.5
0.3
0.4
0.3
CP
1PMB
2PMB
0.2
0.1
5
10
Training time (secs)
0
(c) IJCNN
0.2
0.4
0.6
Training time (secs)
0.8
(d) Letter
Figure 2: Comparison of stochastic gradient methods with the cutting plane (CP) method on Prec@k
maximization tasks. The epoch lengths/buffer sizes for 1PMB and 2PMB were set to 500.
w returned by 1PMB with buffer size s satisfies w.p. 1 ? ?,
s?
`P (x1:n , y1:n , w) ? `P (x1:n , y1:n , w? ) + 2? s,
n
r
s
+O
n
We would like to stress that the above result does not assume i.i.d. data and works for arbitrary
datasets so long as they are randomly ordered. We can show similar guarantees for the two pass
method as well (see Appendix F). Using regularized formulations, we can also exploit logarithmic
regret guarantees [18], offered by online gradient descent, to improve this result ? however we do not
explore those considerations here. Instead, we now look at specific instances of loss functions that
possess the desired uniform convergence properties. As mentioned before, due to the combinatorial
nature of these performance measures, our proofs do not follow from traditional methods.
Theorem 7 (Partial Area under the ROC Curve). For any convex, monotone, Lipschitz, classification
surrogate ? : R ? R+ , the surrogate loss function for the (0, ?)-partial
AUC performance
measure
p
log(1/?)/s :
defined as follows exhibits uniform convergence at the rate ?(s, ?) = O
1
d?n? en+
X
X
>
>
T?
?,t (xj , w) ? ?(xi w ? xj w)
i:yi >0 j:yj <0
See Appendix G for a proof sketch. This result covers a large family of surrogate loss functions such
as hinge loss (5), logistic loss etc. Note that the insistence on including only top ranked negative
points introduces a high degree of non-decomposability into the loss function. A similar result for
the special case ? = 1 is due to [17]. We extend the same to the more challenging case of ? < 1.
Theorem 8 (Structural SVM loss for Prec@k ). The structural SVM surrogate forthe Prec@k per
p
formance measure (see (3)) exhibits uniform convergence at the rate ?(s, ?) = O
log(1/?)/s .
We defer the proof to the full version of the paper. The above result can be extended to a large family
of performances measures introduced in [3] that have been widely adopted [10, 19, 8] such as Fmeasure, G-mean, and PRBEP. The above indicates that our methods are expected to output models
that closely approach the empirical risk minimizer for a wide variety of performance measures. In
the next section we verify that this is indeed the case for several real life and benchmark datasets.
5
Experimental Results
We evaluate the proposed stochastic gradient methods on several real-world and benchmark datasets.
7
2PMB
0.15 (69.6)
0.55 (38.7)
CP
0.39 (62.5)
23.25 (40.8)
0.6
0.54
0.52
0.5
0.48
0.46
0.44
0.42
1PMB
0
Table 1: Comparison of training time (secs) and accuracies (in brackets) of 1PMB, 2PMB and cutting plane
methods for pAUC (in [0, 0.1]) and Prec@k maximization tasks on the KDD Cup 2008 dataset.
Average pAUC
1PMB
0.10 (68.2)
0.49 (42.7)
Average pAUC
Measure
pAUC
Prec@k
10
2
10
Epoch length
4
10
0.55
0.5
0.45
2PMB
0
10
2
10
Epoch length
4
10
Figure 3: Performance of 1PMB and
2PMB on the PPI dataset with varying
epoch/buffer sizes for pAUC tasks.
Performance measures: We consider three measures, 1) partial AUC in the false positive range
[0, 0.1], 2) Prec@k with k set to the proportion of positives (PRBEP), and 3) F-measure.
Algorithms: For partial AUC, we compare against the state-of-the-art cutting plane (CP) and projected subgradient methods (PSG) proposed in [7]; unlike the (online) stochastic methods considered
in this work, the PSG method is a ?batch? algorithm which, at each iteration, computes a subgradientbased update over the entire training set. For Prec@k and F-measure, we compare our methods
against cutting plane methods from [3]. We used structural SVM surrogates for all the measures.
Datasets: We used several data sets for our experiments (see Table 2 in Appendix H); of these,
KDDCup08 is from the KDD Cup 2008 challenge and involves a breast cancer detection task [20],
PPI contains data for a protein-protein interaction prediction task [21], and the remaining datasets
are taken from the UCI repository [22].
Parameters: We used 70% of the data set for training and the remaining for testing, with the results
averaged over 5 random train-test splits. Tunable parameters such as step length scale were chosen
using a small validation set. The epoch lengths/buffer sizes were set to 500 in all experiments. Since
a single iteration of the proposed stochastic methods is very fast in practice, we performed multiple
passes over the training data (see Appendix H for details).
The results for pAUC and Prec@k maximization tasks are shown in the Figures 1 and 2. We found
the proposed stochastic gradient methods to be several orders of magnitude faster than the baseline
methods, all the while achieving comparable or better accuracies. For example, for the pAUC task
on the KDD Cup 2008 dataset, the 1PMB method achieved an accuracy of 64.81% within 0.03
seconds, while even after 0.39 seconds, the cutting plane method could only achieve an accuracy
of 62.52% (see Table 1). As expected, the (online) stochastic gradient methods were faster than
the ?batch? projected subgradient descent method for pAUC as well. We found similar trends on
Prec@k (see Figure 2) and F-measure maximization tasks as well. For F-measure tasks, on the KDD
Cup 2008 dataset, for example, the 1PMB method achieved an accuracy of 35.92 within 12 seconds
whereas, even after 150 seconds, the cutting plane method could only achieve an accuracy of 35.25.
The proposed stochastic methods were also found to be robust to changes in epoch lengths (buffer
sizes) till such a point where excessively long epochs would cause the number of updates as well as
accuracy to dip (see Figure 3). The 2PMB method was found to offer higher accuracies for pAUC
maximization on several datasets (see Table 1 and Figure 1), as well as be more robust to changes
in buffer size (Figure 3). We defer results on more datasets and performance measures to the full
version of the paper.
The cutting plane methods were generally found to exhibit a zig-zag behaviour in performance
across iterates. This is because these methods solve the dual optimization problem for a given performance measure; hence the intermediate models do not necessarily yield good accuracies. On the
other hand, (stochastic) gradient based methods directly offer progress in terms of the primal optimization problem, and hence provide good intermediate solutions as well. This can be advantageous
in scenarios with a time budget in the training phase.
Acknowledgements
The authors thank Shivani Agarwal for helpful comments. They also thank the anonymous reviewers
for their suggestions. HN thanks support from a Google India PhD Fellowship.
8
References
[1] Alexander Rakhlin. Lecture Notes on Online Learning. http://www-stat.wharton.upenn.
edu/?rakhlin/papers/online_learning.pdf, 2009.
[2] Harikrishna Narasimhan and Shivani Agarwal. A Structural SVM Based Approach for Optimizing Partial
AUC. In 30th International Conference on Machine Learning (ICML), 2013.
[3] Thorsten Joachims. A Support Vector Method for Multivariate Performance Measures. In ICML, 2005.
[4] Yisong Yue, Thomas Finley, Filip Radlinski, and Thorsten Joachims. A Support Vector Method for
Optimizing Average Precision. In SIGIR, 2007.
[5] Soumen Chakrabarti, Rajiv Khanna, Uma Sawant, and Chiru Bhattacharyya. Structured Learning for
Non-Smooth Ranking Losses. In KDD, 2008.
[6] Brian McFee and Gert Lanckriet. Metric Learning to Rank. In ICML, 2010.
[7] Harikrishna Narasimhan and Shivani Agarwal. SVMtight
pAUC : A New Support Vector Method for Optimizing
Partial AUC Based on a Tight Convex Upper Bound. In KDD, 2013.
[8] Miroslav Kubat and Stan Matwin. Addressing the Curse of Imbalanced. Training Sets: One-Sided Selection. In 24th International Conference on Machine Learning (ICML), 1997.
[9] Krzysztof Dembczy?nski, Willem Waegeman, Weiwei Cheng, and Eyke H?ullermeier. An Exact Algorithm
for F-Measure Maximization. In NIPS, 2011.
[10] Nan Ye, Kian Ming A. Chai, Wee Sun Lee, and Hai Leong Chieu. Optimizing F-Measures: A Tale of
Two Approaches. In 29th International Conference on Machine Learning (ICML), 2012.
[11] Krzysztof Dembczy?nski, Arkadiusz Jachnik, Wojciech Kotlowski, Willem Waegeman, and Eyke
H?ullermeier. Optimizing the F-Measure in Multi-Label Classification: Plug-in Rule Approach versus
Structured Loss Minimization. In 30th International Conference on Machine Learning (ICML), 2013.
[12] Alexander Rakhlin, Karthik Sridharan, and Ambuj Tewari. Online Learning: Beyond Regret. In 24th
Annual Conference on Learning Theory (COLT), 2011.
[13] Purushottam Kar, Bharath K Sriperumbudur, Prateek Jain, and Harish Karnick. On the Generalization
Ability of Online Learning Algorithms for Pairwise Loss Functions. In ICML, 2013.
[14] Peilin Zhao, Steven C. H. Hoi, Rong Jin, and Tianbao Yang. Online AUC Maximization. In ICML, 2011.
[15] Ofer Dekel, Ran Gilad-Bachrach, Ohad Shamir, and Lin Xiao. Optimal Distributed Online Prediction
Using Mini-Batches. Journal of Machine Learning Research, 13:165?202, 2012.
[16] Yuchen Zhang, John C. Duchi, and Martin J. Wainwright. Communication-Efficient Algorithms for Statistical Optimization. Journal of Machine Learning Research, 14:3321?3363, 2013.
[17] St?ephan Cl?emenc?on, G?abor Lugosi, and Nicolas Vayatis. Ranking and empirical minimization of Ustatistics. Annals of Statistics, 36:844?874, 2008.
[18] Elad Hazan, Adam Kalai, Satyen Kale, and Amit Agarwal. Logarithmic Regret Algorithms for Online
Convex Optimization. In COLT, pages 499?513, 2006.
[19] Sophia Daskalaki, Ioannis Kopanas, and Nikolaos Avouris. Evaluation of Classifiers for an Uneven Class
Distribution Problem. Applied Artificial Intelligence, 20:381?417, 2006.
[20] R. Bharath Rao, Oksana Yakhnenko, and Balaji Krishnapuram. KDD Cup 2008 and the Workshop on
Mining Medical Data. SIGKDD Explorations Newsletter, 10(2):34?38, 2008.
[21] Yanjun Qi, Ziv Bar-Joseph, and Judith Klein-Seetharaman. Evaluation of Different Biological Data and
Computational Classification Methods for Use in Protein Interaction Prediction. Proteins, 63:490?500,
2006.
[22] A. Frank and Arthur Asuncion. The UCI Machine Learning Repository. http://archive.ics.
uci.edu/ml, 2010. University of California, Irvine, School of Information and Computer Sciences.
[23] Ankan Saha, Prateek Jain, and Ambuj Tewari. The interplay between stability and regret in online learning. CoRR, abs/1211.6158, 2012.
[24] Martin Zinkevich. Online Convex Programming and Generalized Infinitesimal Gradient Ascent. In ICML,
pages 928?936, 2003.
[25] Robert J. Serfling. Probability Inequalities for the Sum in Sampling without Replacement. Annals of
Statistics, 2(1):39?48, 1974.
[26] Dimitri P. Bertsekas. Nonlinear Programming: 2nd Edition. Belmont, MA: Athena Scientific, 2004.
9
| 5367 |@word mild:2 repository:2 version:3 polynomial:1 proportion:1 advantageous:1 nd:1 dekel:1 seek:1 crucially:2 mention:1 harder:1 ftrl:7 contains:1 score:1 uma:1 offering:1 interestingly:1 bhattacharyya:1 existing:2 com:1 define1:1 readily:1 john:1 belmont:1 additive:1 j1:1 kdd:8 drop:1 update:7 zik:1 intelligence:1 instantiate:4 plane:12 vanishing:1 characterization:1 provides:1 iterates:1 judith:1 zhang:1 become:1 sophia:1 chakrabarti:1 prove:3 introduce:1 manner:1 pairwise:1 upenn:1 expected:2 indeed:4 behavior:1 themselves:1 frequently:4 multi:1 ming:1 xti:1 curse:1 solver:5 increasing:1 becomes:5 iisc:1 project:1 moreover:1 notation:2 bounded:1 agnostic:1 kubat:1 prateek:3 xe1:2 narasimhan:3 lone:1 hindsight:1 guarantee:3 k2:5 demonstrates:2 classifier:1 control:2 medical:3 appear:1 yn:1 arguably:1 bertsekas:1 positive:11 t1:3 understood:1 before:1 insistence:1 despite:2 accumulates:1 analyzing:1 establishing:1 nonadditive:1 lugosi:1 might:1 yb1:1 g2t:1 collect:4 challenging:1 limited:1 zi0:1 range:3 averaged:1 yj:5 testing:1 practice:1 regret:24 ybt:1 mcfee:1 nondecomposable:1 area:4 empirical:6 significantly:1 yakhnenko:1 protein:4 krishnapuram:1 get:1 cannot:4 onto:1 selection:1 risk:6 seminal:1 optimize:1 zinkevich:1 equivalent:4 demonstrated:1 yt:5 reviewer:1 www:1 straightforward:1 attention:1 starting:1 go:1 convex:16 sigir:1 rajiv:1 bachrach:1 decomposable:27 simplicity:2 emenc:1 rule:1 utilizing:1 stability:5 population:1 notion:4 gert:1 resp:1 pt:8 suppose:5 shamir:1 annals:2 exact:1 programming:3 us:2 designing:1 lanckriet:1 trend:1 satisfying:1 svmtight:1 balaji:1 labeled:2 observed:1 role:1 steven:1 sun:1 ordering:2 valuable:1 zig:1 principled:1 mentioned:2 discriminates:1 convexity:1 complexity:1 environment:1 ran:1 tight:2 solving:1 incur:1 negatively:1 upon:1 matwin:1 po:1 various:1 train:1 jain:3 instantiated:2 fast:1 artificial:1 quite:1 encoded:1 widely:2 valued:1 solve:3 larger:1 say:2 otherwise:1 elad:1 ability:3 statistic:2 satyen:1 itself:3 online:41 interplay:1 kxt:1 sequence:1 took:1 propose:3 interaction:2 product:2 uci:3 rapidly:1 soumen:1 holistic:1 iff:2 till:2 achieve:5 intuitive:1 exploiting:1 convergence:11 chai:1 requirement:1 incremental:1 adam:1 tk:1 help:1 develop:3 tale:1 stat:1 school:1 received:2 progress:1 zit:1 implemented:1 predicted:1 involves:1 come:1 closely:2 owing:1 stochastic:15 exploration:1 enable:1 hoi:1 require:5 behaviour:1 generalization:2 anonymous:1 brian:1 biological:1 extension:1 rong:1 hold:1 considered:2 ic:1 early:1 label:10 combinatorial:1 sensitive:2 bridge:1 minimization:2 aim:1 modified:1 kalai:1 varying:1 focus:4 joachim:2 rank:3 indicates:1 seamlessly:1 contrast:1 sigkdd:1 baseline:1 helpful:1 entire:3 typically:2 abor:1 i1:1 interested:1 provably:3 jachnik:1 arg:9 overall:1 classification:5 dual:2 ziv:1 colt:2 proposes:1 art:2 special:1 ernet:1 fairly:1 oksana:1 wharton:1 having:1 barring:1 sampling:2 kw:3 look:1 icml:9 nearly:1 ullermeier:2 bangalore:1 few:2 saha:1 modern:2 randomly:3 composed:1 wee:1 individual:1 phase:1 ourselves:1 replacement:1 microsoft:2 karthik:1 ab:1 detection:2 interest:3 mining:1 evaluation:2 severe:2 introduces:1 bracket:1 primal:1 fmeasure:1 amenable:1 kt:1 partial:12 arthur:1 minw:5 ohad:1 biometrics:1 conduct:1 yuchen:1 desired:1 miroslav:1 instance:4 boolean:1 rao:1 zj0:4 cover:1 maximization:10 applicability:1 addressing:1 imperative:1 decomposability:2 rare:3 uniform:11 predictor:6 subset:1 dembczy:2 kxi:1 nski:2 thanks:1 st:1 international:4 probabilistic:1 lee:1 receiving:1 w1:2 squared:1 yisong:1 hn:1 zhao:1 style:2 return:3 wojciech:1 dimitri:1 sec:9 ioannis:1 includes:1 satisfy:1 ranking:4 depends:1 stream:10 performed:1 try:1 lot:1 break:1 hazan:1 sup:1 start:2 recover:1 asuncion:1 defer:2 contribution:2 minimize:1 accuracy:10 formance:1 characteristic:2 who:1 efficiently:1 yield:2 ensemble:1 ybi:3 generalize:1 bharath:2 psg:7 email:1 definition:8 infinitesimal:1 against:3 sriperumbudur:1 naturally:1 proof:14 recovers:1 irvine:1 dataset:7 tunable:1 popular:7 recall:2 formalize:1 harikrishna:4 actually:1 higher:2 follow:4 wherein:1 formulation:4 done:1 strongly:1 stage:1 hand:5 sketch:1 replacing:1 ei:2 nonlinear:1 google:1 continuity:2 khanna:1 logistic:2 scientific:1 excessively:1 normalized:1 true:4 remedy:2 counterpart:1 verify:1 hence:2 ye:1 attractive:1 round:2 eyke:2 game:1 interchangeably:1 auc:14 m:1 generalized:1 pdf:1 stress:2 risen:1 performs:1 cp:12 duchi:1 newsletter:1 instantaneous:4 novel:5 recently:2 consideration:1 common:2 extend:2 discussed:1 significant:1 cup:5 rd:4 resorting:1 similarly:1 inclusion:1 stable:1 access:1 etc:7 gt:4 multivariate:1 imbalanced:2 purushottam:2 optimizing:5 scenario:2 store:1 certain:1 buffer:14 inequality:1 kar:2 binary:1 success:1 arkadiusz:1 life:4 yi:12 xe:2 seen:1 converge:3 full:3 desirable:2 multiple:1 smooth:1 faster:4 plug:1 offer:5 long:2 lin:1 qi:1 prediction:8 scalable:3 variant:1 breast:1 metric:1 iteration:2 agarwal:4 achieved:2 gilad:1 vayatis:1 whereas:2 hurdle:1 fine:2 fellowship:1 extra:1 w2:1 unlike:2 posse:1 kotlowski:1 pass:3 comment:1 kwk22:3 yue:1 archive:1 ascent:1 sridharan:1 structural:14 noting:1 yang:1 intermediate:2 split:1 weiwei:1 leong:1 variety:2 xj:9 zi:1 restrict:1 inner:2 ankan:1 handled:1 utility:1 penalty:13 returned:1 yanjun:1 cause:1 constitute:1 deep:1 useful:2 generally:1 clear:2 involve:3 tewari:2 nikolaos:1 amount:1 shivani:3 http:2 kian:1 canonical:1 notice:1 per:3 klein:1 diagnosis:1 shall:5 pauc:36 key:1 waegeman:2 seetharaman:1 nevertheless:1 achieving:1 verified:1 krzysztof:2 subgradient:3 monotone:2 fraction:4 sum:2 letter:2 place:1 family:6 arrive:2 tianbao:1 appendix:11 peilin:1 comparable:3 bound:18 guaranteed:1 nan:1 cheng:1 annual:1 ijcnn:2 sake:1 argument:1 min:5 relatively:1 martin:2 structured:3 developing:1 flush:2 describes:1 slightly:2 across:2 serfling:1 wi:2 partitioned:1 joseph:1 restricted:2 invariant:1 thorsten:2 sided:1 taken:1 ln:2 remains:1 turn:2 mind:1 initiate:1 prajain:1 end:3 adopted:2 ofer:1 available:1 generalizes:1 experimentation:1 willem:2 apply:1 observe:1 progression:1 away:1 prec:31 generic:2 indirectly:1 zi2:1 batch:23 rp:5 existence:1 thomas:1 top:6 assumes:1 include:3 remaining:2 harish:1 hinge:6 ppi:4 medicine:1 exploit:4 prof:1 establish:3 amit:1 objective:3 strategy:2 rt:1 traditional:4 surrogate:15 said:1 exhibit:6 gradient:17 amongst:1 responds:1 hai:1 unable:1 thank:2 athena:1 w0:8 considers:3 trivial:1 provable:1 length:11 index:1 relationship:1 mini:6 difficult:2 robert:1 frank:1 negative:7 danskin:1 design:5 implementation:1 proper:1 zt:3 perform:2 conversion:4 imbalance:4 upper:1 datasets:11 benchmark:4 descent:5 jin:1 situation:3 defining:1 zi1:1 extended:1 communication:1 y1:20 arbitrary:3 community:2 kale:1 ephan:1 introduced:2 vacuous:1 namely:1 pair:1 required:2 extensive:2 z1:3 california:1 nip:1 able:4 adversary:2 beyond:1 below:4 bar:1 challenge:5 ambuj:2 including:3 max:3 memory:2 wainwright:1 ranked:5 natural:1 regularized:4 indicator:3 improve:2 stan:1 finley:1 naive:2 nice:1 understanding:1 literature:1 epoch:14 acknowledgement:1 loss:78 expect:1 permutation:1 lecture:1 sublinear:3 mixed:1 suggestion:1 versus:1 validation:1 incurred:2 degree:1 offered:2 xiao:1 tiny:1 cancer:2 last:1 infeasible:1 enjoys:1 offline:1 guide:1 institute:1 india:3 wide:2 distributed:2 curve:4 calculated:1 xn:1 world:1 dip:1 karnick:1 computes:1 author:1 projected:3 spam:2 pruning:1 approximate:1 cutting:12 ml:1 filip:1 conclude:1 leader:2 xi:9 continuous:1 decomposes:1 chief:1 table:5 learn:1 nature:1 robust:2 nicolas:1 csa:1 complex:1 poly:1 constructing:1 domain:2 necessarily:1 cl:1 main:2 motivation:1 edition:1 x1:23 en:1 roc:4 precision:9 position:2 wish:1 grained:2 hw:1 theorem:8 xt:7 specific:1 jt:1 showing:1 list:4 x:3 admits:1 svm:6 rakhlin:3 workshop:1 false:4 sequential:1 corr:1 importance:1 ci:3 phd:1 magnitude:2 te:1 budget:1 exhausted:2 gap:1 led:1 lt:25 logarithmic:2 explore:1 expressed:1 ordered:2 chieu:1 minimizer:5 satisfies:1 ma:1 chiru:1 goal:3 sorted:2 consequently:1 lipschitz:8 change:2 specifically:1 wt:12 lemma:8 pas:10 experimental:1 player:1 zag:1 hw0:1 exception:1 select:1 uneven:1 support:4 radlinski:1 fulfills:1 alexander:2 indian:1 evaluate:3 |
4,824 | 5,368 | Optimistic planning in Markov decision processes
using a generative model
Bal?azs Sz?or?enyi
INRIA Lille - Nord Europe,
SequeL project, France /
MTA-SZTE Research Group on
Arti?cial Intelligence, Hungary
[email protected]
Gunnar Kedenburg
INRIA Lille - Nord Europe,
SequeL project, France
[email protected]
Remi Munos?
INRIA Lille - Nord Europe,
SequeL project, France
[email protected]
Abstract
We consider the problem of online planning in a Markov decision process with
discounted rewards for any given initial state. We consider the PAC sample complexity problem of computing, with probability 1??, an ?-optimal action using the
smallest possible number of calls to the generative model (which provides reward
and next-state samples). We design an algorithm, called StOP (for StochasticOptimistic Planning), based on the ?optimism in the face of uncertainty? principle. StOP can be used in the general setting, requires only a generative model, and
enjoys a complexity bound that only depends on the local structure of the MDP.
1
Introduction
1.1
Problem formulation
In a Markov decision process (MDP), an agent navigates in a state space X by making decisions
from some action set U . The dynamics of the system are determined by transition probabilities
P : X ? U ? X ? [0, 1] and reward probabilities R : X ? U ? [0, 1] ? [0, 1], as follows: when
the agent chooses action u in state x, then, with probability R(x, u, r), it receives reward r, and with
probability P (x, u, x? ) it makes a transition to a next state x? . This happens independently of all
previous actions, states and rewards?that is, the system possesses the Markov property. See [20, 2]
for a general introduction to MDPs. We do not assume that the transition or reward probabilities
are fully known. Instead, we assume access to the MDP via a generative model (e.g. simulation
software), which, for a state-action (x, u), returns a reward sample r ? R(x, u, ?) and a next-state
sample x? ? P (x, u, ?). We also assume the number of possible next-states to be bounded by N ? N.
We would like to ?nd?
an agent that implements a policy which maximizes the expected cumulative
?
discounted reward E[ t=0 ? t rt ], which we will also refer to as the return. Here, rt is the reward
received at time t and ? ? (0, 1) is the discount factor. Further, we take an online planning approach,
where at each time step, the agent uses the generative model to perform a simulated search (planning)
in the set of policies, starting from the current state. As a result of this search, the agent takes a single
action. An expensive global search for the optimal policy in the whole MDP is avoided.
?
Current af?liation: Google DeepMind
1
To quantify the performance of our algorithm, we consider a PAC (Probably Approximately Correct)
setting, where, given ? > 0 and ? ? (0, 1), our algorithm returns, with probability 1??, an ?-optimal
action (i.e. such that the loss of performing this action and then following an optimal policy instead
of following an optimal policy from the beginning is at most ?). The number of calls to the generative
model required by the planning algorithm is referred to as its sample complexity. The sample and
computational complexities of the planning algorithm introduced here depend on local properties
of the MDP, such as the quantity of near-optimal policies starting from the initial state, rather than
global features like the MDP?s size.
1.2
Related work
The online planning approach and, in particular, its ability to get rid of the dependency on the global
features of the MDP in the complexity bounds (mentioned above, and detailed further below) is
the driving force behind the Monte Carlo Tree Search algorithms [16, 8, 11, 18]. 1 The theoretical analysis of this approach is still far from complete. Some of the earlier algorithms use strong
assumptions, others are applicable only in restricted cases, or don?t adapt to the complexity of the
problem. In this paper we build on ideas used in previous works, and aim at ?xing these issues.
A ?rst related work is the sparse sampling algorithm of [14]. It builds a uniform look-ahead tree of a
given depth (which depends on the precision ?), using for each transition a ?nite number of samples
obtained from a generative model. An estimate of the value function is then built using empirical
averaging instead of expectations in the dynamic programming back-up scheme. This results in an
2 )])
?
? log K+log[1/(?(1??)
log(1/?)
1
algorithm with (problem-independent) sample complexity of order (1??)
3?
(neglecting some poly-logarithmic dependence), where K is the number of actions. In terms of ?,
this bound scales as exp(O([log(1/?)]2 )), which is non-polynomial in 1/?. 2 Another disadvantage
of the algorithm is that the expansion of the look-ahead tree is uniform; it does not adapt to the MDP.
An algorithm which addresses this appears in [21]. It avoids evaluating some unnecessary branches
of the look-ahead tree of the sparse sampling algorithm. However, the provided sample bound does
not improve on the one in [14], and it is possible to show that the bound is tight (for both algorithms).
In fact, the sample complexity turns out to be super-polynomial even in the pure Monte Carlo setting
1
(i.e., when K = 1): 1/?2+(log C)/ log(1/?) , with C ? ?2 (1??)
4.
Close to our contribution are the planning algorithms [13, 3, 5, 15] (see also the survey [18]) that
follow the so-called ?optimism in the face of uncertainty? principle for online planning. This principle has been extensively investigated in the multi-armed bandit literature (see e.g. [17, 1, 4]). In
the planning problem, this approach translates to prioritizing the most promising part of the policy
space during exploration. In [13, 3, 5], the sample complexity depends on a measure of the quantity
of near-optimal policies, which gives a better understanding of the real hardness of the problem than
the uniform bound in [14].
The case of deterministic dynamics and rewards is considered in [13]. The proposed algorithm has
log ?
sample complexity of order (1/?) log(1/?) , where ? ? [1, K] measures (as a branching factor) the
quantity of nodes of the planning tree that belong to near-optimal policies. If all policies are very
good, many nodes need to be explored in order to distinguish the optimal policies from the rest, and
log K
therefore, ? is close to the number of actions K, resulting in the minimax bound of (1/?) log(1/?) .
Now if there is structure in the rewards (e.g. when sub-optimal policies can be eliminated by observing the ?rst rewards along the sequence), then the proportion of near-optimal policies is low,
so ? can be small and the bound is much better. In [3], the case of stochastic rewards have been
considered. However, in that work the performance is not compared to the optimal (closed-loop)
policy, but to the best open-loop policy (i.e. which does not depends on the state but only on the
log(?)
sequence of actions). In that situation, the sample complexity is of order (1/?)max(2, log(1/?) ) .
The deterministic and open-loop settings are relatively simple, since any policy can be identi?ed with
a sequence of actions. In the general MDP case however, a policy corresponds to an exponentially
1
A similar planning approach has been considered in the control literature, such as the model-predictive
control [6] or in the AI community, such as the A? heuristic search [19] and the AO? variant [12].
2
A problem-independent lower bound for the sample complexity, of order (1/?)1/ log(1/?) , is provided too.
2
wide tree, where several branches need to be explored. The closest work to ours in this respect is
[5]. However, it makes the (strong) assumption that a full model of the rewards and transitions is
? ? log(?)
available. The sample complexity achieved is again 1/? log(1/?) , but where ? ? (1, N K] is de?ned
as the branching factor of the set of nodes that simultaneously (1) belong to near-optimal policies,
and (2) whose ?contribution? to the value function at the initial state is non-negligible.
1.3
The main results of the paper
Our main contribution is a planning algorithm, called StOP (for Stochastic Optimistic Planning)
that achieves a polynomial sample complexity in terms of ? (which can be regarded as the leading
parameter in this problem), and which is, in terms of this complexity, competitive to other algorithms
that can exploit more speci?cs of their respective domains. It bene?ts from possible reward or
transition probability structures, and does not require any special restriction or knowledge about the
MDP besides having access to a generative model. The sample complexity bound is more involved
than in previous works, but can be upper-bounded by:
log ?
(1/?)2+ log(1/?) +o(1)
(1)
The important quantity ? ? [1, KN ] plays the role of a branching factor of the set of important
states S ?,? (de?ned precisely later) that ?contribute? in a signi?cant way to near-optimal policies.
These states have a non-negligible probability to be reached when following some near-optimal
policy. This measure is similar (but with some differences illustrated below) to the ? introduced in
the analysis of OP-MDP in [5]. Comparing the two, (1) contains an additional constant of 2 in the
exponent. This is a consequence of the fact that the rewards are random and that we do not have
access to the true probabilities, only to a generative model generating transition and reward samples.
In order to provide intuition about the bound, let us consider several speci?c cases (the derivation of
these bounds can be found in Section E):
? Worst-case. When there is no structure at all, then S ?,? may potentially be the set of
all possible reachable nodes (up to some depth which depends on ?), and its branching
factor is ? = KN . The sample complexity is thus of order (neglecting logarithmic faclog(KN )
tors) (1/?)2+ log(1/?) . This is the same complexity that uniform planning algorithm would
achieve. Indeed, uniform planning would build a tree of depth h with branching factor KN
where from each state-action one would generate m rewards and next-state samples. Then,
dynamic programming would be used with the empirical Bellman operator built from the
samples. Using Chernoff-Hoeffding bound,
? the estimation error is of the order (neglecting
logarithms and (1 ? ?) dependence) of 1/ m. So for a desired error ? we need to choose h
of order log(1/?)/ log(1/?), and m of order 1/?2 leading to a sample complexity of order
log(KN )
m(KN )h = (1/?)2+ log(1/?) . (See also [15]) Note that in the worst-case sense there is no
uniformly better strategy than a uniform planning, which is achieved by StOP. However,
StOP can also do much better in speci?c settings, as illustrated next.
? Case with K0 > 1 actions at the initial state, K1 = 1 actions for all other states, and
arbitrary transition probabilities. Now each branch corresponds to a single policy. In
that case one has ? = 1 (even though N > 1) and the sample complexity of StOP is of
2
?
) with high probability3 . This is the same rate as a Monte-Carlo evalorder O(log(1/?)/?
uation strategy would achieve, by sampling O(log(1/?)/?2 ) random trajectories of length
log(1/?)/ log(1/?). Notice that this result is surprisingly different from OP-MDP which
log N
has a complexity of order (1/?) log(1/?) (in the case when ? = N , i.e., when all transitions
are uniform). Indeed, in the case of uniform transition probabilities, OP-MDP would sample the nodes in breadth-?rst search way, thus achieving this minimax-optimal complexity.
2
?
This does not contradict the O(log(1/?)/?
) bound for StOP (and Monte-Carlo) since this
bound applies to an individual problem and holds in high probability, whereas the bound
for OP-MDP is deterministic and holds uniformly over all problems of this type.
3
We emphasize the dependence on ? here since we want to compare this high-probability bound to the
deterministic bound of OP-MDP.
3
Here we see the potential bene?t of using StOP instead of OP-MDP, even though StOP
only uses a generative model of the MDP whereas OP-MDP requires a full model.
? Highly structured policies. This situation holds when there is a substantial gap between
near optimal policies and other sub-optimal policies. For example if along an optimal
policy, all immediate rewards are 1, whereas as soon as one deviates from it, all rewards
are < 1. Then only a small proportion of the nodes (the ones that contribute to near-optimal
policies) will be expanded by the algorithm. In such cases, ? is very close to 1 and in the
limit, we recover the previous case when K = 1 and the sample complexity is O(1/?)2 .
? Deterministic MDPs. Here N = 1 and we have that ? ? [1, K]. When there is structure in
2
?
). Now when
the rewards (like in the previous case), then ? = 1 and we obtain a rate O(1/?
the MDP is almost deterministic, in the sense that N > 1 but from any state-action, there
is one next-state probability which is close to 1, then we have almost the same complexity
as in the deterministic case (since the nodes that have a small probability to be reached will
not contribute to the set of important nodes S ?,? , which characterizes ?).
? Multi-armed bandit we essentially recover the result of the Action Elimination algorithm
[9] for the PAC setting.
Thus we see that in the worst case StOP is minimax-optimal, and in addition, StOP is able to bene?t
from situations when there is some structure either in the rewards or in the transition probabilities.
We stress that StOP achieves the above mentioned results having no knowledge about ?.
1.4
The structure of the paper
Section 2 describes the algorithm, and introduces all the necessary notions. Section 3 presents the
consistency and sample complexity results. Section 4 discusses run time ef?ciency, and in Section 5
we make some concluding remarks. Finally, the supplementary material provides the missing proofs,
the analysis of the special cases, and the necessary ?xes for the issues with the run-time complexity.
2
StOP: Stochastic Optimistic Planning
Recall that N ? N denotes the number of possible next states. That is, for each state x ? X and each
action u available at x, it holds that P (x, u, x? ) = 0 for all but at most N states x? ? X. Throughout
this section, the state of interest is denoted by x0 , the requested accuracy by ?, and the con?dence
parameter by ?0 . That is, the problem to be solved is to output an action u which is, with probability
at least (1 ? ?0 ), at least ?-optimal in x0 .
The algorithm and the analysis make use of the notion of an (in?nite) planning tree, policies and
trajectories. These notions are introduced in the next subsection.
2.1
Planning trees and trajectories
The in?nite planning tree ?? for a given MDP is a rooted and labeled in?nite tree. Its root is
denoted s0 and is labeled by the state of interest, x0 ? X. Nodes on even levels are called action
nodes (the root is an action node), and have Kd children each on the d-th level of action nodes: each
action u is represented by exactly one child, labeled u. Nodes on odd levels are called transition
nodes and have N children each: if the label of the parent (action) node is x, and the label of the
transition node itself is u, then for each x? ? X with P (x, u, x? ) > 0 there is a corresponding child,
labeled x? . There may be children with probability zero, but no duplicates.
An in?nite policy is a subtree of ?? with the same root, where each action node has exactly one
child and each transition node has N children. It corresponds to an agent having ?xed all its possible
future actions. A (partial) policy ? is a ?nite subtree of ?? , again with the same root, but where
the action nodes have at most one child, each transition node has N children, and all leaves 4 are
on the same level. The number of transition nodes on any path from the root to a leaf is denoted
d(?) and is called the depth of ?. A partial policy corresponds to the agent having its possible
future actions planned for d(?) steps. There is a natural partial order over these policies: a policy
4
Note that leaves are, by de?nition, always action nodes.
4
?? is called descendant policy of a policy ? if ? is a subtree of ?? . If, additionally, it holds that
d(?? ) = d(?) + 1, then ? is called the parent policy of ?? , and ?? the child policy of ?.
A (random) trajectory, or rollout, for some policy ? is a realization ? := (xt , ut , rt )Tt=0 of the
stochastic process that belongs to the policy. A random path is generated from the root by always
following, from a non-leaf action node with label xt , its unique child in ?, then setting ut to the
label of this node, from where, drawing ?rst a label xt+1 from P (xt , ut , ?), one follows the child
with label xt+1 . The reward rt is drawn from the distribution determined by R(xt , ut , ?). The value
?T
of the rollout ? (also called return or payoff in the literature) is v(? ) := t=0 rt ? t , and the value of
?T
the policy ? is v(?) := E[v(? )] = E[ t=0 rt ? t ]. For an action u available at x0 , denote by v(u)
the maximum of the values of the policies having u as the label of the child of root s0 . Denote by v ?
the maximum of these v(u) values. Using this notation, the task of the algorithm is to return, with
high probability, an action u with v(u) ? v ? ? ?.
2.2
The algorithm
StOP (Algorithm 1, see Figure 1 in the supplementary material for an illustration) maintains for each
action u available at x0 a set of active policies Active(u). Initially, it holds that Active(u) = {?u },
where ?u is the shallowest partial policy with the child of the root being labeled u. Also, for each
policy ? that becomes a member of an active set, the algorithm maintains high con?dence lower and
upper bounds for the value v(?) of the policy, denoted ?(?) and b(?), respectively.
In each round t, an optimistic policy ??t,u := argmax??Active(u) b(?) is determined for each action u. Based on this, the current optimistic action u?t := argmaxu b(??t,u ) and secondary action
?
u??
t := argmaxu?=u?t b(?t,u ) are computed. A policy ?t to explore is then chosen: if the policy
that belongs to the secondary action is at least as deeply developed as the policy that belongs to
the optimistic action, the optimistic one is chosen for exploration, otherwise the secondary one.
Note that a smaller depth is equivalent to a larger gap between lower and upper bound, and vice
versa5 . The set Active(ut ) is then updated by replacing the policy ?t by its child policies. Accordingly, the upper and lower bounds for these policies are computed. The algorithm terminates when
?(??t ) + ? ? maxu?=u? b(??t,u )?that is, when, with high con?dence, no policies starting with an
t
action different from u?t have the potential to have signi?cantly higher value.
2.2.1
Number and length of trajectories needed for one partial policy
Fix some integer d > 0 and let ? be a partial policy of depth d. Let, furthermore, ?? be an in?nite
policy that is a descendant of ?. Note that
0 ? v(?? ) ? v(?) ?
?d
1?? .
(2)
d
?
-accurate approximation of the value of ?? . On the other hand, having m
The value of ? is a 1??
trajectories for ?, their average reward v?(?) can be used as an estimate of the value v(?)?
of ?. From
d
ln(1/?)
the Hoeffding bound, this estimate has, with probability at least (1 ? ?), accuracy 1??
1??
2m .
?
d
d
ln(1/?)
?d
( 1??
)2 ? trajectories, 1??
? 1??
holds, so with probWith m := m(d, ?) := ? ln(1/?)
2
1??
2m
?d
?
d
d
ln(1/?)
?
?d
ability at least (1 ? ?), b(?) := v?(?) + 1??
+ 1??
? v?(?) + 2 1??
and ?(?) :=
1??
2m
?
ln(1/?)
1?? d
?d
? v?(?) ? 1?? bound v(?? ) from above and below, respectively. This choice
v?(?) ? 1??
2m
balances the inaccuracy of estimating v(?? ) based on v(?) and the inaccuracy of estimating v(?).
d?
?
6
)/ ln(1/?)?, the smallest integer satisfying 3 1??
? ?/2. Note
Let d? := d? (?, ?) := ?(ln (1??)?
?
that if d(?) = d for any given policy ?, then b(?) ? ?(?) ? ?/2. Because of this, it follows
(see Lemma 3 in the supplementary material) that d? is the maximal length the algorithm ever has
to develop a policy.
5
This approach of using secondary actions is based on the UGapE algorithm [10].
5
Algorithm 1 StOP(s0 , ?0 , ?, ?)
1: for all u available from x0 do
? initialize
2:
?u := smallest policy with the child of s0 labeled u
3:
?1 := (?0 /d? ) ? (K0 )?1
? d(?u ) = 1
4:
(?(?u ), b(?u )) := BoundValue(?u , ?1 )
5:
Active(u) := {?u }
? the set of active policies that follow u in s0
6: for round t=1, 2, . . . do
7:
for all u available at x0 do
8:
??t,u := argmax??Active(u) b(?)
9:
??t := ?? ? , where u?t := argmaxu b(??t,u ),
? optimistic action and policy
t,ut
?
???
t := ?
10:
if
11:
?(??t )
t,u??
t
?
, where u??
t := argmaxu?=u? b(?t,u ),
+? ?
u?t
t
maxu?=u? b(??t,u )
t
then
? secondary action and policy
? termination criterion
return
?
if d(???
? select the policy to evaluate
t ) ? d(?t ) then
?
ut := ut and ?t := ??t
else
??
ut := u??
? action and policy to explore
t and ?t := ?t
Active(ut ) := Active(ut ) \ {?t }
?d(? )?1
?d?1
?
?
? := (?0 /d? ) ? ?=0t (K? )?N
? ?=0 (K? )N = # of policies of depth at most d
?
for all child policy ? of ?t do
(?(?), b(?)) := BoundValue(?? , ?)
Active(ut ) := Active(ut ) ? {?? }
12:
13:
14:
15:
16:
17:
18:
19:
20:
21:
2.2.2
Samples and sample trees
Algorithm StOP aims to aggressively reuse every sample for each transition node and every sample
for each state-action pair, in order to keep the sample complexity as low as possible. Each time the
value of a partial policy is evaluated, all samples that are available for any part of it from previous
rounds are reused. That is, if m trajectories are necessary for assessing the value of some policy
?, and there are m? complete trajectories available and m?? that end at some inner node of ?, then
StOP (more precisely, another algorithm, Sample, called from StOP) samples rewards (using
SampleReward) and transitions (SampleTransition) to generate continuations for the m??
incomplete trajectories and to generate (m ? m? ? m?? ) new trajectories, as described in Section 2.1,
where
? SampleReward(s) for some action node s samples a reward from the distribution
R(x, u, ?), where u is the label of the parent of s and x is the label of the grandparent
of s, and
? SampleTransition(s) for some transition node s samples a next state from the distribution P (x, u, ?), where u is the label of s and x is the label of the parent of s.
To compensate for the sharing of the samples, the con?dences of the estimates are increased, so that
with probability at least (1 ? ?0 ), all of them are valid6 . The samples are organized as a collection of
sample trees, where a sample tree T is a (?nite) subtree of ?? with the property that each transition
node has exactly one child, and that each action node s is associated with some reward rT (s). Note
that the intersection of a policy ? and a sample tree T is always a path. Denote this path by ? (T , ?)
and note that it necessarily starts from the root and ends either in a leaf or in an internal node of ?. In
the former case, this path can be interpreted as a complete trajectory for ?, and in the latter case, as
an initial segment. Accordingly,
?m when the value of a new policy ? needs to be estimated/bounded, it
1
is computed as v?(?) := m
i=1 v(? (Ti , ?)) (see Algorithm 2: BoundValue), where T1 , . . . , Tm
are sample trees constructed by the algorithm. For terseness, these are considered to be global
variables, and are constructed and maintained using algorithm Sample (Algorithm 3).
?
?N ?
In particular, the con?dence is set to 1 ? ?d(?) for policy ?, where ?d = (?0 /d? ) d?1
is ?0
?=0 K?
divided by the number of policies of depth at most d, and by the largest possible depth?see section 2.2.1.
6
6
Algorithm 2 BoundValue(?, ?)
Ensure: with
at least
(1 ? ?), interval [?(?), b(?)] contains v(?)
? probability
?
?2 ?
ln(1/?) 1?? d(?)
1: m :=
2
? d(?)
2: Sample(?,?
s0 , m)
m
1
3: v?(?) := m
i=1 v(? (Ti , ?))
4: ?(?) := v?(?) ?
1?? d(?)
1??
5: b(?) := v?(?) +
?
1??
d(?)
+
?
6: return (?(?), b(?))
? Ensure that at least m trajectories exist for ?
? empirical estimate of v(?)
ln(1/?)
2m
d(?)
1??
1??
?
? Hoeffding bound
ln(1/?)
2m
? . . . and (2)
Algorithm 3 Sample(?, s, m)
Ensure: there are m sample trees T1 , . . . , Tm that contain a complete trajectory for ? (i.e. ? (Ti , ?)
ends in a leaf of ? for i = 1, . . . , m)
1: for i := 1, . . . , m do
2:
if sample tree Ti does not yet exist then
3:
let Ti be a new sample tree of depth 0
4:
let s be the last node of ? (Ti , ?)
? s is an action node
5:
while s is not a leaf of ? do
6:
let s? be the child of s in ? and add it to T as a new child of s
7:
s?? := SampleTransition(s? ),
? s? is a transition node
8:
add s?? to T as a new child of s?
9:
s := s??
10:
rT (s?? ) := SampleReward(s?? )
3
Analysis
Recall that v ? denotes the maximal value of any (possibly in?nite) policy tree. The following theorem formalizes the consistency result for StOP (see the proof in Section C).
Theorem 1. With probability at least (1 ? ?0 ), StOP returns an action with value at least v ? ? ?.
Before stating the sample complexity result, some further notation needs to be introduced.
Let u? denote an optimal action available at state x0 . That is, v(u? ) = v ? . De?ne for u ?= u?
?
?
d(?)
d(?)
Pu? := ? : ? follows u from s0 and v(?) + 3 ?1?? ? v ? ? 3 ?1?? + ? ,
and also de?ne
?
?
d(?)
d(?)
Pu? ? := ? : ? follows u? from s0 , v(?) + 3 ?1?? ? v ? and v(?) ? 6 ?1?? + ? ? max? v(u) .
?
u?=u
?
is the set of ?important? policies that potentially need to be evaluated
Then P :=
in order to determine an ?-optimal action. (See also Lemma 8 in the supplementary material.)
?
Pu? ?
?
u?=u? Pu
Let now p(s) denote the product of the probabilities of the transitions on the path from s0 to s. That
is, for any policy tree ? containing s, a trajectory for ? goes through s with probability p(s). When
estimating the value of some policy ? of depth d, the expected number of trajectories going through
some nodes s of it is p(s)m(d, ?d ). The sample complexity therefore has to take into consideration
for each node s (at least for the ones with ?high? p(s) value) the maximum ?(s) = max{d(?) : ? ?
P ? contains s} of the depth of the relevant policies it is included in. Therefore, the expected number
of trajectories going through s in a given run of StOP is
?
?
? ?
ln(1/??(s) ) 1?? ?(s) 2
(3)
p(s) ? m(?(s), ??(s) ) = p(s)
2
? ?(s)
If (3) is ?large? for some s, it can be used to deduce a high con?dence upper bound on the number of
times s gets sampled. To this end, let S ? denote the set of nodes of the trees in P ? , let N ? denote the
7
??
??
smallest positive integer N satisfying N ? ? s ? S ? : p(s) ? m(?(s), ??(s) ) ? (8/3) ln(2N /?0 ) ?
(obviously N ? ? |S ? |), and de?ne
?
?
S ?,? := s ? S ? : p(s) ? m(?(s), ??(s) ) ? (8/3) ln(2N ? /?0 ) .
S ? is the set of ?important? nodes (P ? is the set of ?important? policies), and S ?,? consists of the
important nodes that, with high probability, are not sampled more than twice as often as expected.
?0
(This high probability is 1 ? 2N
? according to the Bernstein bound, so these upper bounds hold
jointly with probability at least (1 ? ?20 ), as N ? = |S ?,? |. See also Appendix D.)
For s? ? S ? \ S ?,? , the number of times s? gets sampled has a variance that is too high compared
to its expected value (3), so in this case, a different approach is needed in order to derive
? high
con?dence upper bounds. To this end, for a transition node s, let p? (s) := p? (s, ?) := {p(s? ) :
s? is a child of s with p(s? ) ? m(?(s? ), ??(s? ) ) < (8/3) ln(2N ? /?0 )}, and de?ne
?
?
0,
if p? (s) ? 2N ? m(?(s),?
?(s) )
B(s) := B(s, ?) :=
?
?
max(6 ln( 2N
),
2p
(s)m(?(s),
?
))
otherwise.
?(s)
?0
As it will be shown in the proof of Theorem 2 (in Section D), this is a high con?dence upper bound
on the number of trajectories that go through some child s? ? S ? \ S ?,? of some s? ? S ?,? .
Theorem 2. With probability at ?
least (1 ? 2?), StOP outputs a policy of value at least (v?? ? ?), af?
?d
??(s)
ter generating at most s?S ?,? 2p(s)m(?(s), ??(s) ) + B(s) d=d(s)+1 ?=d(s)+1 K? samples,
where d(s) = min{d(?) : s appears in policy ?} is the depth of node s.
Finally, the bound discussed in Section 1 is obtained by setting ? := lim sup??0 max(?1 , ?2 ),
??
?1/d?
?2 (1??)2
2p(s)m(?(s),
?
)
and ?2 := ?2 (?, ?0 , ?) :=
where ?1 := ?1 (?, ?0 , ?) :=
?(s)
s?S ?,? ln(1/?0 )
?
?
? 2
1/d
??(s) ?d
? (1??)2 ?
.
s?S ?,? B(s)
?=d(s) K?
d=d(s)
ln(1/?0 )
4
Ef?ciency
StOP, as presented in Algorithm 1, is not ef?ciently executable. First of all, whenever it evaluates
an optimistic policy, it enumerates all its child policies, which typically has exponential time complexity. Besides that, the sample trees are also treated in an inef?cient way. An ef?cient version of
StOP with all these issues ?xed is presented in Appendix F of the supplementary material.
5
Concluding remarks
In this work, we have presented and analyzed our algorithm, StOP. To the best of our knowledge,
StOP is currently the only algorithm for optimal (i.e. closed loop) online planning with a generative
model that provably bene?ts from local structure both in reward as well as in transition probabilities.
It assumes no knowledge about this structure other than access to the generative model, and does
not impose any restrictions on the system dynamics.
One should note though that the current version of StOP does not support domains with in?nite
N . The sparse sampling algorithm in [14] can easily handle such problems (at the cost of a nonpolynomial (in 1/?) sample complexity), however, StOP has much better sample complexity in case
of ?nite N . An interesting problem for future research is to design adaptive planning algorithms
with sample complexity independent of N ([21] presents such an algorithm, but the complexity
bound provided there is the same as the one in [14]).
Acknowledgments
This work was supported by the French Ministry of Higher Education and Research, and by the
European Community?s Seventh Framework Programme (FP7/2007-2013) under grant agreement
no 270327 (project CompLACS). Author two would like to acknowledge the support of the BMBF
project ALICE (01IB10003B).
8
References
[1] P. Auer, N. Cesa-Bianchi, and P. Fischer. Finite-time analysis of the multiarmed bandit problem. Machine
Learning Journal, 47(2-3):235?256, 2002.
[2] Dimitri P. Bertsekas. Dynamic Programming and Optimal Control. Athena Scienti?c, 2001.
[3] S. Bubeck and R. Munos. Open loop optimistic planning. In Conference on Learning Theory, 2010.
[4] S?ebastien Bubeck and Nicol`o Cesa-Bianchi. Regret analysis of stochastic and nonstochastic multi-armed
bandit problems. Foundations and Trends in Machine Learning, 5(1):1?122, 2012.
[5] Lucian Bus?oniu and R?emi Munos. Optimistic planning for markov decision processes. In Proceedings
15th International Conference on Arti?cial Intelligence and Statistics (AISTATS-12), pages 182?189,
2012.
[6] E. F. Camacho and C. Bordons. Model Predictive Control. Springer-Verlag, 2004.
[7] Nicolo Cesa-Bianchi and Gabor Lugosi. Prediction, Learning, and Games. Cambridge University Press,
New York, NY, USA, 2006.
[8] R?emi Coulom. Ef?cient selectivity and backup operators in Monte-Carlo tree search. In Proceedings
Computers and Games 2006. Springer-Verlag, 2006.
[9] E. Even-Dar, S. Mannor, and Y. Mansour. Action elimination and stopping conditions for reinforcement
learning. In T. Fawcett and N. Mishra, editors, Proceedings of the Twentieth International Conference on
Machine Learning (ICML-2003), pages 162?169, 2003.
[10] Victor Gabillon, Mohammad Ghavamzadeh, and Alessandro Lazaric. Best arm identi?cation: A uni?ed
approach to ?xed budget and ?xed con?dence. In Peter L. Bartlett, Fernando C. N. Pereira, Christopher
J. C. Burges, Lon Bottou, and Kilian Q. Weinberger, editors, NIPS, pages 3221?3229, 2012.
[11] Sylvain Gelly, Yizao Wang, R?emi Munos, and Olivier Teytaud. Modi?cation of UCT with Patterns in
Monte-Carlo Go. Rapport de recherche RR-6062, INRIA, 2006.
[12] Eric A. Hansen and Shlomo Zilberstein. A heuristic search algorithm for Markov decision problems. In
Proceedings Bar-Ilan Symposium on the Foundation of Arti?cial Intelligence, Ramat Gan, Israel, 23?25
June 1999.
[13] J-F. Hren and R. Munos. Optimistic planning of deterministic systems. In Recent Advances in Reinforcement Learning, pages 151?164. Springer LNAI 5323, European Workshop on Reinforcement Learning,
2008.
[14] M. Kearns, Y. Mansour, and A.Y. Ng. A sparse sampling algorithm for near-optimal planning in large
Markovian decision processes. In Machine Learning, volume 49, pages 193?208, 2002.
[15] Gunnar Kedenburg, Raphael Fonteneau, and Remi Munos. Aggregating optimistic planning trees for
solving markov decision processes. In C.J.C. Burges, L. Bottou, M. Welling, Z. Ghahramani, and K.Q.
Weinberger, editors, Advances in Neural Information Processing Systems 26, pages 2382?2390. Curran
Associates, Inc., 2013.
[16] Levente Kocsis and Csaba Szepesv?ari. Bandit based monte-carlo planning. In In: ECML-06. Number
4212 in LNCS, pages 282?293. Springer, 2006.
[17] T. L. Lai and H. Robbins. Asymptotically ef?cient adaptive allocation rules. Advances in Applied Mathematics, 6:4?22, 1985.
[18] R?emi Munos. From bandits to Monte-Carlo Tree Search: The optimistic principle applied to optimization
and planning. Foundation and Trends in Machine Learning, 7(1):1?129, 2014.
[19] N.J. Nilsson. Principles of Arti?cial Intelligence. Tioga Publishing, 1980.
[20] M.L. Puterman. Markov Decision Processes ? Discrete Stochastic Dynamic Programming. John Wiley
& Sons, Inc., New York, NY, 1994.
[21] Thomas J. Walsh, Sergiu Goschin, and Michael L. Littman. Integrating sample-based planning and
model-based reinforcement learning. In Proceedings of the Twenty-Fourth AAAI Conference on Arti?cial
Intelligence, pages 612?617. AAAI Press, 2010.
9
| 5368 |@word version:2 polynomial:3 proportion:2 nd:1 reused:1 open:3 termination:1 simulation:1 arti:5 initial:5 contains:3 ours:1 mishra:1 current:4 comparing:1 yet:1 john:1 shlomo:1 cant:1 camacho:1 generative:12 intelligence:5 leaf:7 accordingly:2 beginning:1 recherche:1 provides:2 mannor:1 node:41 contribute:3 teytaud:1 rollout:2 along:2 constructed:2 symposium:1 descendant:2 consists:1 x0:8 hardness:1 indeed:2 expected:5 planning:32 multi:3 bellman:1 kedenburg:3 discounted:2 armed:3 becomes:1 provided:3 project:5 bounded:3 notation:2 maximizes:1 estimating:3 probability3:1 israel:1 xed:4 interpreted:1 deepmind:1 developed:1 csaba:1 formalizes:1 cial:5 every:2 ti:6 exactly:3 control:4 grant:1 bertsekas:1 t1:2 negligible:2 before:1 local:3 positive:1 aggregating:1 limit:1 consequence:1 path:6 approximately:1 lugosi:1 inria:7 twice:1 alice:1 ramat:1 walsh:1 unique:1 acknowledgment:1 regret:1 implement:1 nite:11 lncs:1 empirical:3 gabor:1 lucian:1 integrating:1 get:3 close:4 operator:2 restriction:2 equivalent:1 deterministic:8 missing:1 go:3 fonteneau:1 starting:3 independently:1 survey:1 pure:1 rule:1 regarded:1 handle:1 notion:3 updated:1 play:1 programming:4 olivier:1 us:2 curran:1 agreement:1 associate:1 trend:2 expensive:1 satisfying:2 labeled:6 role:1 solved:1 wang:1 worst:3 kilian:1 deeply:1 mentioned:2 intuition:1 substantial:1 ugape:1 complexity:34 alessandro:1 reward:28 littman:1 dynamic:7 ghavamzadeh:1 depend:1 tight:1 segment:1 solving:1 predictive:2 eric:1 easily:1 k0:2 represented:1 derivation:1 enyi:1 monte:8 whose:1 heuristic:2 supplementary:5 larger:1 drawing:1 otherwise:2 ability:2 statistic:1 fischer:1 jointly:1 itself:1 online:5 obviously:1 kocsis:1 sequence:3 rr:1 maximal:2 product:1 fr:3 raphael:1 relevant:1 loop:5 hungary:1 realization:1 inef:1 tioga:1 achieve:2 az:1 rst:4 parent:4 assessing:1 generating:2 derive:1 develop:1 stating:1 odd:1 op:7 received:1 strong:2 c:1 signi:2 quantify:1 correct:1 stochastic:6 exploration:2 elimination:2 material:5 education:1 require:1 ao:1 fix:1 hold:8 considered:4 exp:1 maxu:2 driving:1 tor:1 achieves:2 smallest:4 estimation:1 yizao:1 applicable:1 label:11 currently:1 hansen:1 robbins:1 largest:1 vice:1 shallowest:1 always:3 aim:2 super:1 rather:1 zilberstein:1 goschin:1 lon:1 june:1 sense:2 stopping:1 typically:1 lnai:1 initially:1 bandit:6 going:2 france:3 provably:1 issue:3 denoted:4 exponent:1 special:2 initialize:1 having:6 ng:1 sampling:5 eliminated:1 chernoff:1 lille:3 look:3 icml:1 hren:1 future:3 others:1 duplicate:1 modi:1 simultaneously:1 individual:1 argmax:2 interest:2 highly:1 introduces:1 analyzed:1 behind:1 scienti:1 accurate:1 neglecting:3 necessary:3 partial:7 respective:1 tree:26 incomplete:1 logarithm:1 desired:1 theoretical:1 increased:1 earlier:1 markovian:1 planned:1 disadvantage:1 cost:1 uniform:8 seventh:1 too:2 dependency:1 kn:6 chooses:1 international:2 sequel:3 cantly:1 complacs:1 michael:1 gabillon:1 again:2 aaai:2 cesa:3 containing:1 choose:1 possibly:1 hoeffding:3 argmaxu:4 leading:2 return:8 dimitri:1 potential:2 ilan:1 de:8 rapport:1 inc:2 depends:5 later:1 root:9 closed:2 optimistic:14 observing:1 characterizes:1 sup:1 reached:2 recover:2 xing:1 maintains:2 start:1 competitive:1 contribution:3 accuracy:2 variance:1 carlo:8 trajectory:18 cation:2 sharing:1 ed:2 whenever:1 evaluates:1 involved:1 proof:3 associated:1 con:9 oniu:1 stop:28 sampled:3 recall:2 knowledge:4 subsection:1 ut:13 lim:1 organized:1 enumerates:1 auer:1 back:1 appears:2 higher:2 follow:2 formulation:1 evaluated:2 though:3 furthermore:1 uct:1 hand:1 receives:1 replacing:1 christopher:1 google:1 french:1 mdp:20 usa:1 contain:1 true:1 former:1 aggressively:1 illustrated:2 puterman:1 round:3 during:1 branching:5 game:2 rooted:1 maintained:1 bal:1 criterion:1 stress:1 complete:4 tt:1 mohammad:1 consideration:1 ef:6 ari:1 executable:1 exponentially:1 volume:1 belong:2 discussed:1 refer:1 nonpolynomial:1 multiarmed:1 cambridge:1 ai:1 consistency:2 mathematics:1 reachable:1 access:4 europe:3 deduce:1 add:2 navigates:1 pu:4 nicolo:1 closest:1 recent:1 belongs:3 selectivity:1 balazs:1 verlag:2 nition:1 victor:1 ministry:1 additional:1 impose:1 speci:3 determine:1 fernando:1 branch:3 full:2 adapt:2 af:2 compensate:1 divided:1 lai:1 prediction:1 variant:1 essentially:1 expectation:1 fawcett:1 achieved:2 whereas:3 want:1 addition:1 szepesv:1 interval:1 else:1 rest:1 posse:1 probably:1 szorenyi:1 member:1 call:2 integer:3 ciently:1 near:10 ter:1 bernstein:1 nonstochastic:1 inner:1 idea:1 tm:2 translates:1 optimism:2 bartlett:1 reuse:1 peter:1 york:2 action:51 remark:2 dar:1 detailed:1 discount:1 extensively:1 generate:3 continuation:1 exist:2 notice:1 estimated:1 lazaric:1 discrete:1 group:1 gunnar:3 achieving:1 drawn:1 levente:1 breadth:1 asymptotically:1 run:3 uncertainty:2 fourth:1 almost:2 throughout:1 decision:9 appendix:2 sergiu:1 bound:31 distinguish:1 ahead:3 precisely:2 szte:1 software:1 dences:1 dence:8 emi:4 min:1 concluding:2 performing:1 expanded:1 relatively:1 ned:2 structured:1 mta:1 according:1 kd:1 describes:1 smaller:1 terminates:1 son:1 making:1 happens:1 nilsson:1 restricted:1 ln:17 bus:1 turn:1 discus:1 needed:2 fp7:1 end:5 available:9 weinberger:2 thomas:1 denotes:2 assumes:1 ensure:3 gan:1 publishing:1 exploit:1 gelly:1 k1:1 build:3 ghahramani:1 quantity:4 strategy:2 rt:8 dependence:3 grandparent:1 simulated:1 athena:1 besides:2 length:3 illustration:1 coulom:1 balance:1 potentially:2 nord:3 design:2 ebastien:1 policy:81 twenty:1 perform:1 bianchi:3 upper:8 markov:8 acknowledge:1 finite:1 t:2 ecml:1 immediate:1 situation:3 payoff:1 ever:1 mansour:2 arbitrary:1 community:2 prioritizing:1 introduced:4 pair:1 required:1 bene:4 identi:2 inaccuracy:2 nip:1 address:1 able:1 bar:1 below:3 pattern:1 built:2 max:5 natural:1 force:1 treated:1 arm:1 minimax:3 scheme:1 improve:1 mdps:2 ne:4 deviate:1 literature:3 understanding:1 nicol:1 fully:1 loss:1 interesting:1 allocation:1 foundation:3 agent:7 s0:9 principle:5 editor:3 surprisingly:1 last:1 soon:1 supported:1 enjoys:1 burges:2 wide:1 face:2 munos:8 sparse:4 depth:13 transition:24 cumulative:1 avoids:1 evaluating:1 author:1 collection:1 adaptive:2 reinforcement:4 avoided:1 programme:1 far:1 welling:1 contradict:1 emphasize:1 uni:1 keep:1 sz:1 global:4 active:13 rid:1 unnecessary:1 don:1 search:9 additionally:1 promising:1 requested:1 expansion:1 investigated:1 poly:1 necessarily:1 european:2 domain:2 bottou:2 aistats:1 main:2 whole:1 backup:1 child:24 referred:1 cient:4 ny:2 wiley:1 bmbf:1 precision:1 sub:2 pereira:1 ciency:2 exponential:1 theorem:4 uation:1 xt:6 pac:3 explored:2 x:1 workshop:1 subtree:4 budget:1 gap:2 intersection:1 logarithmic:2 remi:3 explore:2 bubeck:2 twentieth:1 applies:1 springer:4 corresponds:4 included:1 determined:3 sylvain:1 uniformly:2 averaging:1 lemma:2 kearns:1 called:10 secondary:5 select:1 internal:1 support:2 latter:1 evaluate:1 |
4,825 | 5,369 | Conditional Swap Regret and
Conditional Correlated Equilibrium
Mehryar Mohri
Courant Institute and Google
251 Mercer Street
New York, NY 10012
Scott Yang
Courant Institute
251 Mercer Street
New York, NY 10012
mohri@cims?nyu?edu
yangs@cims?nyu?edu
Abstract
We introduce a natural extension of the notion of swap regret, conditional swap
regret, that allows for action modi?cations conditioned on the player?s action history. We prove a series of new results for conditional swap regret minimization.
We present algorithms for minimizing conditional swap regret with bounded conditioning history. We further extend these results to the case where conditional
swaps are considered only for a subset of actions. We also de?ne a new notion
of equilibrium, conditional correlated equilibrium, that is tightly connected to the
notion of conditional swap regret: when all players follow conditional swap regret
minimization strategies, then the empirical distribution approaches this equilibrium. Finally, we extend our results to the multi-armed bandit scenario.
1
Introduction
On-line learning has received much attention in recent years. In contrast to the standard batch
framework, the online learning scenario requires no distributional assumption. It can be described
in terms of sequential prediction with expert advice [13] or formulated as a repeated two-player
game between a player (the algorithm) and an opponent with an unknown strategy [7]: at each time
step, the algorithm probabilistically selects an action, the opponent chooses the losses assigned to
each action, and the algorithm incurs the loss corresponding to the action it selected.
The standard measure of the quality of an online algorithm is its regret, which is the difference
between the cumulative loss it incurs after some number of rounds and that of an alternative policy.
The cumulative loss can be compared to that of the single best action in retrospect [13] (external
regret), to the loss incurred by changing every occurrence of a speci?c action to another [9] (internal
regret), or, more generally, to the loss of action sequences obtained by mapping each action to some
other action [4] (swap regret). Swap regret, in particular, accounts for situations where the algorithm
could have reduced its loss by swapping every instance of one action with another (e.g. every time
the player bought Microsoft, he should have bought IBM).
There are many algorithms for minimizing external regret [7], such as, for example, the randomized
weighted-majority algorithm of [13]. It was also shown in [4] and [15] that there exist algorithms for
minimizing internal and swap regret. These regret minimization techniques have been shown to be
useful for approximating game-theoretic equilibria: external regret algorithms for Nash equilibria
and swap regret algorithms for correlated equilibria [14].
By de?nition, swap regret compares a player?s action sequence against all possible modi?cations at
each round, independently of the previous time steps. In this paper, we introduce a natural extension
of swap regret, conditional swap regret, that allows for action modi?cations conditioned on the
player?s action history. Our de?nition depends on the number of past time steps we condition upon.
1
As a motivating example, let us limit this history to just the previous one time step, and suppose we
design an online algorithm for the purpose of investing, where one of our actions is to buy bonds
and another to buy stocks. Since bond and stock prices are known to be negatively correlated, we
should always be wary of buying one immediately after the other ? unless our objective was to pay
for transaction costs without actually modifying our portfolio? However, this does not mean that we
should avoid purchasing one or both of the two assets completely, which would be the only available
alternative in the swap regret scenario. The conditional swap class we introduce provides precisely
a way to account for such correlations between actions. We start by introducing the learning set-up
and the key notions relevant to our analysis (Section 2).
2
Learning set?up and model
We consider the standard online learning set-up with a set of actions N = {1? . . . ? N }. At each
round t ? {1? . . . ? T }, T ? 1, the player selects an action xt ? N according to a distribution pt
over N , in response to which the adversary chooses a function f t : N t ? [0? 1] and causes the
player to incur a loss f t ?xt ? xt?1 ? . . . ? x1 ). The objective of the player is to choose a sequence of
?T
actions ?x1 ? . . . ? xT ) that minimizes his cumulative loss t=1 f t ?xt ? xt?1 ? . . . ? x1 ).
A standard metric used to measure the performance of an online algorithm ? over T rounds is its
?expected) external regret, which measures the player?s expected performance against the best ?xed
action in hindsight:
Reg??? T ) =
Ext
T
?
t=1
?
?xt ?..?x1 )?
?pt ?...?p1 )
[f t ?xt ? ..? x1 )] ? min
j?N
T
?
f t ?j? j? ...? j).
t=1
There are several common modi?cations to the above online learning scenario: (1) we may com?T
t
pare regret against stronger competitor classes: Reg? ??? T ) =
t=1 ?pt ?...?p1 f ?xt ? ..? x1 ) ?
?T
t
min??? t=1 ?pt ?...?p1 [f ???xt )? ??xt?1 )? ...? ??x1 ))] for some function class C ? N N ; (2)
the player may have access to only partial information about the loss, i.e. only knowledge
of f t ?xt ? ..? x1 ) as opposed to f t ?a? xt?1 ? . . . ? x1 )?a ? N (also known as the bandit scenario); (3) the loss function may have bounded memory: f t ?xt ? ...? xt?k ? xt?k?1 ? ...? x1 ) =
f t ?xt ? ...? xt?k ? yt?k?1 ? ...? y1 ), ?xj ? yj ? N .
The scenario where C = N N in (1) is called the swap regret case, and the case where k = 0 in (3) is
referred to as the oblivious adversary. (Sublinear) regret minimization is possible for loss functions
against any competitor class of the form described in (1), with only partial information, and with
at least some level of bounded memory. See [4] and [1] for a reference on (1), [2] and [5] for (2),
and [1] for (3). [6] also provides a detailed summary of the best known regret bounds in all of these
scenarios and more.
The introduction of adversaries with bounded memory naturally leads to an interesting question:
what if we also try to increase the power of the competitor class in this way?
While swap regret is a natural competitor class and has many useful game theoretic consequences
(see [14]), one important missing ingredient is that the competitor class of functions does not have
memory. In fact, in most if not all online learning scenarios and regret minimization algorithms
considered so far, the point of comparison has been against modi?cation of the player?s actions
at each point of time independently of the previous actions. But, as we discussed above in the
?nancial markets example, there exist cases where a player should be measured against alternatives
that depend on the past and the player should take into account the correlations between actions.
Speci?cally, we consider competitor functions of the form ?t : N t ? N t . Let Call = {?t : N t ?
?T
t
N t }?
t=1 denote the class of all such functions. This leads us to the expression:
t=1 ?p1 ?...?pt [f ] ?
?T
t
t
min?t ??all t=1 ?p1 ?...?pt [f ? ? ]. Call is clearly a substantially richer class of competitor functions than traditional swap regret. In fact, it is the most comprehensive class, since we can always
?T
?T
reach t=1 ?p1 ?...?pt [f t ] ? t=1 min?x1 ?..?xt ) f t ?x1 ? ..? xt ) by choosing ?t to map all points to
t
argmin?xt ?..?x1 ) f ?xt ? ...? x1 ). Not surprisingly, however, it is not possible to obtain a sublinear
regret bound against this general class.
2
???
????????
????????
??????
??????
??????
?
???
????????
?
?
????????
????????
???
????????
???
????????
????????
????????
?
?
(a)
(b)
Figure 1: (a) unigram conditional swap class interpreted as a ?nite-state transducer. This is the same
as the usual swap class and has only the trivial state; (b) bigram conditional swap class interpreted as
a ?nite-state transducer. The action at time t ? 1 de?nes the current state and in?uences the potential
swap at time t.
Theorem 1. No algorithm can achieve sublinear regret against the class Call , regardless of the loss
function?s memory.
This result is well-known in the on-line learning community, but, for completeness, we include a
proof in Appendix 9. Theorem 1 suggests examining more reasonable subclasses of Call . To simplify
the notation and proofs that follow in the paper, we will henceforth restrict ourselves to the scenario
of an oblivious adversary, as in the original study of swap regret [4]. However, an application of the
batching technique of [1] should produce analogous results in the non-oblivious case for all of the
theorems that we provide.
Now consider the collection of competitor functions Ck = {? : N k ? N }. Then, a player
who has played actions {as }t?1
s=1 in the past should have his performance compared against
??at ? at?1 ? at?2 ? . . . ? at??k?1) ) at time t, where ? ? Ck . We call this class Ck of functions the
k-gram conditional swap regret class, which also leads us to the regret de?nition:
Reg??? T ) =
?k
T
?
t=1
? t [f t ?xt )] ? min
xt ?p
???k
T
?
t=1
? [f t ???xt ? at?1 ? at?2 ? . . . ? at??k?1) ))].
xt ?pt
Note that this is a direct extension of swap regret to the scenario where we allow for swaps conditioned on the history of the previous ?k ? 1) actions. For k = 1, this precisely coincides with swap
regret.
One important remark about the k-gram conditional swap regret is that it is a random quantity that
depends on the particular sequence of actions played. A natural deterministic alternative would be
of the form:
T
?
t=1
? t [f t ?xt )] ? min
xt ?p
???k
T
?
t=1
?
?xt ?...?x1 )??pt ?...?p1 )
[f t ???xt ? xt?1 ? xt?2 ? . . . ? xt??k?1) ))].
However, by taking the expectation of Reg?k ??? T ) with respect to aT ?1 ? aT2 ? . . . ? a1 and applying
Jensen?s inequality, we obtain
T
T
?
?
Reg??? T )?
? t [f t ?xt )]? min
?k
t=1
xt ?p
???k
t=1
?
?xt ?...?x1 )??pt ?...?p1 )
[f t ???xt ? xt?1 ? xt?2 ? . . . ? xt??k?1) ))]?
and so no generality is lost by considering the randomized sequence of actions in our regret term.
Another interpretation of the bigram conditional swap class is in the context of ?nite-state transducers. Taking a player?s sequence of actions ?x1 ? ...? xT ), we may view each competitor function in
the conditional swap class as an application of a ?nite-state transducer with N states, as illustrated
by Figure 1. Each state encodes the history of actions ?xt?1 ? . . . ? xt??k?1) ) and admits N outgoing
transitions representing the next action along with its possible modi?cation. In this framework, the
original swap regret class is simply a transducer with a single state.
3
3
Full Information Scenario
Here, we prove that it is in fact possible to minimize k-gram conditional swap regret against an
oblivious adversary, starting with the easier to interpret bigram scenario. Our proof constructs a
meta-algorithm using external regret algorithms as subroutines, as in [4]. The key is to attribute
a fraction of the loss to each external regret algorithm, so that these losses sum up to our actual
realized loss and also press the subroutines to minimize regret against each of the conditional swaps.
Theorem 2. There? exists
algorithm ? with bigram swap regret bounded as follows:
?
? an online
Reg?2 ??? T ) ? O N T log N .
Proof. Since the distribution pt at round t is ?nite-dimensional, we can represent it as a vector
pt = ?pt1 ? ...? ptN ). Similarly, since oblivious adversaries take only N arguments, we can write f t
t
as the loss vector f t = ?f1t ? ...? fN
). Let {at }Tt=1 be a sequence of random variables denoting the
player?s actions at each time t, and let ?at t denote the (random) Dirac delta distribution concentrated
at at and applied to variable xt . Then, we can rewrite the bigram swap regret as follows:
Reg??? T ) =
?2
T
?
t=1
=
?t [f t ?xt )] ? min
???2
p
T ?
N
?
t=1 i=1
pti fit ? min
???2
T
?
?
t t?1
t=1 p ??at?1
N
T ?
?
[f t ???xt ? xt?1 )]
t?1
pti ?{a
ft
t?1 =j} ??i?j)
t=1 i?j=1
Our algorithm for achieving sublinear regret is de?ned as follows:
1. At t = 1, initialize N 2 external regret minimizing algorithms Ai?k , ?i? k) ? N 2 .
We can view these in the form of N matrices in RN ?N , {Qt?k }N
k=1 , where for each
is
a
row
vector
consisting
of
the
distribution
weights generated
k ? {1? . . . ? N }, Qt?k
i
by algorithm Ai?k at time t based on losses received at times 1? . . . ? t ? 1.
2. At each time t, let at?1 denote the random action played at time t ? 1 and let ?at?1
denote
t?1
the (random) Dirac delta distribution for this action. De?ne the N ? N matrix Qt =
?N
t?1
t?k
t
k=1 ?{at?1 =k} Q . Q is a Markov chain (i.e., its rows sum up to one), so it admits a
t
stationary distribution p which we we will use as our distribution for time t.
3. When we draw from pt , we play a random action at and receive loss f t . Attribute the
t?1
portion of loss pti ?{a
f t to algorithm Ai?k , and generate distributions Qt?k
i . Notice
t?1 =k}
?N
t t?1
t
t
that i?k=1 pi ?{at?1 =k} f = f , so that the actual realized loss is allocated completely.
Recall that an optimal external regret minimizing algorithm ? (e.g.
majority)
??randomized weighted
?
i?k
i?k
admits a regret bound of the form Ri?k = Ri?k ?Lmin ? T? N ) = O
Lmin log?N ) , where Li?k
min =
?
T
t?i?k
minN
for the sequence of loss vectors {f t?i?k }Tt=1 incurred by the algorithm. Since
j=1
t=1 fj
t
t t
p = p Q is a stationary distribution, we can write:
T
?
t=1
pt ? f t =
N
T ?
?
t=1 j=1
ptj fjt =
N ?
N
T ?
?
pti Qti?j fjt =
t=1 j=1 i=1
N ?
N
T ?
?
t=1 j=1 i=1
4
pti
N
?
k=1
t?1
t
?{i
Qt?k
i?j fj .
t?1 =k}
Rearranging leads to
T
?
t=1
pt ? f t =
?
=
T ?
N ?
N
?
t?1
t
pti ?{i
Qt?k
i?j fj
t?1 =k}
i?k=1 t=1 j=1
N
?
i?k=1
N
?
i?k=1
?? T
?
t?1
pti ?{i
ft
t?1 =k} ??i?k)
t=1
? T
?
t?1
pti ?{i
ft
t?1 =k} ??i?k)
t=1
?
?
?2
T
?
t=1
pt ? f t ? min
???2
+ Ri?k ?Lmin ? T? N )
+
N
?
(for arbitrary ? : N 2 ? N )
Ri?k ?Lmin ? T? N ).
i?k=1
Since ? is arbitrary, we obtain
Reg??? T ) =
?
T ?
N
?
t=1 i?k=1
t?1
pti ?{i
ft
?
t?1 =k} ??i?k)
N
?
Ri?k ?Lmin ? T? N ).
i?k=1
?
log?N ) and that we scaled the losses to algorithm Ai?k by
Using the fact that Ri?k = O
?N ?N
t t?1
pi ?{it?1 =k} , the following inequality holds: k=1 j=1 Lk?j
min ? T . By Jensen?s inequality, this
implies
?
?
?
N
N N ?
N ?
? 1 ?
1 ??
T
k?j
k?j
?
Lmin ?
Lmin ?
?
2
N2
N
N
k=1 j=1
k=1 j=1
?
?N ?N ?
or, equivalently, k=1 j=1 Lk?j
min ? N T . Combining this with our regret bound yields
??
?
N
N
?
? ?
?
?
Li?k
Ri?k ?Lmin ? T? N ) =
O
? O N T log N ?
Reg??? T ) ?
min log N
?2
i?k=1
??
Li?k
min
i?k=1
which concludes the proof.
Remark 1. The computational complexity of a standard external regret minimization algorithm such
as randomized weighted majority per round is in O?N ) ?update the distribution on each of the N
actions multiplicatively and then renormalize), which implies that updating the N 2 subroutines will
cost O?N 3 ) per round. Allocating losses to these subroutines and combining the distributions that
they return will cost an additional O?N 3 ) time. Finding the stationary distribution of a stochastic
3
matrix can be done
?via matrix inversion in O?N )3time. Thus, the total computational complexity
of achieving O?N T log?N )) regret is only O?N T ). We remark that in practice, one often uses
iterative methods to compute dominant eigenvalues ?see [16] for a standard reference and [11] for
recent improvements). [10] has also studied techniques to avoid computing the exact stationary
distribution at every iteration step for similar types of problems.
The meta-algorithm above can be interpreted in three equivalent ways: (1) the player draws an
action xt from distribution pt at time t; (2) the player uses distribution pt to choose among the
N subsets of algorithms Qt1 ? ...? QtN , picking one subset Qtj ; next, after drawing j from pt , the
t?N
t?1
to randomly choose among the algorithms Qt?1
player uses ?{a
j ? ...? Qj , picking algorithm
t?1 =k}
t?a
t?a
Qj t?1 ; after locating this algorithm, the player uses the distribution from algorithm Qj t?1 to draw
t t?1
an action; (3) the player chooses algorithm Qt?k
j with probability pj ?{at?1 =k} and draws an action
from its distribution.
The following more general bound can be given for an arbitrary k-gram swap scenario.
Theorem 3. There??
exists an online
? algorithm ? with k-gram swap regret bounded as follows:
Reg?k ??? T ) ? O N k T log N .
The algorithm used to derive this result is a straightforward extension of the algorithm provided in
the bigram scenario, and the proof is given in Appendix 11.
Remark 2. The computational complexity of achieving the above regret bound is O?N k+1 T ).
5
????????
????????
???
?
?
????????
???
????????
???????? ????????
???
???
?
Figure 2: bigram conditional swap class restricted to a ?nite number of active states. When the
action at time t ? 1 is 1 or 2, the transducer is in the same state, and the swap function is the same.
4
State?Dependent Bounds
In some situations, it may not be relevant to consider conditional swaps for every possible action,
either because of the speci?c problem at hand or simply for the sake of computational ef?ciency.
Thus, for any S ? N 2 , we de?ne the following competitor class of functions:
? for ?i? k) ? S where ?? : N ? N }.
C2?S = {? : N 2 ? N |??i? k) = ??i)
See Figure 2 for a transducer interpretation of this scenario.
We will now show that the algorithm above can be easily modi?ed to derive a tighter bound that
is dependent on the number of states in our competitor class. We will focus on the bigram case,
although a similar result can be shown for the general k-gram conditional swap regret.
?
Theorem 4. There exists an online algorithm ? such that Reg?2?? ??? T )
?
c
O? T ?|S | + N ) log?N )).
The proof of this result is given in Appendix 10. Note that when S = ?, we are in the scenario where
all the previous states matter, and our bound coincides with that of the previous section.
Remark 3. The computational complexity of achieving the above regret bound is O??N ?|?1 ?S)| +
|S c |) + N 3 )T ), where ?1 is projection onto the ?rst component. This follows from the fact that
we allocate the same loss to all {Ai?k }k:?i?k)?S ?i ? ?1 ?S), so we effectively only have to manage
|?1 ?S)| + |S c | subroutines.
5
Conditional Correlated Equilibrium and ??Dominated Actions
It is well-known that regret minimization in on-line learning is related to game-theoretic equilibria
[14]. Speci?cally, when both players in a two-player zero-sum game follow external regret minimizing strategies, then the product of their individual empirical distributions converges to a Nash
equilibrium. Moreover, if all players in a general K-player game follow swap regret minimizing
strategies, then their empirical joint distribution converges to a correlated equilibrium [7].
We will show in this section that when all players follow conditional swap regret minimization
strategies, then the empirical joint distribution will converge to a new stricter type of correlated
equilibrium.
?k)
:S ?
De?nition 1. Let Nk = {1? ...? Nk }, for k ? {1? ...? K} and G = ?S = ?K
k=1 Nk ? {l
K
[0? 1]}k=1 ) denote a K-player game. Let s = ?s1 ? ...? sK ) ? S denote the strategies of all players in
one instance of the game, and let s??k) denote the ?K ? 1)-vector of strategies played by all players
aside from player k. A joint distribution P on two rounds of this game is a conditional correlated
equilibrium if for any player k, actions j? j ? ? Nk , and map ?k : Nk2 ? Nk , we have
?
?
?
P ?s? r) l?k) ?sk ? s??k) ) ? l?k) ??k ?sk ? rk )? s??k) ) ? 0.
?s?r)?S 2 : sk =j?rk =j ?
The standard interpretation of correlated equilibrium, which was ?rst introduced by Aumann, is a
scenario where an external authority assigns mixed strategies to each player in such a way that no
player has an incentive to deviate from the recommendation, provided that no other player deviates
6
from his [3]. In the context of repeated games, a conditional correlated equilibrium is a situation
where an external authority assigns mixed strategies to each player in such a way that no player
has an incentive to deviate from the recommendation in the second round, even after factoring in
information from the previous round of the game, provided that no other player deviates from his.
It is important to note that the concept of conditional correlated equilibrium presented here is different from the notions of extensive form correlated equilibrium and repeated game correlated equilibrium that have been studied in the game theory and economics literature [8, 12].
Notice that when the values taken for ?k are indepndent of its second argument, we retrieve the
familiar notion of correlated equilibrium.
Theorem 5. Suppose that all players in a K-player repeated game follow bigram conditional swap
regret minimizing strategies. Then, the joint empirical distribution of all players converges to a
conditional correlated equilibrium.
Proof. Let I t ? S be a random vector denoting the actions played by all K players in the game
at round t. The empirical joint distribution of every two subsequent rounds of a K-player game
?T ?
played repeatedly for T total rounds has the form P?T = T1 t=1 ?s?r)?S 2 ?{I t =s?I t?1 =r} , where
I = ?I1 ? ..? IK ) and Ik ? p?k) denotes the action played by player k using the mixed strategy p?k) .
t?1
? pt?1??k?1) . Then, the conditional swap regret of each player k,
Let q t??k) denote ?{i
t?1 =k}
reg?k? T ), can be bounded as follows since he is playing with a conditional swap regret minimizing
strategy:
reg?k? T ) =
T
T
?
?
1?
1?
l?k) ?sk ? s??k) ) ? min
?
? T
T t=1 stk ?pt??k)
t=1
?
? ?
log?N )
.
?O N
T
?
?stk ?st?1
)
k
??pt??k) ?q t??k) )
?
?
t
l?k) ???stk ? st?1
k )? s??k) )
De?ne the instantaneous conditional swap regret vector as
? ? ?
?
??
?k)
t
r?t?j0 ?j1 = ?{I t =j0 ?I t?1 =j1 } l?k) I t ? l?k) ?k ?j0 ? j1 )? I??k)
?
?k)
?k)
and the expected instantaneous conditional swap regret vector as
? ?
?
?
??
?k)
t
t
rt?j0 ?j1 = P?stk = j0 )?{I t?1 =j1 } l?k) j0 ? I??k)
? l?k) ?k ?j0 ? j1 )? I??k)
.
?k)
Consider the ?ltration Gt = {information of opponents at time t and of the player?s actions up to
?
? ?k)
?k)
?k)
?k)
time t ? 1}. Then, we see that ? r?t?j0 ?j1 |Gt = rt?j0 ?j1 . Thus, {Rt = rt?j0 ?j1 ? r?t?j0 ?j1 }?
t=1 is a
sequence of bounded martingale differences, and by the Hoeffding-Azuma inequality, we can write
?T
for any ? > 0, that P[| t=1 Rt | > ?] ? 2 exp??C?2 /T ) for some constant C > 0.
? ?
?? ?
? 2 ??
?
?
T
. By our concentration bound, we
Now de?ne the sets AT := ? T1 t=1 Rt ? > C
T log ?T
?
have P ?AT ) ? ?T . Setting ?T = exp?? T ) and applying the Borel-Cantelli lemma, we obtain
?T
that lim supT ?? | T1 t=1 Rt | = 0 a.s..
Finally, since each player followed a conditional swap regret minimizing strategy, we can write
?T
?k)
lim supT ?? T1 t=1 r?t?j0 ?j1 ? 0. Now, if the empirical distribution did not converge to a conditional correlated equilibrium, then by Prokhorov?s theorem, there exists a subsequence {P?Tj }j
satisfying the conditional correlated equilibrium inequality but converging to some limit P ? that is
not a conditional correlated equilibrium. This cannot be true because the inequality is closed under
weak limits.
Convergence to equilibria over the course of repeated game-playing also naturally implies the
scarcity of ?very suboptimal? strategies.
7
De?nition 2. An action pair ?sk ? rk ) ? Nk2 played by player k is conditionally ??dominated if
there exists a map ?k : Nk2 ? Nk such that
l?k) ?sk ? s??k) ) ? l?k) ??k ?sk ? rk )? s??k) ) ? ?.
Theorem 6. Suppose player k follows a conditional swap regret minimizing strategy that produces
a regret R over T instances of the repeated game. Then, on average, an action pair of player k is
R
conditionally ?-dominated at most ?T
fraction of the time.
The proof of this result is provided in Appendix 12.
6
Bandit Scenario
As discussed earlier, the bandit scenario differs from the full-information scenario in that the player
only receives information about the loss of his action f t ?xt ) at each time and not the entire loss
function f t . One standard external regret minimizing algorithm is the Exp3 algorithm introduced
by [2], and it is the base learner off of which we will build a conditional swap regret minimizing
algorithm.
To derive a sublinear conditional swap regret bound, we require an external regret bound on Exp3:
T
?
t=1
? [f t ?xt )] ? min
pt
a?N
T
?
t=1
?
f t ?a) ? 2 Lmin N log?N )?
which can be found in Theorem 3.1 of [5]. Using this estimate, we can derive the following result.
??
?
N 3 log?N )T .
Theorem 7. There exists an algorithm ? such that Reg?2 ?bandit ??? T ) ? O
The proof is given in Appendix 13 and is very similar to the proof for the full information setting.
It can also easily be extended in the analogous way to provide a regret bound for the k-gram regret
in the bandit scenario.
??
?
Theorem 8. There exists an algorithm ? such that Reg?k ?bandit ??? T ) ? O N k+1 log?N )T .
See Appendix 14 for an outline of the algorithm.
7
Conclusion
We analyzed the extent to which on-line learning scenarios are learnable. In contrast to some of
the more recent work that has focused on increasing the power of the adversary (see e.g. [1]), we
increased the power of the competitor class instead by allowing history-dependent action swaps and
thereby extending the notion of swap regret. We proved that this stronger class of competitors can
still be beaten in the sense of sublinear regret as long as the memory of the competitor is bounded.
We also provided a state-dependent bound that gives a more favorable guarantee when only some
parts of the history are considered. In the bigram setting, we introduced the notion of conditional
correlated equilibrium in the context of repeated K-player games, and showed how it can be seen
as a generalization of the traditional correlated equilibrium. We proved that if all players follow
bigram conditional swap regret minimizing strategies, then the empirical joint distribution converges
to a conditional correlated equilibrium and that no player can play very suboptimal strategies too
often. Finally, we showed that sublinear conditional swap regret can also be achieved in the partial
information bandit setting.
8
Acknowledgements
We thank the reviewers for their comments, many of which were very insightful. We are particularly
grateful to the reviewer who found an issue in our discussion on conditional correlated equilibrium
and proposed a helpful resolution. This work was partly funded by the NSF award IIS-1117591. The
material is also based upon work supported by the National Science Foundation Graduate Research
Fellowship under Grant No. DGE 1342536.
8
References
[1] Raman Arora, Ofer Dekel, and Ambuj Tewari. Online bandit learning against an adaptive
adversary: from regret to policy regret. In ICML, 2012.
[2] Peter Auer, Nicol`o Cesa-Bianchi, Yoav Freund, and Robert E. Schapire. The nonstochastic
multiarmed bandit problem. SIAM J. Comput., 32(1):48?77, 2002.
[3] Robert J. Aumann. Subjectivity and correlation in randomized strategies. Journal of Mathematical Economics, 1(1):67?96, March 1974.
[4] Avrim Blum and Yishay Mansour. From external to internal regret. Journal of Machine Learning Research, 8:1307?1324, 2007.
[5] S?ebastien Bubeck and Nicol`o Cesa-Bianchi. Regret analysis of stochastic and nonstochastic
multi-armed bandit problems. CoRR, abs/1204.5721, 2012.
[6] Nicol`o Cesa-Bianchi, Ofer Dekel, and Ohad Shamir. Online learning with switching costs and
other adaptive adversaries. In NIPS, pages 1160?1168, 2013.
[7] Nicol`o Cesa-Bianchi and G?abor Lugosi. Prediction, Learning, and Games. Cambridge University Press, New York, NY, USA, 2006.
[8] Francoise Forges. An approach to communication equilibria. Econometrica, 54(6):pp. 1375?
1385, 1986.
[9] Dean P. Foster and Rakesh V. Vohra. Calibrated learning and correlated equilibrium. Games
and Economic Behavior, 21(12):40 ? 55, 1997.
[10] Amy Greenwald, Zheng Li, and Warren Schudy. More ef?cient internal-regret-minimizing
algorithms. In COLT, pages 239?250. Omnipress, 2008.
[11] N. Halko, P. G. Martinsson, and J. A. Tropp. Finding structure with randomness: Probabilistic
algorithms for constructing approximate matrix decompositions. SIAM Rev., 53(2):217?288,
2011.
[12] Ehud Lehrer. Correlated equilibria in two-player repeated games with nonobservable actions.
Mathematics of Operations Research, 17(1):pp. 175?199, 1992.
[13] Nick Littlestone and Manfred K. Warmuth. The weighted majority algorithm. Inf. Comput.,
108(2):212?261, 1994.
? Tardos, and Vijay V. Vazirani. Algorithmic Game Theory.
[14] Noam Nisan, Tim Roughgarden, Eva
Cambridge University Press, New York, NY, USA, 2007.
[15] Gilles Stoltz and G?abor Lugosi. Learning correlated equilibria in games with compact sets of
strategies. Games and Economic Behavior, 59(1):187?208, 2007.
[16] Lloyd N. Trefethen and David Bau. Numerical Linear Algebra. SIAM: Society for Industrial
and Applied Mathematics, 1997.
9
| 5369 |@word inversion:1 bigram:11 stronger:2 dekel:2 decomposition:1 prokhorov:1 incurs:2 thereby:1 series:1 denoting:2 past:3 current:1 com:1 fn:1 numerical:1 subsequent:1 j1:11 update:1 aside:1 stationary:4 selected:1 warmuth:1 manfred:1 provides:2 completeness:1 authority:2 mathematical:1 along:1 c2:1 direct:1 ik:2 transducer:7 prove:2 introduce:3 market:1 expected:3 behavior:2 p1:8 multi:2 buying:1 actual:2 armed:2 considering:1 increasing:1 provided:5 bounded:9 notation:1 moreover:1 what:1 xed:1 argmin:1 interpreted:3 minimizes:1 substantially:1 hindsight:1 finding:2 guarantee:1 every:6 subclass:1 stricter:1 scaled:1 grant:1 t1:4 limit:3 consequence:1 switching:1 ext:1 lugosi:2 f1t:1 studied:2 suggests:1 schudy:1 lehrer:1 graduate:1 yj:1 lost:1 regret:83 practice:1 differs:1 nite:6 j0:12 empirical:8 projection:1 onto:1 cannot:1 context:3 applying:2 lmin:9 equivalent:1 map:3 deterministic:1 yt:1 missing:1 reviewer:2 straightforward:1 attention:1 regardless:1 independently:2 starting:1 economics:2 focused:1 resolution:1 immediately:1 assigns:2 amy:1 qtj:1 his:5 retrieve:1 notion:8 analogous:2 tardos:1 pt:23 suppose:3 play:2 yishay:1 exact:1 shamir:1 us:4 satisfying:1 particularly:1 updating:1 distributional:1 ft:4 eva:1 connected:1 nash:2 complexity:4 econometrica:1 depend:1 rewrite:1 grateful:1 algebra:1 incur:1 upon:2 negatively:1 learner:1 swap:58 completely:2 easily:2 joint:6 stock:2 choosing:1 trefethen:1 richer:1 pt1:1 drawing:1 qt1:1 online:12 sequence:9 eigenvalue:1 product:1 relevant:2 combining:2 achieve:1 dirac:2 rst:2 convergence:1 extending:1 produce:2 converges:4 tim:1 derive:4 measured:1 qt:8 received:2 implies:3 attribute:2 modifying:1 stochastic:2 material:1 require:1 generalization:1 tighter:1 extension:4 hold:1 considered:3 exp:2 equilibrium:32 mapping:1 algorithmic:1 purpose:1 ptj:1 favorable:1 bond:2 weighted:4 minimization:8 clearly:1 always:2 supt:2 ck:3 avoid:2 probabilistically:1 focus:1 improvement:1 cantelli:1 contrast:2 industrial:1 sense:1 helpful:1 dependent:4 factoring:1 entire:1 abor:2 bandit:11 subroutine:5 selects:2 i1:1 issue:1 among:2 uences:1 colt:1 initialize:1 construct:1 icml:1 simplify:1 oblivious:5 randomly:1 modi:7 tightly:1 comprehensive:1 individual:1 national:1 familiar:1 ourselves:1 consisting:1 microsoft:1 ab:1 zheng:1 analyzed:1 swapping:1 tj:1 chain:1 allocating:1 partial:3 ohad:1 unless:1 stoltz:1 littlestone:1 renormalize:1 instance:3 increased:1 earlier:1 yoav:1 cost:4 introducing:1 subset:3 at2:1 examining:1 too:1 motivating:1 chooses:3 calibrated:1 st:2 randomized:5 siam:3 probabilistic:1 off:1 picking:2 nk2:3 qtn:1 cesa:4 manage:1 opposed:1 choose:3 hoeffding:1 henceforth:1 external:15 expert:1 return:1 li:4 account:3 potential:1 de:12 lloyd:1 matter:1 depends:2 nisan:1 try:1 view:2 closed:1 portion:1 start:1 minimize:2 who:2 yield:1 dean:1 weak:1 vohra:1 asset:1 cation:6 history:8 randomness:1 reach:1 ed:1 against:12 competitor:14 pp:2 subjectivity:1 naturally:2 proof:11 proved:2 recall:1 knowledge:1 lim:2 actually:1 auer:1 courant:2 follow:7 response:1 done:1 generality:1 just:1 correlation:3 retrospect:1 hand:1 receives:1 tropp:1 google:1 quality:1 dge:1 usa:2 concept:1 true:1 assigned:1 illustrated:1 conditionally:2 round:13 game:25 coincides:2 outline:1 theoretic:3 tt:2 qti:1 fj:3 omnipress:1 instantaneous:2 ef:2 common:1 conditioning:1 extend:2 he:2 discussed:2 interpretation:3 interpret:1 martinsson:1 multiarmed:1 cambridge:2 ai:5 mathematics:2 similarly:1 portfolio:1 funded:1 access:1 gt:2 base:1 dominant:1 recent:3 showed:2 inf:1 scenario:22 inequality:6 meta:2 nition:5 seen:1 additional:1 speci:4 converge:2 bau:1 ii:1 full:3 exp3:2 long:1 award:1 a1:1 prediction:2 converging:1 metric:1 expectation:1 iteration:1 represent:1 achieved:1 receive:1 fellowship:1 allocated:1 comment:1 bought:2 call:5 yang:2 xj:1 fit:1 nonstochastic:2 restrict:1 suboptimal:2 economic:2 qj:3 expression:1 allocate:1 locating:1 peter:1 york:4 cause:1 action:52 remark:5 repeatedly:1 generally:1 useful:2 detailed:1 tewari:1 concentrated:1 reduced:1 generate:1 schapire:1 exist:2 nsf:1 notice:2 delta:2 per:2 write:4 incentive:2 key:2 blum:1 achieving:4 changing:1 pj:1 fraction:2 year:1 sum:3 reasonable:1 draw:4 raman:1 appendix:6 bound:15 pay:1 followed:1 played:8 nancial:1 roughgarden:1 precisely:2 ri:7 encodes:1 sake:1 dominated:3 argument:2 min:17 ptn:1 ned:1 according:1 march:1 pti:9 rev:1 s1:1 restricted:1 taken:1 available:1 ofer:2 forge:1 opponent:3 operation:1 batching:1 occurrence:1 batch:1 alternative:4 original:2 denotes:1 include:1 cally:2 build:1 approximating:1 society:1 objective:2 question:1 quantity:1 realized:2 strategy:19 concentration:1 rt:7 usual:1 traditional:2 thank:1 street:2 majority:4 extent:1 trivial:1 minn:1 multiplicatively:1 minimizing:15 equivalently:1 robert:2 noam:1 design:1 ebastien:1 policy:2 unknown:1 fjt:2 allowing:1 bianchi:4 gilles:1 markov:1 situation:3 extended:1 communication:1 y1:1 rn:1 mansour:1 arbitrary:3 community:1 introduced:3 david:1 pair:2 extensive:1 nick:1 nip:1 adversary:9 scott:1 azuma:1 ambuj:1 memory:6 power:3 natural:4 representing:1 pare:1 ne:6 cim:2 lk:2 arora:1 concludes:1 deviate:4 literature:1 acknowledgement:1 nicol:4 freund:1 loss:27 sublinear:7 interesting:1 mixed:3 ingredient:1 foundation:1 incurred:2 purchasing:1 mercer:2 foster:1 playing:2 pi:2 ibm:1 row:2 course:1 mohri:2 summary:1 surprisingly:1 supported:1 warren:1 allow:1 institute:2 taking:2 gram:7 cumulative:3 transition:1 collection:1 aumann:2 adaptive:2 far:1 transaction:1 vazirani:1 approximate:1 compact:1 active:1 buy:2 subsequence:1 iterative:1 investing:1 sk:8 wary:1 rearranging:1 mehryar:1 constructing:1 ehud:1 did:1 n2:1 repeated:8 x1:17 advice:1 referred:1 cient:1 borel:1 martingale:1 ny:4 ciency:1 comput:2 theorem:12 rk:4 xt:50 unigram:1 insightful:1 jensen:2 learnable:1 nyu:2 admits:3 beaten:1 exists:7 avrim:1 sequential:1 effectively:1 corr:1 conditioned:3 nk:6 easier:1 vijay:1 halko:1 simply:2 bubeck:1 recommendation:2 conditional:45 formulated:1 greenwald:1 stk:4 price:1 lemma:1 called:1 total:2 partly:1 player:56 rakesh:1 internal:4 scarcity:1 outgoing:1 reg:15 correlated:25 |
4,826 | 537 | Adaptive Synchronization of
Neural and Physical Oscillators
Kenji Doya
University of California, San Diego
La Jolla, CA 92093-0322, USA
Shuji Yoshizawa
University of Tokyo
Bunkyo-ku, Tokyo 113, Japan
Abstract
Animal locomotion patterns are controlled by recurrent neural networks
called central pattern generators (CPGs). Although a CPG can oscillate
autonomously, its rhythm and phase must be well coordinated with the
state of the physical system using sensory inputs. In this paper we propose
a learning algorithm for synchronizing neural and physical oscillators with
specific phase relationships. Sensory input connections are modified by the
correlation between cellular activities and input signals. Simulations show
that the learning rule can be used for setting sensory feedback connections
to a CPG as well as coupling connections between CPGs.
1
CENTRAL AND SENSORY MECHANISMS IN
LOCOMOTION CONTROL
Patterns of animal locomotion, such as walking, swimming, and fiying, are generated
by recurrent neural networks that are located in segmental ganglia of invertebrates
and spinal cords of vertebrates (Barnes and Gladden, 1985). These networks can
produce basic rhythms of locomotion without sensory inputs and are called central
pattern generators (CPGs). The physical systems of locomotion, such as legs, fins,
and wings combined with physical environments, have their own oscillatory characteristics. Therefore, in order to realize efficient locomotion, the frequency and
the phase of oscillation of a CPG must be well coordinated with the state of the
physical system. For example, the bursting patterns of motoneurons that drive a
leg muscle must be coordinated with the configuration of the leg, its contact with
the ground, and the state of other legs.
109
110
Doya and Yoshizawa
The oscillation pattern of a ePG is largely affected by proprioceptive inputs. It has
been shown in crayfish (Siller et al., 1986) and lamprey (Grillner et aI, 1990) that the
oscillation of a ePG is entrained by cyclic stimuli to stretch sensory neurons over a
wide range of frequency. Both negative and positive feedback pathways are found in
those systems. Elucidation of the function of the sensory inputs to CPGs requires
computational studies of neural and physical dynamical systems. Algorithms for
the learning of rhythmic patterns in recurrent neural networks have been derived by
Doya and Yoshizawa (1989), Pearlmutter (1989), and Williams and Zipser (1989).
In this paper we propose a learning algorithm for synchronizing a neural oscillator
to rhythmic input signals with a specific phase relationship.
It is well known that a coupling between nonlinear oscillators can entrainment their
frequencies. The relative phase between oscillators is determined by the parameters
of coupling and the difference of their intrinsic frequencies. For example, either
in-phase or anti-phase oscillation results from symmetric coupling between neural
oscillators with similar intrinsic frequencies (Kawato and Suzuki, 1980). Efficient
locomotion involves subtle phase relationships between physical variables and motor
commands. Accordingly, our goal is to derive a learning algorithm that can finely
tune the sensory input connections by which the relative phase between physical
and neural oscillators is kept at a specific value required by the task.
2
LEARNING OF SYNCHRONIZATION
We will deal with the following continuous-time model of a CPG network.
d e s
Ti
dtXi(t) = -Xi(t) +
L Wijgj(Xj(t)) + L Vi1:yA:(t) ,
(1)
j=1
1:=1
where Xi(t) and gi(Xi(t)) (i = 1, ... , C) represent the states and the outputs ofCPG
neurons and Y1:(t) (k = 1, ... , S) represents sensory inputs. We assume that the
connection weights W = {Wij} are already established so that the network oscillates
without sensory inputs. The goal oflearning is to find the input connection weights
V
{Vij} that make the network state x(t) (Xl (t), ... ,xc(t))t entrained to the
input signal yet) = (Yl(t), .. . ,Ys(t))t with a specific relative phase.
=
2.1
=
AN OBJECTIVE FUNCTION FOR PHASE-LOCKING
The standard way to derive a learning algorithm is to find out an objective function
to be minimized. If we can approximate the waveforms of Xi(t) and Y1:(t) by sine
waves, a linear relationship
x(t) = Py(t)
specifies a phase-locked oscillation of x(t) and Yet). For example, if we have Yl =
sin wt and Y2 = cos wt, then a matrix P = (~
specifies Xl = v'2 sinewt +1r /4) and
X2 = 2 sine wt + 1r /3). Even when the waveforms are not sinusoidal, minimization of
fi)
an objective function
1 c
1
E(t)
= "2l1x(t) -
py(t)1I2
s
="2 2: {Xi(t) - L Pi1:Y1:(t)}2
i=l
1:=1
(2)
Adaptive Synchronization of Neural and Physical Oscillators
determines a specific relative phase between x(t) and y(t). Thus we call P = {Pik}
a phase-lock matrix.
2.2
LEARNING PROCEDURE
Using the above objective function, we will derive a learning procedure for phaselocked oscillation of x(t) and y(t). First, an appropriate phase-lock matrix P is
identified while the relative phase between x(t) and y(t) changes gradually in time.
Then, a feedback mechanism can be applied so that the network state x(t) is kept
close to the target waveform P y(t).
Suppose we actually have an appropriate phase relationship between x(t) and y(t),
then the phase-lock matrix P can be obtained by gradient descent of E(t) with
respect to PH: as follows (Widrow and Stearns, 1985).
d
dtPik
= -TJ
{}E(t)
{}.
P,k
= TJ {Xi(t) -
S
LPijYj(t)}Yk(t).
(3)
j=1
If the coupling between x(t) and y(t) are weak enough, their relative phase changes
in time unless their intrinsic frequencies are exactly equal and the systems are
completely noiseless. By modulating the learning coefficient TJ by some performance
index of the total system, for example, the speed of locomotion, it is possible to
obtain a matrix P that satisfies the requirement of the task.
Once a phase-lock matrix is derived, we can control x(t) close to Py(t) using the
gradient of E(t) with respect to the network state
{}E(t)
{} X,.()t = Xi(t) -
S
L PikYk(t).
k=1
The simplest feedback algorithm is to add this term to the CPG dynamics as follows.
d e s
Ti dtXi(t) = -Xi(t) + L Wijgj(Xj(t)) - O'{Xi(t) - LPikYk(t)}.
k=1
j=1
The feedback gain 0' (> 0) must be set small enough so that the feedback term
does not destroy the intrinsic oscillation of the CPG. In that case, by neglecting the
small additional decay term O'Xi(t), we have
Tj
d e s
dt Xi(t) -Xj(t) + L Wijgj(Xj (t)) + L O'PikYk(t),
=
j=1
k=1
which is equivalent to the equation (1) with input weights Vik
= O'Pik.
(4)
111
112
Doya and Yoshizawa
3
DELAYED SYNCHRONIZATION
We tested the above learning scheme on a delayed synchronization task; to find
coupling weights between neural oscillators so that they synchronize with a specific
time delay. We used the following coupled CPG model.
c
c
Tdd xi(t) = -xi(t) + L wijyj(t) + ~ Lpi1:y~-n(t),
t
.
J=1
1:=1
yi(t)
= g(xi(t)),
(5)
(i = 1, . .. , C),
=
where superscripts denote the indices of two CPGs (n 1,2). The goal of learning
was to synchronize the waveforms yHt) and y~(t) with a time delay ~T. We used
z(t) = -Iy~(t - ~T) - y~(t)1
as the performance index. The learning coefficient 7] of equation (3) was modulated
by the deviation of z(t) from its running average z(t) using the following equations.
7](t)
= 7]0 {z(t) -
d
z(t)},
Ta dt z(t)
= -z(t) + z(t).
(6)
a
.....
y2
0.0
4. 0
8. 0
12. 0
16. 0
b
20. 0
d
24. 0
28. 0
32. 0
....
y1
y2~rvl\?
O. 0
4. 0
8. 0
12. 0
0.'-;;'O---;4~:0i\""""'""-"8."A'o---:-l-;!-i"""o-~1S: 0
16. 0
c
e
y1 .
y2~y2
0.0
4.0
8.0
12. 0
16. 0
o....'o-----:4:-'-::o::---~8:. 0---:-1-;:1-i-:-0
..
--1~6: 0
Figure 1: Learning of delayed synchronization of neural oscillators. The dotted and
solid curves represent yf(t) and y;(t) respectively. a:without coupling. b:~T = 0.0.
c:~T = 1.0. c:~T = 2.0. d : ~T = 3.0.
Adaptive Synchronization of Neural and Physical Oscillators
First, two CPGs were trained independently to oscillate with sinusoidal waveforms
of period Tl
4.0 and T2
5.0 using continuous-time back-propagation learning
(Doyaand Yoshizawa, 1989). Each CPG was composed of two neurons (C = 2) with
time constants T
1.0 and output functions g()
tanh() . Instead of following
the two step procedure described in the previous section, the network dynamics (5)
and the learning equations (3) and (6) were simulated concurrently with parameters
a = 0.1, '10 = 0.2, and To = 20.0.
=
=
=
=
Figure 1 a shows the oscillation of two CPGs without coupling. Figures 1 b through
e show the phase-locked waveforms after learning for 200 time units with different
desired delay times.
ZERO-LEGGED LOCOMOTION
4
Next we applied the learning rule to the simplest locomotion system that involves a critical phase-lock between the state of the physical system and the motor
command-a zero-legged locomotion system as shown in Figure 2 a.
The physical system is composed of a wheel and a weight that moves back and
forth on a track fixed radially in the wheel. It rolls on the ground by changing its
balance with the displacement of the weight. In order to move the wheel in a given
direction, the weight must be moved at a specific phase with the rotation angle of
the wheel. The motion equations are shown in Appendix.
First, a CPG network was trained to oscillate with a sinusoidal waveform of period
T
1.0 (Doya and Yoshizawa, 1989). The network consisted of one output and
two hidden units (C = 3) with time constants Ti 0.2 and output functions giO =
tanh(). Next, the output of the CPG was used to drive the weight with a force
/max gl(Xl(t?. The position T and the velocity T of the weight and the rotation
angle (cos 0, sin 0) and the angular velocity of the wheel iJ were used as sensory
feedback inputs Yl:(t) (k 1, .. . ,5) after scaling to [-1,1].
=
=
/=
=
In order to eliminate the effect of biases in x(t) and yet), we used the following
learni~g equations.
d
dtPil:
= '1 ((Xi(t) -
S
Xi(t? -
L Pi; (y;(t) -
y;(t?}(Yl:(t) - Yl:(t?,
;=1
d
Ttl:
dt Xi(t) = -Xi(t) + Xi(t),
Ty
dtYl:(t)
d
(7)
=-Yl:(t) + Yl:(t).
The rotation speed of the wheel was employed as the performance index z(t) after
smoothing by the following equation.
T,
d
.
dt z(t) = -z(t) + OCt).
The learning coefficient '1 was modulated by equations (6). The time constants were
Ttl:
4.0, Ty
1.0, T, = 1.0, and To
4.0. Each training run was started from a
random configuration of the wheel and was finished after ten seconds.
=
=
=
113
114
Doya and Yoshizawa
a
sin90
?
cos9O----
9~
b
pos
vel
cos
SID
rot
0.0
/'
-0.5
,
,
,
,
,
1.0
2.0
3.0
4.0
5.0
/'
,perle6.0 0.0
0.0
,
,
1.0
2.0
/'
;-=
5.0
/'
/'
3. 0
/'
4.0
6.0
0.5
c
pos "------'
vel
cos
sm
~
bidS
O. 0
-0.5
/'
1. 0
2. 0
/'
3. 0
/'
4. 0
/'
5. 0
:r----,. ._. '.___,'-,- - -' , . - - '-:-'
6. 0 O. 0
_ , - 1 - '_-::-I'
1. 0
2. 0
3. 0
0.0
Figure 2: Learning of zero-legged locomotion.
4. 0
5. 0
6. 0
0.5
Adaptive Synchronization of Neural and Physical Oscillators
Figure 2 b is an example of the motion of the wheel without sensory feedback.
The rhythms of the CPG and the physical system were not entrained to each other
and the wheel wandered left and right. Figure 2 c shows an example of the wheel
motion after 40 runs of training with parameters Tlo = 0.1 and (}' = 0.2. At first, the
oscillation of the CPG was slowed down by the sensory inputs and then accelerated
with the rotation of the wheel in the right direction.
We compared the patterns of sensory input connections made after learning with
wheels of different sizes. Table 1 shows the connection weights to the output unit.
The positive connection from sin 0 forces the weight to the right-hand side of the
wheel and stabilize clockwise rotation. The negative connection from cos 0 with
smaller radius fastens the rhythm of the CPG when the wheel rotates too fast and
the weight is lifted up. The positive input from r with larger radius makes the
weight stickier to both ends of the track and slows down the rhythm of the CPG.
Table 1: Sensory input weights to the output unit (Plk; k = 1, ... ,5).
radius
2cm
4cm
6cm
8cm
10cm
5
r
r
0.15
0.28
0.67
0.70
0.90
-0.53
-0.55
-0.21
-0.33
-0.12
cosO
-1.35
-1.09
-0.41
-0.40
-0 .30
sinO
1.32
1.22
0.98
0.92
0.93
0
0.07
0.01
0.00
0.03
-0.02
DISCUSSION
The architectures of CPGs in lower vertebrates and invertebrates are supposed to
be determined by genetic information. Nevertheless, the wayan animal utilizes the
sensory inputs must be adaptive to the characteristics of the physical environments
and the changing dimensions of its body parts.
Back-propagation through forward models of physical systems can also be applied
to the learning of sensory feedback (Jordan and Jacobs, 1990). However, learning of
nonlinear dynamics of locomotion systems is a difficult task; moreover, multi-layer
back-propagation is not appropriate as a biological model of learning. The learning
rule (7) is similar to the covariance learning rule (Sejnowski and Stanton, 1990),
which is a biological model of long term potentiation of synapses.
Acknowledgements
The authors thank Allen Selverston, Peter Rowat, and those who gave comments
to our poster at NIPS Conference. This work was partly supported by grants from
the Ministry of Education, Culture, and Science of Japan.
115
116
Doya and Yoshizawa
References
Barnes, W. J. P. & Gladden, M. H. (1985) Feedback and Motor Control in Invertebrates and Vertebrates. Beckenham, Britain: Croom Helm.
Doya, K. & Yoshizawa, S. (1989) Adaptive neural oscillator using continuous-time
back-propagation learning. Neural Networks, 2, 375-386.
Grillner, S. & Matsushima, T. (1991) The neural network underlying locomotion in
Lamprey-Synaptic and cellular mechanisms. Neuron, 7(July), 1-15.
Jordan, M. I. & Jacobs, R. A. (1990) Learning to control an unstable system with
forward modeling. In Touretzky, D. S. (ed.), Advances in Neural Information Processing Systems 2. San Mateo, CA: Morgan Kaufmann.
Kawato, M. & Suzuki, R. (1980) Two coupled neural oscillators as a model of the
circadian pacemaker. Journal of Theoretical Biology, 86, 547-575.
Pearlmutter, B. A. (1989) Learning state space trajectories in recurrent neural networks. Neural Computation, 1, 263-269.
Sejnowski, T. J. & Stanton, P. K. (1990) Covariance storage in the Hippocampus.
In Zornetzer, S. F. et aI. (eds.), An Introduction to Neural and Electronic Networks,
365-377. San Diego, CA: Academic Press.
Siller, K. T., Skorupski, P., Elson, R. C., & Bush, M. H. (1986) Two identified
afferent neurones entrain a central locomotor rhythm generator. Nature, 323, 440443.
Widrow, B. & Stearns, S. D. (1985) Adaptive Signal Processing. Englewood Cliffs,
NJ: Prentice Hall.
Williams, R. J. & Zipser, D. (1989) A learning algorithm for continually running
fully recurrent neural networks. Neural Computation, 1, 270-280.
Appendix
The dynamics of the zero-legged locomotion system:
..
mr
=
.f.(1
JO
+
mR2 sin2 0)
(0
10
- mgc cos
+
mRsin 2 0(r+RcosO?
10
? Ov+2mr(r+RcosO)0'
0'2
10
+mr ,
+m R sm
-loR sin 0 + mgcsinO(r + RcosO) - (v + 2mr(r + RcosO?O,
100
10
Imax
10
1+ MR2
g(Xl(t? -
ur 3 - /Jr,
+ m(r + RcoSO)2.
Parameters: the masses of the weight m = 0.2[kg) and the wheel M = 0.8[kg);
the radius of the wheel R
0.02throughO.l[m)j the inertial moment of the wheel
I
M R2 j the maximum force to the weight 1max
5[N) j the stiffness of the
3
limiter of the weight u
20/ R3 [N/m ); the damping coefficients of the weight
motion /J 0.2/ R [N/(m/s?) and the wheel rotation v 0.05(M +m)R [N/(rad/s?).
=
=t
=
=
=
=
| 537 |@word mr2:2 hippocampus:1 vi1:1 simulation:1 jacob:2 covariance:2 solid:1 moment:1 configuration:2 cyclic:1 tlo:1 genetic:1 yet:3 must:6 realize:1 motor:3 pacemaker:1 accordingly:1 cpg:14 lor:1 gio:1 pathway:1 multi:1 vertebrate:3 moreover:1 underlying:1 mass:1 ttl:2 cm:5 kg:2 selverston:1 nj:1 ti:3 oscillates:1 exactly:1 control:4 unit:4 grant:1 continually:1 positive:3 cliff:1 mateo:1 bursting:1 co:6 range:1 locked:2 procedure:3 displacement:1 epg:2 poster:1 close:2 wheel:18 mgc:1 storage:1 prentice:1 py:3 equivalent:1 britain:1 williams:2 independently:1 rule:4 imax:1 diego:2 target:1 suppose:1 locomotion:15 velocity:2 walking:1 located:1 coso:1 cord:1 l1x:1 autonomously:1 yk:1 environment:2 locking:1 dynamic:4 legged:4 trained:2 ov:1 completely:1 po:2 fast:1 sejnowski:2 larger:1 yht:1 gi:1 superscript:1 propose:2 wijyj:1 supposed:1 forth:1 moved:1 requirement:1 circadian:1 produce:1 coupling:8 recurrent:5 derive:3 widrow:2 ij:1 kenji:1 involves:2 direction:2 waveform:7 radius:4 tokyo:2 education:1 wandered:1 potentiation:1 biological:2 stretch:1 hall:1 ground:2 tanh:2 limiter:1 modulating:1 minimization:1 concurrently:1 modified:1 lifted:1 command:2 derived:2 sin2:1 eliminate:1 hidden:1 wij:1 animal:3 smoothing:1 equal:1 once:1 plk:1 biology:1 represents:1 synchronizing:2 minimized:1 t2:1 stimulus:1 composed:2 delayed:3 phase:23 englewood:1 tj:4 neglecting:1 culture:1 tdd:1 unless:1 damping:1 desired:1 theoretical:1 modeling:1 shuji:1 oflearning:1 deviation:1 delay:3 too:1 combined:1 yl:7 iy:1 jo:1 central:4 wing:1 japan:2 sinusoidal:3 stabilize:1 coefficient:4 coordinated:3 afferent:1 sine:2 wave:1 elson:1 roll:1 kaufmann:1 characteristic:2 largely:1 who:1 weak:1 sid:1 trajectory:1 drive:2 oscillatory:1 synapsis:1 touretzky:1 synaptic:1 ed:2 ty:2 frequency:6 yoshizawa:9 gain:1 radially:1 subtle:1 inertial:1 actually:1 back:5 ta:1 dt:4 vel:2 angular:1 correlation:1 hand:1 nonlinear:2 propagation:4 yf:1 usa:1 effect:1 consisted:1 y2:5 symmetric:1 proprioceptive:1 i2:1 deal:1 sin:4 rhythm:6 dtxi:2 pearlmutter:2 motion:4 allen:1 gladden:2 fi:1 rotation:6 kawato:2 physical:17 spinal:1 ai:2 rot:1 locomotor:1 add:1 segmental:1 own:1 jolla:1 yi:1 muscle:1 morgan:1 motoneuron:1 additional:1 ministry:1 mr:4 employed:1 period:2 signal:4 clockwise:1 july:1 academic:1 long:1 y:1 controlled:1 basic:1 noiseless:1 represent:2 finely:1 comment:1 entrained:3 jordan:2 call:1 zipser:2 enough:2 bid:1 xj:4 gave:1 architecture:1 identified:2 vik:1 peter:1 bunkyo:1 neurones:1 oscillate:3 tune:1 ten:1 ph:1 simplest:2 specifies:2 stearns:2 dotted:1 track:2 affected:1 nevertheless:1 changing:2 kept:2 destroy:1 swimming:1 run:2 angle:2 electronic:1 doya:8 utilizes:1 oscillation:9 pik:2 appendix:2 scaling:1 layer:1 barnes:2 activity:1 x2:1 invertebrate:3 speed:2 pi1:1 jr:1 smaller:1 ur:1 leg:4 slowed:1 gradually:1 equation:8 r3:1 mechanism:3 end:1 stiffness:1 appropriate:3 running:2 lock:5 elucidation:1 lamprey:2 xc:1 contact:1 objective:4 move:2 already:1 gradient:2 thank:1 rotates:1 simulated:1 cellular:2 unstable:1 index:4 relationship:5 balance:1 difficult:1 negative:2 slows:1 neuron:4 sm:2 fin:1 descent:1 anti:1 y1:5 rowat:1 required:1 connection:10 rad:1 california:1 established:1 nip:1 dynamical:1 pattern:8 max:2 critical:1 force:3 synchronize:2 scheme:1 stanton:2 phaselocked:1 finished:1 started:1 coupled:2 acknowledgement:1 relative:6 synchronization:8 fully:1 generator:3 vij:1 pi:1 gl:1 supported:1 cpgs:8 bias:1 side:1 wide:1 stickier:1 rhythmic:2 feedback:10 curve:1 dimension:1 sensory:17 forward:2 suzuki:2 adaptive:7 san:3 made:1 author:1 approximate:1 xi:19 zornetzer:1 continuous:3 helm:1 table:2 ku:1 nature:1 ca:3 grillner:2 body:1 crayfish:1 tl:1 position:1 xl:4 entrain:1 matsushima:1 down:2 specific:7 r2:1 decay:1 intrinsic:4 ganglion:1 determines:1 satisfies:1 oct:1 goal:3 oscillator:14 change:2 determined:2 entrainment:1 wt:3 called:2 total:1 partly:1 la:1 ya:1 modulated:2 accelerated:1 bush:1 tested:1 |
4,827 | 5,370 | Efficient Partial Monitoring with Prior Information
Hastagiri P Vanchinathan
Dept. of Computer Science
ETH Z?urich, Switzerland
[email protected]
G?abor Bart?ok
Dept. of Computer Science
ETH Z?urich, Switzerland
[email protected]
Andreas Krause
Dept. of Computer Science
ETH Z?urich, Switzerland
[email protected]
Abstract
Partial monitoring is a general model for online learning with limited feedback: a
learner chooses actions in a sequential manner while an opponent chooses outcomes.
In every round, the learner suffers some loss and receives some feedback based on
the action and the outcome. The goal of the learner is to minimize her cumulative
loss. Applications range from dynamic pricing to label-efficient prediction to dueling
bandits. In this paper, we assume that we are given some prior information about the
distribution based on which the opponent generates the outcomes. We propose BPM, a
family of new efficient algorithms whose core is to track the outcome distribution with
an ellipsoid centered around the estimated distribution. We show that our algorithm
provably enjoys near-optimal regret rate for locally observable partial-monitoring
problems against stochastic opponents. As demonstrated with experiments on synthetic
as well as real-world data, the algorithm outperforms previous approaches, even for very
uninformed priors, with an order of magnitude smaller regret and lower running time.
1 Introduction
We consider Partial Monitoring, a repeated game where in every time step a learner chooses an action while,
simultaneously, an opponent chooses an outcome. Then the player receives a loss based on the action and
outcome chosen. The learner also receives some feedback based on which she can make better decisions in
subsequent time steps. The goal of the learner is to minimize her cumulative loss over some time horizon.
The performance of the learner is measured by the regret, the excess cumulative loss of the learner
compared to that of the best fixed constant action. If the regret scales linearly with the time horizon, it
means that the learner does not approach the performance of the best action, that is, the learner fails to learn
the problem. On the other hand, sublinear regret indicates that the disadvantage of the learner compared
to the best fixed strategy fades with time.
Games in which the learner receives the outcome as feedback after every time step are called online
learning with full information. This special case of partial monitoring has been addressed by (among
others) Vovk [1] and Littlestone and Warmuth [2], who designed the randomized
? algorithm Exponentially
Weighted Averages (EWA) as a learner strategy. This algorithm achieves ?( T logN) expected regret
against any opponent, where N is the number of actions and T is the time horizon. This regret growth
rate is also proven to be optimal.
Another well-studied special case is the so-called multi-armed bandit problem. In this feedback model,
the learner gets to observe the loss she suffered in every time step. That is, the learner does not receive
any information about losses of actions she did not choose. Asymptotically optimal results were obtained
by Audibert and Bubeck
? [3], who designed the1Implicitly Normalized Forecaster (INF) that achieves the
minimax optimal ?( T N) regret growth rate.
1
The algorithm Exp3 due to Auer et al. [4] achieves the same rate up to a logarithmic factor.
1
However, not all online learning problems have one of the above feedback structures. An important
example for a problem that does not fit in either full-information or bandit problems is dynamic pricing.
Consider the problem of a vendor wanting to sell his products to customers for the best possible price.
When a customer comes in, she (secretly) decides on a maximum price she is willing to buy his product
for, while the vendor has to set a price without knowing the customer?s preferences. The loss of the vendor
is some preset constant if the customer did not buy the product, and an ?opportunity loss?, when the
product was sold cheaper than the customer?s maximum. The feedback, on the other hand, is merely an
indicator whether the transaction happened or not.
Dynamic pricing is just one of the practical applications of partial monitoring. Label efficient prediction,
in its simplest form, has three actions: the first two actions are guesses of a binary outcome but provide
no information, while the third action provides information about the outcome for some unit loss as the
price. This can be thought of an abstract form of spam filtering: the first two actions correspond to putting
an email to the inbox and the spam folder, the third action corresponds to asking the user if the email
is spam or not. Another problem that can be cast as partial monitoring is that of dueling bandits [5, 6]
in which the learner chooses a pair of actions in every time step, the loss she suffers is the average loss
of the two actions, and the feedback is which action was ?better?.
In online learning, we distinguish different models of how the opponent generates the outcomes. In the
mildest version called stochastic or stationary memoryless, the opponent chooses an outcome distribution
before the game starts and then selects outcomes in an iid random manner drawn from the chosen
distribution. The oblivious adversarial opponent chooses the outcomes arbitrarily, but without observing the
actions of the learner. This selection method is equivalent to choosing an outcome sequence ahead of time.
Finally, the non-oblivious or adaptive adversarial opponent chooses outcomes arbitrarily with the possibility
of looking at past actions of the learner. In this work, we focus on strategies against stochastic opponents.
Related work Partial monitoring was first addressed in the seminal paper of Piccolboni and Schindelhauer [7], who designed and analyzed the algorithm FeedExp3. The algorithm?s main idea is to maintain
an unbiased estimate for the loss of each action in every time step, and then use these estimates to run
the full-information algorithm (EWA). Piccolboni and Schindelhauer [7] proved an O(T 3/4) upper bound
on the regret (not taking into account the number of actions) for games for which learning is at all possible.
This bound was later improved by Cesa-Bianchi et al. [8] to O(T 2/3), who also constructed an example
of a problem for which this bound is optimal.
From the above bounds it can be seen that not?all partial-monitoring problems have the same level of
difficulty: while bandit problems enjoy an O( T ) regret rate, some partial-monitoring problems have
?(T 2/3) regret. To this end, Bart?ok et al. [9] showed that partial-monitoring problems with finitely
? many
e T ) regret,
actions and outcomes can be classified into four groups: trivial with zero regret, easy with ?(
hard with ?(T 2/3) regret, and hopeless with linear regret. The distinguishing feature between easy and
hard problems is the local observability condition, an algebraic condition on the feedback structure that can
be efficiently verified for any problem. Bart?ok et al. [9] showed the above classification against stochastic
opponents with the help of algorithm BALATON. This algorithm keeps track of estimates of the loss
difference of ?neighboring? action pairs and eliminates actions that are highly likely to be suboptimal.
?
e T ) regret bound for easy
Since then, several algorithms have been proposed that achieve the O(
games [10, 11]. All these algorithms rely on the core idea of estimating the expected loss difference
between pairs of actions.
Our contributions In this paper, we introduce BPM (Bayes-update Partial Monitoring), a new family of
algorithms against iid stochastic opponents that rely on a novel way of the usage of past observations. Our
algorithms maintain a confidence ellipsoid in the space of outcome distributions, and update the ellipsoid
based on observations following a Bayes-like update. Our approach enjoys better empirical performance
and lower computational overhead; another crucial advantage is that we can incorporate prior information
about the outcome distribution by means of an initial confidence ellipsoid. We prove near-optimal minimax
expected regret bounds for our algorithm, and demonstrate its effectiveness on several partial monitoring
problems on synthetic and real data.
2
2 Problem setup
Partial monitoring is a repeated game where in every round, a learner chooses an action while the opponent
chooses an outcome from some finite action and outcome sets. Then, the learner observes a feedback
signal (from some given set of symbols) and suffers some loss, both of which are deterministic functions
of the action and outcome chosen. In this paper we assume that the opponent chooses the outcomes in
an iid stochastic manner. The goal of the learner is to minimize her cumulative loss.
The following definitions and concepts are mostly taken from Bart?ok et al. [9]. An instance of partial
monitoring is defined by the loss matrix L?RN?M and the feedback table H ??N?M , where N and M
are the cardinality of the action set and the outcome set, respectively, while ? is some alphabet of symbols.
That is, if learner chooses action i while the outcome is j, the loss suffered by the learner is L[i,j], and
the feedback received is H[i,j].
For an action 1 ? i ? N, let `i denote the column vector given by the ith row of L. Let ?M denote the
M-dimensional probability simplex. It is easy to see that for any p??M , if we assume that the opponent
uses p to draw the outcomes (that is, p is the opponent strategy), the expected loss of action i can be
expressed as `>
i p.
We measure the performance of an algorithm with its expected regret, defined as the expected difference
of the cumulative loss of the algorithm and that of the best fixed action in hindsight:
RT = max
1?i?N
T
X
(`It ?`i)>p,
t=1
where T is some time horizon, It (t = 1,...,T ) is the action chosen in time step t, and p is the outcome
distribution the opponent uses.
In this paper, we also assume we have some prior knowledge about the outcome distribution in the form
of a confidence ellipsoid: we are given a distribution p0 ? ?M and a symmetric positive semidefinite
covariance matrix ?0 ?RM?M such that the true outcome distribution p? satisfies
q
?
kp0 ?p?k??1 = (p0 ?p?)>??1
0 (p0 ?p )?1.
0
We use the term ?confidence ellipsoid? even though our condition is not probabilistic; we do not assume
that p? is drawn from a Gaussian distribution before the game starts. On the other hand, the way we track
p? is derived by Bayes updates with a Gaussian conjugate prior, hence the name. We would also like
to note that having the above prior knowledge is without loss of generality. For ?large enough? ?0, the
whole probability simplex is contained in the confidence ellipsoid and thus partial monitoring without
any prior information reduces to our setting.
The following definition reveals how we use the loss matrix to recover the structure of a game.
Definition 1 (Cell decomposition, Bart?ok et al. [9, Definition 2]). For any action 1?i?N, let Ci denote
the set of opponent strategies for which action i is optimal:
Ci = p??M : ?1?j ?N,(`i ?`j )>p?0 .
We call the set Ci the optimality cell of action i. Furthermore, we call the set of optimality cells {C1,...,CN }
the cell decomposition of the game.
Every cell Ci is a convex closed polytope, as it is defined by a linear inequality system. Normally, a cell
has dimension M ?1, which is the same as the dimensionality of the probability simplex. It might happen
however, that a cell is of lower dimensionality. Another possible degeneracy is when two actions share
the same cell. In this paper, for ease of presentation, we assume that these degeneracies do not appear.
For an illustration of cell decomposition, see Figure 1(a).
Now that we know the regions of optimality, we can define when two actions are neighbors. Intuitively,
two actions are neighbors if their optimality cells are neighbors in the strong sense that they not only meet
in ?one corner?.
Definition 2 (Neighbors, Bart?ok et al. [9, page 4]). Two actions i and j are neighbors, if the intersection
of their optimality cells Ci ?Cj is an M ?2-dimensional convex polytope.
3
C5
C3
p?
C1
C4
p?
p?
pt
pt?1
pt?1
C2
(a) Cell decomposition
(b) Before the update
(c) After the update
Figure 1: (a) An example for a cell decomposition with M = 3 outcomes. Under the true outcome distribution
p? , action 3 is optimal. Cells C1 and C3 are neighbors, but C2 and C5 are not. (b) The current estimate pt?1 is far away
from the true distribution, the confidence ellipsoid is large. (c) After updating, pt is closer to the truth, the confidence
ellipsoid shrinks.
To optimize performance, the learner?s primary goal is to find out which cell the opponent strategy lies
in. Then, the learner can choose the action associated with that cell to play optimally. Since the feedback
the learner receives is limited, this task of finding the optimal cell may be challenging.
The next definition enables us to utilize the feedback table H.
Definition 3 (Signal matrix, Bart?ok et al. [9, Definition 1]). Let {?1,?2,...,??i }?? be the set of symbols
appearing in row i of the feedback table H. We define the signal matrix Si ?{0,1}?i ?M of action i as
Si[k,j]=I(H[i,j]=?k ).
In words, Si is the indicator table of observing symbols ?1,...,??i under outcomes 1,...,M given that
the action chosen is i. For an example, consider the case when the ith row of H is (a b a c). Then,
!
1 0 1 0
Si = 0 1 0 0 .
0 0 0 1
A very useful property of the signal matrix is that if we represent outcomes with M-dimensional unit vectors,
then Si can be used as a linear transformation to arrive at the unit-vector representation of the observation.
The following condition condition is key in distinguishing easy and hard games:
Definition 4 (Local observability, Bart?ok et al. [9, Definition 3]). Let actions i and j be neighbors. These
actions are said to be locally observable if `i ? `j ? ImSi> ? ImSj>. Furthermore, a game is locally
observable if all of its neighboring action pairs are locally observable.
Bart?ok et
?al. [9] showed that finite stochastic partial-monitoring problems that admit local observability
e T ) minimax expected regret. In the following, we present our new algorithm family that achieves
have ?(
the same regret rate for locally observable games against stochastic opponents.
3
BPM: New algorithms for Partial Monitoring based on Bayes updates
The algorithms we propose can be decomposed into two main building blocks: the first one keeps track
of a belief about the true outcome distribution and provides us with a set of feasible actions in every round.
The second one is responsible for selecting the action to play from this action set. Pseudocode for the
algorithm family is shown in Algorithm 1.
3.1 Update Rule
The method of updating the belief about the true outcome distribution (p?) is based on the idea that
we pretend that the outcomes are generated from a Gaussian distribution with covariance ? = IM and
unknown mean. We also pretend we have a Gaussian prior for tracking the mean. The parameters of
this prior are denoted by p0 (mean) and ?0 (covariance). In every time step, we perform a Gaussian
Bayes-update using the observation received.
4
Algorithm 1 BPM
input: L,H,p0,?0
initialization: Calculate signal matrices Si
for t=1 to T do
Use selection rule (cf., Sec. 3.2) to choose an action It
Observe feedback Yt
?1
?1
>
> ?1
Update posterior: ??1
t =?t?1 +PIt and pt =?t ?t?1 pt?1 +SIt (SIt SIt ) Yt ;
end for
Full-information case As a gentle start, we explain how the update rule would look like if we had full
information about the outcome in each time step. The update in this case is identical with the standard
Gaussian one-step update:
?1
?t =?t?1 ??t?1(?t?1 +I)
?t =?t ??1
t?1 ?t?1 +Xt
?t?1
or equiv.
or equiv.
?1
??1
t =?t?1 +I,
?t =?t?1 +?t(Xt ??t?1).
Here we use subindex t?1 for the prior parameters and t for the posterior parameters in time step t, and
denote by Xt the outcome (observed in this case), encoded by an M-dimensional unit vector.
General case Moving away from the full-information case, we face the problem of not observing the
outcome, only some symbol that is governed by the signal matrix of the action we chose and the outcome
itself. If we denote, as above, the outcome at time step t by an M-dimensional unit vector Xt, then the
observation symbol can be thought of as a unit vector given by Yt = SiXt, provided the chosen action
is i. It follows that what we observe is a linear transformation of the sample from the outcome distribution.
Following the Bayes update rule and assuming we chose action i at time step t, we derive the
corresponding Gaussian posterior given that the likelihood of the observation is ?(Y |p)?N(Sip,SiSi>).
After some algebraic manipulations we get that the posterior
distribution is Gaussian with covariance
?1
?1
?t =(??1
+P
)
and
mean
p
=?
?
p
+P
X
,
where
Pi =Si>(SiSi>)?1Si is the orthogonal
i
t
t
t?1
i
t
t?1
t?1
>
projection to the image space of Si . Note that even though Xt is not observed, the update can be
performed, since PiXi =Si>(SiSi>)?1SiXt =Si>(SiSi>)?1Yt.
A significant advantage of this method of tracking the outcome distribution as opposed to keeping track
of loss difference estimates (as done in previous works), is that feedback from one action can provide
information about losses across all the actions. We believe that this property has a major role in the
empirical performance improvement over existing methods.
An important part in analyzing our algorithm is to show that, despite the fact that the outcome distribution
is not Gaussian, the update tracks the true outcome distribution well. For an illustration of tracking the
true outcome distribution with the above update, see Figures 1(b) and 1(c).
3.2 Selection rules
For selecting actions given the posterior parameters, we propose two versions for the selection rule:
1. Draw a random sample p from the distribution N(pt?1,?t?1), project the sample to the probability simplex, then choose the action that minimizes the loss for outcome distribution p. This
rule is a close relative of Thompson-sampling. We call this version of the algorithm BPM-TS.
2. Use pt?1 and ?t?1 to build a confidence ellipsoid for p?, enumerate all actions whose cells
intersect with this ellipsoid, then choose the action that was chosen the fewest times so far (called
BPM-LEAST).
Our experiments demonstrate the performance of both versions. We analyze version BPM-LEAST.
5
4 Analysis
We now analyze BPM-LEAST that uses the Gaussian updates, and considers a set of feasible actions based
on the criterion that an action is feasible if its optimality cell intersects with the ellipsoid
(
)
r
1
p:kp?ptk??1 ?1+
NlogMT .
t
2
From these feasible actions, it picks the one that has been chosen the fewest times up to time step t. For
this version of the algorithm, the following regret bound holds.
Theorem 1. Given a locally observable partial-monitoring problem (L,H) with prior information p0,?0,
the algorithm BPM-LEAST achieves expected regret
p
RT ?C T Nlog(MT ),
where C is some problem-dependent constant.
The above constant C depends on two main factors, both of them related to the feedback structure.
The first one is the sum of the smallest eigenvalues of SiSi> for every action i. The second is related
to the local observability condition. As the condition says, for every neighboring action pairs i and j,
`i ?`j ?ImSi> ?ImSj>. This means that there exist vij and vji vectors such that `i ?`j =Si>vij ?Sj>vji.
The constant depends on the maximum 2-norm of these vij vectors.
The proof of the theorem is deferred to the supplementary material. In a nutshell, the proof is divided into
two main parts. First we need to show that the update rule?even though the underlying distribution is not
Gaussian?serves as a good tool for tracking the true outcome distribution. After some algebraic manipulations, the problem reduces to a finding a high probability upper bound for norms of weighted sums of noise
vectors. To this end, we used the martingale version of the matrix Hoeffding inequality [12, Theorem 1.3].
Then, we need to show that the confidence ellipsoid shrinks fast enough that if we only choose actions
whose cell intersect with the ellipsoid, we do not suffer a large regret. In the core of proving this, we arrive
at a term where we need to upper bound k`i ? `j k?t , for some neighboring action pairs (i,j), and we
show that due to local observability?
and the speed at which the posterior covariance shrinks, this term
can be upper bounded by roughly 1/ t.
5 Experiments
First, we run extensive evaluations of BPM on various synthetic datasets and compare the performance
against CBP [10] and FeedExp3 [7]. The datasets used in the simulated experiments are identical to the
ones used by Bart?ok et al. [10] and thus allow us to benchmark against the current state of the art. We also
provide results of BPM on a dataset that was collected by Singla and Krause [13] from real interactions
with many users on the Amazon Mechanical Turk (AMT) [14] crowdsourcing platform. We present the
details of the datasets used and the summarize our results and findings in this section.
5.1 Implementation Details
In order to implement BPM, we made the following implementation choices:
1. To use BPM-LEAST (see Section 3.2), we need to recover the current feasible actions. We
do so by sampling multiple (10000) times from concentric Gaussian ellipsoids centred at the
current mean (pt) and collect feasible actions based on which cells the samples lie in. We resort
to sampling for ease of implementation because otherwise we deal with the problem of finding
the intersection between an ellipsoid and a simplex in M-dimensional space.
2. To implement BPM-TS, we draw p from the distribution N(pt?1,?t?1). We then project it
back to the simplex to obtain a probability distribution on the outcome space.
Primarily due to sampling, both our algorithms are computationally more efficient than the existing
approaches. In particular, BPM-TS is ideally suited for real world tasks as it is several orders of magnitude
faster than existing algorithms during all our experiments. In each iteration, BPM-TS only needs to
draw one sample from a multivariate gaussian and does not need any cell decompositions or expensive
computations to obtain high dimensional intersections.
6
40
20
BPM?TS
BPM?Least
10
FeedExp
0
40
BPM?TS
30
10
2
4
6
Time Steps ? 105
8
2
4
6
Time Steps ? 105
8
0
10
FeedExp
18
CBP
14
Regret ? 103
Regret ? 103
Regret ? 103
2
8
6
BPM?TS
4
12
BPM?TS
10
8
CBP
6
BPM?Least
BPM?Least
BPM?TS
4
FeedExp
16
8
10
10
20
10
CBP
5
7.5
Time Steps ? 105
(c) Effects of priors
FeedExp
0
2.5
(b) Minimax (hard)
4
4
2
BPM?Least
2
2
0
0
misspec. p0,wide ?0
0
0
0
10
8
6
5
BPM?Least
(a) Minimax (easy)
6
accurate p0,wide ?0
accurate p0, tight ?0
20
5
0
0
10
Regret ? 103
Minimax Regret ? 104
Minimax Regret ? 103
CBP
25
misspec. p0,tight ?0
CBP
20
FeedExp
30
15
10
30
35
0
2.5
5
Time Steps ? 105
7.5
10
(d) Single opponent (easy).
0
0
2
4
6
Time Steps ? 105
8
0
10
(e) Single opponent (hard).
0.5
1
1.5
2
Time Steps ? 105
2.5
3
(f) Real data (dynamic pricing).
Figure 2: (a,b,d,e) Comparing BPM on the locally non-observable game ((a,d) benign opponent; (c,e) hard opponent).
Hereby, (a,b) show the pointwise maximal regret over 15 scenarios, and (d,e) show the regret against a single opponent
strategy. (c) shows the effect of a misspecified prior. (f) is the performance of the algorithms on the real dynamic
pricing dataset.
5.2 Simulated dynamic pricing games
Dynamic pricing is a classic example of partial monitoring (see the introduction), and we compare the
performance of the algorithms on the small but not locally observable game described in Bart?ok et al.
[10]. The loss matrix and feedback tables for the dynamic pricing game are given by:
?
?
?
?
0 1 ??? N ?1
y y ??? y
?n y ??? y?
?c 0 ??? N ?2?
?
L= ?
;
H =?
.. ?
? ... . . . . . .
?
? ... . . . . . . ... ?.
.
c
???
c
0
n
???
n
y
Recall that the game models a repeated interaction of a seller with buyers in a market. Each buyer can
either buy the product at the price (signal ?y?) or deny the offer (signal ?n?). The corresponding loss to
the seller is either a known constant c (representing opportunity cost) or the difference between offered
price and the outcome of the customer?s latent valuation of the product (willingness-to-pay). A similar
game models procurement processes as well. Note that this game does not satisfy local observability.
While our theoretical results require this condition, in practice, if the opponent does not intentionally select
harsh regions on the simplex, BPM remains applicable. Under this setting, expected individual regret is a
reasonable measure of performance. That is, we measure the expected regret for fixed opponent strategies.
We also consider the minimax expected regret, which measures worst-case performance (pointwise
maximum) against multiple opponent strategies.
Benign opponent While the dynamic pricing game is not locally observable in general, certain opponent
strategies are easier to compete with than others. Specifically, if the stochastic opponent chooses an
outcome distribution that is away from the intersection of the cells that do not have local observability, the
learning happens in ?non-dangerous? or benign regions. We present results under this setting for simulated
dynamic pricing with N = M = 5. The results shown in Figures 2(a) and 2(d) illustrate the benefits of
both variants of BPM over previous approaches. We achieve an order of magnitude reduction in the regret
suffered w.r.t. both the minimax and the individual regret.
7
Harsh opponent For the same problem, with opponent chooses close to the boundary of the cells of
two non-locally observable actions, the problem becomes harder. Still, BPM dramatically outperforms
the baselines and suffers very little regret as shown in Figures 2(b) and 2(e).
Effect of the prior We study the effects of a misspecified prior in Figure 2(c). As long as the initial
confidence interval specified by the prior covariance is large enough to contain the opponent?s distribution,
an incorrectly specified prior mean does not have an adverse effect on the performance of BPM. As
expected, if the prior confidence ellipse used by BPM does not contain the opponent?s outcome distribution,
however, the regret grows linear in time. Further, if the prior is very informative (accurately specified
prior mean and tight confidence ellipse), very little regret is suffered.
5.3 Results on real data
Dataset description We simulate a procurement game based on real data. Parameter estimation was
done by posting a Human Intelligence Task (HIT) on the Amazon Mechanical Turk (AMT) platform.
Motivated by an application in viral marketing, users were asked about the price they would accept for
(hypothetically) letting us post promotional material to their friends on a social networking site. The survey
also collected features like age, geographic region, number of friends in the social network, activity levels
(year of joining, time spent per day etc.). Note that since the HIT was just a survey and the questions
were about a hypothetical scenario, participants had no incentives to misreport their responses. Complete
responses were collected from approx. 800 participants. See [13] for more details.
The procurement game We simulate a procurement auction by playing back these responses offline.
The game is very similar in structure to dynamic pricing, with the optimal action being the best fixed price
that maximized the marketer?s value or equivalently, minimized the loss. We sampled iid from the survey
data and perturbed the samples slightly to simulate a stream of 300000 potential users. At each iteration,
we simulate a user with a private valuation generated as a function of her attributes. We discretized the
offer prices and the private valuations to be one of 11 values and set the opportunity cost of losing a user
due to low pricing to be 0.5. Thus we recover a partial monitoring game with 11 actions and 11 outcomes
with a 0/1 feedback matrix.
Results We present the results of our evaluation on this dataset in Figure 2(f). Notice that although the
game is not locally observable, the outcome distribution does not seem to be in a difficult region of the cell
decomposition as the adaptive algorithms (CBP and both versions of BPM) perform well. We note that
the total regret suffered by BPM-LEAST is a factor of 10 lower than the regret achieved by CBP on this
dataset. The plots are averaged over 30 runs of the competing algorithms on the stream. To the best of our
knowledge, this is the first time partial monitoring has been evaluated on a real world problem of this size.
6 Conclusions and future work
We introduced a new family of algorithms for locally observable partial-monitoring problems against
stochastic opponents. We also enriched the model of partial monitoring with the possibility of incorporating
prior information about the outcome distribution in the form of a confidence ellipsoid. The new insight
of our approach is that instead of tracking loss differences, we explicitly track the true outcome distribution.
This approach not only eases computational overhead but also helps achieve low regret by being able
to transfer information between actions. In particular, BPM-TS runs orders of magnitude faster than any
existing algorithms, opening the path for the model of partial monitoring to be applied on realistic settings
involving large numbers of actions and outcomes.
Future work includes extending our method for adversarial opponents. Bart?ok [11] already uses the idea
of tracking the true outcome distribution with the help of a confidence parallelotope, which is rather close
to our approach, but has the same shortcomings as other algorithms that track loss differences: it can not
transfer information between actions. Extending our results to problems with large action and outcome
spaces is also an important direction: if we have some prior information about the similarities between
outcomes and/or actions, we have a chance for a reasonable regret.
Acknowledgments This research was supported in part by SNSF grant 200021 137971, ERC StG
307036 and a Microsoft Research Faculty Fellowship.
8
References
[1] V. G. Vovk. Aggregating strategies. In COLT, pages 371?386, 1990.
[2] Nick Littlestone and Manfred K. Warmuth. The weighted majority algorithm. Inf. Comput., 108
(2):212?261, 1994.
[3] Jean-Yves Audibert and S?ebastien Bubeck. Minimax policies for adversarial and stochastic bandits.
In COLT, 2009.
[4] Peter Auer, Nicol`o Cesa-Bianchi, Yoav Freund, and Robert E. Schapire. The nonstochastic
multiarmed bandit problem. SIAM J. Comput., 32(1):48?77, 2002.
[5] Yisong Yue, Josef Broder, Robert Kleinberg, and Thorsten Joachims. The K-armed dueling bandits
problem. Journal of Computer and System Sciences, 78(5):1538?1556, 2012.
[6] Nir Ailon, Thorsten Joachims, and Zohar Karnin. Reducing dueling bandits to cardinal bandits.
arXiv preprint arXiv:1405.3396, 2014.
[7] Antonio Piccolboni and Christian Schindelhauer. Discrete prediction games with arbitrary feedback
and loss. In COLT/EuroCOLT, pages 208?223, 2001.
[8] Nicol`o Cesa-Bianchi, G?abor Lugosi, and Gilles Stoltz. Regret minimization under partial monitoring.
Math. Oper. Res., 31(3):562?580, 2006.
[9] G?abor Bart?ok, D?avid P?al, and Csaba Szepesv?ari. Minimax regret of finite partial-monitoring games
in stochastic environments. Journal of Machine Learning Research - Proceedings Track (COLT),
19:133?154, 2011.
[10] G?abor Bart?ok, Navid Zolghadr, and Csaba Szepesv?ari. An adaptive algorithm for finite stochastic
partial monitoring. In Proceedings of the 29th International Conference on Machine Learning, ICML
2012, Edinburgh, Scotland, UK, June 26 - July 1, 2012, 2012.
[11] G?abor Bart?ok. A near-optimal algorithm for finite partial-monitoring games against adversarial
opponents. In COLT 2013 - The 26th Annual Conference on Learning Theory, June 12-14, 2013,
Princeton University, NJ, USA, pages 696?710, 2013.
[12] Joel A Tropp. User-friendly tail bounds for sums of random matrices. Foundations of Computational
Mathematics, 12(4):389?434, 2012.
[13] Adish Singla and Andreas Krause. Truthful incentives in crowdsourcing tasks using regret
minimization mechanisms. In International World Wide Web Conference (WWW), 2013.
[14] Amazon Mechanical Turk platform. URL https://www.mturk.com.
9
| 5370 |@word private:2 faculty:1 version:8 norm:2 willing:1 forecaster:1 covariance:6 p0:10 decomposition:7 pick:1 harder:1 reduction:1 initial:2 selecting:2 outperforms:2 past:2 existing:4 current:4 comparing:1 com:1 si:12 subsequent:1 happen:1 informative:1 benign:3 realistic:1 enables:1 christian:1 designed:3 plot:1 update:19 bart:15 stationary:1 intelligence:1 guess:1 warmuth:2 scotland:1 ith:2 core:3 manfred:1 provides:2 math:1 preference:1 cbp:8 constructed:1 c2:2 prove:1 overhead:2 introduce:1 manner:3 market:1 expected:12 roughly:1 multi:1 discretized:1 eurocolt:1 decomposed:1 kp0:1 little:2 armed:2 cardinality:1 becomes:1 provided:1 estimating:1 project:2 underlying:1 bounded:1 what:1 minimizes:1 sisi:5 hindsight:1 finding:4 transformation:2 csaba:2 nj:1 every:12 hypothetical:1 friendly:1 growth:2 nutshell:1 sip:1 rm:1 hit:2 uk:1 unit:6 normally:1 enjoy:1 appear:1 grant:1 before:3 positive:1 schindelhauer:3 local:7 aggregating:1 despite:1 joining:1 analyzing:1 meet:1 path:1 lugosi:1 might:1 chose:2 initialization:1 studied:1 collect:1 challenging:1 pit:1 ease:2 limited:2 range:1 averaged:1 practical:1 responsible:1 acknowledgment:1 practice:1 regret:46 block:1 implement:2 inbox:1 intersect:2 empirical:2 eth:3 thought:2 projection:1 confidence:14 word:1 get:2 close:3 selection:4 seminal:1 optimize:1 equivalent:1 deterministic:1 demonstrated:1 customer:6 yt:4 secretly:1 www:2 urich:3 convex:2 thompson:1 survey:3 amazon:3 fade:1 rule:8 insight:1 his:2 proving:1 classic:1 pt:11 play:2 user:7 losing:1 distinguishing:2 us:4 expensive:1 updating:2 observed:2 role:1 preprint:1 worst:1 calculate:1 region:5 observes:1 environment:1 ideally:1 asked:1 dynamic:11 seller:2 tight:3 learner:26 various:1 intersects:1 alphabet:1 fewest:2 fast:1 shortcoming:1 kp:1 outcome:59 choosing:1 whose:3 encoded:1 supplementary:1 jean:1 say:1 otherwise:1 itself:1 online:4 ptk:1 sequence:1 advantage:2 eigenvalue:1 propose:3 nlog:1 interaction:2 product:6 maximal:1 neighboring:4 achieve:3 description:1 gentle:1 extending:2 help:3 derive:1 illustrate:1 friend:2 spent:1 uninformed:1 measured:1 finitely:1 received:2 strong:1 come:1 switzerland:3 direction:1 attribute:1 stochastic:13 centered:1 human:1 material:2 require:1 equiv:2 im:1 hold:1 around:1 major:1 achieves:5 smallest:1 estimation:1 applicable:1 label:2 singla:2 tool:1 weighted:3 minimization:2 snsf:1 gaussian:13 rather:1 derived:1 focus:1 misreport:1 joachim:2 she:6 improvement:1 june:2 indicates:1 likelihood:1 adversarial:5 stg:1 baseline:1 sense:1 dependent:1 promotional:1 accept:1 abor:5 her:4 bandit:10 bpm:35 selects:1 provably:1 josef:1 among:1 classification:1 colt:5 logn:1 denoted:1 art:1 special:2 platform:3 karnin:1 having:1 sampling:4 identical:2 sell:1 look:1 icml:1 future:2 simplex:7 others:2 minimized:1 cardinal:1 opening:1 primarily:1 oblivious:2 simultaneously:1 individual:2 cheaper:1 bartok:1 maintain:2 microsoft:1 possibility:2 highly:1 evaluation:2 joel:1 deferred:1 analyzed:1 semidefinite:1 accurate:2 closer:1 partial:29 orthogonal:1 stoltz:1 littlestone:2 re:1 theoretical:1 instance:1 column:1 asking:1 disadvantage:1 yoav:1 cost:2 optimally:1 perturbed:1 synthetic:3 chooses:14 broder:1 randomized:1 siam:1 international:2 eas:1 probabilistic:1 cesa:3 yisong:1 opposed:1 choose:6 hoeffding:1 corner:1 admit:1 resort:1 oper:1 account:1 potential:1 centred:1 sec:1 includes:1 satisfy:1 explicitly:1 audibert:2 depends:2 stream:2 later:1 performed:1 closed:1 observing:3 analyze:2 start:3 bayes:6 recover:3 participant:2 contribution:1 minimize:3 yves:1 who:4 efficiently:1 maximized:1 correspond:1 accurately:1 iid:4 monitoring:29 classified:1 explain:1 networking:1 suffers:4 email:2 definition:10 against:12 turk:3 intentionally:1 hereby:1 associated:1 proof:2 degeneracy:2 sampled:1 proved:1 dataset:5 recall:1 knowledge:3 dimensionality:2 cj:1 auer:2 back:2 ok:15 day:1 response:3 improved:1 done:2 though:3 shrink:3 generality:1 furthermore:2 just:2 marketing:1 evaluated:1 hand:3 receives:5 tropp:1 web:1 willingness:1 pricing:12 believe:1 grows:1 building:1 usage:1 name:1 normalized:1 unbiased:1 concept:1 true:10 piccolboni:3 hence:1 effect:5 contain:2 geographic:1 memoryless:1 symmetric:1 deal:1 round:3 game:28 during:1 criterion:1 complete:1 demonstrate:2 auction:1 image:1 novel:1 ari:2 misspecified:2 viral:1 pseudocode:1 mt:1 exponentially:1 tail:1 significant:1 multiarmed:1 approx:1 mathematics:1 deny:1 erc:1 had:2 moving:1 similarity:1 etc:1 posterior:6 multivariate:1 showed:3 inf:4 manipulation:2 scenario:2 certain:1 inequality:2 binary:1 arbitrarily:2 seen:1 vanchinathan:1 truthful:1 signal:8 july:1 full:6 multiple:2 reduces:2 faster:2 exp3:1 offer:2 long:1 divided:1 post:1 navid:1 prediction:3 variant:1 involving:1 mturk:1 arxiv:2 iteration:2 represent:1 achieved:1 cell:25 ewa:2 receive:1 c1:3 fellowship:1 krause:3 szepesv:2 addressed:2 interval:1 suffered:5 crucial:1 eliminates:1 yue:1 effectiveness:1 seem:1 call:3 near:3 easy:7 enough:3 fit:1 nonstochastic:1 competing:1 suboptimal:1 andreas:2 idea:4 observability:7 knowing:1 cn:1 avid:1 whether:1 motivated:1 url:1 suffer:1 peter:1 algebraic:3 action:77 antonio:1 enumerate:1 useful:1 subindex:1 dramatically:1 locally:12 simplest:1 schapire:1 http:1 exist:1 notice:1 happened:1 estimated:1 track:9 per:1 discrete:1 incentive:2 group:1 putting:1 four:1 key:1 drawn:2 verified:1 utilize:1 asymptotically:1 merely:1 sum:3 year:1 run:4 compete:1 arrive:2 family:5 reasonable:2 feedexp3:2 draw:4 decision:1 bound:10 pay:1 distinguish:1 annual:1 activity:1 ahead:1 dangerous:1 generates:2 kleinberg:1 speed:1 simulate:4 optimality:6 ailon:1 conjugate:1 smaller:1 across:1 slightly:1 happens:1 vji:2 intuitively:1 thorsten:2 taken:1 computationally:1 vendor:3 remains:1 mechanism:1 know:1 letting:1 end:3 serf:1 opponent:38 observe:3 away:3 appearing:1 running:1 cf:1 opportunity:3 zolghadr:1 pretend:2 build:1 ellipse:2 question:1 already:1 strategy:11 primary:1 rt:2 said:1 simulated:3 majority:1 polytope:2 valuation:3 considers:1 collected:3 trivial:1 assuming:1 pointwise:2 ellipsoid:17 illustration:2 equivalently:1 setup:1 mostly:1 difficult:1 robert:2 implementation:3 ebastien:1 policy:1 unknown:1 perform:2 bianchi:3 upper:4 gilles:1 observation:6 datasets:3 sold:1 benchmark:1 finite:5 t:10 incorrectly:1 looking:1 rn:1 arbitrary:1 concentric:1 introduced:1 cast:1 pair:6 mechanical:3 c3:2 extensive:1 specified:3 nick:1 c4:1 zohar:1 able:1 summarize:1 max:1 belief:2 dueling:4 difficulty:1 rely:2 indicator:2 wanting:1 minimax:11 representing:1 harsh:2 nir:1 prior:23 nicol:2 relative:1 freund:1 loss:32 sublinear:1 filtering:1 proven:1 age:1 foundation:1 krausea:1 offered:1 usa:1 vij:3 playing:1 share:1 pi:1 row:3 adish:1 hopeless:1 supported:1 keeping:1 enjoys:2 offline:1 allow:1 neighbor:7 wide:3 taking:1 face:1 benefit:1 edinburgh:1 feedback:21 dimension:1 boundary:1 world:4 cumulative:5 c5:2 folder:1 adaptive:3 made:1 spam:3 far:2 social:2 transaction:1 excess:1 sj:1 observable:12 keep:2 decides:1 buy:3 reveals:1 latent:1 table:5 learn:1 transfer:2 did:2 main:4 linearly:1 whole:1 noise:1 repeated:3 enriched:1 site:1 martingale:1 fails:1 comput:2 lie:2 governed:1 third:2 posting:1 procurement:4 theorem:3 xt:5 symbol:6 sit:3 incorporating:1 sequential:1 ci:5 magnitude:4 horizon:4 easier:1 suited:1 intersection:4 logarithmic:1 likely:1 bubeck:2 expressed:1 contained:1 tracking:6 marketer:1 ch:3 corresponds:1 truth:1 satisfies:1 amt:2 chance:1 goal:4 presentation:1 price:9 sixt:2 feasible:6 hard:6 hastagiri:2 adverse:1 specifically:1 reducing:1 vovk:2 preset:1 called:4 total:1 buyer:2 player:1 select:1 hypothetically:1 ethz:3 incorporate:1 dept:3 princeton:1 crowdsourcing:2 |
4,828 | 5,371 | Nonparametric Bayesian inference on multivariate
exponential families
William Vega-Brown, Marek Doniec, and Nicholas Roy
Massachusetts Institute of Technology
Cambridge, MA 02139
{wrvb, doniec, nickroy}@csail.mit.edu
Abstract
We develop a model by choosing the maximum entropy distribution from the
set of models satisfying certain smoothness and independence criteria; we show
that inference on this model generalizes local kernel estimation to the context of
Bayesian inference on stochastic processes. Our model enables Bayesian inference in contexts when standard techniques like Gaussian process inference are
too expensive to apply. Exact inference on our model is possible for any likelihood function from the exponential family. Inference is then highly efficient,
requiring only O (log N ) time and O (N ) space at run time. We demonstrate our
algorithm on several problems and show quantifiable improvement in both speed
and performance relative to models based on the Gaussian process.
1
Introduction
Many learning problems can be formulated in terms of inference on predictive stochastic models.
These models are distributions p(y|x) over possible observation values y drawn from some observation set Y, conditioned on a known input value x from an input set X . The supervised learning
problem is then to infer a distribution p(y|x? , D) over possible observations for some target input
x? , given a sequence of N independent observations D = {(x1 , y 1 ), . . . , (xN , y N )}.
It is often convenient to associate latent parameters ? ? ? with each input x, where p(y|?) is a
known likelihood function. By inferring a distribution over target parameters ? ? associated with x? ,
we can infer a distribution over y.
Z
p(y|x? , D) =
d? ? p(y|? ? )p(? ? |x? , D)
(1)
?
For instance, regression problems can be formulated as the inference of an unknown but deterministic underlying function ?(x) given noisy observations, so that p(y|x) = N (y; ?(x), ? 2 ), where
? 2 is a known noise variance. If we can specify a joint prior over the parameters corresponding to
different inputs, we can infer p(? ? |x? , D) using Bayes? rule.
"N
#
Z
Y
p(? ? |x? , D) ?
d? i p(y i |? i ) p(? 1:N , ? ? |x? , x1:N )
(2)
?N
i=1
The notation x1:N indicates the sample inputs {x1 , . . . , xN }; this model is depicted graphically in
figure 1a. Although the choice of likelihood is often straightforward, specifying a prior can be more
difficult. Ideally, we want a prior which is expressive, in the sense that it can accurately capture all
prior knowledge, and which permits efficient and accurate inference.
A powerful motivating example for specifying problems in terms of generative models is the Gaussian process [1], which specifies the prior p(? 1:N |x1:N ) as a multivariate Gaussian with a covariance
parameterized by x1:N . This prior can express complex and subtle relationships between inputs and
1
xn
xn
?n
?n
yn
N
yn
?
x?
i, j
??
y?
(b) Inference model
(a) Stochastic process
Figure 1: Figure 1a models any stochastic process with fully connected latent parameters. Figure 1b is our
approximate model, used for inference; we assume that the latent parameters are independent given the target
parameters. Shaded nodes are observed.
observations, and permits efficient exact inference for a Gaussian likelihood with known variance.
Extensions exist to perform approximate inference with other likelihood functions [2, 3, 4, 5].
However, the assumptions of the Gaussian process are not the only set of reasonable assumptions,
and are not always appropriate. Very large datasets require complex sparsification techniques to be
computationally tractable [6]. Likelihood functions with many coupled parameters, such as the parameters of a categorical distribution or of the covariance matrix of a multivariate Gaussian, require
the introduction of large numbers of latent variables which must be inferred approximately. As an
example, the generalized Wishart process developed by Wilson and Ghahramani [7] provides a distribution over covariance matrices using a sum of Gaussian processes. Inference of the the posterior
distribution over the covariance can only be performed approximately: no exact inference procedure
is known.
Historically, an alternative approach to estimation has been to use local kernel estimation techniques
[8,
are often developed from a weighted parameter likelihood of the form p(?|D) =
Q 9, 10], which
wi
i p(y i |?) . Algorithms for determining the maximum likelihood parameters of such a model
are easy to implement and very fast in practice; various techniques, such as dual trees [11] or the
improved fast Gauss transform [12] allow the computation of kernel estimates in logarithmic or
constant time. This choice of model is often principally motivated by the computational convenience
of resulting algorithms. However, it is not clear how to perform Bayesian inference on such models.
Most instantiations instead return a point estimate of the parameters.
In this paper, we bridge the gap between the local kernel estimators and Bayesian inference. Rather
than perform approximate inference on an exact generative model, we formulate a simplified model
for which we can efficiently perform exact inference. Our simplification is to choose the maximum
entropy distribution from the set of models satisfying certain smoothness and independence criteria. We then show that for any likelihood function in the exponential family, our process model
has a conjugate prior, which permits us to perform Bayesian inference in closed form. This motivates many of the local kernel estimators from a Bayesian perspective, and generalizes them to new
problem domains. We demonstrate the usefulness of this model on multidimensional regression
problems with coupled observations and input-dependent noise, a setting which is difficult to model
using Gaussian process inference.
2
The kernel process model
Given a likelihood function, a generative model can be specified by a prior p(? 1:N , ? ? |x? , x1:N ).
For almost all combinations of prior and likelihood, inference is analytically intractable. We relax
the requirement that the model be generative, and instead require only that the prior be well-formed
for a given query x? . To facilitate inference, we make the strong assumption that the latent training
parameters ? 1:N are conditionally independent given the target parameters ? ? .
"N
#
Y
?
?
?
?
p(? 1:N , ? |x1:N , x ) =
p(? i |? , xi , x ) p(? ? |x? )
(3)
i=1
This restricted structure is depicted graphically in figure 1b. Essentially, we assume that interactions between latent parameters are unimportant relative to interactions between the latent and target
parameters; this will allow us to build models based on exponential family likelihood functions
which will permit exact inference. Note that models with this structure will not correspond exactly
to probabilistic generative models; the prior distribution over the latent parameters associated with
the training inputs varies depending on the target input. Instead of approximating inference on our
2
model, we make our approximation at the stage of model selection; in doing so, we enable fast,
exact inference. Note that the class of models with the structure of equation (3) is quite rich, and
as we demonstrate in section 3.2, performs well on many problems. We discuss in section 4 the
ramifications of this assumption and when it is appropriate.
This assumption is closely related to known techniques for sparsifying Gaussian process inference.
Qui?nonero-Candela and Rasmussen [6] provide a summary of many sparsification techniques, and
describe which correspond to generative models. One of the most successful sparsification techniques, the fully independent training conditional approximation of Snelson [13], assumes all training examples are independent given a specified set of inducing inputs. Our assumption extends this
to the case of a single inducing input equal to the target input.
Note that by marginalizing the parameters ? 1:N , we can directly relate the observations y 1:N to the
target parameters ? ? . Combining equations (2) and (3),
"N Z
#
Y
? ?
?
p(? |x , D) ?
d? i p(y i |? i )p(? i |? , xi , x? ) p(? ? |x? )
(4)
i=1
?
and marginalizing the latent parameters ? 1:N , we observe that the posterior factors into a product
over likelihoods p(y i |? ? , x, x? ) and a prior over ? ? .
"N
#
Y
?
?
=
p(y i |? , xi , x ) p(? ? |x? )
(5)
i=1
Note that we can equivalently specify either p(?|? ? , x, x? ) or p(y|? ? , x, x? ), without loss of generality. In other words, we can equivalently describe the interaction between input variables either
in the likelihood function or in the prior.
2.1
The extended likelihood function
By construction, we know the distribution p(y i |? i ). After making the transformation to equation (5), much of the problem of model specification has shifted to specifying the distribution
p(y i |? ? , xi , x? ). We call this distribution the extended likelihood distribution. Intuitively, these distributions should be related; if x? = xi , then we expect ? i = ? ? and p(y i |? ? , xi , x? ) = p(y i |? i ).
We therefore define the extended likelihood function in terms of the known likelihood p(y i |? i ).
Typically, we prefer smooth models: we expect similar inputs to lead to a similar distribution over
outputs. In the absence of a smoothness constraint, any inference method can perform arbitrarily
poorly [14]. However, the notion of smoothness is not well-defined in the context of probability
distributions. Denote g(y i ) = p(y i |? ? , xi , x? ), and f (y i ) = p(y i |? i ). We can formalize a smooth
model as one in which the information divergence of the likelihood distribution f from the extended
likelihood distribution g is bounded by some function ? : X ? X ? R+ .
DKL (gkf ) ? ?(x? , xi )
(6)
Since the divergence is a premetric, ?(?, ?) must also satisfy the properties of a premetric: ?(x, x) =
0 ?x and ?(x1 , x2 ) ? 0 ?x1 , x2 . For example, if X = Rn , we may draw an analogy to Lipschitz
continuity and choose ?(x1 , x2 ) = Kkx1 ? x2 k, with K a positive constant. The class of models
with bounded divergence has the property that g ? f as x0 ? x, and it does so smoothly provided
?(?, ?) is smooth. Note that this bound is a constraint on possible g, not an objective to be minimized;
in particular, we do not minimize the divergence between g and f to develop an approximation, as is
common in the approximate inference literature. Note also that this constraint has a straightforward
information-theoretic interpretation; ?(x1 , x2 ) is a bound on the amount of information we would
lose if we were to assume an observation y 1 were taken at x2 instead of at x1 .
The assumptions of equations (3) and (6) define a class of models for a given likelihood function,
but are insufficient for specifying a well-defined prior. We therefore use the principle of maximum
entropy and choose the maximum entropy distribution from among that class. In our attached supporting material, we prove the following.
Theorem 1 The maximum entropy distribution g satisfying DKL (gkf ) ? ?(x? , x) has the form
?
g(y) ? f (y)k(x ,x)
(7)
where k : X ? X ? [0, 1] is a kernel function which can be uniquely determined from ?(?, ?) and
f (?).
3
There is an equivalence relationship between functions k(?, ?) and ?(?, ?); as either is uniquely determined by the other, it may more convenient to select a kernel function than a smoothness bound,
and doing so implies no loss in generality or correctness. Note it is neither necessary nor sufficient
that the kernel function k(?, ?) be positive definite. It is necessary only that k(x, x) = 1?x and that
k(x, x0 ) ? [0, 1]?x, x0 . This includes the possibility of asymmetric kernel functions. We discuss
in the attached supporting material the mapping between valid kernel functions k(?, ?) and bounding
functions ?(?, ?).
It follows from equation (7) that the maximum entropy distribution satisfying a bound of ?(x, x? )
on the divergence of the observation distribution p(y|? ? , x, x? ) from the known distribution
p(y|?, x, x? ) = p(y|?) is
?
p(y|? ? , x, x? ) ? p(y|?)k(x,x ) .
(8)
By combining equations (5) and (6), we can fully specify a stochastic model with a likelihood
p(y|?), a pointwise marginal prior p(?|x), and a kernel function k : X ? X ? [0, 1]. To perform
inference, we must evaluate
p(?|x, D) ?
N
Y
p(y i |?)k(x,xi ) p(?|x)
(9)
i=1
This can be done in closed form if we can normalize the terms on the right side of the equality.
In certain limiting cases with uninformative priors, our model can be reduced to known frequentist
estimators. For instance, if we employ an uninformative prior p(?|x) ? 1 and choose the maximum? ? = arg max p(? ? |x? , D), we recover the weighted maximumlikelihood target parameters ?
likelihood estimator, detailed by Wang [15]. If the function k(x, x0 ) is local, in the sense that it goes
to zero if the distance kx ? x0 k is large, then choosing maximum likelihood parameter estimates for
an uninformative prior gives the locally weighted maximum-likelihood estimator, described in the
context of regression by Cleveland [16] and for generalized linear models by Tibshirani and Hastie
[10]. However, our result is derived from a Bayesian interpretation of statistics, and we infer a full
distribution over the parameters; we are not limited to a point estimate. The distinction is of both
academic and practical interest; in addition to providing insight into to the meaning of the weighting
function and the validity of the inferred parameters, by inferring a posterior distribution we provide
a principled way to reason about our knowledge and to insert prior knowledge of the underlying
process.
2.2
Kernel inference on the exponential family
Equation (8) is particularly useful if we choose our likelihood model p(y|?) from the exponential
family.
p(y|?) = h(y) exp ? > T (y) ? A(?)
(10)
A member of an exponential family remains in the same family when raised to the power of
k(x, xi ). Because every exponential family has a conjugate prior, we may choose our point-wise
prior p(? ? |x? ) to be conjugate to our chosen likelihood. We denote this conjugate prior p? (?, ?),
where ? and ? are hyperparameters.
p(?|x? ) = p? (?(x? ), ?(x? )) = f (?(x? ), ?(x? )) exp (? ? ?(x? ) ? ?(x? )A(?))
(11)
Therefore, our posterior as defined by equation (9) may be evaluated in closed form.
N
N
X
X
p(? ? |x? , D) = p? (
k(x? , xi )T (y i ) + ?(x? ),
k(x? , xi ) + ?(x? ))
i=1
The prior predictive distribution p(y|x) is given by
Z
p(y|x) = p(y|?)p? (?|?(x? ), ?(x? ))
= h(y)
(12)
i=1
f (?(x? ), ?(x? ))
f (?(x? ) + T (y), ?(x? ) + 1)
4
(13)
(14)
and the posterior predictive distribution is
PN
PN
f ( i=1 k(x? , xi )T (y i ) + ?(x? ), i=1 k(x? , xi ) + ?(x? ))
p(y|x? , D) = h(y) PN
PN
f ( i=1 k(x? , xi )T (y i ) + ?(x? ) + T (y), i=1 k(x? , xi ) + ?(x? ) + 1)
(15)
This is a general formulation of the posterior distribution over the parameters of any likelihood
model belonging to the exponential family. Note that given a function k(x? , x), we may evaluate this
posterior without sampling, in time linear in the number of samples. Moreover, for several choices
of kernels the relevant sums can be evaluates in sub-linear time; a sum over squared exponential
kernels, for instance, can be evaluated in logarithmic time.
3
Local inference for multivariate Gaussian
We now discuss in detail the application of equation (12) to the case of a multivariate Gaussian
likelihood model with unknown mean ? and unknown covariance ?.
p(y|?, ?) = N (y; ?, ?)
(16)
We present the conjugate prior, posterior, and predictive distributions without derivation; see [17],
for example, for a derivation. The conjugate prior for a multivariate Gaussian with unknown mean
and covariance is the normal-inverse Wishart distribution, with hyperparameter functions ?0 (x? ),
?(x? ), ?(x? ), and ?(x? ).
?
?
?
p(?, ?|x ) = N ?; ?0 (x ),
? W ?1 (?; ?(x? ), ?(x? ))
(17)
?(x? )
The hyperparameter functions have intuitive interpretations; ?0 (x? ) is our initial belief of the mean
function, while ?(x? ) is our confidence in that belief, with ?(x? ) = 0 indicating no confidence
in the region near x? , and ?(x? ) ? ? indicating a state of perfect knowledge. Likewise, ?(x? )
indicates the expected covariance, and ?(x? ) represents the confidence in that estimate, much like
?. Given a dataset D, we can compute a posterior over the mean and covariance, represented by
updated parameters ?00 (x? ), ?0 (x? ), ?0 (x? ), and ? 0 (x? ).
?0 (x? ) = ?(x? ) + k(x? )
?(x? )?0 (x? ) + y
?00 (x? ) =
?(x? ) + k(x? )
?0 (x? ) = ?(x? ) + S(x? ) +
? 0 (x? ) = ?(x? ) + k(x? )
(18)
?
?
?(x )k(x )
E(x? )
?(x? ) + k(x? )
where
k(x? ) =
N
X
N
k(x? , xi )
y(x? ) =
i=1
S(x? ) =
N
X
1 X
k(x? , xi )y i
k(x? ) i=1
>
k(x? , xi ) y i ? y(x? ) y i ? y(x? )
(19)
i=1
>
E(x? ) = y(x? ) ? ?0 (x? ) y(x? ) ? ?0 (x? )
The resulting posterior predictive distribution is a multivariate Student-t distribution.
?0 (x? ) + 1
0
?
0
?
?
p(y|x ) = t? 0 (x? ) ?0 (x ), 0 ? 0 ? ? (x )
? (x )? (x )
3.1
(20)
Special cases
Two special cases of the multivariate Gaussian are worth mentioning. First, a fixed, known co?
)
variance ?(x? ) can be described by the hyperparameters ?(x? ) = lim??? ?(x
? . The resulting
posterior distribution is then
1
?
0
?
p(?|x , D) = N ?0 , 0 ? ?(x )
(21)
? (x )
5
with predictive distribution
p(?|x? , D) = N
1 + ?0 (x? )
?
?00 ,
?(x
)
?0 (x? )
(22)
In the limit as ? goes to 0, when the prior is uninformative, the mean and mode of the predictive
distribution approaches the Nadaraya-Watson [8, 9] estimate.
PN
k(x? , xi )yi
?
(23)
?N W (x ) = Pi=1
N
?
i=1 k(x , xi )
The complementary case of known mean ?(x? ) and unknown covariance ?(x? ) is described by the
limit ? ? ?. In this case, the posterior distribution is
N
N
X
X
p(?|x? , D) = W ?1 ?(x? ) +
ki (y i ? ?(x? ))(y i ? ?(x? ))> , ?(x? ) +
ki
(24)
i=1
i=1
with predictive distribution
?
p(y|x ) = t? 0 (x? )
N
X
1
ki (y i ? ?(x? ))(y i ? ?(x? ))>
?(x ), 0 ? ?(x? ) +
? (x )
i=1
?
!
(25)
In the limit as ? goes to 0, the maximum likelihood covariance estimate is
?ML (x? ) =
N
X
ki (y i ? ?(x? ))(y i ? ?(x? ))>
(26)
i=1
which is precisely the result of our prior work [18, 19]. In both cases, our method yields distributions
over parameters, rather than point estimates; moreover, the use of Bayesian inference naturally
handles the case of limited or no available samples.
3.2
Experimental results
We evaluate our approach on several regression problems, and compare the results with alternative nonparametric Bayesian models. In all experiments, we use the squared-exponential kernel
k(y, y 0 ) = exp( 2c ky ? y 0 k2 ). This function meets both the requirements of our algorithm and is
positive-definite and thus a suitable covariance function for models based on the Gaussian process.
We set the kernel scale c by maximum likelihood for each model.
We compare our approach to covariance prediction to the generalized Wishart process (GWP) of [7].
First, we sample a synthetic dataset; the output is a two-dimensional observation set Y = R2 , where
samples are drawn from a zero-mean normal distribution with a covariance that rotates over time.
>
cos(t) ? sin(t)
4 0
cos(t) ? sin(t)
?(t) =
(27)
0 10
sin(t) cos(t)
sin(t) cos(t)
Second, we predict the covariances of the returns on two currency exchanges?the Euro to US dollar,
and the Japanese yen to US dollar?over the past four years. Following Wilson and Ghahramani, we
), where Pt is the exchange rate on day t. Illustrative results are provided
define a return as log( PPt+1
t
in figure 2. To compare these results quantitatively, one natural measure is the mean of the logarithm
of the likelihood of the predicted model given the data.
MLL =
N
1 X 1 > ? ?1
? i)
? (y ? y + log det ?
N i=1 2 i i i
(28)
? i is the maximum likelihood covariance predicted for the ith sample.
Here, ?
In addition to how well our model describes the available data, we may also be interested in how
accurately we recover the distribution used to generate the data. This is a measure of how closely
the inferred ellipses in figure 2 approximate the true covariance ellipses. One measure of the quality
of the inferred distribution is the KL divergence of the inferred distribution from the true distribution
6
?=0.83
?=1.20
?=1.57
?=1.94
?=2.31
5
5
5
5
5
0
?5
?10
?10
0
?5
?5
0
x1
5
10
?10
?10
0
?5
?5
0
5
x1
Ground
Truth
10
?10
?10
x2
10
x2
10
x2
10
x2
10
x2
10
0
?5
?5
0
x1
Inverse Wishart
5
10
?10
?10
0
?5
?5
0
5
x
Generalised Wishart 1Process
10
?10
?10
?5
0
x1
5
10
(a) Synthetic periodic data
0.05
0.05
JPY/USD
2014/1/10
0.05
JPY/USD
2013/3/22
0.05
JPY/USD
2012/6/4
0.05
JPY/USD
2011/8/17
JPY/USD
2010/10/29
0
0
?0.05
?0.05
?0.05?0.025 0 0.025 0.05 ?0.05?0.025
EUR/USD
0
0
?0.05
0.025 0.05 ?0.05?0.025
0
Inverse Wishart
0
?0.05
0.025 0.05 ?0.05?0.025
Generalised Wishart Process
0
0
?0.05
0.025 0.05 ?0.05?0.025 0 0.025 0.05
EUR/USD
(b) Exchange data
Figure 2: Comparison of covariances predicted by our kernel inverse Wishart process and the generalized
Wishart process for the problems described in section 3.2. The true covariance used to generate data is provided
for comparison. The samples used are plotted so that the area of the circle is proportional to the weight assigned
by the kernel. The kernel inverse Wishart process outperforms the generalized Wishart process, both in terms
of the likelihood of the training data, and in terms of the divergence of the inferred distribution from the true
distribution.
used to generate the data. Note we cannot evaluate this quantity on the exchange dataset, as we do
not know the true distribution. We present both the mean likelihood and the KL divergence of both
algorithms, along with running times, in table 1.
By both metrics, our algorithm outperforms the GWP by a significant margin; the running time
advantage of kernel estimation over the GWP is even more dramatic. It is important to note that
running times are difficult to compare, as they depend heavily on implementation and hardware
details; the numbers reported should be considered qualitatively. Both algorithms were implemented
in the MATLAB programming language, with the likelihood functions for the GWP implemented in
heavily optimized c code in an effort to ensure a fair competition. Despite this, the GWP took over
a thousand times longer than our method to generate predictions.
Periodic
Exchange
kNIW
GWP
kNIW
GWP
ttr (s)
0.022
7.08
0.520
15.7
tev (ms)
0.003
0.135
0.020
1.708
MLL
-10.43
-19.79
7.73
7.56
DKL (?
pkp)
0.0138
0.0248
?
?
Table 1: Comparison of the performance of two models of covariance prediction, based on time required to
make predictions at evaluation, the mean log likelihood and the KL divergence between the predicted covariance
and the ground truth covariance.
We next evaluate our approach on heteroscedastic regression problems. First, we generate 100
samples from the distribution described by Yuan and Wahba [20], which has mean ?(x? ) =
2 exp(?30(x? ? 0.25)2 ) + sin(?(x? )2 ) and variance ? 2 (x? ) = exp(2 ? sin(2?x? )). Second,
we test on the motorcycle dataset of Silverman et al. [21]. We compare our approach to a variety
of Gaussian process based regression algorithms, including a standard homoscedastic Gaussian process, the variational heteroscedastic Gaussian process of L?azaro-Gredilla and Titsias [4], and the
maximum likelihood heteroscedastic Gaussian process of Quadrianto et al. [22]. All algorithms are
implemented in MATLAB, using the authors? own code. Running times are presented with the same
caveat as in the previous experiments, and a similar conclusion holds: our method provides results
which are as good or better than methods based upon the Gaussian process, and does so in a fraction
of the time. Figure 3 illustrates the predictions made by our method on the heteroscedastic motor7
100
100
50
50
0
a
a
0
?50
?50
?100
?100
?150
10
15
20
25
30
35
40
45
?150
50
10
15
20
25
t
30
35
40
45
50
t
(a) kNIW
(b) VHGP
Figure 3: Comparison of the distributions inferred using the kernel normal inverse Wishart process and the
variational heteroscedastic Gaussian process to model Silverman?s motorcycle dataset. Both models capture
the time-varying nature of the measurement noise; as is typical, the kernel model is much less smooth and has
more local structure than the Gaussian process model. Both models perform well according to most metrics,
but the kernel model can be computed in a fraction of the time.
cycle dataset of Silverman. For reference, we provide the distribution generated by the variational
heteroscedastic Gaussian process.
Motorcycle
Periodic
kNIW
GP
VHGP
MLHGP
kNIW
GP
VHGP
MLHGP
ttr (s)
0.124
0.52
3.12
2.39
0.68
3.41
26.4
38.3
tev (ms)
2.95
3.52
7.53
5.83
7.94
22
54.4
29.1
NMSE
0.2
0.202
0.202
0.204
0.0708
0.0822
0.0827
0.0827
MLL
-4.04
-4.51
-4.07
-4.03
-2.07
-2.56
-1.85
-2.38
Table 2: Comparison of the performance of various models of heteroscedastic processes, based on time required to train, time required to make predictions at evaluation, the normalized mean squared error, and the
mean log likelihood. Note how the normal-inverse Wishart process obtains performance as good or better than
the other algorithms in a fraction of the time.
4
Discussion
We have presented a family of stochastic models which permit exact inference for any likelihood
function from the exponential family. Algorithms for performing inference on this model include
many local kernel estimators, and extend them to probabilistic contexts. We showed the instantiation
of our model for a multivariate Gaussian likelihood; due to lack of space, we do not present others,
but the approach is easily extended to tasks like classification and counting. The models we develop
are built on a strong assumption of independence; this assumption is critical to enabling efficient
exact inference. We now explore the costs of this assumption, and when it is inappropriate.
First, while the kernel function in our model does not need to be positive definite?or even
symmetric?we lose an important degree of flexibility relative to the covariance functions employed
in a Gaussian process. Covariance functions can express a number of complex concepts, such as a
prior over functions with a specified additive or hierarchical structure [23]; these concepts cannot
be easily formulated in terms of smoothness. Second, by neglecting the relationships between latent
parameters, we lose the ability to extrapolate trends in the data, meaning that in places where data
is sparse we cannot expect good performance. Thus, for a problem like time series forecasting, our
approach will likely be unsuccessful. Our approach is suitable in situations where we are likely to
see similar inputs many times, which is often the case. Moreover, regardless of the family of models
used, extrapolation to regions of sparse data can perform very poorly if the prior does not model the
true process well. Our approach is particularly effective when data is readily available, but computation is expensive; the gains in efficiency due to an independence assumption allow us to scale to
larger much larger datasets, improving predictive performance with less design effort.
Acknowledgements
This research was funded by the Office of Naval Research under contracts N00014-09-1-1052 and
N00014-10-1-0936. The support of Behzad Kamgar-Parsi and Tom McKenna is gratefully acknowledged.
8
References
[1] C. E. Rasmussen and C. Williams, Gaussian processes for machine learning. Cambridge,
MA: MIT Press, Apr. 2006, vol. 14, no. 2.
[2] Q. Le, A. Smola, and S. Canu, ?Heteroscedastic Gaussian process regression,? in Proc. ICML,
2005, pp. 489?496.
[3] K. Kersting, C. Plagemann, P. Pfaff, and W. Burgard, ?Most-Likely Heteroscedastic Gaussian
Process Regression,? in Proc. ICML, Corvallis, OR, USA, June 2007, pp. 393?400.
[4] M. L?azaro-Gredilla and M. Titsias, ?Variational heteroscedastic Gaussian process regression,?
in Proc. ICML, 2011.
[5] L. Shang and A. B. Chan, ?On approximate inference for generalized gaussian process models,? arXiv preprint arXiv:1311.6371, 2013.
[6] J. Qui?nonero-Candela and C. Rasmussen, ?A unifying view of sparse approximate Gaussian
process regression,? The Journal of Machine Learning Research, vol. 6, pp. 1939?1959, 2005.
[7] A. Wilson and Z. Ghahramani, ?Generalised Wishart processes,? in Proc. UAI, 2011, pp. 736?
744.
[8] E. Nadaraya, ?On estimating regression,? Theory of Probability & Its Applications, vol. 9,
no. 1, pp. 141?143, 1964.
[9] G. Watson, ?Smooth regression analysis,? Sankya: The Indian Journal of Statistics, Series A,
vol. 26, no. 4, pp. 359?372, 1964.
[10] R. Tibshirani and T. Hastie, ?Local likelihood estimation,? Journal of the American Statistical
Association, vol. 82, no. 398, pp. 559?567, 1987.
[11] A. G. Gray and A. W. Moore, ?N-body?problems in statistical learning,? in NIPS, vol. 4. Citeseer, 2000, pp. 521?527.
[12] C. Yang, R. Duraiswami, N. A. Gumerov, and L. Davis, ?Improved fast gauss transform and
efficient kernel density estimation,? in Computer Vision, 2003. Proceedings. Ninth IEEE International Conference on. IEEE, 2003, pp. 664?671.
[13] E. Snelson, ?Flexible and efficient Gaussian process models for machine learning,? PhD thesis,
University of London, 2007.
[14] L. Gyorfi, M. Kohler, A. Krzyzak, and H. Walk, A Distribution Free Theory of Nonparametric
Regression. New York, NY: Springer, 2002.
[15] S. Wang, ?Maximum weighted likelihood estimation,? PhD thesis, University of British
Columbia, 2001.
[16] W. S. Cleveland, ?Robust locally weighted regression and smoothing scatterplots,? Journal of
the American statistical association, vol. 74, no. 368, pp. 829?836, 1979.
[17] K. Murphy, ?Conjugate Bayesian analysis of the Gaussian distribution,? 2007.
[18] W. Vega-Brown, ?Predictive Parameter Estimation for Bayesian Filtering,? SM Thesis, Massachusetts Institute of Technology, 2013.
[19] W. Vega-Brown and N. Roy, ?CELLO-EM: Adaptive Sensor Models without Ground Truth,?
in Proc. IROS, Tokyo, Japan, 2013.
[20] M. Yuan and G. Wahba, ?Doubly penalized likelihood estimator in heteroscedastic regression,?
Statistics & probability letters, vol. 69, no. 1, pp. 11?20, 2004.
[21] B. W. Silverman et al., ?Some aspects of the spline smoothing approach to non-parametric
regression curve fitting,? Journal of the Royal Statistical Society, Series B, vol. 47, no. 1, pp.
1?52, 1985.
[22] N. Quadrianto, K. Kersting, M. Reid, T. Caetano, and W. Buntine, ?Most-Likely Heteroscedastic Gaussian Process Regression,? in Proc. ICDM, Miami, FL, USA, December 2009.
[23] D. Duvenaud, H. Nickisch, and C. E. Rasmussen, ?Additive Gaussian processes,? in Advances
in Neural Information Processing Systems 24, Granada, Spain, 2011, pp. 226?234.
9
| 5371 |@word covariance:23 citeseer:1 dramatic:1 initial:1 series:3 past:1 outperforms:2 must:3 readily:1 additive:2 enables:1 generative:6 ith:1 caveat:1 provides:2 node:1 along:1 yuan:2 prove:1 doubly:1 fitting:1 x0:5 expected:1 nor:1 inappropriate:1 provided:3 cleveland:2 underlying:2 notation:1 bounded:2 moreover:3 estimating:1 spain:1 developed:2 sparsification:3 transformation:1 every:1 multidimensional:1 exactly:1 k2:1 yn:2 reid:1 positive:4 generalised:3 local:9 limit:3 despite:1 meet:1 approximately:2 pkp:1 equivalence:1 specifying:4 shaded:1 heteroscedastic:12 co:5 mentioning:1 limited:2 nadaraya:2 gyorfi:1 practical:1 practice:1 implement:1 definite:3 silverman:4 procedure:1 area:1 convenient:2 word:1 confidence:3 convenience:1 cannot:3 selection:1 context:5 deterministic:1 graphically:2 straightforward:2 go:3 regardless:1 williams:1 formulate:1 rule:1 estimator:7 insight:1 handle:1 notion:1 limiting:1 updated:1 target:9 construction:1 pt:1 heavily:2 exact:9 programming:1 associate:1 trend:1 roy:2 satisfying:4 expensive:2 particularly:2 asymmetric:1 observed:1 preprint:1 wang:2 capture:2 thousand:1 region:2 connected:1 cycle:1 caetano:1 principled:1 ideally:1 depend:1 predictive:10 titsias:2 upon:1 efficiency:1 easily:2 joint:1 various:2 represented:1 derivation:2 train:1 fast:4 describe:2 effective:1 london:1 query:1 choosing:2 quite:1 larger:2 plagemann:1 relax:1 ability:1 statistic:3 gp:2 transform:2 noisy:1 sequence:1 advantage:1 behzad:1 took:1 interaction:3 product:1 relevant:1 nonero:2 combining:2 ramification:1 motorcycle:3 poorly:2 flexibility:1 intuitive:1 inducing:2 normalize:1 ky:1 quantifiable:1 competition:1 requirement:2 perfect:1 depending:1 develop:3 parsi:1 strong:2 implemented:3 predicted:4 implies:1 closely:2 tokyo:1 stochastic:6 enable:1 material:2 tev:2 require:3 exchange:5 extension:1 insert:1 hold:1 miami:1 considered:1 ground:3 normal:4 exp:5 duvenaud:1 mapping:1 predict:1 homoscedastic:1 estimation:8 proc:6 lose:3 bridge:1 correctness:1 weighted:5 mit:2 sensor:1 gaussian:35 always:1 rather:2 pn:5 kersting:2 varying:1 wilson:3 office:1 derived:1 june:1 naval:1 improvement:1 likelihood:44 indicates:2 sense:2 dollar:2 inference:38 dependent:1 typically:1 interested:1 arg:1 dual:1 among:1 classification:1 flexible:1 raised:1 special:2 smoothing:2 marginal:1 equal:1 sampling:1 represents:1 icml:3 minimized:1 others:1 spline:1 quantitatively:1 employ:1 ppt:1 divergence:9 mll:3 murphy:1 william:1 interest:1 highly:1 possibility:1 evaluation:2 accurate:1 neglecting:1 necessary:2 tree:1 logarithm:1 walk:1 circle:1 plotted:1 instance:3 cost:1 burgard:1 usefulness:1 successful:1 too:1 motivating:1 reported:1 buntine:1 varies:1 periodic:3 synthetic:2 nickisch:1 eur:2 cello:1 density:1 international:1 csail:1 probabilistic:2 contract:1 squared:3 thesis:3 choose:6 usd:7 wishart:14 american:2 return:3 japan:1 student:1 includes:1 satisfy:1 performed:1 view:1 extrapolation:1 closed:3 candela:2 doing:2 bayes:1 recover:2 yen:1 minimize:1 formed:1 variance:4 efficiently:1 likewise:1 correspond:2 yield:1 bayesian:12 accurately:2 worth:1 evaluates:1 pp:13 naturally:1 associated:2 gain:1 dataset:6 massachusetts:2 knowledge:4 lim:1 subtle:1 formalize:1 supervised:1 day:1 tom:1 specify:3 improved:2 duraiswami:1 formulation:1 done:1 evaluated:2 generality:2 stage:1 smola:1 expressive:1 lack:1 continuity:1 mode:1 quality:1 gray:1 facilitate:1 usa:2 validity:1 brown:3 requiring:1 true:6 normalized:1 concept:2 analytically:1 equality:1 assigned:1 symmetric:1 moore:1 conditionally:1 sin:6 uniquely:2 davis:1 illustrative:1 criterion:2 generalized:6 ttr:2 m:2 mlhgp:2 theoretic:1 demonstrate:3 performs:1 meaning:2 snelson:2 wise:1 vega:3 variational:4 common:1 attached:2 extend:1 interpretation:3 association:2 significant:1 measurement:1 corvallis:1 cambridge:2 smoothness:6 canu:1 language:1 gratefully:1 funded:1 specification:1 longer:1 multivariate:9 posterior:12 own:1 showed:1 perspective:1 chan:1 certain:3 n00014:2 arbitrarily:1 watson:2 yi:1 employed:1 full:1 currency:1 infer:4 smooth:5 academic:1 icdm:1 gkf:2 dkl:3 ellipsis:2 prediction:6 regression:17 essentially:1 metric:2 vision:1 arxiv:2 kernel:27 addition:2 want:1 uninformative:4 member:1 december:1 call:1 near:1 counting:1 yang:1 easy:1 variety:1 independence:4 hastie:2 wahba:2 pfaff:1 det:1 motivated:1 effort:2 forecasting:1 krzyzak:1 jpy:5 york:1 matlab:2 useful:1 clear:1 unimportant:1 detailed:1 amount:1 nonparametric:3 locally:2 hardware:1 reduced:1 generate:5 specifies:1 exist:1 shifted:1 tibshirani:2 hyperparameter:2 vol:9 express:2 sparsifying:1 four:1 acknowledged:1 drawn:2 neither:1 iros:1 fraction:3 sum:3 year:1 run:1 inverse:7 parameterized:1 powerful:1 letter:1 extends:1 family:13 reasonable:1 almost:1 place:1 draw:1 prefer:1 qui:2 bound:4 ki:4 fl:1 simplification:1 constraint:3 precisely:1 x2:11 aspect:1 speed:1 gwp:7 performing:1 gredilla:2 according:1 combination:1 conjugate:7 belonging:1 describes:1 em:1 wi:1 making:1 intuitively:1 restricted:1 principally:1 taken:1 computationally:1 equation:9 remains:1 discus:3 know:2 tractable:1 generalizes:2 available:3 permit:5 apply:1 observe:1 hierarchical:1 appropriate:2 nicholas:1 frequentist:1 alternative:2 assumes:1 running:4 ensure:1 include:1 unifying:1 ghahramani:3 build:1 approximating:1 society:1 objective:1 quantity:1 parametric:1 distance:1 rotates:1 evaluate:5 reason:1 code:2 pointwise:1 relationship:3 insufficient:1 providing:1 equivalently:2 difficult:3 relate:1 implementation:1 design:1 motivates:1 unknown:5 perform:9 gumerov:1 observation:11 datasets:2 sm:1 enabling:1 supporting:2 situation:1 extended:5 rn:1 ninth:1 inferred:7 required:3 specified:3 kl:3 optimized:1 distinction:1 nip:1 built:1 royal:1 max:1 including:1 marek:1 belief:2 power:1 suitable:2 critical:1 natural:1 unsuccessful:1 technology:2 historically:1 categorical:1 coupled:2 columbia:1 prior:29 literature:1 acknowledgement:1 determining:1 relative:3 marginalizing:2 fully:3 loss:2 expect:3 proportional:1 filtering:1 analogy:1 degree:1 sufficient:1 principle:1 granada:1 pi:1 summary:1 penalized:1 rasmussen:4 free:1 side:1 allow:3 institute:2 sparse:3 curve:1 xn:4 valid:1 rich:1 author:1 qualitatively:1 made:1 adaptive:1 simplified:1 approximate:7 obtains:1 ml:1 instantiation:2 uai:1 xi:21 latent:10 table:3 nature:1 robust:1 improving:1 complex:3 japanese:1 domain:1 apr:1 bounding:1 noise:3 hyperparameters:2 quadrianto:2 fair:1 complementary:1 x1:17 nmse:1 body:1 euro:1 ny:1 scatterplots:1 sub:1 inferring:2 exponential:12 weighting:1 theorem:1 british:1 mckenna:1 r2:1 intractable:1 phd:2 conditioned:1 illustrates:1 kx:1 margin:1 gap:1 entropy:6 depicted:2 logarithmic:2 smoothly:1 azaro:2 explore:1 likely:4 springer:1 truth:3 ma:2 conditional:1 formulated:3 lipschitz:1 absence:1 determined:2 typical:1 shang:1 gauss:2 experimental:1 indicating:2 select:1 support:1 maximumlikelihood:1 indian:1 kohler:1 extrapolate:1 |
4,829 | 5,372 | Fast Kernel Learning for Multidimensional Pattern
Extrapolation
Andrew Gordon Wilson?
CMU
Elad Gilboa?
WUSTL
Arye Nehorai
WUSTL
John P. Cunningham
Columbia
Abstract
The ability to automatically discover patterns and perform extrapolation is an essential quality of intelligent systems. Kernel methods, such as Gaussian processes,
have great potential for pattern extrapolation, since the kernel flexibly and interpretably controls the generalisation properties of these methods. However, automatically extrapolating large scale multidimensional patterns is in general difficult, and developing Gaussian process models for this purpose involves several
challenges. A vast majority of kernels, and kernel learning methods, currently
only succeed in smoothing and interpolation. This difficulty is compounded by
the fact that Gaussian processes are typically only tractable for small datasets, and
scaling an expressive kernel learning approach poses different challenges than
scaling a standard Gaussian process model. One faces additional computational
constraints, and the need to retain significant model structure for expressing the
rich information available in a large dataset. In this paper, we propose a Gaussian
process approach for large scale multidimensional pattern extrapolation. We recover sophisticated out of class kernels, perform texture extrapolation, inpainting,
and video extrapolation, and long range forecasting of land surface temperatures,
all on large multidimensional datasets, including a problem with 383,400 training
points. The proposed method significantly outperforms alternative scalable and
flexible Gaussian process methods, in speed and accuracy. Moreover, we show
that a distinct combination of expressive kernels, a fully non-parametric representation, and scalable inference which exploits existing model structure, are critical
for large scale multidimensional pattern extrapolation.
1
Introduction
Our ability to effortlessly extrapolate patterns is a hallmark of intelligent systems: even with large
missing regions in our field of view, we can see patterns and textures, and we can visualise in our
mind how they generalise across space. Indeed machine learning methods aim to automatically
learn and generalise representations to new situations. Kernel methods, such as Gaussian processes
(GPs), are popular machine learning approaches for non-linear regression and classification [1, 2, 3].
Flexibility is achieved through a kernel function, which implicitly represents an inner product of
arbitrarily many basis functions. The kernel interpretably controls the smoothness and generalisation
properties of a GP. A well chosen kernel leads to impressive empirical performances [2].
However, it is extremely difficult to perform large scale multidimensional pattern extrapolation with
kernel methods. In this context, the ability to learn a representation of the data entirely depends
on learning a kernel, which is a priori unknown. Moreover, kernel learning methods [4] are not
typically intended for automatic pattern extrapolation; these methods often involve hand crafting
combinations of Gaussian kernels (for smoothing and interpolation), for specific applications such
as modelling low dimensional structure in high dimensional data. Without human intervention,
the vast majority of existing GP models are unable to perform pattern discovery and extrapolation.
?
Authors contributed equally.
1
While recent approaches such as [5] enable extrapolation on small one dimensional datasets, it is
difficult to generalise these approaches for larger multidimensional situations. These difficulties
arise because Gaussian processes are computationally intractable on large scale data, and while
scalable approximate GP methods have been developed [6, 7, 8, 9, 10, 11, 12, 13], it is uncertain
how to best scale expressive kernel learning approaches. Furthermore, the need for flexible kernel
learning on large datasets is especially great, since such datasets often provide more information to
automatically learn an appropriate statistical representation.
In this paper, we introduce GPatt, a flexible, non-parametric, and computationally tractable approach
to kernel learning for multidimensional pattern extrapolation, with particular applicability to data
with grid structure, such as images, video, and spatial-temporal statistics. Specifically:
? We extend fast Kronecker-based GP inference (e.g., [14, 15]) to account for non-grid data. Our
experiments include data where more than 70% of the training data are not on a grid. Indeed most
applications where one would want to exploit Kronecker structure involve missing and non-grid
data ? caused by, e.g., water, government boundaries, missing pixels and image artifacts. By
adapting expressive spectral mixture kernels to the setting of multidimensional inputs and KroP +1
necker structure, we achieve exact inference and learning costs of O(P N P ) computations and
2
O(P N P ) storage, for N datapoints and P input dimensions, compared to the standard O(N 3 )
computations and O(N 2 ) storage associated with GPs.
? We show that i) spectral mixture kernels (adapted for Kronecker structure); ii) scalable inference based on Kronecker methods (adapted for incomplete grids); and, iii) truly non-parametric
representations, when used in combination (to form GPatt) distinctly enable large-scale multidimensional pattern extrapolation with GPs. We demonstrate this through a comparison with
various expressive models and inference techniques: i) spectral mixture kernels with arguably the
most popular scalable GP inference method (FITC) [10]; ii) a flexible and efficient recent spectral
based kernel learning method (SSGP) [6]; and, iii) the most popular GP kernels with Kronecker
based inference.
? The information capacity of non-parametric methods grows with the size of the data. A truly
non-parametric GP must have a kernel that is derived from an infinite basis function expansion.
We find that a truly non-parametric representation is necessary for pattern extrapolation on large
datasets, and provide insights into this surprising result.
? GPatt is highly scalable and accurate. This is the first time, as far as we are aware, that highly
expressive non-parametric kernels with in some cases hundreds of hyperparameters, on datasets
exceeding N = 105 training instances, can be learned from the marginal likelihood of a GP, in
only minutes. Such experiments show that one can, to some extent, solve kernel selection, and
automatically extract useful features from the data, on large datasets, using a special combination
of expressive kernels and scalable inference.
? We show the proposed methodology provides a distinct approach to texture extrapolation and inpainting; it was not previously known how to make GPs work for these fundamental applications.
? Moreover, unlike typical inpainting approaches, such as patch-based methods (which work by
recursively copying pixels or patches into a gap in an image, preserving neighbourhood similarities), GPatt is not restricted to spatial inpainting. This is demonstrated on a video extrapolation
example, for which standard inpainting methods would be inapplicable [16]. Similarly, we apply
GPatt to perform large-scale long range forecasting of land surface temperatures, through learning
a sophisticated correlation structure across space and time. This learned correlation structure also
provides insights into the underlying statistical properties of these data.
? We demonstrate that GPatt can precisely recover sophisticated out-of-class kernels automatically.
2
Spectral Mixture Product Kernels for Pattern Discovery
The spectral mixture kernel has recently been introduced [5] to offer a flexible kernel that can learn
any stationary kernel. By appealing to Bochner?s theorem [17] and building a scale mixture of A
Gaussian pairs in the spectral domain, [5] produced the spectral mixture kernel
kSM (? ) =
A
X
wa2 exp{?2? 2 ? 2 ?a2 } cos(2?? ?a ) ,
a=1
2
(1)
which they applied to one-dimensional input data with a small number of points. For tractability
with multidimensional inputs and large data, we propose a spectral mixture product (SMP) kernel:
kSMP (? |?) =
P
Y
kSM (?p |? p ) ,
(2)
p=1
P
where ?p is the pth component of ? = x ? x0 ? R , ? p are the hyperparameters {?a , ?a2 , wa2 }A
a=1 of
the pth spectral mixture kernel in the product of Eq. (2), and ? = {? p }P
are
the
hyperparameters
p=1
of the SMP kernel. The SMP kernel of Eq. (2) has Kronecker structure which we exploit for scalable
and exact inference in section 2.1. With enough components A, the SMP kernel of Eq. (2) can model
any stationary product kernel to arbitrary precision, and is flexible even with a small number of
components, since scale-location Gaussian mixture models can approximate many spectral densities.
We use SMP-A as shorthand for an SMP kernel with A components in each dimension (for a total
of 3P A kernel hyperparameters and 1 noise hyperparameter). Wilson [18, 19] contains detailed
discussions of spectral mixture kernels.
Critically, a GP with an SMP kernel is not a finite basis function method, but instead corresponds to
a finite (A component) mixture of infinite basis function expansions. Therefore such a GP is a truly
nonparametric method. This difference between a truly nonparametric representation ? namely a
mixture of infinite bases ? and a parametric kernel method, a finite basis expansion corresponding
to a degenerate GP, is critical both conceptually and practically, as our results will show.
2.1
Fast Exact Inference with Spectral Mixture Product Kernels
Gaussian process inference and learning requires evaluating (K +? 2 I)?1 y and log |K +? 2 I|, for an
N ? N covariance matrix K, a vector of N datapoints y, and noise variance ? 2 , as described in the
supplementary material. For this purpose, it is standard practice to take the Cholesky decomposition
of (K + ? 2 I) which requires O(N 3 ) computations and O(N 2 ) storage, for a dataset of size N .
However, many real world applications are engineered for grid structure, including spatial statistics,
sensor arrays, image analysis, and time sampling. [14] has shown that the Kronecker structure
2
in product kernels can be exploited for exact inference and hyperparameter learning in O(P N P )
P +1
storage and O(P N P ) operations, so long as the inputs x ? X are on a multidimensional grid,
meaning X = X1 ? ? ? ? ? XP ? RP . Details are in the supplement.
Here we relax this grid assumption. Assuming we have a dataset of M observations which are
not necessarily on a grid, we propose to form a complete grid using W imaginary observations,
yW ? N (f W , ?1 IW ), ? 0. The total observation vector y = [yM , yW ]> has N = M + W
entries: y = N (f , DN ), where the noise covariance matrix DN = diag(DM , ?1 IW ), DM =
? 2 IM . The imaginary observations yW have no corrupting effect on inference: the moments of
the resulting predictive distribution are exactly the same as for the standard predictive distribution,
namely lim?0 (KN + DN )?1 y = (KM + DM )?1 yM (proof in the supplement).
?1
For inference, we must evaluate (KN + DN ) y. Since DN is not a scaled identity (as is the usual
case in Kronecker methods), we cannot efficiently decompose KN + DN , but we can efficiently
take matrix vector products involving KN and DN . We therefore use preconditioned conjugate gra?1
dients (PCG) [20] to compute (KN + DN ) y, an iterative method involving only matrix vector
?1/2
products. We use the preconditioning matrix C = DN
to solve C > (KN + DN ) Cz = C > y.
The preconditioning matrix C speeds up convergence by ignoring the imaginary observations yW .
P +1
Exploiting the fast multiplication of Kronecker matrices, PCG takes O(JP N P ) total operations
?1
(where the number of iterations J N ) to compute (KN + DN ) y to convergence within machine precision (supplement). This procedure can also be used to handle heteroscedastic noise.
For learning (hyperparameter training) we must evaluate the marginal likelihood (supplement). We
cannot efficiently compute the log |KM + DM | complexity penalty in the marginal likelihood, because KM is not a Kronecker matrix. We approximate the complexity penalty as
M
M
X
X
2
?M + ?2 ) ,
log |KM + DM | =
log(?M
+
?
)
?
log(?
(3)
i
i
i=1
2
i=1
We approximate the eigenvalues ?M
i
for noise variance ? .
of KM using the eigenvalues of KN such
? M = M ?N for i = 1, . . . , M , which is particularly effective for large M (e.g. M > 1000)
that ?
i
N i
3
[7]. [21] proves this eigenvalue approximation is asymptotically consistent (e.g., converges in the
limit of large M ), and [22] shows how one can bound the true eigenvalues by their approximation
using PCA. Notably, only the log determinant (complexity penalty) term in the marginal likelihood
undergoes a small approximation, and inference remains exact.
All remaining terms in the marginal likelihood can be computed exactly and efficiently using PCG.
The total runtime cost of hyperparameter learning and exact inference with an incomplete grid is thus
P +1
O(P N P ). In image problems, for example, P = 2, and so the runtime complexity reduces to
1.5
O(N ). Although the proposed inference can handle non-grid data, this inference is most suited
to inputs where there is some grid structure ? images, video, spatial statistics, etc. If there is no
such grid structure (e.g., none of the training data fall onto a grid), then the computational expense
necessary to augment the data with imaginary grid observations can be prohibitive. Although incomplete grids have been briefly considered in, e.g. [23], such approaches generally involve costly
and numerically unstable rank 1 updates, inducing inputs, and separate (and restricted) treatments
of ?missing? and ?extra? data. Moreover, the marginal likelihood, critical for kernel learning, is not
typically considered in alternate approaches to incomplete grids.
3
Experiments
In our experiments we combine the SMP kernel of Eq. (2) with the fast exact inference and learning
procedures of section 2.1, in a GP method we henceforth call GPatt1,2 .
We contrast GPatt with many alternative Gaussian process kernel methods. We are particularly
interested in kernel methods, since they are considered to be general purpose regression methods, but
conventionally have difficulty with large scale multidimensional pattern extrapolation. Specifically,
we compare to the recent sparse spectrum Gaussian process regression (SSGP) [6] method, which
provides fast and flexible kernel learning. SSGP models the kernel spectrum (spectral density)
as a sum of point masses, such that SSGP is a finite basis function (parametric) model, with as
many basis functions as there are spectral point masses. SSGP is similar to the recent models of
Le et al. [8] and Rahimi and Recht [9], except it learns the locations of the point masses through
marginal likelihood optimization. We use the SSGP implementation provided by the authors at
http://www.tsc.uc3m.es/?miguel/downloads.php.
To further test the importance of the fast inference (section 2.1) used in GPatt, we compare to a GP
which uses the SMP kernel of section 2 but with the popular fast FITC [10, 24] inference, which
uses inducing inputs, and is implemented in GPML (http://www.gaussianprocess.org/
gpml). We also compare to GPs with the popular squared exponential (SE), rational quadratic (RQ)
and Mat?ern (MA) (with 3 degrees of freedom) kernels, catalogued in Rasmussen and Williams [1],
respectively for smooth, multi-scale, and finitely differentiable functions. Since GPs with these
kernels cannot scale to the large datasets we consider, we combine these kernels with the same fast
inference techniques that we use with GPatt, to enable a comparison.3 Moreover, we stress test each
of these methods in terms of speed and accuracy, as a function of available data and extrapolation
range, and number of components. All of our experiments contain a large percentage of non-grid
data, and we test accuracy and efficiency as a function of the percentage of missing data.
In all experiments we assume Gaussian noise, to express the marginal likelihood of the data p(y|?)
solely as a function of kernel hyperparameters ?. To learn ? we optimize the marginal likelihood
using BFGS. We use a simple initialisation scheme: any frequencies {?a } are drawn from a uniform
distribution from 0 to the Nyquist frequency (1/2 the sampling rate), length-scales {1/?a } from a
truncated Gaussian distribution, with mean proportional to the range of the data, and weights {wa }
are initialised as the empirical standard deviation of the data divided by the number of components
used in the model. In general, we find GPatt is robust to initialisation, particularly for N > 104
datapoints. We show a representative initialisation in the experiments.
This range of tests allows us to separately understand the effects of the SMP kernel, a non-parametric
representation, and the proposed inference methods of section 2.1; we will show that all are required
for good extrapolation performance.
1
We write GPatt-A when GPatt uses an SMP-A kernel.
Experiments were run on a 64bit PC, with 8GB RAM and a 2.8 GHz Intel i7 processor.
3
We also considered the model of [25], but this model is intractable for the datasets we considered and is
not structured for the fast inference of section 2.1.
2
4
3.1
Extrapolating Metal Tread Plate and Pores Patterns
We extrapolate the missing region, shown in Figure 1a, on a real metal tread plate texture. There
are 12675 training instances (Figure 1a), and 4225 test instances (Figure 1b). The inputs are pixel
locations x ? R2 (P = 2), and the outputs are pixel intensities. The full pattern is shown in Figure
1c. This texture contains shadows and subtle irregularities, no two identical diagonal markings, and
patterns that have correlations across both input dimensions.
(a) Train
(g) GP-SE
(k) Train
(b) Test
(h) GP-MA
(l) GPatt
(c) Full
(d) GPatt
(i) GP-RQ
(m) GP-MA
(e) SSGP
(f) FITC
(j) GPatt Initialisation
(n) Train
(o) GPatt
(p) GP-MA
Figure 1: (a)-(j): Extrapolation on a Metal Tread Plate Pattern. Missing data are shown in black. a)
Training region (12675 points), b) Testing region (4225 points), c) Full tread plate pattern, d) GPatt30, e) SSGP with 500 basis functions, f) FITC with 500 inducing (pseudo) inputs, and the SMP-30
kernel, and GPs with the fast exact inference in section 2.1, and g) squared exponential (SE), h)
Mat?ern (MA), and i) rational quadratic (RQ) kernels. j) Initial and learned hyperparameters using
GPatt using simple initialisation. During training, weights of extraneous components automatically
shrink to zero. (k)-(h) and (n)-(p): Extrapolation on tread plate and pore patterns, respectively, with
added artifacts and non-stationary lighting changes.
To reconstruct the missing and training regions, we use GPatt-30. The GPatt reconstruction shown
in Fig 1d is as plausible as the true full pattern shown in Fig 1c, and largely automatic. Without hand
crafting of kernel features to suit this image, exposure to similar images, or a sophisticated initialisation, GPatt has automatically discovered the underlying structure of this image, and extrapolated
that structure across a large missing region, even though the structure of this pattern is not independent across the two spatial input dimensions. Indeed the separability of the SMP kernel represents
only a soft prior assumption, and does not rule out posterior correlations between input dimensions.
The reconstruction in Figure 1e was produced with SSGP, using 500 basis functions. In principle
SSGP can model any spectral density (and thus any stationary kernel) with infinitely many components (basis functions). However, since these components are point masses (in frequency space),
each component has highly limited expressive power. Moreover, with many components SSGP experiences practical difficulties regarding initialisation, over-fitting, and computation time (scaling
quadratically with the number of basis functions). Although SSGP does discover some interesting
structure (a diagonal pattern), and has equal training and test performance, it is unable to capture
enough information for a convincing reconstruction, and we did not find that more basis functions
improved performance. Likewise, FITC with an SMP-30 kernel and 500 inducing (pseudo) inputs
cannot capture the necessary information to interpolate or extrapolate. On this example, FITC ran
for 2 days, and SSGP-500 for 1 hour, compared to GPatt which took under 5 minutes.
GPs with SE, MA, and RQ kernels are all truly Bayesian nonparametric models ? these kernels
are derived from infinite basis function expansions. Therefore, as seen in Figure 1 g), h), i), these
methods are completely able to capture the information in the training region; however, these kernels
do not have the proper structure to reasonably extrapolate across the missing region ? they simply
act as smoothing filters. Moreover, this comparison is only possible because we have implemented
these GPs using the fast exact inference techniques introduced in section 2.1.
5
(b) Accuracy Stress Test
?
50
1
k3
0.5
0
0
(a) Runtime Stress Test
1
k2
k1
1
0.5
0
0
?
50
True
Recovered
0.5
0
0
?
50
(c) Recovering Sophisticated Kernels
Figure 2: Stress Tests. a) Runtime Stress Test. We show the runtimes in seconds, as a function
of training instances, for evaluating the log marginal likelihood, and any relevant derivatives, for a
standard GP with SE kernel (as implemented in GPML), FITC with 500 inducing (pseudo) inputs
and SMP-25 and SMP-5 kernels, SSGP with 90 and 500 basis functions, and GPatt-100, GPatt-25,
and GPatt-5. Runtimes are for a 64bit PC, with 8GB RAM and a 2.8 GHz Intel i7 processor, on the
cone pattern (P = 2), shown in the supplement. The ratio of training inputs to the sum of imaginary
and training inputs for GPatt is 0.4 and 0.6 for the smallest two training sizes, and 0.7 for all other
training sets. b) Accuracy Stress Test. MSLL as a function of holesize on the metal pattern of
Figure 1. The values on the horizontal axis represent the fraction of missing (testing) data from
the full pattern (for comparison Fig 1a has 25% missing data). We compare GPatt-30 and GPatt-15
with GPs with SE, MA, and RQ kernels (and the inference of section 2.1), and SSGP with 100 basis
functions. The MSLL for GPatt-15 at a holesize of 0.01 is ?1.5886. c) Recovering Sophisticated
Kernels. A product of three kernels (shown in green) was used to generate a movie of 112,500
training points. From this data, GPatt-20 reconstructs these component kernels (the learned SMP-20
kernel is shown in blue). All kernels are a function of ? = x ? x0 and have been scaled by k(0).
Overall, these results indicate that both expressive nonparametric kernels, such as the SMP kernel,
and the specific fast inference in section 2.1, are needed to extrapolate patterns in these images.
We note that the SMP-30 kernel used with GPatt has more components than needed for this problem.
However, as shown in Fig. 1j, if the model is overspecified, the complexity penalty in the marginal
likelihood shrinks the weights ({wa } in Eq. (1)) of extraneous components, as a proxy for model
selection ? an effect similar to automatic relevance determination [26]. Components which do not
significantly contribute to model fit are automatically pruned, as shrinking the weights decreases the
eigenvalues of K and thus minimizes the complexity penalty (a sum of log eigenvalues). The simple
GPatt initialisation in Fig 1j is used in all experiments and is especially effective for N > 104 .
In Figure 1 (k)-(h) and (n)-(p) we use GPatt to extrapolate on treadplate and pore patterns with added
artifacts and lighting changes. GPatt still provides a convincing extrapolation ? able to uncover both
local and global structure. Alternative GPs with the inference of section 2.1 can interpolate small
artifacts quite accurately, but have trouble with larger missing regions.
3.2
Stress Tests and Recovering Complex 3D Kernels from Video
We stress test GPatt and alternative methods in terms of speed and accuracy, with varying datasizes, extrapolation ranges, basis functions, inducing (pseudo) inputs, and components. We assess
accuracy using standardised mean square error (SMSE) and mean standardized log loss (MSLL) (a
scaled negative log likelihood), as defined in Rasmussen and Williams [1] on page 23. Using the
empirical mean and variance to fit the data would give an SMSE and MSLL of 1 and 0 respectively.
Smaller SMSE and more negative MSLL values correspond to better fits of the data.
The runtime stress test in Figure 2a shows that the number of components used in GPatt does not
significantly affect runtime, and that GPatt is much faster than FITC (using 500 inducing inputs) and
SSGP (using 90 or 500 basis functions), even with 100 components (601 kernel hyperparameters).
The slope of each curve roughly indicates the asymptotic scaling of each method. In this experiment,
the standard GP (with SE kernel) has a slope of 2.9, which is close to the cubic scaling we expect. All
other curves have a slope of 1 ? 0.1, indicating linear scaling with the number of training instances.
However, FITC and SSGP are used here with a fixed number of inducing inputs and basis functions.
More inducing inputs and basis functions should be used when there are more training instances ?
and these methods scale quadratically with inducing inputs and basis functions for a fixed number
of training instances. GPatt, on the other hand, can scale linearly in runtime as a function of training
6
Table 1: We compare the test performance of GPatt-30 with SSGP (using 100 basis functions), and
GPs using SE, MA, and RQ kernels, combined with the inference of section 3.2, on patterns with a
train test split as in the metal treadplate pattern of Figure 1. We show the results as SMSE (MSLL).
train, test
GPatt
SSGP
SE
MA
RQ
Rubber mat
Tread plate
Pores
Wood
Chain mail
12675, 4225
12675, 4225
12675, 4225
14259, 4941
14101, 4779
0.31 (?0.57)
0.65 (?0.21)
0.97 (0.14)
0.86 (?0.069)
0.89 (0.039)
0.45 (?0.38)
1.06 (0.018)
0.90 (?0.10)
0.88 (?0.10)
0.90 (?0.10)
0.0038 (?2.8)
1.04 (?0.024)
0.89 (?0.21)
0.88 (?0.24)
0.88 (?0.048)
0.015 (?1.4)
0.19 (?0.80)
0.64 (1.6)
0.43 (1.6)
0.077 (0.77)
0.79 (?0.052)
1.1 (0.036)
1.1 (1.6)
0.99 (0.26)
0.97 (?0.0025)
size, without any deterioration in performance. Furthermore, the fixed 2-3 orders of magnitude
GPatt outperforms the alternatives is as practically important as asymptotic scaling.
The accuracy stress test in Figure 2b shows extrapolation (MSLL) performance on the metal tread
plate pattern of Figure 1c with varying holesizes, running from 0% to 60% missing data for testing
(for comparison the hole in Fig 1a has 25% missing data). GPs with SE, RQ, and MA kernels (and
the fast inference of section 2.1) all steadily increase in error as a function of holesize. Conversely,
SSGP does not increase in error as a function of holesize ? with finite basis functions SSGP cannot
extract as much information from larger datasets as the alternatives. GPatt performs well relative to
the other methods, even with a small number of components. GPatt is particularly able to exploit the
extra information in additional training instances: only when the holesize is so large that over 60%
of the data are missing does GPatt?s performance degrade to the same level as alternative methods.
In Table 1 we compare the test performance of GPatt with SSGP, and GPs using SE, MA, and RQ
kernels, for extrapolating five different patterns, with the same train test split as for the tread plate
pattern in Figure 1. All patterns are shown in the supplement. GPatt consistently has the lowest
SMSE and MSLL. Note that many of these datasets are sophisticated patterns, containing intricate
details which are not strictly periodic, such as lighting irregularities, metal impurities, etc. Indeed
SSGP has a periodic kernel (unlike the SMP kernel which is not strictly periodic), and is capable of
modelling multiple periodic components, but does not perform as well as GPatt on these examples.
We also consider a particularly large example, where we use GPatt-10 to perform learning and exact
inference on the Pores pattern, with 383,400 training points, to extrapolate a large missing region
with 96,600 test points. The SMSE is 0.077, and the total runtime was 2800 seconds. Images of the
successful extrapolation are shown in the supplement.
We end this section by showing that GPatt can accurately recover a wide range of kernels, even using
a small number of components. To test GPatt?s ability to recover ground truth kernels, we simulate
a 50 ? 50 ? 50 movie of data (e.g. two spatial input dimensions, one temporal) using a GP with
kernel k = k1 k2 k3 (each component kernel in this product operates on a different input dimension),
where k1 = kSE + kSE ? kPER , k2 = kMA ? kPER + kMA ? kPER , and k3 = (kRQ + kPER ) ? kPER + kSE .
(kPER (? ) = exp[?2 sin2 (? ? ?)/`2 ], ? = x ? x0 ). We use 5 consecutive 50 ? 50 slices for testing,
leaving a large number N = 112500 of training points, providing much information to learn the true
generating kernels. Moreover, GPatt-20 reconstructs these complex out of class kernels in under 10
minutes, as shown in Fig 2c. In the supplement, we show true and predicted frames from the movie.
3.3
Wallpaper and Scene Reconstruction and Long Range Temperature Forecasting
Although GPatt is a general purpose regression method, it can also be used for inpainting: image
restoration, object removal, etc. We first consider a wallpaper image stained by a black apple mark,
shown in Figure 3. To remove the stain, we apply a mask and then separate the image into its
three channels (red, green, and blue), resulting in 15047 pixels in each channel for training. In each
channel we ran GPatt using SMP-30. We then combined the results from each channel to restore the
image without any stain, which is impressive given the subtleties in the pattern and lighting.
In our next example, we wish to reconstruct a natural scene obscured by a prominent rooftop, shown
in the second row of Figure 3a). By applying a mask, and following the same procedure as for
the stain, this time with 32269 pixels in each channel for training, GPatt reconstructs the scene
without the rooftop. This reconstruction captures subtle details, such as waves, with only a single
7
(a) Inpainting
1
0.8
0.6
0.4
0.2
0.5
0
0
50
X [Km]
0
Correlations
Correlations
1
0.5
50
Y [Km]
0
0
0.8
0.6
0.4
0.2
0
50
100
Time [mon]
(b) Learned GPatt Kernel for Temperatures
0.8
0.6
0.4
0.2
20
40
X [Km]
0
0.8
0.6
0.4
0.2
50
Y [Km]
0
5
Time [mon]
(c) Learned GP-SE Kernel for Temperatures
Figure 3: a) Image inpainting with GPatt. From left to right: A mask is applied to the original image,
GPatt extrapolates the mask region in each of the three (red, blue, green) image channels, and the
results are joined to produce the restored image. Top row: Removing a stain (train: 15047 ? 3).
Bottom row: Removing a rooftop to restore a natural scene (train: 32269?3). We do not extrapolate
the coast. (b)-(c): Kernels learned for land surface temperatures using GPatt and GP-SE.
training image. In fact this example has been used with inpainting algorithms which were given
access to a repository of thousands of similar images [27]. The results emphasized that conventional
inpainting algorithms and GPatt have profoundly different objectives, which are sometimes even at
cross purposes: inpainting attempts to make the image look good to a human (e.g., the example in
[27] placed boats in the water), while GPatt is a general purpose regression algorithm, which simply
aims to make accurate predictions at test input locations, from training data alone. For example,
GPatt can naturally learn temporal correlations to make predictions in the video example of section
3.2, for which standard patch based inpainting methods would be inapplicable [16].
Similarly, we use GPatt to perform long range forecasting of land surface temperatures. After training on 108 months (9 years) of temperature data across North America (299,268 training points; a
71 ? 66 ? 108 completed grid, with missing data for water), we forecast 12 months (1 year) ahead
(33,252 testing points). The runtime was under 30 minutes. The learned kernels using GPatt and GPSE are shown in Figure 3 b) and c). The learned kernels for GPatt are highly non-standard ? both
quasi periodic and heavy tailed. These learned correlation patterns provide insights into features
(such as seasonal influences) which affect how temperatures vary in space and time. Indeed learning
the kernel allows us to discover fundamental properties of the data. The temperature forecasts using
GPatt and GP-SE, superimposed on maps of North America, are shown in the supplement.
4
Discussion
Large scale multidimensional pattern extrapolation problems are of fundamental importance in machine learning, where we wish to develop scalable models which can make impressive generalisations. However, there are many obstacles towards applying popular kernel methods, such as Gaussian processes, to these fundamental problems. We have shown that a combination of expressive
kernels, truly Bayesian nonparametric representations, and inference which exploits model structure, can distinctly enable a kernel approach to these problems. Moreover, there is much promise
in further exploring Bayesian nonparametric kernel methods for large scale pattern extrapolation.
Such methods can be extremely expressive, and expressive methods are most needed for large scale
problems, which provide relatively more information for automatically learning a rich statistical
representation of the data.
Acknowledgements AGW thanks ONR grant N000141410684 and NIH grant R01GM093156. JPC
thanks Simons Foundation grants SCGB #325171, #325233, and the Grossman Center at Columbia.
8
References
[1] C.E. Rasmussen and C.K.I. Williams. Gaussian processes for Machine Learning. The MIT Press, 2006.
[2] C.E. Rasmussen. Evaluation of Gaussian Processes and Other Methods for Non-linear Regression. PhD
thesis, University of Toronto, 1996.
[3] A. O?Hagan. Curve fitting and optimal design for prediction. Journal of the Royal Statistical Society, B
(40):1?42, 1978.
[4] M. G?onen and E. Alpayd?n. Multiple kernel learning algorithms. Journal of Machine Learning Research,
12:2211?2268, 2011.
[5] A.G. Wilson and R.P. Adams. Gaussian process kernels for pattern discovery and extrapolation. International Conference on Machine Learning, 2013.
[6] M. L?azaro-Gredilla, J. Qui?nonero-Candela, C.E. Rasmussen, and A.R. Figueiras-Vidal. Sparse spectrum
Gaussian process regression. Journal of Machine Learning Research, 11:1865?1881, 2010.
[7] C.K.I. Williams and M. Seeger. Using the Nystr?om method to speed up kernel machines. In Advances in
Neural Information Processing Systems, pages 682?688. MIT Press, 2001.
[8] Q. Le, T. Sarlos, and A. Smola. Fastfood-computing Hilbert space expansions in loglinear time. In
International Conference on Machine Learning, pages 244?252, 2013.
[9] A. Rahimi and B. Recht. Random features for large-scale kernel machines. In Neural Information Processing Systems, 2007.
[10] E. Snelson and Z. Ghahramani. Sparse gaussian processes using pseudo-inputs. In Advances in neural
information processing systems, volume 18, page 1257. MIT Press, 2006.
[11] J. Hensman, N. Fusi, and N.D. Lawrence. Gaussian processes for big data. In Uncertainty in Artificial
Intelligence (UAI). AUAI Press, 2013.
[12] M. Seeger, C.K.I. Williams, and N.D. Lawrence. Fast forward selection to speed up sparse Gaussian
process regression. In Workshop on AI and Statistics, volume 9, 2003.
[13] J. Qui?nonero-Candela and C.E. Rasmussen. A unifying view of sparse approximate Gaussian process
regression. The Journal of Machine Learning Research, 6:1939?1959, 2005.
[14] Y. Saatc?i. Scalable Inference for Structured Gaussian Process Models. PhD thesis, University of Cambridge, 2011.
[15] E. Gilboa, Y. Saatc?i, and J.P. Cunningham. Scaling multidimensional inference for structured Gaussian
processes. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2013.
[16] C. Guillemot and O. Le Meur. Image inpainting: Overview and recent advances. Signal Processing
Magazine, IEEE, 31(1):127?144, 2014.
[17] S. Bochner. Lectures on Fourier Integrals, volume 42. Princeton University Press, 1959.
[18] A.G. Wilson. A process over all stationary kernels. June, 2012. Technical Report, University of Cambridge. http://www.cs.cmu.edu/?andrewgw/spectralkernel.pdf.
[19] A.G. Wilson. Covariance Kernels for Fast Automatic Pattern Discovery and Extrapolation with Gaussian Processes. PhD thesis, University of Cambridge, 2014. URL http://www.cs.cmu.edu/
?andrewgw/andrewgwthesis.pdf.
[20] K.E. Atkinson. An introduction to numerical analysis. John Wiley & Sons, 2008.
[21] C.T.H. Baker. The numerical treatment of integral equations. 1977.
[22] C.K.I. Williams and J. Shawe-Taylor. The stability of kernel principal components analysis and its relation
to the process eigenspectrum. In Advances in Neural Information Processing Systems, volume 15, page
383. MIT Press, 2003.
[23] Y. Luo and R. Duraiswami. Fast near-grid Gaussian process regression. In International Conference on
Artificial Intelligence and Statistics, 2013.
[24] A. Naish-Guzman and S. Holden. The generalized FITC approximation. In Advances in Neural Information Processing Systems, pages 1057?1064, 2007.
[25] D. Duvenaud, J.R. Lloyd, R. Grosse, J.B. Tenenbaum, and Z. Ghahramani. Structure discovery in nonparametric regression through compositional kernel search. In International Conference on Machine
Learning, 2013.
[26] D.J.C MacKay. Bayesian nonlinear modeling for the prediction competition. Ashrae Transactions, 100
(2):1053?1062, 1994.
[27] J. Hays and A. Efros. Scene completion using millions of photographs. Communications of the ACM, 51
(10):87?94, 2008.
9
| 5372 |@word determinant:1 briefly:1 repository:1 km:9 covariance:3 decomposition:1 inpainting:13 nystr:1 recursively:1 moment:1 initial:1 contains:2 initialisation:8 outperforms:2 existing:2 imaginary:5 recovered:1 surprising:1 luo:1 must:3 john:2 numerical:2 remove:1 extrapolating:3 update:1 stationary:5 alone:1 prohibitive:1 intelligence:3 ksm:2 provides:4 contribute:1 location:4 toronto:1 org:1 five:1 dn:11 shorthand:1 combine:2 fitting:2 introduce:1 x0:3 mask:4 intricate:1 notably:1 indeed:5 roughly:1 multi:1 automatically:10 provided:1 discover:3 moreover:9 underlying:2 baker:1 mass:4 lowest:1 minimizes:1 r01gm093156:1 developed:1 wa2:2 temporal:3 pseudo:5 multidimensional:15 act:1 auai:1 runtime:9 exactly:2 scaled:3 k2:3 control:2 grant:3 intervention:1 arguably:1 local:1 limit:1 solely:1 interpolation:2 black:2 downloads:1 conversely:1 heteroscedastic:1 co:1 limited:1 range:9 practical:1 testing:5 practice:1 irregularity:2 procedure:3 empirical:3 significantly:3 adapting:1 wustl:2 cannot:5 onto:1 selection:3 close:1 storage:4 context:1 applying:2 influence:1 www:4 optimize:1 conventional:1 demonstrated:1 missing:18 map:1 center:1 williams:6 exposure:1 flexibly:1 sarlos:1 insight:3 rule:1 array:1 datapoints:3 stability:1 handle:2 magazine:1 exact:10 gps:14 us:3 stain:4 particularly:5 hagan:1 bottom:1 capture:4 thousand:1 region:11 decrease:1 ran:2 rq:9 complexity:6 nehorai:1 impurity:1 predictive:2 inapplicable:2 efficiency:1 basis:22 preconditioning:2 completely:1 various:1 america:2 train:8 distinct:2 fast:17 effective:2 artificial:2 quite:1 mon:2 elad:1 larger:3 solve:2 supplementary:1 relax:1 reconstruct:2 plausible:1 ability:4 statistic:5 gp:24 eigenvalue:6 differentiable:1 took:1 propose:3 reconstruction:5 product:11 relevant:1 nonero:2 interpretably:2 flexibility:1 achieve:1 degenerate:1 inducing:10 competition:1 figueiras:1 exploiting:1 convergence:2 produce:1 generating:1 adam:1 converges:1 object:1 andrew:1 develop:1 pose:1 miguel:1 completion:1 finitely:1 eq:5 implemented:3 recovering:3 involves:1 shadow:1 indicate:1 predicted:1 c:2 filter:1 human:2 engineered:1 enable:4 material:1 government:1 decompose:1 standardised:1 im:1 strictly:2 exploring:1 effortlessly:1 practically:2 considered:5 ground:1 duvenaud:1 exp:2 great:2 k3:3 lawrence:2 efros:1 vary:1 consecutive:1 a2:2 smallest:1 purpose:6 currently:1 iw:2 gaussianprocess:1 mit:4 sensor:1 gaussian:29 aim:2 varying:2 wilson:5 gpml:3 derived:2 june:1 seasonal:1 guillemot:1 consistently:1 rank:1 modelling:2 likelihood:12 indicates:1 superimposed:1 contrast:1 seeger:2 sin2:1 inference:35 typically:3 holden:1 cunningham:2 relation:1 quasi:1 interested:1 pixel:6 overall:1 classification:1 flexible:7 pcg:3 priori:1 augment:1 extraneous:2 smoothing:3 spatial:6 special:1 mackay:1 marginal:11 field:1 equal:1 aware:1 sampling:2 runtimes:2 identical:1 represents:2 look:1 report:1 guzman:1 gordon:1 intelligent:2 interpolate:2 intended:1 suit:1 attempt:1 freedom:1 highly:4 evaluation:1 mixture:14 truly:7 pc:2 chain:1 accurate:2 integral:2 capable:1 necessary:3 experience:1 incomplete:4 taylor:1 obscured:1 uncertain:1 ashrae:1 instance:8 soft:1 obstacle:1 modeling:1 restoration:1 applicability:1 cost:2 deviation:1 entry:1 tractability:1 hundred:1 uniform:1 successful:1 kn:8 periodic:5 combined:2 recht:2 density:3 fundamental:4 thanks:2 international:4 retain:1 pore:5 ym:2 squared:2 thesis:3 reconstructs:3 containing:1 henceforth:1 derivative:1 grossman:1 account:1 potential:1 bfgs:1 lloyd:1 north:2 caused:1 depends:1 view:2 extrapolation:29 candela:2 red:2 wave:1 recover:4 slope:3 simon:1 ass:1 square:1 php:1 accuracy:8 om:1 variance:3 largely:1 efficiently:4 likewise:1 correspond:1 krq:1 necker:1 conceptually:1 bayesian:4 accurately:2 produced:2 critically:1 none:1 lighting:4 apple:1 processor:2 meur:1 overspecified:1 frequency:3 initialised:1 steadily:1 dm:5 naturally:1 associated:1 proof:1 rational:2 arye:1 dataset:3 treatment:2 popular:6 lim:1 hilbert:1 subtle:2 sophisticated:7 uncover:1 uc3m:1 day:1 methodology:1 improved:1 duraiswami:1 shrink:2 though:1 furthermore:2 smola:1 correlation:8 hand:3 horizontal:1 expressive:12 nonlinear:1 saatc:2 undergoes:1 artifact:4 quality:1 grows:1 building:1 effect:3 contain:1 true:5 andrewgw:2 during:1 generalized:1 prominent:1 tsc:1 plate:8 stress:10 pdf:2 complete:1 demonstrate:2 performs:1 temperature:10 hallmark:1 image:23 meaning:1 coast:1 recently:1 snelson:1 nih:1 smp:21 overview:1 jp:1 volume:4 visualise:1 extend:1 million:1 numerically:1 significant:1 expressing:1 cambridge:3 ai:1 smoothness:1 automatic:4 grid:21 similarly:2 shawe:1 access:1 impressive:3 surface:4 similarity:1 etc:3 base:1 posterior:1 recent:5 hay:1 onr:1 arbitrarily:1 exploited:1 preserving:1 seen:1 additional:2 bochner:2 signal:1 ii:2 full:5 multiple:2 reduces:1 rahimi:2 compounded:1 smooth:1 faster:1 determination:1 technical:1 offer:1 long:5 cross:1 jpc:1 divided:1 equally:1 prediction:4 scalable:10 regression:11 involving:2 cmu:3 iteration:1 kernel:113 cz:1 represent:1 deterioration:1 achieved:1 sometimes:1 kma:2 want:1 separately:1 leaving:1 extra:2 wallpaper:2 unlike:2 call:1 near:1 iii:2 enough:2 split:2 naish:1 affect:2 fit:3 inner:1 regarding:1 i7:2 pca:1 gb:2 url:1 forecasting:4 nyquist:1 penalty:5 compositional:1 useful:1 generally:1 detailed:1 involve:3 yw:4 se:14 nonparametric:7 tenenbaum:1 http:4 generate:1 percentage:2 stained:1 blue:3 write:1 hyperparameter:4 mat:3 promise:1 profoundly:1 express:1 drawn:1 vast:2 asymptotically:1 ram:2 fraction:1 sum:3 cone:1 wood:1 run:1 year:2 uncertainty:1 gra:1 patch:3 fusi:1 scaling:8 qui:2 bit:2 entirely:1 bound:1 atkinson:1 quadratic:2 adapted:2 extrapolates:1 ahead:1 constraint:1 kronecker:10 precisely:1 scene:5 fourier:1 speed:6 simulate:1 extremely:2 pruned:1 relatively:1 ern:2 structured:3 developing:1 marking:1 alternate:1 gredilla:1 combination:5 conjugate:1 across:7 smaller:1 son:1 separability:1 appealing:1 restricted:2 computationally:2 rubber:1 equation:1 previously:1 remains:1 needed:3 mind:1 tractable:2 end:1 available:2 operation:2 vidal:1 apply:2 appropriate:1 spectral:16 neighbourhood:1 alternative:7 ssgp:23 rp:1 original:1 standardized:1 remaining:1 include:1 trouble:1 running:1 top:1 completed:1 unifying:1 exploit:5 k1:3 especially:2 prof:1 ghahramani:2 society:1 crafting:2 objective:1 added:2 restored:1 parametric:10 costly:1 usual:1 diagonal:2 loglinear:1 unable:2 separate:2 capacity:1 majority:2 degrade:1 mail:1 extent:1 unstable:1 eigenspectrum:1 water:3 preconditioned:1 assuming:1 alpayd:1 length:1 copying:1 ratio:1 convincing:2 providing:1 onen:1 difficult:3 expense:1 negative:2 implementation:1 design:1 proper:1 unknown:1 perform:8 contributed:1 observation:6 datasets:12 finite:5 kse:3 truncated:1 situation:2 communication:1 frame:1 discovered:1 arbitrary:1 intensity:1 introduced:2 pair:1 namely:2 required:1 learned:10 quadratically:2 hour:1 able:3 pattern:45 smse:6 challenge:2 including:2 green:3 video:6 royal:1 power:1 critical:3 difficulty:4 natural:2 restore:2 boat:1 msll:8 fitc:10 scheme:1 movie:3 axis:1 conventionally:1 extract:2 columbia:2 prior:1 discovery:5 removal:1 acknowledgement:1 multiplication:1 asymptotic:2 relative:1 fully:1 loss:1 expect:1 lecture:1 interesting:1 proportional:1 foundation:1 degree:1 metal:7 xp:1 consistent:1 proxy:1 principle:1 corrupting:1 land:4 heavy:1 row:3 extrapolated:1 placed:1 rasmussen:6 gilboa:2 understand:1 generalise:3 fall:1 wide:1 face:1 sparse:5 distinctly:2 ghz:2 slice:1 boundary:1 dimension:7 curve:3 evaluating:2 world:1 rich:2 hensman:1 author:2 forward:1 pth:2 far:1 transaction:2 approximate:5 implicitly:1 global:1 uai:1 scgb:1 spectrum:3 search:1 iterative:1 tailed:1 table:2 learn:7 reasonably:1 robust:1 channel:6 ignoring:1 expansion:5 necessarily:1 complex:2 domain:1 diag:1 did:1 fastfood:1 linearly:1 big:1 noise:6 arise:1 hyperparameters:7 x1:1 fig:7 representative:1 intel:2 cubic:1 grosse:1 wiley:1 precision:2 shrinking:1 exceeding:1 wish:2 exponential:2 learns:1 minute:4 theorem:1 removing:2 specific:2 emphasized:1 showing:1 r2:1 essential:1 intractable:2 workshop:1 importance:2 supplement:9 texture:5 magnitude:1 phd:3 hole:1 forecast:2 gap:1 suited:1 photograph:1 simply:2 azaro:1 infinitely:1 joined:1 subtlety:1 corresponds:1 truth:1 acm:1 ma:11 succeed:1 identity:1 month:2 towards:1 change:2 generalisation:3 specifically:2 infinite:4 typical:1 except:1 operates:1 principal:1 total:5 e:1 indicating:1 cholesky:1 mark:1 relevance:1 evaluate:2 princeton:1 extrapolate:8 |
4,830 | 5,373 | Mind the Nuisance: Gaussian Process
Classification using Privileged Noise
Daniel Hern?andez-Lobato
Universidad Aut?onoma de Madrid
Madrid, Spain
Viktoriia Sharmanska
IST Austria
Klosterneuburg, Austria
[email protected]
[email protected]
Kristian Kersting
TU Dortmund
Dortmund, Germany
Christoph H. Lampert
IST Austria
Klosterneuburg, Austria
[email protected]
[email protected]
Novi Quadrianto
SMiLe CLiNiC, University of Sussex
Brighton, United Kingdom
[email protected]
Abstract
The learning with privileged information setting has recently attracted a lot of attention within the machine learning community, as it allows the integration of additional knowledge into the training process of a classifier, even when this comes
in the form of a data modality that is not available at test time. Here, we show
that privileged information can naturally be treated as noise in the latent function
of a Gaussian process classifier (GPC). That is, in contrast to the standard GPC
setting, the latent function is not just a nuisance but a feature: it becomes a natural
measure of confidence about the training data by modulating the slope of the GPC
probit likelihood function. Extensive experiments on public datasets show that the
proposed GPC method using privileged noise, called GPC+, improves over a standard GPC without privileged knowledge, and also over the current state-of-the-art
SVM-based method, SVM+. Moreover, we show that advanced neural networks
and deep learning methods can be compressed as privileged information.
1
Introduction
Prior knowledge is a crucial component of any learning system as without a form of prior knowledge learning is provably impossible [1]. Many forms of integrating prior knowledge into machine
learning algorithms have been developed: as a preference of certain prediction functions over others,
as a Bayesian prior over parameters, or as additional information about the samples in the training
set used for learning a prediction function. In this work, we rely on the last of these setups, adopting
Vapnik and Vashist?s learning using privileged information (LUPI), see e.g. [2, 3]: we want to learn
a prediction function, e.g. a classifier, and in addition to the main data modality that is to be used for
prediction, the learning system has access to additional information about each training example.
This scenario has recently attracted considerable interest within the machine learning community
because it reflects well the increasingly relevant situation of learning as a service: an expert trains a
machine learning system for a specific task on request from a customer. Clearly, in order to achieve
the best result, the expert will use all the information available to him or her, not necessarily just the
1
information that the system itself will have access to during its operation after deployment. Typical
scenarios for learning as a service include visual inspection tasks, in which a classifier makes realtime decisions based on the input from its sensor, but at training time, additional sensors could be
made use of, and the processing time per training example plays less of a role. Similarly, a classifier
built into a robot or mobile device operates under strong energy constraints, while at training time,
energy is less of a problem, so additional data can be generated and made use of. A third scenario is
when the additional data is confidential, as e.g. in health care applications. Specifically, a diagnosis
system may be improved when more information is available at training time, e.g., specific blood
tests, genetic sequences, or drug trials, for the subjects that form the training set. However, the same
data may not be available at test time, as obtaining it could be impractical, unethical, or illegal.
We propose a novel method for using privileged information based on the framework of Gaussian
process classifiers (GPCs). The privileged data enters the model in form of a latent variable, which
modulates the noise term of the GPC. Because the noise is integrated out before obtaining the final
model, the privileged information is only required at training time, not at prediction time. The most
interesting aspect of the proposed model is that by this procedure, the influence of the privileged
information becomes very interpretable: its role is to model the confidence that the GPC has about
any training example, which can be directly read off from the slope of the probit likelihood. Instances
that are easy to classify by means of their privileged data cause a faster increasing probit, which
means the GP trusts the training example and tried to fit it well. Instances that are hard to classify
result in a slowly increasing slope, so that the GPC considers them less reliable and does not put a
lot of effort in fitting their label well. Our experiments on multiple datasets show that this procedure
leads not just to more interpretable models, but also to better prediction accuracy.
Related work: The LUPI framework was originally proposed by Vapnik and Vashist [2], inspired
by a thought-experiment: when training a soft-margin SVM, what if an oracle would provide us
with the optimal values of the slack variables? As it turns out, this would actually provably reduce
the amount of training data needed, and consequently, Vapnik and Vashist proposed the SVM+
classifier that uses privileged data to predict values for the slack variables, which led to improved
performance on several categorisation tasks and found applications, e.g., in finance [4]. This setup
was subsequently improved, by a faster training algorithm [5], better theoretical characterisation [3],
and it was generalised, e.g., to the learning to rank setting [6], clustering [7], metric learning [8] and
multi-class data classification [9]. Recently, however, it was shown that the main effect of the SVM+
procedure is to assign a data-dependent weight to each training example in the SVM objective [10].
The proposed method, GPC+, constitutes the first Bayesian treatment of classification using privileged information. The resulting privileged noise approach is related to input-modulated noise
commonly done in the regression task, where several Bayesian treatments of heteroscedastic regression using GPs have been proposed. Since the predictive density and marginal likelihood are no
longer analytically tractable, most works deal with approximate inference, i.e., techniques such as
Markov Chain Monte Carlo [11], maximum a posteriori [12], and variational Bayes [13]. To our
knowledge, however, there is no prior work on heteroscedastic classification using GPs ? we will
elaborate the reasons in Section 2.1 ? and this work is the first to develop approximate inference
based on expectation propagation for the heteroscedastic noise case in the context of classification.
2
GPC+: Gaussian process classification with privileged noise
For self-consistency we first review the GPC model [14] with an emphasis on the noise-corrupted
latent Gaussian process view. Then, we show how to treat privileged information as heteroscedastic
noise in this process. An elegant aspect of this view is how the privileged noise is able to distinguish
between easy and hard samples and to re-calibrate the uncertainty on the class label of each instance.
2.1
Gaussian process classifier with noisy latent process
Consider a set of N input-output data points or samples D = {(x1 , y1 ), . . . , (xN , yN )} ? Rd ?
{0, 1}. Assume that the class label yi of the sample xi has been generated as yi = I[ f?(xi ) ? 0 ],
where f?(?) is a noisy latent function and I[?] is the Iverson?s bracket notation, i.e., I[ P ] = 1 when
the condition P is true, and 0 otherwise. Induced by the label generation process, we adopt the
2
following form of likelihood function for ?
f = (f?(x1 ), . . . , f?(xN ))> :
YN
YN
Pr(y|?
f , X = (x1 , . . . , xN )> ) =
Pr(yn = 1|xn , f?) =
n=1
n=1
I[ f?(xn ) ? 0 ],
(1)
where f?(xn ) = f (xn ) + n with f (xn ) being the noise-free latent function. The noise term n
is assumed to be independent and normally distributed with zero mean and variance ? 2 , that is
n ? N (n |0, ? 2 ). To make inference about f?(xn ), we need to specify a prior over this function.
We proceed by imposing a zero mean Gaussian process prior [14] on the noise-free latent function,
that is f (xn ) ? GP(0, k(xn , ?)) where k(?, ?) is a positive-definite kernel function [15] that specifies
prior properties of f (?). A typical kernel function that allows for non-linear smooth functions is the
2
1
squared exponential kernel kf (xn , xm ) = ? exp(? 2l
kxn ? xm k`2 ), where ? controls the prior
amplitude of f (?) and l controls its prior smoothness. The prior and the likelihood are combined
using Bayes? rule to get the posterior of f?(?). Namely, Pr(?
f |X, y) = Pr(y|?
f , X)Pr(?
f )/Pr(y|X).
We can simplify the above noisy latent process view by integrating out the noise term n and writing
down the individual likelihood at sample xn in terms of the noise-free latent function f (?). Namely,
Z
Pr(yn = 1|xn , f ) = I[f?(xn ) ? 0]N (n |0, ? 2 )dn = ?(0,?2 ) (f (xn )),
(2)
where we have used that f?(xn ) = f (xn ) + n and ?(?,?2 ) (?) is a Gaussian cumulative distribution
function (CDF) with mean ? and variance ? 2 . Typically the standard Gaussian CDF is used, that is
?(0,1) (?), in the likelihood of (2). Coupled with a Gaussian process prior on the latent function f (?),
this results in the widely adopted noise-free latent Gaussian process view with probit likelihood.
The equivalence between a noise-free latent process with probit likelihood and a noisy latent process
with step-function likelihood is widely known [14]. It is also widely accepted that the function f?(?)
(or the functionf (?)) is a nuisance function as we do not observe its value and its sole purpose is
for a convenient formulation of the model [14]. However, in this paper, we show that by using
privileged information as the noise term, the latent function f? now plays a crucial role. The latent
function with privileged noise adjusts the slope transition in the Gaussian CDF to be faster or slower
corresponding to more certainty or more uncertainty about the samples in the original input space.
2.2
Introducing privileged information into the nuisance function
In the learning under privileged information (LUPI) paradigm [2], besides input data points
?
{x1 , . . . , xN } and associated labels {y1 , . . . , yN }, we are given additional information x?n ? Rd
about each training instance xn . However, this privileged information will not be available for unseen test instances. Our goal is to exploit the additional data x? to influence our choice of the latent
function f?(?). This needs to be done while making sure that the function does not directly use the
privileged data as input, as it is simply not available at test time. We achieve this naturally by treating
the privileged information as a heteroscedastic (input-dependent) noise in the latent process.
Our classification model with privileged noise is then as follows:
Likelihood model : Pr(yn = 1|xn , f?) = I[ f?(xn ) ? 0 ] ,
Assume : f?(xn ) = f (xn ) + n
where xn ? Rd
(3)
(4)
?
i.i.d.
Privileged noise model : n ? N (n |0, z(x?n ) = exp(g(x?n ))) , where x?n ? Rd
GP prior model : f (xn ) ? GP(0, kf (xn , ?)) and g(x?n ) ? GP(0, kg (x?n , ?)).
(5)
(6)
In the above, the function exp(?) is needed to ensure positivity of the noise variance. The term kg (?, ?)
is a positive-definite kernel function that specifies the prior properties of another latent function g(?),
which is evaluated in the privileged space x? . Crucially, the noise term n is now heteroscedastic,
that is, it has a different variance z(x?n ) at each input point xn . This is in contrast to the standard GPC
approach discussed in Section 2.1 where the noise term is homoscedastic, n ? N (n |0, z(x?n ) =
? 2 ). An input-dependent noise term is very common in regression tasks with continuous output
values yn ? R, resulting in heteroscedastic regression models, which have been proven more flexible
in numerous applications as already touched upon in the section on related work. However, to our
knowledge, there is no prior work on heteroscedastic classification models. This is not surprising as
the nuisance view of the latent function renders a flexible input-dependent noise point-less.
3
Posterior mean of
for an easy instance
= 1.0
= 5.0
0.98
= 0.5
0.8
exp(g(x?n ))
exp(g(x?n ))
exp(g(x?n ))
0.84
0.6
0.6
0.58
0.4
0.4
0.0
?10
0.2
0.2
?5
0
1
5
Posterior mean of
for a difficult instance
0.0
?(0,exp(g(x?n))(f (xn))
0.8
1.0
AwA (DeCAF) / Chimpanzee v. Giant Panda
1.0
10
f (xn)
?2
?1
0
1
2
Figure 1: Effects of privileged noise on the nuisance function. (Left) On synthetic data. Suppose for an input
xn , the latent function value is f (xn ) = 1. Now also assume that the associated privileged information x?n for
the n-th data point deems the sample as difficult, say exp(g(x?n )) = 5.0. Then the likelihood will reflect this
uncertainty Pr(yn = 1|f, g, xn , x?n ) = 0.58. In contrast, if the associated privileged information considers the
sample as easy, say e.g. exp(g(x?n )) = 0.5, the likelihood is very certain Pr(yn = 1|f, g, xn , x?n ) = 0.98.
(Right) On real data taken from our experiments in Sec. 4. The posterior means of the ?(?) function (solid)
and its 1-standard deviation confidence interval (dash-dot) for easy (blue) and difficult (black) instances of the
Chimpanzee v. Giant Panda binary task on the Animals with Attributes (AwA) dataset. (Best viewed in color).
In the context of privileged information heteroscedastic classification is a very sensible idea, which is
best illustrated when investigating the effect of privileged information in the equivalent formulation
of a noise free latent process, i.e., when one integrates out the privileged input-dependent noise term:
Z
Pr(yn = 1|xn , x?n , f, g) = I[ f?(xn ) ? 0 ]N (n |0, exp(g(x?n ))dn
p
= ?(0,exp(g(x?n ))) (f (xn )) = ?(0,1) (f (xn )/ exp(g(x?n )).
(7)
This equation shows that the privileged information adjusts the slope transition of the Gaussian
CDF through the latent function g(?). For difficult samples the latent function g(?) will be high,
the slope transition will be slower, and thus more uncertainty will be in the likelihood Pr(yn =
1|xn , x?n , f, g). For easy samples, however, g(?) will be low, the slope transition will be faster,
and thus less uncertainty will be in the likelihood term. This behaviour is illustrated in Figure 1.
For non-informative samples in the privileged space, the value of g for those samples should be
equal to a global noise value, as in a standard GPC. Thus, privileged information should in principle
never hurt. Proving this theoretically is, however, an interesting and challenging research direction.
Experimentally, however, we observe in the section on experiments the scenario described.
2.3
Posterior and prediction on test data
Define g = (g(x?1 ), . . . , g(x?n ))T and X? = (x?1 , . . . , x?n )T .
Given the likelihood
QN
?
?
Pr(y|X, X , f , g) = n=1 Pr(yn = 1|f, g, xn , xn ) with the individual term Pr(yn |f, g, xn , x?n )
given in (7) and the Gaussian process priors on functions, the posterior for f and g is:
Pr(f , g|y, X, X? ) =
Pr(y|X, X? , f , g)Pr(f )Pr(g)
,
Pr(y|X, X? )
(8)
where Pr(y|X, X? ) can be maximised with respect to a set of hyper-parameter values such as the
amplitude ? and the smoothness l of the kernel functions [14]. For a previously unseen test point
xnew ? Rd , the predictive distribution for its label ynew is given as:
Z
Pr(ynew = 1|y, X, X? ) = I[ f?(xnew ) ? 0 ]Pr(fnew |f )Pr(f , g|y, X, X? )df dgdfnew ,
(9)
where Pr(fnew |f ) is a Gaussian conditional distribution. We note that in (9) we do not consider the
privileged information x?new associated to xnew . The interpretation is that we consider homoscedastic
4
noise at test time. This is a reasonable approach as there is no additional information for increasing
or decreasing our confidence in the newly observed data xnew . Finally, we predict the label for a test
point via Bayesian decision theory: the label being predicted is the one with the largest probability.
3
Expectation propagation with numerical quadrature
Unfortunately, as for most interesting Bayesian models, inference in the GPC+ model is very challenging. Already in the homoscedastic case, the predictive density and marginal likelihood are
not tractable. Here, we therefore adapt Minka?s expectation propagation (EP) [16] with numerical
quadrature for approximate inference. Our choice is supported on the fact that EP is the preferred
method for approximate inference in GPCs, in terms of accuracy and computational cost [17, 18].
Consider the joint distribution of f , g and y, Pr(y|X, X? , f , g)Pr(f )Pr(g), where Pr(f ) and Pr(g)
QN
are Gaussian process priors and the likelihood Pr(y|X, X? , f , g) equals n=1 Pr(yn |xn , x?n , f, g),
with Pr(yn |xn , x?n , f, g) given by (7). EP approximates each non-normal factor in this distribution
by an un-normalised bi-variate normal distribution of f and g (we assume independence between f
and g). The only non-normal factors are those of the likelihood, which are approximated as:
Pr(yn |xn , x?n , f, g) ? ? n (f, g) = z n N (f (xn )|mf , v f )N (g(x?n )|mg , v g ) ,
(10)
where the parameters with the super-script are to be found by EP. The posterior approximation Q
computed by EP results from normalising with respect to f and g the EP approximate joint. That is,
Q is obtained by replacing each likelihood factor by the corresponding approximate factor ? n :
YN
Pr(f , g|X, X? , y) ? Q(f , g) := Z ?1 [
?(f, g)]Pr(f )Pr(g) ,
(11)
n=1
where Z is a normalisation constant that approximates the model evidence, Pr(y|X, X? ). The
normal distribution belongs to the exponential family of probability distributions and is closed under
the product and division. It is hence possible to show that Q is the product of two multi-variate
normals [19]. The first normal approximates the posterior for f and the second the posterior for g.
EP tries to fix the parameters of ? n so that it is similar to the exact factor Pr(yn |xn , x?n , f, g) in
regions of high posterior probability [16]. For this,
EP iteratively updates each ? n until convergence
to hminimise KLi Pr(yn |xn , x?n , f, g)Qold /Zn ||Q , where Qold is a normal distribution proportional
Q
?
to
n0 6=n ? n0 Pr(f )Pr(g) with all variables different from f (xn ) and g(xn ) marginalised out, Zn
is simply a normalisation constant and KL(?||?) denotes the Kullback-Leibler divergence between
probability distributions. Assume Qnew is the distribution minimising the previous divergence. Then,
? n ? Qnew /Qold and the parameter z n of ? n is fixed to guarantee that ? n integrates the same as
the exact factor with respect to Qold . The minimisation of the KL divergence involves matching
expected sufficient statistics (mean and variance) between Pr(yn |xn , x?n , f, g)Qold /Zn and Qnew .
These expectations can be obtained from the derivatives of log Zn with respect to the (natural)
parameters of Qold [19]. Unfortunately, the computation of log Zn in closed form is intractable. We
show here that it can be approximated by a one dimensional quadrature. Denote by mf , vf , mg and
vg the means and variances of Qold for f (xn ) and g(x?n ), respectively. Then,
Z
q
Zn = ?(0,1) yn mf / vf + exp(g(x?n )) N (g(x?n )|mg , vg )dg(x?n ) .
(12)
Thus, EP only requires five quadratures to update each ? n . One to compute log Zn and four extras
to compute its derivatives with respect to mf , vf , mg and vg . After convergence, Q can be used
to approximate predictive distributions and the normalisation constant Z can be maximised to find
good values for the model?s hyper-parameters. In particular, it is possible to compute the gradient
of Z with respect to the parameters of the Gaussian process priors for f and g [19]. An R language
implementation of GPC+ using EP for approximate inference is found in the supplementary material.
4
Experiments
We investigate the performance of GPC+. To this aim we considered three types of binary classification tasks corresponding to different privileged information using two real-world datasets: Attribute
Discovery and Animals with Attributes. We detail these experiments in turn in the following sections.
5
Methods: We compared our proposed GPC+ method with the well-established LUPI method based
on SVM, SVM+ [5]. As a reference, we also fit standard GP and SVM classifiers when learning on
the original space Rd (GPC and SVM baselines). For all four methods, we used a squared exponential
kernel with amplitude parameter ? and smoothness parameter l. For simplicity, we set ? = 1.0 in
all cases. There are two hyper-parameters in GPC (smoothness parameter l and noise variance ? 2 )
and also two in GPC+ (smoothness parameters l of kernel kf (?, ?) and of kernel kg (?, ?)). In GPC
and GPC+, we used type II-maximum likelihood for finding all hyper-parameters. SVM has two
knobs, i.e., smoothness and regularisation, and SVM+ has four knobs, two smoothness and two
regularisation parameters. In SVM we used a grid search guided by cross-validation to set all hyperparameters. However, this procedure was too expensive for finding the best parameters in SVM+.
Thus, we used the performance on a separate validation set to guide the search. This means that we
give a competitive advantage to SVM+ over the other methods, which do not use the validation set.
Evaluation metric: To evaluate the performance of each method we used the classification error
measured on an independent test set. We performed 100 repeats of all the experiments to get the
better statistics of the performance and we report the mean and the standard deviation of the error.
4.1
Attribute discovery dataset
The data set was collected from a website that aggregates product data from a variety of e-commerce
sources and includes both images and associated textual descriptions [20]. The images and texts are
grouped into 4 broad shopping categories: bags, earrings, ties, and shoes. We used 1800 samples
from this dataset. We generated 6 binary classification tasks for each pair of the 4 classes with 200
samples for training, 200 samples for validation, and the rest of the samples for testing performance.
Neural networks on texts as privileged information: We used images as the original domain and
texts as the privileged domain. This setting was also explored in [6]. However, we used a different
dataset because textual descriptions of the images used in [6] are sparse and contain duplicates. More
precisely, we extracted more advanced text features instead of simple term frequency (TF) features.
For the images representation, we extracted SURF descriptors [21] and constructed a codebook of
100 visual words using the k-means clustering. For the text representation, we extracted 200 dimensional continuous word-vectors using a neural network skip-gram architecture [22]1 . To convert this
word representation into a fixed-length sentence representation, we constructed a codebook of 100
word-vectors using again k-means clustering. We note that a more elaborate approach to transform
word to sentence or document features has recently been developed [23], and we are planning to
explore this in the future. We performed PCA for dimensionality reduction in the original and privileged domains and only kept the top 50 principal components. Finally, we standardised the data so
that each feature had zero mean and unit standard deviation.
The experimental results are summarised in Table 1. On average over 6 tasks, SVM with hinge loss
outperforms GPC with probit likelihood. However, GPC+ significantly improves over GPC providing the best results on average. This clearly shows that GPC+ is able to employ the neural network
textual representation as privileged information. In contrast, SVM+ produced the same result as
SVM. We suspect this is due to the fact that that SVM has already shown strong performance on
the original image space coupled with the difficulties of finding the best values of the four hyperparameters of SVM+. Keep in mind that in SVM+ we discretised the hyper-parameter search space
over 625 (5 ? 5 ? 5 ? 5) possible combination values and used a separate validation set to estimate
the resulting prediction performance.
4.2
Animals with attributes (AwA) dataset
The dataset was collected by querying image search engines for each of the 50 animals categories
which have complimentary high level descriptions of their semantic properties such as shape, colour,
or habitat information among others [24]. The semantic attributes per animal class were retrieved
from a prior psychological study. We focused on the 10 categories corresponding to the test set of this
dataset for which the predicted attributes are provided based on the probabilistic DAP model [24].
The 10 classes are: chimpanzee, giant panda, leopard, persian cat, pig, hippopotamus, humpback
whale, raccoon, rat, seal, which have 6180 images associated in total. As in Section 4.1 and also in
1
https://code.google.com/p/word2vec/
6
Table 1: Average error rate in % (the lower the better) on the Attribute Discovery dataset over 100 repetitions.
We used images as the original domain and neural networks word-vector representation on texts as the privileged domain. The best method for each binary task is highlighted in boldface. An average rank equal to one
means that the corresponding method has the smallest error on the 6 tasks.
bags v. earrings
bags v. ties
bags v. shoes
earrings v. ties
earrings v. shoes
ties v. shoes
average error on each task
average ranking
GPC
9.79?0.12
10.36?0.16
9.66?0.13
10.84?0.14
7.74?0.11
15.51?0.16
10.65?0.11
3.0
GPC+ (Ours)
9.50?0.11
10.03?0.15
9.22?0.11
10.56?0.13
7.33?0.10
15.54?0.16
10.36?0.12
1.8
SVM
9.89?0.14
9.44?0.16
9.31?0.12
11.15?0.16
7.75?0.13
14.90?0.21
10.41?0.11
2.7
SVM+
9.89?0.13
9.47?0.13
9.29?0.14
11.11?0.16
7.63?0.13
15.10?0.18
10.42?0.11
2.5
[6], we generated 45 binary classification tasks for each pair of the 10 classes with 200 samples for
training, 200 samples for validation, and the rest of samples for testing the predictive performance.
Neural networks on images as privileged information: Deep learning methods have gained an increased attention within the machine learning and computer vision community over the recent years.
This is due to their capability in extracting informative features and delivering strong predictive performance in many classification tasks. As such, we are interested to explore the use of deep learning
based features as privileged information so that their predictive power can be used even if we do not
have access to them at prediction time. We used the standard SURF features [21] with 2000 visual
words as the original domain and the recently proposed DeCAF features [25] extracted from the
activation of a deep convolutional network trained in a fully supervised fashion as the privileged domain. The DeCAF features have 4096 dimensions. All features are provided with the AwA dataset2 .
We again performed PCA for dimensionality reduction in the original and privileged domains and
only kept the top 50 principal components, as well as standardised the data.
Attributes as privileged information: Following the experimental setting of [6], we also used
images as the original domain and attributes as the privileged domain. Images were represented by
2000 visual words based on SURF descriptors and attributes were in the form of 85 dimensional
predicted attributes based on probabilistic binary classifiers [24]. As previously, we also performed
PCA and kept the top 50 principal components in the original domain and standardised the data.
The results of these experiments are shown in Figure 2 in terms of pairwise comparisons over 45
binary tasks between GPC+ and the main baselines, GPC and SVM+. The complete results with
the error of each method GPC, GPC+, SVM, and SVM+ on each problem are relegated to the
supplementary material. In contrast to the results on the attribute discovery dataset, on the AwA
dataset it is clear that GPC outperforms SVM in almost all of the 45 binary classification tasks
(see the supplementary material). The average error of GPC over 4500 (45 tasks and 100 repeats
per task) experiments is much lower than SVM. On the AwA dataset, SVM+ can take advantage
of privileged information ? be it deep belief DeCAF features or semantic attributes ? and shows
significant performance improvement over SVM. However, GPC+ still shows the best overall results
and further improves the already strong performance of GPC. As illustrated in Figure 1 (right), the
privileged information modulates the slope of the probit likelihood function differently for easy
and difficult examples: easy examples gain slope and hence importance whereas difficult ones lose
importance in the classification. In this dataset we analysed our experimental results using the
multiple dataset statistical comparison method described in [26]3 . The results of the statistical tests
are summarised in Figure 3. When DeCAF attributes are used as privileged information, there is
statistical evidence supporting that GPC+ performs best among the four methods, while when the
semantic attributes are used as privileged information, GPC+ still performs best but there is not
enough evidence to reject that GPC+ performs comparable to GPC.
2
http://attributes.kyb.tuebingen.mpg.de
Note that we are not able to use this method on the results of the attribute discovery dataset in Table 1
because the number of methods compared (i.e., 4) is almost equal to the number of tasks or datasets (i.e., 6).
3
7
(DeCAF as privileged)
(Attributes as privileged)
Figure 2: Pairwise comparison of the proposed GPC+ method and main baselines is shown via the relative
difference of the error rate (top: GPC+ versus GPC, bottom: GPC+ versus SVM+). The length of the 45 bars
corresponds to relative difference of the error rate over 45 cases. Average error rates of each method on the
AwA dataset across each of the 45 tasks are found in the supplementary material. (Best viewed in color).
Critical Distance
GPC
SVM+
GPC+
1
Critical Distance
SVM
2
3
GPC
SVM+
GPC+
SVM
1
4
(DeCAF as privileged)
2
3
4
(Attributes as privileged)
Figure 3: Average rank (the lower the better) of the four methods and critical distance for statistically significant differences (see [26]) on the AwA dataset. An average rank equal to one means that particular method
has the smallest error on the 45 tasks. Whenever the average ranks differ by more than the critical distance,
there is statistical evidence (p-value < 10%) supporting a difference in the average ranks and hence in the
performance. We also link two methods with a solid line if they are not statistically different from each other
(p-value > 10%). When the DeCAF features are used as privileged information, there is statistical evidence
supporting that GPC+ performs best among the four methods considered. When the attributes are used, GPC+
still performs best, but there is not enough evidence to reject that GPC+ performs comparable to GPC.
5
Conclusions and future work
We presented the first treatment of the learning with privileged information paradigm under the
Gaussian process classification (GPC) framework, and called it GPC+. In GPC+ privileged information is used in the latent noise layer, resulting in a data-dependent modulation of the slope of the
likelihood. The training time of GPC+ is about twice times the training time of a standard Gaussian
process classifier. The reason is that GPC+ must train two latent functions, f and g, instead of only
one. Nevertheless, our results show that GPC+ is an effective way to use privileged information,
which manifest itself in significantly better prediction accuracy. Furthermore, to our knowledge,
this is the first time that a heteroscedastic noise term is used to improve GPC. We have also shown
that recent advances in continuous word-vector neural networks representations [23] and deep convolutional networks for image representations [25] can be used as privileged information. For future
work, we plan to extend the GPC+ framework to the multi-class case and to speed up computation
by devising a quadrature-free expectation propagation method, similar to the ones in [27, 28].
Acknowledgement: D. Hern?andez-Lobato is supported by Direcci?on General de Investigaci?on MCyT and by
Consejer??a de Educaci?on CAM (projects TIN2010-21575-C02-02, TIN2013-42351-P and S2013/ICE-2845).
V. Sharmanska is funded by the European Research Council under the ERC grant agreement no 308036.
References
[1] D.H. Wolpert. The lack of a priori distinctions between learning algorithms. Neural computation, 8:1341?
1390, 1996.
8
[2] V. Vapnik and A. Vashist. A new learning paradigm: Learning using privileged information. Neural
Networks, 22:544 ? 557, 2009.
[3] D. Pechyony and V. Vapnik. On the theory of learnining with privileged information. In Advances in
Neural Information Processing Systems (NIPS), pages 1894?1902, 2010.
[4] B. Ribeiro, C. Silva, A. Vieira, A. Gaspar-Cunha, and J.C. das Neves. Financial distress model prediction
using SVM+. In International Joint Conference on Neural Networks (IJCNN), 2010.
[5] D. Pechyony and V. Vapnik. Fast optimization algorithms for solving SVM+. In Statistical Learning and
Data Science, 2011.
[6] V. Sharmanska, N. Quadrianto, and C. H. Lampert. Learning to rank using privileged information. In
International Conference on Computer Vision (ICCV), 2013.
[7] J. Feyereisl and U. Aickelin. Privileged information for data clustering. Information Sciences, 194:4?23,
2012.
[8] S. Fouad, P. Tino, S. Raychaudhury, and P. Schneider. Incorporating privileged information through
metric learning. IEEE Transactions on Neural Networks and Learning Systems, 24:1086 ? 1098, 2013.
[9] V. Sharmanska, N. Quadrianto, and C. H. Lampert. Learning to transfer privileged information, 2014.
arXiv:1410.0389 [cs.CV].
[10] M. Lapin, M. Hein, and B. Schiele. Learning using privileged information: SVM+ and weighted SVM.
Neural Networks, 53:95?108, 2014.
[11] P. W. Goldberg, C. K. I. Williams, and C. M. Bishop. Regression with input-dependent noise: A Gaussian
process treatment. In Advances in Neural Information Processing Systems (NIPS), 1998.
[12] N. Quadrianto, K. Kersting, M. D. Reid, T. S. Caetano, and W. L. Buntine. Kernel conditional quantile
estimation via reduction revisited. In International Conference on Data Mining (ICDM), 2009.
[13] M. L?azaro-Gredilla and M. K. Titsias. Variational heteroscedastic Gaussian process regression. In International Conference on Machine Learning (ICML), 2011.
[14] C. E. Rasmussen and C. K. I. Williams. Gaussian Processes for Machine Learning (Adaptive Computation
and Machine Learning). The MIT Press, 2006.
[15] B. Scholkopf and A. J. Smola. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond. MIT Press, Cambridge, MA, USA, 2001.
[16] T. P. Minka. A Family of Algorithms for Approximate Bayesian Inference. PhD thesis, Massachusetts
Institute of Technology, 2001.
[17] H. Nickisch and C. E. Rasmussen. Approximations for Binary Gaussian Process Classification. Journal
of Machine Learning Research, 9:2035?2078, 2008.
[18] M. Kuss and C. E. Rasmussen. Assessing approximate inference for binary Gaussian process classification. Journal of Machine Learning Research, 6:1679?1704, 2005.
[19] M. Seeger. Expectation propagation for exponential families. Technical report, Department of EECS,
University of California, Berkeley, 2006.
[20] T. L. Berg, A. C. Berg, and J. Shih. Automatic attribute discovery and characterization from noisy web
data. In European Conference on Computer Vision (ECCV), 2010.
[21] H. Bay, A. Ess, T. Tuytelaars, and L. Van Gool. Speeded-up robust features (SURF). Computer Vision
and Image Understanding, 110:346?359, 2008.
[22] T. Mikolov, I. Sutskever, K. Chen, G. Corrado, and J. Dean. Distributed representations of words and
phrases and their compositionality. In Advances in Neural Information Processing Systems (NIPS), 2013.
[23] Q. V. Le and T. Mikolov. Distributed representations of sentences and documents. In International
Conference on Machine Learning (ICML), 2014.
[24] C. H. Lampert, H. Nickisch, and S. Harmeling. Attribute-based classification for zero-shot visual object
categorization. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36:453?465, 2014.
[25] J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng, and T. Darrell. Decaf: A deep convolutional activation feature for generic visual recognition. In International Conference on Machine Learning
(ICML), 2014.
[26] J. Dem?sar. Statistical comparisons of classifiers over multiple data sets. Journal of Machine Learning
Research, 7:1?30, 2006.
[27] J. Riihim?aki, P. Jyl?anki, and A. Vehtari. Nested Expectation Propagation for Gaussian Process Classification with a Multinomial Probit Likelihood. Journal of Machine Learning Research, 14:75?109, 2013.
[28] D. Hern?andez-Lobato, J. M. Hern?andez-Lobato, and P. Dupont. Robust multi-class Gaussian process
classification. In Advances in Neural Information Processing Systems (NIPS), 2011.
9
| 5373 |@word trial:1 seal:1 crucially:1 tried:1 deems:1 solid:2 shot:1 reduction:3 united:1 daniel:2 genetic:1 document:2 ours:1 outperforms:2 current:1 com:1 surprising:1 analysed:1 activation:2 riihim:1 attracted:2 must:1 numerical:2 informative:2 shape:1 kyb:1 dupont:1 treating:1 interpretable:2 update:2 n0:2 intelligence:1 device:1 website:1 devising:1 inspection:1 maximised:2 es:1 gpcs:2 normalising:1 characterization:1 s2013:1 codebook:2 revisited:1 preference:1 zhang:1 five:1 iverson:1 constructed:2 scholkopf:1 fitting:1 pairwise:2 theoretically:1 expected:1 mpg:1 planning:1 multi:4 inspired:1 decreasing:1 increasing:3 becomes:2 spain:1 provided:2 moreover:1 notation:1 project:1 what:1 kg:3 complimentary:1 developed:2 finding:3 giant:3 impractical:1 guarantee:1 certainty:1 berkeley:1 finance:1 tie:4 classifier:12 anki:1 control:2 uk:1 normally:1 unit:1 yn:22 grant:1 reid:1 before:1 service:2 generalised:1 positive:2 treat:1 ice:1 hernandez:1 modulation:1 black:1 emphasis:1 twice:1 equivalence:1 christoph:1 heteroscedastic:11 deployment:1 challenging:2 bi:1 statistically:2 speeded:1 commerce:1 harmeling:1 testing:2 definite:2 procedure:4 drug:1 thought:1 illegal:1 convenient:1 matching:1 confidence:4 integrating:2 word:10 significantly:2 reject:2 get:2 put:1 context:2 impossible:1 influence:2 writing:1 equivalent:1 dean:1 customer:1 lobato:4 williams:2 attention:2 focused:1 simplicity:1 onoma:1 rule:1 adjusts:2 financial:1 proving:1 hurt:1 sar:1 play:2 suppose:1 exact:2 gps:2 us:1 goldberg:1 agreement:1 approximated:2 expensive:1 recognition:1 observed:1 role:3 ep:10 bottom:1 enters:1 region:1 caetano:1 learnining:1 vehtari:1 schiele:1 cam:1 trained:1 solving:1 predictive:7 titsias:1 upon:1 division:1 joint:3 differently:1 cat:1 represented:1 train:2 fast:1 effective:1 monte:1 aggregate:1 hyper:5 widely:3 supplementary:4 say:2 otherwise:1 compressed:1 statistic:2 unseen:2 tuytelaars:1 gp:6 transform:1 itself:2 noisy:5 final:1 highlighted:1 sequence:1 mg:4 advantage:2 propose:1 product:3 tu:2 relevant:1 achieve:2 description:3 sutskever:1 convergence:2 chl:1 darrell:1 assessing:1 categorization:1 klosterneuburg:2 object:1 develop:1 ac:3 measured:1 sole:1 strong:4 c:2 predicted:3 come:1 involves:1 skip:1 differ:1 direction:1 guided:1 attribute:23 subsequently:1 public:1 qold:7 material:4 behaviour:1 assign:1 fix:1 andez:4 shopping:1 standardised:3 fnew:2 leopard:1 considered:2 normal:7 exp:13 predict:2 adopt:1 smallest:2 homoscedastic:3 purpose:1 estimation:1 integrates:2 lose:1 bag:4 label:8 council:1 modulating:1 him:1 largest:1 grouped:1 tf:1 fouad:1 repetition:1 reflects:1 hoffman:1 weighted:1 mit:2 clearly:2 sensor:2 gaussian:26 aim:1 super:1 hippopotamus:1 kersting:2 mobile:1 minimisation:1 knob:2 improvement:1 rank:7 likelihood:25 contrast:5 seeger:1 baseline:3 posteriori:1 inference:9 dependent:7 humpback:1 gaspar:1 integrated:1 typically:1 investigaci:1 her:1 relegated:1 interested:1 germany:1 provably:2 overall:1 classification:22 flexible:2 among:3 priori:1 plan:1 art:1 integration:1 animal:5 neve:1 marginal:2 equal:5 tzeng:1 never:1 whale:1 broad:1 novi:1 constitutes:1 icml:3 future:3 report:2 others:2 simplify:1 duplicate:1 employ:1 mcyt:1 dg:1 divergence:3 individual:2 interest:1 normalisation:3 investigate:1 mining:1 evaluation:1 bracket:1 chain:1 word2vec:1 ynew:2 re:1 hein:1 theoretical:1 psychological:1 instance:8 classify:2 soft:1 jyl:1 increased:1 zn:7 calibrate:1 phrase:1 cost:1 introducing:1 deviation:3 too:1 buntine:1 corrupted:1 eec:1 synthetic:1 combined:1 nickisch:2 density:2 international:6 probabilistic:2 universidad:1 off:1 squared:2 reflect:1 again:2 thesis:1 slowly:1 positivity:1 expert:2 derivative:2 de:5 sec:1 includes:1 dem:1 ranking:1 script:1 view:5 lot:2 closed:2 try:1 performed:4 competitive:1 bayes:2 panda:3 capability:1 slope:10 jia:1 accuracy:3 convolutional:3 variance:7 descriptor:2 educaci:1 bayesian:6 vashist:4 produced:1 dortmund:3 carlo:1 pechyony:2 earring:4 kuss:1 whenever:1 energy:2 frequency:1 minka:2 naturally:2 associated:6 gain:1 newly:1 dataset:16 treatment:4 massachusetts:1 austria:4 feyereisl:1 knowledge:8 color:2 improves:3 dimensionality:2 manifest:1 amplitude:3 vieira:1 actually:1 originally:1 supervised:1 specify:1 improved:3 formulation:2 done:2 evaluated:1 furthermore:1 just:3 smola:1 until:1 web:1 trust:1 replacing:1 propagation:6 google:1 lack:1 usa:1 effect:3 contain:1 true:1 regularization:1 analytically:1 hence:3 kxn:1 read:1 iteratively:1 leibler:1 semantic:4 illustrated:3 deal:1 during:1 self:1 nuisance:6 tino:1 sussex:2 aki:1 rat:1 brighton:1 dap:1 complete:1 performs:6 silva:1 image:14 variational:2 novel:1 recently:5 common:1 multinomial:1 discussed:1 interpretation:1 approximates:3 extend:1 significant:2 cambridge:1 imposing:1 cv:1 smoothness:7 rd:6 automatic:1 consistency:1 grid:1 similarly:1 erc:1 distress:1 language:1 had:1 dot:1 funded:1 access:3 robot:1 longer:1 posterior:10 recent:2 retrieved:1 belongs:1 scenario:4 certain:2 binary:10 dataset2:1 yi:2 aut:1 additional:9 care:1 schneider:1 paradigm:3 corrado:1 ii:1 multiple:3 persian:1 smooth:1 technical:1 faster:4 adapt:1 minimising:1 cross:1 icdm:1 privileged:70 prediction:11 regression:6 vision:4 metric:3 expectation:7 df:1 arxiv:1 kernel:10 adopting:1 addition:1 want:1 whereas:1 interval:1 source:1 modality:2 crucial:2 extra:1 rest:2 sure:1 subject:1 induced:1 elegant:1 suspect:1 chimpanzee:3 smile:1 extracting:1 easy:8 enough:2 variety:1 independence:1 fit:2 variate:2 architecture:1 reduce:1 idea:1 pca:3 colour:1 effort:1 render:1 proceed:1 cause:1 deep:7 gpc:61 delivering:1 clear:1 awa:8 amount:1 category:3 http:2 specifies:2 per:3 blue:1 diagnosis:1 summarised:2 ist:4 four:7 shih:1 nevertheless:1 blood:1 characterisation:1 kept:3 convert:1 year:1 uncertainty:5 family:3 reasonable:1 almost:2 c02:1 realtime:1 decision:2 vf:3 comparable:2 layer:1 lupi:4 dash:1 distinguish:1 xnew:4 oracle:1 ijcnn:1 constraint:1 categorisation:1 precisely:1 aspect:2 speed:1 mikolov:2 department:1 gredilla:1 request:1 combination:1 across:1 increasingly:1 making:1 iccv:1 pr:43 taken:1 equation:1 previously:2 hern:4 slack:2 turn:2 needed:2 mind:2 tractable:2 qnew:3 adopted:1 available:6 operation:1 uam:1 observe:2 generic:1 slower:2 original:10 denotes:1 clustering:4 include:1 ensure:1 top:4 hinge:1 exploit:1 quantile:1 objective:1 already:4 gradient:1 distance:4 separate:2 link:1 sensible:1 evaluate:1 considers:2 collected:2 tuebingen:1 reason:2 boldface:1 besides:1 length:2 code:1 providing:1 kingdom:1 setup:2 difficult:6 unfortunately:2 implementation:1 datasets:4 markov:1 supporting:3 situation:1 y1:2 confidential:1 sharmanska:4 community:3 compositionality:1 namely:2 required:1 kl:2 extensive:1 pair:2 sentence:3 discretised:1 engine:1 california:1 distinction:1 textual:3 established:1 nip:4 able:3 bar:1 beyond:1 pattern:1 xm:2 pig:1 built:1 reliable:1 belief:1 gool:1 power:1 critical:4 treated:1 natural:2 rely:1 difficulty:1 advanced:2 marginalised:1 improve:1 technology:1 habitat:1 numerous:1 coupled:2 health:1 text:6 prior:19 review:1 discovery:6 acknowledgement:1 kf:3 understanding:1 regularisation:2 relative:2 probit:8 loss:1 fully:1 interesting:3 generation:1 proportional:1 tin2010:1 proven:1 viktoriia:1 vg:3 querying:1 versus:2 validation:6 clinic:1 sufficient:1 principle:1 eccv:1 supported:2 last:2 free:7 repeat:2 rasmussen:3 guide:1 normalised:1 institute:1 sparse:1 distributed:3 van:1 dimension:1 xn:52 transition:4 cumulative:1 world:1 qn:2 gram:1 made:2 commonly:1 adaptive:1 ribeiro:1 transaction:2 approximate:10 preferred:1 kullback:1 keep:1 global:1 investigating:1 assumed:1 xi:2 continuous:3 latent:26 un:1 search:4 bay:1 table:3 learn:1 transfer:1 robust:2 obtaining:2 necessarily:1 european:2 domain:11 da:1 surf:4 main:4 noise:38 lampert:4 hyperparameters:2 quadrianto:5 quadrature:5 x1:4 direcci:1 madrid:2 elaborate:2 fashion:1 exponential:4 third:1 donahue:1 touched:1 down:1 specific:2 bishop:1 explored:1 svm:39 evidence:6 intractable:1 incorporating:1 vapnik:6 modulates:2 decaf:9 gained:1 importance:2 phd:1 margin:1 chen:1 mf:4 wolpert:1 led:1 simply:2 explore:2 raccoon:1 azaro:1 shoe:4 visual:6 vinyals:1 kristian:1 corresponds:1 nested:1 extracted:4 cdf:4 ma:1 conditional:2 goal:1 viewed:2 kli:1 consequently:1 unethical:1 considerable:1 hard:2 experimentally:1 typical:2 specifically:1 operates:1 principal:3 called:2 total:1 accepted:1 e:1 experimental:3 berg:2 support:1 modulated:1 lapin:1 |
4,831 | 5,374 | Automated Variational Inference
for Gaussian Process Models
Edwin V. Bonilla
The University of New South Wales
[email protected]
Trung V. Nguyen
ANU & NICTA
[email protected]
Abstract
We develop an automated variational method for approximate inference in Gaussian process (GP) models whose posteriors are often intractable. Using a mixture
of Gaussians as the variational distribution, we show that (i) the variational objective and its gradients can be approximated efficiently via sampling from univariate Gaussian distributions and (ii) the gradients wrt the GP hyperparameters can
be obtained analytically regardless of the model likelihood. We further propose
two instances of the variational distribution whose covariance matrices can be
parametrized linearly in the number of observations. These results allow gradientbased optimization to be done efficiently in a black-box manner. Our approach is
thoroughly verified on five models using six benchmark datasets, performing as
well as the exact or hard-coded implementations while running orders of magnitude faster than the alternative MCMC sampling approaches. Our method can be
a valuable tool for practitioners and researchers to investigate new models with
minimal effort in deriving model-specific inference algorithms.
1
Introduction
Gaussian processes (GPs, [1]) are a popular choice in practical Bayesian non-parametric modeling.
The most straightforward application of GPs is the standard regression model with Gaussian likelihood, for which the posterior can be computed in closed form. However, analytical tractability is
no longer possible when having non-Gaussian likelihoods, and inference must be carried out via approximate methods, among which Markov chain Monte Carlo (MCMC, see e.g. [2]) and variational
inference [3] are arguably the two techniques most widely used.
MCMC algorithms provide a flexible framework for sampling from complex posterior distributions
of probabilistic models. However, their generality comes at the expense of very high computational
cost as well as cumbersome convergence analysis. Furthermore, methods such as Gibbs sampling
may perform poorly when there are strong dependencies among the variables of interest. Other
algorithms such as the elliptical slice sampling (ESS) developed in [4] are more effective at drawing
samples from strongly correlated Gaussians. Nevertheless, while improving upon generic MCMC
methods, the sampling cost of ESS remains a major challenge for practical usages.
Alternative to MCMC is the deterministic approximation approach via variational inference, which
has been used in numerous applications with some empirical success ( see e.g. [5, 6, 7, 8, 9, 10, 11]).
The main insight from variational methods is that optimizing is generally easier than integrating.
Indeed, they approximate a posterior by optimizing a lower bound of the marginal likelihood, the
so-called evidence lower bound (ELBO). While variational inference can be considerably faster
than MCMC, it lacks MCMC?s broader applicability as it requires derivations of the ELBO and its
gradients on a model-by-model basis.
This paper develops an automated variational inference technique for GP models that not only reduces the overhead of the tedious mathematical derivations inherent to variational methods but also
1
allows their application to a wide range of problems. In particular, we consider Gaussian process
models that satisfy the following properties: (i) factorization across latent functions and (ii) factorization across observations. The former assumes that, when there is more than one latent function,
they are generated from independent GPs. The latter assumes that, given the latent functions, the
observations are conditionally independent. Existing GP models, such as regression [1], binary and
multi-class classification [6, 12], warped GPs [13], log Gaussian Cox process [14], and multi-output
regression [15], all fall into this class of models. We note, however, that our approach goes beyond
standard settings for which elaborate learning machinery has been developed, as we only require
access to the likelihood function in a black-box manner.
Our automated deterministic inference method uses a mixture of Gaussians as the approximating
posterior distribution and exploits the decomposition of the ELBO into a KL divergence term and
an expected log likelihood term. In particular, we derive an analytical lower bound for the KL term;
and we show that the expected log likelihood term and its gradients can be computed efficiently by
sampling from univariate Gaussian distributions, without explicitly requiring gradients of the likelihood. Furthermore, we optimize the GP hyperparameters within the same variational framework by
using their analytical gradients, irrespective of the specifics of the likelihood models.
Additionally, we exploit the efficient parametrization of the covariance matrices in the models, which
is linear in the number of observations, along with variance-reduction techniques in order to provide an automated inference framework that is useful in practice. We verify the effectiveness of
our method with extensive experiments on 5 different GP settings using 6 benchmark datasets. We
show that our approach performs as well as exact GPs or hard-coded deterministic inference implementations, and that it can be up to several orders of magnitude faster than state-of-the-art MCMC
approaches.
Related work
Black box variational inference (BBVI, [16]) has recently been developed for general latent variable
models. Due to this generality, it under-utilizes the rich amount of information available in GP models that we previously discussed. For example, BBVI approximates the KL term of the ELBO, but
this is computed analytically in our method. A clear disadvantage of BBVI is that it does not provide
an analytical or practical way of learning the covariance hyperparameters of GPs ? in fact, these are
set to fixed values. In principle, these values can be learned in BBVI using stochastic optimization, but experimentally, we have found this to be problematic, ineffectual, and time-consuming. In
contrast, our method optimizes the hyperparameters using their exact gradients.
An approach more closely related to ours is in [17], which investigates variational inference for
GP models with one latent function and factorial likelihood. Their main result is an efficient
parametrization when using a standard variational Gaussian distribution. Our method is more general in that it allows multiple latent functions, hence being applicable to settings such as multi-class
classification and multi-output regression. Furthermore, our variational distribution is a mixture of
Gaussians, with the full Gaussian distribution being a particular case. Another recent approach to
deterministic approximate inference is the Integrated Nested Laplace Approximation (INLA, [18]).
INLA uses numerical integration to approximate the marginal likelihood, which makes it unsuitable
for GP models that contain a large number of hyperparameters.
2
A family of GP models
We consider supervised learning problems with a dataset of N training inputs x = {xn }N
n=1 and
their corresponding targets y = {yn }N
n=1 . The mapping from inputs to outputs is established via
Q underlying latent functions, and our objective is to reason about these latent functions from the
observed data. We specify a class of GP models for which the priors and the likelihoods have the
following structure:
Q
Q
Y
Y
p(f |?0 ) =
p(f?j |?0 ) =
N (f?j ; 0, Kj ),
(1)
j=1
p(y|f , ?1 ) =
N
Y
j=1
p(yn |fn? , ?1 ),
n=1
2
(2)
where f is the set of all latent function values; f?j = {fj (xn )}N
n=1 denotes the values of the latent
Q
function j; fn? = {fj (xn )}j=1 is the set of latent function values which yn depends upon; Kj is
the covariance matrix evaluated at every pair of inputs induced by the covariance function kj (?, ?);
and ?0 and ?1 are covariance hyperparameters and likelihood parameters, respectively.
In other words, the class of models specified by Equations (1) and (2) satisfy the following two criteria: (a) factorization of the prior over the latent functions and (b) factorization of the conditional
likelihood over the observations. Existing GP models including GP regression [1], binary classification [6, 12], warped GPs [13], log Gaussian Cox processes [14], multi-class classification [12], and
multi-output regression [15] all belong to this family of models.
3
Automated variational inference for GP models
This section describes our automated inference framework for posterior inference of the latent functions for the given family of models. Apart from Equations (1) and (2), we only require access to
the likelihood function in a black-box manner, i.e. specific knowledge of its shape or its gradient is
not needed. Posterior inference for general (non-Gaussian) likelihoods is analytically intractable.
We build our posterior approximation framework upon variational inference principles. This entails
positing a tractable family of distributions and finding the member of the family that is ?closest?
to the true posterior in terms of their KL divergence. Herein we choose the family of mixture of
Gaussians (MoG) with K components, defined as
K
K Q
1 X
1 XY
q(f |?) =
qk (f |mk , Sk ) =
N (f?j ; mkj , Skj ),
K
K
j=1
k=1
? = {mkj , Skj },
(3)
k=1
where qk (f |mk , Sk ) is the component k with variational parameters mk = {mkj }Q
j=1 and Sk =
{Skj }Q
.
Less
general
MoG
with
isotropic
covariances
have
been
used
with
variational
inference
j=1
in [7, 19]. Note that within each component, the posteriors over the latent functions are independent.
Minimizing the divergence KL[q(f |?)||p(f |y)] is equivalent to maximizing the evidence lower
bound (ELBO) given by:
K
1 X
?
Eq [log p(y|f )] = L.
log p(y) ? Eq [? log q(f |?)] + Eq [log p(f )] +
{z
} K k=1 k
|
(4)
?KL[q(f |?)||p(f )]
Observe that the KL term in Equation (4) does not depend on the likelihood. The remaining term,
called the expected log likelihood (ELL), is the only contribution of the likelihood to the ELBO. We
can thus address the technical difficulties regarding each component and their derivatives separately
using different approaches. In particular, we can obtain a lower bound of the first term (KL) and
approximate the second term (ELL) via sampling. Due to the limited space, we only show the main
results and refer the reader to the supplementary material for derivation details.
3.1
A lower bound of ?KL[q(f |?)||p(f )]
The first component of the KL divergence term is the entropy of a Gaussian mixture which is not
analytically tractable. However, a lower bound of this entropy can be obtained using Jensen?s inequality (see e.g. [20]) giving:
Eq [? log q(f |?)] ? ?
K
K
X
X
1
1
log
N (mk ; ml , Sk + Sl ).
K
K
k=1
(5)
l=1
The second component of the KL term is a negative cross-entropy between a Gaussian mixture and
a Gaussian, which can be computed analytically giving:
Eq [log p(f )] = ?
K Q
1 XX
?1
N log 2? + log |Kj | + mTkj K?1
j mkj + tr (Kj Skj ) .
2K
j=1
(6)
k=1
The gradients of the two terms in Equations (5) and (6) wrt the variational parameters can be computed analytically and are given in the supplementary material.
3
3.2
An approximation to the expected log likelihood (ELL)
It is clear from Equation (4) that the ELL can be obtained via the ELLs of the individual mixture
components Eqk [log p(y|f )]. Due to the factorial assumption of p(y|f ), the expectation becomes:
Eqk [log p(y|f )] =
N
X
Eqk(n) [log p(yn |fn? )],
(7)
n=1
where qk(n) = qk(n) (fn? |?k(n) ) is the marginal posterior with variational parameters ?k(n) that
correspond to fn? . The gradients of these individual ELL terms wrt the variational parameters ?k(n)
are given by:
??k(n) Eqk(n) [log p(yn |fn? )] =Eqk(n) ??k(n) log qk(n) (fn? |?k(n) ) log p(yn |fn? ).
(8)
Using Equations (7) and (8) we establish the following theorem regarding the computation of the
ELL and its gradients.
Theorem 1. The expected log likelihood and its gradients can be approximated using samples from
univariate Gaussian distributions.
The proof is in Section 1 of the supplementary material. A less general result, for the case of
one latent function and the variational Gaussian posterior, was obtained in [17] using a different
derivation. Note that when Q > 1, qk(n) is not a univariate marginal. Nevertheless, it has a diagonal
covariance matrix due to the factorization of the latent posteriors so the theorem still holds.
3.3
Learning of the variational parameters and other model parameters
In order to learn the parameters of the model we use gradient-based optimization of the ELBO. For
this we require the gradients of the ELBO wrt all model parameters.
Variational parameters. The noisy gradients of the ELBO w.r.t. the variational means mk(n) and
variances Sk(n) corresponding to data point n are given by:
S
X
i
? m L ? ?m Lent + ?m Lcross + 1 s?1 ?
?
(f i ? mk(n) ) log p(yn |fn?
),
k(n)
k(n)
k(n)
KS k(n) i=1 n?
(9)
? S L ? ?S Lent + ?S Lcross
?
k(n)
k(n)
k(n)
S
X
1
?1
?1
?1
i
i
i
+
dg
s
?s
? (fn? ? mk(n) ) ? (fn? ? mk(n) ) ? sk(n) log p(yn |fn?
) (10)
2KS i=1 k(n) k(n)
i
where ? is the entrywise Hadamard product; {fn?
}Si=1 are samples from qk(n) (fn? |mk(n) , sk(n) );
?1
sk(n) is the diagonal of Sk(n) and sk(n) is the element-wise inverse of sk(n) ; dg turns a vector to a
diagonal matrix; and Lent = Eq [? log q(f |?)] and Lcross = Eq [log p(f )] are given by Equations (5)
and (6). The control variates technique described in [16] is also used to further reduce the variance
of these estimators.
Covariance hyperparameters. The ELBO in Equation (4) reveals a remarkable property: the hyperparameters depend only on the negative cross-entropy term Eq [log p(f )] whose exact expression
was derived in Equation (6). This has a significant practical implication: despite using black-box
inference, the hyperparameters are optimized wrt the true evidence lower bound (given fixed variational parameters). This is an additional and crucial advantage of our automated inference method
over other generic inference techniques [16] that seem incapable of hyperparameter learning, in
part because there are not yet techniques for reducing the variance of the gradient estimators. The
gradient of the ELBO wrt any hyperparameter ? of the j-th covariance function is given by:
?? L = ?
K
1 X
?1
?1
T
tr K?1
j ?? Kj ? Kj ?? Kj Kj (mkj mkj + Sj ) .
2K
k=1
4
(11)
Likelihood parameters The noisy gradients w.r.t. the likelihood parameters can also be estimated
via samples from univariate marginals:
??1 L ?
K N
S
1 XXX
k,i
k,i
??1 log p(yn |f(n)
, ?1 ), where f(n)
? qk(n) (fn? |mk(n) , sk(n) ).
KS
n=1 i=1
(12)
k=1
3.4
Practical variational distributions
The gradients from the previous section may be used for automated variational inference for GP
models. However, the mixture of Gaussians (MoG) requires O(N 2 ) variational parameters for each
covariance matrix, i.e. we need to estimate a total of O(QKN 2 ) parameters. This causes difficulties
for learning when these parameters are optimized simultaneously. This section introduces two special members of the MoG family that improve the practical tractability of our inference framework.
Full Gaussian posterior. This instance is the mixture with only 1 component and is thus a Gaussian distribution. Its covariance matrix has block diagonal structure, where each block is a full
covariance corresponding to that of a single latent function posterior. We thus refer to it as the
full Gaussian posterior. As stated in the following theorem, full Gaussian posteriors can still be
estimated efficiently in our variational framework.
Theorem 2. Only O(QN ) variational parameters are required to parametrize the latent posteriors
with full covariance structure.
The proof is given Section 2 of the supplementary material. This result has been stated previously
(see e.g. [6, 7, 17]) but for specific models that belong to the class of GP models considered here.
Mixture of diagonal Gaussians posterior. Our second practical variational posterior is a Gaussian mixture with diagonal covariances, yielding two immediate benefits. Firstly, only O(QN )
parameters are required for each mixture component. Secondly, computation is more efficient as
inverting a diagonal covariance can be done in linear time. Furthermore, as a result of the following
theorem, optimization will typically converge faster when using a mixture of diagonal Gaussians.
Theorem 3. The estimator of the gradients wrt the variational parameters using the mixture of
diagonal Gaussians has a lower variance than the full Gaussian posterior?s.
The proof is in Section 3 of the supplementary material and is based on the Rao-Blackwellization
technique [21]. We note that this result is different to that in [16]. In particular, our variational
distribution is a mixture, thus multi-modal. The theorem is only made possible due to the analytical
tractability of the KL term in the ELBO.
Given the noisy gradients, we use off-the-shelf, gradient-based optimizers, such as conjugate gradient, to learn the model parameters. Note that stochastic optimization may also be used, but it may
require significant time and effort in tuning the learning rates.
3.5
Prediction
Given the MoG posterior, the predictive distribution for new test points x? is given by:
p(Y? |x? ) =
Z
K Z
1 X
p(Y? |f? ) p(f? |f )qk (f )df df? .
K
(13)
k=1
The inner integral is the predictive distribution of the latent values f? and it is a Gaussian since
both qk (f ) and p(f? |f ) are Gaussian. The probability of the test points taking values y? (e.g. in
classification) can thus be readily estimated via Monte Carlo sampling. The predictive means and
variances of a MoG can be obtained from that of the individual mixture components as described in
Section 6 of the supplementary material.
5
Table 1: Datasets, their statistics, and the corresponding likelihood functions and models used in the
experiments, where Ntrain , Ntest , and D are the training size, testing size, and the input dimension,
respectively. See text for detailed description of the models.
Dataset Ntrain Ntest D Likelihood p(y|f )
Model
Mining disasters 811
0
1
?y exp(??)/y!
Log Gausian Cox process
Boston housing 300
206
13 N (y; f, ? 2 )
Standard regression
Creep 800
1266 30 ?y t(y)N (t(y); f, ? 2 ) Warped Gaussian processes
Abalone 1000
3177 8
same as above
Warped Gaussian processes
Breast cancer 300
383
9
1/(1 + exp(?f
Binary classification
P ))
USPS 1233
1232 256 exp(fc )/ i=1 exp(fi ) Multi-class classification
4
Experiments
We perform experiments with five GP models: standard regression [1], warped GPs [13], binary
classification [6, 12], multi-class classification [12], and log Gaussian Cox processes [14] on six
datasets (see Table 1) and repeat the experiments five times using different data subsets.
Experimental settings. The squared exponential covariance function with automatic relevance determination (see Ch. 4 in [1]) is used with the GP regression and warped GPs. The isotropic covariance is used with all other models. The noisy gradients of the ELBO are approximated with
2000 samples and 200 samples are used with control variates to reduce the variance of the gradient
estimators. The model parameters (variational, covariance hyperparameters and likelihood parameters) are learned by iteratively optimizing one set while fixing the others until convergence, which is
determined when changes are less than 1e-5 for the ELBO or 1e-3 for the variational parameters.
Evaluation metrics. To assess the predictive accuracy, we use the standardized squared error (SSE)
for the regression tasks and the classification error rates for the classification tasks. The negative log
predictive density (NLPD) is also used to evaluate the confidence of the prediction. For all of the
metrics, smaller figures are better.
Notations. We call our method AGP and use AGP-FULL, AGP-MIX and AGP-MIX2 when
using the full Gaussian and the mixture of diagonal Gaussians with 1 and 2 components, respectively.
Details of these two posteriors were given in Section 3.4. On the plots, we use the shorter notations,
FULL, MIX, and MIX2 due to the limited space.
Reading the box plots. We used box plots to give a more complete picture of the predictive performance. Each plot corresponds to the distribution of a particular metric evaluated at all test points
for a given task. The edges of a box are the q1 = 25th and q3 = 75th percentiles and the central
mark is the median. The dotted line marks the limit of extreme points that are greater than the 97.5th
percentile. The whiskers enclose the points in the range (q1 ? 1.5(q3 ? q1 ), q3 + 1.5(q3 ? q1 )), which
amounts to approximately ?2.7? if the data is normally distributed. The points outside the whiskers
and below the dotted line are outliers and are plotted individually.
4.1
Standard regression
First we consider the standard Gaussian process regression for which the predictive distribution can
be computed analytically. We compare with this exact inference method (GPR) using the Boston
housing dataset [22]. The results in Figure 1 show that AGP-FULL achieves nearly identical performance as GPR. This is expected as the analytical posterior is a full Gaussian. AGP-MIX and
AGP-MIX2 also give comparable performance in terms of the median SSE and NLPD.
4.2
Warped Gaussian processes (WGP)
The WGP allows for non-Gaussian processes and non-Gaussian noises. The likelihood for each
target yn is attained by warping it through a nonlinear monotonic transformation t(y) giving
p(yn |fn ) = ?yn t(yn )N (t(yn )|fn , ? 2 ). We used the same neural net style transformation as in
[13]. We fixed the warp parameters and used the same procedure for making analytical approximations to the predicted means and variances for all methods.
6
Boston housing
Boston housing
8
0.8
7
NLPD
SSE
0.6
0.4
6
5
4
3
0.2
2
0
FULL
MIX
MIX2
GPR
FULL
MIX
MIX2
GPR
Figure 1: The distributions of SSE and NLPD of all methods on the regression task. Compared to the
exact inference method GPR, the performance of AGP-FULL is identical while that of AGP-MIX
and AGP-MIX2 are comparable.
Creep
Abalone
Creep
Abalone
5
0.4
3
7
0.1
5
1.5
4
0.5
2
FULL
MIX
MIX2
GPR
WGP
3
2
1
3
0
NLPD
2
SSE
0.2
4
2.5
6
NLPD
SSE
0.3
1
0
FULL
MIX
MIX2
GPR
FULL
WGP
MIX
MIX2
GPR
WGP
FULL
MIX
MIX2
GPR
WGP
Figure 2: The distributions of SSE and NLPD of all methods on the regression task with warped
GPs. The AGP methods (FULL, MIX and MIX 2) give comparable performance to exact inference
with WGP and slightly outperform GPR which has narrower ranges of predictive variances.
We compare with the exact implementation of [13] and the standard GP regression (GPR) on the
Creep [23] and Abalone [22] datasets. The results in Figure 2 show that the AGP methods give
comparable performance to the exact method WGP and slightly outperform GPR. The prediction
by GPR exhibits characteristically narrower ranges of predictive variances which can be attributed
to its Gaussian noise assumption.
4.3
Classification
For binary classification, we use the logistic likelihood and experiment with the Wisconsin breast
cancer dataset [22]. We compare with the variational bounds (VBO) and the expectation propagation
(EP) methods. Details of VBO and EP can be found in [6]. All methods use the same analytical
approximations when making prediction.
For multi-class classification, we use the softmax likelihood and experiment with a subset of the
USPS dataset [1] containing the digits 4, 7, and 9. We compare with a variational inference method
(VQ) which constructs the ELBO via a quadratic lower bound to the likelihood terms [5]. Prediction
is made by squashing the samples from the predictive distributions of the latent values at test points
through the softmax likelihood for all methods.
Breast cancer
USPS
0.06
1
1
0.05
0.03
0.02
0.8
0.6
NLPD
Error rates
VQ
FULL
MIX
MIX2
VBO
EP
NLPD
0.8
0.04
0.4
0.2
0.6
0.4
0.2
0.01
0
0
Breast cancer
USPS
0
FULL
MIX
MIX2
VBO
EP
FULL
MIX
MIX2
VQ
Figure 3: Left plot: classification error rates averaged over 5 runs (the error bars show two standard
deviations). The AGP methods have classification errors comparable to the hard-coded implementations. Middle and right plots: the distribution of NLPD of all methods on the binary and multi-class
classification tasks, respectively. The hard-coded methods are slightly better than AGP.
7
4
Posteriors of the latent intensity
0.5
Intensity
Event counts
3
2
0.4
FULL
MIX
HMC & ESS
0.3
0.2
1
0.1
0
1860 1880 1900 1920 1940 1960
Time
0
1860 1880 1900 1920 1940 1960
Time
2.5
Log10 speed?up factor
0.6
2
1.5
1
FULL
MIX
ESS
0.5
0
Time comparison against HMC
Figure 4: Left plot: the true event counts during the given time period. Middle plot: the posteriors
(estimated intensities) inferred by all methods. For each method, the middle line is the posterior
mean and the two remaining lines enclose 90% interval. AGP-FULL infers the same posterior as
HMC and ESS while AGP-MIX obtains the same mean but underestimates the variance. Right
plot: speed-up factors against the HMC method. The AGP methods run more than 2 orders of
magnitude faster than the sampling methods.
The classification error rates and the NLPD are shown in Figure 3 for both tasks. For binary classification, the AGP methods give comparable performance to the hard-coded implementations, VBO
and EP. The latter is often considered the best approximation method for this task [6]. Similar
results can be observed for the multi-class classification problem.
We note that the running times of our methods are comparable to that of the hard-coded methods.
For example, the average training times for VBO, EP, MIX, and FULL are 76s, 63s, 210s, and 480s
respectively, on the Wisconsin dataset.
4.4
Log Gaussian Cox process (LGCP)
The LGCP is an inhomogeneous Poisson process with the log-intensity function being a shifted
?yn exp(?? )
draw from a Gaussian process. Following [4], we use the likelihood p(yn |fn ) = n yn ! n ,
where ?n = exp(fn + m) is the mean of a Poisson distribution and m is the offset to the log mean.
The data concerns coal-mining disasters taken from a standard dataset for testing point processes
[24]. The offset m and the covariance hyperparameters are set to the same values as in [4].
We compare AGP with the Hybrid Monte Carlo (HMC, [25]) and elliptical slice sampling (ESS,
[4]) methods, where the latter is designed specifically for GP models. We collected every 100th
sample for a total of 10k samples after a burn-in period of 5k samples; the Gelman-Rubin potential
scale reduction factors [26] are used to check for convergence. The middle plot of Figure 4 shows
the posteriors learned by all methods. We see that the posterior by AGP-FULL is similar to that
by HMC and ESS. AGP-MIX obtains the same posterior mean but it underestimates the variance.
The right plot shows the speed-up factors of all methods against the slowest method HMC. The
AGP methods run more than two orders of magnitude faster than HMC, thus confirming the computational advantages of our method to the sampling approaches. Training time was measured on a
desktop with Intel(R) i7-2600 3.40GHz CPU with 8GB of RAM using Matlab R2012a.
5
Discussion
We have developed automated variational inference for Gaussian process models (AGP). AGP performs as well as the exact or hard-coded implementations when testing on five models using six real
world datasets. AGP has the potential to be a powerful tool for GP practitioners and researchers
when devising models for new or existing problems for which variational inference is not yet available. In the future we will address the scalability of AGP to deal with very large datasets.
Acknowledgements
NICTA is funded by the Australian Government through the Department of Communications and
the Australian Research Council through the ICT Centre of Excellence Program.
8
References
[1] Carl Edward Rasmussen and Christopher K. I. Williams. Gaussian processes for machine learning. The
MIT Press, 2006.
[2] Radford M. Neal. Probabilistic inference using Markov chain Monte Carlo methods. Technical report,
Department of Computer Science, University of Toronto, 1993.
[3] Michael I Jordan, Zoubin Ghahramani, Tommi S Jaakkola, and Lawrence K Saul. An introduction to
variational methods for graphical models. Springer, 1998.
[4] Iain Murray, Ryan Prescott Adams, and David J.C. MacKay. Elliptical slice sampling. In AISTATS, 2010.
[5] Mohammad E. Khan, Shakir Mohamed, Benjamin M. Marlin, and Kevin P. Murphy. A stick-breaking
likelihood for categorical data analysis with latent Gaussian models. In AISTATS, pages 610?618, 2012.
[6] Hannes Nickisch and Carl Edward Rasmussen. Approximations for binary Gaussian process classification. Journal of Machine Learning Research, 9(10), 2008.
[7] Trung V. Nguyen and Edwin Bonilla. Efficient variational inference for Gaussian process regression
networks. In AISTATS, pages 472?480, 2013.
[8] Mohammad E. Khan, Shakir Mohamed, and Kevin P. Murphy. Fast Bayesian inference for non-conjugate
Gaussian process regression. In NIPS, pages 3149?3157, 2012.
[9] Miguel L?azaro-Gredilla. Bayesian warped Gaussian processes. In NIPS, pages 1628?1636, 2012.
[10] Mark Girolami and Simon Rogers. Variational Bayesian multinomial probit regression with Gaussian
process priors. Neural Computation, 18(8):1790?1817, 2006.
[11] Miguel L?azaro-Gredilla and Michalis Titsias. Variational heteroscedastic Gaussian process regression. In
ICML, 2011.
[12] Christopher K.I. Williams and David Barber. Bayesian classification with Gaussian processes. Pattern
Analysis and Machine Intelligence, IEEE Transactions on, 20(12):1342?1351, 1998.
[13] Edward Snelson, Carl Edward Rasmussen, and Zoubin Ghahramani. Warped Gaussian processes. In
NIPS, 2003.
[14] Jesper M?ller, Anne Randi Syversveen, and Rasmus Plenge Waagepetersen. Log Gaussian Cox processes.
Scandinavian journal of statistics, 25(3):451?482, 1998.
[15] Andrew G. Wilson, David A. Knowles, and Zoubin Ghahramani. Gaussian process regression networks.
In ICML, 2012.
[16] Rajesh Ranganath, Sean Gerrish, and David M. Blei. Black box variational inference. In AISTATS, 2014.
[17] Manfred Opper and C?edric Archambeau. The variational Gaussian approximation revisited. Neural
Computation, 21(3):786?792, 2009.
[18] H?avard Rue, Sara Martino, and Nicolas Chopin. Approximate Bayesian inference for latent Gaussian
models by using integrated nested Laplace approximations. Journal of the royal statistical society: Series
b (statistical methodology), 71(2):319?392, 2009.
[19] Samuel J. Gershman, Matthew D. Hoffman, and David M. Blei. Nonparametric variational inference. In
ICML, 2012.
[20] Marco F. Huber, Tim Bailey, Hugh Durrant-Whyte, and Uwe D. Hanebeck. On entropy approximation
for Gaussian mixture random vectors. In IEEE International Conference on Multisensor Fusion and
Integration for Intelligent Systems, 2008.
[21] George Casella and Christian P. Robert. Rao-Blackwellisation of sampling schemes. Biometrika, 1996.
[22] K. Bache and M. Lichman. UCI machine learning repository, 2013.
[23] D. Cole, C. Martin-Moran, A.G. Sheard, H.K.D.H. Bhadeshia, and D.J.C. MacKay. Modelling creep
rupture strength of ferritic steel welds. Science and Technology of Welding & Joining, 5(2):81?89, 2000.
[24] R.G. Jarrett. A note on the intervals between coal-mining disasters. Biometrika, 66(1):191?193, 1979.
[25] Simon Duane, Anthony D. Kennedy, Brian J. Pendleton, and Duncan Roweth. Hybrid Monte Carlo.
Physics letters B, 195(2):216?222, 1987.
[26] Andrew Gelman and Donald B Rubin. Inference from iterative simulation using multiple sequences.
Statistical science, pages 457?472, 1992.
9
| 5374 |@word cox:6 middle:4 repository:1 tedious:1 simulation:1 covariance:20 decomposition:1 q1:4 tr:2 edric:1 reduction:2 series:1 lichman:1 ours:1 existing:3 elliptical:3 com:1 lgcp:2 anne:1 si:1 yet:2 must:1 readily:1 fn:19 numerical:1 confirming:1 shape:1 christian:1 plot:11 designed:1 intelligence:1 devising:1 ntrain:2 trung:2 desktop:1 isotropic:2 es:7 parametrization:2 manfred:1 blei:2 revisited:1 toronto:1 firstly:1 five:4 positing:1 mathematical:1 along:1 wale:1 overhead:1 manner:3 coal:2 excellence:1 huber:1 expected:6 indeed:1 lcross:3 multi:12 blackwellization:1 cpu:1 becomes:1 xx:1 underlying:1 notation:2 developed:4 finding:1 transformation:2 marlin:1 every:2 biometrika:2 stick:1 control:2 normally:1 yn:17 arguably:1 limit:1 despite:1 joining:1 approximately:1 black:6 burn:1 au:2 k:3 sara:1 heteroscedastic:1 archambeau:1 factorization:5 limited:2 range:4 jarrett:1 averaged:1 practical:7 testing:3 practice:1 block:2 optimizers:1 digit:1 procedure:1 empirical:1 word:1 integrating:1 confidence:1 prescott:1 donald:1 zoubin:3 mkj:6 gelman:2 optimize:1 equivalent:1 deterministic:4 maximizing:1 straightforward:1 regardless:1 go:1 williams:2 insight:1 estimator:4 iain:1 deriving:1 sse:7 laplace:2 target:2 exact:10 gps:10 us:2 carl:3 element:1 approximated:3 skj:4 bache:1 observed:2 ep:6 wgp:8 valuable:1 benjamin:1 gausian:1 depend:2 predictive:10 titsias:1 upon:3 basis:1 edwin:2 usps:4 bhadeshia:1 derivation:4 fast:1 effective:1 jesper:1 monte:5 kevin:2 outside:1 pendleton:1 whose:3 widely:1 supplementary:6 drawing:1 elbo:15 statistic:2 gp:21 noisy:4 shakir:2 eqk:5 advantage:2 housing:4 sequence:1 analytical:8 net:1 propose:1 product:1 uci:1 hadamard:1 poorly:1 description:1 scalability:1 convergence:3 adam:1 tim:1 derive:1 develop:1 andrew:2 fixing:1 miguel:2 measured:1 eq:8 edward:4 strong:1 predicted:1 enclose:2 come:1 australian:2 girolami:1 tommi:1 whyte:1 inhomogeneous:1 closely:1 stochastic:2 material:6 rogers:1 require:4 government:1 brian:1 ryan:1 secondly:1 hold:1 gradientbased:1 marco:1 considered:2 exp:6 lawrence:1 mapping:1 inla:2 matthew:1 major:1 achieves:1 applicable:1 council:1 individually:1 cole:1 tool:2 hoffman:1 mit:1 gaussian:54 shelf:1 broader:1 jaakkola:1 wilson:1 derived:1 q3:4 martino:1 modelling:1 likelihood:33 check:1 slowest:1 contrast:1 inference:38 ferritic:1 integrated:2 typically:1 vbo:6 chopin:1 uwe:1 among:2 flexible:1 classification:22 art:1 integration:2 ell:7 special:1 marginal:4 softmax:2 construct:1 mackay:2 having:1 sampling:14 identical:2 icml:3 nearly:1 future:1 others:1 report:1 develops:1 inherent:1 plenge:1 intelligent:1 dg:2 simultaneously:1 divergence:4 individual:3 murphy:2 interest:1 investigate:1 mining:3 bbvi:4 evaluation:1 introduces:1 mixture:18 extreme:1 yielding:1 chain:2 implication:1 rajesh:1 integral:1 edge:1 xy:1 shorter:1 machinery:1 plotted:1 minimal:1 mk:10 roweth:1 instance:2 modeling:1 rao:2 disadvantage:1 tractability:3 cost:2 applicability:1 subset:2 deviation:1 dependency:1 considerably:1 nickisch:1 thoroughly:1 density:1 international:1 hugh:1 probabilistic:2 off:1 physic:1 michael:1 squared:2 central:1 containing:1 choose:1 warped:10 derivative:1 style:1 potential:2 satisfy:2 bonilla:3 explicitly:1 depends:1 closed:1 randi:1 simon:2 contribution:1 ass:1 accuracy:1 variance:12 qk:10 efficiently:4 correspond:1 bayesian:6 carlo:5 researcher:2 kennedy:1 cumbersome:1 casella:1 against:3 underestimate:2 mohamed:2 proof:3 attributed:1 dataset:7 popular:1 knowledge:1 infers:1 sean:1 attained:1 supervised:1 xxx:1 methodology:1 specify:1 modal:1 entrywise:1 hannes:1 done:2 box:9 strongly:1 generality:2 furthermore:4 evaluated:2 syversveen:1 until:1 lent:3 christopher:2 nonlinear:1 lack:1 propagation:1 logistic:1 usage:1 requiring:1 verify:1 contain:1 true:3 former:1 analytically:7 hence:1 iteratively:1 neal:1 deal:1 conditionally:1 during:1 percentile:2 samuel:1 abalone:4 criterion:1 complete:1 mohammad:2 performs:2 fj:2 variational:49 wise:1 snelson:1 recently:1 fi:1 multinomial:1 discussed:1 belong:2 approximates:1 marginals:1 refer:2 significant:2 gibbs:1 tuning:1 automatic:1 centre:1 funded:1 access:2 entail:1 longer:1 scandinavian:1 posterior:31 closest:1 recent:1 optimizing:3 optimizes:1 apart:1 inequality:1 binary:8 success:1 incapable:1 creep:5 additional:1 greater:1 george:1 converge:1 period:2 ller:1 ii:2 multiple:2 full:28 mix:20 reduces:1 technical:2 faster:6 determination:1 cross:2 coded:7 prediction:5 regression:20 breast:4 mog:6 expectation:2 df:2 metric:3 poisson:2 disaster:3 separately:1 interval:2 median:2 crucial:1 south:1 induced:1 member:2 effectiveness:1 seem:1 practitioner:2 call:1 jordan:1 automated:10 variate:2 reduce:2 regarding:2 inner:1 i7:1 six:3 expression:1 gb:1 effort:2 cause:1 matlab:1 useful:1 generally:1 clear:2 detailed:1 factorial:2 amount:2 nonparametric:1 sl:1 outperform:2 problematic:1 shifted:1 dotted:2 estimated:4 hyperparameter:2 nevertheless:2 verified:1 ram:1 run:3 inverse:1 letter:1 powerful:1 family:7 reader:1 agp:26 knowles:1 utilizes:1 draw:1 duncan:1 investigates:1 comparable:7 bound:10 quadratic:1 strength:1 weld:1 speed:3 performing:1 martin:1 department:2 gredilla:2 conjugate:2 across:2 describes:1 smaller:1 slightly:3 making:2 outlier:1 taken:1 equation:9 vq:3 remains:1 previously:2 turn:1 count:2 wrt:7 needed:1 tractable:2 available:2 gaussians:10 parametrize:1 observe:1 generic:2 bailey:1 alternative:2 assumes:2 running:2 denotes:1 remaining:2 standardized:1 graphical:1 michalis:1 log10:1 unsuitable:1 exploit:2 giving:3 ghahramani:3 build:1 establish:1 approximating:1 murray:1 society:1 warping:1 objective:2 parametric:1 diagonal:10 exhibit:1 gradient:25 parametrized:1 barber:1 collected:1 multisensor:1 reason:1 nicta:3 characteristically:1 nlpd:11 rasmus:1 minimizing:1 hmc:8 robert:1 expense:1 negative:3 stated:2 steel:1 implementation:6 unsw:1 perform:2 observation:5 datasets:7 markov:2 benchmark:2 immediate:1 communication:1 hanebeck:1 intensity:4 inferred:1 david:5 inverting:1 pair:1 required:2 kl:12 extensive:1 specified:1 ineffectual:1 optimized:2 khan:2 learned:3 herein:1 established:1 nip:3 address:2 beyond:1 bar:1 below:1 pattern:1 reading:1 challenge:1 program:1 including:1 royal:1 event:2 difficulty:2 hybrid:2 scheme:1 improve:1 technology:1 numerous:1 picture:1 irrespective:1 carried:1 categorical:1 rupture:1 kj:9 text:1 prior:3 ict:1 acknowledgement:1 wisconsin:2 probit:1 whisker:2 gershman:1 remarkable:1 rubin:2 principle:2 squashing:1 cancer:4 repeat:1 rasmussen:3 blackwellisation:1 allow:1 warp:1 wide:1 fall:1 taking:1 saul:1 waagepetersen:1 benefit:1 slice:3 distributed:1 dimension:1 xn:3 ghz:1 world:1 rich:1 qn:2 opper:1 sheard:1 made:2 nguyen:3 welding:1 transaction:1 ranganath:1 sj:1 approximate:7 obtains:2 ml:1 reveals:1 consuming:1 latent:23 iterative:1 sk:12 table:2 additionally:1 learn:2 nicolas:1 improving:1 complex:1 anthony:1 rue:1 aistats:4 main:3 linearly:1 noise:2 hyperparameters:11 qkn:1 intel:1 elaborate:1 exponential:1 breaking:1 durrant:1 gpr:13 theorem:8 specific:4 jensen:1 moran:1 offset:2 evidence:3 concern:1 intractable:2 fusion:1 magnitude:4 anu:1 easier:1 boston:4 entropy:5 fc:1 azaro:2 univariate:5 monotonic:1 radford:1 ch:1 nested:2 corresponds:1 springer:1 gerrish:1 duane:1 conditional:1 narrower:2 hard:7 experimentally:1 change:1 determined:1 specifically:1 reducing:1 called:2 total:2 ntest:2 experimental:1 mark:3 latter:3 relevance:1 evaluate:1 mcmc:8 correlated:1 |
4,832 | 5,375 | Variational Gaussian Process State-Space Models
Roger Frigola, Yutian Chen and Carl E. Rasmussen
Department of Engineering
University of Cambridge
{rf342,yc373,cer54}@cam.ac.uk
Abstract
State-space models have been successfully used for more than fifty years in different areas of science and engineering. We present a procedure for efficient variational Bayesian learning of nonlinear state-space models based on sparse Gaussian
processes. The result of learning is a tractable posterior over nonlinear dynamical
systems. In comparison to conventional parametric models, we offer the possibility to straightforwardly trade off model capacity and computational cost whilst
avoiding overfitting. Our main algorithm uses a hybrid inference approach combining variational Bayes and sequential Monte Carlo. We also present stochastic
variational inference and online learning approaches for fast learning with long
time series.
1
Introduction
State-space models (SSMs) are a widely used class of models that have found success in applications
as diverse as robotics, ecology, finance and neuroscience (see, e.g., Brown et al. [3]). State-space
models generalize other popular time series models such as linear and nonlinear auto-regressive
models: (N)ARX, (N)ARMA, (G)ARCH, etc. [21].
In this article we focus on Bayesian learning of nonparametric nonlinear state-space models. In
particular, we use sparse Gaussian processes (GPs) [19] as a convenient method to encode general
assumptions about the dynamical system such as continuity or smoothness. In contrast to conventional parametric methods, we allow the user to easily trade off model capacity and computation
time. Moreover, we present a variational training procedure that allows very complex models to be
learned without risk of overfitting.
Our variational formulation leads to a tractable approximate posterior over nonlinear dynamical
systems. This approximate posterior can be used to compute fast probabilistic predictions of future
trajectories of the dynamical system. The computational complexity of our learning approach is
linear in the length of the time series. This is possible thanks to the use of variational sparse GPs [22]
which lead to a smoothing problem for the latent state trajectory in a simpler auxiliary dynamical
system. Smoothing in this auxiliary system can be carried out with any conventional technique (e.g.
sequential Monte Carlo). In addition, we present a stochastic variational inference procedure [10] to
accelerate learning for long time series and we also present an online learning scheme.
This work is useful in situations where: 1) it is important to know how uncertain future predictions
are, 2) there is not enough knowledge about the underlying nonlinear dynamical system to create
a principled parametric model, and 3) it is necessary to have an explicit model that can be used
to simulate the dynamical system into the future. These conditions arise often in engineering and
finance. For instance, consider an autonomous aircraft adapting its flight control when carrying a
large external load of unknown weight and aerodynamic characteristics. A model of the nonlinear
dynamics of the new system can be very useful in order to automatically adapt the control strategy.
When few data points are available, there is high uncertainty about the dynamics. In this situation,
1
a model that quantifies its uncertainty can be used to synthesize control laws that avoid the risks of
overconfidence.
The problem of learning flexible models of nonlinear dynamical systems has been tackled from
multiple perspectives. Ghahramani and Roweis [9] presented a maximum likelihood approach to
learn nonlinear SSMs based on radial basis functions. This work was later extended by using a
parameterized Gaussian process point of view and developing tailored filtering algorithms [6, 7, 23].
Approximate Bayesian learning has also been developed for parameterized nonlinear SSMs [5, 24].
Wang et al. [25] modeled the nonlinear functions in SSMs using Gaussian processes (GP-SSMs) and
found a MAP estimate of the latent variables and hyperparameters. Their approach preserved the
nonparametric properties of Gaussian processes. Despite using MAP learning over state trajectories,
overfitting was not an issue since it was applied in a dimensionality reduction context where the
latent space of the SSM was much smaller than the observation space. In a similar vein, [4, 12]
presented a hierarchical Gaussian process model that could model linear dynamics and nonlinear
mappings from latent states to observations. More recently, Frigola et al. [8] learned GP-SSMs
in a fully Bayesian manner by employing particle MCMC methods to sample from the smoothing
distribution. However, their approach led to predictions with a computational cost proportional to
the length of the time series.
In the rest of this article, we present an approach to variational Bayesian learning of flexible nonlinear state-space models which leads to a simple representation of the posterior over nonlinear
dynamical systems and results in predictions having a low computational complexity.
2
Gaussian Process State-Space Models
We consider discrete-time nonlinear state-space models built with deterministic functions and additive noise
xt+1 = f (xt ) + vt ,
(1a)
yt = g(xt ) + et .
(1b)
The dynamics of the system are defined by the state transition function f (xt ) and independent
additive noise vt (process noise). The states xt ? RD are latent variables such that all future
variables are conditionally independent on the past given the present state. Observations yt ? RE
are linked to the state via another deterministic function g(xt ) and independent additive noise et
(observation noise). State-space models are stochastic dynamical processes that are useful to model
time series y , {y1 , ..., yT }. The deterministic functions in (1) can also take external known inputs
(such as control signals) as an argument but, for conciseness, we will omit those in our notation.
A traditional approach to learn f and g is to restrict them to a family of parametric functions. This is
particularly appropriate when the dynamical system is very well understood, e.g. orbital mechanics
of a spacecraft. However, in many applications, it is difficult to specify a class of parametric models
that can provide both the ability to model complex functions and resistance to overfitting thanks to an
easy to specify prior or regularizer. Gaussian processes do have these properties: they can represent
functions of arbitrary complexity and provide a straightforward way to specify assumptions about
those unknown functions, e.g. smoothness. In the light of this, it is natural to place Gaussian process
priors over both f and g [25]. However, the extreme flexibility of the two Gaussian processes
leads to severe nonidentifiability and strong correlations between the posteriors of the two unknown
functions. In the rest of this paper we will focus on a model with a GP prior over the transition
function and a parametric likelihood. However, our variational formulation can also be applied to
the double GP case (see supplementary material).
A probabilistic state-space model with a Gaussian process prior over the transition function and a
parametric likelihood is specified by
f (x) ? GP mf (x), kf (x, x0 ) ,
(2a)
xt | ft ? N (xt | ft , Q),
x0 ? p(x0 )
yt | xt ? p(yt | xt , ? y ),
(2b)
(2c)
(2d)
where we have used ft , f (xt?1 ). Since f (x) ? RD , we use the convention that the covariance
function kf returns a D ? D matrix. We group all hyperparameters into ? , {? f , ? y , Q}. Note that
2
states
0
0
time
0
time
0
time
time
Figure 1: State trajectories from four 2-state nonlinear dynamical systems sampled from a GP-SSM
prior with fixed hyperparameters. The same prior generates systems with qualitatively different
behaviors, e.g. the leftmost panel shows behavior similar to that of a non-oscillatory linear system
whereas the rightmost panel appears to have arisen from a limit cycle in a nonlinear system.
we are not restricting the likelihood (2d) to any particular form. The joint distribution of a GP-SSM
is
T
Y
p(y, x, f ) = p(x0 )
p(yt |xt )p(xt |ft )p(ft |f1:t?1 , x0:t?1 ),
(3)
t=1
where we use the convention f1:0 = ? and omit the conditioning on ? in the notation. The GP on
the transition function induces a distribution over the latent function values with the form of a GP
predictive:
p(ft |f1:t?1 , x0:t?1 ) = N mf (xt?1 ) + Kt?1,0:t?2 K?1
0:t?2,0:t?2 (f1:t?1 ? mf (x0:t?2 )),
?1
Kt?1,t?1 ? Kt?1,0:t?2 K0:t?2,0:t?2 K>
(4)
t?1,0:t?2 ,
where the subindices of the kernel matrices indicate the arguments to the covariance function necessary to build each matrix, e.g. Kt?1,0:t?2 = [kf (xt?1 , x0 ) . . . kf (xt?1 , xt?2 )]. When t = 1, the
distribution is that of a GP marginal p(f1 |x0 ) = N (mf (x0 ), kf (x0 , x0 )).
Equation (3) provides a sequential procedure to sample state trajectories and observations. GPSSMs are doubly stochastic models in the sense that one could, at least notionally, first sample a
state transition dynamics function from eq. (2a) and then, conditioned on that function, sample the
state trajectory and observations.
GP-SSMs are a very rich prior over nonlinear dynamical systems. In Fig. 1 we illustrate this concept
by showing state trajectories sampled from a GP-SSM with fixed hyperparameters. The dynamical
systems associated with each of these trajectories are qualitatively very different from each other. For
instance, the leftmost panel shows the dynamics of an almost linear non-oscillatory system whereas
the rightmost panel corresponds to a limit cycle in a nonlinear system. Our goal in this paper is
to use this prior over dynamical systems and obtain a tractable approximation to the posterior over
dynamical systems given the data.
3
Variational Inference in GP-SSMs
Since the GP-SSM is a nonparametric model, in order to define a posterior distribution over f (x) and
make probabilistic predictions it is necessary to first find the smoothing distribution p(x0:T |y1:T ).
Frigola et al. [8] obtained samples from the smoothing distribution that could be used to define a
predictive density via Monte Carlo integration. This approach is expensive since it requires averaging over L state trajectory samples of length T . In this section we present an alternative approach
that aims to find a tractable distribution over the state transition function that is independent of the
length of the time series. We achieve this by using variational sparse GP techniques [22].
3.1
Augmenting the Model with Inducing Variables
As a first step to perform variational inference in a GP-SSM, we augment the model with M inducing
points u , {ui }M
i=1 . Those inducing points are jointly Gaussian with the latent function values. In
the case of a GP-SSM, the joint probability density becomes
p(y, x, f , u) = p(x, f |u) p(u)
T
Y
t=1
3
p(yt |xt ),
(5)
where
p(u) = N (u | 0, Ku,u )
p(x, f |u) = p(x0 )
T
Y
(6a)
p(ft |f1:t?1 , x0:t?1 , u)p(xt |ft ),
(6b)
t=1
T
Y
?1
>
p(ft |f1:t?1 , x0:t?1 , u) = N f1:T | K0:T ?1,u K?1
u,u u, K0:T ?1 ? K0:T ?1,u Ku,u K0:T ?1,u . (6c)
t=1
Kernel matrices relating to the inducing points depend on a set of inducing inputs {zi }M
i=1 in such
a way that Ku,u is an M D ? M D matrix formed with blocks kf (zi , zj ) having size D ? D. For
brevity, we use a zero mean function and we omit conditioning on the inducing inputs in the notation.
3.2
Evidence Lower Bound of an Augmented GP-SSM
Variational inference [1] is a popular method for approximate Bayesian inference based on making
assumptions about the posterior over latent variables that lead to a tractable lower bound on the
evidence of the model (sometimes referred to as ELBO). Maximizing this lower bound is equivalent
to minimizing the Kullback-Leibler divergence between the approximate posterior and the exact
one. Following standard variational inference methodology, [1] we obtain the evidence lower bound
of a GP-SSM augmented with inducing points
QT
Z
p(u)p(x0 ) t=1 p(ft |f1:t?1 , x0:t?1 , u)p(yt |xt )p(xt |ft )
log p(y|?) ?
q(x, f , u) log
.
(7)
q(x, f , u)
x,f ,u
In order to achieve tractability, we use a variational distribution that factorizes as
q(x, f , u) = q(u)q(x)
T
Y
p(ft |f1:t?1 , x0:t?1 , u),
(8)
t=1
where q(u) and q(x) can take any form but the terms relating to f are taken to match those of the
prior (3). As a consequence, the difficult p(ft |...) terms inside the log cancel out and lead to the
following lower bound
Z
L(q(u), q(x),?) = ?KL(q(u)kp(u)) + H(q(x)) + q(x) log p(x0 )
x
+
T Z
X
t=1
q(x)q(u)
x,u
ft
|
Z
Z
p(ft |xt?1 , u) log p(xt |ft ) +
{z
}
q(x) log p(yt |xt )
(9)
x
?(xt ,xt?1 ,u)
where KL denotes the Kullback-Leibler divergence and H the entropy. The integral with respect to
ft can be solved analytically: ?(xt , xt?1 , u) = ? 21 tr(Q?1 Bt?1 ) + log N (xt |At?1 u, Q) where
?1
At?1 = Kt?1,u K?1
u,u , and Bt?1 = Kt?1,t?1 ? Kt?1,u Ku,u Ku,t?1 .
As in other variational sparse GP methods, the choice of variational distribution (8) gives the ability to precisely learn the latent function at the locations of the inducing inputs. Away from those
locations, the posterior takes the form of the prior conditioned on the inducing variables. By increasing the number of inducing variables, the ELBO can only become tighter [22]. This offers a
straightforward trade-off between model capacity and computation cost without increasing the risk
of overfitting.
3.3
Optimal Variational Distribution for u
The optimal distribution of q(u) can be found by setting to zero the functional derivative of the
evidence lower bound with respect to q(u)
q ? (u) ? p(u)
T
Y
exp{hlog N (xt |At?1 u, Q)iq(x) },
t=1
4
(10)
where h?iq(x) denotes an expectation with respect to q(x). The optimal variational distribution
q ? (u) is, conveniently, a multivariate Gaussian distribution. If, for simplicity of notation, we restrict
ourselves to D = 1 the natural parameters of the optimal distribution are
?1
?1 = Q
T
X
hATt?1 xt iq(xt ,xt?1 ) ,
t=1
1
?2 = ?
2
K?1
uu
?1
+Q
T
X
hATt?1 At?1 iq(xt?1 )
!
. (11)
t=1
The mean and covariance matrix of q ? (u), denoted as ? and ? respectively, can be computed as
? = ?? 1 and ? = (?2? 2 )?1 . Note that the optimal q(u) depends on the sufficient statistics
PT
PT
?1 = t=1 hKTt?1,u xt iq(xt ,xt?1 ) and ?2 = t=1 hKTt?1,u Kt?1,u iq(xt?1 ) .
3.4
Optimal Variational Distribution for x
In an analogous way as for q ? (u), we can obtain the optimal form of q(x)
q ? (x) ? p(x0 )
T
Y
1
p(yt |xt ) exp{? tr Q?1 (Bt?1 + At?1 ?ATt?1 ) } N (xt |At?1 ?, Q), (12)
2
t=1
where, in the second equation, we have used q(u) = N (u|?, ?).
The optimal distribution q ? (x) is equivalent to the smoothing distribution of an auxiliary parametric
state-space model. The auxiliary model is simpler than the original one in (3) since the latent states
factorize with a Markovian structure. Equation (12) can be interpreted as a nonlinear state-space
model with a Gaussian state transition density, N (xt |At?1 ?,
Q), and a likelihood augmented with
an additional term: exp{? 21 tr Q?1 (Bt?1 + At?1 ?ATt?1 ) }.
Smoothing in nonlinear Markovian state-space models is a standard problem in the context of time
series modeling. There are various existing strategies to find the smoothing distribution which could
be used depending on the characteristics of each particular problem [20]. For instance, in a mildly
nonlinear system with Gaussian noise, an extended Kalman smoother can have very good performance. On the other hand, problems with severe nonlinearities and/or non-Gaussian likelihoods can
lead to heavily multimodal smoothing distributions that are better represented using particle methods. We note that the application of sequential Monte Carlo (SMC) is particularly straightforward
in the present auxiliary model.
3.5
Optimizing the Evidence Lower Bound
Algorithm 1 presents a procedure to maximize the evidence lower bound by alternatively sampling
from the smoothing distribution and taking steps both in ? and in the natural parameters of q ? (u).
We propose a hybrid variational-sampling approach whereby approximate samples from q ? (x) are
obtained with a sequential Monte Carlo smoother. However, as discussed in section 3.4, depending
on the characteristics of the dynamical system, other smoothing methods could be more appropriate
[20]. As an alternative to smoothing on the auxiliary dynamical system in (12), one could force a
q(x) from a particular family of distributions and optimise the evidence lower bound with respect
to its variational parameters. For instance, we could posit a Gaussian q(x) with a sparsity pattern in
the covariance matrix assuming zero covariance between non-neighboring states and maximize the
ELBO with respect to the variational parameters.
We use stochastic gradient descent [10] to maximize the ELBO (where we have plugged in the
optimal q ? (u) [22]) by using its gradient with respect to the hyperparameters. Both quantities are
stochastic in our hybrid approach due to variance introduced by the sampling of q ? (x). In fact,
vanilla sequential Monte Carlo methods will result in biased estimators of the gradient and the
parameters of q ? (u). However, in our experiments this has not been an issue. Techniques such as
particle MCMC would be a viable alternative to conventional sequential Monte Carlo [13].
5
Algorithm 1 Variational learning of GP-SSMs with particle smoothing. Batch mode (i.e. non-SVI)
is the particular case where the mini-batch is the whole dataset.
Require: Observations y1:T . Initial values for ?, ? 1 and ? 2 . Schedules for ? and ?. i = 1.
repeat
y? :? 0 ? S AMPLE M INI BATCH(y1:T )
{x? :? 0 }L
sample from eq. (12)
l=1 ? G ET S AMPLES O PTIMAL QX(y? :? 0 , ?, ? 1 , ? 2 )
?? L ? G ET T HETAG RADIENT({x? :? 0 }L
,
?)
supp. material
l=1
? ?1 , ? ?2 ? G ET O PTIMAL QU({x? :? 0 }L
,
?)
eq.
(11) or (14)
l=1
? 1 ? ? 1 + ?i (? ?1 ? ? 1 )
? 2 ? ? 2 + ?i (? ?2 ? ? 2 )
? ? ? + ? i ?? L
i?i+1
until ELBO convergence
3.6
Making Predictions
One of the most appealing properties of our variational approach to learning GP-SSMs is that the
approximate predictive distribution of the state transition function can be cheaply computed
Z
Z
p(f? |x? , y) =
p(f? |x? , x, u) p(x|u, y) p(u|y) ?
p(f? |x? , u) p(x|u, y) q(u)
x,u
x,u
Z
=
p(f? |x? , u) q(u) = N (f? |A? ?, B? + A? ?A>
(13)
? ).
u
The derivation in eq. (13) contains two approximations: 1) predictions at new test points are considered to depend only on the inducing variables, and 2) the posterior distribution over u is approximated by a variational distribution.
After pre-computations, the cost of each prediction is O(M ) for the mean and O(M 2 ) for the
variance. This contrasts with the O(T L) and O(T 2 L) complexity
of approaches based on sampling
R
from the smoothing distribution where p(f? |x? , y) = x p(f? |x? , x) p(x|y) is approximated with
L samples from p(x|y) [8]. The variational approach condenses the learning of the latent function
on the inducing points u and does not explicitly need the smoothing distribution p(x|y) to make
predictions.
4
Stochastic Variational Inference
Stochastic variational inference (SVI) [10] can be readily applied using our evidence lower bound.
When the observed time series is long, it can be expensive to compute q ? (u) or the gradient of L with
?L
respect to the hyperparameters and inducing inputs. Since both q ? (u) and ??/z
depend linearly
1:M
on q(x) via sufficient statistics that contain a summation over all elements in the state trajectory,
we can obtain unbiased estimates of these sufficient statistics by using one or multiple segments
of the sequence that are sampled uniformly at random. However, obtaining q(x) also requires a
time complexity of O(T ). Yet, in practice, q(x) can be approximated by running the smoothing
algorithm locally around those segments. This can be justified by the fact that in a time series
context, the smoothing distribution at a particular time is not largely affected by measurements that
are far into the past or the future [20]. The natural parameters of q ? (u) can be estimated by using a
portion of the time series of length S
?
?
?0
?0
X
X
1
T
?1 T
? 1 = Q?1
hAT xt iq(xt ,xt?1 ) , ? 2 = ? ?K?1
hAT At?1 iq(xt?1 ) ? . (14)
uu + Q
S t=? t?1
2
S t=? t?1
5
Online Learning
Our variational approach to learn GP-SSMs also leads naturally to an online learning implementation. This is of particular interest in the context of dynamical systems as it is often the case that data
arrives in a sequential manner, e.g. a robot learning the dynamics of different objects by interacting
6
Table 1: Experimental evaluation of 1D nonlinear system. Unless otherwise stated, training times
are reported for a dataset with T = 500 and test times are given for a test set with 105 data points. All
pre-computations independent on test data are performed before timing the ?test time?. Predictive
log likelihoods are the average over the full test set. * our PMCMC code did not use fast updatesdowndates of the Cholesky factors during training. This does not affect test times.
Variational GP-SSM
Var. GP-SSM (SVI, T = 104 )
PMCMC GP-SSM [8]
GP-NARX [17]
GP-NARX + FITC [17, 18]
Linear (N4SID, [16])
Test RMSE
test
tr
log p(xtest
t+1 |xt , y0:T )
1.15
1.07
1.12
1.46
1.47
2.35
-1.61
-1.47
-1.57
-1.90
-1.90
-2.30
Train time
Test time
2.14 min
4.12 min
547 min*
0.22 min
0.17 min
0.01 min
0.14 s
0.14 s
421 s
3.85 s
0.23 s
0.11 s
with them. Online learning in a Bayesian setting consists in sequential application of Bayes rule
whereby the posterior after observing data up to time t becomes the prior at time t + 1 [2, 15].
In our case, this involves replacing the prior p(u) = N (u|0, Ku,u ) by the approximate posterior
N (u|?, ?) obtained in the previous step. The expressions for the update of the natural parameters
of q ? (u) with a new mini batch y? :? 0 are
?0
?0
X
1 ?1 X T
0
?1
T
0
hAt?1 At?1 iq(xt?1 ) .
(15)
?1 = ?1 + Q
hAt?1 xt iq(xt ,xt?1 ) , ? 2 = ? 2 ? Q
2
t=?
t=?
6
Experiments
The goal of this section is to showcase the ability of variational GP-SSMs to perform approximate
Bayesian learning of nonlinear dynamical systems. In particular, we want to demonstrate: 1) the
ability to learn the inherent nonlinear dynamics of a system, 2) the application in cases where the
latent states have higher dimensionality than the observations, and 3) the use of non-Gaussian likelihoods.
6.1 1D Nonlinear System
We apply our variational learning procedure presented above to the one-dimensional nonlinear system described by p(xt+1 |xt ) = N (f (xt ), 1) and p(yt |xt ) = N (xt , 1) where the transition function
is xt + 1 if x < 4 and ?4xt + 21 if x ? 4. Its pronounced kink makes it challenging to learn. Our
goal is to find a posterior distribution over this function using a GP-SSM with Mat?ern covariance
function. To solve the expectations with respect to the approximate smoothing distribution q(x) we
use a bootstrap particle fixed-lag smoother with 1000 particles and a lag of 10.
In Table 1, we compare our method (Variational GP-SSM) against the PMCMC sampling procedure from [8] taking 100 samples and 10 burn in samples. As in [8], the sampling exhibited very
good mixing with 20 particles. We also compare to an auto-regressive model based on Gaussian
process regression [17] of order 5 with Mat?ern ARD covariance function with and without FITC
approximation. Finally, we use a linear subspace identification method (N4SID, [16]) as a baseline for comparison. The PMCMC training offers the best test performance from all methods using
500 training points at the cost of substantial train and test time. However, if more data is available
(T = 104 ) the stochastic variational inference procedure can be very attractive since it improves test
performance while having a test time that is independent of the training set size. The reported SVI
performance has been obtained with mini-batches of 100 time-steps.
6.2 Neural Spike Train Recordings
We now turn to the use of SSMs to learn a simple model of neural activity in rats? hippocampus. We
use data in neuron cluster 1 (the most active) from experiment ec013.717 in [14]. In some regions
of the time series, the action potential spikes show a clear pattern where periods of rapid spiking
are followed by periods of very little spiking. We wish to model this behaviour as an autonomous
nonlinear dynamical system (i.e. one not driven by external inputs). Many parametric models of
nonlinear neuron dynamics have been proposed [11] but our goal here is to learn a model from data
7
30
20
0
10
940
940.5
941
940
time [s]
940.5
30
states
spike counts
40
states
spike counts
40
20
0
10
941
0
time [s]
0.5
1
0
prediction time [s]
0.5
1
prediction time [s]
x(1)
0
x(2)
0
x(2)
x(2)
x(2)
Figure 2: From left to right: 1) part of the observed spike count data, 2) sample from the corresponding smoothing distribution, 3) predictive distribution of spike counts obtained by simulating
the posterior dynamical from an initial state, and 4) corresponding latent states.
0
x(1)
0
x(1)
(2)
xt+1
x(1)
(1)
(2)
f (xt , xt ),
Figure 3: Contour plots of the state transition function
=
and trajectories in
state space. Left: mean posterior function and trajectory from smoothing distribution. Other three
panels: transition functions sampled from the posterior and trajectories simulated conditioned on
the corresponding sample. Those simulated trajectories start inside the limit cycle and are naturally
attracted towards it. Note how function samples are very similar in the region of the limit cycle.
without using any biological insight. We use a GP-SSM with a structure such that it is the discretetime analog of a second order nonlinear ordinary differential equation: two states one of which
is the derivative of the other. The observations are spike counts in temporal bins of 0.01 second
width. We use a Poisson likelihood relating the spike counts to the second latent state yt |xt ?
(2)
Poisson(exp(?xt + ?)).
We find a posterior distribution for the state transition function using our variational GP-SSM approach. Smoothing is done with a fixed-lag particle smoother and training until convergence takes
approximately 50 iterations of Algorithm 1. Figure 2 shows a part of the raw data together with an
approximate sample from the smoothing distribution during the same time interval. In addition, we
show the distribution over predictions made by chaining 1-step-ahead predictions. To make those
predictions we have switched off process noise (Q = 0) to show more clearly the effect of uncertainty in the state transition function. Note how the frequency of roughly 6 Hz present in the data is
well captured. Figure 3 shows how the limit cycle corresponding to a nonlinear dynamical system
has been captured (see caption for details).
7
Discussion and Future Work
We have derived a tractable variational formulation to learn GP-SSMs: an important class of models of nonlinear dynamical systems that is particularly suited to applications where a principled
parametric model of the dynamics is not available. Our approach makes it possible to learn very expressive models without risk of overfitting. In contrast to previous approaches [4, 12, 25], we have
demonstrated the ability to learn a nonlinear state transition function in a latent space of greater
dimensionality than the observation space. More crucially, our approach yields a tractable posterior
over nonlinear systems that, as opposed to those based on sampling from the smoothing distribution
[8], results in a computation time for the predictions that does not depend on the length of the time
series.
Given the interesting capabilities of variational GP-SSMs, we believe that future work is warranted.
In particular, we want to focus on structured variational distributions q(x) that could eliminate the
need to solve the smoothing problem in the auxiliary dynamical system at the cost of having more
variational parameters to optimize. On a more theoretical side, we would like to better characterize
GP-SSM priors in terms of their dynamical system properties: stability, equilibria, limit cycles, etc.
8
References
[1] C. M. Bishop. Pattern Recognition and Machine Learning. Springer, 2006.
[2] Tamara Broderick, Nicholas Boyd, Andre Wibisono, Ashia C Wilson, and Michael Jordan. Streaming
variational Bayes. In Advances in Neural Information Processing Systems 26, pages 1727?1735. Curran
Associates, Inc., 2013.
[3] Emery N. Brown, Loren M. Frank, Dengda Tang, Michael C. Quirk, and Matthew A. Wilson. A statistical
paradigm for neural spike train decoding applied to position prediction from ensemble firing patterns of
rat hippocampal place cells. The Journal of Neuroscience, 18(18):7411?7425, 1998.
[4] Andreas C. Damianou, Michalis Titsias, and Neil D. Lawrence. Variational Gaussian process dynamical
systems. In Advances in Neural Information Processing Systems 24, pages 2510?2518. 2011.
[5] J. Daunizeau, K.J. Friston, and S.J. Kiebel. Variational Bayesian identification and prediction of stochastic
nonlinear dynamic causal models. Physica D: Nonlinear Phenomena, 238(21):2089 ? 2118, 2009.
[6] M. P. Deisenroth and S. Mohamed. Expectation Propagation in Gaussian process dynamical systems. In
Advances in Neural Information Processing Systems (NIPS) 25, pages 2618?2626. 2012.
[7] M. P. Deisenroth, R. D. Turner, M. F. Huber, U. D. Hanebeck, and C. E. Rasmussen. Robust filtering and
smoothing with Gaussian processes. IEEE Transactions on Automatic Control, 57(7):1865 ?1871, 2012.
[8] Roger Frigola, Fredrik Lindsten, Thomas B. Sch?on, and Carl E. Rasmussen. Bayesian inference and
learning in Gaussian process state-space models with particle MCMC. In Advances in Neural Information
Processing Systems (NIPS) 26. 2013.
[9] Z. Ghahramani and S. Roweis. Learning nonlinear dynamical systems using an EM algorithm. In Advances in Neural Information Processing Systems (NIPS) 11. MIT Press, 1999.
[10] Matthew D Hoffman, David M Blei, Chong Wang, and John Paisley. Stochastic variational inference. The
Journal of Machine Learning Research, 14(1):1303?1347, 2013.
[11] Eugene M Izhikevich. Neural excitability, spiking and bursting. International Journal of Bifurcation and
Chaos, 10(06):1171?1266, 2000.
[12] Neil D. Lawrence and Andrew J. Moore. The hierarchical Gaussian process latent variable model. In
Zoubin Ghahramani, editor, Proceedings of the 24th International Conference on Machine Learning
(ICML), 2007.
[13] Fredrik Lindsten and Thomas B. Sch?on. Backward simulation methods for Monte Carlo statistical inference. Foundations and Trends in Machine Learning, 6(1):1?143, 2013.
[14] K. Mizuseki, A. Sirota, E. Pastalkova, K. Diba, and G. Buzski. Multiple single unit recordings from
different rat hippocampal and entorhinal regions while the animals were performing multiple behavioral
tasks. CRCNS.org. http://dx.doi.org/10.6080/K09G5JRZ, 2013.
[15] Manfred Opper. A bayesian approach to on-line learning. In David Saad, editor, On-Line Learning in
Neural Networks. Cambridge University Press, 1998.
[16] Van Overschee P. and De Moor B. Subspace Identification for Linear Systems, Theory, Implementation,
Applications. Kluwer Academic Publishers, 1996.
[17] J. Qui?nonero Candela, A Girard, J. Larsen, and C.E. Rasmussen. Propagation of uncertainty in Bayesian
kernel models - application to multiple-step ahead forecasting. In Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP ?03). 2003 IEEE International Conference on, volume 2, pages
II?701?4 vol.2, April 2003.
[18] J. Qui?nonero-Candela and C.E. Rasmussen. A unifying view of sparse approximate Gaussian process
regression. Journal of Machine Learning Research, 6:1939?1959, 2005.
[19] C. E. Rasmussen and C. K. I. Williams. Gaussian Processes for Machine Learning. MIT Press, 2006.
[20] S. S?arkk?a. Bayesian Filtering and Smoothing. Cambridge University Press, 2013.
[21] R. H. Shumway and D. S. Stoffer. Time Series Analysis and Its Applications. Springer, 3rd edition, 2011.
[22] Michalis Titsias. Variational learning of inducing variables in sparse Gaussian processes. In Proceedings
of the 12th International Conference on Artificial Intelligence and Statistics (AISTATS), 2009.
[23] R. Turner, M. P. Deisenroth, and C. E. Rasmussen. State-space inference and learning with Gaussian
processes. In Yee Whye Teh and Mike Titterington, editors, 13th International Conference on Artificial
Intelligence and Statistics, volume 9 of W&CP, pages 868?875, Chia Laguna, Sardinia, Italy, 2010.
[24] Harri Valpola and Juha Karhunen. An unsupervised ensemble learning method for nonlinear dynamic
state-space models. Neural Computation, 14(11):2647?2692, 2002.
[25] J.M. Wang, D.J. Fleet, and A. Hertzmann. Gaussian process dynamical models. In Advances in Neural
Information Processing Systems (NIPS) 18, pages 1441?1448. MIT Press, Cambridge, MA, 2006.
9
| 5375 |@word aircraft:1 hippocampus:1 simulation:1 crucially:1 covariance:7 xtest:1 tr:4 reduction:1 initial:2 series:14 att:2 contains:1 rightmost:2 past:2 existing:1 arkk:1 yet:1 dx:1 attracted:1 readily:1 kiebel:1 john:1 additive:3 plot:1 update:1 intelligence:2 manfred:1 regressive:2 blei:1 provides:1 location:2 ssm:17 simpler:2 org:2 pastalkova:1 become:1 differential:1 viable:1 consists:1 doubly:1 behavioral:1 inside:2 manner:2 x0:21 orbital:1 spacecraft:1 huber:1 rapid:1 roughly:1 behavior:2 mechanic:1 automatically:1 little:1 increasing:2 becomes:2 moreover:1 underlying:1 notation:4 panel:5 interpreted:1 developed:1 lindsten:2 whilst:1 titterington:1 temporal:1 finance:2 uk:1 control:5 unit:1 omit:3 before:1 engineering:3 understood:1 timing:1 limit:6 consequence:1 laguna:1 despite:1 firing:1 approximately:1 burn:1 dengda:1 bursting:1 challenging:1 smc:1 practice:1 block:1 svi:4 bootstrap:1 procedure:8 area:1 adapting:1 convenient:1 boyd:1 pre:2 radial:1 zoubin:1 risk:4 context:4 yee:1 optimize:1 conventional:4 map:2 deterministic:3 yt:12 maximizing:1 equivalent:2 straightforward:3 demonstrated:1 williams:1 simplicity:1 estimator:1 rule:1 insight:1 stability:1 autonomous:2 analogous:1 pt:2 heavily:1 user:1 exact:1 caption:1 carl:2 us:1 gps:2 curran:1 associate:1 synthesize:1 element:1 expensive:2 particularly:3 approximated:3 recognition:1 trend:1 showcase:1 vein:1 observed:2 ft:17 mike:1 wang:3 solved:1 region:3 cycle:6 trade:3 principled:2 substantial:1 complexity:5 ui:1 broderick:1 hertzmann:1 cam:1 dynamic:12 carrying:1 depend:4 segment:2 yutian:1 predictive:5 titsias:2 basis:1 easily:1 accelerate:1 joint:2 k0:5 multimodal:1 various:1 represented:1 icassp:1 regularizer:1 harri:1 derivation:1 train:4 fast:3 monte:8 kp:1 doi:1 artificial:2 lag:3 widely:1 supplementary:1 solve:2 elbo:5 otherwise:1 ability:5 statistic:5 neil:2 gp:36 jointly:1 online:5 sequence:1 propose:1 neighboring:1 combining:1 nonero:2 mixing:1 flexibility:1 achieve:2 roweis:2 subindices:1 radient:1 inducing:14 pronounced:1 kink:1 convergence:2 double:1 cluster:1 emery:1 object:1 illustrate:1 andrew:1 iq:10 augmenting:1 quirk:1 ac:1 depending:2 ard:1 qt:1 eq:4 strong:1 auxiliary:7 involves:1 indicate:1 uu:2 convention:2 fredrik:2 ptimal:2 posit:1 stochastic:11 material:2 bin:1 require:1 behaviour:1 f1:10 tighter:1 biological:1 summation:1 physica:1 around:1 considered:1 exp:4 equilibrium:1 mapping:1 lawrence:2 matthew:2 create:1 successfully:1 moor:1 nonidentifiability:1 hoffman:1 mit:3 clearly:1 gaussian:30 aim:1 avoid:1 factorizes:1 wilson:2 encode:1 derived:1 focus:3 likelihood:9 contrast:3 baseline:1 sense:1 inference:15 streaming:1 bt:4 eliminate:1 issue:2 flexible:2 augment:1 denoted:1 animal:1 smoothing:26 integration:1 bifurcation:1 marginal:1 having:4 sampling:7 cancel:1 icml:1 unsupervised:1 future:7 inherent:1 few:1 mizuseki:1 divergence:2 ourselves:1 ecology:1 interest:1 cer54:1 possibility:1 evaluation:1 severe:2 stoffer:1 chong:1 extreme:1 arrives:1 light:1 kt:8 integral:1 necessary:3 unless:1 plugged:1 arma:1 re:1 causal:1 theoretical:1 uncertain:1 instance:4 modeling:1 markovian:2 ordinary:1 cost:6 tractability:1 characterize:1 reported:2 straightforwardly:1 thanks:2 density:3 loren:1 international:5 probabilistic:3 off:4 decoding:1 michael:2 together:1 opposed:1 external:3 derivative:2 return:1 supp:1 potential:1 nonlinearities:1 amples:1 de:1 inc:1 explicitly:1 depends:1 later:1 view:2 performed:1 candela:2 linked:1 observing:1 portion:1 start:1 bayes:3 capability:1 rmse:1 formed:1 variance:2 characteristic:3 largely:1 ensemble:2 yield:1 generalize:1 n4sid:2 bayesian:13 identification:3 raw:1 carlo:8 trajectory:14 oscillatory:2 damianou:1 andre:1 against:1 frequency:1 tamara:1 mohamed:1 larsen:1 naturally:2 associated:1 conciseness:1 sampled:4 dataset:2 popular:2 knowledge:1 dimensionality:3 improves:1 schedule:1 appears:1 higher:1 methodology:1 specify:3 april:1 formulation:3 done:1 roger:2 arch:1 correlation:1 until:2 flight:1 hand:1 replacing:1 expressive:1 nonlinear:38 propagation:2 continuity:1 mode:1 believe:1 izhikevich:1 effect:1 brown:2 concept:1 contain:1 unbiased:1 analytically:1 excitability:1 leibler:2 moore:1 conditionally:1 attractive:1 during:2 width:1 whereby:2 chaining:1 rat:3 leftmost:2 hippocampal:2 ini:1 whye:1 demonstrate:1 cp:1 variational:46 chaos:1 recently:1 functional:1 spiking:3 conditioning:2 volume:2 discussed:1 analog:1 relating:3 kluwer:1 measurement:1 cambridge:4 paisley:1 smoothness:2 rd:3 vanilla:1 automatic:1 particle:9 robot:1 etc:2 posterior:19 multivariate:1 perspective:1 optimizing:1 italy:1 driven:1 success:1 vt:2 captured:2 additional:1 greater:1 ssms:15 arx:1 maximize:3 period:2 paradigm:1 signal:2 ii:1 smoother:4 multiple:5 full:1 match:1 adapt:1 academic:1 offer:3 long:3 chia:1 prediction:17 regression:2 expectation:3 poisson:2 iteration:1 represent:1 tailored:1 arisen:1 kernel:3 robotics:1 sometimes:1 cell:1 preserved:1 addition:2 whereas:2 justified:1 want:2 interval:1 publisher:1 sch:2 fifty:1 rest:2 biased:1 exhibited:1 daunizeau:1 saad:1 recording:2 gpssms:1 hz:1 hatt:2 ample:1 jordan:1 enough:1 easy:1 affect:1 zi:2 restrict:2 andreas:1 fleet:1 expression:1 forecasting:1 resistance:1 speech:1 action:1 useful:3 clear:1 nonparametric:3 locally:1 induces:1 discretetime:1 http:1 zj:1 neuroscience:2 estimated:1 diverse:1 discrete:1 mat:2 vol:1 affected:1 group:1 four:1 backward:1 year:1 parameterized:2 uncertainty:4 place:2 family:2 almost:1 qui:2 bound:10 followed:1 tackled:1 aerodynamic:1 activity:1 ahead:2 precisely:1 generates:1 simulate:1 argument:2 min:6 performing:1 ern:2 department:1 developing:1 structured:1 smaller:1 em:1 y0:1 appealing:1 qu:1 making:2 taken:1 equation:4 turn:1 count:6 know:1 tractable:7 available:3 apply:1 hierarchical:2 away:1 appropriate:2 simulating:1 nicholas:1 alternative:3 batch:5 hat:4 original:1 thomas:2 denotes:2 running:1 michalis:2 unifying:1 narx:2 ghahramani:3 build:1 quantity:1 spike:9 parametric:10 strategy:2 traditional:1 gradient:4 subspace:2 valpola:1 simulated:2 capacity:3 assuming:1 length:6 kalman:1 modeled:1 code:1 mini:3 minimizing:1 difficult:2 hlog:1 frank:1 ashia:1 stated:1 implementation:2 unknown:3 perform:2 teh:1 observation:10 neuron:2 juha:1 descent:1 situation:2 extended:2 y1:4 interacting:1 arbitrary:1 hanebeck:1 introduced:1 david:2 specified:1 kl:2 acoustic:1 learned:2 nip:4 dynamical:30 pattern:4 sparsity:1 built:1 optimise:1 overschee:1 natural:5 hybrid:3 force:1 friston:1 turner:2 scheme:1 fitc:2 carried:1 auto:2 prior:13 eugene:1 sardinia:1 kf:6 law:1 shumway:1 fully:1 frigola:4 interesting:1 filtering:3 proportional:1 var:1 foundation:1 switched:1 sufficient:3 article:2 editor:3 pmcmc:4 repeat:1 rasmussen:7 side:1 allow:1 taking:2 sparse:7 van:1 opper:1 transition:14 rich:1 contour:1 qualitatively:2 made:1 employing:1 far:1 qx:1 transaction:1 approximate:12 kullback:2 overfitting:6 active:1 factorize:1 alternatively:1 latent:16 quantifies:1 table:2 learn:11 ku:6 robust:1 obtaining:1 warranted:1 complex:2 did:1 aistats:1 main:1 linearly:1 whole:1 noise:7 arise:1 hyperparameters:6 edition:1 girard:1 augmented:3 fig:1 overconfidence:1 referred:1 crcns:1 position:1 explicit:1 wish:1 tang:1 load:1 xt:62 bishop:1 showing:1 evidence:8 restricting:1 sequential:9 entorhinal:1 conditioned:3 karhunen:1 chen:1 mildly:1 mf:4 suited:1 entropy:1 led:1 cheaply:1 conveniently:1 springer:2 corresponds:1 ma:1 goal:4 towards:1 uniformly:1 averaging:1 experimental:1 deisenroth:3 cholesky:1 brevity:1 avoiding:1 wibisono:1 mcmc:3 phenomenon:1 |
4,833 | 5,376 | Gaussian Process Volatility Model
Jos?e Miguel Hern?andez Lobato
Cambridge University
[email protected]
Yue Wu
Cambridge University
[email protected]
Zoubin Ghahramani
Cambridge University
[email protected]
Abstract
The prediction of time-changing variances is an important task in the modeling of
financial data. Standard econometric models are often limited as they assume rigid
functional relationships for the evolution of the variance. Moreover, functional
parameters are usually learned by maximum likelihood, which can lead to overfitting. To address these problems we introduce GP-Vol, a novel non-parametric
model for time-changing variances based on Gaussian Processes. This new model
can capture highly flexible functional relationships for the variances. Furthermore,
we introduce a new online algorithm for fast inference in GP-Vol. This method
is much faster than current offline inference procedures and it avoids overfitting
problems by following a fully Bayesian approach. Experiments with financial data
show that GP-Vol performs significantly better than current standard alternatives.
1
Introduction
Time series of financial returns often exhibit heteroscedasticity, that is the standard deviation or
volatility of the returns is time-dependent. In particular, large returns (either positive or negative) are
often followed by returns that are also large in size. The result is that financial time series frequently
display periods of low and high volatility. This phenomenon is known as volatility clustering [1].
Several univariate models have been proposed in the literature for capturing this property. The best
known and most popular is the Generalised Autoregressive Conditional Heteroscedasticity model
(GARCH) [2]. An alternative to GARCH are stochastic volatility models [3]. However, there is no
evidence that SV models have better predictive performance than GARCH [4, 5, 6].
GARCH has further inspired a host of variants and extensions. A review of many of these models
can be found in [7]. Most of these GARCH variants attempt to address one or both limitations of
GARCH: a) the assumption of a linear dependency between current and past volatilities, and b)
the assumption that positive and negative returns have symmetric effects on volatility. Asymmetric
effects are often observed, as large negative returns often send measures of volatility soaring, while
this effect is smaller for large positive returns [8, 9]. Finally, there are also extensions that use
additional data besides daily closing prices to improve volatility predictions [10].
Most solutions proposed in these variants of GARCH involve: a) introducing nonlinear functional
relationships for the evolution of volatility, and b) adding asymmetric effects in these functional
relationships. However, the GARCH variants do not fundamentally address the problem that the
specific functional relationship of the volatility is unknown. In addition, these variants can have a
high number of parameters, which may lead to overfitting when using maximum likelihood learning.
More recently, volatility modeling has received attention within the machine learning community,
with the development of copula processes [11] and heteroscedastic Gaussian processes [12]. These
1
models leverage the flexibility of Gaussian Processes [13] to model the unknown relationship between the variances. However, these models do not address the asymmetric effects of positive and
negative returns on volatility.
We introduce a new non-parametric volatility model, called the Gaussian Process Volatility Model
(GP-Vol). This new model is more flexible, as it is not limited by a fixed functional form. Instead, a
non-parametric prior distribution is placed on possible functions, and the functional relationship is
learned from the data. This allows GP-Vol to explicitly capture the asymmetric effects of positive
and negative returns on volatility. Our new volatility model is evaluated in a series of experiments
with real financial returns, and compared against popular econometric models, namely, GARCH,
EGARCH [14] and GJR-GARCH [15]. In these experiments, GP-Vol produces the best overall
predictions. In addition to this, we show that the functional relationship learned by GP-Vol often
exhibits the nonlinear and asymmetric features that previous models attempt to capture.
The second main contribution of the paper is the development of an online algorithm for learning
GP-Vol. GP-Vol is an instance of a Gaussian Process State Space Model (GP-SSM). Previous work
on GP-SSMs [16, 17, 18] has mainly focused on developing approximation methods for filtering
and smoothing the hidden states in GP-SSM, without jointly learning the GP transition dynamics.
Only very recently have Frigola et al. [19] addressed the problem of learning both the hidden states
and the transition dynamics by using Particle Gibbs with Ancestor Sampling (PGAS) [20]. In this
paper, we introduce a new online algorithm for performing inference on GP-SSMs. Our algorithm
has similar predictive performance as PGAS on financial data, but is much faster.
2
Review of GARCH and GARCH variants
The standard variance model for financial data is GARCH. GARCH assumes a Gaussian observation
model and a linear transition function for the variance: the time-varying variance ?t2 is linearly
dependent on p previous variance values and q previous squared time series values, that is,
Pq
Pp
2
xt ? N (0, ?t2 ) ,
and
?t2 = ?0 + j=1 ?j x2t?j + i=1 ?i ?t?i
,
(1)
where xt are the values of the return time series being modeled. This model is flexible and can
produce a variety of clustering behaviors of high and low volatility periods for different settings
of ?1 , . . . , ?q and ?1 , . . . , ?p . However, it has several limitations. First, only linear relationships
2
between ?t?p:t?1
and ?t2 are allowed. Second, past positive and negative returns have the same
effect on ?t2 due to the quadratic term x2t?j . However, it is often observed that large negative returns
lead to larger rises in volatility than large positive returns [8, 9].
A more flexible and often cited GARCH extension is Exponential GARCH (EGARCH) [14]. The
equation for ?t2 is now:
Pq
Pp
2
log(?t2 ) = ?0 + j=1 ?j g(xt?j ) + i=1 ?i log(?t?i
) , where g(xt ) = ?xt + ? |xt | . (2)
Asymmetry in the effects of positive and negative returns is introduced through the function g(xt ). If
the coefficient ? is negative, negative returns will increase volatility, while the opposite will happen
if ? is positive. Another GARCH extension that models asymmetric effects is GJR-GARCH [15]:
Pq
Pp
Pr
2
?t2 = ?0 + j=1 ?j x2t?j + i=1 ?i ?t?i
+ k=1 ?k x2t?k It?k ,
(3)
where It?k = 0 if xt?k ? 0 and It?k = 1 otherwise. The asymmetric effect is now captured by
It?k , which is nonzero if xt?k < 0.
3
Gaussian process state space models
GARCH, EGARCH and GJR-GARCH can be all represented as General State-Space or Hidden
Markov models (HMM) [21, 22], with the unobserved dynamic variances being the hidden states.
Transition functions for the hidden states are fixed and assumed to be linear in these models. The
linear assumption limits the flexibility of these models.
More generally, a non-parametric approach can be taken where a Gaussian Process (GP) prior is
placed on the transition function, so that its functional form can be learned from data. This Gaussian
Process state space model (GP-SSM) is a generalization of HMM. GP-SSM and HMM differ in two
main ways. First, in HMM the transition function has a fixed functional form, while in GP-SSM
2
3
2.5
truth
GP?Vol 5%
GP?Vol 95%
2.5
2
1.5
1.5
1
1
0.5
0.5
0
0
0
20
40
60
80
Number of Observations
truth
GP?Vol 5%
GP?Vol 95%
2
100
?0.5
0
20
40
60
80
100
Number of Observations
Figure 1: Left, graphical model for GP-Vol. The transitions of the hidden states vt is represented by
the unknown function f . f takes as inputs the previous state vt?1 and previous observation xt?1 .
Middle, 90% posterior interval for a. Right, 90% posterior interval for b.
it is represented by a GP. Second, in GP-SSM the states do not have Markovian structure once the
transition function is marginalized out.
The flexibility of GP-SSMs comes at a cost: inference in GP-SSMs is computationally challenging.
Because of this, most of the previous work on GP-SSMs [16, 17, 18] has focused on filtering and
smoothing the hidden states in GP-SSM, without jointly learning the GP dynamics. Note that in
[18], the authors learn the dynamics, but using a separate dataset in which both input and target
values for the GP model are observed. A few papers considered learning both the GP dynamics and
the hidden states for special cases of GP-SSMs. For example, [23] applied EM to obtain maximum
likelihood estimates for parametric systems that can be represented by GPs. A general method has
been recently proposed for joint inference on the hidden states and the GP dynamics using Particle
Gibbs with Ancestor Sampling (PGAS) [20, 19]. However, PGAS is a batch MCMC inference
method that is computationally very expensive.
4
Gaussian process volatility model
Our new Gaussian Process Volatility Model (GP-Vol) is an instance of GP-SSM:
xt ? N (0, ?t2 ) ,
vt := log(?t2 ) = f (vt?1 , xt?1 ) + t ,
t ? N (0, ?n2 ) .
(4)
Note that we model the logarithm of the variance, which has real support. Equation (4) defines
a GP-SMM. We place a GP prior on the transition function f . Let zt = (vt , xt ). Then f ?
GP(m, k) where m(zt ) and k(zt , zt0 ) are the GP mean and covariance functions, respectively. The
mean function can encode prior knowledge of the system dynamics. The covariance function gives
the prior covariance between function values: k(zt , zt0 ) = Cov(f (zt ), f (zt0 )) . Intuitively if zt and zt0
are close to each other, the covariances between the corresponding function values should be large:
f (zt ) and f (zt0 ) should be highly correlated.
The graphical model for GP-Vol is given in Figure 1. The explicit dependence of transition function
values on the previous return xt?1 enables GP-Vol to model the asymmetric effects of positive and
negative returns on the variance evolution. GP-Vol can be extended to depend on p previous log
variances and q past returns like in GARCH(p,q). In this case, the transition would be of the form
vt = f (vt?1 , vt?2 , ..., vt?p , xt?1 , xt?2 , ..., xt?q ) + t .
5
Bayesian inference in GP-Vol
In the standard GP regression setting, the inputs and targets are fully observed and f can be learned
using exact Bayesian inference [13]. However, this is not the case in GP-Vol, where the unknown
{vt } form part of the inputs and all the targets. Let ? denote the model hyper-parameters and let
f = [f (v1 ), . . . , f (vT )]. Directly learning the joint posterior of the unknown variables f , v1:T and
? is a challenging task. Fortunately, the posterior p(vt |?, x1:t ), where f has been marginalized out,
can be approximated with particles [24]. We first describe a standard sequential Monte Carlo (SMC)
particle filter to learn this posterior.
i
Let {v1:t?1
}N
i=1 be particles representing chains of states up to t ? 1 with corresponding normalized
i
weights Wt?1 . The posterior p(v1:t?1 |?, x1:t?1 ) is then approximated by
PN
i
i
p?(v1:t?1 |?, x1:t?1 ) = i=1 Wt?1
?v1:t?1
(v1:t?1 ) .
(5)
3
The corresponding posterior for v1:t can be approximated by propagating these particles forward.
For this, we propose new states from the GP-Vol transition model and then we importance-weight
j
them according to the GP-Vol observation model. Specifically, we resample particles v1:t?1
from
j
(5) according to their weights Wt?1 , and propagate the samples forward. Then, for each of the
j
, x1:t?1 ), which is the GP predictive
particles propagated forward, we propose vtj from p(vt |?, v1:t?1
distribution. The proposed particles are then importance-weighted according to the observation
model, that is, Wtj ? p(xt |?, vtj ) = N (xt |0, exp{vtj }).
The above setup assumes that ? is known. To learn these hyper-parameters, we can also encode them
in particles and filter them together with the hidden states. However, since ? is constant across time,
naively filtering such particles without regeneration will fail due to particle impoverishment, where
a few or even one particle receives all the weight. To solve this problem, the Regularized Auxiliary
Particle Filter (RAPF) regenerates parameter particles by performing kernel smoothing operations
[25]. This introduces artificial dynamics and estimation bias. Nevertheless, RAPF has been shown
to produce state-of-the-art inference in multivariate parametric financial models [6].
RAPF was designed for HMMs, but GP-Vol is non-Markovian once f is marginalized out. Therefore,
we design a new version of RAPF for non-Markovian systems and refer to it as the Regularized
Auxiliary Particle Chain Filter (RAPCF), see Algorithm 1. There are two main parts in RAPCF.
First, there is the Auxiliary Particle Filter (APF) part in lines 5, 6 and 7 of the pseudocode [26].
This part selects particles associated with high expected likelihood, as given by the new expected
state in (7) and the corresponding resampling weight in (8). This bias towards particles with high
expected likelihood is eliminated when the final importance weights are computed in (9). The most
promising particles are propagated forward in lines 8 and 9. The main difference between RAPF and
i
RAPCF is in the effect that previous states v1:t?1
have in the propagation of particles. In RAPCF
all the previous states determine the probabilities of the particles being propagated, as the model is
i
non-Markovian, while in RAPF these probabilities are only determined by the last state vt?1
. The
second part of RAPCF avoids particle impoverishment in ?. For this, new particles are generated
in line 10 by sampling from a Gaussian kernel. The over-dispersion introduced by these artificial
dynamics is eliminated in (6) by shrinking the particles towards their empirical average. We fix the
shrinking parameter ? to be 0.95. In practice, we found little difference in predictions when we
varied ? from 0.99 to 0.95.
RAPCF has limitations similar to those of RAPF. First, it introduces bias as sampling from the
kernel adds artificial dynamics. Second, RAPCF only filters forward and does not smooth backward.
Consequently, there will be impoverishment in distant ancestors vt?L , since these states are not
regenerated. When this occurs, GP-Vol will consider the collapsed ancestor states as inputs with
little uncertainty and the predictive variance near these inputs will be underestimated. These issues
can be addressed by adopting a batch MCMC approach. In particular, Particle Markov Chain Monte
Carlo (PMCMC) procedures [24] established a framework for learning the states and the parameters
in general state space models. Additionally, [20] developed a PMCMC algorithm called Particle
Gibbs with ancestor sampling (PGAS) for learning non-Markovian state space models. PGAS was
applied by [19] to learn GP-SSMs. These batch MCMC methods are computationally much more
expensive than RAPCF. Furthermore, our experiments show that in the GP-Vol model, RAPCF and
PGAS have similar empirical performance, while RAPCF is orders of magnitude faster than PGAS.
This indicates that the aforementioned issues have limited impact in practice.
6
Experiments
We performed three sets of experiments. First, we tested on synthetic data whether we can jointly
learn the hidden states and transition dynamics in GP-Vol using RAPCF. Second, we compared
the performance of GP-Vol against standard econometric models GARCH, EGARCH and GJRGARCH on fifty real financial time series. Finally, we compared RAPCF with the batch MCMC
method PGAS in terms of accuracy and execution time. The code for RAPCF in GP-Vol is publicly
available at http://jmhl.org.
6.1
Experiments with synthetic data
We generated ten synthetic datasets of length T = 100 according to (4). The transition function f is
sampled from a GP prior specified with a linear mean function and a squared exponential covariance
4
Algorithm 1 RAPCF
1: Input: data x1:T , number of particles N , shrinkage parameter 0 < ? < 1, prior p(?).
2: Sample N parameter particles from the prior: {?0i }i=1,...,N ? p(?).
3: Set initial importance weights, W0i = 1/N .
4: for t = 1 to T do
PN
i
i
5:
Shrink parameter particles towards their empirical mean ??t?1 = i=1 Wt?1
?t?1
by setting
i
i
e
?
? = ??
+ (1 ? ?)?t?1 .
(6)
t
t?1
Compute the new expected states:
i
?it = E(vt |?eti , v1:t?1
, x1:t?1 ) .
(7)
Compute importance weights proportional to the likelihood of the new expected states:
i
gti ? Wt?1
p(xt |?it , ?eti ) .
(8)
Resample N auxiliary indices {j} according to weights {gti }.
j
}j?J .
Propagate the corresponding chains of hidden states forward, that is, {v1:t?1
Add jitter: ?tj ? N (?etj , (1 ? ?2 )Vt?1 ), where Vt?1 is the empirical covariance of ?t?1 .
j
Propose new states vtj ? p(vt |?tj , v1:t?1
, x1:t?1 ).
Compute importance weights adjusting for the modified proposal:
(9)
Wtj ? p(xt |vtj , ?tj )/p(xt |?jt , ?etj ) ,
end for
j
Output: particles for chains of states v1:T
, particles for parameters ?tj and particle weights Wtj .
6:
7:
8:
9:
10:
11:
12:
13:
14:
function. The linear mean function is E(vt ) = m(vt?1 , xt?1 ) = avt?1 + bxt?1 . The squared
exponential covariance function is k(y, z) = ? exp(?0.5|y ? z|2 /l2 ) where l is the length-scale
parameter and ? is the amplitude parameter.
We used RAPCF to learn the hidden states v1:T and the hyper-parameters ? = (a, b, ?n , ?, l) using
non-informative diffuse priors for ?. In these experiments, RAPCF successfully recovered the state
and the hyper-parameter values. For the sake of brevity, we only include two typical plots of the 90%
posterior intervals for hyper-parameters a and b in the middle and right of Figures 1. The intervals
are estimated from the filtered particles for a and b at each time step t. In both plots, the posterior
intervals eventually concentrate around the true parameter values, shown as dotted blue lines.
6.2
Experiments with real data
We compared the predictive performances of GP-Vol, GARCH, EGARCH and GJR-GARCH on real
financial datasets. We used GARCH(1,1), EGARCH(1,1) and GJR-GARCH(1,1,1) models since
these variants have the least number of parameters and are consequently less affected by overfitting
problems. We considered fifty datasets, consisting of thirty daily Equity and twenty daily foreign
exchange (FX) time series. For the Equity series, we used daily closing prices. For FX, which
operate 24h a day, with no official daily closing prices, we cross-checked different pricing sources
and took the consensus price up to 4 decimal places at 10am New York, which is the time with
most market liquidity. Each of the resulting time series contains a total of T = 780 observations
from January 2008 to January 2011. The price data p1:T was pre-processed to eliminate prices
corresponding to times when markets were closed or not liquid. After this, prices were converted
into logarithmic returns, xt = log(pt /pt?1 ). Finally, the resulting returns were standardized to have
zero mean and unit standard deviation.
During the experiments, each method receives an initial time series of length 100. The different
models are trained on that data and then a one-step forward prediction is made. The performance of
each model is measured in terms of the predictive log-likelihood on the first return out of the training
set. Then the training set is augmented with the new observation and the training and prediction steps
are repeated. The whole process is repeated sequentially until no further data is received.
GARCH, EGARCH and GJR-GARCH were implemented using numerical optimization routines
provided by Kevin Sheppard 1 . A relatively long initial time series of length 100 was needed to
to train these models. Using shorter initial data resulted in wild jumps in the maximum likelihood
1
http:///www.kevinsheppard.com/wiki/UCSD_GARCH/
5
Nemenyi Test
CD
GJR
GP?VOL
GARCH
EGARCH
1
2
3
4
Figure 2: Comparison between GP-Vol, GARCH, EGARCH and GJR-GARCH via a Nemenyi test.
The figure shows the average rank across datasets of each method (horizontal axis). The methods
whose average ranks differ more than a critical distance (segment labeled CD) show significant
differences in performance at this confidence level. When the performances of two methods are
statistically different, their corresponding average ranks appear disconnected in the figure.
estimates of the model parameters. These large fluctuations produced very poor one-step forward
predictions. By contrast, GP-Vol is less susceptible to overfitting since it approximates the posterior
distribution using RAPCF instead of finding point estimates of the model parameters. We placed
broad non-informative priors on ? = (a, b, ?n , ?, l) and used N = 200 particles and shrinkage
parameter ? = .95 in RAPCF.
Dataset
GARCH EGARCH GJR
AUDUSD
BRLUSD
CADUSD
CHFUSD
CZKUSD
EURUSD
GBPUSD
IDRUSD
JPYUSD
KRWUSD
MXNUSD
MYRUSD
NOKUSD
NZDUSD
PLNUSD
SEKUSD
SGDUSD
TRYUSD
TWDUSD
ZARUSD
?1.303
?1.203
?1.402
?1.375
?1.422
?1.418
?1.382
?1.223
?1.350
?1.189
?1.220
?1.394
?1.416
?1.369
?1.395
?1.403
?1.382
?1.224
?1.384
?1.318
?1.514
?1.227
?1.409
?1.404
?1.473
?2.120
?3.511
?1.244
?2.704
?1.168
?3.438
?1.412
?1.567
?3.036
?1.385
?3.705
?2.844
?1.461
?1.377
?1.344
?1.305
?1.201
?1.402
?1.404
?1.422
?1.426
?1.386
?1.209
?1.355
?1.209
?1.278
?1.395
?1.419
?1.379
?1.382
?1.402
?1.398
?1.238
?1.388
?1.301
Table 1: FX series.
GP-Vol
?1.297
?1.180
?1.386
?1.359
?1.456
?1.403
?1.385
?1.039
?1.347
?1.154
?1.167
?1.392
?1.416
?1.389
?1.393
?1.407
?1.393
?1.236
?1.294
?1.304
Dataset GARCH EGARCH GJR
A
AA
AAPL
ABC
ABT
ACE
ADBE
ADI
ADM
ADP
ADSK
AEE
AEP
AES
AET
?1.304
?1.228
?1.234
?1.341
?1.295
?1.084
?1.335
?1.373
?1.228
?1.229
?1.345
?1.292
?1.151
?1.237
?1.285
?1.449
?1.280
?1.358
?1.976
?1.527
?2.025
?1.501
?1.759
?1.884
?1.720
?1.604
?1.282
?1.177
?1.319
?1.302
?1.281
?1.230
?1.219
?1.344
?1.3003
?1.106
?1.386
?1.352
?1.223
?1.205
?1.340
?1.263
?1.146
?1.234
?1.269
GP-Vol
?1.282
?1.218
?1.212
?1.337
?1.302
?1.073
?1.302
?1.356
?1.223
?1.211
?1.316
?1.166
?1.142
?1.197
?1.246
Table 2: Equity series 1-15.
Dataset GARCH EGARCH
AFL
AGN
AIG
AIV
AIZ
AKAM
AKS
ALL
ALTR
AMAT
AMD
AMGN
AMP
AMT
AMZN
?1.057
?1.270
?1.151
?1.111
?1.423
?1.230
?1.030
?1.339
?1.286
?1.319
?1.342
?1.191
?1.386
?1.206
?1.206
?1.126
?1.338
?1.256
?1.147
?1.816
?1.312
?1.034
?3.108
?1.443
?1.465
?1.348
?1.542
?1.444
?1.820
?1.567
GJR
GP-Vol
?1.061
?1.261
?1.195
?1.1285
?1.469
?1.229
?1.052
?1.316
?1.277
?1.332
?1.332
?1.1772
?1.365
?1.3658
?1.3537
?0.997
?1.274
?1.069
?1.133
?1.362
?1.246
?1.015
?1.327
?1.282
?1.310
?1.243
?1.189
?1.317
?1.210
?1.342
Table 3: Equity series 16-30.
We show the average predictive log-likelihood of GP-Vol, GARCH, EGARCH and GJR-GARCH in
tables 1, 2 and 3 for the FX series, the first 15 Equity series and the last 15 Equity series, respectively.
The results of the best performing method in each dataset have been highlighted in bold. These tables
show that GP-Vol obtains the highest predictive log-likelihood in 29 of the 50 analyzed datasets. We
perform a statistical test to determine whether differences among GP-Vol, GARCH, EGARCH and
GJR-GARCH are significant. These methods are compared against each other using the multiple
comparison approach described by [27]. In this comparison framework, all the methods are ranked
according to their performance on different tasks. Statistical tests are then applied to determine
whether the differences among the average ranks of the methods are significant. In our case, each of
the 50 datasets analyzed represents a different task. A Friedman rank sum test rejects the hypothesis
that all methods have equivalent performance at ? = 0.05 with p-value less than 10?15 . Pairwise
comparisons between all the methods with a Nemenyi test at a 95% confidence level are summarized
in Figure 2. The Nemenyi test shows that GP-Vol is significantly better than the other methods.
The other main advantage of GP-Vol over existing models is that it can learn the functional relationship f between the new log variance vt and the previous log variance vt?1 and previous return xt?1 .
We plot a typical log variance surface in the left of Figure 3. This surface is generated by plotting the
mean predicted outputs vt against a grid of inputs for vt?1 and xt?1 . For this, we use the functional
dynamics learned with RAPCF on the AUDUSD time series. AUDUSD stands for the amount of
US dollars that an Australian dollar can buy. The grid of inputs is designed to contain a range of
values experienced by AUDUSD from 2008 to 2011, which is the period covered by the data. The
surface is colored according to the standard deviation of the posterior predictive distribution for the
log variance. Large standard deviations correspond to uncertain predictions, and are redder.
6
Cross section vt vs xt?1
Cross section vt vs vt?1
Log Variance Surface for AUDUSD
0.35
4
4
2
0.25
0
0.2
?2
0.15
?4
1.5
1
0
vt
2
vt
output, vt
2
3
0.3
1
?1
?2
0.5
?3
5
0
input, xt?1
?5
?2
0
2
4
input, vt?1
0.1
?4
0
?5
?6
?4
?2
0
vt?1
2
4
6
?6
?4
?2
0
2
4
6
xt?1
Figure 3: Left, surface generated by plotting the mean predicted outputs vt against a grid of inputs
for vt?1 and xt?1 . Middle, predicted vt ? 2 s.d. for inputs (0, xt?1 ). Right, predicted vt ? 2 s.d.
for inputs (0, xt?1 ).
The plot in the left of Figure 3 shows several patterns. First, there is an asymmetric effect of positive
and negative previous returns xt?1 . This can be seen in the skewness and lack of symmetry of the
contour lines with respect to the vt?1 axis. Second, the relationship between vt?1 and vt is slightly
non-linear because the distance between consecutive contour lines along the vt?1 axis changes as we
move across those lines, especially when xt?1 is large. In addition, the relationship between xt?1
and vt is nonlinear, but some sort of skewed quadratic function. These two patterns confirm the
asymmetric effect and the nonlinear transition function that EGARCH and GJR-GARCH attempt
to model. Third, there is a dip in predicted log variance for vt?1 < ?2 and ?1 < xt?1 < 2.5.
Intuitively this makes sense, as it corresponds to a calm market environment with low volatility.
However, as xt?1 becomes more extreme the market becomes more turbulent and vt increases.
To further understand the transition function f we study cross sections of the log variance surface.
First, vt is predicted for a grid of vt?1 and xt?1 = 0 in the middle plot of Figure 3. Next, vt is
predicted for various xt?1 and vt?1 = 0 in the right plot of Figure 3. The confidence bands in the
figures correspond to the mean prediction ?2 standard deviations. These cross sections confirm the
nonlinearity of the transition function and the asymmetric effect of positive and negative returns on
the log variance. The transition function is slightly non-linear as a function of vt?1 as the band in
the middle plot of Figure 3 passes through (?2, ?2) and (0, 0), but not (2, 2). Surprisingly, we
observe in the right plot of Figure 3 that large positive xt?1 produces larger vt when vt?1 = 0 since
the band is slightly higher at xt?1 = 6 than at xt?1 = ?6. However, globally, the highest predicted
vt occurs when vt?1 > 5 and xt?1 < ?5, as shown in the surface plot.
6.3
Comparison between RAPCF and PGAS
We now analyze the potential shortcomings of RAPCF that were discussed in Section 5. For this,
we compare RAPCF against PGAS on the twenty FX time series from the previous section in terms
of predictive log-likelihood and execution times. The RAPCF setup is the same as in Section 6.2.
For PGAS, which is a batch method, the algorithm is run on initial training data x1:L , with L = 100,
and a one-step forward prediction is made. The predictive log-likelihood is evaluated on the next
observation out of the training set. Then the training set is augmented with the new observation
and the batch training and prediction steps are repeated. The process is repeated sequentially until
no further data is received. For these experiments we used shorter time series with T = 120 since
PGAS is computationally very expensive. Note that we cannot simply learn the GP-SSM dynamics
on a small set of training data and then predict on a large test dataset, as it was done in [19]. These
authors were able to predict forward as they were using synthetic data with known ?hidden? states.
We analyze different settings of RAPCF and PGAS. In RAPCF we use N = 200 particles since that
number was used to compare against GARCH, EGARCH and GJR-GARCH in the previous section.
PGAS has two parameters: a) N , the number of particles and b) M , the number of iterations.
Three combinations of these settings were used. The resulting average predictive log-likelihoods for
RAPCF and PGAS are shown in Table 4. On each dataset, the results of the best performing method
7
have been highlighted in bold. The average rank of each method across the analyzed datasets is
shown in Table 5. From these tables, there is no evidence that PGAS outperforms RAPCF on these
financial datasets, since there is no clear predictive edge of any PGAS setting over RAPCF.
Dataset
AUDUSD
BRLUSD
CADUSD
CHFUSD
CZKUSD
EURUSD
GBPUSD
IDRUSD
JPYUSD
KRWUSD
MXNUSD
MYRUSD
NOKUSD
NZDUSD
PLNUSD
SEKUSD
SGDUSD
TRYUSD
TWDUSD
ZARUSD
RAPCF PGAS.1
N = 200 N = 10
M = 100
?1.1205 ?1.0571
?1.0102 ?1.0043
?1.4174 ?1.4778
?1.8431 ?1.8536
?1.2263 ?1.2357
?1.3837 ?1.4586
?1.1863 ?1.2106
?0.5446 ?0.5220
?2.0766 ?1.9286
?1.0566 ?1.1212
?0.2417 ?0.2731
?1.4615 ?1.5464
?1.3095 ?1.3443
?1.2254 ?1.2101
?0.8972 ?0.8704
?1.0085 ?1.0085
?1.6229 ?1.9141
?1.8336 ?1.8509
?1.7093 ?1.7178
?1.3236 ?1.3326
PGAS.2
N = 25
M = 100
?1.0699
?0.9959
?1.4514
?1.8453
?1.2424
?1.3717
?1.1790
?0.5388
?2.1585
?1.2032
?0.2271
?1.4745
?1.3048
?1.2366
?0.8708
?1.0505
?1.7566
?1.8352
?1.8315
?1.3440
PGAS.3
N = 10
M = 200
?1.0936
?0.9759
?1.4077
?1.8478
?1.2093
?1.4064
?1.1729
?0.5463
?2.1658
?1.2066
?0.2538
?1.4724
?1.3169
?1.2373
?0.8704
?1.0360
?1.7837
?1.8553
?1.7257
?1.3286
Method
RAPCF
PGAS.1
PGAS.2
PGAS.3
Configuration
N = 200
N = 10, M = 100
N = 25, M = 100
N = 10, M = 200
Rank
2.025
2.750
2.550
2.675
Table 5: Average ranks.
Method
RAPCF
PGAS.1
PGAS.2
PGAS.3
Configuration
N = 200
N = 10, M = 100
N = 25, M = 100
N = 10, M = 200
Avg. Time
6
732
1832
1465
Table 6: Avg. running time.
Table 4: Results for RAPCF vs. PGAS.
As mentioned above, there is little difference between the predictive accuracies of RAPCF and
PGAS. However, PGAS is computationally much more expensive. We show average execution times
in minutes for RAPCF and PGAS in Table 6. Note that RAPCF is up to two orders of magnitude
faster than PGAS. The cost of this latter method could be reduced by using fewer particles N or
fewer iterations M , but this would also reduce its predictive accuracy. Even after doing so, PGAS
would still be more costly than RAPCF. RAPCF is also competitive with GARCH, EGARCH and
GJR, whose average training times are in this case 2.6, 3.5 and 3.1 minutes, respectively. A naive
implementation of RAPCF has cost O(N T 4 ), since at each time step t there is a O(T 3 ) cost from
the inversion of the GP covariance matrix. On the other hand, the cost of applying PGAS naively is
O(N M T 5 ), since for each batch of data x1:t there is a O(N M T 4 ) cost. These costs can be reduced
to be O(N T 3 ) and O(N M T 4 ) for RAPCF and PGAS respectively by doing rank one updates of
the inverse of the GP covariance matrix at each time step. The costs can be further reduced by a
factor of T 2 by using sparse GPs [28].
7
Summary and discussion
We have introduced a novel Gaussian Process Volatility model (GP-Vol) for time-varying variances
in financial time series. GP-Vol is an instance of a Gaussian Process State-Space model (GP-SSM)
which is highly flexible and can model nonlinear functional relationships and asymmetric effects of
positive and negative returns on time-varying variances. In addition, we have presented an online
inference method based on particle filtering for GP-Vol called the Regularized Auxiliary Particle
Chain Filter (RAPCF). RAPCF is up to two orders of magnitude faster than existing batch Particle
Gibbs methods. Results for GP-Vol on 50 financial time series show significant improvements in
predictive performance over existing models such as GARCH, EGARCH and GJR-GARCH. Finally,
the nonlinear transition functions learned by GP-Vol can be easily analyzed to understand the effect
of past volatility and past returns on future volatility.
For future work, GP-Vol can be extended to learn the functional relationship between a financial
instrument?s volatility, its price and other market factors, such as interest rates. The functional
relationship thus learned can be useful in the pricing of volatility derivatives on the instrument.
Additionally, the computational efficiency of RAPCF makes it an attractive choice for inference in
other GP-SSMs different from GP-Vol. For example, RAPCF could be more generally applied to
learn the hidden states and the dynamics in complex control systems.
8
References
[1] R. Cont. Empirical properties of asset returns: Stylized facts and statistical issues. Quantitative Finance,
1(2):223?236, 2001.
[2] T. Bollerslev. Generalized autoregressive conditional heteroskedasticity. Journal of econometrics,
31(3):307?327, 1986.
[3] A. Harvey, E. Ruiz, and N. Shephard. Multivariate stochastic variance models. The Review of Economic
Studies, 61(2):247?264, 1994.
[4] S. Kim, N. Shephard, and S. Chib. Stochastic volatility: likelihood inference and comparison with ARCH
models. The Review of Economic Studies, 65(3):361?393, 1998.
[5] S.H. Poon and C. Granger. Practical issues in forecasting volatility. Financial Analysts Journal, 61(1):45?
56, 2005.
[6] Y. Wu, J. M. Hern?andez-Lobato, and Z. Ghahramani. Dynamic covariance models for multivariate financial time series. In ICML, pages 558?566, 2013.
[7] L. Hentschel. All in the family nesting symmetric and asymmetric GARCH models. Journal of Financial
Economics, 39(1):71?104, 1995.
[8] G. Bekaert and G. Wu. Asymmetric volatility and risk in equity markets. Review of Financial Studies,
13(1):1?42, 2000.
[9] J.Y. Campbell and L. Hentschel. No news is good news: An asymmetric model of changing volatility in
stock returns. Journal of financial Economics, 31(3):281?318, 1992.
[10] M.W. Brandt and C.S. Jones. Volatility forecasting with range-based EGARCH models. Journal of
Business & Economic Statistics, 24(4):470?486, 2006.
[11] A. Wilson and Z. Ghahramani. Copula processes. In Advances in Neural Information Processing Systems
23, pages 2460?2468. 2010.
[12] M. L?azaro-Gredilla and M. K. Titsias. Variational heteroscedastic Gaussian process regression. In ICML,
pages 841?848, 2011.
[13] C.E. Rasmussen and C.K.I. Williams. Gaussian processes for machine learning. Springer, 2006.
[14] D.B. Nelson. Conditional heteroskedasticity in asset returns: A new approach. Econometrica, 59(2):347?
370, 1991.
[15] L.R. Glosten, R. Jagannathan, and D.E. Runkle. On the relation between the expected value and the
volatility of the nominal excess return on stocks. The Journal of Finance, 48(5):1779?1801, 1993.
[16] J. Ko and D. Fox. GP-BayesFilters: Bayesian filtering using Gaussian process prediction and observation
models. Autonomous Robots, 27(1):75?90, 2009.
[17] M. P. Deisenroth, M. F. Huber, and U. D. Hanebeck. Analytic moment-based Gaussian process filtering.
In ICML, pages 225?232. ACM, 2009.
[18] M. Deisenroth and S. Mohamed. Expectation Propagation in Gaussian Process Dynamical Systems. In
Advances in Neural Information Processing Systems 25, pages 2618?2626, 2012.
[19] R. Frigola, F. Lindsten, T. B. Sch?on, and C. E. Rasmussen. Bayesian inference and learning in Gaussian
process state-space models with particle MCMC. In NIPS, pages 3156?3164. 2013.
[20] F. Lindsten, M. Jordan, and T. Sch?on. Ancestor Sampling for Particle Gibbs. In Advances in Neural
Information Processing Systems 25, pages 2600?2608, 2012.
[21] L.E. Baum and T. Petrie. Statistical inference for probabilistic functions of finite state Markov chains.
The Annals of Mathematical Statistics, 37(6):1554?1563, 1966.
[22] A. Doucet, N. De Freitas, and N. Gordon. Sequential Monte Carlo methods in practice. Springer Verlag,
2001.
[23] R. D. Turner, M. P. Deisenroth, and C. E. Rasmussen. State-space inference and learning with Gaussian
processes. In AISTATS, pages 868?875, 2010.
[24] C. Andrieu, A. Doucet, and R. Holenstein. Particle Markov chain Monte Carlo methods. Journal of the
Royal Statistical Society: Series B (Statistical Methodology), 72(3):269?342, 2010.
[25] J. Liu and M. West. Combined parameter and state estimation in simulation-based filtering. Institute of
Statistics and Decision Sciences, Duke University, 1999.
[26] M.K. Pitt and N. Shephard. Filtering via simulation: Auxiliary particle filters. Journal of the American
Statistical Association, pages 590?599, 1999.
[27] J. Dem?sar. Statistical comparisons of classifiers over multiple data sets. Journal of Machine Learning
Research, 7:1?30, 2006.
[28] J. Qui?nonero-Candela and C.E. Rasmussen. A unifying view of sparse approximate Gaussian process
regression. The Journal of Machine Learning Research, 6:1939?1959, 2005.
9
| 5376 |@word version:1 middle:5 inversion:1 simulation:2 propagate:2 eng:1 covariance:10 moment:1 initial:5 configuration:2 series:24 contains:1 liu:1 liquid:1 amp:1 past:5 existing:3 outperforms:1 current:3 recovered:1 com:1 freitas:1 afl:1 distant:1 happen:1 informative:2 numerical:1 enables:1 analytic:1 designed:2 plot:9 update:1 resampling:1 v:3 fewer:2 filtered:1 colored:1 ssm:10 org:1 brandt:1 bayesfilters:1 mathematical:1 along:1 wild:1 introduce:4 pairwise:1 huber:1 market:6 expected:6 behavior:1 p1:1 frequently:1 inspired:1 globally:1 little:3 becomes:2 provided:1 moreover:1 skewness:1 developed:1 lindsten:2 unobserved:1 finding:1 quantitative:1 finance:2 classifier:1 uk:2 control:1 unit:1 appear:1 positive:14 generalised:1 limit:1 ak:1 fluctuation:1 challenging:2 heteroscedastic:2 hmms:1 limited:3 smc:1 range:2 statistically:1 practical:1 thirty:1 practice:3 procedure:2 empirical:5 significantly:2 reject:1 pre:1 confidence:3 zoubin:2 cannot:1 close:1 collapsed:1 applying:1 risk:1 www:1 equivalent:1 lobato:2 send:1 williams:1 attention:1 economics:2 baum:1 focused:2 nesting:1 financial:19 jmh233:1 fx:5 autonomous:1 sar:1 annals:1 target:3 pt:2 nominal:1 exact:1 duke:1 gps:2 hypothesis:1 jmhl:1 harvard:1 expensive:4 approximated:3 asymmetric:15 econometrics:1 labeled:1 observed:4 capture:3 news:2 highest:2 mentioned:1 environment:1 econometrica:1 cam:2 dynamic:16 trained:1 depend:1 segment:1 heteroscedasticity:2 predictive:16 wtj:3 titsias:1 efficiency:1 easily:1 joint:2 stylized:1 stock:2 represented:4 various:1 train:1 fast:1 describe:1 shortcoming:1 monte:4 artificial:3 hyper:5 kevin:1 whose:2 ace:1 larger:2 solve:1 regeneration:1 otherwise:1 cov:1 statistic:3 gp:82 jointly:3 highlighted:2 final:1 online:4 advantage:1 vtj:5 propose:3 took:1 nonero:1 poon:1 flexibility:3 bollerslev:1 etj:2 asymmetry:1 produce:4 volatility:34 ac:2 propagating:1 miguel:1 measured:1 received:3 shephard:3 auxiliary:6 implemented:1 predicted:8 come:1 australian:1 differ:2 concentrate:1 filter:8 stochastic:3 exchange:1 fix:1 andez:2 generalization:1 extension:4 around:1 considered:2 exp:2 predict:2 pitt:1 consecutive:1 resample:2 estimation:2 successfully:1 weighted:1 eti:2 gaussian:23 modified:1 pn:2 shrinkage:2 varying:3 wilson:1 encode:2 improvement:1 rank:9 likelihood:14 mainly:1 indicates:1 contrast:1 kim:1 sense:1 am:1 dollar:2 inference:15 dependent:2 rigid:1 foreign:1 eliminate:1 hidden:15 relation:1 ancestor:6 pgas:35 selects:1 overall:1 issue:4 flexible:5 impoverishment:3 aforementioned:1 adm:1 among:2 development:2 smoothing:3 special:1 copula:2 art:1 once:2 sampling:6 eliminated:2 represents:1 broad:1 jones:1 icml:3 future:2 regenerated:1 t2:10 fundamentally:1 gordon:1 few:2 abt:1 chib:1 resulted:1 consisting:1 attempt:3 zt0:5 friedman:1 interest:1 highly:3 introduces:2 analyzed:4 extreme:1 tj:4 chain:8 edge:1 daily:5 shorter:2 fox:1 logarithm:1 uncertain:1 instance:3 modeling:2 markovian:5 cost:8 introducing:1 deviation:5 dependency:1 sv:1 synthetic:4 combined:1 cited:1 probabilistic:1 jos:1 together:1 squared:3 american:1 derivative:1 return:31 converted:1 potential:1 de:1 bold:2 summarized:1 dem:1 coefficient:1 explicitly:1 performed:1 view:1 closed:1 candela:1 analyze:2 doing:2 competitive:1 sort:1 contribution:1 publicly:1 accuracy:3 variance:25 correspond:2 bayesian:5 produced:1 carlo:4 asset:2 holenstein:1 checked:1 against:7 pp:3 mohamed:1 associated:1 redder:1 propagated:3 sampled:1 dataset:8 adjusting:1 popular:2 knowledge:1 amplitude:1 routine:1 campbell:1 higher:1 day:1 methodology:1 bxt:1 evaluated:2 shrink:1 done:1 furthermore:2 arch:1 until:2 hand:1 receives:2 horizontal:1 nonlinear:6 propagation:2 aiv:1 lack:1 defines:1 pricing:2 gti:2 effect:17 normalized:1 true:1 contain:1 evolution:3 andrieu:1 symmetric:2 nonzero:1 attractive:1 during:1 skewed:1 jagannathan:1 generalized:1 performs:1 variational:1 novel:2 recently:3 petrie:1 pseudocode:1 functional:16 discussed:1 association:1 approximates:1 adp:1 refer:1 significant:4 cambridge:3 gibbs:5 grid:4 closing:3 particle:45 nonlinearity:1 pq:3 robot:1 surface:7 heteroskedasticity:2 add:2 posterior:11 multivariate:3 verlag:1 harvey:1 vt:53 aee:1 garch:45 captured:1 seen:1 additional:1 fortunately:1 ssms:8 determine:3 period:3 multiple:2 smooth:1 faster:5 cross:5 long:1 host:1 post:1 impact:1 prediction:12 variant:7 regression:3 ko:1 expectation:1 iteration:2 kernel:3 adopting:1 proposal:1 addition:4 addressed:2 interval:5 underestimated:1 source:1 sch:2 fifty:2 operate:1 pass:1 yue:1 jordan:1 near:1 leverage:1 variety:1 opposite:1 reduce:1 economic:3 whether:3 forecasting:2 york:1 generally:2 useful:1 covered:1 involve:1 clear:1 amount:1 ten:1 band:3 processed:1 reduced:3 http:2 wiki:1 dotted:1 estimated:1 blue:1 vol:47 affected:1 nevertheless:1 changing:3 backward:1 econometric:3 v1:16 sum:1 aig:1 run:1 inverse:1 uncertainty:1 jitter:1 place:2 family:1 wu:3 decision:1 qui:1 capturing:1 followed:1 display:1 quadratic:2 diffuse:1 sake:1 performing:4 relatively:1 developing:1 according:7 gredilla:1 combination:1 disconnected:1 poor:1 smaller:1 across:4 em:1 slightly:3 intuitively:2 aet:1 pr:1 taken:1 computationally:5 equation:2 hern:2 eventually:1 x2t:4 fail:1 needed:1 avt:1 turbulent:1 granger:1 instrument:2 end:1 available:1 operation:1 observe:1 calm:1 alternative:2 batch:8 assumes:2 clustering:2 include:1 standardized:1 running:1 graphical:2 marginalized:3 unifying:1 ghahramani:3 especially:1 society:1 move:1 occurs:2 parametric:6 costly:1 dependence:1 exhibit:2 distance:2 separate:1 hmm:4 nelson:1 amd:1 consensus:1 sheppard:1 aes:1 analyst:1 besides:1 code:1 modeled:1 relationship:15 length:4 index:1 decimal:1 cont:1 setup:2 susceptible:1 negative:14 rise:1 design:1 implementation:1 zt:7 unknown:5 twenty:2 perform:1 observation:11 dispersion:1 markov:4 datasets:8 finite:1 january:2 extended:2 varied:1 community:1 hanebeck:1 introduced:3 namely:1 specified:1 learned:8 established:1 nip:1 address:4 able:1 usually:1 pattern:2 gjr:17 dynamical:1 royal:1 apf:1 critical:1 ranked:1 business:1 regularized:3 turner:1 representing:1 improve:1 axis:3 naive:1 review:5 literature:1 prior:10 l2:1 fully:2 frigola:2 limitation:3 filtering:8 proportional:1 plotting:2 cd:2 summary:1 pmcmc:2 placed:3 last:2 surprisingly:1 rasmussen:4 offline:1 bias:3 understand:2 institute:1 amat:1 sparse:2 liquidity:1 dip:1 transition:19 avoids:2 stand:1 autoregressive:2 contour:2 author:2 forward:10 made:2 w0i:1 jump:1 avg:2 excess:1 approximate:1 obtains:1 smm:1 confirm:2 doucet:2 overfitting:5 sequentially:2 buy:1 assumed:1 table:12 additionally:2 promising:1 learn:10 symmetry:1 adi:1 aep:1 complex:1 official:1 aistats:1 main:5 linearly:1 whole:1 n2:1 allowed:1 repeated:4 x1:9 augmented:2 west:1 shrinking:2 experienced:1 explicit:1 exponential:3 third:1 ruiz:1 minute:2 specific:1 xt:43 jt:1 evidence:2 naively:2 adding:1 sequential:2 importance:6 magnitude:3 execution:3 logarithmic:1 simply:1 univariate:1 azaro:1 springer:2 aa:1 corresponds:1 truth:2 amt:1 abc:1 acm:1 conditional:3 consequently:2 towards:3 price:8 agn:1 change:1 specifically:1 determined:1 typical:2 wt:5 called:3 total:1 equity:7 deisenroth:3 support:1 latter:1 glosten:1 brevity:1 mcmc:5 tested:1 phenomenon:1 correlated:1 |
4,834 | 5,377 | Bandit Convex Optimization: Towards Tight Bounds
Kfir Y. Levy
Technion?Israel Institute of Technology
Haifa 32000, Israel
[email protected]
Elad Hazan
Technion?Israel Institute of Technology
Haifa 32000, Israel
[email protected]
Abstract
Bandit Convex Optimization (BCO) is a fundamental framework for decision
making under uncertainty, which generalizes many problems from the realm of online and statistical learning. While the special case of linear cost functions is well
understood, a gap on the attainable regret for BCO with nonlinear losses remains
an important open question. In this paper we take a step towards understanding
the best attainable regret bounds for BCO: we give an efficient and near-optimal
regret algorithm for BCO with strongly-convex and smooth loss functions. In contrast to previous works on BCO that use time invariant exploration schemes, our
method employs an exploration scheme that shrinks with time.
1
Introduction
The power of Online Convex Optimization (OCO) framework is in its ability to generalize many
problems from the realm of online and statistical learning, and supply universal tools to solving
them. Extensive investigation throughout the last decade has yield efficient algorithms with worst
case guarantees. This has lead many practitioners to embrace the OCO framework in modeling and
solving real world problems.
One of the greatest challenges in OCO is finding tight bounds to the problem of Bandit Convex
Optimization (BCO). In this ?bandit? setting the learner observes the loss function only at the point
that she has chosen. Hence, the learner has to balance between exploiting the information she has
gathered and between exploring the new data. The seminal work of [5] elegantly resolves this
?exploration-exploitation? dilemma by devising a combined explore-exploit gradient descent algorithm. They obtain a bound of O(T 3/4 ) on the expected regret for the general case of an adversary
playing bounded and Lipschitz-continuous convex losses.
In this paper we investigate the BCO setting assuming that the adversary is limited to inflicting
strongly-convex and smooth losses and the player may choose points from a constrained
decision
?
? T ). This rate is
set. In this setting we devise an efficient algorithm that achieves a regret of O(
the best possible up?to logarithmic factors as implied by a recent work of [11], cleverly obtaining a
lower bound of ?( T ) for the same setting.
During our analysis, we develop a full-information algorithm that takes advantage of the strongconvexity of loss functions and uses a self-concordant barrier as a regularization term. This algorithm enables us to perform ?shrinking exploration? which is a key ingredient in our BCO algorithm.
Conversely, all previous works on BCO use a time invariant exploration scheme.
This paper is organized as follows. In Section 2 we introduce our setting and review necessary
preliminaries regarding self-concordant barriers. In Section 3 we discuss schemes to perform single1
Setting
Full-Info.
BCO
Convex
? 3/4 )
O(T
Linear
?
?( ?T )
? T)
O(
Smooth
Str.-Convex
? 2/3 )
O(T
?
?( T )
Str.-Convex & Smooth
?(log
? T)
? T ) [Thm. 10]
O(
Table 1: Known regret bounds in the Full-Info./ BCO setting. Our new result is highlighted, and
? 2/3 ) bound.
improves upon the previous O(T
point gradient estimations, then we define first-order online methods and analyze the performance
of such methods receiving noisy gradient estimates. Our main result is described and analyzed in
Section 4; Section 5 concludes.
1.1
Prior work
For BCO with general convex loss functions, almost simultaneously to [5], a bound of O(T 3/4 )
was also obtained by [7] for the setting of?
Lipschitz-continuous convex losses. Conversely, the best
known lower bound for this problem is ?( T ) proved for the easier full-information setting.
In case the adversary is limited to using linear losses, it can be shown that the player does not
?pay? for exploration; this property was
? used by [4] to devise the Geometric Hedge algorithm that
? T ). Later [1], inspired by interior point methods, devised the
achieves an optimal regret rate of O(
first efficient algorithm that attains the same nearly-optimal regret rate for this setup of bandit linear
optimization.
For some special classes of nonlinear convex losses, there are several works that lean on ideas
from [5] to achieve improved upper bounds for BCO. In the case of convex and smooth losses [9]
? 2/3 ). The same regret rate of O(T
? 2/3 ) was achieved by [2] in the
attained an upper bound of O(T
case of strongly-convex losses. For the special?case of unconstrained BCO with strongly-convex
? T ). A recent paper by Shamir [11], significantly
and smooth losses, [2] obtained a regret of O(
?
advanced our understanding of BCO by devising a lower bound of ?( T ) for the setting of stronglyconvex and smooth BCO. The latter implies the tightness of our bound.
A comprehensive survey by Bubeck and Cesa-Bianchi [3], provides a review of the bandit optimization literature in both stochastic and online setting.
2
Setting and Background
Notation: During this paper we denote by || ? || the `2 norm when referring to vectors, and use
the same notation for the spectral norm when referring to matrices. We denote by Bn and Sn the
n-dimensional euclidean unit ball and unit sphere, and by v ? Bn and u ? Sn random variables
chosen uniformly from these sets. The symbol I is used for the identity matrix (its dimension will
be clear from the context). For a positive definite matrix A 0 we denote by A1/2 the matrix B
such that B > B = A, and by A?1/2 the inverse of B. Finally, we denote [N ] := {1, . . . , N }.
2.1
Bandit Convex Optimization
We consider a repeated game of T rounds between a player and an adversary, at each round t ?
[T ]
1. player chooses a point xt ? K.
2. adversary independently chooses a loss function ft ? F.
3. player suffers a loss ft (xt ) and receives a feedback Ft .
2
In the OCO (Online Convex Optimization) framework we assume that the decision set K is convex and that all functions in F are convex. Our paper focuses on adversaries limited to choosing
functions from the set F?,? ; the set off all ?-strongly-convex and ?-smooth functions.
We also limit ourselves to oblivious adversaries where the loss sequence {ft }Tt=1 is predetermined
and is therefore independent of the player?s choices. Mind that in this case the best point in hindsight
is also independent of the player?s choices. We also assume that the loss functions are defined over
the entire space Rn and are strongly-convex and smooth there; yet the player may only choose points
from a constrained set K.
Let us define the regret of A, and its regret with respect to a comparator w ? K:
RegretA
T =
T
X
t=1
ft (xt ) ? min
?
w ?K
T
X
RegretA
T (w) =
ft (w? ),
t=1
T
X
t=1
ft (xt ) ?
T
X
ft (w)
t=1
A player aims at minimizing his regret, and we are interested in players that ensure an o(T ) regret
for any loss sequence that the adversary may choose.
The player learns through the feedback Ft received in response to his actions. In the full informations
setting, he receives the loss function ft itself as a feedback, usually by means of a gradient oracle i.e. the decision maker has access to the gradient of the loss function at any point in the decision set.
Conversely, in the BCO setting the given feedback is ft (xt ), i.e., the loss function
only
at the point
that he has chosen; and the player aims at minimizing his expected regret, E RegretA
T .
2.2
Strong Convexity and Smoothness
As mentioned in the last subsection we consider an adversary limited to choosing loss functions
from the set F?,? , the set of ?-strongly convex and ?-smooth functions, here we define these properties.
Definition 1. (Strong Convexity) We say that a function f : Rn ? R is ?-strongly convex over the
set K if for all x, y ? K it holds that,
f (y) ? f (x) + ?f (x)> (y ? x) +
?
||x ? y||2
2
(1)
Definition 2. (Smoothness) We say that a convex function f : Rn ? R is ?-smooth over the set K
if the following holds:
f (y) ? f (x) + ?f (x)> (y ? x) +
2.3
?
||x ? y||2 ,
2
?x, y ? K
(2)
Self Concordant Barriers
Interior point methods are polynomial time algorithms to solving constrained convex optimization
programs. The main tool in these methods is a barrier function that encodes the constrained set and
enables the use of a fast unconstrained optimization machinery. More on this subject can be found
in [8].
Let K ? Rn be a convex set with a non empty interior int(K)
Definition 3. A function R : int(K) ? R is called ?-self-concordant if:
1. R is three times continuously differentiable and convex, and approaches infinity along any
sequence of points approaching the boundary of K.
2. For every h ? Rn and x ? int(K) the following holds:
|?3 R(x)[h, h, h]| ? 2(?2 R(x)[h, h])3/2
3
and
|?R(x)[h]| ? ? 1/2 (?2 R(x)[h, h])1/2
here, ?3 R(x)[h, h, h] :=
?3
?t1 ?t2 ?t3 R(x
+ t1 h + t2 h + t3 h)
t1 =t2 =t3 =0
.
?
Our algorithm requires a ?-self-concordant barrier over K, and its regret depends on ?. It is well
n
known that any convex set in R admits a ? = O(n) such barrier (? might be much smaller), and that
most interesting convex sets admit a self-concordant barrier that is efficiently represented.
The Hessian of a self-concordant barrier induces a local norm at every x ? int(K), we denote this
norm by || ? ||x and its dual by || ? ||?x and define ?h ? Rn :
q
q
||h||x = h> ?2 R(x)h,
||h||?x = h> (?2 R(x))?1 h
we assume that ?2 R(x) always has a full rank.
The following fact is a key ingredient in the sampling scheme of BCO algorithms [1, 9]. Let R is
be self-concordant barrier and x ? int(K) then the Dikin Ellipsoide,
W1 (x) := {y ? Rn : ||y ? x||x ? 1}
(3)
i.e. the || ? ||x -unit ball centered around x, is completely contained in K.
Our regret analysis requires a bound on R(y) ? R(x); hence, we will find the following lemma
useful:
Lemma 4. Let R be a ?-self-concordant function over K, then:
R(y) ? R(x) ? ? log
1
,
1 ? ?x (y)
where ?x (y) = inf{t ? 0 : x + t?1 (y ? x) ? K},
?x, y ? int(K)
?x, y ? int(K)
Note that ?x (y) is called the Minkowsky function and it is always in [0, 1]. Moreover, as y approaches the boundary of K then ?x (y) ? 1.
3
3.1
Single Point Gradient Estimation and Noisy First-Order Methods
Single Point Gradient Estimation
A main component of BCO algorithms is a randomized sampling scheme for constructing gradient estimates. Here, we survey the previous schemes as well as the more general scheme that we
use.
Spherical estimators: Flaxman et al. [5] introduced a method that produces single point gradient
estimates through spherical sampling. These estimates are then inserted into a full-information procedure that chooses the next decision point for the player. Interestingly, these gradient estimates are
unbiased predictions for the gradients of a smoothed version function which we next define.
Let ? > 0 and v ? Bn , the smoothed version of a function f : Rn ? R is defined as follows:
f?(x) = E[f (x + ?v)]
(4)
The next lemma of [5] ties between the gradients of f? and an estimate based on samples of f :
Lemma 5. Let u ? Sn , and consider the smoothed version f? defined in Equation (4), then the
following applies:
n
(5)
?f?(x) = E[ f (x + ?u)u]
?
Therefore, n? f (x + ?u)u is an unbiased estimator for the gradients of the smoothed version.
4
x
x
x
t
K
K
(a) Eigenpoles Sampling
(b) Continuous Sampling
K
(c) Shrinking Sampling
Figure 1: Dikin Ellipsoide Sampling Schemes
Ellipsoidal estimators: Abernethy et al. [1] introduced the idea of sampling from an ellipsoid
(specifically the Dikin ellipsoid) rather than a sphere in the context of BCO. They restricted the
sampling to the eigenpoles of the ellipsoid (Fig. 1a). A more general method of sampling continuously from an ellipsoid was introduced in [9] (Fig. 1b). We shall see later that our?algorithm
? T ) regret
uses a ?shrinking-sampling? scheme (Fig. 1c), which is crucial in achieving the O(
bound.
The following lemma of [9] shows that we can sample f non uniformly over all directions and create
an unbiased gradient estimate of a respective smoothed version:
Corollary 6. Let f : Rn ? R be a continuous function, let A ? Rn?n be invertible, and v ? Bn ,
u ? Sn . Define the smoothed version of f with respect to A:
f?(x) = E[f (x + Av)]
(6)
Then the following holds:
?f?(x) = E[nf (x + Au)A?1 u]
(7)
Note that if A 0 then {Au : u ? Sn } is an ellipsoid?s boundary.
Our next lemma shows that the smoothed version preserves the strong-convexity of f , and that we
can measure the distance between f? and f using the spectral norm of A2 :
Lemma 7. Consider a function f : Rn ? R, and a positive definite matrix A ? Rn?n . Let f? be
the smoothed version of f with respect to A as defined in Equation (6). Then the following holds:
? If f is ?-strongly convex then so is f?.
? If f is convex and ?-smooth, and ?max be the largest eigenvalue of A then:
?
?
0 ? f?(x) ? f (x) ? ||A2 ||2 = ?2max
2
2
(8)
Remark: Lemma 7 also holds if we define the smoothed version of f as f?(x) = Eu?Sn [f (x + Au)]
i.e. an average of the original function values over the unit sphere rather than the unit ball as defined
in Equation (6). Proof is similar to the one of Lemma 7.
3.2
Noisy First-Order Methods
Our algorithm utilizes a full-information online algorithm, but instead of providing this method with
exact gradient values we insert noisy estimates of the gradients. In what follows we define first-order
online algorithms, and present a lemma that analyses the regret of such algorithm receiving noisy
gradients.
5
Definition 8. (First-Order Online Algorithm) Let A be an OCO algorithm receiving an arbitrary
sequence of differential convex loss functions f1 , . . . , fT , and providing points x1 ? A and xt ?
A(f1 , . . . , ft?1 ). Given that A requires all loss functions to belong to some set F0 . Then A is called
first-order online algorithm if the following holds:
? Adding a linear function to a member of F0 remains in F0 ; i.e., for every f ? F0 and
a ? Rn then also f + a> x ? F0
? The algorithm?s choices depend only on its gradient values taken in the past choices of A,
i.e. :
A(f1 , . . . , ft?1 ) = A(?f1 (x1 ), . . . , ?ft?1 (xt?1 )),
?t ? [T ]
The following is a generalization of Lemma 3.1 from [5]:
Lemma 9. Let w be a fixed point in K. Let A be a first-order online algorithm receiving a sequence
of differential convex loss functions f1 , . . . , fT : K ? R (ft+1 possibly depending on z1 , . . . zt ).
Where z1 . . . zT are defined as follows: z1 ? A, zt ? A(g1 , . . . , gt?1 ) where gt ?s are vector valued
random variables such that:
E[gt z1 , f1 , . . . , zt , ft ] = ?ft (zt )
Then if A ensures a regret bound of the form: RegretA
T ? BA (?f1 (x1 ), . . . , ?fT (xT )) in the full
information case then, in the case of noisy gradients it ensures the following bound:
T
T
X
X
E[
ft (zt )] ?
ft (w) ? E[BA (g1 , . . . , gT )]
t=1
4
t=1
Main Result and Analysis
Following is the main theorem of this paper:
Theorem 10. Let K be a convex set with diameter DK and R be a ?-self-concordant barrier over
K. Then in the BCO setting where the adversary is limited to choosing ?-smooth and ?-stronglyconvex
functions and |ft (x)| ? L, ?x ? K, then the expected regret of Algorithm 1 with ? =
q
(?+2?/?) log T
2n2 L2 T
is upper bounded as
s
E[RegretT ] ? 4nL
2?
?+
?
2
?DK
T log T + 2L +
=O
2
r
??
T log T
?
!
whenever T / log T ? 2 (? + 2?/?).
Algorithm 1 BCO Algorithm for Str.-convex & Smooth losses
Input: ? > 0, ? > 0, ?-self-concordant barrier R
Choose x1 = arg minx?K R(x)
for t = 1, 2 . . . T do
?1/2
Define Bt = ?2 R(xt ) + ??tI
Draw u ? Sn
Play yt = xt + Bt u
Observe ft (xt + Bt u) and define gt= nft (xt + Bt u)Bt?1
u ?1
Pt
?
>
2
Update xt+1 = arg minx?K ? =1 g? x + 2 ||x ? x? || + ? R(x)
end for
Algorithm 1 shrinks the exploration magnitude with time (Fig. 1c); this is enabled thanks to the
strong-convexity of the losses. It also updates according to a full-information first-order algorithm
6
denoted FTARL-?, which is defined below. This algorithm is a variant of the FTRL methodology
as defined in [6, 10].
Algorithm 2 FTARL-?
Input: ? > 0, ?-self concordant barrier R
Choose x1 = arg minx?K R(x)
for t = 1, 2 . . . T do
Receive ?ht (xt )
Pt
Output xt+1 = arg minx?K ? =1 ?h? (x? )> x + ?2 ||x ? x? ||2 + ? ?1 R(x)
end for
Next we give a proof sketch of Theorem 10
Proof sketch of Therorem 10. Let us decompose the expected regret of Algorithm 1 with respect to
w ? K:
PT
E [RegretT (w)] := t=1 E [ft (yt ) ? ft (w)]
PT
= t=1 E [ft (yt ) ? ft (xt )]
(9)
h
i
PT
+ t=1 E ft (xt ) ? f?t (xt )
(10)
h
i
PT
? t=1 E ft (w) ? f?t (w)
(11)
h
i
PT
+ t=1 E f?t (xt ) ? f?t (w)
(12)
where expectation is taken with respect to the player?s choices, and f?t is defined as
f?t (x) = E[ft (x + Bt v)],
?x ? K
here v ? Bn and the smoothing matrix Bt is defined in Algorithm 1.
The sampling scheme used by Algorithm 1 yields an unbiased gradient estimate gt of the smoothed
version f?t , which is then inserted to FTARL-? (Algorithm 2). We can therefore interpret Algorithm 1 as performing noisy first-order method (FTARL-?) over the smoothed versions. The xt ?s
in Algorithm 1 are the outputs of FTARL-?, thus the term in Equation (12) is associated with ?exploitation?. The other terms in Equations (9)-(11) measure the cost of sampling away from xt , and
the distance between the smoothed version and the original function, hence these term are associated
with ?exploration?.
In what follows we analyze these terms separately and show that Algorithm 1
?
? T ) regret.
achieves O(
The Exploration Terms: The next hold by the remark that follows Lemma 7 and by the lemma
itself:
E[ft (yt ) ? ft (xt )] = E Eu [ft (xt + Bt u)] ? ft (xt )xt ] ? 0.5?E ||Bt2 ||2 ? ?/2??t (13)
h
i
? E[ft (w) ? f?t (w)] = E E[f?t (w) ? ft (w)xt ] ? 0.5?E ||Bt2 ||2 ? ?/2??t
(14)
h
i
E[ft (xt ) ? f?t (xt )] = E E[ft (xt ) ? f?t (xt )xt ] ? 0
(15)
where ||Bt2 ||2 ? 1/??t follows by the definition of Bt and by the fact that ?2 R(xt ) is positive
definite.
7
The Exploitation Term: The next Lemma bounds the regret of FTARL-? in the full-information
setting:
Lemma 11. Let R be a self-concordant barrier over a convex set K, and ? > 0. Consider an
online player receiving ?-strongly-convex loss functions h1 , . . . , hT and choosing points according
to FTARL-? (Algorithm 2), and ?||?ht (xt )||?t ? 1/2, ?t ? [T ]. Then the player?s regret is upper
bounded as follows:
T
X
t=1
ht (xt ) ?
T
X
t=1
ht (w) ? 2?
T
X
t=1
2
(||?ht (xt )||?t ) + ? ?1 R(w),
?z ? K
here (||a||?t )2 = aT (?2 R(xt ) + ??tI)?1 a
Note that Algorithm 1 uses the estimates gt as inputs into FTARL-?. Using Corollary 6 we can
show that the gt ?s are unbiased estimates for the gradients of the smoothed versions f?t ?s. Using the
regret bound of the above lemma, and the unbiasedness of the gt ?s, Lemma 9 ensures us:
T
X
t=1
T
i
X
?
?
E ft (xt ) ? ft (w) ? 2?
E[(||gt ||?t )2 ] + ? ?1 R(w)
h
(16)
t=1
By the definitions of gt and Bt , and recalling |ft (x)| ? L, ?x ? K, we can bound:
h
?1 ?1 i
2
E[(||gt ||?t )2 xt ] = E n2 (ft (xt + Bt u)) u> Bt?1 ?2 R(xt ) + ??tI
Bt uxt ? (nL)2
Concluding: Plugging the latter into Equation (16) and combining Equations (9)-(16) we get:
E[RegretT (w)] ? 2?(nL)2 T + ? ?1 R(w) + 2?? ?1 log T
(17)
Recall that x1 = arg minx?K R(x) and assume w.l.o.g. that R(x1 ) = 0 (we can always add
R a constant). Thus, for a point w ? K such that ?x1 (w) ? 1 ? T ?1 Lemma 4 ensures us that
R(w) ? ? log T . Combining the latter
p with Equation (17) and the choice of ? in Theorem 10 assures
an expected regret bounded by 4nL (? + 2?? ?1 ) T log T . For w ? K such that ?x1 (w) > 1?T ?1
we can always find w0 ? K such that ||w ? w0 || ? O(T ?1 ) and ?x1 (w0 ) ? 1 ? T ?1 , using the
Lipschitzness of the ft ?s, Theorem 10 holds.
Correctness:
Note that Algorithm 1 chooses points from the set {xt +
?1/2
2
? R(xt ) + ??tI
u, u ? Sn } which is inside the Dikin ellipsoid and therefore belongs to K
(the Dikin Eliipsoid is always in K).
5
Summary and open questions
We have presented an efficient algorithm that attains near optimal regret for the setting of BCO with
strongly-convex and smooth losses, advancing our understanding of optimal regret rates for bandit
learning.
Perhaps the most important question in bandit learning remains the resolution of the attainable regret
bounds for smooth but non-strongly-convex, or vice versa, and generally convex cost functions (see
Table 1). Ideally, this should be accompanied by an efficient algorithm, although understanding the
optimal rates up to polylogarithmic factors would be a significant advancement by itself.
Acknowledgements
The research leading to these results has received funding from the European Union?s Seventh Framework Programme (FP7/2007-2013) under grant agreement n? 336078 ? ERCSUBLRN.
8
References
[1] Jacob Abernethy, Elad Hazan, and Alexander Rakhlin. Competing in the dark: An efficient
algorithm for bandit linear optimization. In COLT, pages 263?274, 2008.
[2] Alekh Agarwal, Ofer Dekel, and Lin Xiao. Optimal algorithms for online convex optimization
with multi-point bandit feedback. In COLT, pages 28?40, 2010.
[3] S?ebastien Bubeck and Nicolo Cesa-Bianchi. Regret analysis of stochastic and nonstochastic multi-armed bandit problems. Foundations and Trends in Machine Learning, 5(1):1?122,
2012.
[4] Varsha Dani, Thomas P. Hayes, and Sham Kakade. The price of bandit information for online
optimization. In NIPS, 2007.
[5] Abraham Flaxman, Adam Tauman Kalai, and H. Brendan McMahan. Online convex optimization in the bandit setting: gradient descent without a gradient. In SODA, pages 385?394,
2005.
[6] Elad Hazan. A survey: The convex optimization approach to regret minimization. In Suvrit
Sra, Sebastian Nowozin, and Stephen J. Wright, editors, Optimization for Machine Learning,
pages 287?302. MIT Press, 2011.
[7] Robert D Kleinberg. Nearly tight bounds for the continuum-armed bandit problem. In NIPS,
volume 17, pages 697?704, 2004.
[8] Arkadii Nemirovskii. Interior point polynomial time methods in convex programming. Lecture
Notes, 2004.
[9] Ankan Saha and Ambuj Tewari. Improved regret guarantees for online smooth convex optimization with bandit feedback. In AISTATS, pages 636?642, 2011.
[10] Shai Shalev-Shwartz. Online learning and online convex optimization. Foundations and Trends
in Machine Learning, 4(2):107?194, 2011.
[11] Ohad Shamir. On the complexity of bandit and derivative-free stochastic convex optimization.
In Conference on Learning Theory, pages 3?24, 2013.
9
| 5377 |@word exploitation:3 version:13 polynomial:2 norm:5 dekel:1 open:2 bn:5 jacob:1 attainable:3 ftrl:1 interestingly:1 past:1 dikin:5 yet:1 predetermined:1 enables:2 update:2 devising:2 advancement:1 provides:1 along:1 differential:2 supply:1 stronglyconvex:2 inside:1 introduce:1 expected:5 multi:2 inspired:1 spherical:2 resolve:1 armed:2 str:3 bounded:4 notation:2 moreover:1 israel:4 what:2 finding:1 hindsight:1 lipschitzness:1 guarantee:2 every:3 nf:1 ti:4 tie:1 unit:5 grant:1 positive:3 t1:3 understood:1 local:1 limit:1 might:1 au:3 conversely:3 limited:5 union:1 regret:33 definite:3 procedure:1 universal:1 significantly:1 get:1 interior:4 inflicting:1 context:2 seminal:1 yt:4 independently:1 convex:48 survey:3 resolution:1 estimator:3 his:3 enabled:1 shamir:2 play:1 pt:7 exact:1 programming:1 us:3 agreement:1 trend:2 lean:1 ft:44 inserted:2 worst:1 ensures:4 eu:2 observes:1 mentioned:1 convexity:4 complexity:1 ideally:1 depend:1 tight:3 solving:3 dilemma:1 upon:1 learner:2 completely:1 represented:1 tx:1 fast:1 choosing:4 shalev:1 abernethy:2 elad:3 valued:1 say:2 tightness:1 ability:1 g1:2 highlighted:1 noisy:7 itself:3 online:18 advantage:1 sequence:5 differentiable:1 eigenvalue:1 combining:2 achieve:1 exploiting:1 empty:1 produce:1 adam:1 depending:1 develop:1 ac:2 received:2 strong:4 implies:1 direction:1 stochastic:3 exploration:9 centered:1 f1:7 generalization:1 investigation:1 preliminary:1 decompose:1 exploring:1 insert:1 hold:9 around:1 wright:1 achieves:3 continuum:1 a2:2 estimation:3 maker:1 largest:1 correctness:1 create:1 vice:1 tool:2 minimization:1 dani:1 mit:1 always:5 aim:2 rather:2 kalai:1 corollary:2 focus:1 she:2 rank:1 contrast:1 brendan:1 attains:2 entire:1 bt:13 bandit:17 interested:1 arg:5 dual:1 colt:2 denoted:1 constrained:4 special:3 smoothing:1 sampling:13 oco:5 nearly:2 t2:3 employ:1 oblivious:1 saha:1 simultaneously:1 preserve:1 comprehensive:1 ourselves:1 recalling:1 investigate:1 bt2:3 analyzed:1 nl:4 therorem:1 kfir:1 ehazan:1 necessary:1 respective:1 ohad:1 machinery:1 euclidean:1 haifa:2 modeling:1 cost:3 technion:4 seventh:1 combined:1 varsha:1 referring:2 chooses:4 thanks:1 fundamental:1 randomized:1 unbiasedness:1 ie:1 off:1 receiving:5 invertible:1 continuously:2 w1:1 cesa:2 choose:5 possibly:1 admit:1 derivative:1 leading:1 accompanied:1 int:7 depends:1 bco:23 later:2 h1:1 hazan:3 analyze:2 shai:1 il:2 efficiently:1 yield:2 gathered:1 t3:3 generalize:1 suffers:1 whenever:1 sebastian:1 definition:6 proof:3 associated:2 proved:1 recall:1 realm:2 subsection:1 improves:1 organized:1 attained:1 methodology:1 response:1 improved:2 shrink:2 strongly:12 sketch:2 receives:2 nonlinear:2 kfiryl:1 perhaps:1 unbiased:5 hence:3 regularization:1 arkadii:1 round:2 during:2 self:13 game:1 tt:1 funding:1 volume:1 belong:1 he:2 interpret:1 significant:1 versa:1 smoothness:2 unconstrained:2 access:1 f0:5 alekh:1 gt:12 add:1 nicolo:1 recent:2 inf:1 belongs:1 suvrit:1 devise:2 stephen:1 full:11 sham:1 smooth:17 sphere:3 lin:1 devised:1 a1:1 plugging:1 prediction:1 variant:1 expectation:1 agarwal:1 achieved:1 receive:1 background:1 separately:1 crucial:1 subject:1 member:1 practitioner:1 near:2 nonstochastic:1 approaching:1 competing:1 regarding:1 idea:2 ankan:1 hessian:1 action:1 remark:2 regrett:3 useful:1 generally:1 clear:1 tewari:1 dark:1 ellipsoidal:1 induces:1 diameter:1 shall:1 key:2 achieving:1 ht:6 advancing:1 inverse:1 uncertainty:1 soda:1 throughout:1 almost:1 utilizes:1 draw:1 decision:6 bound:22 pay:1 oracle:1 infinity:1 encodes:1 kleinberg:1 min:1 concluding:1 performing:1 embrace:1 according:2 ball:3 cleverly:1 smaller:1 kakade:1 making:1 invariant:2 restricted:1 taken:2 equation:8 remains:3 assures:1 discus:1 mind:1 fp7:1 end:2 generalizes:1 ofer:1 observe:1 away:1 spectral:2 original:2 thomas:1 ensure:1 exploit:1 implied:1 question:3 gradient:23 minx:5 distance:2 w0:3 assuming:1 ellipsoid:6 providing:2 balance:1 minimizing:2 setup:1 robert:1 info:2 ba:2 ebastien:1 zt:6 perform:2 bianchi:2 upper:4 av:1 descent:2 nemirovskii:1 rn:13 smoothed:13 arbitrary:1 thm:1 introduced:3 extensive:1 z1:4 polylogarithmic:1 nip:2 adversary:10 usually:1 below:1 challenge:1 program:1 ambuj:1 max:2 power:1 greatest:1 advanced:1 scheme:11 technology:2 concludes:1 flaxman:2 sn:8 review:2 understanding:4 prior:1 geometric:1 literature:1 l2:1 acknowledgement:1 loss:29 lecture:1 interesting:1 ingredient:2 foundation:2 regreta:4 xiao:1 editor:1 playing:1 nowozin:1 summary:1 last:2 free:1 institute:2 barrier:13 tauman:1 feedback:6 dimension:1 boundary:3 world:1 programme:1 hayes:1 shwartz:1 continuous:4 decade:1 table:2 sra:1 obtaining:1 european:1 constructing:1 elegantly:1 aistats:1 main:5 abraham:1 n2:2 repeated:1 x1:10 fig:4 shrinking:3 mcmahan:1 levy:1 learns:1 theorem:5 xt:43 symbol:1 rakhlin:1 dk:2 admits:1 adding:1 magnitude:1 gap:1 easier:1 logarithmic:1 explore:1 bubeck:2 contained:1 applies:1 hedge:1 comparator:1 identity:1 towards:2 lipschitz:2 price:1 specifically:1 uniformly:2 lemma:19 called:3 concordant:13 player:16 latter:3 alexander:1 |
4,835 | 5,378 | Stochastic Multi-Armed-Bandit Problem
with Non-stationary Rewards
Yonatan Gur
Stanford University
Stanford, CA
[email protected]
Omar Besbes
Columbia University
New York, NY
[email protected]
Assaf Zeevi
Columbia University
New York, NY
[email protected]
Abstract
In a multi-armed bandit (MAB) problem a gambler needs to choose at each round
of play one of K arms, each characterized by an unknown reward distribution.
Reward realizations are only observed when an arm is selected, and the gambler?s
objective is to maximize his cumulative expected earnings over some given horizon of play T . To do this, the gambler needs to acquire information about arms
(exploration) while simultaneously optimizing immediate rewards (exploitation);
the price paid due to this trade off is often referred to as the regret, and the main
question is how small can this price be as a function of the horizon length T . This
problem has been studied extensively when the reward distributions do not change
over time; an assumption that supports a sharp characterization of the regret, yet is
often violated in practical settings. In this paper, we focus on a MAB formulation
which allows for a broad range of temporal uncertainties in the rewards, while still
maintaining mathematical tractability. We fully characterize the (regret) complexity of this class of MAB problems by establishing a direct link between the extent
of allowable reward ?variation? and the minimal achievable regret, and by establishing a connection between the adversarial and the stochastic MAB frameworks.
1
Introduction
Background and motivation. In the presence of uncertainty and partial feedback on rewards, an
agent that faces a sequence of decisions needs to judiciously use information collected from past
observations when trying to optimize future actions. A widely studied paradigm that captures this
tension between the acquisition cost of new information (exploration) and the generation of instantaneous rewards based on the existing information (exploitation), is that of multi armed bandits
(MAB), originally proposed in the context of drug testing by [1], and placed in a general setting
by [2]. The original setting has a gambler choosing among K slot machines at each round of play,
and upon that selection observing a reward realization. In this classical formulation the rewards
are assumed to be independent and identically distributed according to an unknown distribution
that characterizes each machine. The objective is to maximize the expected sum of (possibly discounted) rewards received over a given (possibly infinite) time horizon. Since their inception, MAB
problems with various modifications have been studied extensively in Statistics, Economics, Operations Research, and Computer Science, and are used to model a plethora of dynamic optimization
problems under uncertainty; examples include clinical trials ([3]), strategic pricing ([4]), investment
in innovation ([5]), packet routing ([6]), on-line auctions ([7]), assortment selection ([8]), and on1
line advertising ([9]), to name but a few. For overviews and further references cf. the monographs
by [10], [11] for Bayesian / dynamic programming formulations, and [12] that covers the machine
learning literature and the so-called adversarial setting. Since the set of MAB instances in which one
can identify the optimal policy is extremely limited, a typical yardstick to measure performance of a
candidate policy is to compare it to a benchmark: an oracle that at each time instant selects the arm
that maximizes expected reward. The difference between the performance of the policy and that of
the oracle is called the regret. When the growth of the regret as a function of the horizon T is sublinear, the policy is long-run average optimal: its long run average performance converges to that of
the oracle. Hence the first order objective is to develop policies with this characteristic. The precise
rate of growth of the regret as a function of T provides a refined measure of policy performance.
[13] is the first paper that provides a sharp characterization of the regret growth rate in the context of
the traditional (stationary random rewards) setting, often referred to as the stochastic MAB problem.
Most of the literature has followed this path with the objective of designing policies that exhibit the
?slowest possible? rate of growth in the regret (often referred to as rate optimal policies).
In many application domains, several of which were noted above, temporal changes in the reward
distribution structure are an intrinsic characteristic of the problem. These are ignored in the traditional stochastic MAB formulation, but there have been several attempts to extend that framework.
The origin of this line of work can be traced back to [14] who considered a case where only the state
of the chosen arm can change, giving rise to a rich line of work (see, e.g., [15], and [16]). In particular, [17] introduced the term restless bandits; a model in which the states (associated with reward
distributions) of arms change in each step according to an arbitrary, yet known, stochastic process.
Considered a hard class of problems (cf. [18]), this line of work has led to various approximations
(see, e.g., [19]), relaxations (see, e.g., [20]), and considerations of more detailed processes (see, e.g.,
[21] for irreducible Markov process, and [22] for a class of history-dependent rewards).
Departure from the stationarity assumption that has dominated much of the MAB literature raises
fundamental questions as to how one should model temporal uncertainty in rewards, and how to
benchmark performance of candidate policies. One view, is to allow the reward realizations to be
selected at any point in time by an adversary. These ideas have their origins in game theory with the
work of [23] and [24], and have since seen significant development; [25] and [12] provide reviews
of this line of research. Within this so called adversarial formulation, the efficacy of a policy over a
given time horizon T is often measured relative to a benchmark defined by the single best action one
could have taken in hindsight (after seeing all reward realizations). The single best action benchmark
represents a static oracle, as it is constrained to a single (static) action. This static oracle can perform
quite poorly relative to a dynamic oracle that follows the optimal dynamic sequence of actions, as
the latter optimizes the (expected) reward at each time instant over all possible actions.1 Thus, a
potential limitation of the adversarial framework is that even if a policy has a ?small? regret relative
to a static oracle, there is no guarantee with regard to its performance relative to the dynamic oracle.
Main contributions. The main contribution of this paper lies in fully characterizing the (regret)
complexity of a broad class of MAB problems with non-stationary reward structure by establishing
a direct link between the extent of reward ?variation? and the minimal achievable regret. More
specifically, the paper?s contributions are along four dimensions. On the modeling side we formulate
a class of non-stationary reward structure that is quite general, and hence can be used to realistically
capture a variety of real-world type phenomena, yet is mathematically tractable. The main constraint
that we impose on the evolution of the mean rewards is that their variation over the relevant time
horizon is bounded by a variation budget VT ; a concept that was recently introduced in [26] in the
context of non-stationary stochastic approximation. This limits the power of nature compared to
the adversarial setup discussed above where rewards can be picked to maximally affect the policy?s
performance at each instance within {1, . . . , T }. Nevertheless, this constraint allows for a rich
class of temporal changes, extending most of the treatment in the non-stationary stochastic MAB
literature, which mainly focuses on a finite number of changes in the mean rewards, see, e.g., [27]
and references therein. We further discuss connections with studied non-stationary instances in ?6.
The second dimension of contribution lies in the analysis domain. For a general class of
non-stationary reward distributions we establish lower bounds on the performance of any nonanticipating policy relative to the dynamic oracle, and show that these bounds can be achieved,
1
Under non-stationary rewards it is immediate that the single best action may be sub-optimal in many
decision epochs, and the performance gap between the static and the dynamic oracles can grow linearly with T .
2
uniformly over the class of admissible reward distributions, by a suitable policy construction. The
term ?achieved? is meant in the sense of the order of the regret as a function of the time horizon T ,
the variation budget VT , and the number of arms K. Our policies are shown to be minimax optimal
up to a term that is logarithmic in the number of arms, and the regret is sublinear and is of order
1/3
(KVT ) T 2/3 . Our analysis complements studied non-stationary instances by treating a broad and
flexible class of temporal changes in the reward distributions, yet still establishing optimality results
and showing that sublinear regret is achievable. Our results provide a spectrum of orders of the
minimax regret ranging between order T 2/3 (when VT is a constant independent of T ) and order T
(when VT grows linearly with T ), mapping allowed variation to best achievable performance.
With the analysis described above we shed light on the exploration-exploitation trade off that characterizes the non-stationary reward setting, and the change in this trade off compared to the stationary
setting. In particular, our results highlight the tension that exists between the need to ?remember?
and ?forget.? This is characteristic of several algorithms that have been developed in the adversarial MAB literature, e.g., the family of exponential weight methods such as EXP3, EXP3.S and the
like; see, e.g., [28], and [12]. In a nutshell, the fewer past observations one retains, the larger the
stochastic error associated with one?s estimates of the mean rewards, while at the same time using
more past observations increases the risk of these being biased.
One interesting observation drawn in this paper connects between the adversarial MAB setting, and
the non-stationary environment studied here. In particular, as in [26], it is seen that an optimal policy
in the adversarial setting may be suitably calibrated to perform near-optimally in the non-stationary
stochastic setting. This will be further discussed after the main results are established.
2
Problem Formulation
Let K = {1, . . . , K} be a set of arms. Let T = {1, 2, . . . , T } denote a sequence of decision epochs
faced by a decision maker. At any epoch t ? T , the decision-maker pulls one of the K arms. When
pulling arm k ? K at epoch t ? T , a reward Xtk ? [0, 1] is obtained, where Xtk is a random
variable with expectation ?kt = E Xtk . We denote the best possible expected reward at decision
epoch t by ??t , i.e., ??t = maxk?K ?kt .
Changes in the expected rewards of the arms. We assume the expected reward of each arm ?kt
may change at any decision epoch. We denote by ?k the sequence of expected rewards of arm k:
T
?k = ?kt t=1 . In addition, we denote by ? the sequence of vectors of all K expected rewards:
k K
? = ? k=1 . We assume that the expected reward of each arm can change an arbitrary number of
times, but bound the total variation of the expected rewards:
T
?1
X
sup ?kt ? ?kt+1 .
(1)
t=1 k?K
Let {Vt : t = 1, 2, . . .} be a non-decreasing sequence of positive real numbers such that V1 = 0,
KVt ? t for all t, and for normalization purposes set V2 = 2 ? K ?1 . We refer to VT as the variation
budget over T . We define the corresponding temporal uncertainty set, as the set of reward vector
sequences that are subject to the variation budget VT over the set of decision epochs {1, . . . , T }:
(
)
T
?1
X
k
K?T
k
V = ? ? [0, 1]
:
sup ?t ? ?t+1 ? VT .
t=1 k?K
The variation budget captures the constraint imposed on the non-stationary environment faced by
the decision-maker. While limiting the possible evolution in the environment, it allows for numerous forms in which the expected rewards may change: continuously, in discrete shocks, and of a
changing rate (Figure 1 depicts two different variation patterns that correspond to the same variation
budget). In general, the variation budget VT is designed to depend on the number of pulls T .
Admissible policies, performance, and regret. Let U be a random variable defined over a probability space (U, U, Pu ). Let ?1 : U ? K and ?t : [0, 1]t?1 ? U ? K for t = 2, 3, . . . be measurable
functions. With some abuse of notation we denote by ?t ? K the action at time t, that is given by
?1 (U )
t = 1,
?t =
?
?t Xt?1
, . . . , X1? , U
t = 2, 3, . . . ,
3
Figure 1: Two instances of variation in the mean rewards: (Left) A fixed variation budget (that equals 3) is
?spent? over the whole horizon. (Right) The same budget is ?spent? in the first third of the horizon.
The mappings {?t : t = 1, . . . , T } together with the distribution Pu define the class of admissible
policies. We denote this class by P. We further denote by {H
1, . . . ,T } the filtration associt , t =
t?1
ated with a policy ? ? P, such that H1 = ? (U ) and Ht = ? Xj? j=1 , U for all t ? {2, 3, . . .}.
Note that policies in P are non-anticipating, i.e., depend only on the past history of actions and observations, and allow for randomized strategies via their dependence on U .
We define the regret under policy ? ? P compared to a dynamic oracle as the worst-case difference
between the expected performance of pulling at each epoch t the arm which has the highest expected
reward at epoch t (the dynamic oracle performance) and the expected performance under policy ?:
( T
" T
#)
X
X
?
?
?
?
R (V, T ) = sup
?t ? E
?t
,
??V
t=1
t=1
?
where the expectation E [?] is taken with respect to the noisy rewards, as well as to the policy?s
actions. In addition, we denote by R? (V, T ) the minimal worst-case regret that can be guaranteed
by an admissible policy ? ? P, that is, R? (V, T ) = inf ??P R? (V, T ). Then, R? (V, T ) is the best
achievable performance. In the following sections we study the magnitude of R? (V, T ). We analyze
the magnitude of this quantity by establishing upper and lower bounds; in these bounds we refer to
a constant C as absolute if it is independent of K, VT , and T .
3
Lower bound on the best achievable performance
We next provide a lower bound on the the best achievable performance.
Theorem 1 Assume that rewards have a Bernoulli distribution. Then, there is some
absolute con
stant C > 0 such that for any policy ? ? P and for any T ? 1, K ? 2 and VT ? K ?1 , K ?1 T ,
1/3
R? (V, T ) ? C (KVT )
T 2/3 .
We note that when reward distributions
are stationary, there are known policies such as UCB1 ([29])
?
that achieve regret of order T in the stochastic setup. When the reward structure is non-stationary
and defined by the class V, then no policy may achieve such a performance and the best performance
must incur a regret of at least order T 2/3 . This additional complexity embedded in the non-stationary
stochastic MAB problem compared to the stationary one will be further discussed in ?6. We note
that Theorem 1 also holds when VT is increasing with T . In particular, when the variation budget is
linear in T , the regret grows linearly and long run average optimality is not achievable.
The driver of the change in the best achievable performance relative to the one established in a
stationary environment, is a second tradeoff (over the tension between exploring different arms and
capitalizing on the information already collected) introduced by the non-stationary environment,
between ?remembering? and ?forgetting?: estimating the expected rewards is done based on past
observations of rewards. While keeping track of more observations may decrease the variance of
mean rewards estimates, the non-stationary environment implies that ?old? information is potentially
less relevant due to possible changes in the underlying rewards. The changing rewards give incentive
to dismiss old information, which in turn encourages enhanced exploration. The proof of Theorem 1
emphasizes the impact of these tradeoffs on the achievable performance.
4
Key ideas in the proof. At a high level the proof of Theorem 1 builds on ideas of identifying
a worst-case ?strategy? of nature (e.g., [28], proof of Theorem 5.1) adapting them to our setting.
While the proof is deferred to the online companion (as supporting material), we next describe the
key ideas when VT = 1.2 We define a subset of vector sequences V 0 ? V and show that when
1/3
? is drawn randomly from V 0 , any admissible policy must incur regret of order (KVT ) T 2/3 .
? T each (except,
We define a partition of the decision horizon T into batches T1 , . . . , Tm of size ?
possibly the last batch):
n
n
oo
? T + 1 ? t ? min j ?
?T,T
Tj = t : (j ? 1)?
, for all j = 1, . . . , m,
(2)
? T e is the number of batches. In V 0 , in every batch there is exactly one ?good?
where m = dT /?
arm with expected reward 1/2 + ? for some 0 < ? ? 1/4, and all the other arms have expected
reward 1/2. The ?good? arm is drawn independently in the beginning of each batch according to
a discrete uniform distribution over {1, . . . , K}. Thus, the identity of the ?good? arm can change
? T ? VT , any ? ? V 0 is composed of expected
only between batches. By selecting ? such that ?T /?
reward sequences with a variation of at most VT , and therefore V 0 ? V. Given the draws under which
expected reward sequences are generated, nature prevents any accumulation of information from one
batch to another, since at the beginning of each batch a new ?good?parm is drawn independently of
? T no admissible policy can
the history. The proof of Theorem 1 establishes that when ? ? 1/ ?
? T epochs in each
identify the ?good? arm with high probability within a batch. Since there are p
?
?
?
batch, the regret that
p any policy must incur
p along a batch is of order ?T ? ? ? ?T , which yields
?
?
?
a regret of order ?T ? T /?T ? T / ?T throughout the whole horizon. Selecting the smallest
? T such that the variation budget constraint is satisfied leads to ?
? T ? T 2/3 , yielding a
feasible ?
2/3
regret of order T
throughout the horizon.
4
A near-optimal policy
We apply the ideas underlying the lower bound in Theorem 1 to develop a rate optimal policy for
the non-stationary stochastic MAB problem with a variation budget. Consider the following policy:
Rexp3. Inputs: a positive number ?, and a batch size ?T .
1. Set batch index j = 1
2. Repeat while j ? dT /?T e:
(a) Set ? = (j ? 1) ?T
(b) Initialization: for any k ? K set wtk = 1
(c) Repeat for t = ? + 1, . . . , min {T, ? + ?T }:
? For each k ? K, set
wk
pkt = (1 ? ?) PK t
k0
k0 =1 wt
+
?
K
K
? Draw an arm k 0 from K according to the distribution pkt k=1
0
? Receive a reward Xtk
? k0 = X k0 /pk0 , and for any k 6= k 0 set X
? k = 0. For all k ? K update:
? For k 0 set X
t
t
t
t
(
)
? tk
?
X
k
wt+1
= wtk exp
K
(d) Set j = j + 1, and return to the beginning of step 2
Clearly ? ? P. The Rexp3 policy uses Exp3, a policy introduced by [30] for solving a worst-case
sequential allocation problem, as a subroutine, restarting it every ?T epochs.
2
For the sake of simplicity, the discussion in this paragraph assumes a variation budget that is fixed and
independent of T ; the proof of Theorem 3 details a general treatment for a budget that depends on T .
5
l
m
1/3
2/3
Theorem 2 Let ? be the Rexp3 policy with a batch size ?T = (K log K) (T /VT )
and
q
n
o
K log K
?
with ? = min 1 ,
(e?1)?T . Then, there is some absolute constant C such that for every T ? 1,
?1
K ? 2, and VT ? K , K ?1 T :
1/3
R? (V, T ) ? C? (K log K ? VT ) T 2/3 .
Theorem 2 is obtained by establishing a connection between the regret relative to the single best
action in the adversarial setting, and the regret with respect to the dynamic oracle in non-stationary
stochastic setting with variation budget. Several classes of policies, such as exponential-weight
?
(including Exp3) and polynomial-weight policies, have been shown to achieve regret of order T
with respect to the single best action in the adversarial setting (see chapter 6 of [12] for a review).
While in general these policies tend to perform well numerically, there is no guarantee for their
performance relative to the dynamic oracle studied in this paper, since the single best action itself
may incur linear regret relative to the dynamic oracle; see also [31] for a study of the empirical
performance of?
one class of algorithms. The proof of Theorem 2 shows that any policy that achieves
regret of order T with respect to the single best action in the adversarial setting, can be used as a
subroutine to obtain near-optimal performance with respect to the dynamic oracle in our setting.
Rexp3 emphasizes the two tradeoffs discussed in the previous section. The first tradeoff, information
acquisition versus capitalizing on existing information, is captured by the subroutine policy Exp3. In
fact, any policy that achieves a good performance compared to a single best action benchmark in the
adversarial setting must balance exploration and exploitation. The second tradeoff, ?remembering?
versus ?forgetting,? is captured by restarting Exp3 and forgetting any acquired information every
?T pulls. Thus, old information that may slow down the adaptation to the changing environment
is being discarded. Theorem 1 and Theorem 2 together characterize the minimax regret (up to a
multiplicative factor, logarithmic in the number of arms) in a full spectrum of variations VT :
R? (V, T ) (KVT )
1/3
T 2/3 .
Hence, we have quantified the impact of the extent of change in the environment on the best achievable performance in this broad class of problems. For example, for the case in which VT = C ? T ? ,
for some absolute constant C and 0 ? ? < 1 the best achievable regret is of order T (2+?)/3 .
We finally note that restarting is only one way of adapting policies from the adversarial MAB setting
to achieve near optimality in the non-stationary stochastic setting; a way that articulates well the
principles leading to near optimality. In the online companion we demonstrate that near optimality
can be achieved by other adaptation methods, showing that the Exp3.S policy (given in [28]) can be
1/3
to achieve near optimality in our setting, without restarting.
tuned by ? = T1 and ? ? (KVT /T )
5
Proof of Theorem 2
The structure of the proof is as follows. First, we break the horizon to a sequence of batches of size
?T each, and analyze the performance gap between the single best action and the dynamic oracle
in each batch. Then, we plug in a known performance guarantee for Exp3 relative to the single best
action, and sum over batches to establish the regret of Rexp3 relative to the dynamic oracle.
Step 1 (Preliminaries). Fix T ? 1, K ? 2, and VT ? K ?1 , K ?1 T . Let ? be the Rexp3 policy,
o
n q
K log K
and ?T ? {1, . . . , T } (to be specified later on). We break the
tuned by ? = min 1 ,
(e?1)?T
horizon T into a sequence of batches T1 , . . . , Tm of size ?T each (except, possibly Tm ) according
to (2). Let ? ? V, and fix j ? {1, . . . , m}. We decomposition the regret in batch j:
?
?
E ?
?
X
t?Tj
(??t
?
??t )?
?
=
X
t?Tj
|
??t
? E ?max
k?K
?
?X
?
Xtk
t?Tj
{z
J1,j
??
?
??
?
?
?
?X
?
?
X ?
k
?
? + E ?max
Xt ? ? E ?
?t ? .
k?K ?
?
?
t?Tj
t?Tj
} |
{z
}
J2,j
(3)
The first component, J1,j , is the expected loss associated with using a single action over batch j.
The second component, J2,j , is the expected regret relative to the best static action in batch j.
6
Step 2 (Analysis of J1,j and J2,j ). Defining ?kT +1 = ?kT for all k ? K,
we denote the variation in
P
k
k
expected rewards along batch Tj by Vj = t?Tj maxk?K ?t+1 ? ?t . We note that:
m
X
Vj =
j=1
m X
X
j=1 t?Tj
max ?kt+1 ? ?kt ? VT .
(4)
k?K
o
nP
k
. Then,
?
Let k0 be an arm with best expected performance over Tj : k0 ? arg maxk?K
t
t?Tj
?
?
??
?
?
?
?
?X ?
?
?X
X
X
max
?kt
=
?kt 0 = E ?
(5)
Xtk0 ? ? E ?max
Xtk ? ,
?
?
k?K ?
k?K ?
t?Tj
t?Tj
t?Tj
t?Tj
and therefore, one has:
?
J1,j
X
=
??t ? E ?max
k?K
t?Tj
?
?X
?
t?Tj
??
? (a) X
??t ? ?kt 0
Xtk ? ?
?
t?Tj
n
o (b)
? ?T max ??t ? ?kt 0 ? 2Vj ?T ,
(6)
t?Tj
for any ? ? V and j ? {1, . . . , m}, where (a) holds by (5) and (b) holds by the following argument:
otherwise there is an epoch t0 ? Tj for which ??t0 ? ?kt00 > 2Vj . Indeed, let k1 = arg maxk?K ?kt0 .
In such case, for all t ? Tj one has ?kt 1 ? ?kt01 ? Vj > ?kt00 + Vj ? ?kt 0 , since Vj is the maximal
variation in batch Tj . This however, contradicts the optimality of k0 at epoch t, and thus (6) holds.
In addition,
q Corollary
n
o 3.2 in [28] points out that the regret incurred by Exp3 (tuned by ? =
K log K
min 1 ,
(e?1)?T ) along ?T batches, relative to the single best action, is bounded by
?
?
2 e ? 1 ?T K log K. Therefore, for each j ? {1, . . . , m} one has
?
?
?
?
??
?X
?
X
(a)
p
?
J2,j = E ?max
Xtk ? E? ?
(7)
??t ?? ? 2 e ? 1 ?T K log K,
?
k?K ?
t?Tj
t?Tj
for any ? ? V, where (a) holds since within each batch arms are pulled according to Exp3(?).
Step 3 (Regret throughout the horizon). Summing over m = dT /?T e batches we have:
( T
" T
#)
m
X
X
(a) X
p
?
?
?
?
?
R (V, T ) = sup
?t ? E
?t
?
2 e ? 1 ?T K log K + 2Vj ?T
??V
(b)
?
=
t=1
t=1
j=1
?
p
T
+ 1 ? 2 e ? 1 ?T K log K + 2?T VT .
?T
?
?
p
?
2 e ? 1 K log K ? T
?
+ 2 e ? 1 ?T K log K + 2?T VT ,
?T
where: l (a) holds by (3), (6),m and (7); and (b) follows from (4).
1/3
2/3
, we establish:
?T = (K log K) (T /VT )
R? (V, T )
?
(a)
?
Finally, selecting
?
1/3
2 e ? 1 (K log K ? VT ) T 2/3
r
?
1/3
2/3
+2 e ? 1
(K log K) (T /VT ) + 1 K log K
1/3
2/3
+2 (K log K) (T /VT ) + 1 VT
? ?
1/3
2+2 2
e ? 1 + 4 (K log K ? VT ) T 2/3 ,
where (a) follows from T ? K ? 2, and VT ? K ?1 , K ?1 T . This concludes the proof.
7
(8)
6
Discussion
Unknown variation budget. The Rexp3 policy relies on prior knowledge of VT , but predictions
of VT may be inaccurate (such estimation can be maintained from historical data if actions are
occasionally randomized, for example, by fitting VT = T ? ). Denoting the ?true? variation budget
by VT and the estimate that is used by the agent when tuning Rexp3 by V?T , one may observe that
the analysis in the proof of Theorem 2 holds until equation (8), but then ?T will be tuned using V?T .
This implies that when VT and V?T are ?close,? Rexp3 still guarantees long-run average optimality.
For example, suppose that Rexp3 is tuned by V?T = T ? , but the variation is VT = T ?+? . Then
sublinear regret (of order T 2/3+?/3+? ) is guaranteed as long as ? < (1 ? ?)/3; e.g., if ? = 0 and
? = 1/4, Rexp3 guarantees regret of order T 11/12 (accurate tuning would have guaranteed order
T 3/4 ). Since there are no restrictions on the rate at which the variation budget can be spent, an
interesting and potentially challenging open problem is to delineate to what extent it is possible to
design adaptive policies that do not use prior knowledge of VT , yet guarantee ?good? performance.
Contrasting with traditional (stationary) ?
MAB problems. The characterized minimax regret in
the stationary stochastic setting is of order T when expected rewards can be arbitrarily close to
each other, and of order log T when rewards are ?well separated? (see [13] and [29]). Contrast1/3
ing the minimax regret (of order VT T 2/3 ) we have established in the stochastic non-stationary
MAB problem with those established in stationary settings allows one to quantify the ?price of nonstationarity,? which mathematically captures the added complexity embedded in changing rewards
versus stationary ones (as a function of the allowed variation). Clearly, additional complexity is
introduced even when the allowed variation is fixed and independent of the horizon length.
Contrasting with other non-stationary MAB instances. The class of MAB problems with nonstationary rewards that is formulated in the current chapter extends other MAB formulations that
allow rewards to change in a more structured manner. For example, [32] consider a setting where
rewards evolve according to a Brownian motion and regret is linear in T ; our results (when VT is
linear in T ) are consistent with theirs. Two other representative studies are those of [27], that study a
stochastic MAB problems in which expected rewards may change a finite number of times, and [28]
that formulate an adversarial MAB problem in which the identity of the best arm may change a finite
number of times. Both studies suggest policies?that, utilizing the prior knowledge that the number of
changes must be finite, achieve regret of order T relative to the best sequence of actions. However,
the performance of these policies can deteriorate to regret that is linear in T when the number of
changes is allowed to depend on T . When there is a finite variation (VT is fixed and independent of
T ) but not necessarily a finite number of changes, we establish that the best achievable performance
deteriorate to regret of order T 2/3 . In that respect, it is not surprising that the ?hard case? used to
establish the lower bound in Theorem 1 describes a nature?s strategy that allocates variation over a
large (as a function of T ) number of changes in the expected rewards.
Low variation rates. While our formulation focuses on ?significant? variation in the mean rewards,
our established bounds also hold for ?smaller? variation scales;
when VT decreases from O(1) to
?
O(T ?1/2 ) the minimax regret rate decreases from T 2/3 to T . Indeed, when the variation scale is
O(T ?1/2 ) or smaller, the rate of regret coincides with that of the classical stochastic MAB setting.
References
[1] W. R. Thompson. On the likelihood that one unknown probability exceeds another in view of the evidence
of two samples. Biometrika, 25:285?294, 1933.
[2] H. Robbins. Some aspects of the sequential design of experiments. Bulletin of the American Mathematical
Society, 55:527?535, 1952.
[3] M. Zelen. Play the winner rule and the controlled clinical trials. Journal of the American Statistical
Association, 64:131?146, 1969.
[4] D. Bergemann and J. Valimaki. Learning and strategic pricing. Econometrica, 64:1125?1149, 1996.
[5] D. Bergemann and U. Hege. The financing of innovation: Learning and stopping. RAND Journal of
Economics, 36 (4):719?752, 2005.
8
[6] B. Awerbuch and R. D. Kleinberg. Addaptive routing with end-to-end feedback: distributed learning and
geometric approaches. In Proceedings of the 36th ACM Symposiuim on Theory of Computing (STOC),
pages 45?53, 2004.
[7] R. D. Kleinberg and T. Leighton. The value of knowing a demand curve: Bounds on regret for online
posted-price auctions. In Proceedings of the 44th Annual IEEE Symposium on Foundations of Computer
Science (FOCS), pages 594?605, 2003.
[8] F. Caro and G. Gallien. Dynamic assortment with demand learning for seasonal consumer goods. Management Science, 53:276?292, 2007.
[9] S. Pandey, D. Agarwal, D. Charkrabarti, and V. Josifovski. Bandits for taxonomies: A model-based
approach. In SIAM International Conference on Data Mining, 2007.
[10] D. A. Berry and B. Fristedt. Bandit problems: sequential allocation of experiments. Chapman and Hall,
1985.
[11] J. C. Gittins. Multi-Armed Bandit Allocation Indices. John Wiley and Sons, 1989.
[12] N. Cesa-Bianchi and G. Lugosi. Prediction, Learning, and Games. Cambridge University Press, Cambridge, UK, 2006.
[13] T. L. Lai and H. Robbins. Asymptotically efficient adaptive allocation rules. Advances in Applied Mathematics, 6:4?22, 1985.
[14] J. C. Gittins and D. M. Jones. A dynamic allocation index for the sequential design of experiments.
North-Holland, 1974.
[15] J. C. Gittins. Bandit processes and dynamic allocation indices (with discussion). Journal of the Royal
Statistical Society, Series B, 41:148?177, 1979.
[16] P. Whittle. Arm acquiring bandits. The Annals of Probability, 9:284?292, 1981.
[17] P. Whittle. Restless bandits: Activity allocation in a changing world. Journal of Applied Probability,
25A:287?298, 1988.
[18] C. H. Papadimitriou and J. N. Tsitsiklis. The complexity of optimal queueing network control. In Structure
in Complexity Theory Conference, pages 318?322, 1994.
[19] D. Bertsimas and J. Nino-Mora. Restless bandits, linear programming relaxations, and primal dual index
heuristic. Operations Research, 48(1):80?90, 2000.
[20] S. Guha and K. Munagala. Approximation algorithms for partial-information based stochastic control
with markovian rewards. In 48th Annual IEEE Symposium on Fundations of Computer Science (FOCS),
pages 483?493, 2007.
[21] R. Ortner, D. Ryabko, P. Auer, and R. Munos. Regret bounds for restless markov bandits. In Algorithmic
Learning Theory, pages 214?228. Springer Berlin Heidelberg, 2012.
[22] M. G. Azar, A. Lazaric, and E. Brunskill. Stochastic optimization of a locally smooth function under
correlated bandit feedback. arXiv preprint arXiv:1402.0562, 2014.
[23] D. Blackwell. An analog of the minimax theorem for vector payoffs. Pacific Journal of Mathematics,
6:1?8, 1956.
[24] J. Hannan. Approximation to bayes risk in repeated plays, Contributions to the Theory of Games, Volume
3. Princeton University Press, Cambridge, UK, 1957.
[25] D. P. Foster and R. V. Vohra. Regret in the on-line decision problem. Games and Economic Behaviour,
29:7?35, 1999.
[26] O. Besbes, Y. Gur, and A. Zeevi. Non-stationary stochastic optimization. Working paper, 2014.
[27] A. Garivier and E. Moulines. On upper-confidence bound policies for switching bandit problems. In
Algorithmic Learning Theory, pages 174?188. Springer Berlin Heidelberg, 2011.
[28] P. Auer, N. Cesa-Bianchi, Y. Freund, and R. E. Schapire. The non-stochastic multi-armed bandit problem.
SIAM journal of computing, 32:48?77, 2002.
[29] P. Auer, N. Cesa-Bianchi, and P. Fischer. Finite-time analysis of the multiarmed bandit problem. Machine
Learning, 47:235?246, 2002.
[30] Y. Freund and R. E. Schapire. A decision-theoretic generalization of on-line learning and an application
to boosting. J. Comput. System Sci., 55:119?139, 1997.
[31] C. Hartland, S. Gelly, N. Baskiotis, O. Teytaud, and M. Sebag. Multi-armed bandit, dynamic environments
and meta-bandits. NIPS-2006 workshop, Online trading between exploration and exploitation, Whistler,
Canada, 2006.
[32] A. Slivkins and E. Upfal. Adapting to a changing environment: The brownian restless bandits. In Proceedings of the 21st Annual Conference on Learning Theory (COLT), pages 343?354, 2008.
9
| 5378 |@word trial:2 exploitation:5 achievable:13 leighton:1 polynomial:1 suitably:1 open:1 decomposition:1 paid:1 series:1 efficacy:1 selecting:3 tuned:5 denoting:1 past:5 existing:2 current:1 surprising:1 yet:5 must:5 john:1 partition:1 j1:4 treating:1 designed:1 update:1 stationary:32 selected:2 fewer:1 beginning:3 characterization:2 provides:2 boosting:1 earnings:1 teytaud:1 mathematical:2 along:4 direct:2 driver:1 symposium:2 focs:2 fitting:1 assaf:2 paragraph:1 manner:1 deteriorate:2 acquired:1 forgetting:3 indeed:2 expected:27 multi:6 moulines:1 discounted:1 decreasing:1 armed:6 increasing:1 estimating:1 bounded:2 notation:1 maximizes:1 gsb:1 underlying:2 what:1 developed:1 contrasting:2 hindsight:1 guarantee:6 temporal:6 remember:1 every:4 growth:4 shed:1 nutshell:1 exactly:1 biometrika:1 uk:2 control:2 positive:2 t1:3 limit:1 switching:1 establishing:6 path:1 abuse:1 lugosi:1 therein:1 studied:7 initialization:1 quantified:1 challenging:1 josifovski:1 limited:1 range:1 practical:1 testing:1 investment:1 regret:51 empirical:1 drug:1 adapting:3 confidence:1 seeing:1 suggest:1 close:2 selection:2 context:3 risk:2 optimize:1 measurable:1 imposed:1 accumulation:1 restriction:1 pk0:1 economics:2 independently:2 thompson:1 formulate:2 simplicity:1 identifying:1 rule:2 utilizing:1 pull:3 gur:2 his:1 mora:1 variation:36 limiting:1 annals:1 construction:1 play:5 enhanced:1 suppose:1 programming:2 us:1 designing:1 origin:2 observed:1 preprint:1 capture:4 worst:4 ryabko:1 trade:3 highest:1 decrease:3 monograph:1 environment:10 complexity:7 reward:70 econometrica:1 dynamic:19 raise:1 depend:3 solving:1 incur:4 upon:1 k0:7 various:2 chapter:2 separated:1 describe:1 choosing:1 refined:1 quite:2 heuristic:1 stanford:3 widely:1 larger:1 otherwise:1 statistic:1 fischer:1 noisy:1 itself:1 online:4 sequence:13 maximal:1 adaptation:2 j2:4 relevant:2 realization:4 poorly:1 achieve:6 realistically:1 plethora:1 extending:1 gittins:3 converges:1 tk:1 spent:3 oo:1 develop:2 measured:1 received:1 implies:2 trading:1 quantify:1 stochastic:22 exploration:6 packet:1 routing:2 munagala:1 material:1 behaviour:1 fix:2 generalization:1 preliminary:1 mab:25 mathematically:2 exploring:1 hold:8 considered:2 hall:1 exp:1 mapping:2 algorithmic:2 zeevi:2 kvt:6 achieves:2 smallest:1 purpose:1 estimation:1 maker:3 robbins:2 establishes:1 clearly:2 corollary:1 focus:3 seasonal:1 bernoulli:1 likelihood:1 mainly:1 slowest:1 adversarial:14 sense:1 dependent:1 stopping:1 inaccurate:1 bandit:19 subroutine:3 selects:1 arg:2 among:1 flexible:1 dual:1 colt:1 development:1 constrained:1 equal:1 chapman:1 represents:1 broad:4 jones:1 future:1 papadimitriou:1 np:1 few:1 irreducible:1 ortner:1 randomly:1 composed:1 simultaneously:1 connects:1 attempt:1 stationarity:1 mining:1 deferred:1 yielding:1 light:1 primal:1 tj:24 kt:16 accurate:1 partial:2 allocates:1 old:3 minimal:3 instance:6 modeling:1 markovian:1 cover:1 retains:1 tractability:1 cost:1 strategic:2 subset:1 uniform:1 guha:1 characterize:2 optimally:1 calibrated:1 st:1 fundamental:1 randomized:2 siam:2 international:1 off:3 together:2 continuously:1 satisfied:1 management:1 cesa:3 choose:1 possibly:4 american:2 leading:1 return:1 potential:1 whittle:2 wk:1 north:1 depends:1 ated:1 multiplicative:1 h1:1 view:2 picked:1 break:2 observing:1 characterizes:2 sup:4 analyze:2 bayes:1 contribution:5 variance:1 characteristic:3 who:1 correspond:1 identify:2 yield:1 bayesian:1 emphasizes:2 advertising:1 vohra:1 history:3 nonstationarity:1 acquisition:2 associated:3 proof:12 static:6 con:1 fristedt:1 treatment:2 knowledge:3 anticipating:1 auer:3 back:1 originally:1 dt:3 tension:3 maximally:1 rand:1 formulation:8 done:1 delineate:1 inception:1 until:1 working:1 dismiss:1 pricing:2 grows:2 pulling:2 name:1 concept:1 true:1 evolution:2 hence:3 awerbuch:1 round:2 game:4 encourages:1 maintained:1 noted:1 coincides:1 trying:1 allowable:1 theoretic:1 demonstrate:1 motion:1 auction:2 ranging:1 instantaneous:1 consideration:1 recently:1 overview:1 winner:1 volume:1 caro:1 extend:1 discussed:4 association:1 analog:1 numerically:1 theirs:1 significant:2 refer:2 multiarmed:1 cambridge:3 tuning:2 mathematics:2 associ:1 pu:2 brownian:2 optimizing:1 optimizes:1 inf:1 occasionally:1 yonatan:1 meta:1 arbitrarily:1 vt:42 seen:2 captured:2 additional:2 remembering:2 impose:1 maximize:2 paradigm:1 full:1 hannan:1 ing:1 exceeds:1 smooth:1 characterized:2 exp3:10 clinical:2 long:5 plug:1 lai:1 controlled:1 impact:2 prediction:2 expectation:2 stant:1 arxiv:2 normalization:1 agarwal:1 achieved:3 receive:1 background:1 addition:3 grow:1 biased:1 subject:1 tend:1 nonstationary:1 near:7 presence:1 besbes:2 identically:1 variety:1 affect:1 xj:1 economic:1 idea:5 tm:3 knowing:1 tradeoff:5 judiciously:1 gambler:4 t0:2 york:2 action:22 ignored:1 detailed:1 extensively:2 locally:1 schapire:2 lazaric:1 track:1 discrete:2 incentive:1 key:2 four:1 nevertheless:1 traced:1 drawn:4 queueing:1 changing:6 garivier:1 ht:1 shock:1 v1:1 asymptotically:1 relaxation:2 bertsimas:1 sum:2 run:4 uncertainty:5 extends:1 family:1 throughout:3 bergemann:2 draw:2 decision:12 on1:1 bound:13 followed:1 guaranteed:3 oracle:18 annual:3 activity:1 constraint:4 sake:1 dominated:1 kleinberg:2 aspect:1 argument:1 extremely:1 optimality:8 min:5 xtk:8 structured:1 pacific:1 according:7 describes:1 smaller:2 son:1 contradicts:1 modification:1 wtk:2 taken:2 equation:1 discus:1 turn:1 tractable:1 end:2 capitalizing:2 operation:2 assortment:2 apply:1 observe:1 v2:1 batch:26 original:1 assumes:1 include:1 cf:2 maintaining:1 instant:2 giving:1 gelly:1 k1:1 build:1 establish:5 classical:2 society:2 objective:4 question:2 quantity:1 already:1 added:1 strategy:3 dependence:1 traditional:3 exhibit:1 link:2 berlin:2 sci:1 omar:1 extent:4 collected:2 consumer:1 length:2 index:5 balance:1 acquire:1 innovation:2 setup:2 potentially:2 stoc:1 taxonomy:1 rise:1 filtration:1 design:3 policy:50 unknown:4 perform:3 bianchi:3 upper:2 observation:7 markov:2 discarded:1 benchmark:5 finite:7 supporting:1 immediate:2 maxk:4 defining:1 payoff:1 precise:1 articulates:1 sharp:2 arbitrary:2 canada:1 princeton:1 introduced:5 complement:1 blackwell:1 specified:1 connection:3 slivkins:1 established:5 nip:1 adversary:1 pattern:1 departure:1 including:1 max:8 royal:1 power:1 suitable:1 arm:28 minimax:7 numerous:1 concludes:1 columbia:4 faced:2 review:2 literature:5 epoch:13 prior:3 geometric:1 evolve:1 berry:1 relative:14 freund:2 embedded:2 fully:2 loss:1 highlight:1 sublinear:4 generation:1 limitation:1 interesting:2 allocation:7 versus:3 foundation:1 upfal:1 incurred:1 agent:2 consistent:1 principle:1 foster:1 placed:1 last:1 keeping:1 repeat:2 tsitsiklis:1 side:1 allow:3 pulled:1 face:1 characterizing:1 bulletin:1 absolute:4 munos:1 distributed:2 regard:1 feedback:3 dimension:2 curve:1 world:2 cumulative:1 rich:2 adaptive:2 historical:1 restarting:4 summing:1 assumed:1 nino:1 spectrum:2 pandey:1 pkt:2 nature:4 ca:1 heidelberg:2 necessarily:1 posted:1 domain:2 vj:8 pk:1 main:5 linearly:3 motivation:1 whole:2 azar:1 sebag:1 allowed:4 repeated:1 x1:1 referred:3 representative:1 depicts:1 ny:2 slow:1 wiley:1 sub:1 brunskill:1 exponential:2 comput:1 candidate:2 lie:2 third:1 admissible:6 theorem:17 companion:2 down:1 xt:2 showing:2 evidence:1 intrinsic:1 exists:1 workshop:1 sequential:4 magnitude:2 budget:18 horizon:16 demand:2 restless:5 gap:2 ucb1:1 forget:1 led:1 logarithmic:2 prevents:1 holland:1 acquiring:1 springer:2 relies:1 acm:1 slot:1 identity:2 formulated:1 price:4 feasible:1 hard:2 change:23 infinite:1 typical:1 specifically:1 uniformly:1 except:2 wt:2 called:3 total:1 whistler:1 support:1 latter:1 meant:1 yardstick:1 violated:1 baskiotis:1 later:1 phenomenon:1 correlated:1 |
4,836 | 5,379 | Extreme bandits
Alexandra Carpentier
Statistical Laboratory, CMS
University of Cambridge, UK
Michal Valko
SequeL team
INRIA Lille - Nord Europe, France
[email protected]
[email protected]
Abstract
In many areas of medicine, security, and life sciences, we want to allocate limited resources to different sources in order to detect extreme values. In this paper,
we study an efficient way to allocate these resources sequentially under limited
feedback. While sequential design of experiments is well studied in bandit theory,
the most commonly optimized property is the regret with respect to the maximum
mean reward. However, in other problems such as network intrusion detection, we
are interested in detecting the most extreme value output by the sources. Therefore, in our work we study extreme regret which measures the efficiency of an algorithm compared to the oracle policy selecting the source with the heaviest tail.
We propose the E XTREME H UNTER algorithm, provide its analysis, and evaluate
it empirically on synthetic and real-world experiments.
1
Introduction
We consider problems where the goal is to detect outstanding events or extreme values in domains
such as outlier detection [1], security [18], or medicine [17]. The detection of extreme values is
important in many life sciences, such as epidemiology, astronomy, or hydrology, where, for example,
we may want to know the peak water flow. We are also motivated by network intrusion detection
where the objective is to find the network node that was compromised, e.g., by seeking the one
creating the most number of outgoing connections at once. The search for extreme events is typically
studied in the field of anomaly detection, where one seeks to find examples that are far away from
the majority, according to some problem-specific distance (cf. the surveys [8, 16]).
In anomaly detection research, the concept of anomaly is ambiguous and several definitions exist [16]: point anomalies, structural anomalies, contextual anomalies, etc. These definitions are
often followed by heuristic approaches that are seldom analyzed theoretically. Nonetheless, there
exist some theoretical characterizations of anomaly detection. For instance, Steinwart et al. [19]
consider the level sets of the distribution underlying the data, and rare events corresponding to rare
level sets are then identified as anomalies. A very challenging characteristic of many problems in
anomaly detection is that the data emitted by the sources tend to be heavy-tailed (e.g., network traffic [2]) and anomalies come from the sources with the heaviest distribution tails. In this case, rare
level sets of [19] correspond to distributions? tails and anomalies to extreme values. Therefore, we
focus on the kind of anomalies that are characterized by their outburst of events or extreme values,
as in the setting of [22] and [17].
Since in many cases, the collection of the data samples emitted by the sources is costly, it is important to design adaptive-learning strategies that spend more time sampling sources that have a
higher risk of being abnormal. The main objective of our work is the active allocation of the sampling resources for anomaly detection, in the setting where anomalies are defined as extreme values.
Specifically, we consider a variation of the common setting of minimal feedback also known as
the bandit setting [14]: the learner searches for the most extreme value that the sources output by
probing the sources sequentially. In this setting, it must carefully decide which sources to observe
1
because it only receives the observation from the source it chooses to observe. As a consequence,
it needs to allocate the sampling time efficiently and should not waste it on sources that do not have
an abnormal character. We call this specific setting extreme bandits, but it is also known as max-k
problem [9, 21, 20]. We emphasize that extreme bandits are poles apart from classical bandits, where
the objective is to maximize the sum of observations [3]. An effective algorithm for the classical
bandit setting should focus on the source with the highest mean, while an effective algorithm for the
extreme bandit problem should focus on the source with the heaviest tail. It is often the case that
a heavy-tailed source has a small mean, which implies that the classical bandit algorithms perform
poorly for the extreme bandit problem.
The challenging part of our work dwells in the active sampling strategy to detect the heaviest tail
under the limited bandit feedback. We proffer E XTREME H UNTER, a theoretically founded algorithm, that sequentially allocates the resources in an efficient way, for which we prove performance
guarantees. Our algorithm is efficient under a mild semi-parametric assumption common in extreme value theory, while known results by [9, 21, 20] for the extreme bandit problem only hold in
a parametric setting (see Section 4 for a detailed comparison).
2
Learning model for extreme bandits
In this section, we formalize the active (bandit) setting and characterize the measure of performance
for any algorithm ?. The learning setting is defined as follows. Every time step, each of the K arms
(sources) emits a sample Xk,t ? Pk , unknown to the learner. The precise characteristics of Pk are
defined in Section 3. The learner ? then chooses some arm It and then receives only the sample
XIt ,t . The performance of ? is evaluated by the most extreme value found and compared to the
most extreme value possible. We define the reward of a learner ? as:
G?n = max XIt ,t
t?n
The optimal oracle strategy is the one that chooses at each time the arm with the highest potential
revealing the highest value, i.e., the arm ? with the heaviest tail. Its expected reward is then:
?
E [Gn ] = max E max Xk,t
k?K
t?n
The goal of learner ? is to get as close as possible to the optimal oracle strategy. In other words, the
aim of ? is to minimize the expected extreme regret:
Definition 1. The extreme regret in the bandit setting is defined as:
?
?
?
E [Rn ] = E [Gn ] ? E [Gn ] = max E max Xk,t ? E max XIt ,t
k?K
3
t?n
t?n
Heavy-tailed distributions
In this section, we formally define our observation model. Let X1 , . . . , Xn be n i.i.d. observations
from a distribution P . The behavior of the statistic maxi?n Xi is studied by extreme value theory.
One of the main results is the Fisher-Tippett-Gnedenko theorem [11, 12] that characterizes the limiting distribution of this maximum as n converges to infinity. Specifically, it proves that a rescaled
version of this maximum converges to one of the three possible distributions: Gumbel, Fr?echet, or
Weibull. This rescaling factor depends on n. To be concise, we write ?maxi?n Xi converges to a
distribution? to refer to the convergence of the rescaled version to a given distribution. The Gumbel distribution corresponds to the limiting distribution of the maximum of ?not too heavy tailed?
distributions, such as sub-Gaussian or sub-exponential distributions. The Weibull distribution coincides with the behaviour of the maximum of some specific bounded random variables. Finally,
the Fr?echet distribution corresponds to the limiting distribution of the maximum of heavy-tailed
random variables. As many interesting problems concern heavy-tailed distributions, we focus on
Fr?echet distributions in this work. The distribution function of a Fr?echet random variable is defined
for x ? m, and for two parameters ?, s as:
?
P (x) = exp ? x?m
.
s
2
In this work, we consider positive distributions P : [0, ?) ? [0, 1]. For ? > 0, the FisherTippett-Gnedenko theorem also states that the statement ?P converges to an ?-Fr?echet distribution?
is equivalent to the statement ?1 ? P is a ?? regularly varying function in the tail?. These statements
are slightly less restrictive than the definition of approximately ?-Pareto distributions1 , i.e., that there
exists C such that P verifies:
|1 ? P (x) ? Cx?? |
lim
= 0,
(1)
x??
x??
or equivalently that P (x) = 1 ? Cx?? + o(x?? ). If and only if 1 ? P is ?? regularly varying in
the tail, then the limiting distribution of maxi Xi is an ?-Fr?echet distribution. The assumption of
?? regularly varying in the tail is thus the weakest possible assumption that ensures that the (properly rescaled) maximum of samples emitted by a heavy tailed distributions has a limit. Therefore,
the very related assumption of approximate Pareto is almost minimal, but it is (provably) still not
restrictive enough to ensure a convergence rate. For this reason, it is natural to introduce an assumption that is slightly stronger than (1). In particular, we assume, as it is common in the extreme value
literature, a second order Pareto condition also known as the Hall condition [13].
Definition 2. A distribution P is (?, ?, C, C 0 )-second order Pareto (?, ?, C, C 0 > 0) if for x ? 0:
1 ? P (x) ? Cx?? ? C 0 x??(1+?)
By this definition, P (x) = 1 ? Cx?? + O x??(1+?) , which is stronger than the assumption
P (x) = 1 ? Cx?? + o(x?? ), but similar for small ?.
Remark 1. In the definition above, ? defines the rate of the convergence (when x diverges to infinity)
of the tail of P to the tail of a Pareto distribution 1 ? Cx?? . The parameter ? characterizes the
heaviness of the tail: The smaller the ?, the heavier the tail. In the reminder of the paper, we will be
therefore concerned with learning the ? and identifying the smallest one among the sources.
4
Related work
There is a vast body of research in offline anomaly detection which looks for examples that deviate
from the rest of the data, or that are not expected from some underlying model. A comprehensive
review of many anomaly detection approaches can be found in [16] or [8]. There has been also some
work in active learning for anomaly detection [1], which uses a reduction to classification. In online
anomaly detection, most of the research focuses on studying the setting where a set of variables is
monitored. A typical example is the monitoring of cold relief medications, where we are interested
in detecting an outbreak [17]. Similarly to our focus, these approaches do not look for outliers in a
broad sense but rather for the unusual burst of events [22].
In the extreme values settings above, it is often assumed, that we have full information about each
variable. This is in contrast to the limited feedback or a bandit setting that we study in our work.
There has been recently some interest in bandit algorithms for heavy-tailed distributions [4]. However the goal of [4] is radically different from ours as they maximize the sum of rewards and not
the maximal reward. Bandit algorithms have been already used for network intrusion detection [15],
but they typically consider classical or restless setting. [9, 21, 20] were the first to consider the
extreme bandits problem, where our setting is defined as the max-k problem. [21] and [9] consider a fully parametric setting. The reward distributions are assumed to be exactly generalized
extreme value distributions. Specifically, [21] assumes that the distributions are exactly Gumbel,
P (x) = exp(?(x ? m)/s)), and [9], that the distributions are exactly of Gumbel or Fr?echet
P (x) = exp(?(x ? m)? /(s?))). Provided that these assumptions hold, they propose an algorithm for which the regret is asymptotically negligible when compared to the optimal oracle reward.
These results are interesting since they are the first for extreme bandits, but their parametric assumption is unlikely to hold in practice and the asymptotic nature of their bounds limits their impact.
Interestingly, the objective of [20] is to remove the parametric assumptions of [21, 9] by offering
the T HRESHOLDA SCENT algorithm. However, no analysis of this algorithm for extreme bandits is
provided. Nonetheless, to the best of our knowledge, this is the closest competitor for E XTREME H UNTER and we empirically compare our algorithm to T HRESHOLDA SCENT in Section 7.
1
We recall the definition of the standard Pareto distribution as a distribution P , where for some constants ?
and C, we have that for x ? C 1/? , P = 1 ? Cx?? .
3
In this paper we also target the extreme bandit setting, but contrary to [9, 21, 20], we only make a
semi-parametric assumption on the distribution; the second order Pareto assumption (Definition 2),
which is standard in extreme value theory (see e.g., [13, 10]). This is light-years better and significantly weaker than the parametric assumptions made in the prior works for extreme bandits.
Furthermore, we provide a finite-time regret bound for our more general semi-parametric setting
(Theorem 2), while the prior works only offer asymptotic results. In particular, we provide an upper bound on the rate at which the regret becomes negligible when compared to the optimal oracle
reward (Definition 1).
5
Extreme Hunter
In this section, we present our main results. In particular, we present the algorithm and the main
theorem that bounds its extreme regret. Before that, we first provide an initial result on the expectation of the maximum of second order Pareto random variables which will set the benchmark for the
oracle regret. We first characterize the expectation of the maximum of second order Pareto distributions. The following lemma states that the expectation of the maximum of i.i.d. second order Pareto
samples is equal, up to a negligible term, to the expectation of the maximum of i.i.d. Pareto samples.
This result is crucial for assessing the benchmark for the regret, in particular the expected value of
the maximal oracle sample. Theorem 1 is based on Lemma 3, both provided in the appendix.
Theorem 1. Let X1 , . . . , Xn be n i.i.d. samples drawn according to (?, ?, C, C 0 )-second order
Pareto distribution P (see Definition 2). If ? > 1, then:
2C 0 D
1/?
1/?
2
(nC)1/? + B = o (nC)
,
+ C ?+1?+1
E(max Xi ) ? (nC)1/? ? 1? ?1 ? 4D
n (nC)
n?
i
where D2 , D1+? > 0 are some universal constants, and B is defined in the appendix (9).
Theorem 1 implies that the optimal strategy in hindsight attains the following expected reward:
h
i
1/?
E [G?n ] ? max (Ck n) k ? 1? ?1
k
Our objective is therefore to find a learner ? Algorithm 1 E XTREME H UNTER
such that E [G?n ] ? E [G?n ] is negligible when
Input:
compared to E[G?n ], i.e., when compared to
K: number of arms
?
?
(nC ? )1/? ? 1? ?1? ? n1/? where ? is the
n: time horizon
optimal arm.
b: where b ? ?k for all k ? K
N : minimum number of pulls of each arm
From the discussion above, we know that the
Initialize:
minimization of the extreme regret is linked
Tk ? 0 for all k ? K
with the identification of the arm with the heav? ? exp(? log2 n)/(2nK)
iest tail. Our E XTREME H UNTER algorithm is
Run:
based on a classical idea in bandit theory: opfor t = 1 to n do
timism in the face of uncertainty. Our stratfor k = 1 to K do
egy is to estimate E [maxt?n Xk,t ] for any k
if Tk ? N then
and to pull the arm which maximizes its upBk,t ? ?
per bound. From Definition 2, the estimation
else
of this quantity relies heavily on an efficient esestimate b
hk,t that verifies (2)
timation of ?k and Ck , and on associated confibk,t using (3)
estimate C
dence widths. This topic is a classic problem in
update Bk,t using (5) with (2) and (4)
extreme value theory, and such estimators exist
end if
provided that one knows a lower bound b on ?k
end for
[10, 6, 7]. From now on we assume that a conPlay arm kt ? arg maxk Bk,t
stant b > 0 such that b ? mink ?k is known
Tkt ? Tkt + 1
to the learner. As we argue in Remark 2, this
end for
assumption is necessary .
Since our main theoretical result is a finite-time upper bound, in the following exposition we carefully describe all the constants and stress what quantities they depend on. Let Tk,t be the number of
samples drawn from arm k at time t. Define ? = exp(? log2 n)/(2nK) and consider an estimator
4
b
hk,t of 1/?k at time t that verifies the following condition with probability 1 ? ?, for Tk,t larger than
some constant N2 that depends only on ?k , Ck , C 0 and b:
p
1
?b/(2b+1)
hk,t ? D log(1/?)Tk,t
= B1 (Tk,t ),
(2)
?k ? b
where D is a constant that also depends only on ?k , Ck , C 0 , and b. For instance, the estimator
in [6] (Theorem 3.7) verifies this property and provides D and N2 but other estimators are possible.
Consider the associated estimator for Ck :
?
?
Tk,t n
o
X
b
1
h /(2b+1) ?
bk,t = T 1/(2b+1) ?
C
1 Xk,u ? Tk,tk,t
(3)
k,t
Tk,t u=1
For this estimator, we know [7] with probability 1 ? ? that for Tk,t ? N2 :
q
bk,t ? E log(Tk,t /?) log(Tk,t )T ?b/(2b+1) = B2 (Tk,t ),
Ck ? C
k,T
(4)
where E is derived in [7] in the proof of Theorem 2. Let N = max A log(n)2(2b+1)/b , N2 where
A depends on (?k , Ck )k , b, D, E, and C 0 , and is such that:
(2b+1)/b
?
2D log(n)2
max (2B1 (N ), 2B2 (N )/Ck ) ? 1, N ? (2D log2 n)(2b+1)/b , and N > 1?maxk 1/?k
This inspires Algorithm 1, which first pulls each arm N times and then, at each time t > KN , pulls
the arm that maximizes Bk,t , which we define as:
bhk,t +B1 (Tk,t )
? b
bk,t + B2 (Tk,t ) n
?
hk,t , B1 (Tk,t ) ,
(5)
C
? y) = ?(1
? ? x ? y), where we set ?
? = ? for any x > 0 and +? otherwise.
where ?(x,
Remark 2. A natural question is whether it is possible to learn ?k as well. In fact, this is not possible
for this model and a negative result was proved by [7]. The result states that in this setting it is not
possible to test between two fixed values of ? uniformly over the set of distributions. Thereupon, we
define b as a lower bound for all ?k . With regards to the Pareto distribution, ? = ? corresponds to
the exact Pareto distribution, while ? = 0 for such distribution that is not (asymptotically) Pareto.
We show that this algorithm meets the desired properties. The following theorem states our main
result by upper-bounding the extreme regret of E XTREME H UNTER.
Theorem 2. Assume that the distributions of the arms are respectively (?k , ?k , Ck , C 0 ) second
order Pareto (see Definition 2) with mink ?k > 1. If n ? Q, the expected extreme regret of E X TREME H UNTER is bounded from above as:
?
(2b+1)/b
? log(n)(1?1/?? )
?b/((b+1)?? )
E [Rn ] ? L(nC ? )1/? K
log(n)
+
n
+
n
= E [G?n ] o(1),
n
where L, Q > 0 are some constants depending only on (?k , Ck )k , C 0 , and b (Section 6).
Theorem 2 states that the E XTREME H UNTER strategy performs almost as well as the best (oracle)
strategy, up to a term that is negligible when compared to the performance of the oracle strategy.
?
Indeed, the regret is negligible when compared to (nC ? )1/? , which is the order of magnitude of the
performance of the best oracle strategy E [G?n ] = maxk?K E [maxt?n Xk,t ]. Our algorithm thus
detects the arm that has the heaviest tail.
For n large enough (as a function of (?k , ?k , Ck )k , C 0 and K), the two first terms in the regret
become negligible when compared to the third one, and the regret is then bounded as:
?
E [Rn ] ? E [G?n ] O n?b/((b+1)? )
We make two observations: First, the larger the b, the tighter this bound is, since the model is then
closer to the parametric case. Second, smaller ?? also tightens the bound, since the best arm is then
very heavy tailed and much easier to recognize.
5
6
Analysis
In this section, we prove an upper bound on the extreme regret of Algorithm 1 stated in Theorem 2.
Before providing the detailed proof, we give a high-level overview and the intuitions.
In Step 1, we define the (favorable) high probability event ? of interest, useful for analyzing the
mechanism of the bandit algorithm. In Step 2, given ?, we bound the estimates of ?k and Ck , and
use them to bound the main upper confidence bound. In Step 3, we upper-bound the number of pulls
of each suboptimal arm: we prove that with high probability we do not pull them too often. This
enables us to guarantee that the number of pulls of the optimal arms ? is on ? equal to n up to a
negligible term.
The final Step 4 of the proof is concerned with using this lower bound on the number of pulls of
the optimal arm in order to lower bound the expectation of the maximum of the collected samples.
Such step is typically straightforward in the classical (mean-optimizing) bandits by the linearity of
the expectation. It is not straightforward in our setting. We therefore prove Lemma 2, in which we
show that the expected value of the maximum of the samples in the favorable event ? will be not too
far away from the one that we obtain without conditioning on ?.
Step 1: High probability event. In this step, we define the favorable event ?. We set
def
? = exp(? log2 n)/(2nK) and consider the event ? such that for any k ? K, N ? T ? n:
p
1
? k (T ) ? D log(1/?)T ?b/(2b+1) ,
?k ? h
p
Ck ? C?k (T ) ? E log(T /?)T ?b/(2b+1) ,
? k (T ) and C?k (T ) are the estimates of 1/?k and Ck respectively using the first T samples.
where h
bk,t which are the estimates of the same quantities at time
Notice, they are not the same as b
hk,t and C
t for the algorithm, and thus with Tk,t samples. The probability of ? is larger than 1 ? 2nK? by a
union bound on (2) and (4).
Step 2: Bound on Bk,t . The following lemma holds on ? for upper- and lower-bounding Bk,t .
Lemma 1. (proved in the appendix) On ?, we have that for any k ? K, and for Tk,t ? N :
p
1
1
?b/(2b+1)
(Ck n) ?k ? 1? ?1k ? Bk,t ? (Ck n) ?k ? 1? ?1k
1 + F log(n) log(n/?)Tk,t
(6)
Step 3: Upper bound on the number of pulls of a suboptimal arm. We proceed by using the
bounds on Bk,t from the previous step to upper-bound the number of suboptimal pulls. Let ? be the
best arm. Assume that at round t, some arm k 6= ? is pulled. Then by definition of the algorithm
B?,t ? Bk,t , which implies by Lemma 1:
p
?b/(2b+1)
1/??
1/?
(C ? n)
? 1? ?1? ? (Ck n) k ? 1? ?1k
1 + F log(n) log(n/?)Tk,t
Rearranging the terms we get:
1/??
(C ? n)
(Ck n)
1/?k
? 1? ?1?
? 1? ?1k
p
?b/(2b+1)
? 1 + F log(n) log(n/?)Tk,t
(7)
We now define ?k which is analogous to the gap in the classical bandits:
1/??
(C ? n)
? 1? ?1?
?k =
?1
1/?
(Ck n) k ? 1? ?1k
Since Tk,t ? n, (7) implies for some problem dependent constants G and G0 dependent only on
(?k , Ck )k , C 0 and b, but independent of ? that:
2
(2b+1)/(2b)
(2b+1)(2b)
log(n/?)
Tk,t ? N + G0 log n ?
? N + G log2 n log(n/?)
2
k
6
This implies that number T ? of pulls of arm ? is with probability 1 ? ? 0 , at least
X
(2b+1)/(2b)
n?
G log2 n log(2nK/? 0 )
? KN,
k6=?
where ? 0 = 2nK?. Since n is larger than
Q ? 2KN + 2GK log2 n log (2nK/? 0 )
we have that T ? ?
n
2
(2b+1)/(2b)
,
as a corollary.
Step 4: Bound on the expectation. We start by lower-bounding the expected gain:
E[Gn ] = E max XIt ,Tk,t ? E max XIt ,Tk,t 1{?} ? E max X?,T?,t 1{?} = E max? Xi 1{?}
t?n
t?n
t?n
i?T
The next lemma links the expectation of maxt?T ? X?,t with the expectation of maxt?T ? X?,t 1{?}.
Lemma 2. (proved in the appendix) Let X1 , . . . , XT be i.i.d. samples from an (?, ?, C, C 0 )-second
order Pareto distribution F . Let ? 0 be an event of probability larger than 1 ? ?. Then for ? < 1/2
and for T ? Q large enough so that c max 1/T, 1/T ? ? 1/4 for a given constant c > 0, that
1/?
depends only on C, C 0 and ?, and also for T ? log(2) max C (2C 0 )
, 8 log (2) :
1/? 1?1/?
1/?
8
(T C)
?
E max Xt 1{?} ? (T C)
? 1? ?1 ? 4 + ??1
t?T
2C 0 D1+?
1/?
1/?
2
? 2 4D
(T
C)
+
(T
C)
+
B
.
1+?
?
T
C
T
Since n is large enough so that 2n2 K? 0 = 2n2 K exp ? log2 n ? 1/2, where ? 0 = exp ? log2 n ,
and the probability of ? is larger than 1 ? ? 0 , we can use Lemma 2 for the optimal arm:
01? 1 8D2
1
0
Dmax
8
2B
? ?? ? T ? ? (C4C
,
E max? X?,t 1{?} ? (T ? C ? ) ?? ? 1? ?1? ? 4+ ??1
?
1
? )1+b (T ? )b
?
? ?
t?T
(T C ) ?
def
where Dmax = maxi D1+?i . Using Step 3, we bound the above with a function of n. In particular,
?
we lower-bound the last three terms in the brackets using T ? ? n2 and the (T ? C ? )1/? factor as:
?
(T ? C ? )1/? ? (nC ? )1/?
?
1?
GK
n
2b+1
log(2n2 K/? 0 ) 2b ? KN
n
We are now ready to relate the lower bound on the gain of E XTREME H UNTER with the upper bound
of the gain of the optimal policy (Theorem 1), which brings us the upper bound for the regret:
E [Rn ] = E [G?n ] ? E [Gn ] ? E [G?n ] ? E max? Xi ? E [G?n ] ? E max? X?,t 1{?}
i?T
t?T
2b+1
?
2
0
01?1/??
KN
B
2b
,
? H(nC ? )1/? n1 + (nC1? )b + GK
log(2n
K/?
)
+
+
?
+
?
n
n
(nC ? )1/?
where H is a constant that depends on (?k , Ck )k , C 0 , and b. To bound the last term, we use the
?
?
?
?
definition of B (9) to get the n?? /((? +1)? ) term, upper-bounded by n?b/((b+1)? ) as b ? ? ? .
?1
?b
?
Notice that this final term also eats up n and n terms since b/((b + 1)? ) ? min(1, b).
We finish by using ? 0 = exp ? log2 n and grouping the problem-dependent constants into L to get
the final upper bound:
?
?
?
(2b+1)/b
E [Rn ] ? L(nC ? )1/? K
+ n? log(n)(1?1/? ) + n?b/((b+1)? )
n log(n)
7
Comparison of extreme bandit strategies (K=3)
Comparison of extreme bandit strategies on the network data K=5
250
ExtremeHunter
UCB
ThresholdAscent
ExtremeHunter
UCB
ThresholdAscent
9000
Comparison of extreme bandit strategies (K=3)
2500
10000
ExtremeHunter
UCB
ThresholdAscent
200
2000
8000
6000
5000
4000
extreme regret
extreme regret
extreme regret
7000
1500
1000
150
100
3000
2000
50
500
1000
0
0
1000
2000
3000
4000
5000
6000
time t
7000
8000
9000
10000
0
0
1000
2000
3000
4000
5000
time t
6000
7000
8000
9000
10000
0
0
1000
2000
3000
4000
5000
6000
7000
8000
9000
10000
time t
Figure 1: Extreme regret as a function of time for the exact Pareto distributions (left), approximate
Pareto (middle) distributions, and the network traffic data (right).
7
Experiments
In this section, we empirically evaluate E XTREME H UNTER on synthetic and real-world data. The
measure of our evaluation is the extreme regret from Definition 1. Notice that even thought we
evaluate the regret as a function of time T , the extreme regret is not cumulative and it is more in the
spirit of simple regret [5]. We compare our E XTREME H UNTER with T HRESHOLDA SCENT [20].
Moreover, we also compare to classical U CB [3], as an example of the algorithm that aims for the
arm with the highest mean as opposed to the heaviest tail. When the distribution of a single arm
has both the highest mean and the heaviest-tail, both E XTREME H UNTER and U CB are expected to
perform the same with respect to the extreme regret. In the light of Remark 2, we set b = 1 to
consider a wide class of distributions.
Exact Pareto Distributions In the first experiment, we consider K = 3 arms with the distributions
Pk (x) = 1?x??k , where ? = [5, 1.1, 2]. Therefore, the most heavy-tailed distribution is associated
with the arm k = 2. Figure 1 (left) displays the averaged result of 1000 simulations with the time
horizon T = 104 . We observe that E XTREME H UNTER eventually keeps allocating most of the
pulls to the arm of the interest. Since in this case, the arm with the heaviest tail is also the arm
with the largest mean, U CB also performs well and it is even able to detect the best arm earlier.
T HRESHOLDA SCENT, on the other way, was not always able to allocate the pulls properly in 104
steps. This may be due to the discretization of the rewards that this algorithm is using.
Approximate Pareto Distributions For the exact Pareto distributions, the smaller the tail index
the higher the mean and even UCB obtains a good performance. However, this is no longer necessarily the case for the approximate Pareto distributions. For this purpose, we perform the second
experiment where we mix an exact Pareto distribution with a Dirac distribution in 0. We consider
K = 3 arms. Two of the arms follow the exact Pareto distributions with ?1 = 1.5 and ?3 = 3.
On the other hand, the second arm has a mixture weight of 0.2 for the exact Pareto distribution with
?2 = 1.1 and 0.8 mixture weight of the Dirac distribution in 0. For this setting, the second arm
is the most heavy-tailed but the first arms has the largest mean. Figure 1 (middle) shows the result. We see that U CB performs worse since it eventually focuses on the arm with the largest mean.
T HRESHOLDA SCENT performs better than U CB but not as good as E XTREME H UNTER.
Computer Network Traffic Data In this experiment, we evaluate E XTREME H UNTER on heavytailed network traffic data which was collected from user laptops in the enterprise environment [2].
The objective is to allocate the sampling capacity among the computer nodes (arms), in order to find
the largest outbursts of the network activity. This information then serves an IT department to further
investigate the source of the extreme network traffic. For each arm, a sample at the time t corresponds to the number of network activity events for 4 consecutive seconds. Specifically, the network
events are the starting times of packet flows. In this experiment, we selected K = 5 laptops (arms),
where the recorded sequences were long enough. Figure 1 (right) shows that E XTREME H UNTER
again outperforms both T HRESHOLDA SCENT and U CB.
Acknowledgements We would like to thank John Mark Agosta and Jennifer Healey for the network traffic data. The research presented in this paper was supported by Intel Corporation, by
French Ministry of Higher Education and Research, and by European Community?s Seventh Framework Programme (FP7/2007-2013) under grant agreement no 270327 (CompLACS).
8
References
[1] Naoki Abe, Bianca Zadrozny, and John Langford. Outlier Detection by Active Learning. In
Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and
Data Mining, pages 504?509, 2006.
[2] John Mark Agosta, Jaideep Chandrashekar, Mark Crovella, Nina Taft, and Daniel Ting. Mixture models of endhost network traffic. In IEEE Proceedings of INFOCOM,, pages 225?229.
[3] Peter Auer, Nicol`o Cesa-Bianchi, and Paul Fischer. Finite-time Analysis of the Multiarmed
Bandit Problem. Machine Learning, 47(2-3):235?256, 2002.
[4] S?ebastien Bubeck, Nicol`o Cesa-Bianchi, and G?abor Lugosi. Bandits With Heavy Tail. Information Theory, IEEE Transactions on, 59(11):7711?7717, 2013.
[5] S?ebastien Bubeck, R?emi Munos, and Gilles Stoltz. Pure Exploration in Multi-armed Bandits
Problems. Algorithmic Learning Theory, pages 23?37, 2009.
[6] Alexandra Carpentier and Arlene K. H. Kim. Adaptive and minimax optimal estimation of the
tail coefficient. Statistica Sinica, 2014.
[7] Alexandra Carpentier and Arlene K. H. Kim. Honest and adaptive confidence interval for the
tail coefficient in the Pareto model. Electronic Journal of Statistics, 2014.
[8] Varun Chandola, Arindam Banerjee, and Vipin Kumar. Anomaly detection: A survey. ACM
Comput. Surv., 41(3):15:1?15:58, July 2009.
[9] Vincent A. Cicirello and Stephen F. Smith. The max k-armed bandit: A new model of exploration applied to search heuristic selection. AAAI Conference on Artificial Intelligence, 2005.
[10] Laurens de Haan and Ana Ferreira. Extreme Value Theory: An Introduction. Springer Series
in Operations Research and Financial Engineering. Springer, 2006.
[11] Ronald Aylmer Fisher and Leonard Henry Caleb Tippett. Limiting forms of the frequency
distribution of the largest or smallest member of a sample. Mathematical Proceedings of the
Cambridge Philosophical Society, 24:180, 1928.
[12] Boris Gnedenko. Sur la distribution limite du terme maximum d?une s?erie al?eatoire. The
Annals of Mathematics, 44(3):423?453, 1943.
[13] Peter Hall and Alan H. Welsh. Best Attainable Rates of Convergence for Estimates of Parameters of Regular Variation. The Annals of Statistics, 12(3):1079?1084, 1984.
[14] Tze L. Lai and Herbert Robbins. Asymptotically efficient adaptive allocation rules. Advances
in Applied Mathematics, 6(1):4?22, 1985.
[15] Keqin Liu and Qing Zhao. Dynamic Intrusion Detection in Resource-Constrained Cyber Networks. In IEEE International Symposium on Information Theory Proceedings, 2012.
[16] Markos Markou and Sameer Singh. Novelty detection: a review, part 1: statistical approaches.
Signal Process., 83(12):2481?2497, 2003.
[17] Daniel B. Neill and Gregory F. Cooper. A multivariate Bayesian scan statistic for early event
detection and characterization. Machine Learning, 79:261?282, 2010.
[18] Carey E. Priebe, John M. Conroy, David J. Marchette, and Youngser Park. Scan Statistics on
Enron Graphs. In Computational and Mathematical Organization Theory, volume 11, pages
229?247, 2005.
[19] Ingo Steinwart, Don Hush, and Clint Scovel. A Classification Framework for Anomaly Detection. Journal of Machine Learning Research, 6:211?232, 2005.
[20] Matthew J. Streeter and Stephen F. Smith. A Simple Distribution-Free Approach to the Max
k-Armed Bandit Problem. In Principles and Practice of Constraint Programming, volume
4204, pages 560?574, 2006.
[21] Matthew J. Streeter and Stephen F. Smith. An Asymptotically Optimal Algorithm for the Max
k-Armed Bandit Problem. In AAAI Conference on Artificial Intelligence Intelligence, pages
135?142, 2006.
[22] Ryan Turner, Zoubin Ghahramani, and Steven Bottone. Fast online anomaly detection using
scan statistics. IEEE Workshop on Machine Learning for Signal Processing, 2010.
9
| 5379 |@word mild:1 version:2 middle:2 stronger:2 d2:2 seek:1 simulation:1 attainable:1 concise:1 reduction:1 initial:1 liu:1 series:1 selecting:1 daniel:2 offering:1 ours:1 interestingly:1 outperforms:1 scovel:1 contextual:1 michal:2 discretization:1 must:1 john:4 ronald:1 enables:1 remove:1 update:1 intelligence:3 selected:1 une:1 xk:6 smith:3 detecting:2 characterization:2 node:2 provides:1 mathematical:2 burst:1 enterprise:1 become:1 symposium:1 prove:4 scent:6 introduce:1 theoretically:2 indeed:1 expected:9 behavior:1 multi:1 detects:1 armed:4 becomes:1 provided:4 underlying:2 bounded:4 maximizes:2 linearity:1 moreover:1 laptop:2 what:1 cm:1 kind:1 weibull:2 astronomy:1 hindsight:1 corporation:1 guarantee:2 every:1 exactly:3 ferreira:1 uk:2 grant:1 positive:1 negligible:8 before:2 naoki:1 engineering:1 limit:2 consequence:1 analyzing:1 meet:1 jaideep:1 clint:1 approximately:1 lugosi:1 inria:2 studied:3 challenging:2 limited:4 averaged:1 practice:2 regret:27 union:1 cold:1 area:1 universal:1 significantly:1 revealing:1 thought:1 word:1 confidence:2 regular:1 zoubin:1 get:4 agosta:2 close:1 selection:1 risk:1 dwells:1 equivalent:1 straightforward:2 starting:1 survey:2 identifying:1 pure:1 estimator:6 rule:1 pull:13 financial:1 classic:1 variation:2 analogous:1 limiting:5 annals:2 target:1 heavily:1 user:1 anomaly:21 gnedenko:3 exact:7 us:1 programming:1 agreement:1 surv:1 steven:1 nc1:1 ensures:1 highest:5 rescaled:3 intuition:1 environment:1 reward:10 cam:1 dynamic:1 depend:1 singh:1 efficiency:1 learner:7 fast:1 effective:2 describe:1 artificial:2 heuristic:2 spend:1 larger:6 otherwise:1 statistic:6 fischer:1 final:3 online:2 sequence:1 propose:2 maximal:2 fr:8 poorly:1 dirac:2 convergence:4 diverges:1 assessing:1 boris:1 converges:4 tk:26 depending:1 ac:1 come:1 implies:5 laurens:1 exploration:2 packet:1 ana:1 education:1 behaviour:1 taft:1 tighter:1 ryan:1 hold:4 hall:2 exp:9 cb:6 algorithmic:1 matthew:2 consecutive:1 smallest:2 early:1 heavytailed:1 purpose:1 estimation:2 favorable:3 robbins:1 largest:5 minimization:1 eats:1 gaussian:1 always:1 aim:2 rather:1 ck:21 varying:3 corollary:1 derived:1 focus:7 xit:5 properly:2 intrusion:4 contrast:1 sigkdd:1 medication:1 attains:1 detect:4 sense:1 kim:2 hk:5 aylmer:1 dependent:3 typically:3 unlikely:1 abor:1 bandit:36 france:1 interested:2 provably:1 arg:1 among:2 classification:2 k6:1 constrained:1 initialize:1 field:1 once:1 equal:2 sampling:5 lille:1 look:2 broad:1 park:1 recognize:1 comprehensive:1 qing:1 relief:1 welsh:1 n1:2 detection:21 organization:1 interest:3 investigate:1 mining:1 evaluation:1 analyzed:1 extreme:51 bracket:1 mixture:3 light:2 heaviness:1 kt:1 allocating:1 crovella:1 closer:1 necessary:1 unter:16 allocates:1 stoltz:1 desired:1 theoretical:2 minimal:2 instance:2 earlier:1 gn:5 pole:1 rare:3 seventh:1 inspires:1 too:3 characterize:2 kn:5 gregory:1 synthetic:2 chooses:3 epidemiology:1 peak:1 international:2 sequel:1 complacs:1 again:1 aaai:2 heaviest:9 recorded:1 opposed:1 cesa:2 worse:1 creating:1 zhao:1 rescaling:1 potential:1 de:1 b2:3 waste:1 healey:1 coefficient:2 chandola:1 depends:6 infocom:1 linked:1 traffic:7 characterizes:2 start:1 carey:1 minimize:1 characteristic:2 efficiently:1 correspond:1 identification:1 vincent:1 bayesian:1 hunter:1 monitoring:1 definition:16 competitor:1 markou:1 nonetheless:2 echet:7 frequency:1 hydrology:1 associated:3 proof:3 monitored:1 emits:1 gain:3 proved:3 recall:1 lim:1 reminder:1 knowledge:2 formalize:1 carefully:2 auer:1 higher:3 varun:1 follow:1 evaluated:1 furthermore:1 langford:1 hand:1 steinwart:2 receives:2 banerjee:1 french:1 defines:1 brings:1 alexandra:3 concept:1 laboratory:1 statslab:1 round:1 width:1 ambiguous:1 vipin:1 coincides:1 generalized:1 stress:1 performs:4 arindam:1 recently:1 common:3 empirically:3 overview:1 tightens:1 conditioning:1 volume:2 tail:22 refer:1 multiarmed:1 cambridge:2 seldom:1 mathematics:2 similarly:1 henry:1 europe:1 longer:1 marchette:1 etc:1 closest:1 multivariate:1 optimizing:1 apart:1 life:2 herbert:1 minimum:1 ministry:1 novelty:1 maximize:2 signal:2 july:1 semi:3 full:1 mix:1 stephen:3 sameer:1 alan:1 characterized:1 offer:1 long:1 lai:1 tippett:2 impact:1 expectation:9 stant:1 want:2 interval:1 else:1 source:18 crucial:1 rest:1 enron:1 tend:1 cyber:1 member:1 regularly:3 flow:2 contrary:1 spirit:1 call:1 emitted:3 structural:1 distributions1:1 enough:5 concerned:2 finish:1 identified:1 suboptimal:3 idea:1 honest:1 whether:1 motivated:1 heavier:1 allocate:5 peter:2 proceed:1 remark:4 useful:1 detailed:2 exist:3 notice:3 per:1 write:1 drawn:2 carpentier:4 vast:1 asymptotically:4 graph:1 sum:2 year:1 run:1 uncertainty:1 almost:2 decide:1 electronic:1 appendix:4 abnormal:2 bound:30 def:2 followed:1 display:1 neill:1 oracle:10 activity:2 infinity:2 constraint:1 dence:1 emi:1 min:1 kumar:1 department:1 according:2 erie:1 smaller:3 slightly:2 character:1 outbreak:1 outlier:3 tkt:2 resource:5 jennifer:1 dmax:2 eventually:2 mechanism:1 know:4 fp7:1 end:3 unusual:1 serf:1 studying:1 operation:1 observe:3 away:2 assumes:1 cf:1 ensure:1 log2:10 medicine:2 restrictive:2 ting:1 prof:1 ghahramani:1 classical:8 society:1 seeking:1 objective:6 g0:2 already:1 quantity:3 question:1 strategy:12 costly:1 parametric:9 distance:1 link:1 thank:1 capacity:1 majority:1 topic:1 argue:1 collected:2 water:1 reason:1 nina:1 sur:1 index:1 providing:1 equivalently:1 nc:11 sinica:1 statement:3 relate:1 nord:1 gk:3 mink:2 negative:1 stated:1 priebe:1 design:2 ebastien:2 policy:2 unknown:1 perform:3 bianchi:2 upper:13 gilles:1 observation:5 benchmark:2 finite:3 ingo:1 zadrozny:1 maxk:3 team:1 precise:1 rn:5 community:1 abe:1 bk:12 david:1 optimized:1 connection:1 security:2 philosophical:1 conroy:1 hush:1 able:2 bottone:1 max:25 event:14 natural:2 valko:2 turner:1 arm:41 minimax:1 ready:1 deviate:1 review:2 literature:1 prior:2 acknowledgement:1 discovery:1 nicol:2 asymptotic:2 fully:1 interesting:2 allocation:2 principle:1 pareto:27 heavy:12 maxt:4 supported:1 last:2 free:1 offline:1 weaker:1 pulled:1 wide:1 face:1 munos:1 limite:1 regard:1 feedback:4 xn:2 world:2 cumulative:1 commonly:1 collection:1 adaptive:4 made:1 founded:1 far:2 programme:1 transaction:1 approximate:4 emphasize:1 obtains:1 keep:1 sequentially:3 active:5 b1:4 assumed:2 xi:6 don:1 search:3 compromised:1 timation:1 tailed:11 streeter:2 nature:1 learn:1 rearranging:1 du:1 necessarily:1 european:1 domain:1 pk:3 main:7 statistica:1 bounding:3 paul:1 n2:8 verifies:4 x1:3 body:1 intel:1 bianca:1 cooper:1 probing:1 sub:2 exponential:1 comput:1 third:1 theorem:14 specific:3 xt:2 maxi:4 concern:1 weakest:1 exists:1 grouping:1 workshop:1 sequential:1 magnitude:1 horizon:2 gumbel:4 restless:1 nk:7 egy:1 easier:1 gap:1 cx:7 tze:1 bubeck:2 springer:2 corresponds:4 radically:1 relies:1 acm:2 goal:3 exposition:1 leonard:1 fisher:2 specifically:4 typical:1 uniformly:1 lemma:9 la:1 ucb:4 formally:1 mark:3 scan:3 outstanding:1 bhk:1 evaluate:4 outgoing:1 d1:3 |
4,837 | 538 | Iterative Construction of
Sparse Polynomial Approximations
Terence D. Sanger
Massachusetts Institute
of Technology
Room E25-534
Cambridge, MA 02139
[email protected]
Richard S. Sutton
GTE Laboratories
Incorporated
40 Sylvan Road
Waltham, MA 02254
[email protected]
Christopher J. Matheus
GTE Laboratories
Incorporated
40 Sylvan Road
Waltham, MA 02254
[email protected]
Abstract
We present an iterative algorithm for nonlinear regression based on construction of sparse polynomials. Polynomials are built sequentially from
lower to higher order. Selection of new terms is accomplished using a novel
look-ahead approach that predicts whether a variable contributes to the
remaining error. The algorithm is based on the tree-growing heuristic in
LMS Trees which we have extended to approximation of arbitrary polynomials of the input features. In addition, we provide a new theoretical
justification for this heuristic approach. The algorithm is shown to discover a known polynomial from samples, and to make accurate estimates
of pixel values in an image-processing task.
1
INTRODUCTION
Linear regression attempts to approximate a target function by a model that is a
linear combination of the input features. Its approximation ability is thus limited
by the available features. We describe a method for adding new features that are
products or powers of existing features. Repeated addition of new features leads
to the construction of a polynomial in the original inputs, as in (Gabor 1961).
Because there is an infinite number of possible product terms, we have developed
a new method for predicting the usefulness of entire classes of features before they
are included. The resulting nonlinear regression will be useful for approximating
functions that can be described by sparse polynomials.
1064
Iterative Construction of Sparse Polynomial Approximations
f
Xn
Figure 1: Network depiction of linear regression on a set of features Xi.
2
THEORY
Let {xdi=l be the set of features already included in a model that attempts to
predict the function f . The output of the model is a linear combination
n
i = LCiXi
i=l
where the Ci'S are coefficients determined using linear regression. The model can
also be depicted as a single-layer network as in figure 1. The approximation error
is e
f - j, and we will attempt to minimize E[e 2 ] where E is the expectation
operator.
=
The algorithm incrementally creates new features that are products of existing
features. At each step, the goal is to select two features xp and Xq already in the
model and create a new feature XpXq (see figure 2). Even if XpXq does not decrease
the approximation error, it is still possible that XpXqXr will decrease it for some
X r . So in order to decide whether to create a new feature that is a product with
x p , the algorithm must "look-ahead" to determine if there exists any polynomial a
in the xi's such that inclusion ofax p would significantly decrease the error. If no
such polynomial exists, then we do not need to consider adding any features that
are products with xp.
=
Define the inner product between two polynomials a and b as (alb)
E[ab] where
the expected value is taken with respect to a probability measure I-" over the (zeroE[a 2 ], and let P be the set of
mean) input values. The induced norm is IIal12
polynomials with finite norm. {P, (?I?)} is then an infinite-dimensional linear vector
space. The Weierstrass approximation theorem proves that P is dense in the set of
all square-integrable functions over 1-", and thus justifies the assumption that any
function of interest can be approximated by a member of P.
=
Assume that the error e is a polynomial in P. In order to test whether
ipates in e for any polynomial a E P, we write
e
=
apxp
+ bp
axp
partic-
1065
1066
Sanger, Sutton, and Matheus
f
Figure 2: Incorporation of a new product term into the model.
where ap and bp are polynomials, and ap is chosen to minimize lIapxp - ell 2
E[( apxp - e )2]. The orthogonality principle then shows that apxp is the projection
of the polynomial e onto the linear subspace of polynomials xpP. Therefore, bp is
orthogonal to xpP, so that E[bpg] = 0 for all g in xpP.
We now write
E[e 2]
= E[a;x;] + 2E[apxpbp] + E[b;] = E[a;x;] + E[b;]
since E[apxpbp] = 0 by orthogonality. If apxp were included in the model, it would
thus reduce E[e 2] by E[a;x;], so we wish to choose xp to maximize E[a;x;]. Unfortunately, we have no dIrect measurement of ap ?
3
METHODS
Although E[a;x;] cannot be measured directly, Sanger (1991) suggests choosing xp
to maximize E[e2x~] instead, which is directly measurable. Moreover, note that
E[e 2x;]
=
E[a;x;] + 2E[apx;bp] + E[x;b;]
=
E[a;x;]
and thus E[e 2x;] is related to the desired but unknown value E[a;x;]. Perhaps
better would be to use
E[e 2x 2]
E[x~] -
~=-::-:p-
E[a 2x4 ]
p
p
E[x~]
which can be thought of as the regression of (a;x~)xp against xp'
More recently, (Sutton and Matheus 1991) suggest using the regression coefficients
of e2 against
for all i as the basis for comparison. The regression coefficients Wi
are called "potentials", and lead to a linear approximation of the squared error:
xr
(1)
Iterative Construction of Sparse Polynomial Approximations
If a new term apxp were included in the model of f, then the squared error would
be b; which is orthogonal to any polynomial in xpP and in particular to x;. Thus
the coefficient of x; in (1) would be zero after inclusion of apxp, and wpE[x;] is an
approximation to the decrease in mean-squared error E[e 2 ] - E[b;] which we can
expect from inclusion of apxp. We thus choose xp by maximizing wpE[x;].
This procedure is a form of look-ahead which allows us to predict the utility of a
high-order term apxp without actually including it in the regression. This is perhaps
most useful when the term is predicted to make only a small contribution for the
optimal a p , because in this case we can drop from consideration any new features
that include xp.
We can choose a different variable Xq similarly, and test the usefulness of incorporating the product XpXq by computing a "joint potential" Wpq which is the regression of
the squared error against the model including a new term x~x~. The joint potential
attempts to predict the magnitude of the term E[a~qx;xi].
We now use this method to choose a single new feature XpXq to include in the model.
For all pairs XiXj such that Xi and Xj individually have high potentials, we perform
a third regression to determine the joint potentials of the product terms XiXj. Any
term with a high joint potential is likely to participate in f. We choose to include the
new term XpXq with the largest joint potential. In the network model, this results in
the construction of a new unit that computes the product of xp and x q, as in figure
2. The new unit is incorporated into the regression, and the resulting error e will
be orthogonal to this unit and all previous units. Iteration of this technique leads
to the successive addition of new regression terms and the successive decrease in
mean-squared error E[e 2 ]. The process stops when the residual mean-squared error
drops below a chosen threshold, and the final model consists of a sparse polynomial
in the original inputs.
We have implemented this algorithm both in a non-iterative version that computes
coefficients and potentials based on a fixed data set, and in an iterative version that
uses the LMS algorithm (Widrow and Hoff 1960) to compute both coefficients and
potentials incrementally in response to continually arriving data. In the iterative
version, new terms are added at fixed intervals and are chosen by maximizing over
the potentials approximated by the LMS algorithm. The growing polynomial is
efficiently represented as a tree-structure, as in (Sanger 1991a).
Although the algorithm involves three separate regressions, each is over only O( n)
terms, and thus the iterative version of the algorithm is only of O(n) complexity
per input pattern processed.
4
RELATION TO OTHER ALGORITHMS
Approximation of functions over a fixed monomial basis is not a new technique
(Gabor 1961, for example) . However, it performs very poorly for high-dimensional
input spaces, since the set of all monomials (even of very low order) can be prohibitively large. This has led to a search for methods which allow the generation of
sparse polynomials. A recent example and bibliography are provided in (Grigoriev
et al. 1990), which describes an algorithm applicable to finite fields (but not to
1067
1068
Sanger, Sutton, and Matheus
j
Figure 3: Products of hidden units in a sigmoidal feedforward network lead to a
polynomial in the hidden units themselves.
real-valued random variables).
The GMDH algorithm (Ivakhnenko 1971, Ikeda et al. 1976, Barron et al. 1984)
incrementally adds new terms to a polynomial by forming a second (or higher)
order polynomial in 2 (or more) of the current terms, and including this polynomial
as a new term if it correlates with the error. Since GMDH does not use look-ahead,
it risks avoiding terms which would be useful at future steps. For example, if the
polynomial to be approximated is xyz where all three variables are independent,
then no polynomial in x and y alone will correlate with the error, and thus the
term xy may never be included. However, x 2y2 does correlate with x 2y2 Z2, so
the look-ahead algorithm presented here would include this term, even though the
error did not decrease until a later step. Although GMDH can be extended to
test polynomials of more than 2 variables, it will always be testing a finite-order
polynomial in a finite number of variables, so there will always exist target functions
which it will not be able to approximate.
Although look-ahead avoids this problem, it is not always useful. For practical
purposes, we may be interested in the best Nth-order approximation to a function,
so it may not be helpful to include terms which participate in monomials of order
greater than N, even if these monomials would cause a large decrease in error.
For example, the best 2nd-order approximation to x 2 + ylOOO + zlOOO may be x 2 ,
even though the other two terms contribute more to the error. In practice, some
combination of both infinite look-ahead and GMDH-type heuristics may be useful.
5
APPLICATION TO OTHER STRUCTURES
These methods have a natural application to other network structures. The inputs
to the polynomial network can be sinusoids (leading to high-dimensional Fourier
representations), Gaussians (leading to high-dimensional Radial Basis Functions)
or other appropriate functions (Sanger 1991a, Sanger 1991b). Polynomials can
Iterative Construction of Sparse Polynomial Approximations
even be applied with sigmoidal networks as input, so that
Xi
=
(T
(I:
SijZj )
where the z;'s are the original inputs, and the Si;'S are the weights to a sigmoidal
hidden unit whose value is the polynomial term Xi. The last layer of hidden units
in a multilayer network is considered to be the set of input features Xi to a linear
output unit, and we can compute the potentials of these features to determine the
hidden unit xp that would most decrease the error if apxp were included in the
model (for the optimal polynomial ap ). But a p can now be approximated using a
subnetwork of any desired type. This subnetwork is used to add a new hidden unit
C&pxp that is the product of xp with the subnetwork output C&p, as in figure 3.
In order to train the C&p subnetwork iteratively using gradient descent, we need to
compute the effect of changes in C&p on the network error ? E[(f - j)2]. We have
=
where S 4pXp is the weight from the new hidden unit to the outpu t. Without loss of
1 by including this factor within C&p. Thus the error
generality we can set S4pXp
term for iteratively training the subnetwork ap is
=
which can be used to drive a standard backpropagation-type gradient descent algorithm. This gives a method for constructing new hidden nodes and a learning
algorithm for training these nodes. The same technique can be applied to deeper
layers in a multilayer network.
6
EXAMPLES
We have applied the algorithm to approximation of known polynomials in the presence of irrelevant noise variables, and to a simple image-processing task.
Figure 4 shows the results of applying the algorithm to 200 samples of the polynomial 2 + 3XIX2 + 4X3X4X5 with 4 irrelevant noise variables. The algorithm correctly
finds the true polynomial in 4 steps, requiring about 5 minutes on a Symbolics Lisp
Machine. Note that although the error did not decrease after cycle 1, the term X4X5
was incorporated since it would be useful in a later step to reduce the error as part
of X3X4X5 in cycle 2.
The image processing task is to predict a pixel value on the succeeding scan line
from a 2x5 block of pixels on the preceding 2 scan lines. If successful, the resulting
polynomial can be used as part of a DPCM image coding strategy. The network
was trained on random blocks from a single face image, and tested on a different
image. Figure 5 shows the original training and test images, the pixel predictions,
and remaining error . Figure 6 shows the resulting 55-term polynomial. Learning
this polynomial required less than 10 minutes on a Sun Sparcstation 1.
1069
1070
Sanger, Sutton, and Matheus
+
+
200 sa.mples of IJ = 2
3z1 z2
4x3 z4 Zs
with 4 additional irrelevant inputs, z6-z9
Original MSE: 1.0
Cycle 1 :
MSE:
Terms:
Coeffs:
Po ten tials:
Top Pairs:
New Term:
0.967
X2
X4
Xl
X3
-0.19
0.14
0.24
0.31
0.22
0.24
0.2S
0 . 32
(S 4) (5 3) (43) (4 4)
XIO =X4 X S
Cycle 2:
MSE:
Terms:
Coeffs:
Potentials:
Top Pairs:
New Term:
0.966
Xl
X2
X3
X4
-0.19
0.14
0.24
0.30
0.25
0.22
0.2S
O.OS
(103) (101) (102) (10 10)
Xu =X10 X 3 =X3 X 4 X S
Cycle 3:
MSE:
Terms:
Coeffs:
Potentials:
Top Pairs:
New Term:
0.349
Xl
X2
X4
X3
0. 04 -0.26
0.09
0.37
0.02
0.S2
0.S9
0.03
(2 1) (2 9) (22) (1 9)
Xu =X1 X 2
Cycle 4:
MSE:
Terms:
Coeffs:
Solution:
0.000
Xl
X2
-0. 00
-0.00
2
X3
-0.00
X4
0.00
Xs
0. 17
0 .33
X6
0.48
0.01
X7
0.03
0.08
X8
O.OS
0. 01
X9
0.S8
0.05
Xs
0.18
0 .02
X6
0.48
0.03
X7
0.03
0.08
X8
O.OS
0.02
X9
0.S7
0.03
XlO
O.OS
0 .47
Xs
-0.04
-0.08
X6
0.27
0. 03
X7
0.10
-O.OS
X8
0 .22
-0.06
X9
0.42
0.05
X10
-0.26
-O.OS
Xll
4.07
O. OS
Xs
-0.00
X6
0 .00
X7
0.00
X8
0.00
X9
0.00
X10
-0.00
Xu
4.00
X l2
3.00
+ 3X1 X2 + 4X3X4X5
Figure 4: A simple example of polynomial learning.
Figure 5: Original, predicted, and error images. The top row is the training image
(RMS error 8.4), and the bottom row is the test image (RMS error 9.4).
Iterative Construction of Sparse Polynomial Approximations
-40? 1z0 + -23.9z1 + -5.4z2 + -17?1z3+
(1.1z 5 + 2.4z8 + -1.1z2 + -1.5z0 + -2.0Z1 + 1.3z 4 + 2.3z6 + 3?1z7 + -25 .6)z4 +
(
(-2.9z9 + 3.0z 8 + -2.9z4 + -2.8z3 + -2 .9z2 + -1.9z5 + -6. 3%0 + -5.2%1 + 2.5z6 + 6.7z7 + 1.1)z9+
(3 . 9z 8 + Z5 + 3.3z4 + 1. 6z3 + 1.1z2 + 2 .9z 6 + 5.0Z7 + 16 .1)z8+
-2.3%3 + -2 .1%2 + -1.6.%1 + 1.1z4 + 2?1z6 + 3.5%7 + 28 .6)z5+
87 ?1z6 + 128.1%7 + 80 .5%8+
(
(-2?6.%9 + -2.4%5 + -4.5%0 + -3 .9%1 + 3.4%6 + 7 .3%7 + -2.5)%9+
21.7%8 + -16 . 0%4 + -12?1z3 + -8.8%2 + 31.4)%9+
2.6
Figure 6: 55-term polynomial used to generate figure 5.
Acknowledgments
We would like to thank Richard Brandau for his helpful comments and suggestions
on an earlier draft of this paper. This report describes research done both at
GTE Laboratories Incorporated, in Waltham MA, and at the laboratory of Dr.
Emilio Bizzi in the department of Brain and Cognitive Sciences at MIT. T. Sanger
was supported during this work by a National Defense Science and Engineering
Graduate Fellowship, and by NIH grants 5R37 AR26710 and 5R01NS09343 to Dr.
Bizzi.
References
BarronR. L., Mucciardi A. N., CookF. J., CraigJ. N., Barron A. R., 1984, Adaptive
learning networks: Development and application in the United States of algorithms
related to GMDH, In Farlow S. J., ed., Self-Organizing Methods in Modeling, pages
25-65, Marcel Dekker, New York.
Gabor D., 1961, A universal nonlinear filter, predictor, and simulator which optimizes itself by a learning process, Proc. lEE, 108B:422-438.
Grigoriev D. Y., Karpinski M., Singer M. F., 1990, Fast parallel algorithms for
sparse polynomial interpolation over finite fields, SIAM J. Computing, 19(6):10591063.
Ikeda S., Ochiai M., Sawaragi Y., 1976, Sequential GMDH algorithm and its application to river flow prediction, IEEE Trans. Systems, Man, and Cybernetics,
SMC-6(7):473-479.
Ivakhnenko A. G., 1971, Polynomial theory of complex systems, IEEE Trans.
Systems, Man, and Cybernetics, SMC-1( 4):364-378.
Sanger T. D., 1991a, Basis-function trees as a generalization of local variable selection methods for function approximation, In Lippmann R. P., Moody J. E.,
Touretzky D. S., ed.s, Advances in Neural Information Processing Systems 3, pages
700-706, Morgan Kaufmann, Proc. NIPS'90, Denver CO.
Sanger T. D., 1991b, A tree-structured adaptive network for function approximation
in high dimensional spaces, IEEE Trans. Neural Networks, 2(2):285-293.
Sutton R. S., Matheus C. J., 1991, Learning polynomial functions by feature construction, In Proc. Eighth Inti. Workshop on Machine Learning, Chicago.
Widrow B., Hoff M. E., 1960, Adaptive switching circuits, In IRE WESCON Conv.
Record, Part 4, pages 96-104.
1071
| 538 |@word version:4 polynomial:47 norm:2 nd:1 dekker:1 united:1 existing:2 current:1 com:2 z2:6 si:1 must:1 ikeda:2 chicago:1 drop:2 succeeding:1 alone:1 record:1 weierstrass:1 draft:1 ire:1 contribute:1 node:2 successive:2 sigmoidal:3 direct:1 consists:1 expected:1 themselves:1 growing:2 simulator:1 brain:1 conv:1 provided:1 discover:1 moreover:1 circuit:1 z:1 developed:1 sparcstation:1 prohibitively:1 unit:12 grant:1 continually:1 before:1 engineering:1 local:1 switching:1 sutton:7 farlow:1 interpolation:1 ap:5 suggests:1 co:1 limited:1 smc:2 graduate:1 practical:1 acknowledgment:1 testing:1 practice:1 block:2 backpropagation:1 x3:6 xr:1 procedure:1 z7:3 universal:1 gabor:3 significantly:1 projection:1 thought:1 road:2 radial:1 suggest:1 onto:1 cannot:1 selection:2 operator:1 s9:1 risk:1 applying:1 measurable:1 maximizing:2 his:1 justification:1 construction:9 target:2 mples:1 us:1 approximated:4 predicts:1 bottom:1 cycle:6 sun:1 r37:1 decrease:9 complexity:1 xio:1 trained:1 creates:1 basis:4 ar26710:1 po:1 joint:5 represented:1 train:1 fast:1 describe:1 choosing:1 whose:1 heuristic:3 valued:1 ability:1 apx:1 itself:1 final:1 product:12 bpg:1 organizing:1 poorly:1 gmdh:6 widrow:2 measured:1 ij:1 sa:1 implemented:1 predicted:2 involves:1 marcel:1 waltham:3 filter:1 dpcm:1 generalization:1 considered:1 predict:4 lm:3 bizzi:2 purpose:1 proc:3 applicable:1 individually:1 largest:1 create:2 mit:2 always:3 partic:1 helpful:2 entire:1 hidden:8 relation:1 interested:1 pixel:4 development:1 ell:1 hoff:2 field:2 never:1 tds:1 x4:6 look:7 future:1 report:1 richard:2 national:1 attempt:4 ab:1 interest:1 accurate:1 xy:1 orthogonal:3 tree:5 desired:2 theoretical:1 earlier:1 modeling:1 coeffs:4 monomials:3 predictor:1 usefulness:2 successful:1 xdi:1 siam:1 river:1 lee:1 terence:1 e25:1 moody:1 squared:6 x9:4 choose:5 dr:2 cognitive:1 leading:2 potential:13 coding:1 coefficient:6 later:2 parallel:1 pxp:2 contribution:1 minimize:2 square:1 kaufmann:1 efficiently:1 sylvan:2 drive:1 cybernetics:2 touretzky:1 ed:2 against:3 e2:1 stop:1 massachusetts:1 actually:1 z9:3 higher:2 axp:1 x6:4 response:1 done:1 though:2 generality:1 until:1 christopher:1 nonlinear:3 o:7 incrementally:3 perhaps:2 alb:1 effect:1 requiring:1 y2:2 true:1 sinusoid:1 laboratory:4 iteratively:2 x5:1 during:1 self:1 performs:1 image:10 consideration:1 novel:1 recently:1 nih:1 denver:1 s8:1 measurement:1 cambridge:1 ai:1 similarly:1 inclusion:3 z4:5 depiction:1 add:2 recent:1 irrelevant:3 optimizes:1 accomplished:1 integrable:1 morgan:1 greater:1 additional:1 preceding:1 determine:3 maximize:2 emilio:1 x10:3 grigoriev:2 z5:3 prediction:2 regression:14 multilayer:2 expectation:1 iteration:1 karpinski:1 addition:3 fellowship:1 interval:1 comment:1 induced:1 member:1 flow:1 lisp:1 presence:1 feedforward:1 xj:1 inner:1 reduce:2 whether:3 rms:2 defense:1 utility:1 s7:1 york:1 cause:1 useful:6 ten:1 processed:1 generate:1 exist:1 per:1 correctly:1 write:2 threshold:1 decide:1 layer:3 ahead:7 incorporation:1 orthogonality:2 bp:4 x2:5 bibliography:1 x7:4 fourier:1 wpq:1 department:1 structured:1 combination:3 describes:2 xixj:2 wi:1 inti:1 taken:1 xyz:1 singer:1 available:1 gaussians:1 barron:2 appropriate:1 ochiai:1 original:6 symbolics:1 remaining:2 include:5 top:4 sanger:11 prof:1 approximating:1 already:2 added:1 strategy:1 subnetwork:5 gradient:2 subspace:1 separate:1 thank:1 participate:2 z3:4 unfortunately:1 wpe:2 e2x:1 unknown:1 perform:1 xll:1 finite:5 descent:2 matheus:7 extended:2 incorporated:5 arbitrary:1 pair:4 required:1 z1:3 nip:1 trans:3 able:1 below:1 pattern:1 eighth:1 ivakhnenko:2 built:1 including:4 power:1 natural:1 predicting:1 residual:1 nth:1 technology:1 apxp:9 x8:4 xq:2 l2:1 loss:1 expect:1 generation:1 suggestion:1 xp:11 principle:1 row:2 supported:1 last:1 arriving:1 monomial:1 allow:1 deeper:1 institute:1 face:1 sparse:10 xn:1 avoids:1 z8:2 computes:2 adaptive:3 qx:1 correlate:3 approximate:2 lippmann:1 sequentially:1 wescon:1 xi:7 search:1 iterative:10 z6:5 contributes:1 mse:5 complex:1 constructing:1 did:2 dense:1 s2:1 noise:2 repeated:1 xu:3 x1:2 wish:1 xl:4 third:1 theorem:1 minute:2 z0:2 outpu:1 x:4 exists:2 incorporating:1 workshop:1 adding:2 sequential:1 ci:1 magnitude:1 justifies:1 depicted:1 led:1 likely:1 forming:1 xlo:1 ma:4 xix2:1 goal:1 room:1 man:2 change:1 included:6 infinite:3 determined:1 gte:5 called:1 select:1 scan:2 tested:1 avoiding:1 |
4,838 | 5,380 | Discovering, Learning and Exploiting Relevance
Mihaela van der Schaar
Electrical Engineering Department
University of California Los Angeles
[email protected]
Cem Tekin
Electrical Engineering Department
University of California Los Angeles
[email protected]
Abstract
In this paper we consider the problem of learning online what is the information
to consider when making sequential decisions. We formalize this as a contextual
multi-armed bandit problem where a high dimensional (D-dimensional) context
vector arrives to a learner which needs to select an action to maximize its expected
reward at each time step. Each dimension of the context vector is called a type. We
assume that there exists an unknown relation between actions and types, called the
relevance relation, such that the reward of an action only depends on the contexts
of the relevant types. When the relation is a function, i.e., the reward of an action
only depends on the context of a single type, and the expected reward of an action
is Lipschitz continuous in the context of its relevant type, we propose an algo?
? ? ) regret with a high probability, where ? = 2/(1 + 2).
rithm that achieves O(T
Our algorithm achieves this by learning the unknown relevance relation, whereas
prior contextual bandit algorithms that do not exploit the existence of a relevance
? (D+1)/(D+2) ) regret. Our algorithm alternates between exrelation will have O(T
ploring and exploiting, it does not require reward observations in exploitations,
and it guarantees with a high probability that actions with suboptimality greater
than are never selected in exploitations. Our proposed method can be applied to
a variety of learning applications including medical diagnosis, recommender systems, popularity prediction from social networks, network security etc., where at
each instance of time vast amounts of different types of information are available
to the decision maker, but the effect of an action depends only on a single type.
1
Introduction
In numerous learning problems the decision maker is provided with vast amounts of different types
of information which it can utilize to learn how to select actions that lead to high rewards. The
value of each type of information can be regarded as the context on which the learner acts, hence
all the information can be encoded in a context vector. We focus on problems where this context
vector is high dimensional but the reward of an action only depends on a small subset of types. This
dependence is given in terms of a relation between actions and types, which is called the relevance
relation. For an action set A and a type set D, the relevance relation is given by R = {R(a)}a?A ,
where R(a) ? D. Expected reward of an action a only depends on the values of the relevant
types of contexts. Hence, for a context vector x, action a?s expected reward is equal to ?(a, xR(a) ),
where xR(a) is the context vector corresponding to the types in R(a). Several examples of relevance
relations and their effect on expected action rewards are given in Fig. 1. The problem of finding
the relevance relation is important especially when maxa?A |R(a)| << |D|.1 In this paper we
consider the case when the relevance relation is a function, i.e., |R(a)| = 1, for all a ? A, which is
an important special case. We discuss the extension of our framework to the more general case in
Section 3.3.
1
For a set A, |A| denotes its cardinality.
1
Figure 1: Examples of relevance relations: (i) general relevance relation, (ii) linear relevance relation, (iii) relevance function. In this paper we only consider (iii), while our methods can easily be
generalized to (i) and (ii).
Relevance relations exists naturally in many practical applications. For example, when sequentially
treating patients with a particular disease, many types of information (contexts) are usually available
- the patients? age, weight, blood tests, scans, medical history etc. If a drug?s effect on a patient is
caused by only one of the types, then learning the relevant type for the drug will result in significantly
faster learning for the effectiveness of the drug for the patients.2 Another example is recommender
systems, where recommendations are made based on the high dimensional information obtained
from the browsing and purchase histories of the users. A user?s response to a product recommendation will depend on the user?s gender, occupation, history of past purchases etc., while his/her
response to other product recommendations may depend on completely different information about
the user such as the age and home address.
Traditional contextual bandit solutions disregard existence of such relations, hence have regret
bounds that scale exponentially with the dimension of the context vector [1, 2]. In order to solve the
curse of dimensionality problem, a new approach which learns the relevance relation in an online
way is required. The algorithm we propose simultaneously learns the relevance relation (when it is a
function) and the action rewards by comparing sample mean rewards of each action for context pairs
of different types that are calculated based on the context and reward observations so far. The only
assumption we make about actions and contexts is the Lipschitz continuity of expected reward of an
action in the context of its relevant type. Our main contributions can be summarized as follows:
? We propose the Online Relevance Learning with Controlled Feedback (ORL-CF) algorithm
that alternates between exploration
and exploitation phases, which achieves a regret bound
?
? 3
?
of O(T ), with ? = 2/(1 + 2), when the relevance relation is a function.
? We derive separate bounds on the regret incurred in exploration and exploitation phases.
ORL-CF only needs to observe the reward in exploration phases, hence the reward feedback
is controlled. ORL-CF achieves the same time order of regret even when observing the
reward has a non-zero cost.
? Given any ? > 0, which is an input to ORL-CF, suboptimal actions will never be selected
in exploitation steps with probability at least 1 ? ?. This is very important, perhaps vital in
numerous applications where the performance needs to be guaranteed, such as healthcare.
Due to the limited space, numerical results on the performance of our proposed algorithm is included
in the supplementary material.
2
Even when there are multiple relevant types for each action, but there is one dominant type whose effect
on the reward of the action is significantly larger than the effects of other types, assuming that the relevance
relation is a function will be a good approximation.
3
? is the same as O(?) except it hides terms that have polylogarithmic growth.
O(?) is the Big O notation, O(?)
2
2
Problem Formulation
A is the set of actions, D is the dimension of the context vector, D := {1, 2, . . . , D} is the set of
types, and R = {R(a)}a?A : A ? D is the relevance function, which maps every a ? A to a
unique d ? D. At each time step t = 1, 2, . . ., a context vector xt arrives to the learner. After
observing xt the learner selects an action a ? A, which results in a random reward rt (a, xt ). The
learner may choose to observe this reward by paying cost cO ? 0. The goal of the learner is to
maximize the sum of the generated rewards minus costs of observations for any time horizon T .
Each xt consists of D types of contexts, and can be written as xt = (x1,t , x2,t , . . . , xD,t ) where xi,t
is called the type i context. Xi denotes the space of type i contexts and X := X1 ? X2 ? . . . ? XD
denotes the space of context vectors. At any t, we have xi,t ? Xi for all i ? D. For the sake of
notational simplicity we take Xi = [0, 1] for all i ? D, but all our results can be generalized to
the case when Xi is a bounded subset of the real line. For x = (x1 , x2 , . . . , xD ) ? X , rt (a, x)
is generated according to an i.i.d. process with distribution F (a, xR(a) ) with support in [0, 1] and
expected value ?(a, xR(a) ).
The following assumption gives a similarity structure between the expected reward of an action and
the contexts of the type that is relevant to that action.
Assumption 1. For all a ? A, x, x0 ? X , we have |?(a, xR(a) )??(a, x0R(a) )| ? L|xR(a) ?x0R(a) |,
where L > 0 is the Lipschitz constant.
We assume that the learner knows the L given in Assumption 1. This is a natural assumption in
contextual bandit problems [1, 2]. Given a context vector x = (x1 , x2 , . . . , xD ), the optimal action
is a? (x) := arg maxa?A ?(a, xR(a) ), but the learner does not know it since it does not know R,
F (a, xR(a) ) and ?(a, xR(a) ) for a ? A, x ? X a priori. In order to assess the learner?s loss due to
unknowns, we compare its performance with the performance of an oracle benchmark which knows
a? (x) for all x ? X . Let ?t (a) := ?(a, xR(a),t ). The action chosen by the learner at time t is
denoted by ?t . The learner also decides whether to observe the reward or not, and this decision of
the learner at time t is denoted by ?t ? {0, 1}, where ?t = 1 implies that the learner chooses to
observe the reward and ?t = 0 implies that the learner does not observe the reward. The learner?s
performance loss with respect to the oracle benchmark is defined as the regret, whose value at time
T is given by
T
T
X
X
(?t (?t ) ? cO ?t ).
(1)
?t (a? (xt )) ?
R(T ) :=
t=1
t=1
A regret that grows sublinearly in T , i.e., O(T ? ), ? < 1, guarantees convergence in terms of the
average reward, i.e., R(T )/T ? 0. We are interested in achieving sublinear growth with a rate
independent of D.
3
3.1
Online Relevance Learning with Controlled Feedback
Description of the algorithm
In this section we propose the algorithm Online Relevance Learning with Controlled Feedback
(ORL-CF), which learns the best action for each context vector by simultaneously learning the relevance relation, and then estimating the expected reward of each action. The feedback, i.e., reward
observations, is controlled based on the past context vector arrivals, in a way that reward observations are only made for actions for which the uncertainty in the reward estimates are high for
the current context vector. The controlled feedback feature allows ORL-CF to operate as an active
learning algorithm. Operation of ORL-CF can be summarized as follows:
? Adaptively discretize (partition) the context space of each type to learn action rewards of
similar contexts together.
? For an action, form reward estimates for pairs of intervals corresponding to pairs of types.
Based on the accuracy of these estimates, either choose to explore and observe the reward
or choose to exploit the best estimated action for the current context vector.
? In order to choose the best action, compare the reward estimates for pairs of intervals for
which one interval belongs to type i, for each type i and action a. Conclude that type i
3
is relevant to a if the variation of the reward estimates does not greatly exceed the natural
variation of the expected reward of action a over the interval of type i (calculated using
Assumption 1).
Online Relevance Learning with Controlled Feedback (ORL-CF):
1: Input: L, ?, ?.
2: Initialization: Pi,1 = {[0, 1]}, i ? D. Run Initialize(i, Pi,1 , 1), i ? D.
3: while t ? 1 do
4:
Observe xS
t , find pt that xt belongs to.
5:
Set Ut := i?D Ui,t , where Ui,t (given in (3)), is the set of under explored actions for type i.
6:
if Ut 6= ? then
7:
(Explore) ?t = 1, select ?t randomly from Ut , observe rt (?t , xt ).
8:
Update pairwise sample means: for all q ? Qt , given in (2).
r?ind(q) (q, ?t ) = (S ind(q) (q, ?t )?
rind(q) (q, ?t ) + rt (?t , xt ))/(S ind(q) (q, ?t ) + 1).
9:
Update counters: for all q ? Qt , S ind(q) (q, ?t ) + +.
10:
else
11:
(Exploit) ?t = 0, for each a ? A calculate the set of candidate relevant contexts Relt (a) given
in (4).
12:
for a ? A do
13:
if Relt (a) = ? then
14:
Randomly select c?t (a) from D.
15:
else
16:
For each i ? Relt (a), calculate Vart (i, a) given in (5).
17:
Set c?t (a) = arg mini?Relt (a) Vart (i, a).
18:
end if
c
? (a)
19:
Calculate r?t t (a) as given in (6).
20:
end for
c
? (a)
21:
Select ?t = arg maxa?A r?t t (pc?t (a),t , a).
22:
end if
23:
for i ? D do
24:
N i (pi,t ) + +.
25:
if N i (pi,t ) ? 2?l(pi,t ) then
26:
Create two new level l(pi,t ) + 1 intervals p, p0 whose union gives pi,t .
27:
Pi,t+1 = Pi,t ? {p, p0 } ? {pi,t }.
28:
Run Initialize(i, {p, p0 }, t).
29:
else
30:
Pi,t+1 = Pi,t .
31:
end if
32:
end for
33:
t=t+1
34: end while
Initialize(i, B, t):
1: for p ? B do
2:
Set N i (p) = 0, r?i,j (p, pj , a) = r?j,i (pj , p, a) = 0, S i,j (p, pj , a) = S j,i (pj , p, a) = 0 for all a ? A,
j ? D?i and pj ? Pj,t .
3: end for
Figure 2: Pseudocode for ORL-CF.
Since the number of contexts is infinite, learning the reward of an action for each context is not
feasible. In order to learn fast, ORL-CF exploits the similarities between the contexts of the relevant
type given in Assumption 1 to estimate the rewards of the actions. The key to success of our algorithm is that this estimation is good enough. ORL-CF adaptively forms the partition of the space
for each type in D, where the partition for the context space of type i at time t is denoted by Pi,t .
All the elements of Pi,t are disjoint intervals of Xi = [0, 1] whose lengths are elements of the set
{1, 2?1 , 2?2 , . . .}.4 An interval with length 2?l , l ? 0 is called a level l interval, and for an interval
p, l(p) denotes its level, s(p) denotes its length. By convention, intervals are of the form (a, b], with
the only exception being the interval containing 0, which is of the form [0, b].5 Let pi,t ? Pi,t be the
interval that xi,t belongs to, pt := (p1,t , . . . , pD,t ) and P t := (P1,t , . . . , PD,t ).
4
Setting interval lengths to powers of 2 is for presentational simplicity. In general, interval lengths can be
set to powers of any real number greater than 1.
5
Endpoints of intervals will not matter in our analysis, so our results will hold even when the intervals have
common endpoints.
4
The pseudocode of ORL-CF is given in Fig. 2. ORL-CF starts with Pi,1 = {Xi } = {[0, 1]} for
each i ? D. As time goes on and more contexts arrive for each type i, it divides Xi into smaller and
smaller intervals. The idea is to combine the past observations made in an interval to form sample
mean reward estimates for each interval, and use it to approximate the expected rewards of actions
for contexts lying in these intervals. The intervals are created in a way to balance the variation of
the sample mean rewards due to the number of past observations that are used to calculate them and
the variation of the expected rewards in each interval.
We also call Pi,t the set of active intervals for type i at time t. Since the partition of each type is
adaptive, as time goes on, new intervals become active while old intervals are deactivated, based on
how contexts arrive. For a type i interval p, let Nti (p) be the number of times xi,t0 ? p ? Pi,t0 for
t0 ? t. The duration of time that an interval remains active, i.e., its lifetime, is determined by an input
parameter ? > 0, which is called the duration parameter. Whenever the number of arrivals to an
interval p exceeds 2?l(p) , ORL-CF deactivates p and creates two level l(p)+1 intervals, whose union
gives p. For example, when pi,t = (k2?l , (k + 1)2?l ] for some 0 < k ? 2l ? 1 if Nti (pi,t ) ? 2?l ,
ORL-CF sets
Pi,t+1 = Pi,t ? {(k2?l , (k + 1/2)2?l ], ((k + 1/2)2?l , (k + 1)2?l ]} ? {pi,t }.
Otherwise Pi,t+1 remains the same as Pi,t . It is easy to see that the lifetime of an interval increases
exponentially in its duration parameter.
We next describe the counters, control numbers and sample mean rewards the learner keeps for each
pair of intervals corresponding to a pair of types to determine whether to explore or exploit and how
to exploit. Let D?i := D ? {i}. For type i, let Qi,t := {(pi,t , pj,t ) : j ? D?i } be the pair of
intervals that are related to type i at time t, and let
[
Qt :=
Qi,t .
(2)
i?D
To denote an element of Qi,t or Qt we use index q. For any q ? Qt , the corresponding pair of types
is denoted by ind(q). For example, ind((pi,t , pj,t )) = i, j. The decision to explore or exploit at time
t is solely
T based on pt . For events A1 , . . . , AK , let I(A1 , . . . , Ak ) denote the indicator function of
event k=1:K Ak . For p ? Pi,t , p0 ? Pj,t , let
Sti,j (p, p0 , a) :=
t?1
X
I (?t0 = a, ?t = 1, pi,t0 = p, pj,t0 = p0 ) ,
t0 =1
be the number of times a is selected and the reward is observed when the type i context is in p and
type j context is in p0 , summed over times when both intervals are active. Also for the same p and
p0 let
!
t?1
X
i,j
0
0
r?t (p, p , a) :=
rt (a, xt )I (?t0 = a, ?t = 1, pi,t0 = p, pj,t0 = p ) /(Sti,j (p, p0 , a)),
t0 =1
be the pairwise sample mean reward of action a for pair of intervals (p, p0 ).
At time t, ORL-CF assigns a control number to each i ? D denoted by
Di,t :=
2 log(tD|A|/?)
,
(Ls(pi,t ))2
which depends on the cardinality of A, the length of the active interval that type i context is in
at time t and a confidence parameter ? > 0, which controls the accuracy of sample mean reward
estimates. Then, it computes the set of under-explored actions for type i as
ind(q)
Ui,t := {a ? A : St
(q, a) < Di,t for some q ? Qi (t)},
(3)
S
and then, the set of under-explored actions as Ut := i?D Ui,t . The decision to explore or exploit is
based on whether or not Ut is empty.
(i) If Ut 6= ?, ORL-CF randomly selects an action ?t ? Ut to explore, and observes its reward
rt (?t , xt ). Then, it updates the pairwise sample mean rewards and pairwise counters for all q ? Qt ,
ind(q)
r?t+1 (q, ?t ) =
ind(q)
St
ind(q)
(q,?t )?
rt+1 (q,?t )+rt (?t ,xt )
ind(q)
St
(q,?t )+1
ind(q)
ind(q)
, St+1 (q, ?t ) = St
5
(q, ?t ) + 1.
(ii) If Ut = ?, ORL-CF exploits by estimating the relevant type c?t (a) for each a ? A and forming
sample mean reward estimates for action a based on c?t (a). It first computes the set of candidate
relevant types for each a ? A,
Relt (a) := {i ? D : |?
rti,j (pi,t , pj,t , a) ? r?ti,k (pi,t , pk,t , a)| ? 3Ls(pi,t ), ?j, k ? D?i }.
(4)
The intuition is that if i is the type that is relevant to a, then independent of the values of the contexts
of the other types, the variation of the pairwise sample mean reward of a over pi,t must be very close
to the variation of the expected reward of a in that interval.
If Relt (a) is empty, this implies that ORL-CF failed to identify the relevant type, hence c?t (a) is
randomly selected from D. If Relt (a) is nonempty, ORL-CF computes the maximum variation
Vart (i, a) := max |?
rti,j (pi,t , pj,t , a) ? r?ti,k (pi,t , pk,t , a)|,
j,k?D?i
(5)
for each i ? Relt (a). Then it sets c?t (a) = mini?Relt (a) Vart (i, a). This way, whenever the type
relevant to action a is in Relt (a), even if it is not selected as the estimated relevant type, the sample
mean reward of a calculated based on the estimated relevant type will be very close to the sample
mean of its reward calculated according to the true relevant type. After finding the estimated relevant
types, the sample mean reward of each action is computed based on its estimated relevant type as
P
c? (a),j
c? (a),j
r? t
(pc?t (a),t , pj,t , a)St t
(pc?t (a),t , pj,t , a)
j?D??
ct (a) t
c?t (a)
r?t
(a) :=
.
(6)
P
c? (a),j
St t
(pc?t (a),t , pj,t , a)
j?D??
c (a)
t
c? (a)
arg maxa?A r?t t (pc?t (a),t , a).
Then, ORL-CF selects ?t =
Since the reward is not observed in
exploitations, pairwise sample mean rewards and counters are not updated.
3.2
Regret analysis of ORL-CF
Let ? (T ) ? {1, 2, . . . , T } be the set of time steps in which ORL-CF exploits by time T . ? (T ) is a
random set which depends on context arrivals and the randomness of the action selection of ORLCF. The regret R(T ) defined in (1) can be written as a sum of the regret incurred during explorations
(denoted by RO (T )) and the regret incurred during exploitations (denoted by RI (T )). The following
theorem gives a bound on the regret of ORL-CF in exploitation steps.
Theorem 1. Let ORL-CF run with duration parameter ? > 0, confidence parameter ? > 0 and
control numbers Di,t := 2 log(t|A|D/?)
(Ls(pi,t ))2 , for i ? D. Let Rinst (t) be the instantaneous regret at time t,
which is the loss in expected reward at time t due to not selecting a? (xt ). Then, with probability at
least 1 ? ?, we have
Rinst (t) ? 8L(s(pR(?t ),t ) + s(pR(a? (xt )),t )),
for all t ? ? (T ), and the total regret in exploitation steps is bounded above by
X
RI (T ) ? 8L
(s(pR(?t ),t + s(pR(a? (xt )),t )) ? 16L22? T ?/(1+?) ,
t?? (T )
for arbitrary context vectors x1 , x2 , . . . , xT .
Theorem 1 provides both context arrival process dependent and worst case bounds on the exploitation regret of ORL-CF. By choosing ? arbitrarily close to zero, RI (T ) can be made O(T ? ) for any
? > 0. While this is true, the reduction in regret for smaller ? not only comes from increased accuracy, but it is also due to the reduction in the number of time steps in which ORL-CF exploits, i.e.,
|? (T )|. By definition time t is an exploitation step if
Sti,j (pi,t , pj,t , a) ?
L2
22 max{l(pi,t ),l(pj,t )}+1 log(t|A|D/?)
2 log(t|A|D/?)
=
,
2
2
min{s(pi,t ) , s(pj,t ) }
L2
for all q = (pi,t , pj,t ) ? Qt , i, j ? D. This implies that for any q ? Qi,t which has the interval
? 2l ) explorations are required before any exploitation can take
with maximum level equal to l, O(2
place. Since the time a level l interval can stay active is 2?l , it is required that ? ? 2 so that ? (T ) is
nonempty.
The next theorem gives a bound on the regret of ORL-CF in exploration steps.
6
Theorem 2. Let ORL-CF run with ?, ? and Di,t , i ? D values as stated in Theorem 1. Then,
RO (T ) ?
960D2 (cO + 1) log(T |A|D/?) 4/? 64D2 (cO + 1) 2/?
T
+
T ,
7L2
3
with probability 1, for arbitrary context vectors x1 , x2 , . . . , xT .
Based on the choice of the duration parameter ?, which determines how long an interval will stay
active, it is possible to get different regret bounds for explorations and exploitations. Any ? > 4 will
give a sublinear regret bound for both explorations and exploitations. The regret in exploitations
increases in ? while the regret in explorations decreases in ?.
?
Theorem 3. Let ORL-CF run with ? and Di,t , i ? D values as stated in Theorem 1 and ? = 2+2 2.
Then, the time order of exploration and exploitation regrets are?balanced up to logaritmic orders.
?
? 2/(1+ 2) ) and RO (T ) = O(T
? 2/(1+ 2) ) .
With probability at least 1 ? ? we have both RI (T ) = O(T
Remark 1. Prior work on contextual bandits focused on balancing the regret due to exploration
and exploitation. For example in [1, 2], for a D-dimensional context vector algorithms are shown
? (D+1)/(D+2) ) regret.6 Also in [1] a O(T (D+1)/(D+2) ) lower bound on the regret
to achieve O(T
is proved. An interesting question is to find the tightest lower bound for contextual bandits with
relevance function. One trivial lower bound is O(T 2/3 ), which corresponds to D = 1. However,
since finding the action with the highest expected reward for a context vector requires comparisons
of estimated rewards of actions with different relevant types, which requires accurate sample mean
reward estimates for 2 dimensions of the context space corresponding to those types, we conjecture
that a tighter lower bound is O(T 3/4 ). Proving this is left as future work.
Another interesting case is when actions with suboptimality greater than > 0 must never be chosen
in any exploitation step by time T . When such a condition is imposed, ORL-CF can start with
partitions Pi,1 that have sets with high levels such that it explores more at the beginning to have
more accurate reward estimates before any exploitation. The following theorem gives the regret
bound of ORL-CF for this case.
Theorem 4. Let ORL-CF run with duration parameter ? > 0, confidence parameter ? > 0, control
numbers Di,t := 2 log(t|A|D/?)
(Ls(pi,t ))2 , and with initial partitions Pi,1 , i ? D consisting of intervals of
length lmin = dlog2 (3L/(2))e. Then, with probability 1 ? ?, Rinst (t) ? for all t ? ? (T ),
RI (T ) ? 16L22? T ?/(1+?) and
81L4 960D2 (cO + 1) log(T |A|D/?) 4/? 64D2 (cO + 1) 2/?
RO (T ) ? 4
T
+
T
,
7L2
3
for arbitrary
context vectors x1 , x2 , . . . , xT . Bounds on RI (T ) and RO (T ) are balanced for ? =
?
2 + 2 2.
3.3
Future Work
In this paper we only considered the relevance relations that are functions. Similar learning methods
can be developed for more general relevance relations such as the ones given in Fig. 1 (i) and (ii).
For example, for the general case in Fig. 1 (i), if |R(a)| ? Drel << D, for all a ? A, and Drel is
known by the learner, the following variant of ORL-CF can be used to achieve regret whose time
order depends only on Drel but not on D.
? Instead of keeping pairwise sample mean reward estimates, keep sample mean reward estimates of actions for Drel + 1 tuples of intervals of Drel + 1 types.
? For a Drel tuple of types i, let Qi,t be the Drel + 1 tuples of intervals that are related to i
at time t, and Qt be the union of Qi,t over all Drel tuples of types. Similar to ORL-CF,
compute the set of under-explored actions Ui,t , and the set of candidate relevant Drel tuples
of types Relt (a), using the newly defined sample mean reward estimates.
6
The results are shown in terms of the covering dimension which reduces to Euclidian dimension for our
problem.
7
? In exploitation, set c?t (a) to be the Drel tuple of types with the minimax variation, where the
variation of action a for a tuple i is defined similar to (5), as the maximum of the distance
between the sample mean rewards of action a for Drel +1 tuples that are in Qi,t .
Another interesting case is when the relevance relation is linear as given in Fig. 1 (ii). For example,
for action a if there is a type i that is much more relevant compared to other types j ? D?i , i.e.,
wa,i >> wa,j , where the weights wa,i are given in Fig. 1, then ORL-CF is expected to have good
performance (but not sublinear regret with respect to the benchmark that knows R).
4
Related Work
Contextual bandit problems are studied by many others in the past [3, 4, 1, 2, 5, 6]. The problem we consider in this paper is a special case of the Lipschitz contextual bandit problem [1, 2],
where the only assumption is the existence of a known similarity metric between the expected rewards of actions for different contexts. It is known that the lower bound on regret for this problem
? (D+1)/(D+2) ) regret [1, 2].
is O(T (D+1)/(D+2) ) [1], and there exists algorithms that achieve O(T
Compared to the prior work above, ORL-CF only needs to observe rewards in explorations and has
a regret whose time order is independent of D. Hence it can still learn the optimal actions fast
enough in settings where observations are costly and context vector is high dimensional.
Examples of related works that consider limited observations are KWIK learning [7, 8] and label
efficient learning [9, 10, 11]. For example, [8] considers a bandit model where the reward function
comes from a parameterized family of functions and gives bound on the average regret. An online
prediction problem is considered in [9, 10, 11], where the predictor (action) lies in a class of linear
predictors. The benchmark of the context is the best linear predictor. This restriction plays a crucial
role in deriving regret bounds whose time order does not depend on D. Similar to these works,
ORL-CF can guarantee with a high probability that actions with large suboptimalities will never
be selected in exploitation steps. However, we do not have any assumptions on the form of the
expected reward function other than the Lipschitz continuity and that it depends on a single type for
each action.
In [12] graphical bandits are proposed where the learner takes an action vector a which includes
actions from several types that consitute a type set T . The expected reward of a for context vector x
can be decomposed into sum of reward functions each of which only depends on a subset of D ? T .
However, it is assumed that the form of decomposition is known but the functions are not known.
Another work [13] proposes a fast learning algorithm for an i.i.d. contextual bandit problem in
which the rewards for contexts and actions are sampled from a joint probability distribution. In this
work the authors consider learning
? the best policy from a finite set of policies with oracle access,
and prove a regret bound of O( T ) which is also logarithmic in the size of the policy space. In
contrast, in our problem (i) contexts arrive according to an arbitrary exogenous process, and the
action rewards are sampled from an i.i.d. distribution given the context value, (ii) the set of policies
that the learner can adopt is not restricted.
Large dimensional action spaces, where the rewards depend on a subset of the types of actions are
considered in [14] and [15]. [14] considers the problem when the reward is H?older continuous in
an unknown low-dimensional tuple of types, and uses a special discretization of the action space
to achieve dimension independent bounds on the regret. This discretization can be effectively used
since the learner can select the actions, as opposed to our case where the learner does not have any
control over contexts. [15] considers the problem of optimizing high dimensional functions that
have an unknown low dimensional structure from noisy observations.
5
Conclusion
In this paper we formalized the problem of learning the best action through learning the relevance
relation between types of contexts and actions. For the case when the relevance relation is a function,
we proposed an algorithm that (i) has sublinear regret with time order independent of D, (ii) only
requires reward observations in explorations, (iii) for any > 0, does not select any suboptimal
actions in exploitations with a high probability. In the future we will extend our results to the linear
and general relevance relations illustrated in Fig. 1.
8
References
[1] T. Lu, D. P?al, and M. P?al, ?Contextual multi-armed bandits,? in International Conference on
Artificial Intelligence and Statistics (AISTATS), 2010, pp. 485?492.
[2] A. Slivkins, ?Contextual bandits with similarity information,? in Conference on Learning Theory (COLT), 2011.
[3] E. Hazan and N. Megiddo, ?Online learning with prior knowledge,? in Learning Theory.
Springer, 2007, pp. 499?513.
[4] J. Langford and T. Zhang, ?The epoch-greedy algorithm for contextual multi-armed bandits,?
Advances in Neural Information Processing Systems (NIPS), vol. 20, pp. 1096?1103, 2007.
[5] W. Chu, L. Li, L. Reyzin, and R. E. Schapire, ?Contextual bandits with linear payoff functions,?
in International Conference on Artificial Intelligence and Statistics (AISTATS), 2011, pp. 208?
214.
[6] M. Dudik, D. Hsu, S. Kale, N. Karampatziakis, J. Langford, L. Reyzin, and T. Zhang, ?Efficient
optimal learning for contextual bandits,? arXiv preprint arXiv:1106.2369, 2011.
[7] L. Li, M. L. Littman, T. J. Walsh, and A. L. Strehl, ?Knows what it knows: a framework for
self-aware learning,? Machine Learning, vol. 82, no. 3, pp. 399?443, 2011.
[8] K. Amin, M. Kearns, M. Draief, and J. D. Abernethy, ?Large-scale bandit problems and KWIK
learning,? in International Conference on Machine Learning (ICML), 2013, pp. 588?596.
[9] N. Cesa-Bianchi, C. Gentile, and F. Orabona, ?Robust bounds for classification via selective
sampling,? in International Conference on Machine Learning (ICML), 2009, pp. 121?128.
[10] S. M. Kakade, S. Shalev-Shwartz, and A. Tewari, ?Efficient bandit algorithms for online multiclass prediction,? in International Conference on Machine Learning (ICML), 2008, pp. 440?
447.
[11] E. Hazan and S. Kale, ?Newtron: an efficient bandit algorithm for online multiclass prediction.?
in Advances in Neural Information Processing Systems (NIPS), 2011, pp. 891?899.
[12] K. Amin, M. Kearns, and U. Syed, ?Graphical models for bandit problems,? in Conference on
Uncertainty in Artificial Intelligence (UAI), 2011.
[13] A. Agarwal, D. Hsu, S. Kale, J. Langford, L. Li, and R. E. Schapire, ?Taming the monster: A
fast and simple algorithm for contextual bandits,? arXiv preprint arXiv:1402.0555, 2014.
[14] H. Tyagi and B. Gartner, ?Continuum armed bandit problem of few variables in high dimensions,? in Workshop on Approximation and Online Algorithms (WAOA), 2014, pp. 108?119.
[15] J. Djolonga, A. Krause, and V. Cevher, ?High-dimensional Gaussian process bandits,? in Advances in Neural Information Processing Systems (NIPS), 2013, pp. 1025?1033.
9
| 5380 |@word exploitation:22 d2:4 decomposition:1 p0:10 euclidian:1 minus:1 reduction:2 rind:1 initial:1 selecting:1 past:5 current:2 contextual:15 comparing:1 discretization:2 mihaela:2 chu:1 written:2 must:2 numerical:1 partition:6 treating:1 update:3 intelligence:3 discovering:1 selected:6 greedy:1 beginning:1 provides:1 zhang:2 become:1 consists:1 prove:1 combine:1 newtron:1 pairwise:7 x0:1 sublinearly:1 expected:19 p1:2 multi:3 decomposed:1 td:1 armed:4 curse:1 cardinality:2 provided:1 estimating:2 notation:1 bounded:2 what:2 maxa:4 developed:1 finding:3 guarantee:3 every:1 act:1 ti:2 xd:4 growth:2 megiddo:1 draief:1 ro:5 k2:2 control:6 healthcare:1 medical:2 before:2 engineering:2 ak:3 solely:1 initialization:1 studied:1 co:6 limited:2 walsh:1 practical:1 unique:1 union:3 regret:37 xr:10 drug:3 significantly:2 confidence:3 get:1 close:3 selection:1 context:60 lmin:1 restriction:1 map:1 imposed:1 go:2 kale:3 duration:6 l:4 focused:1 tekin:1 simplicity:2 formalized:1 assigns:1 regarded:1 deriving:1 his:1 proving:1 variation:9 updated:1 pt:3 play:1 user:4 us:1 element:3 schaar:1 observed:2 role:1 monster:1 preprint:2 electrical:2 worst:1 calculate:4 counter:4 decrease:1 highest:1 observes:1 disease:1 intuition:1 pd:2 balanced:2 ui:5 reward:76 littman:1 depend:4 algo:1 creates:1 learner:21 completely:1 easily:1 joint:1 fast:4 describe:1 artificial:3 choosing:1 shalev:1 abernethy:1 whose:8 encoded:1 supplementary:1 solve:1 larger:1 otherwise:1 statistic:2 noisy:1 online:11 propose:4 product:2 relevant:23 reyzin:2 achieve:4 deactivates:1 amin:2 description:1 los:2 exploiting:2 convergence:1 empty:2 derive:1 qt:8 paying:1 implies:4 come:2 convention:1 exploration:13 material:1 require:1 tighter:1 extension:1 hold:1 lying:1 considered:3 achieves:4 adopt:1 continuum:1 estimation:1 label:1 maker:2 create:1 gaussian:1 tyagi:1 focus:1 notational:1 karampatziakis:1 greatly:1 contrast:1 dependent:1 her:1 bandit:23 relation:26 selective:1 selects:3 interested:1 arg:4 classification:1 colt:1 denoted:7 priori:1 proposes:1 special:3 initialize:3 summed:1 equal:2 aware:1 never:4 sampling:1 icml:3 purchase:2 future:3 others:1 djolonga:1 few:1 randomly:4 simultaneously:2 phase:3 consisting:1 arrives:2 pc:5 accurate:2 tuple:4 divide:1 old:1 cevher:1 instance:1 increased:1 cost:3 subset:4 predictor:3 chooses:1 adaptively:2 st:7 explores:1 international:5 stay:2 together:1 presentational:1 cesa:1 containing:1 choose:4 l22:2 opposed:1 li:3 summarized:2 includes:1 matter:1 caused:1 depends:10 exogenous:1 observing:2 hazan:2 start:2 contribution:1 ass:1 accuracy:3 identify:1 lu:1 randomness:1 history:3 whenever:2 definition:1 pp:11 naturally:1 di:6 sampled:2 newly:1 proved:1 hsu:2 knowledge:1 ut:8 dimensionality:1 formalize:1 response:2 formulation:1 lifetime:2 langford:3 continuity:2 perhaps:1 grows:1 effect:5 true:2 hence:6 logaritmic:1 illustrated:1 ind:13 during:2 self:1 covering:1 suboptimality:2 generalized:2 drel:11 instantaneous:1 common:1 pseudocode:2 endpoint:2 exponentially:2 extend:1 access:1 similarity:4 etc:3 kwik:2 dominant:1 hide:1 optimizing:1 belongs:3 success:1 arbitrarily:1 der:1 greater:3 gentile:1 dudik:1 determine:1 maximize:2 ii:7 multiple:1 reduces:1 exceeds:1 faster:1 rti:2 long:1 a1:2 controlled:7 qi:8 prediction:4 variant:1 patient:4 metric:1 arxiv:4 agarwal:1 whereas:1 krause:1 interval:42 else:3 suboptimalities:1 crucial:1 operate:1 effectiveness:1 call:1 ee:1 exceed:1 iii:3 vital:1 enough:2 easy:1 variety:1 suboptimal:2 idea:1 multiclass:2 angeles:2 t0:11 whether:3 action:72 remark:1 tewari:1 amount:2 schapire:2 estimated:6 disjoint:1 popularity:1 diagnosis:1 vol:2 key:1 blood:1 achieving:1 pj:20 utilize:1 vast:2 sum:3 run:6 sti:3 parameterized:1 uncertainty:2 arrive:3 place:1 family:1 home:1 decision:6 orl:38 bound:20 ct:1 guaranteed:1 oracle:3 x2:7 ri:6 sake:1 ucla:2 min:1 conjecture:1 department:2 according:3 alternate:2 smaller:3 kakade:1 making:1 vart:4 restricted:1 pr:4 gartner:1 remains:2 discus:1 nonempty:2 know:7 end:7 available:2 operation:1 tightest:1 observe:9 existence:3 denotes:5 cf:38 graphical:2 exploit:11 especially:1 question:1 costly:1 dependence:1 rt:8 traditional:1 distance:1 separate:1 considers:3 trivial:1 assuming:1 length:7 index:1 mini:2 balance:1 stated:2 policy:4 unknown:5 bianchi:1 recommender:2 discretize:1 observation:11 benchmark:4 finite:1 payoff:1 arbitrary:4 pair:9 required:3 slivkins:1 security:1 california:2 x0r:2 polylogarithmic:1 nti:2 nip:3 address:1 usually:1 including:1 max:2 deactivated:1 power:2 event:2 syed:1 natural:2 indicator:1 minimax:1 older:1 numerous:2 created:1 taming:1 prior:4 epoch:1 l2:4 occupation:1 loss:3 sublinear:4 interesting:3 age:2 incurred:3 pi:46 balancing:1 strehl:1 keeping:1 van:1 feedback:7 dimension:8 calculated:4 computes:3 author:1 made:4 adaptive:1 far:1 social:1 approximate:1 keep:2 dlog2:1 cem:1 sequentially:1 decides:1 active:8 uai:1 conclude:1 assumed:1 tuples:5 xi:11 shwartz:1 continuous:2 learn:4 robust:1 aistats:2 pk:2 main:1 big:1 arrival:4 x1:7 fig:7 rithm:1 candidate:3 lie:1 learns:3 theorem:10 xt:18 explored:4 x:1 exists:3 workshop:1 sequential:1 effectively:1 horizon:1 browsing:1 logarithmic:1 explore:6 forming:1 failed:1 recommendation:3 springer:1 gender:1 corresponds:1 determines:1 goal:1 orabona:1 lipschitz:5 feasible:1 included:1 infinite:1 except:1 determined:1 kearns:2 called:6 total:1 disregard:1 relt:11 exception:1 select:7 l4:1 support:1 scan:1 relevance:31 |
4,839 | 5,381 | Online combinatorial optimization with stochastic
decision sets and adversarial losses
Gergely Neu
Michal Valko
SequeL team, INRIA Lille ? Nord Europe, France
{gergely.neu,michal.valko}@inria.fr
Abstract
Most work on sequential learning assumes a fixed set of actions that are available
all the time. However, in practice, actions can consist of picking subsets of readings from sensors that may break from time to time, road segments that can be
blocked or goods that are out of stock. In this paper we study learning algorithms
that are able to deal with stochastic availability of such unreliable composite actions. We propose and analyze algorithms based on the Follow-The-PerturbedLeader prediction method for several learning settings differing in the feedback
provided to the learner. Our algorithms rely on a novel loss estimation technique
that we call Counting Asleep Times. We deliver regret bounds for our algorithms
for the previously studied full information and (semi-)bandit settings, as well as a
natural middle point between the two that we call the restricted information setting. A special consequence of our results is a significant improvement of the best
known performance guarantees achieved by an efficient algorithm for the sleeping
bandit problem with stochastic availability. Finally, we evaluate our algorithms
empirically and show their improvement over the known approaches.
1
Introduction
In online learning problems [4] we aim to sequentially select actions from a given set in order to
optimize some performance measure. However, in many sequential learning problems we have to
deal with situations when some of the actions are not available to be taken. A simple and wellstudied problem where such situations arise is that of sequential routing [8], where we have to select
every day an itinerary for commuting from home to work so as to minimize the total time spent
driving (or even worse, stuck in a traffic jam). In this scenario, some road segments may be blocked
for maintenance, forcing us to work with the rest of the road network. This problem is isomorphic to
packet routing in ad-hoc computer networks where some links might not be always available because
of a faulty transmitter or a depleted battery. Another important class of sequential decision-making
problems where the decision space might change over time is recommender systems [11]. Here,
some items may be out of stock or some service may not be applicable at some time (e.g., a movie
not shown that day, bandwidth issues in video streaming services). In these cases, the advertiser may
refrain from recommending unavailable items. Other reasons include a distributor being overloaded
with commands or facing shipment problems.
Learning problems with such partial-availability restrictions have been previously studied in the
framework of prediction with expert advice. Freund et al. [7] considered the problem of online
prediction with specialist experts, where some experts? predictions might not be available from time
to time, and the goal of the learner is to minimize regret against the best mixture of experts. Kleinberg
et al. [15] proposed a stronger notion of regret measured against the best ranking of experts and gave
efficient algorithms that work under stochastic assumptions on the losses, referring to this setting as
prediction with sleeping experts. They have also introduced the notion of sleeping bandit problems
where the learner only gets partial feedback about its decisions. They gave an inefficient algorithm
1
for the non-stochastic case, with some hints that it might be difficult to learn efficiently in this
general setting. This was later reaffirmed by Kanade and Steinke [14], who reduce the problem of
PAC learning of DNF formulas to a non-stochastic sleeping experts problem, proving the hardness
of learning in this setup. Despite these negative results, Kanade et al. [13] have shown that there
is still hope to obtain efficient algorithms in adversarial environments, if one introduces a certain
stochastic a assumption on the decision set.
In this paper, we extend the work of Kanade et al. [13] to combinatorial settings where the action
set of the learner is possibly huge, but has a compact representation. We also assume stochastic
action availability: in each decision period, the decision space is drawn from a fixed but unknown
probability distribution independently of the history of interaction between the learner and the environment. The goal of the learner is to minimize the sum of losses associated with its decisions. As
usual in online settings, we measure the performance of the learning algorithm by its regret defined
as the gap between the total loss of the best fixed decision-making policy from a pool of policies
and the total loss of the learner. The choice of this pool, however, is a rather delicate question in our
problem: the usual choice of measuring regret against the best fixed action is meaningless, since not
all actions are available in all time steps. Following Kanade et al. [13] (see also [15]), we consider
the policy space composed of all mappings from decision sets to actions within the respective sets.
We study the above online combinatorial optimization setting under three feedback assumptions.
Besides the full-information and bandit settings considered by Kanade et al. [13], we also consider a
restricted feedback scheme as a natural middle ground between the two by assuming that the learner
gets to know the losses associated only with available actions. This extension (also studied by [15])
is crucially important in practice, since in most cases it is unrealistic to expect that an unavailable
expert would report its loss. Finally, we also consider a generalization of bandit feedback to the
combinatorial case known as semi-bandit feedback.
Our main contributions in this paper are two algorithms called S LEEPING C AT and S LEEPING C ATBANDIT that work in the restricted and semi-bandit information schemes, respectively. The best
known competitor of our algorithms is the BSFPL algorithm of Kanade et al. [13] that works in
two phases. First, an initial phase is dedicated to the estimation of the distribution of the available
actions. Then, in the main phase, BSFPL randomly alternates between exploration and exploitation.
Our technique improves over the FPL-based method of Kanade et al. [13] by removing the costly exploration phase dedicated to estimate the availability probabilities, and also the explicit exploration
steps in their main phase. This is achieved by a cheap alternative loss estimation procedure called
Counting Asleep Times (or CAT) that does not require estimating the distribution of the action sets.
This technique improves the regret bound of [13] ?
after T steps from O(T 4/5 ) to O(T 2/3 ) in their
setting, and also provides a regret guarantee of O( T ) in the restricted setting.1
2
Background
We now give the formal definition of the learning problem. We consider a sequential interaction
scheme between a learner and an environment where in each round t ? [T ] = {1, 2, . . . , T }, the
d
learner has to choose an action Vt from a subset St of a known decision set S ? {0, 1} with
kvk1 ? m for all v ? S. We assume that the environment selects St according to some fixed (but
unknown) distribution P, independently of the interaction history. Unaware of the learner?s decision,
the environment also decides on a loss vector `t ? [0, 1]d that will determine the loss suffered by the
learner, which is of the form Vt> `t . We make no assumptions on how the environment generates
the sequence of loss vectors, that is, we are interested in algorithms that work in non-oblivious (or
adaptive) environments. At the end of each round, the learner receives some feedback based on the
loss vector and the action of the learner. The goal of the learner is pick its actions so as to minimize
the losses it accumulates by the end of the T ?th round. This setup generalizes the setting of online
combinatorial optimization considered by Cesa-Bianchi and Lugosi [5], Audibert et al. [1], where
the decision set is assumed to be fixed throughout the learning procedure. The interaction protocol
is summarized on Figure 1 for reference.
1
While not explicitly proved by Kanade et al. [13], their technique can be extended to work in the restricted
setting, where it can be shown to guarantee a regret of O(T 3/4 ).
2
Parameters:
d
full set of decision vectors S = {0, 1} , number of rounds T , unknown distribution P ? ?2S
For all t = 1, 2, . . . , T repeat
1. The environment draws a set of available actions St ? P and picks a loss vector
`t ? [0, 1]d .
2. The set St is revealed to the learner.
3. Based on its previous observations (and possibly some source of randomness), the
learner picks an action Vt ? St .
4. The learner suffers loss Vt> `t and gets some feedback:
(a) in the full information setting, the learner observes `t ,
(b) in the restricted setting, the learner observes `t,i for all i ? Dt ,
(c) in the semi-bandit setting, the learner observes `t,i for all i such that Vt,i = 1.
Figure 1: The protocol of online combinatorial optimization with stochastic action availability.
We distinguish between three different feedback schemes, the simplest one being the full information
scheme where the loss vectors are completely revealed to the learner at the end of each round. In
the restricted-information scheme, we make a much milder assumption that the learner is informed
about the losses of the available actions. Precisely, we define the set of available components as
Dt = {i ? [d] : ?v ? St : vi = 1}
and assume that the learner can observe the i-th component of the loss vector `t if and only if i ? Dt .
This is a sensible assumption in a number of practical applications, e.g., in sequential routing problems where components are associated with links in a network. Finally, in the semi-bandit scheme,
we assume that the learner only observes losses associated with the components of its own decision,
that is, the feedback is `t,i for all i such that Vt,i = 1. This is the case in in online advertising settings where components of the decision vectors represent customer-ad allocations. The observation
history Ft is defined as the sigma-algebra generated by the actions chosen by the learner and the
decision sets handed out by the environment by the end of round t: Ft = ?(Vt , St , . . . , V1 , S1 ).
The performance of the learner is measured with respect to the best fixed policy (otherwise known
as a choice function in discrete choice theory [16]) of the form ? : 2S ? S. In words, a policy
? ? S? whenever the environment selects action set S.
? The (total expected)
? will pick action ?(S)
regret of the learner is defined as
RT = max
?
T
h
i
X
>
E (Vt ? ?(St )) `t .
(1)
t=1
Note that the above expectation integrates over both the randomness injected by the learner and the
stochastic process generating the decision sets. The attentive reader might notice that this regret
criterion is very similar to that of Kanade et al. [13], who study the setting of prediction with expert
advice (where m = 1) and measure regret against the best fixed ranking of experts. It is actually
easy to show that the optimal policy in their setting belongs to the set of ranking policies, making
our regret definition equivalent to theirs.
3
Loss estimation by Counting Asleep Times
In this section, we describe our method used for estimating unobserved losses that works without
having to explicitly learn the availability distribution P. To explain the concept on a high level, let
us now consider our simpler partial-observability setting, the restricted-information setting. For the
formal treatment of the problem, let us fix any component i ? [d] and define At,i = 1{i?Dt } and
ai = E [At,i |Ft?1 ]. Had we known the observation probability ai , we would be able to estimate
the i?th component of the loss vector `t by `??t,i = (`t,i At,i )/ai , as the quantity `t,i At,i is observable.
It is easy to see that the estimate `??t,i is unbiased by definition ? but, unfortunately, we do not
know ai , so we have no hope to compute it. A simple idea used by Kanade et al. [13] is to devote
3
the first T0 rounds of interaction solely to the purpose of estimating ai by the sample mean a
?i =
PT0
( t=1 At,i )/T0 . While this trick gets the job done, it is obviously wasteful as we have to throw
away all loss observations before the estimates are sufficiently concentrated. 2
We take a much simpler approach based on the observation that the ?asleep-time? of component i
is a geometrically distributed random variable with parameter ai . The asleep-time of component i
starting from time t is formally defined as
Nt,i = min {n > 0 : i ? Dt+n } ,
which is the number of rounds until the next observation of the loss associated with component i.
Using the above definition, we construct our loss estimates as the vector `?t whose i-th component is
`?t,i = `t,i At,i Nt,i .
(2)
It is easy to see that the above loss estimates are unbiased as
1
= `t,i
ai
for any i. We will refer to this loss-estimation method as Counting Asleep Times (CAT).
E [`t,i At,i Nt,i |Ft?1 ] = `t,i E [At,i |Ft?1 ] E [Nt,i |Ft?1 ] = `t,i ai ?
Looking at the definition (2), the attentive reader might worry that the vector `?t depends on future
realizations of the random decision sets and thus could be useless for practical use. However, observe that there is no reason that the learner should use the estimate `?t,i before component i wakes
up in round t + Nt,i ? which is precisely the time when the estimate becomes well-defined. This
suggests a very simple implementation of CAT: whenever a component is not available, estimate its
loss by the last observation from that component! More formally, set
(
`t,i ,
if i ? Dt
`?t,i = ?
`t?1,i , otherwise.
It is easy to see that at the beginning of any round t, the two alternative definitions match for all
components i ? Dt . In the next section, we confirm that this property is sufficient for running our
algorithm.
4
Algorithms & their analyses
For all information settings, we base our learning algorithms on the Follow-the-Perturbed-Leader
(FPL) prediction method of Hannan [9], as popularized by Kalai and Vempala [12]. This algorithm
works by additively perturbing the total estimated loss of each component, and then running an optimization oracle over the perturbed losses to choose the next action. More precisely, our algorithms
b t = Pt `?t and pick the action
maintain the cumulative sum of their loss estimates L
s=1
b t?1 ? Zt ,
Vt = arg min v T ? L
v?St
where Zt is a perturbation vector with independent exponentially distributed components with unit
expectation, generated independently of the history, and ? > 0 is a parameter of the algorithm. Our
algorithms for the different information settings will be instances of FPL that employ different loss
estimates suitable for the respective settings. In the first part of this section, we present the main
tools of analysis that will be used for each resulting method.
As usual for analyzing FPL-based methods [12, 10, 18], we start by defining a hypothetical foree with standard exponential components
caster that uses a time-independent perturbation vector Z
and peeks one step into the future. However, we need an extra trick to deal with the randomness of
the decision set: we introduce the time-independent decision set Se ? P (drawn independently of
the filtration (Ft )t ) and define
bt ? Z
e .
Vet = arg min v T ? L
v?Se
2
Notice that we require ?sufficient concentration? from 1/?
ai and not only from a
?i ! The deviation of such
quantities is rather difficult to control, as demonstrated by the complicated analysis of Kanade et al. [13].
4
Clearly, this forecaster is infeasible as it uses observations from the future. Also observe that
Vet?1 ? Vt given Ft?1 . The following two lemmas show how analyzing this forecaster can help in
establishing the performance of our actual algorithms.
Lemma 1. For any sequence of loss estimates, the expected regret of the hypothetical forecaster
against any fixed policy ? : 2S ? S satisfies
" T
#
T
X
e `?t ? m (log d + 1) .
E
Vet ? ?(S)
?
t=1
The statement is easily proved by applying
the follow-the-leader/be-the-leader
lemma3 (see, e.g., [4,
e
Lemma 3.1]) and using the upper bound E
Z
? log d + 1.
?
The following result can be extracted from the proof of Theorem 1 of Neu and Bart?ok [18].
Lemma 2. For any sequence of nonnegative loss estimates,
i
h
2
T
E (Vet?1 ? Vet )T `?t Ft?1 ? ? E Vet?1
`?t Ft?1 .
In the next subsections, we apply these results to obtain bounds for the three information settings.
4.1
Algorithm for full information
In the simplest setting, we can use `?t = `t , which yields the following theorem:
Theorem 1. Define
(
" T
#
)
X
?
T
LT = max min E
?(St ) `t , 4(log d + 1) .
?
Setting ? =
p
t=1
(log d + 1)/L?T , the regret of FPL in the full information scheme satisfies
q
RT ? 2m 2L?T (log d + 1).
As this result is comparable to the best available bounds for FPL [10, 18] in the full information
setting and a fixed decision set, it reinforces the observation of Kanade et al. [13], who show that the
sleeping experts problem with full information and stochastic availability is no more difficult than
the standard experts problem. The proof of Theorem 1 follows directly from combining Lemmas 1
and 2 with some standard tricks. For completeness, details are provided in Appendix A.
4.2
Algorithm for restricted feedback
In this section, we use the CAT loss estimate defined in Equation (2) as `?t in FPL, and call the
resulting method S LEEPING C AT. The following theorem gives the performance guarantee for this
algorithm.
Pd
Theorem 2. Define Qt =
i=1 E [ Vt,i | i ? Dt ]. The total expected regret of S LEEPING C AT
against the best fixed policy is upper bounded as
T
X
m(log d + 1)
RT ?
+ 2?m
Qt .
?
t=1
e T `?t = E [?(St )T `t ], where we used that `?t is independent of
Proof. We start by observing E ?(S)
e The proof is completed by combining
Se and is an unbiased estimate of `t , and also that St ? S.
this with Lemmas 1 and 2, and the bound
2
T
?
e
E Vt?1 `t Ft?1 ? 2mQt .
The proof of this last statement follows from a tedious calculation that we defer to Appendix B.
3
e allowing the necessary
This lemma can be proved in the current case by virtue of the fixed decision set S,
recursion steps to go through.
5
Below, we provide two ways of further bounding the regret under various assumptions. The first one
provides a universal upper bound that holds without any further assumptions.
p
Corollary 1. Setting ? = (log d + 1)/(2dT ), the regret of S LEEPING C AT against the best fixed
policy is bounded as
p
RT ? 2m 2dT (log d + 1).
The proof follows from the fact that
? Qt ? d no matter what P is. A somewhat surprising feature
of our bound is its scaling with d log d, which is much worse than the logarithmic dependence
exhibited in the full information case. It is easy to see, however, that this bound is not improvable in
general ? see Appendix D for a simple example. The next bound, however, shows that it is possible
to improve this bound by assuming that most components are reliable in some sense, which is the
case in many practical settings.
p
Corollary 2. Assuming ai ? ? for all i, we have Qt ? 1/?, and setting ? = ?(log d + 1)/(2T )
guarantees that the regret of S LEEPING C AT against the best fixed policy is bounded as
s
2T (log d + 1)
RT ? 2m
.
?
4.3
Algorithm for semi-bandit feedback
We now turn our attention to the problem of learning with semi-bandit feedback where the learner
only gets to observe the losses associated with its own decision. Specifically, we assume that the
learner observes all components i of the loss vector such that Vt,i = 1. The extra difficulty in this
setting is that our actions influence the feedback that we receive, so we have to be more careful when
defining our loss estimates. Ideally, we would like to work with unbiased estimates of the form
X
`t,i
?
?
P(S)E
Vt,i Ft?1 , St = S? . (3)
`??t,i = ? Vt,i ,
where
qt,i
= E [ Vt,i | Ft?1 ] =
qt,i
? S
S?2
for all i ? [d]. Unfortunately though, we are in no position to compute these estimates, as this would
require perfect knowledge of the availability distribution P! Thus we have to look for another way
to compute reliable loss estimates. A possible idea is to use
qt,i ? ai = E [ Vt,i | Ft?1 , St ] ? P [i ? Dt ] .
?
qt,i
instead of
in Equation 3 to normalize the observed losses. This choice yields another unbiased
loss estimate as
`t,i Vt,i
`t,i
Vt,i
`t,i
E
F
F
,
S
F
=
E
E
E [ At,i | Ft?1 ] = `t,i ,
(4)
t?1
t?1
t t?1 =
qt,i ai
ai
qt,i
ai
which leaves us with the problem of computing qt,i and ai . While this also seems to be a tough
challenge, we now show to estimate this quantity by generalizing the CAT technique presented in
Section 3.
Besides our trick used for estimating the 1/ai ?s by the random variables Nt,i , we now also have to
face the problem of not being able to find a closed-form expression for the qt,i ?s. Hence, we follow
the geometric resampling approach of Neu and Bart?ok [18] and draw an additional sequence of M
perturbation vectors Zt0 (1), . . . , Zt0 (M ) and use them to compute
n
o
b t?1 ? Z 0 (k)
Vt0 (k) = arg min ? L
t
v?St
for all k ? [M ]. Using these simulated actions, we define
0
Kt,i = min k ? [M ] : Vt,i
(k) = Vt,i ? {M } .
and
`?t,i = `t,i Kt,i Nt,i Vt,i
(5)
`t,i Vt,i
qt,i ai
in expectation, yielding yet
for all i. Setting M = ? makes this expression equivalent to
another unbiased estimator for the losses. Our analysis, however, crucially relies on setting M to
6
a finite value so as to control the variance of the loss estimates. We are not aware of any other
work that achieves a similar variance-reduction effect without explicitly exploring the action space
[17, 6, 5, 3], making this alternative bias-variance tradeoff a unique feature of our analysis. We call
the algorithm resulting from using the loss estimates above S LEEPING C AT BANDIT. The following
theorem gives the performance guarantee for this algorithm.
Pd
Theorem 3. Define Qt = i=1 E [ Vt,i | i ? Dt ]. The total expected regret of S LEEPING C AT BAN DIT against the best fixed policy is bounded as
RT ?
T
X
m(log d + 1)
dT
+ 2?M m
.
Qt +
?
eM
t=1
Proof. First, observe that E `?t,i Ft?1
? `t,i as E [ Kt,i Vt,i | Ft?1 , St ] ? At,i and
e T `?t ? E [?(St )T `t ] by a simiE [ At,i Nt,i | Ft?1 ] = 1 by definition. Thus, we can get E ?(S)
lar argument that we used in the proof of Theorem 2. After yet another long and tedious calculation
(see Appendix C), we can prove
2
T
?
e
E Vt?1 `t Ft?1 ? 2M mQt .
(6)
The proof is concluded by combining this bound with Lemmas 1 and 2 and the upper bound
h
i
d
T
,
E [ VtT `t | Ft?1 ] ? E Vet?1
`?t Ft?1 +
eM
which can be proved by following the proof of Theorem 1 in Neu and Bart?ok [18].
?
2/3
1/3
m(log d+1)
dT
and M = ?1e ? ?2m(log
guarantees that the
Corollary 3. Setting ? =
2dT
d+1)
regret of S LEEPING C AT BANDIT against the best fixed policy is bounded as
RT ? (2mdT )2/3 ? (log d + 1)1/3 .
The proof of the corollary follows from bounding Qt ? d and plugging the parameters into the
bound of Theorem 3. Similarly to the improvement of Corollary 2, it is possible to replace the factor
d2/3 by (d/?)1/3 if we assume that ai ? ? for all i and some ? > 0.
This corollary implies that S LEEPING C AT BANDIT achieves a regret of (2KT )2/3 ? (log K + 1)1/3
in the case when S = [K], that is, in the K-armed sleeping bandit problem considered by Kanade
et al. [13]. This improves their bound of O((KT )4/5 log T ) by a large margin, thanks to the fact
that we did not have to explicitly learn the distribution P.
5
Experiments
In this section we present the empirical evaluation of our algorithms for bandit and semi-bandit
settings, and compare them to its counterparts [13]. We demonstrate that the wasteful exploration of
BSFPL does not only result in worse regret bounds but also degrades its empirical performance.
For the bandit case, we evaluate S LEEPING C AT BANDIT using the same setting as Kanade et al. [13].
We consider an experiment with T = 10, 000 and 5 arms, each of which are available independenly
of each other with probability p. Losses for each arm are constructed as random walks with Gaussian
increments of standard deviation 0.002, initialized uniformly on [0, 1]. Losses outside [0, 1] are truncated. In our first experiment (Figure 2, left), we study the effect of changing p on the performance
of BSFPL and S LEEPING C AT BANDIT. Notice that when p is very low, there are few or no arms to
choose from. In this case the problems are easy by design and all algorithms suffer low regret. As
p increases, the policy space starts to blow up and the problem becomes more difficult. When p approaches one, it collapses into the set of single arms and the problem gets easier again. Observe that
the behavior of S LEEPING C AT BANDIT follows this trend. On the other hand, the performance of
BSFPL steadily decreases with increasing availability. This is due to the explicit exploration rounds
in the main phase of BSFPL, that suffers the loss of the uniform policy scaled by the exploration
probability. The performance of the uniform policy is plotted for reference.
7
sleeping bandits, 5 arms, varying availabity, average over 20 runs
cumulative regret at time T = 10000
BSFPL
0.25
SleepingCat
RandomGuess
0.2
0.15
0.1
0.05
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
availabity
Figure 2: Left: Multi-arm bandits - varying availabilities. Middle: Shortest paths on a 3 ? 3 grid.
Right: Shortest paths on a 10 ? 10 grid.
To evaluate S LEEPING C AT BANDIT in the semi-bandit setting, we consider the shortest path problem on grids of 3 ? 3 and 10 ? 10 nodes, which amounts to 12 and 180 edges respectively. For
each edge, we generate a random-walk loss sequence in the same way as in our first experiment.
In each round t, the learner has to choose a path from the lower left corner to the upper right one
composed from available edges. The individual availability of each edge is sampled with probability
0.9, independently of the others. Whenever an edge gets disconnected from the source, it becomes
unavailable itself, resuling in a quite complicated action-availability distribution. Once a learner
chooses a path, the losses of chosen road segments are revealed and the learner suffers their sum.
Since [13] does not provide a combinatorial version of their approach, we compare against C OMB BSFPL, a straightforward extension of BSFPL. As in BSFPL, we dedicate an initial phase to estimate the availabilities of each component, requiring d oracle calls per step. In the main phase, we
follow BSFPL and alternate between exploration and exploitation. In exploration rounds, we test
for the reachability of a randomly sampled edge and update the reward estimates as in BSFPL.
Figure 2 (middle and right) shows the performance of C OMB BSFPL and S LEEPING C AT BANDIT
for a fixed loss sequence, averaged over 20 samples of the component availabilities. We also plot the
performance of a random policy that follows the perturbed leader with all-zero loss estimates. First
observe that the initial exploration phase sets back the performance of C OMB BSFPL significantly.
The second drawback of C OMB BSFPL is the explicit separation of exploration and the exploitation
rounds. This drawback is far more apparent when the number of components increases, as it is the
case for the 10 ? 10 grid graph with 180 components. As C OMB BSFPL only estimates the loss of
one edge per exploration step, sampling each edge as few as 50 times eats up 9, 000 rounds from
the available 10, 000. S LEEPING C AT BANDIT does not suffer from this problem as it uses all its
observations in constructing the loss estimates.
6
Conclusions & future work
In this paper, we studied the problem of online combinatorial optimization with changing decision
sets. Our main contribution is a novel loss-estimation technique that enabled us to prove strong
regret bounds under various partial-feedback schemes. In particular, our results largely improve on
the best known results for the sleeping bandit problem [13], which suffers large losses from both
from an initial exploration phase and from explicit exploration rounds in the main phase. These
findings are also supported by our experiments.
?
Still, one might ask if it is possible to efficiently achieve a regret of order T under semi-bandit
feedback. While the E XP 4 algorithm of Auer et al. [2] can be used to obtain such regret guarantee,
running this algorithm is out of question as its time and space complexity can be double-exponential
in d (see also the comments in [15]). Had we had access to the loss estimates (3), we would be able
to control the regret of FPL as the term on the right hand side of
? Equation (6) could be replaced
by md, which is sufficient for obtaining a regret bound of O(m dT log d). In fact, it seems that
learning in the bandit setting requires significantly more knowledge about P than the knowledge of
the ai ?s. The question if we can extend the CAT technique to estimate all the relevant quantities of
P is an interesting problem left for future investigation.
Acknowledgements The research presented in this paper was supported by French Ministry
of Higher Education and Research, by European Community?s Seventh Framework Programme
(FP7/2007-2013) under grant agreement no 270327 (CompLACS), and by FUI project Herm`es.
8
References
[1] Audibert, J. Y., Bubeck, S., and Lugosi, G. (2014). Regret in online combinatorial optimization.
Mathematics of Operations Research. to appear.
[2] Auer, P., Cesa-Bianchi, N., Freund, Y., and Schapire, R. E. (2002). The nonstochastic multiarmed bandit problem. SIAM J. Comput., 32(1):48?77.
[3] Bubeck, S., Cesa-Bianchi, N., and Kakade, S. M. (2012). Towards minimax policies for online
linear optimization with bandit feedback. In COLT 2012, pages 1?14.
[4] Cesa-Bianchi, N. and Lugosi, G. (2006). Prediction, Learning, and Games. Cambridge University Press, New York, NY, USA.
[5] Cesa-Bianchi, N. and Lugosi, G. (2012). Combinatorial bandits. Journal of Computer and
System Sciences, 78:1404?1422.
[6] Dani, V., Hayes, T. P., and Kakade, S. (2008). The price of bandit information for online
optimization. In NIPS-20, pages 345?352.
[7] Freund, Y., Schapire, R., Singer, Y., and Warmuth, M. (1997). Using and combining predictors
that specialize. In Proceedings of the 29th Annual ACM Symposium on the Theory of Computing,
pages 334?343. ACM Press.
[8] Gy?orgy, A., Linder, T., Lugosi, G., and Ottucs?ak, Gy.. (2007). The on-line shortest path problem
under partial monitoring. Journal of Machine Learning Research, 8:2369?2403.
[9] Hannan, J. (1957). Approximation to bayes risk in repeated play. Contributions to the theory of
games, 3:97?139.
[10] Hutter, M. and Poland, J. (2004). Prediction with expert advice by following the perturbed
leader for general weights. In ALT, pages 279?293.
[11] Jannach, D., Zanker, M., Felfernig, A., and Friedrich, G. (2010). Recommender Systems: An
Introduction. Cambridge University Press.
[12] Kalai, A. and Vempala, S. (2005). Efficient algorithms for online decision problems. Journal
of Computer and System Sciences, 71:291?307.
[13] Kanade, V., McMahan, H. B., and Bryan, B. (2009). Sleeping experts and bandits with stochastic action availability and adversarial rewards. In AISTATS 2009, pages 272?279.
[14] Kanade, V. and Steinke, T. (2012). Learning hurdles for sleeping experts. In Proceedings of
the 3rd Innovations in Theoretical Computer Science Conference (ITCS 12), pages 11?18. ACM.
[15] Kleinberg, R. D., Niculescu-Mizil, A., and Sharma, Y. (2008). Regret bounds for sleeping
experts and bandits. In COLT 2008, pages 425?436.
[16] Koshevoy, G. A. (1999). Choice functions and abstract convex geometries. Mathematical
Social Sciences, 38(1):35?44.
[17] McMahan, H. B. and Blum, A. (2004). Online geometric optimization in the bandit setting
against an adaptive adversary. In COLT 2004, pages 109?123.
[18] Neu, G. and Bart?ok, G. (2013). An efficient algorithm for learning with semi-bandit feedback.
In ALT 2013, pages 234?248.
9
| 5381 |@word exploitation:3 version:1 middle:4 stronger:1 seems:2 tedious:2 d2:1 additively:1 crucially:2 forecaster:3 pick:5 reduction:1 initial:4 pt0:1 current:1 michal:2 nt:8 surprising:1 yet:2 cheap:1 plot:1 update:1 bart:4 resampling:1 leaf:1 item:2 warmuth:1 beginning:1 provides:2 completeness:1 node:1 simpler:2 mathematical:1 constructed:1 symposium:1 prove:2 specialize:1 distributor:1 introduce:1 expected:4 hardness:1 behavior:1 vtt:1 multi:1 actual:1 armed:1 increasing:1 becomes:3 provided:2 estimating:4 bounded:5 project:1 what:1 informed:1 differing:1 unobserved:1 finding:1 guarantee:8 every:1 hypothetical:2 scaled:1 control:3 unit:1 grant:1 appear:1 before:2 service:2 consequence:1 despite:1 accumulates:1 ak:1 analyzing:2 establishing:1 solely:1 path:6 lugosi:5 inria:2 might:7 studied:4 suggests:1 collapse:1 averaged:1 practical:3 unique:1 omb:5 practice:2 regret:31 procedure:2 universal:1 empirical:2 significantly:2 composite:1 word:1 road:4 get:8 faulty:1 risk:1 applying:1 influence:1 optimize:1 restriction:1 equivalent:2 customer:1 demonstrated:1 go:1 attention:1 starting:1 independently:5 straightforward:1 convex:1 estimator:1 enabled:1 proving:1 notion:2 increment:1 pt:1 play:1 us:3 agreement:1 trick:4 trend:1 observed:1 ft:21 decrease:1 observes:5 environment:10 pd:2 complexity:1 reward:2 ideally:1 battery:1 segment:3 algebra:1 deliver:1 learner:35 completely:1 easily:1 stock:2 cat:6 various:2 dnf:1 describe:1 mqt:2 outside:1 whose:1 quite:1 apparent:1 otherwise:2 itself:1 online:14 obviously:1 hoc:1 sequence:6 propose:1 interaction:5 fr:1 relevant:1 combining:4 realization:1 achieve:1 normalize:1 double:1 generating:1 perfect:1 spent:1 help:1 measured:2 qt:16 job:1 strong:1 throw:1 reachability:1 implies:1 drawback:2 stochastic:12 exploration:13 packet:1 routing:3 jam:1 education:1 require:3 fix:1 generalization:1 investigation:1 extension:2 exploring:1 hold:1 sufficiently:1 considered:4 ground:1 mapping:1 driving:1 achieves:2 purpose:1 estimation:6 integrates:1 applicable:1 combinatorial:10 tool:1 hope:2 dani:1 clearly:1 sensor:1 always:1 gaussian:1 aim:1 eats:1 rather:2 kalai:2 varying:2 command:1 kvk1:1 corollary:6 improvement:3 transmitter:1 adversarial:3 sense:1 milder:1 streaming:1 niculescu:1 bt:1 bandit:38 france:1 selects:2 interested:1 issue:1 arg:3 colt:3 special:1 construct:1 aware:1 having:1 once:1 sampling:1 lille:1 look:1 future:5 report:1 others:1 hint:1 oblivious:1 zanker:1 employ:1 few:2 randomly:2 composed:2 individual:1 itinerary:1 phase:11 replaced:1 geometry:1 delicate:1 maintain:1 zt0:2 huge:1 evaluation:1 wellstudied:1 introduces:1 mixture:1 yielding:1 kt:5 edge:8 partial:5 necessary:1 respective:2 walk:2 initialized:1 plotted:1 theoretical:1 hutter:1 handed:1 instance:1 measuring:1 deviation:2 subset:2 uniform:2 predictor:1 seventh:1 perturbed:4 chooses:1 referring:1 st:17 thanks:1 siam:1 sequel:1 picking:1 pool:2 complacs:1 fui:1 gergely:2 again:1 cesa:5 choose:4 possibly:2 worse:3 corner:1 expert:16 inefficient:1 blow:1 gy:2 summarized:1 availability:16 matter:1 explicitly:4 ranking:3 ad:2 audibert:2 vi:1 later:1 break:1 depends:1 closed:1 analyze:1 traffic:1 observing:1 start:3 bayes:1 complicated:2 vt0:1 defer:1 contribution:3 minimize:4 improvable:1 variance:3 who:3 efficiently:2 largely:1 yield:2 itcs:1 advertising:1 monitoring:1 randomness:3 history:4 explain:1 suffers:4 whenever:3 neu:6 definition:7 against:12 competitor:1 attentive:2 steadily:1 associated:6 proof:11 sampled:2 proved:4 treatment:1 ask:1 subsection:1 knowledge:3 improves:3 actually:1 back:1 auer:2 worry:1 ok:4 higher:1 dt:16 day:2 follow:5 done:1 though:1 until:1 hand:2 receives:1 french:1 lar:1 usa:1 effect:2 concept:1 unbiased:6 requiring:1 counterpart:1 hence:1 perturbedleader:1 deal:3 round:16 game:2 criterion:1 demonstrate:1 dedicated:2 novel:2 empirically:1 perturbing:1 exponentially:1 extend:2 theirs:1 significant:1 blocked:2 refer:1 multiarmed:1 cambridge:2 ai:19 rd:1 grid:4 mathematics:1 similarly:1 had:3 access:1 europe:1 base:1 own:2 belongs:1 forcing:1 scenario:1 certain:1 refrain:1 vt:26 ministry:1 additional:1 somewhat:1 determine:1 shortest:4 advertiser:1 period:1 sharma:1 semi:11 full:10 hannan:2 match:1 calculation:2 long:1 plugging:1 prediction:9 maintenance:1 expectation:3 represent:1 achieved:2 sleeping:11 receive:1 background:1 hurdle:1 wake:1 source:2 suffered:1 peek:1 concluded:1 extra:2 rest:1 meaningless:1 exhibited:1 comment:1 tough:1 call:5 counting:4 depleted:1 revealed:3 easy:6 gave:2 nonstochastic:1 bandwidth:1 reduce:1 observability:1 idea:2 tradeoff:1 t0:2 expression:2 suffer:2 york:1 action:30 se:3 amount:1 concentrated:1 simplest:2 dit:1 generate:1 schapire:2 notice:3 estimated:1 per:2 reinforces:1 bryan:1 discrete:1 blum:1 drawn:2 wasteful:2 changing:2 v1:1 graph:1 geometrically:1 sum:3 run:1 injected:1 throughout:1 reader:2 separation:1 home:1 draw:2 decision:26 appendix:4 scaling:1 comparable:1 bound:19 distinguish:1 oracle:2 nonnegative:1 annual:1 precisely:3 kleinberg:2 generates:1 argument:1 min:6 vempala:2 according:1 alternate:2 popularized:1 disconnected:1 em:2 kakade:2 making:4 s1:1 restricted:9 taken:1 equation:3 previously:2 turn:1 singer:1 know:2 fp7:1 end:4 available:15 generalizes:1 operation:1 apply:1 observe:7 away:1 specialist:1 alternative:3 assumes:1 running:3 include:1 completed:1 reaffirmed:1 question:3 quantity:4 mdt:1 degrades:1 costly:1 rt:7 usual:3 concentration:1 dependence:1 devote:1 md:1 link:2 simulated:1 sensible:1 reason:2 ottucs:1 assuming:3 besides:2 useless:1 innovation:1 difficult:4 setup:2 unfortunately:2 statement:2 nord:1 sigma:1 negative:1 filtration:1 implementation:1 design:1 zt:2 policy:18 unknown:3 bianchi:5 recommender:2 upper:5 observation:10 allowing:1 commuting:1 finite:1 truncated:1 situation:2 extended:1 looking:1 team:1 leeping:16 defining:2 perturbation:3 community:1 overloaded:1 introduced:1 friedrich:1 nip:1 able:4 adversary:1 below:1 reading:1 challenge:1 max:2 reliable:2 video:1 unrealistic:1 suitable:1 natural:2 rely:1 difficulty:1 valko:2 recursion:1 arm:6 minimax:1 scheme:9 improve:2 movie:1 mizil:1 fpl:8 poland:1 geometric:2 acknowledgement:1 freund:3 loss:57 expect:1 interesting:1 allocation:1 facing:1 sufficient:3 xp:1 ban:1 repeat:1 last:2 supported:2 infeasible:1 formal:2 bias:1 side:1 steinke:2 face:1 distributed:2 feedback:18 cumulative:2 unaware:1 stuck:1 adaptive:2 programme:1 far:1 dedicate:1 social:1 compact:1 observable:1 unreliable:1 confirm:1 sequentially:1 decides:1 hayes:1 assumed:1 recommending:1 herm:1 leader:5 vet:7 kanade:16 learn:3 obtaining:1 unavailable:3 orgy:1 european:1 constructing:1 protocol:2 did:1 aistats:1 main:8 bounding:2 arise:1 repeated:1 advice:3 ny:1 position:1 explicit:4 exponential:2 comput:1 mcmahan:2 formula:1 removing:1 theorem:11 pac:1 alt:2 virtue:1 consist:1 sequential:6 margin:1 gap:1 easier:1 generalizing:1 lt:1 logarithmic:1 bubeck:2 satisfies:2 relies:1 extracted:1 asleep:6 acm:3 goal:3 careful:1 towards:1 replace:1 price:1 change:1 specifically:1 uniformly:1 lemma:8 total:7 called:2 isomorphic:1 e:1 select:2 formally:2 linder:1 evaluate:3 |
4,840 | 5,382 | Multilabel Structured Output Learning with Random
Spanning Trees of Max-Margin Markov Networks
Hongyu Su
Helsinki Institute for Information Technology
Dept of Information and Computer Science
Aalto University, Finland
[email protected]
Mario Marchand
D?epartement d?informatique et g?enie logiciel
Universit?e Laval
Qu?ebec (QC), Canada
[email protected]
Emilie Morvant?
LaHC, UMR CNRS 5516
Univ. of St-Etienne, France
[email protected]
Juho Rousu
Helsinki Institute for Information Technology
Dept of Information and Computer Science
Aalto University, Finland
[email protected]
John Shawe-Taylor
Department of Computer Science
University College London
London, UK
[email protected]
Abstract
We show that the usual score function for conditional Markov networks can be
written as the expectation over the scores of their spanning trees. We also show
that a small random sample of these output trees can attain a significant fraction
of the margin obtained by the complete graph and we provide conditions under
which we can perform tractable inference. The experimental results confirm that
practical learning is scalable to realistic datasets using this approach.
1
Introduction
Finding an hyperplane that minimizes the number of misclassifications is N P-hard. But the support
vector machine (SVM) substitutes the hinge for the discrete loss and, modulo a margin assumption,
can nonetheless efficiently find a hyperplane with a guarantee of good generalization. This paper
investigates whether the problem of inference over a complete graph in structured output prediction
can be avoided in an analogous way based on a margin assumption.
We first show that the score function for the complete output graph can be expressed as the expectation over the scores of random spanning trees. A sampling result then shows that a small random
sample of these output trees can attain a significant fraction of the margin obtained by the complete
graph. Together with a generalization bound for the sample of trees, this shows that we can obtain
good generalization using the average scores of a sample of trees in place of the complete graph.
We have thus reduced the intractable inference problem to a convex optimization not dissimilar to
a SVM. The key inference problem to enable learning with this ensemble now becomes finding the
maximum violator for the (finite sample) average tree score. We then provide the conditions under
which the inference problem is tractable. Experimental results confirm this prediction and show that
?
Most of this work was carried out while E. Morvant was affiliated with IST Austria, Klosterneurburg.
1
practical learning is scalable to realistic datasets using this approach with the resulting classification
accuracy enhanced over more naive ways of training the individual tree score functions.
The paper aims at exploring the potential ramifications of the random spanning tree observation
both theoretically and practically. As such, we think that we have laid the foundations for a fruitful
approach to tackle the intractability of inference in a number of scenarios. Other attractive features
are that we do not require knowledge of the output graph?s structure, that the optimization is convex,
and that the accuracy of the optimization can be traded against computation. Our approach is firmly
rooted in the maximum margin Markov network analysis [1]. Other ways to address the intractability
of loopy graph inference have included using approximate MAP inference with tree-based and LP
relaxations [2], semi-definite programming convex relaxations [3], special cases of graph classes for
which inference is efficient [4], use of random tree score functions in heuristic combinations [5].
Our work is not based on any of these approaches, despite superficial resemblances to, e.g., the
trees in tree-based relaxations and the use of random trees in [5]. We believe it represents a distinct
approach to a fundamental problem of learning and, as such, is worthy of further investigation.
2
Definitions and Assumptions
We consider supervised learning problems where the input space X is arbitrary and the output space
def
Y consists of the set of all `-dimensional multilabel vectors (y1 , . . . , y` ) = y where each yi ?
{1, . . . , ri } for some finite positive integer ri . Each example (x, y) ? X ?Y is mapped to a joint
feature vector ? (x, y). Given a weight vector w in the space of joint feature vectors, the predicted
output yw (x) at input x ? X , is given by the output y maximizing the score F (w, x, y), i.e.,
def
yw (x) = argmax F (w, x, y)
;
where
def
F (w, x, y) = hw, ? (x, y)i ,
(1)
y?Y
and where h?, ?i denotes the inner product in the joint feature space. Hence, yw (x) is obtained by
solving the so-called inference problem, which is known to be N P-hard for many output feature
maps [6, 7]. Consequently, we aim at using an output feature map for which the inference problem can be solved by a polynomial time algorithm such as dynamic programming. The margin
?(w, x, y) achieved by predictor w at example (x, y) is defined as,
def
?(w, x, y) =
min [F (w, x, y) ? F (w, x, y0 )] .
y0 6=y
We consider the case where the feature map ? is a potential function for a Markov network defined
by a complete graph G with ` nodes and `(` ? 1)/2 undirected edges. Each node i of G represents
an output variable yi and there exists an edge (i, j) of G for each pair (yi , yj ) of output variables.
For any example (x, y) ? X ? Y, its joint feature vector is given by
? (x, y) = ? i,j (x, yi , yj ) (i,j)?G = ? (x) ? ? i,j (yi , yj ) (i,j)?G ,
where ? is the Kronecker product. Hence, any predictor w can be written as w = (wi,j )(i,j)?G
where wi,j is w?s weight on ? i,j (x, yi , yj ). Therefore, for any w and any (x, y), we have
X
X
F (w, x, y) = hw, ? (x, y)i =
hwi,j , ? i,j (x, yi , yj )i =
Fi,j (wi,j , x, yi , yj ) ,
(i,j)?G
(i,j)?G
where we denote by Fi,j (wi,j , x, yi , yj ) = hwi,j , ? i,j (x, yi , yj ) the score of labeling the edge (i, j)
by (yi , yj ) given input x.
For any vector a, let kak denote its L2 norm. Throughout the paper, we make the assumption that
?(x, y)k = 1 for all (x, y)
we have a normalized joint feature space such that k?
? X ? Y and
?i,j (x, yi , yj )k is the same for all (i, j) ? G. Since the complete graph G has 2` edges, it follows
k?
?1
?i,j (x, yi , yj )k2 = 2`
that k?
for all (i, j) ? G.
def
We also have a training set S = {(x1 , y1 ), . . . , (xm , ym )} where each example is generated independently according to some unknown distribution D. Mathematically, we do not assume the
existence of a predictor w achieving some positive margin ?(w, x, y) on each (x, y) ? S. Indeed,
2
for some S, there might not exist any w where ?(w, x, y) > 0 for all (x, y) ? S. However, the
generalization guarantee will be best when w achieves a large margin on most training points.
Given any ? > 0, and any (x, y) ? X ? Y, the hinge loss (at scale ?) incurred on (x, y) by a unit L2
norm predictor w that achieves a (possibly negative) margin ?(w, x, y) is given by L? (?(w, x, y)),
def
where the so-called hinge loss function L? is defined as L? (s) = max (0, 1 ? s/?) ?s ? R . We
def
will also make use of the ramp loss function A? defined by A? (s) = min(1, L? (s)) ?s ? R .
The proofs of all the rigorous results of this paper are provided in the supplementary material.
3
Superposition of Random Spanning Trees
Given a complete graph G of ` nodes (representing the Markov network), let S(G) denote the set of
all ``?2 spanning trees of G. Recall that each spanning tree of G has ` ? 1 edges. Hence, forany
edge (i, j) ? G, the number of trees in S(G) covering that edge (i, j) is given by ``?2 (`?1)/ 2` =
(2/`)``?2 . Therefore, for any function f of the edges of G we have
X X
2 X
f ((i, j)) = ``?2
f ((i, j)) .
`
T ?S(G) (i,j)?T
(i,j)?G
Given any spanning tree T of G and given any predictor w, let wT denote the projection of w on the
edges of T . Namely, (wT )i,j = wi,j if (i, j) ? T , and (wT )i,j = 0 otherwise. Let us also denote
?T (x, y))i,j = ? i,j (x, yi , yj )
by ? T (x, y), the projection of ? (x, y) on the edges of T . Namely, (?
?1
?T (x, y))i,j = 0 otherwise. Recall that k?
?i,j (x, yi , yj )k2 = 2`
if (i, j) ? T , and (?
?(i, j) ? G.
Thus, for all (x, y) ? X ? Y and for all T ? S(G), we have
X
`?1
2
?T (x, y)k2 =
?i,j (x, yi , yj )k2 = ` = .
k?
k?
`
2
(i,j)?T
We now establish how F (w, x, y) can be written as an expectation over all the spanning trees of G.
def
def
? =
?T k. Let U(G) denote the uniform distribution on
? T = wT /kwT k, ?
Lemma 1. Let w
? T /k?
T
S(G). Then, we have
r
`
def
?
? T , ? T (x, y)i, where aT =
kwT k .
F (w, x, y) =
E aT hw
2
T ?U (G)
Moreover, for any w such that kwk = 1, we have:
E
T ?U (G)
a2T = 1, and
E
T ?U (G)
aT ? 1 .
def
Let T = {T1 , . . . , Tn } be a sample of n spanning trees of G where each Ti is sampled independently
according to U(G). Given any unit L2 norm predictor w on the complete graph G, our task is to
investigate how the margins ?(w, x, y), for each (x, y) ? X ?Y, will be modified if we approximate
the (true) expectation over all spanning trees by an average over the sample T .
For this task, we consider any (x, y) and any w of unit L2 norm. Let FT (w, x, y) denote the
estimation of F (w, x, y) on the tree sample T ,
n
1X
def
? (x, y)i ,
? Ti , ?
FT (w, x, y) =
aT hw
Ti
n i=1 i
and let ?T (w, x, y) denote the estimation of ?(w, x, y) on the tree sample T ,
def
?T (w, x, y) = min
[FT (w, x, y) ? FT (w, x, y0 )] .
0
y 6=y
The following lemma states how ?T relates to ?.
Lemma 2. Consider any unit L2 norm predictor w on the complete graph G that achieves a margin
of ?(w, x, y) for each (x, y) ? X ? Y, then we have
?T (w, x, y) ? ?(w, x, y) ? 2 ?(x, y) ? X ? Y ,
whenever we have |FT (w, x, y) ? F (w, x, y)| ? for all (x, y) ? X ? Y.
3
Lemma 2 has important consequences whenever |FT (w, x, y) ? F (w, x, y)| ? for all (x, y) ?
X ? Y. Indeed, if w achieves a hard margin ?(w, x, y) ? ? > 0 for all (x, y) ? S, then we have
that w also achieves a hard margin of ?T (w, x, y) ? ? ? 2 on each (x, y) ? S when using the tree
sample T instead of the full graph G. More generally, if w achieves a ramp loss of A? (?(w, x, y))
for each (x, y) ? X ? Y, then w achieves a ramp loss of A? (?T (w, x, y)) ? A? (?(w, x, y) ? 2)
for all (x, y) ? X ? Y when using the tree sample T instead of the full graph G. This last property
follows directly from the fact that A? (s) is a non-increasing function of s.
?
The next lemma tells us that, apart from a slow ln2 ( n) dependence, a sample of n ? ?(`2 /2 )
spanning trees is sufficient to assure that the condition of Lemma 2 holds with high probability for all
(x, y) ? X ? Y. Such a fast convergence rate was made possible by using PAC-Bayesian methods
which, in our case, prevented us of using the union bound over all possible y ? Y.
Lemma 3. Consider any > 0 and any unit L2 norm predictor w for the complete graph G acting
on a normalized joint feature space. For any ? ? (0, 1), let
? 2
`2 1
1 8 n
n ? 2
+ ln
.
(2)
16 2
?
Then with probability of at least 1 ? ?/2 over all samples T generated according to U(G)n , we
have, simultaneously for all (x, y) ? X ? Y, that |FT (w, x, y) ? F (w, x, y)| ? .
def
? T1 , . . . , w
? Tn }
Given a sample T of n spanning trees of G, we now consider an arbitrary set W = {w
? (x, y).
? Ti operates on a unit L2 norm feature vector ?
of unit L2 norm weight vectors where each w
Ti
For any T and any such set W, we consider an arbitrary unit L2 norm conical combination of each
Pn
def
weight in W realized by a n-dimensional weight vector q = (q1 , . . . , qn ), where i=1 qi2 = 1 and
each qi ? 0. Given any (x, y) and any T , we define the score FT (W, q, x, y) achieved on (x, y)
by the conical combination (W, q) on T as
n
X
def 1
? (x, y)i ,
? Ti , ?
qi hw
FT (W, q, x, y) = ?
Ti
n i=1
(3)
?
where
n denominator ensures
Pthe
? that we always have FT (W, q, x, y) ? 1 in view of the fact
n
that i=1 qi can be as large as n. Note also that FT (W, q, x, y) is the score of the feature vector
obtained by the concatenation of all the weight vectors in W (and weighted by q) acting on a feature
? multiplied by 1/?n. Hence, given T , we define the
vector obtained by concatenating each ?
Ti
margin ?T (W, q, x, y) achieved on (x, y) by the conical combination (W, q) on T as
def
?T (W, q, x, y) = min
[FT (W, q, x, y) ? FT (W, q, x, y0 )] .
0
y 6=y
(4)
For any unit L2 norm predictor w that achieves a margin of ?(w, x, y) for all (x, y) ? X ? Y, we
now show that there exists, with high probability, a unit L2 norm conical combination (W, q) on T
achieving margins that are not much smaller than ?(w, x, y).
Theorem 4. Consider any unit L2 norm predictor w for the complete graph G, acting on a normalized joint feature space, achieving a margin of ?(w, x, y) for each (x, y) ? X ? Y. Then for any
> 0, and any n satisfying Lemma 3, for any ? ? (0, 1], with probability of at least 1 ? ? over all
samples T generated according to U(G)n , there exists a unit L2 norm conical combination (W, q)
on T such that, simultaneously for all (x, y) ? X ? Y, we have
?T (W, q, x, y) ? ?
1
[?(w, x, y) ? 2] .
1+
From Theorem 4, and since A? (s) is a non-increasing function of s, it follows that, with probability at least 1 ? ? over the random draws of T ? U(G)n , there exists (W, q) on T such that,
simultaneously for all ?(x, y) ? X ? Y, for any n satisfying Lemma 3 we have
A? (?T (W, q, x, y)) ? A? [?(w, x, y) ? 2] (1 + )?1/2 .
Hence, instead of searching for a predictor w for the complete graph G that achieves a small expected ramp loss E(x,y)?D A? (?(w, x, y), Theorem 4 tells us that we can settle the search for a
4
unit L2 norm conical combination (W, q) on a sample T of randomly-generated spanning trees of
G that achieves small E(x,y)?D A? (?T (W, q, x, y)). But recall that ?T (W, q, x, y)) is the margin
of a weight vector obtained by the concatenation of all the weight vectors in W (weighted by q) on
? ?
a feature vector obtained by the concatenation of the n feature vectors (1/ n)?
Ti . It thus follows
that any standard risk bound for the SVM applies directly to E(x,y)?D A? (?T (W, q, x, y)). Hence,
by adapting the SVM risk bound of [8], we have the following result.
Theorem 5. Consider any sample T of n spanning trees of the complete graph G. For any ? > 0
and any 0 < ? ? 1, with probability of at least 1 ? ? over the random draws of S ? Dm ,
simultaneously for all unit L2 norm conical combinations (W, q) on T , we have
r
m
1 X ?
2
ln(2/?)
?
A (?T (W, q, xi , yi )) + ? + 3
.
E A (?T (W, q, x, y)) ?
m i=1
2m
? m
(x,y)?D
Hence, according to this theorem, the conical combination (W, q) having the best generalization
guarantee is the one which minimizes the sum of the first two terms on the right hand side of
the inequality. Note that the theorem is still valid if we replace, in the empirical risk term, the
non-convex ramp loss A? by the convex hinge loss L? . This provides the theoretical basis of the
proposed optimization problem for learning (W, q) on the sample T .
4
A L2 -Norm Random Spanning Tree Approximation Approach
def
If we introduce the usual
variables ?k = ? ? L? (?T (W, q, xk , yk ), Theorem 5 suggests that
Pslack
m
1
we should minimize ? k=1 ?k for some fixed margin value ? > 0. Rather than performing this
task for several values of ?, we show in the supplementary material that we can, equivalently, solve
the following optimization problem for several values of C > 0.
Definition 6. Primal L2 -norm Random Tree Approximation.
m
n
min
wTi ,?k
s.t.
X
1X
2
?k
||wTi ||2 + C
2 i=1
k=1
n
X
? (xk , yk )i ? max
hwTi , ?
Ti
y6=yk
i=1
n
X
? (xk , y)i ? 1 ? ?k ,
hwTi , ?
Ti
i=1
?k ? 0 , ? k ? {1, . . . , m},
where {wTi |Ti ? T } are the feature weights to be learned on each tree, ?k is the margin slack
allocated for each xk , and C is the slack parameter that controls the amount of regularization.
This primal form has the interpretation of maximizing the joint margins from individual trees between (correct) training examples and all the other (incorrect) examples.
The key for the efficient optimization is solving the ?argmax? problem efficiently. In particular, we
note that the space of all multilabels is exponential in size, thus forbidding exhaustive enumeration
over it. In the following, we show how exact inference over a collection T of trees can be implemented in ?(Kn`) time per data point, where K is the smallest number such that the average score
Pn
def
? (x, y)i.
of the K?th best multilabel for each tree of T is at most FT (x, y) = n1 i=1 hwTi , ?
Ti
Whenever K is polynomial in the number of labels, this gives us exact polynomial-time inference
over the ensemble of trees.
4.1
Fast inference over a collection of trees
It is well known that the exact solution to the inference problem
def
? (x, y)i,
? Ti (x) = argmax FwTi (x, y) = argmax hwTi , ?
y
Ti
y?Y
(5)
y?Y
on an individual tree Ti can be obtained in ?(`) time by dynamic programming. However, there is
? Ti of Equation (5) is also a maximizer of FT . In practice, y
? Ti
no guarantee that the maximizer y
5
can differ for each spanning tree Ti ? T . Hence, instead of using only the best scoring multil? Ti from each individual Ti ? T , we consider the set of the K highest scoring multilabels
abel y
? Ti ,K } of FwTi (x, y). In the supplementary material we describe a dynamic
YTi ,K = {?
yTi ,1 , ? ? ? , y
programming to find the K highest multilabels in ?(K`) time. Running this algorithm for all of the
trees gives us a candidate set of ?(Kn) multilabels YT ,K = YT1 ,K ? ? ? ? ? YTn ,K . We now state a
key lemma that will enable us to verify if the candidate set contains the maximizer of FT .
?
Lemma 7. Let yK
= argmax FT (x, y) be the highest scoring multilabel in YT ,K . Suppose that
y?YT ,K
n
1X
def
FwTi (x, yTi ,K ) = ?x (K).
n i=1
?
It follows that FT (x, yK
) = maxy?Y FT (x, y).
?
FT (x, yK
)?
?
We can use any K satisfying the lemma as the length of K-best lists, and be assured that yK
is a
maximizer of FT .
We now examine the conditions under which the highest scoring multilabel is present in our candef
def
? = yw (x) =
didate set YT ,K with high probability. For any x ? X and any predictor w, let y
argmax F (w, x, y) be the highest scoring multilabel in Y for predictor w on the complete graph G.
y?Y
def
For any y ? Y, let KT (y) be the rank of y in tree T and let ?T (y) = KT (y)/|Y| be the normalized
rank of y in tree T . We then have 0 < ?T (y) ? 1 and ?T (y0 ) = miny?Y ?T (y) whenever y0 is a
highest scoring multilabel in tree T . Since w and x are arbitrary and fixed, let us drop them momendef
def
tarily from the notation and let F (y) = F (w, x, y), and FT (y) = FwT (x, y). Let U(Y) denote the
def
def
uniform distribution of multilabels on Y. Then, let ?T = Ey?U (Y) FT (y) and ? = ET ?U (G) ?T .
Let T ? U(G)n be a sample of n spanning trees of G. Since the scoring function FT of each tree
T of G is bounded in absolute value, it follows that FT is a ?T -sub-Gaussian random variable for
some ?T > 0. We now show that, with high probability, there exists a tree T ? T such that ?T (?
y)
def
is decreasing exponentially rapidly with (F (?
y) ? ?)/?, where ? 2 = ET ?U (G) ?T2 .
Lemma 8. Let the scoring function FT of each spanning tree of G be a ?T -sub-Gaussian random
variable under the uniform distribution of labels; i.e., for each T on G, there exists ?T > 0 such
that for any ? > 0 we have
?2
def
Let ? 2 =
E
T ?U (G)
2
e?(FT (y)??T ) ? e 2 ?T .
y?U (Y)
def
?T2 , and let ? = Pr
?T ? ? ? FT (?
y) ? F (?
y) ? ?T2 ? ? 2 . Then,
E
T ?U (G)
Pr
T ?U (G)n
? 12
?T ? T : ?T (?
y) ? e
(F (?
y)??)2
?2
? 1 ? (1 ? ?)n .
Thus, even for very small ?, when n is large enough, there exists, with high probability, a tree T ? T
? has a small ?T (?
such that y
y) whenever [F (?
y) ? ?]/? is large for G. For example, when |Y| = 2`
n
(the multiple binary classification case), we have with probability
? of at least 1 ? (1 ? ?) , that there
exists T ? T such that KT (?
y) = 1 whenever F (?
y) ? ? ? ? 2` ln 2.
4.2
Optimization
To optimize the L2 -norm RTA problem (Definition 6) we convert it to the marginalized dual form
(see the supplementary material for the derivation), which gives us a polynomial-size problem (in
the number of microlabels) and allows us to use kernels to tackle complex input spaces efficiently.
Definition 9. L2 -norm RTA Marginalized Dual
1 X
1 X
maxm
?(k, e, ue ) ?
?(k, e, ue )KTe (xk , ue ; x0k , u0e )?(k 0 , e, u0e ) ,
? ?M
|ET |
2
e,k,ue
e,k,ue ,
k0 ,u0e
where ET is the union of the sets of edges appearing in T , and ? ? Mm are the marginal dual
def
variables ? = (?(k, e, ue ))k,e,ue , with the triplet (k, e, ue ) corresponding to labeling the edge
6
DATASET
M ICROLABEL L OSS (%)
SVM
E MOTIONS
Y EAST
S CENE
E NRON
C AL 500
22.4
20.0
9.8
6.4
13.7
F INGERPRINT 10.3
NCI60
15.3
M EDICAL
2.6
C IRCLE 10
4.7
C IRCLE 50
5.7
0/1 L OSS (%)
MTL MMCRF MAM
RTA
SVM
MTL
20.2
20.7
11.6
6.5
13.8
17.3
16.0
2.6
6.3
6.2
18.8
19.8
8.8
5.3
13.8
10.7
14.9
2.1
0.6
3.8
77.8
85.9
47.2
99.6
100.0
99.0
56.9
91.8
28.9
69.8
74.5
88.7
55.2
99.6
100.0
100.0
53.0
91.8
33.2
72.3
20.1
21.7
18.4
6.2
13.7
10.5
14.6
2.1
2.6
1.5
19.5
20.1
17.0
5.0
13.7
10.5
14.3
2.1
2.5
2.1
MMCRF MAM
71.3
93.0
72.2
92.7
100.0
99.6
63.1
63.8
20.3
38.8
69.6
86.0
94.6
87.9
100.0
99.6
60.0
63.1
17.7
46.2
RTA
66.3
77.7
30.2
87.7
100.0
96.7
52.9
58.8
4.0
52.8
Table 1: Prediction performance of each algorithm in terms of microlabel loss and 0/1 loss. The best
performing algorithm is highlighted with boldface, the second best is in italic.
e = (v, v 0 ) ? ET of the output graph by ue = (uv , uv0 ) ? Yv ?Yv0 for the training example xk . Also,
Mm is the marginal dual feasible set and
def
KTe (xk , ue ; xk0 , u0e ) =
NT (e)
K(xk , xk0 ) ? e (ykv , ykv0 ) ? ? e (uv , uv0 ), ? e (yk0 v , yk0 v0 ) ? ? e (u0v , u0v0 )
|ET |2
is the joint kernel of input features and the differences of output features of true and competing
multilabels (yk , u), projected to the edge e. Finally, NT (e) denotes the number of times e appears
among the trees of the ensemble.
The master algorithm described in the supplementary material iterates over each training example
until convergence. The processing of each training example xk proceeds by finding the worst violating multilabel of the ensemble defined as
def
? k = argmax FT (xk , y) ,
y
(6)
y6=yk
using the K-best inference approach of the previous section, with the modification that the correct
? k is mapped to a vertex
multilabel is excluded from the K-best lists. The worst violator y
? (xk ) = C ? ([?
?
ye = ue ])e,ue ? Mk
corresponding to the steepest feasible ascent direction (c.f, [9]) in the marginal dual feasible set Mk
of example xk , thus giving us a subgradient of the objective of Definition 9. An exact line search is
?.
used to find the saddle point between the current solution and ?
5
Empirical Evaluation
We compare our method RTA to Support Vector Machine (SVM) [10, 11], Multitask Feature Learning (MTL) [12], Max-Margin Conditional Random Fields (MMCRF) [9] which uses the loopy belief propagation algorithm for approximate inference on the general graph, and Maximum Average
Marginal Aggregation (MAM) [5] which is a multilabel ensemble model that trains a set of random
tree based learners separately and performs the final approximate inference on a union graph of the
edge potential functions of the trees. We use ten multilabel datasets from [5]. Following [5], MAM
is constructed with 180 tree based learners, and for MMCRF a consensus graph is created by pooling edges from 40 trees. We train RTA with up to 40 spanning trees and with K up to 32. The linear
kernel is used for methods that require kernelized input. Margin slack parameters are selected from
{100, 50, 10, 1, 0.5, 0.1, 0.01}. We use 5-fold cross-validation to compute the results.
Prediction performance. Table 1 shows the performance in terms of microlabel loss and 0/1 loss.
The best methods are highlighted in ?boldface? and the second best in ?italics? (see supplementary
material for full results). RTA quite often improves over MAM in 0/1 accuracy, sometimes with
noticeable margin except for Enron and Circle50. The performances in microlabel accuracy are
quite similar while RTA is slightly above the competition. This demonstrates the advantage of RTA
that gains by optimizing on a collection of trees simultaneously rather than optimizing on individual
trees as MAM. In addition, learning using approximate inference on a general graph seems less
7
?
?
1
3
?
10
?
?
32
K (% of |Y|)
100
316
1000
?
?
1
3
?
?
10
?
32
?
?
K (% of |Y|)
100
?
100
316
1000
?
?
??
?
60
80
?
40
60
80
?
20
?
Y* being verified (% of data)
?
??
Emotions
Yeast
Scene
Enron
Cal500
Fingerprint
NCI60
Medical
Circle10
Circle50
??
?
0
20
40
?
?
40
60
80
?
|T| = 40
?
?
20
?
Y* being verified (% of data)
??
0
?
100
|T| = 10
?
?
0
Y* being verified (% of data)
100
|T| = 5
1
3
?
10
?
?
32
?
K (% of |Y|)
?
?
100
316
1000
Figure 1: Percentage of examples with provably optimal y? being in the K-best lists plotted as a
function of K, scaled with respect to the number of microlabels in the dataset.
favorable as the tree-based methods, as MMCRF quite consistently trails to RTA and MAM in
both microlabel and 0/1 error, except for Circle50 where it outperforms other models. Finally, we
notice that SVM, as a single label classifier, is very competitive against most multilabel methods for
microlabel accuracy.
Exactness of inference on the collection of trees. We now study the empirical behavior of the
inference (see Section 4) on the collection of trees, which, if taken as a single general graph, would
call for solving an N P-hard inference problem. We provide here empirical evidence that we can
perform exact inference on most examples in most datasets in polynomial time.
We ran the K-best inference on eleven datasets where the RTA models were trained with different
amounts of spanning trees |T | = {5, 10, 40} and values for K = {2, 4, 8, 16, 32, 40, 60}. For each parameter combination and for each example, we recorded whether the K-best inference was provably
exact on the collection (i.e., if Lemma 7 was satisfied). Figure 1 plots the percentage of examples
where the inference was indeed provably exact. The values are shown as a function of K, expressed
as the percentage of the number of microlabels in each dataset. Hence, 100% means K = `, which
denotes low polynomial (?(n`2 )) time inference in the exponential size multilabel space.
We observe, from Figure 1, on some datasets (e.g., Medical, NCI60), that the inference task is very
easy since exact inference can be computed for most of the examples even with K values that are
below 50% of the number of microlabels. By setting K = ` (i.e., 100%) we can perform exact
inference for about 90% of the examples on nine datasets with five trees, and eight datasets with
40 trees. On two of the datasets (Cal500, Circle50), inference is not (in general) exact with low
values of K. Allowing K to grow superlinearly on ` would possibly permit exact inference on these
datasets. However, this is left for future studies.
Finally, we note that the difficulty of performing provably exact inference slightly increases when
more spanning trees are used. We have observed that, in most cases, the optimal multilabel y? is
still on the K-best lists but the conditions of Lemma 7 are no longer satisfied, hence forbidding us
to prove exactness of the inference. Thus, working to establish alternative proofs of exactness is a
worthy future research direction.
6
Conclusion
The main theoretical result of the paper is the demonstration that if a large margin structured output
predictor exists, then combining a small sample of random trees will, with high probability, generate
a predictor with good generalization. The key attraction of this approach is the tractability of the
inference problem for the ensemble of trees, both indicated by our theoretical analysis and supported
by our empirical results. However, as a by-product, we have a significant added benefit: we do not
need to know the output structure a priori as this is generated implicitly in the learned weights
for the trees. This is used to significant advantage in our experiments that automatically leverage
correlations between the multiple target outputs to give a substantive increase in accuracy. It also
suggests that the approach has enormous potential for applications where the structure of the output
is not known but is expected to play an important role.
8
References
[1] Ben Taskar, Carlos Guestrin, and Daphne Koller. Max-margin markov networks. In S. Thrun,
L.K. Saul, and B. Sch?olkopf, editors, Advances in Neural Information Processing Systems 16,
pages 25?32. MIT Press, 2004.
[2] Martin J. Wainwright, Tommy S. Jaakkola, and Alan S. Willsky. MAP estimation via agreement on trees: message-passing and linear programming. IEEE Transactions on Information
Theory, 51(11):3697?3717, 2005.
[3] Michael I. Jordan and Martin J Wainwright. Semidefinite relaxations for approximate inference
on graphs with cycles. In S. Thrun, L.K. Saul, and B. Sch?olkopf, editors, Advances in Neural
Information Processing Systems 16, pages 369?376. MIT Press, 2004.
[4] Amir Globerson and Tommi S. Jaakkola. Approximate inference using planar graph decomposition. In B. Sch?olkopf, J.C. Platt, and T. Hoffman, editors, Advances in Neural Information
Processing Systems 19, pages 473?480. MIT Press, 2007.
[5] Hongyu Su and Juho Rousu. Multilabel classification through random graph ensembles. Machine Learning, dx.doi.org/10.1007/s10994-014-5465-9, 2014.
[6] Robert G. Cowell, A. Philip Dawid, Steffen L. Lauritzen, and David J. Spiegelhalter. Probabilistic Networks and Expert Systems. Springer, New York, 1999.
[7] Thomas G?artner and Shankar Vembu. On structured output training: hard cases and an efficient
alternative. Machine Learning, 79:227?242, 2009.
[8] John Shawe-Taylor and Nello Cristianini. Kernel Methods for Pattern Analysis. Cambridge
University Press, 2004.
[9] J. Rousu, C. Saunders, S. Szedmak, and J. Shawe-Taylor. Efficient algorithms for max-margin
structured classification. Predicting Structured Data, pages 105?129, 2007.
[10] Kristin P. Bennett. Combining support vector and mathematical programming methods for
classifications. In B. Sch?olkopf, C. J. C. Burges, and A. J. Smola, editors, Advances in Kernel
Methods?Support Vector Learning, pages 307?326. MIT Press, Cambridge, MA, 1999.
[11] Nello Cristianini and John Shawe-Taylor. An Introduction to Support Vector Machines and
Other Kernel-Based Learning Methods. Cambridge University Press, Cambridge, U.K., 2000.
[12] Andreas Argyriou, Theodoros Evgeniou, and Massimiliano Pontil. Convex multi-task feature
learning. Machine Learning, 73(3):243?272, 2008.
[13] Yevgeny Seldin, Franc?ois Laviolette, Nicol`o Cesa-Bianchi, John Shawe-Taylor, and Peter
Auer. PAC-Bayesian inequalities for martingales. IEEE Transactions on Information Theory,
58:7086?7093, 2012.
[14] Andreas Maurer. A note on the PAC Bayesian theorem. CoRR, cs.LG/0411099, 2004.
[15] David McAllester. PAC-Bayesian stochastic model selection. Machine Learning, 51:5?21,
2003.
[16] Juho Rousu, Craig Saunders, Sandor Szedmak, and John Shawe-Taylor. Kernel-based learning of hierarchical multilabel classification models. Journal of Machine Learning Research,
7:1601?1626, December 2006.
9
| 5382 |@word multitask:1 polynomial:6 norm:19 seems:1 yv0:1 decomposition:1 q1:1 didate:1 ytn:1 epartement:1 contains:1 score:13 outperforms:1 current:1 nt:2 forbidding:2 written:3 dx:1 john:5 realistic:2 eleven:1 drop:1 plot:1 selected:1 amir:1 xk:12 steepest:1 provides:1 iterates:1 node:3 theodoros:1 org:1 daphne:1 five:1 mathematical:1 constructed:1 incorrect:1 consists:1 prove:1 artner:1 tommy:1 introduce:1 theoretically:1 expected:2 indeed:3 behavior:1 os:2 examine:1 multi:1 steffen:1 decreasing:1 automatically:1 enumeration:1 increasing:2 becomes:1 provided:1 moreover:1 notation:1 bounded:1 ykv:1 superlinearly:1 minimizes:2 finding:3 guarantee:4 ti:22 tackle:2 ebec:1 universit:1 classifier:1 k2:4 demonstrates:1 uk:2 control:1 unit:14 medical:2 scaled:1 platt:1 positive:2 t1:2 consequence:1 despite:1 might:1 umr:1 suggests:2 practical:2 globerson:1 yj:14 union:3 practice:1 definite:1 pontil:1 empirical:5 attain:2 adapting:1 projection:2 selection:1 shankar:1 risk:3 a2t:1 optimize:1 fruitful:1 map:5 yt:4 maximizing:2 nron:1 independently:2 convex:6 qc:1 attraction:1 searching:1 analogous:1 s10994:1 enhanced:1 suppose:1 target:1 modulo:1 exact:12 programming:6 play:1 us:1 trail:1 agreement:1 assure:1 dawid:1 satisfying:3 observed:1 ft:29 role:1 taskar:1 solved:1 worst:2 ensures:1 cycle:1 highest:6 yk:9 ran:1 abel:1 miny:1 cristianini:2 dynamic:3 multilabel:16 trained:1 solving:3 learner:2 basis:1 joint:9 k0:1 derivation:1 train:2 univ:2 informatique:1 distinct:1 fast:2 london:2 describe:1 doi:1 massimiliano:1 labeling:2 tell:2 exhaustive:1 saunders:2 quite:3 heuristic:1 supplementary:6 solve:1 ramp:5 otherwise:2 think:1 highlighted:2 final:1 advantage:2 ucl:1 product:3 fr:1 combining:2 ramification:1 rapidly:1 pthe:1 competition:1 olkopf:4 convergence:2 ben:1 ac:1 lauritzen:1 noticeable:1 implemented:1 predicted:1 ois:1 c:1 differ:1 direction:2 tommi:1 correct:2 stochastic:1 enable:2 settle:1 mcallester:1 material:6 require:2 generalization:6 investigation:1 mathematically:1 exploring:1 hold:1 practically:1 mm:2 traded:1 finland:2 achieves:10 smallest:1 estimation:3 favorable:1 label:3 superposition:1 maxm:1 weighted:2 hoffman:1 kristin:1 exactness:3 mit:4 gaussian:2 always:1 aim:2 modified:1 rather:2 pn:2 jaakkola:2 consistently:1 rank:2 aalto:4 rigorous:1 inference:37 cnrs:1 kernelized:1 koller:1 france:1 provably:4 classification:6 dual:5 among:1 priori:1 ircle:2 special:1 marginal:4 field:1 emotion:1 evgeniou:1 having:1 sampling:1 represents:2 y6:2 future:2 hongyu:3 t2:3 franc:1 randomly:1 simultaneously:5 kwt:2 individual:5 argmax:7 n1:1 message:1 investigate:1 evaluation:1 hwi:2 semidefinite:1 primal:2 logiciel:1 kt:3 edge:15 tree:69 taylor:7 maurer:1 plotted:1 theoretical:3 mk:2 loopy:2 tractability:1 vertex:1 predictor:15 uniform:3 kn:2 st:2 fundamental:1 probabilistic:1 michael:1 together:1 ym:1 cesa:1 recorded:1 satisfied:2 possibly:2 expert:1 potential:4 view:1 mario:2 kwk:1 yv:1 aggregation:1 competitive:1 carlos:1 minimize:1 accuracy:6 efficiently:3 ensemble:7 bayesian:4 craig:1 emilie:2 whenever:6 definition:5 against:2 nonetheless:1 dm:1 proof:2 sampled:1 gain:1 dataset:3 austria:1 knowledge:1 recall:3 improves:1 auer:1 appears:1 supervised:1 mtl:3 violating:1 planar:1 smola:1 until:1 correlation:1 hand:1 working:1 su:3 maximizer:4 propagation:1 indicated:1 resemblance:1 yeast:1 believe:1 ye:1 verify:1 true:2 normalized:4 hence:10 regularization:1 excluded:1 attractive:1 ue:12 rooted:1 covering:1 kak:1 ulaval:1 ln2:1 complete:15 tn:2 performs:1 motion:1 fi:4 uv0:2 laval:1 exponentially:1 interpretation:1 vembu:1 significant:4 cambridge:4 uv:2 shawe:7 fingerprint:1 longer:1 v0:1 optimizing:2 apart:1 scenario:1 cene:1 inequality:2 binary:1 yi:17 scoring:8 guestrin:1 ey:1 semi:1 relates:1 full:3 multiple:2 alan:1 cross:1 prevented:1 qi:3 prediction:4 scalable:2 denominator:1 expectation:4 rousu:5 yk0:2 kernel:7 sometimes:1 achieved:3 addition:1 separately:1 grow:1 allocated:1 sch:4 enron:2 ascent:1 pooling:1 undirected:1 december:1 jordan:1 integer:1 call:1 leverage:1 enough:1 easy:1 misclassifications:1 wti:3 competing:1 inner:1 andreas:2 whether:2 peter:1 passing:1 nine:1 york:1 generally:1 yw:4 amount:2 ten:1 reduced:1 generate:1 qi2:1 exist:1 percentage:3 notice:1 per:1 discrete:1 ist:1 key:4 enormous:1 achieving:3 verified:3 graph:29 relaxation:4 fwti:3 fraction:2 sum:1 convert:1 subgradient:1 master:1 place:1 laid:1 throughout:1 yt1:1 draw:2 investigates:1 bound:4 def:32 conical:8 fold:1 marchand:2 kronecker:1 helsinki:2 ri:2 scene:1 min:5 performing:3 martin:2 structured:6 department:1 according:5 combination:10 fwt:1 smaller:1 slightly:2 y0:6 wi:5 lp:1 qu:1 modification:1 maxy:1 pr:2 taken:1 ln:3 equation:1 slack:3 know:1 tractable:2 permit:1 multiplied:1 eight:1 observe:1 juho:4 hierarchical:1 appearing:1 alternative:2 existence:1 substitute:1 thomas:1 denotes:3 running:1 mam:7 hinge:4 marginalized:2 etienne:2 laviolette:1 giving:1 establish:2 objective:1 sandor:1 added:1 realized:1 dependence:1 usual:2 italic:2 mapped:2 thrun:2 concatenation:3 philip:1 nello:2 consensus:1 spanning:22 boldface:2 willsky:1 substantive:1 length:1 demonstration:1 equivalently:1 lg:1 robert:1 negative:1 xk0:2 affiliated:1 unknown:1 perform:3 allowing:1 bianchi:1 observation:1 markov:6 datasets:10 finite:2 y1:2 worthy:2 arbitrary:4 canada:1 david:2 pair:1 namely:2 learned:2 address:1 proceeds:1 below:1 pattern:1 xm:1 max:6 belief:1 wainwright:2 difficulty:1 predicting:1 representing:1 technology:2 firmly:1 spiegelhalter:1 created:1 carried:1 naive:1 szedmak:2 l2:19 nicol:1 x0k:1 loss:13 validation:1 foundation:1 incurred:1 sufficient:1 editor:4 intractability:2 ift:1 supported:1 last:1 side:1 burges:1 institute:2 saul:2 absolute:1 benefit:1 valid:1 qn:1 made:1 collection:6 projected:1 avoided:1 transaction:2 approximate:7 implicitly:1 confirm:2 xi:1 search:2 triplet:1 table:2 superficial:1 ca:1 rta:11 complex:1 assured:1 main:1 yevgeny:1 x1:1 martingale:1 slow:1 sub:2 concatenating:1 exponential:2 candidate:2 hw:5 theorem:8 pac:4 list:4 svm:8 evidence:1 intractable:1 exists:9 corr:1 cal500:2 margin:28 saddle:1 seldin:1 expressed:2 applies:1 cowell:1 springer:1 violator:2 ma:1 conditional:2 consequently:1 replace:1 bennett:1 yti:3 hard:6 feasible:3 included:1 except:2 operates:1 hyperplane:2 wt:4 acting:3 lemma:15 called:2 experimental:2 east:1 college:1 support:5 morvant:3 dissimilar:1 multilabels:6 dept:2 argyriou:1 |
4,841 | 5,383 | Metric Learning for Temporal Sequence Alignment
Damien Garreau ? ?
ENS
[email protected]
R?emi Lajugie ? ?
INRIA
[email protected]
Sylvain Arlot ?
CNRS
[email protected]
Francis Bach ?
INRIA
[email protected]
Abstract
In this paper, we propose to learn a Mahalanobis distance to perform alignment
of multivariate time series. The learning examples for this task are time series for
which the true alignment is known. We cast the alignment problem as a structured
prediction task, and propose realistic losses between alignments for which the
optimization is tractable. We provide experiments on real data in the audio-toaudio context, where we show that the learning of a similarity measure leads to
improvements in the performance of the alignment task. We also propose to use
this metric learning framework to perform feature selection and, from basic audio
features, build a combination of these with better alignment performance.
1
Introduction
The problem of aligning temporal sequences is ubiquitous in applications ranging from bioinformatics [5, 1, 23] to audio processing [4, 6]. The goal is to align two similar time series that have the
same global structure, but local temporal differences. Most alignments algorithms rely on similarity measures, and having a good metric is crucial, especially in the high-dimensional setting where
some features of the signals can be irrelevant to the alignment task. The goal of this paper is to show
how to learn this similarity measure from annotated examples in order to improve the relevance of
the alignments.
For example, in the context of music information retrieval, alignment is used in two different cases:
(1) audio-to-audio alignment and (2) audio-to-score alignment. In the first case, the goal is to match
two audio interpretations of the same piece that are potentially different in rythm, whereas audio-toscore alignment focuses on matching an audio signal to a symbolic representation of the score. In
the second case, there are some attempts to learn from annotated data a measure for performing the
alignment. Joder et al. [12] propose to fit a generative model in that context, and Keshet et al. [13]
learn this measure in a discriminative setting.
Similarly to Keshet et al. [13], we use a discriminative loss to learn the measure, but our work focuses
on audio-to-audio alignment. In that context, the set of authorized alignments is much larger, and we
explicitly cast the problem as a structured prediction task, that we solve using off-the-shelf stochastic
optimization techniques [15] but with proper and significant adjustments, in particular in terms of
losses. The ideas of alignment are also very relevant to the community of speech recognition since
the pioneering work of Sakoe and Chiba [19].
?
?
Contributed equally
SIERRA project-team, D?epartement d?Informatique de l?Ecole Normale Sup?erieure (CNRS, INRIA, ENS)
1
The need for metric learning goes far beyond unsupervised partitioning problems. Weinberger and
Saul [26] proposed a large-margin framework for learning a metric in nearest-neighbour algorithms
based on sets of must-link/must-not-link constraints. Lajugie et al. [16] proposed to use a large
margin framework to learn a Mahalanobis metric in the context of partitioning problems. Since
structured SVM have been proposed by Tsochantaridis et al. [25] and Taskar et al. [22], they have
successfully been used to solve many learning problems, for instance to learn weights for graph
matching [3] or a metric for ranking tasks [17]. They have also been used to learn graph structures
using graph cuts [21].
We make the following five contributions:
? We cast the learning of a Mahalanobis metric in the context of alignment as a structured prediction
problem.
? We show that on real musical datasets this metric improves the performance of alignment algorithms using high-level features.
? We propose to use the metric learning framework to learn combinations of basic audio features
and get good alignment performances.
? We show experimentally that the standard Hamming loss, although tractable computationnally,
does not permit to learn a relevant similarity measure in some real world settings.
? We propose a new loss, closer to the true evaluation loss for alignments, leading to a tractable
learning task, and derive an efficient Frank-Wolfe-based algorithm to deal with this new loss.
That loss solves some issues encountered with the Hamming loss.
2
Matricial formulation of alignment problems
2.1 Notations
In this paper, we consider the alignment problem between two multivariate time series sharing the
same dimension p, but possibly of different lengths TA and TB , namely A ? RTA ?p and B ?
RTB ?p . We refer to the rows of A as a1 , . . . , aTA ? Rp and those of B as b1 , . . . , bTB ? Rp as
column vectors. From now on, we denote by X the pair of signals (A, B).
Let C(X) ? RTA ?TB be an arbitrary pairwise affinity matrix associated to the pair X, that is,
C(X)i,j encodes the affinity between ai and bj . Note that our framework can be extended to the
case where A and B are multivariate signals of different dimensions, as long as C(X) is welldefined. The goal of the alignment task is to find two non-decreasing sequences of indices ? and ?
of same length u ? max(TA , TB ) and to match each
index ?(i) in the time series A to the time
Ptime
u
index ?(i) in the time series B, in such a way that i=1 C(X)?(i),?(i) is maximal, and that (?, ?)
satisfies:
?
?(1) = ?(1) = 1
(matching beginnings)
?
?(u) = TA , ?(u) = TB
(matching endings)
(1)
?
?i, (?(i + 1), ?(i + 1)) ? (?(i), ?(i)) ? {(1, 0), (0, 1), (1, 1)} (three type of moves)
For a given (?, ?), we define the binary matrix Y ? {0, 1}TA ?TB such that Y?(i),?(i) = 1 for every
i ? {1, . . . , u} and 0 otherwise. We denote by Y(X) the set of such matrices, which is uniquely
determined by TA and TB . An example is given in Fig. 1. A vertical move in the Y matrix means
that the signal B is waiting for A, whereas an horizontal one means that A is waiting for B, and a
diagonal move means that they move together. In this sense the time reference is ?warped?.
When C(X) is known, the alignment task can be cast as the following linear program (LP) over the
set Y(X):
max Tr(C(X)> Y ).
(2)
Y ?Y(X)
Our goal is to learn how to form the affinity matrix: once we have learned C(X), the alignment is
obtained from Eq. (2). The optimization problem in Eq. (2) will be referred to as the decoding of
our model.
Dynamic time warping. Given the affinity matrix C(X) associated with the pair of signals X =
(A, B), finding the alignment that solves the LP of Eq. (2) can be done efficiently in O(TA TB ) using
2
Figure 1: Example of two valid alignments encoded by matrices Y 1 and Y 2 . Red upper triangles
1
2
show the (i, j) such that Yi,j
= 1, and the blue lower ones show the (i, j) such that Yi,j
= 1. The
1
2
grey zone corresponds to the area loss ?abs between Y and Y .
a dynamic programming algorithm. It is often referred to as dynamic time warping [5, 18]. This
algorithm is described in Alg. 1 of the supplementary material. Various additional constraints may
be used in the dynamic time warping algorithm [18], which we could easily add to Alg. 1.
The cardinality of the set Y(X) is huge: it corresponds to the number of paths on a rectangular grid
from the southwest (1, 1) to the northeast corner (TA , TB ) with vertical, horizontal and diagonal
moves allowed. This is the definition of the Delannoy numbers?[2]. As noted in [24], when t =
t
? ?2) .
TA = TB goes to infinity, and one can show that #Yt,s ? ?(3+2
?t
2.2
3 2?4
The Mahalanobis metric
In many applications (see, e.g., [6]), for a pair X = (A, B), the affinity matrix is computed by
C(A, B)i,j = ?kai,k ? bj,k k2 . In this paper we propose to learn the metric to compare ai and bj
instead of using the plain Euclidean metric. That is, C(X) is parametrized by a matrix W ? W ?
Rp?p , where W ? Rp?p is the set of semi-definite positive matrices, and we use the corresponding
Mahalanobis metric to compute the pairwise affinity between ai and bj :
C(X; W )i,j = ?(ai ? bj )> W (ai ? bj ).
(3)
Note that the decoding of Eq. (2) is the maximization of a linear function in the parameter W :
max Tr(C(X; W )> Y )
Y ?Y(X)
?
max Tr(W > ?(X, Y )),
Y ?Y(X)
(4)
if we define the joint feature map
?(X, Y ) = ?
TA X
TB
X
Yi,j (ai ? bj )(ai ? bj )> ? Rp?p .
(5)
i=1 j=1
3
Learning the metric
From now on, we assume that we are given n pairs of training instances1 (X i , Y i ) =
i
i
i
i
((Ai , B i ), Y i ) ? RTA ?p ? RTB ?p ? {0, 1}TA ?TB , i = 1, . . . , n. Our goal is to find a matrix
W such that the predicted alignments are close to the groundtruth on these examples, as well as
on unseen examples. We first define a loss between alignments, in order to quantify the proximity
between alignments.
1
We will see that it is necessary to have fully labelled instances, which means that for each pair X i we need
an exact alignment Y i between Ai and B i . Partial alignment might be dealt with by alternating between metric
learning and constrained alignment.
3
3.1 Losses between alignments
In our framework, the alignments are encoded by matrices in Y(X), thus we P
are interested in func2
tions ` : Y(X) ? Y(X) ? R+ . The Frobenius norm is defined by kM k2F = i,j Mi,j
.
Hamming loss. A simple loss between matrices is the Frobenius norm of their difference, which
turns out to be the unnormalized Hamming loss [9] for 0/1-valued matrices. For two matrices
Y1 , Y2 ? Y(X), it is defined as:
`H (Y1 , Y2 ) = kY1 ? Y2 k2F = Tr(Y1> Y1 ) + Tr(Y2> Y2 ) ? 2 Tr(Y1> Y2 )
>
>
= Tr(Y1 1TB 1>
TA ) + Tr(Y2 1TB 1TA ) ? 2 Tr(Y1 Y2 ),
(6)
T
where 1T is the vector of R with all coordinates equal to 1. The last line of Eq. (6) comes from
the fact that the Yi have 0/1-values; that makes the Hamming loss affine in Y1 and Y2 . This loss is
often used in other structured prediction tasks [15]; in the audio-to-score setting, Keshet et al. [13]
use a modified version of this loss, which is the average number of times the difference between the
two alignments is greater than a fixed threshold.
This loss is easy to optimize since, it is linear in our parametrization of the alignement problem, but
not optimal for audio-to-audio alignment. Indeed, a major drawback of the Hamming loss is that, for
alignments of fixed length, it depends only on the number of ?crossings? between alignment paths:
one can easily find Y1 , Y2 , Y3 such that `H (Y2 , Y1 ) = `H (Y3 , Y1 ) but Y2 is much closer to Y1 than
Y3 (see Fig. 2). It is important to notice this is often the case when the length of the signals grows.
Area loss. A more natural loss can be computed as the mean distance beween the paths depicted by
two matrices Y 1 , Y 2 ? Y(X). This loss corresponds to the area between the paths of two matrices
Y , as represented by the grey zone on Fig. 1.
1
2
Formally, as in Fig. 1, for each t ? {1, . . . , TB } we put ?t = | min{k, Yt,k
= 1}?min{k, Yt,k
= 1}|.
Then the area loss is the mean of the ?t . In the audio literature [14], this loss is sometimes called the
?mean absolute deviation? loss and is noted ?abs (Y 1 , Y 2 ).
Unfortunately, for the general alignment problem, ?abs is not linear in the matrices Y . But in the
context of alignment of sequences of two different natures, one of the signal is a reference and
thus the index sequence ? defined in Eq. (1) is increasing, e.g., for the audio-to-partition alignment
problem [12]. This loss is then linear in each of its arguments. More precisely, if we introduce the
matrix LTA ? RTA ?TA which is lower triangular with ones (including on the diagonal), we can
write the loss as
`O = kLTA (Y1 ? Y2 )k2F
(7)
>
> >
= Tr(LTA Y1 1TB 1>
TA ) + Tr(LTA Y2 1TB 1TA ) ? 2 Tr(LTA Y1 Y2 LTA ).
We now prove that this loss corresponds to the area loss in this special case. Let Y be an alignment,
P
Pi
then it is easy see that (LTA Y )i,j = k (LTA )i,k Yk,j = k=1 Yk,j . If Y does not have vertical
moves, i.e., for eachPj there is an unique kj such that Ykj ,j = 1, we have that (LTA Y )i,j = 1 if and
only if i ? kj . So i,j (LTA Y )i,j = #{(i, j), i ? kj }, which is exactly the area under the curve
determined by the path of Y . In all our experiments, we use ?abs for evaluation but not for training.
Approximation of the area loss: the symmetrized area loss. In many real world applications [14], a meaningful loss to assess the quality of an alignment is the area loss. As shown by
our experiments, if the Hamming loss is sufficient in some simple situations and allows to learn a
metric that leads to good alignment performance in terms of area loss, on more challenging datasets
it does not work at all (see Sec. 5). This is due to the fact that two alignments that are very close
in terms of area loss can suffer a big Hamming loss (cf. Fig. 2). Thus it is natural to extend the
formulation of Eq. (7) to matrices in Y(X). We start by symmetrizing the formulation of Eq. (7) to
overcome problems of overpenalization of vertical vs. horizontal moves. We define, for any couple
of binary matrices (Y 1 , Y 2 ),
1
(8)
`S (Y1 , Y2 ) = kLTA (Y1 ? Y2 )k2F + k(Y1 ? Y2 )LTB )k2F
2
h
1
>
> >
=
Tr(Y1> L>
TA LTA Y1 ) + Tr(LTA Y2 1TB 1TA ) ? 2 Tr(Y2 LTA LTA Y1 )
2
i
>
>
>
>
+ Tr(Y1 LTB L>
.
TB Y ) + Tr(Y2 1TA 1TB LTB LTB Y2 ) ? 2 Tr(Y2 LTB LTB Y1
4
1600
Most violated constraint for Hamming Loss
1400
1200
Most violated constraint for lS
Groundruth alignment
tB
1000
800
600
400
200
0
0
200
400
600
800
tA
1000
1200
1400
1600
1800
Figure 2: On the real world Bach chorales dataset, we have represented a groundtruth alignment
together with two others. In term of Hamming loss, both alignments are as far from the groundtruth
whereas for the area loss, they are not. In the structured prediction setting described in Sec. 4,
the depicted alignment are the so-called ?most violated constraint?, namely the output of the loss
augmented decoding step (see Sec. 4).
We propose now to make this loss concave over the convex hull of Y(X) that we denote from now
2
on Y(X). Let us introduce DT = ?max (L>
T LT )IT ?T with ?max (U ) the largest eigenvalue of U .
For any binary matrices Y1 , Y2 , we have
1
>
>
`S (Y1 , Y2 ) =
Tr(Y1> (L>
TA LTA ? DTA )Y1 ) + Tr(DTA Y1 1TB 1TA ) + Tr(LTA Y2 1TB 1TA )
2
>
? 2 Tr(Y2> (L>
TA L ? DTA )Y1 ) + Tr(Y1 (LTB LTB ? DTB )Y )
i
>
>
>
>
+ Tr(Y1 DTB 1TB 1>
)
+
Tr(Y
L
L
Y
)
?
2
Tr(Y
L
L
Y
)
,
T
2
2
T
TA
2
TB
TB 1
B
B
and we get a concave function over Y(X) that coincides with `S on Y(X).
3.2 Empirical loss minimization
Recall that we are given n alignment examples (X i , Y i )1?i?n . For a fixed loss `, our goal is now
to solve the following minimization problem in W :
?
?
n
?
?1 X
min
` Y i , argmax Tr(C(X i ; W )> Y ) + ??(W ) ,
(9)
W ?W ? n
?
Y ?YT i ,T i
i=1
A
B
where ? = ?2 kW k2F is a convex regularizer preventing from overfitting, with ? ? 0.
4
Large margin approach
In this section we describe a large margin approach to solve a surrogate to the problem in Eq. (9),
which is untractable. As shown in Eq. (4), the decoding task is the maximum of a linear function
in the parameter W and aims at predicting an output over a large and discrete space (the space
of potential alignments with respect to the constraints in Eq. (1)). Learning W thus falls into the
structured prediction framework [25, 22]. We define the hinge loss, a convex surrogate, by
n
o
L(X, Y ; W ) = 0max
`(Y, Y 0 ) ? Tr(W > [?(X, Y ) ? ?(X, Y 0 )]) .
(10)
Y ?Y(X)
2
For completeness, in our experiments, we also try to set the matrices DT with minimal trace that dominate
L>
T LT by solving a semidefinite program (SDP). We report the associated result in Fig 4. Note also that
other matrices could have been chosen. In particular, since our matrices LT are pointwise positive, the matrix
>
Diag(L>
T LT ) ? LT LT is such that the loss is concave.
5
The evaluation of L is usually referred to as ?loss-augmented decoding?, see [25]. If we define Yb i
as the argmax in Eq. (10) when (X, Y ) = (X i , Y i ), then elementary computations show that
Yb i = argmin Tr((U > ? 2Y i> ? C(X i ; W )> )Y ),
Y ?Y(X)
TA ?TB
where U = 1TB 1>
.
TB ? R
We now aim at solving the following problem, sometimes called the margin-rescaled problem:
min
W ?W
n
n
o
?
1X
kW k2F +
max `(Y, Y i ) ? Tr(W > ?(X i , Y i ) ? ?(X i , Y ) ) .
2
n i=1 Y ?Y(X)
(11)
Hamming loss case. From Eq. (4), one can notice that our joint feature map is linear in Y . Thus,
if we take a loss that is linear in the first argument of `, for instance the Hamming loss, the lossaugmented decoding is the maximization of a linear function over the spaces Y(X) that we can
solve efficiently using dynamic programming algorithms (see Sec. 2.1 and supplementary material).
That way, plugging the Hamming loss (Eq. (6)) in Eq. (11) leads to a convex structured prediction
problem. This problem can be solved using standard techniques such as cutting plane methods [11],
stochastic gradient descent [20], or block-coordinate Frank-Wolfe in the dual [15]. Note that we
adapted the standard unconstrained optimization methods to our setting, where W 0.
Optimization using the symmetrized area loss. The symmetrized area loss is concave in its first
argument, thus the problem of Eq. (11) is in a min/max form and deriving a dual is straightforward.
Details can be found in the supplementary material. If we plug the symmetrized area loss `S (SAL)
defined in Eq. (8) into our problem (11), we can show that the dual of (11) has the following form:
Pn
Pn
P
1
1
i
i
T 2
min
i=1 ?
i=1 `S (Z, Z ), (12)
j,k (Yi ? Z )j,k (aj ? bk )(aj ? bk ) kF ? n
2?n2 k
(Z 1 ,...,Z n )?Y
if we denote by Y(X i ) the convex hull of the sets Y(X i ), and by Y the cartesian product over all
the training examples i of such sets. Note that we recover a similar result as [15]. Since the SAL
loss is concave, the aforementioned problem is convex.
The problem (12) is a quadratic program over the compact set Y. Thus we can use a Frank-Wolfe [7]
algorithm. Note that it is similar to the one proposed by Lacoste-Julien et al. [15] but with an
additional term due to the concavity of the loss.
5
Experiments
We applied our method to the task of learning a good similarity measure for aligning audio signals. In this field researchers have spent a lot of efforts in designing well-suited and meaningful
features [12, 4]. But the problem of combining these features for aligning temporal sequences is
still challenging. For simplicity, we took W diagonal for our experiments.
5.1
Dataset of Kirchhoff and Lerch [14]
Dataset description. First, we applied our method on the dataset of Kirchhoff and Lerch [14]. In
this dataset, pairs of aligned examples (Ai , B i ) are artificially created by stretching an original audio
signal. That way, the groundtruth alignment Y i is known and thus the data falls into our setting A
more precise description of the dataset can be found in [14].
The N = 60 pairs are stretched along two different tempo curves. Each signal is made of 30s
of music divided in frames of 46ms with a hopsize of 23ms, thus leading to a typical length of the
signals of T ? 1300 in our setting. We keep p = 11 features that are simple to implement and known
to perform well for alignment tasks [14]. Those were: five MFCC [8] (labeled M 1, . . . , M 5 in
Fig. 3), the spectral flatness (SF), the spectral centroid (SC), the spectral spread (SS), the maximum
of the envelope (Max), and the power level of each frame (Pow), see [14] for more details on the
computation of the features. We normalize each feature by subtracting the median value and dividing
by the standard deviation to the median, as audio data are subject to outliers.
6
0.2
?abs (s)
0.15
0.1
0.05
0
W M PowM1 SC M4 SR SF M3 Max SS M2 M5
Figure 3: Comparison of performance between individual features and the learned metric. Error
bars for the performance of the learned metric were determined with the best and the worst performance on 5 different experiments. W denotes the learned combination using our method, and M
the best MFCC combination.
Experiments. We conducted the following experiment: for each individual feature, we perform
alignment using dynamic time warping algorithm and evaluate the performance of this single feature
in terms of losses typically used to asses performance in this setting [14]. In Fig. 3, we report the
results of these experiments.
Then, we plug these data into our method, using the Hamming loss to learn a linear positive combination of these features. The result is reported in Fig. 3. Thus, combining these features on this
dataset yields to better performances than only considering a single feature.
For completeness, we also conducted the experiments using the standard 13 first MFCCs coefficients
and their first and second order derivatives as features. These results competed with the best learned
combination of the handcrafted features. Namely, in terms of the ?abs loss, they perform at 0.046
seconds. Note that these results are slightly worse than the best single handcrafted feature, but better
than the best MFCC coefficient used as a feature.
As a baseline, we also compared ourselves against the uniform combination of handcrafted features
(the metric being the identity matrix). The results are off the charts on Fig. 3 with ?abs at 4.1 seconds
(individual values ranging from 1.4 seconds to 7.4 seconds).
5.2
Chorales dataset
Dataset. The Bach 10 dataset3 consists in ten J. S. Bach?s Chorales (small quadriphonic pieces).
For each Chorale, a MIDI reference file corresponding to the ?score?, or basically a representation
of the partition. The alignments between the MIDI files and the audio file are given, thus we have
converted these MIDI files into audio following what is classically done for alignment (see e.g, [10]).
That way we fall into the audio-to-audio framework in which our technique apply. Each piece of
music is approximately 25s long, leading to similar signal length (T ? 1300).
Experiments. We use the same features as in Sec. 5.1. As depicted in Fig. 4, the optimization with
Hamming loss performs poorly on this dataset. In fact, the best individual feature performance is far
better than the performance of the learned W . Thus metric learning with the ?practical? Hamming
loss performs much worse than the best single feature.
Then, we conducted the same learning experiment with the symetrized area loss `S . The resulting
learned parameter is far better than the one learned using the Hamming loss. We get a performance
that is similar to the one of the best feature. Note that these features were handcrafted and reaching
their performance on this hard task with only a few training instances is already challenging.
3
http://music.cs.northwestern.edu/data/Bach10.html.
7
8
4
?
abs
(s)
6
2
0
(1) (2) (3) (4) (5) (6)
Figure 4: Performance of our algorithms on the Chorales dataset. From left to right: (1) Best
single feature, (2) Best learned combination of features using the symmetrized area loss `S , (3)
Best combination of MFCC using SAL and DT obtained via SDP (see footnote in section 3) (4)
Best combination of MFCC and derivatives learned with `S , (5) Best combination of MFCCs and
derivatives learned with Hamming loss, (6) Best combination of features of [14] using Hamming
loss.
In Fig. 2, we have depicted the result, for a learned parameter W , of the loss augmented decoding
performed either using the area. As it is known for structured SVM, this represents the most violated
constraint [25]. We can see that the most violated constraint for the Hamming loss leads to an alignment which is totally unrelated to the groundtruth alignment whereas the one for the symmetrized
area loss is far closer and much more discriminative.
5.3
Feature selection
Last, we conducted feature selection experiments over the same datasets. Starting from low level
features, namely the 13 leading MFCCs coefficients and their first two derivatives, we learn a linear
combination of these that achieves good alignment performance in terms of the area loss. Note
that very little musical prior knowledge is put into these. Moreover we either improve on the best
handcrafted feature on the dataset of [14] or perform similarly. On both datasets, the performance
of learned combination of handcrafted features performed similarly to the combination of these 39
MFCCs coefficients.
6
Conclusion
In this paper, we have presented a structured prediction framework for learning the metric for temporal alignment problems. We are able to combine hand-crafted features, as well as building automatically new state-of-the-art features from basic low-level information with little expert knowledge.
Technically, this is made possible by considering a loss beyond the usual Hamming loss which is
typically used because it is ?practical? within a structured prediction framework (linear in the output
representation).
The present work may be extended in several ways, the main one being to consider cases where
only partial information about the alignments is available. This is often the case in music [4] or
bioinformatics applications. Note that, similarly to Lajugie et al. [16] a simple alternating optimization between metric learning and constrained alignment provide a simple first solution, which could
probably be improved upon.
Acknowledgements. The authors acknowledge the support of the European Research Council
(SIERRA project 239993), the GARGANTUA project funded by the Mastodons program of CNRS
and the Airbus foundation through a PhD fellowship. Thanks to Piotr Bojanowski, for helpful
discussions. Warm thanks go to Arshia Cont and Philippe Cuvillier for sharing their knowledge
about audio processing, and to Holger Kirchhoff and Alexander Lerch for their dataset.
8
References
[1] J. Aach and G. M. Church. Aligning gene expression time series with time warping algorithms. Bioinformatics, 17(6):495?508, 2001.
[2] C. Banderier and S. Schwer. Why Delannoy numbers? Journal of statistical planning and inference, 135
(1):40?54, 2005.
[3] T. S. Caetano, J. J. McAuley, L. Cheng, Q. V. Le, and A. J. Smola. Learning graph matching. IEEE Trans.
on PAMI, 31(6):1048?1058, 2009.
[4] A. Cont, D. Schwarz, N. Schnell, C. Raphael, et al. Evaluation of real-time audio-to-score alignment. In
Proc. ISMIR, 2007.
[5] M. Cuturi, J.-P. Vert, O. Birkenes, and T. Matsui. A kernel for time series based on global alignments. In
Proc. ICASSP, volume 2, pages II?413. IEEE, 2007.
[6] S. Dixon and G. Widmer. Match: A music alignment tool chest. In Proc. ISMIR, pages 492?497, 2005.
[7] M. Frank and P. Wolfe. An algorithm for quadratic programming. Naval research logistics quarterly, 3
(1-2):95?110, 1956.
[8] B. Gold, N. Morgan, and D. Ellis. Speech and audio signal processing: processing and perception of
speech and music. John Wiley & Sons, 2011.
[9] R. Hamming. Error detecting and error correcting codes. Bell system technical journal, 29(2), 1950.
[10] N. Hu, R. B. Dannenberg, and G. Tzanetakis. Polyphonic audio matching and alignment for music
retrieval. Computer Science Department, page 521, 2003.
[11] T. Joachims, T. Finley, and C.-N. J. Yu. Cutting-plane training of structural SVMs. Machine Learning,
77(1):27?59, 2009.
[12] C. Joder, S. Essid, and G. Richard. Learning optimal features for polyphonic audio-to-score alignment.
IEEE Trans. on Audio, Speech, and Language Processing, 21(10):2118?2128, 2013.
[13] J. Keshet, S. Shalev-Shwartz, Y. Singer, and D. Chazan. A large margin algorithm for speech-to-phoneme
and music-to-score alignment. IEEE Transactions on Audio, Speech, and Language Processing, 15(8):
2373?2382, 2007.
[14] H. Kirchhoff and A. Lerch. Evaluation of features for audio-to-audio alignment. Journal of New Music
Research, 40(1):27?41, 2011.
[15] S. Lacoste-Julien, M. Jaggi, M. Schmidt, P. Pletscher, et al. Block-coordinate Frank-Wolfe optimization
for structural SVMs. In Proc. ICML, 2013.
[16] R. Lajugie, F. Bach, and S. Arlot. Large-margin metric learning for constrained partitioning problems. In
Proc. ICML, 2014.
[17] B. McFee and G. R. Lanckriet. Metric learning to rank. In Proc. ICML, pages 775?782, 2010.
[18] M. M?uller. Information retrieval for music and motion. Springer, 2007.
[19] H. Sakoe and S. Chiba. Dynamic programming algorithm optimization for spoken word recognition.
Acoustics, Speech and Signal Processing, IEEE Transactions on, 26(1):43?49, 1978.
[20] S. Shalev-Shwartz, Y. Singer, N. Srebro, and A. Cotter. Pegasos: Primal estimated sub-gradient solver for
SVM. Mathematical Programming, 127(1):3?30, 2011.
[21] M. Szummer, P. Kohli, and D. Hoiem. Learning CRFs using graph cuts. In Proc. CVPR. 2008.
[22] B. Taskar, D. Koller, and C. Guestrin. Max-margin Markov networks. Adv. NIPS, 2003.
[23] J. D. Thompson, F. Plewniak, and O. Poch. Balibase: a benchmark alignment database for the evaluation
of multiple alignment programs. Bioinformatics, 15(1):87?88, 1999.
[24] A. Torres, A. Cabada, and J. J. Nieto. An exact formula for the number of alignments between two dna
sequences. Mitochondrial DNA, 14(6):427?430, 2003.
[25] I. Tsochantaridis, T. Joachims, T. Hofmann, Y. Altun, and Y. Singer. Large margin methods for structured
and interdependent output variables. Journal of Machine Learning Research, 6(9):1453?1484, 2005.
[26] K. Q. Weinberger and L. K. Saul. Distance metric learning for large margin nearest neighbor classification.
Journal of Machine Learning Research, 10:207?244, 2009.
9
| 5383 |@word kohli:1 version:1 norm:2 km:1 grey:2 hu:1 tr:30 mcauley:1 epartement:1 series:8 score:7 hoiem:1 ecole:1 must:2 john:1 realistic:1 partition:2 hofmann:1 polyphonic:2 v:1 generative:1 plane:2 beginning:1 parametrization:1 completeness:2 detecting:1 five:2 mathematical:1 along:1 welldefined:1 prove:1 consists:1 combine:1 introduce:2 sakoe:2 pairwise:2 indeed:1 planning:1 sdp:2 decreasing:1 automatically:1 nieto:1 little:2 cardinality:1 increasing:1 considering:2 project:3 totally:1 notation:1 unrelated:1 moreover:1 solver:1 what:1 argmin:1 spoken:1 finding:1 temporal:5 every:1 y3:3 concave:5 mitochondrial:1 exactly:1 k2:1 partitioning:3 arlot:3 positive:3 local:1 path:5 approximately:1 pami:1 inria:5 might:1 challenging:3 matsui:1 unique:1 practical:2 block:2 definite:1 implement:1 mcfee:1 area:20 empirical:1 bell:1 vert:1 matching:6 word:1 symbolic:1 get:3 pegasos:1 close:2 selection:3 tsochantaridis:2 altun:1 put:2 context:7 optimize:1 map:2 yt:4 crfs:1 go:3 straightforward:1 starting:1 l:1 convex:6 rectangular:1 thompson:1 simplicity:1 correcting:1 m2:1 dominate:1 deriving:1 coordinate:3 exact:2 programming:5 designing:1 lanckriet:1 wolfe:5 crossing:1 recognition:2 cut:2 labeled:1 database:1 taskar:2 solved:1 worst:1 caetano:1 adv:1 rescaled:1 yk:2 sal:3 cuturi:1 schnell:1 dynamic:7 solving:2 technically:1 upon:1 triangle:1 easily:2 joint:2 kirchhoff:4 icassp:1 various:1 represented:2 regularizer:1 informatique:1 describe:1 sc:2 shalev:2 encoded:2 larger:1 solve:5 supplementary:3 kai:1 valued:1 otherwise:1 s:2 triangular:1 cvpr:1 unseen:1 sequence:7 eigenvalue:1 took:1 propose:8 subtracting:1 maximal:1 product:1 fr:4 raphael:1 relevant:2 combining:2 aligned:1 poorly:1 gold:1 description:2 frobenius:2 normalize:1 sierra:2 tions:1 derive:1 spent:1 damien:2 nearest:2 eq:17 solves:2 dividing:1 predicted:1 c:1 come:1 quantify:1 drawback:1 annotated:2 stochastic:2 hull:2 bojanowski:1 func2:1 material:3 southwest:1 tzanetakis:1 elementary:1 proximity:1 delannoy:2 bj:8 major:1 achieves:1 proc:7 council:1 schwarz:1 largest:1 successfully:1 tool:1 cotter:1 minimization:2 uller:1 aim:2 modified:1 normale:1 reaching:1 pn:2 shelf:1 focus:2 competed:1 improvement:1 naval:1 joachim:2 rank:1 centroid:1 baseline:1 sense:1 helpful:1 inference:1 cnrs:3 typically:2 koller:1 interested:1 issue:1 dual:3 aforementioned:1 html:1 classification:1 constrained:3 special:1 art:1 equal:1 once:1 field:1 having:1 piotr:1 kw:2 represents:1 holger:1 unsupervised:1 k2f:7 yu:1 icml:3 others:1 report:2 richard:1 few:1 neighbour:1 airbus:1 individual:4 m4:1 argmax:2 ourselves:1 attempt:1 ab:8 huge:1 evaluation:6 alignment:75 semidefinite:1 primal:1 closer:3 partial:2 necessary:1 dataset3:1 euclidean:1 minimal:1 instance:4 column:1 elli:1 balibase:1 maximization:2 deviation:2 uniform:1 gargantua:1 northeast:1 conducted:4 reported:1 thanks:2 off:2 decoding:7 together:2 possibly:1 classically:1 worse:2 corner:1 warped:1 derivative:4 leading:4 expert:1 potential:1 converted:1 de:1 sec:5 coefficient:4 dixon:1 explicitly:1 ranking:1 depends:1 piece:3 performed:2 try:1 lot:1 francis:2 sup:1 red:1 start:1 recover:1 contribution:1 ass:2 chart:1 musical:2 stretching:1 efficiently:2 phoneme:1 yield:1 dealt:1 basically:1 mastodon:1 mfcc:5 researcher:1 footnote:1 sharing:2 definition:1 against:1 associated:3 mi:1 hamming:22 couple:1 dataset:13 recall:1 knowledge:3 improves:1 ubiquitous:1 ta:25 dt:3 improved:1 yb:2 formulation:3 done:2 smola:1 hand:1 horizontal:3 quality:1 aj:2 rtb:2 grows:1 building:1 true:2 y2:27 poch:1 ltb:8 alternating:2 deal:1 widmer:1 mahalanobis:5 uniquely:1 noted:2 unnormalized:1 coincides:1 m:2 m5:1 btb:1 performs:2 motion:1 ranging:2 handcrafted:6 volume:1 extend:1 interpretation:1 significant:1 refer:1 ai:10 stretched:1 unconstrained:1 erieure:1 grid:1 similarly:4 language:2 mfccs:4 funded:1 similarity:5 align:1 aligning:4 add:1 jaggi:1 multivariate:3 irrelevant:1 binary:3 yi:5 morgan:1 guestrin:1 additional:2 greater:1 signal:15 semi:1 ii:1 multiple:1 flatness:1 technical:1 match:3 plug:2 bach:6 long:2 retrieval:3 divided:1 equally:1 a1:1 plugging:1 prediction:9 basic:3 pow:1 metric:26 lossaugmented:1 sometimes:2 kernel:1 whereas:4 fellowship:1 median:2 plewniak:1 crucial:1 envelope:1 sr:1 file:4 probably:1 subject:1 chest:1 structural:2 easy:2 lerch:4 fit:1 idea:1 expression:1 effort:1 suffer:1 speech:7 ptime:1 ten:1 svms:2 dna:2 http:1 notice:2 estimated:1 blue:1 write:1 discrete:1 waiting:2 threshold:1 lacoste:2 graph:5 groundtruth:5 ismir:2 cheng:1 quadratic:2 encountered:1 adapted:1 constraint:8 infinity:1 precisely:1 encodes:1 emi:1 argument:3 min:6 performing:1 structured:12 department:1 combination:15 dta:3 slightly:1 son:1 lp:2 outlier:1 beween:1 turn:1 lta:15 singer:3 symmetrizing:1 tractable:3 available:1 permit:1 apply:1 quarterly:1 spectral:3 dtb:2 tempo:1 schmidt:1 weinberger:2 ky1:1 symmetrized:6 rp:5 original:1 denotes:1 cf:1 hinge:1 music:11 build:1 especially:1 warping:5 move:7 already:1 usual:1 diagonal:4 surrogate:2 affinity:6 gradient:2 distance:3 link:2 parametrized:1 length:6 code:1 index:4 pointwise:1 cont:2 unfortunately:1 potentially:1 frank:5 trace:1 proper:1 perform:6 contributed:1 upper:1 vertical:4 datasets:4 markov:1 benchmark:1 acknowledge:1 descent:1 philippe:1 logistics:1 situation:1 extended:2 team:1 precise:1 y1:31 frame:2 arbitrary:1 community:1 bk:2 cast:4 namely:4 pair:8 acoustic:1 learned:13 nip:1 trans:2 beyond:2 bar:1 able:1 usually:1 perception:1 pioneering:1 tb:28 program:5 max:12 including:1 power:1 chazan:1 natural:2 rely:1 warm:1 predicting:1 pletscher:1 chorale:5 improve:2 julien:2 created:1 church:1 finley:1 kj:3 prior:1 literature:1 acknowledgement:1 interdependent:1 kf:1 loss:73 fully:1 northwestern:1 srebro:1 foundation:1 affine:1 sufficient:1 pi:1 row:1 ata:1 last:2 saul:2 fall:3 neighbor:1 absolute:1 chiba:2 dimension:2 plain:1 world:3 ending:1 valid:1 curve:2 overcome:1 preventing:1 concavity:1 made:2 author:1 far:5 transaction:2 compact:1 cutting:2 midi:3 keep:1 gene:1 global:2 overfitting:1 b1:1 discriminative:3 shwartz:2 why:1 learn:15 nature:1 rta:4 garreau:2 alg:2 european:1 artificially:1 diag:1 spread:1 main:1 big:1 n2:1 allowed:1 augmented:3 fig:12 referred:3 crafted:1 en:4 torres:1 wiley:1 sub:1 sf:2 untractable:1 formula:1 svm:3 keshet:4 phd:1 cartesian:1 margin:10 authorized:1 suited:1 depicted:4 lt:6 remi:1 adjustment:1 springer:1 corresponds:4 satisfies:1 goal:7 identity:1 labelled:1 experimentally:1 hard:1 sylvain:2 determined:3 typical:1 called:3 m3:1 meaningful:2 zone:2 formally:1 support:1 szummer:1 alexander:1 bioinformatics:4 relevance:1 violated:5 evaluate:1 audio:33 ykj:1 |
4,842 | 5,384 | Proximal Quasi-Newton for Computationally
Intensive `1-regularized M -estimators
Kai Zhong 1
Ian E.H. Yen 2
Inderjit S. Dhillon 2
Pradeep Ravikumar 2
2
Institute for Computational Engineering & Sciences
Department of Computer Science
University of Texas at Austin
[email protected], {ianyen,inderjit,pradeepr}@cs.utexas.edu
1
Abstract
We consider the class of optimization problems arising from computationally intensive `1 -regularized M -estimators, where the function or gradient values are
very expensive to compute. A particular instance of interest is the `1 -regularized
MLE for learning Conditional Random Fields (CRFs), which are a popular class
of statistical models for varied structured prediction problems such as sequence
labeling, alignment, and classification with label taxonomy. `1 -regularized MLEs
for CRFs are particularly expensive to optimize since computing the gradient values requires an expensive inference step. In this work, we propose the use of a
carefully constructed proximal quasi-Newton algorithm for such computationally
intensive M -estimation problems, where we employ an aggressive active set selection technique. In a key contribution of the paper, we show that the proximal
quasi-Newton method is provably super-linearly convergent, even in the absence
of strong convexity, by leveraging a restricted variant of strong convexity. In our
experiments, the proposed algorithm converges considerably faster than current
state-of-the-art on the problems of sequence labeling and hierarchical classification.
1
Introduction
`1 -regularized M -estimators have attracted considerable interest in recent years due to their ability
to fit large-scale statistical models, where the underlying model parameters are sparse. The optimization problem underlying these `1 -regularized M -estimators takes the form:
min f (w) := ?kwk1 + `(w),
w
(1)
where `(w) is a convex differentiable loss function. In this paper, we are particularly interested in the
case where the function or gradient values are very expensive to compute; we refer to these functions
as computationally intensive functions, or CI functions in short. A particular case of interest are `1 regularized MLEs for Conditional Random Fields (CRFs), where computing the gradient requires
an expensive inference step.
There has been a line of recent work on computationally efficient methods for solving (1), including
[2, 8, 13, 21, 23, 4]. It has now become well understood that it is key to leverage the sparsity
of the optimal solution by maintaining sparse intermediate iterates [2, 5, 8]. Coordinate Descent
(CD) based methods, like CDN [8], maintain the sparsity of intermediate iterates by focusing on an
active set of working variables. A caveat with such methods is that, for CI functions, each coordinate
update typically requires a call of inference oracle to evaluate partial derivative for single coordinate.
One approach adopted in [16] to address this is using Blockwise Coordinate Descent that updates
a block of variables at a time by ignoring the second-order effect, which however sacrifices the
convergence guarantee. Newton-type methods have also attracted a surge of interest in recent years
[5, 13], but these require computing the exact Hessian or Hessian-vector product, which is very
1
expensive for CI functions. This then suggests the use of quasi-Newton methods, popular instances
of which include OWL-QN [23], which is adapted from `2 -regularized L-BFGS, as well as Projected
Quasi-Newton (PQN) [4]. A key caveat with OWL-QN and PQN however is that they do not exploit
the sparsity of the underlying solution. In this paper, we consider the class of Proximal QuasiNewton (Prox-QN) methods, which we argue seem particularly well-suited to such CI functions, for
the following three reasons. Firstly, it requires gradient evaluations only once in each outer iteration.
Secondly, it is a second-order method, which has asymptotic superlinear convergence. Thirdly, it
can employ some active-set strategy to reduce the time complexity from O(d) to O(nnz), where d
is the number of parameters and nnz is the number of non-zero parameters.
While there has been some recent work on Prox-QN algorithms [2, 3], we carefully construct an
implementation that is particularly suited to CI `1 -regularized M -estimators. We carefully maintain the sparsity of intermediate iterates, and at the same time reduce the gradient evaluation time.
A key facet of our approach is our aggressive active set selection (which we also term a ?shrinking strategy?) to reduce the number of active variables under consideration at any iteration, and
correspondingly the number of evaluations of partial gradients in each iteration. Our strategy is
particularly aggressive in that it runs over multiple epochs, and in each epoch, chooses the next
working set as a subset of the current working set rather than the whole set; while at the end of an
epoch, allows for other variables to come in. As a result, in most iterations, our aggressive shrinking
strategy only requires the evaluation of partial gradients in the current working set. Moreover, we
adapt the L-BFGS update to the shrinking procedure such that the update can be conducted without
any loss of accuracy caused by aggressive shrinking. Thirdly, we store our data in a feature-indexed
structure to combine data sparsity as well as iterate sparsity.
[26] showed global convergence and asymptotic superlinear convergence for Prox-QN methods under the assumption that the loss function is strongly convex. However, this assumption is known to
fail to hold in high-dimensional sampling settings, where the Hessian is typically rank-deficient, or
indeed even in low-dimensional settings where there are redundant features. In a key contribution
of the paper, we provide provable guarantees of asymptotic superlinear convergence for Prox-QN
method, even without assuming strong-convexity, but under a restricted variant of strong convexity, termed Constant Nullspace Strong Convexity (CNSC), which is typically satisfied by standard
M -estimators.
To summarize, our contributions are twofold. (a) We present a carefully constructed proximal quasiNewton method for computationally intensive (CI) `1 -regularized M -estimators, which we empirically show to outperform many state-of-the-art methods on CRF problems. (b) We provide the
first proof of asymptotic superlinear convergence for Prox-QN methods without strong convexity,
but under a restricted variant of strong convexity, satisfied by typical M -estimators, including the
`1 -regularized CRF MLEs.
2
Proximal Quasi-Newton Method
A proximal quasi-Newton approach to solve M -estimators of the form (1) proceeds by iteratively
constructing a quadratic approximation of the objective function (1) to find the quasi-Newton direction, and then conducting a line search procedure to obtain the next iterate.
Given a solution estimate wt at iteration t, the proximal quasi-Newton method computes a descent
direction by minimizing the following regularized quadratic model,
1
dt = arg min g Tt ? + ?T Bt ? + ?kwt + ?k1
?
2
(2)
where g t = g(wt ) is the gradient of `(wt ) and Bt is an approximation to the Hessian of `(w). Bt
is usually formulated by the L-BFGS algorithm. This subproblem (2) can be efficiently solved by
randomized coordinate descent algorithm as shown in Section 2.2.
The next iterate is obtained from the backtracking line search procedure, wt+1 = wt + ?t dt , where
the step size ?t is tried over {? 0 , ? 1 , ? 2 , ...} until the Armijo rule is satisfied,
f (wt + ?t dt ) ? f (wt ) + ?t ??t ,
where 0 < ? < 1, 0 < ? < 1 and ?t = g Tt dt + ?(kwt + dt k1 ? kwt k1 ).
2
2.1
BFGS update formula
Bt can be efficiently updated by the gradients of the previous iterations according to the BFGS
update [18],
Bt?1 st?1 sTt?1 Bt?1
y t?1 y Tt?1
Bt = Bt?1 ?
+
(3)
sTt?1 Bt?1 st?1
y Tt?1 st?1
where st = wt+1 ? wt and y t = g t+1 ? g t
We use the compact formula for Bt [18],
?
Bt = B0 ? QRQT = B0 ? QQ,
where
?1
StT B0 St
Lt
? := RQT
,Q
Q := [ B0 St Yt ] , R :=
LTt
?Dt
St = [s0 , s1 , ..., st?1 ] , Yt = y 0 , y 1 , ..., y t?1
T
si?1 y j?1 if i > j
T
T
Dt = diag[s0 y 0 , ..., st?1 y t?1 ] and (Lt )i,j =
0
otherwise
In practical implementation, we apply Limited-memory-BFGS. It only uses the information of the
? have only size, d ? 2m and 2m ? d, respectively. B0 is
most recent m gradients, so that Q and Q
usually set as ?t I for computing Bt , where ?t = y Tt?1 st?1 /sTt?1 st?1 [18]. As will be discussed in
? is updated just on the rows(columns) corresponding to the working set, A. The
Section 2.3, Q(Q)
time complexity for L-BFGS update is O(m2 |A| + m3 ).
2.2
Coordinate Descent for Inner Problem
Randomized coordinate descent is carefully employed to solve the inner problem (2) by Tang and
Scheinberg [2]. In the update for coordinate j, d ? d + z ? ej , z ? is obtained by solving the onedimensional problem,
1
z ? = arg min (Bt )jj z 2 + ((g t )j + (Bt d)j )z + ?|(wt )j + dj + z|
z 2
This one-dimensional problem has a closed-form solution, z ? = ?c + S(c ? b/a, ?/a) ,where S is
the soft-threshold function and a = (Bt )jj , b = (g t )j + (Bt d)j and c = (wt )j + dj . For B0 = ?t I,
? j , where q Tj is the j-th row of Q and q
?j
the diagonal of Bt can be computed by (Bt )jj = ?t ? q Tj q
?
is the j-th column of Q. And the second term in b, (Bt d)j can be computed by,
?
? = ?t dj ? q T d,
(Bt d)j = ?t dj ? q Tj Qd
j
? := Qd.
? has only 2m dimension, it is fast to update (Bt d)j by q and d.
? In each
? Since d
where d
j
? d
??d
?+q
?j z?.
inner iteration, only dj is updated, so we have the fast update of d,
Since we only update the coordinates in the working set, the above algorithm has only computation
complexity O(m|A| ? inner iter), where inner iter is the number of iterations used for solving
the inner problem.
2.3
Implementation
In this section, we discuss several key implementation details used in our algorithm to speed up the
optimization.
Shrinking Strategy
In each iteration, we select an active or working subset A of the set of all variables: only the variables
in this set are updated in the current iteration. The complementary set, also called the fixed set, has
only values of zero and is not updated. The use of such a shrinking strategy reduces the overall
complexity from O(d) to O(|A|). Specifically, we (a) update the gradients just on the working set,
? just on the rows(columns) corresponding to the working set, and (c) compute the
(b) update Q (Q)
latest entries in Dt , ?t , Lt and StT St by just using the corresponding working set rather than the
whole set.
3
The key facet of our ?shrinking strategy? however is in aggressively shrinking the active set: at the
next iteration, we set the active set to be a subset of the previous active set, so that At ? At?1 . Such
an aggressive shrinking strategy however is not guaranteed to only weed out irrelevant variables.
Accordingly, we proceed in epochs. In each epoch, we progressively shrink the active set as above,
till the iterations seem to converge. At that time, we then allow for all the ?shrunk? variables to
come back and start a new epoch. Such a strategy was also called an -cooling strategy by Fan et
al. [14], where the shrinking stopping criterion is loose at the beginning, and progressively becomes
more strict each time all the variables are brought back. For L-BFGS update, when a new epoch
starts, the memory of L-BFGS is cleaned to prevent any loss of accuracy.
Because at the first iteration of each new epoch, the entire gradient over all coordinates is evaluated, the computation time for those iterations accounts for a significant portion of the total time
complexity. Fortunately, our experiments show that the number of epochs is typically between 3-5.
Inexact inner problem solution
Like many other proximal methods, e.g. GLMNET and QUIC, we solve the inner problem inexactly.
This reduces the time complexity of the inner problem dramatically. The amount of inexactness is
based on a heuristic method which aims to balance the computation time of the inner problem in each
outer iteration. The computation time of the inner problem is determined by the number of inner
iterations and the size of working set. Thus, we let the number of inner iterations, inner iter =
min{max inner, bd/|A|c}, where max inner = 10 in our experiment.
Data Structure for both model sparsity and data sparsity
In our implementation we take two sparsity patterns into consideration: (a) model sparsity, which
accounts for the fact that most parameters are equal to zero in the optimal solution; and (b) data
sparsity, wherein most feature values of any particular instance are zeros. We use a feature-indexed
data structure to take advantage of both sparsity patterns. Computations involving data will be timeconsuming if we compute over all the instances including those that are zero. So we leverage the
sparsity of data in our experiment by using vectors of pairs, whose members are the index and its
value. Traditionally, each vector represents an instance and the indices in its pairs are the feature
indices. However, in our implementation, to take both model sparsity and data sparsity into account,
we use an inverted data structure, where each vector represents one feature (feature-indexed) and
the indices in its pairs are the instance indices. This data structure facilitates the computation of the
gradient for a particular feature, which involves only the instances related to this feature.
We summarize these steps in the algorithm below. And a detailed algorithm is in Appendix 2.
Algorithm 1 Proximal Quasi-Newton Algorithm (Prox-QN)
Input: Dataset {x(i) , y (i) }i=1,2,...,N , termination criterion , ? and L-BFGS memory size m.
Output: w? converging to arg minw f (w).
? ? ?.
1: Initialize w ? 0, g ? ?`(w)/?w, working set A ? {1, 2, ...d}, and S, Y , Q, Q
2: while termination criterion is not satisfied or working set doesn?t contain all the variables do
3:
Shrink working set.
4:
if Shrinking stopping criterion is satisfied then
5:
Take all the shrunken variables back to working set and clean the memory of L-BFGS.
6:
Update Shrinking stopping criterion and continue.
7:
end if
8:
Solve inner problem (2) over working set and obtain the new direction d.
9:
Conduct line search based on Armijo rule and obtain new iterate w.
? and related matrices over working set.
10:
Update g, s, y, S, Y , Q, Q
11: end while
3
Convergence Analysis
In this section, we analyze the convergence behavior of proximal quasi-Newton method in the superlinear convergence phase, where the unit step size is chosen. To simplify the analysis, in this section,
we assume the inner problem is solved exactly and no shrinking strategy is employed. We also
provide the global convergence proof for Prox-QN method with shrinking strategy in Appendix 1.5.
In current literature, the analysis of proximal Newton-type methods relies on the assumption of
4
strongly convex objective function to prove superlinear convergence [3]; otherwise, only sublinear
rate can be proved [25]. However, our objective (1) is not strongly convex when the dimension is
very large or there are redundant features. In particular, the Hessian matrix H(w) of the smooth
function `(w) is not positive-definite. We thus leverage a recently introduced restricted variant of
strong convexity, termed Constant Nullspace Strong Convexity (CNSC) in [1]. There the authors
analyzed the behavior of proximal gradient and proximal Newton methods under such a condition.
The proximal quasi-Newton procedure in this paper however requires a subtler analysis, but in a key
contribution of the paper, we are nonetheless able to show asymptotic superlinear convergence of
the Prox-QN method under this restricted variant of strong convexity.
Definition 1 (Constant Nullspace Strong Convexity (CNSC)). A composite function (1) is said to
have Constant Nullspace Strong Convexity restricted to space T (CNSC-T ) if there is a constant
vector space T s.t. `(w) depends only on projT (w), i.e. `(w) = `(projT (w)), and its Hessian
satisfies
(4)
mkvk2 ? v T H(w)v ? M kvk2 , ?v ? T , ?w ? Rd
for some M ? m > 0, and
(5)
H(w)v = 0, ?v ? T ? , ?w ? Rd ,
where projT (w) is the projection of w onto T and T ? is the complementary space orthogonal to
T.
This condition can be seen to be an algebraic condition that is satisfied by typical M -estimators considered in high-dimensional settings. In this paper, we will abuse the use of CNSC-T for symmetric
matrices. We say a symmetric matrix H satisfies CNSC-T condition if H satisfies (4) and (5). In
?
the following theorems, we will denote the orthogonal basis of T as U ? Rd?d , where d? ? d is
the dimensionality of T space and U T U = I. Then the projection to T space can be written as
projT (w) = U U T w.
Theorem 1 (Asymptotic Superlinear Convergence). Assume ?2 `(w) and ?`(w) are Lipschitz continuous. Let Bt be the matrices generated by BFGS update (3). Then if `(w) and Bt satisfy CNSC-T
condition, the proximal quasi-Newton method has q-superlinear convergence:
kz t+1 ? z ? k ? o (kz t ? z ? k) ,
T
?
T ?
where z t = U wt , z = U w and w? is an optimal solution of (1).
The proof is given in Appendix 1.4. We prove it by exploiting the CNSC-T property. First, we
?
re-build our problem and algorithm on the reduced space Z = {z ? Rd |z = U T w}, where
the strong-convexity property holds. Then we prove the asymptotic superlinear convergence on Z
following Theorem 3.7 in [26].
Theorem 2. For Lipschitz continuous `(w), the sequence {wt } produced by the proximal quasiNewton Method in the super-linear convergence phase has
f (wt ) ? f (w? ) ? Lkz t ? z ? k,
(6)
?
T
?
T ?
where L = L` + ? d, L` is the Lipschitz constant of `(w), z t = U wt and z = U w .
The proof is also in Appendix 1.4. It is proved by showing that both the smooth part and the nondifferentiable part satisfy the modified Lipschitz continuity.
4
Application to Conditional Random Fields with `1 Penalty
In CRF problems, we are interested in learning a conditional distribution of labels y ? Y given
observation x ? X , where y has application-dependent structure such as sequence, tree, or table in
which label assignments have inter-dependency. The distribution is of the form
( d
)
X
1
Pw (y|x) =
exp
wk fk (y, x) ,
Zw (x)
k=1
where fk is the feature functions, wk is the associated weight, d is the number of feature functions
and Zw (x) is the partition function. Given a training data set {(xi , y i )}N
i=1 , our goal is to find the
optimal weights w such that the following `1 -regularized negative log-likelihood is minimized.
min f (w) = ?kwk1 ?
w
N
X
i=1
5
log Pw (y (i) |x(i) )
(7)
Since |Y|, the number of possible values y takes, can be exponentially large, the evaluation of
`(w) and the gradient ?`(w) needs application-dependent oracles to conduct the summation over
Y. For example, in sequence labeling problem, a dynamic programming oracle, forward-backward
algorithm, is usually employed to compute ?`(w). Such an oracle can be very expensive. In ProxQN algorithm for sequence labeling problem, the forward-backward algorithm takes O(|Y |2 N T ?
exp) time, where exp is the time for the expensive exponential computation, T is the sequence
length and Y is the possible label set for a symbol in the sequence. Then given the obtained oracle,
the evaluation of the partial gradients over the working set A has time complexity, O(Dnnz |A|T ),
where Dnnz is the average number of instances related to a feature. Thus when O(|Y |2 N T ? exp +
Dnnz |A|T ) > O(m3 + m2 |A|), the gradients evaluation time will dominate.
The following theorem gives that the `1 -regularized CRF MLEs satisfy the CNSC-T condition.
PN
Theorem 3. With `1 penalty, the CRF loss function, `(w) = ? i=1 log Pw (y (i) |x(i) ), satisfies
the CNSC-T condition with T = N ? , where N = {v ? Rd |?T v = 0} is a constant subspace of
Rd and ? ? Rd?(N |Y|) is defined as below,
h
i
?kn = fk (y l , x(i) ) ? E fk (y, x(i) )
where n = (i ? 1)|Y| + l, l = 1, 2, ...|Y| and E is the expectation over the conditional probability
Pw (y|x(i) ).
According to the definition of CNSC-T condition, the `1 -regularized CRF MLEs don?t satisfy
the classical strong-convexity condition when N has non-zero members, which happens in the
following two cases: (i) the exponential representation is not minimal [27], i.e. for any instance i there exist a non-zero vector a and a constant bi such that ha, ?(y, x(i) )i = bi , where
?(y, x) = [f1 (y, x(i) ), f2 (y, x(i) ), ..., fd (y, x(i) )]T ; (ii) d > N |Y|, i.e., the number of feature
functions is very large. The first case holds in many problems, like the sequence labeling and hierarchical classification discussed in Section 6, and the second case will hold in high-dimensional
problems.
5
Related Methods
There have been several methods proposed for solving `1 -regularized M -estimators of the form in
(7). In this section, we will discuss these in relation to our method.
Orthant-Wise Limited-memory Quasi-Newton (OWL-QN) introduced by Andrew and Gao [23]
extends L-BFGS to `1 -regularized problems. In each iteration, OWL-QN computes a generalized
gradient called pseudo-gradient to determine the orthant and the search direction, then does a line
search and a projection of the new iterate back to the orthant. Due to its fast convergence, it is
widely implemented by many software packages, such as CRF++, CRFsuite and Wapiti. But OWLQN does not take advantage of the model sparsity in the optimization procedure, and moreover Yu
et al. [22] have raised issues with its convergence proof.
Stochastic Gradient Descent (SGD) uses the gradient of a single sample as the search direction
at each iteration. Thus, the computation for each iteration is very fast, which leads to fast convergence at the beginning. However, the convergence becomes slower than the second-order method
when the iterate is close to the optimal solution. Recently, an `1 -regularized SGD algorithm proposed by Tsuruoka et al.[21] is claimed to have faster convergence than OWL-QN. It incorporates
`1 -regularization by using a cumulative `1 penalty, which is close to the `1 penalty received by the
parameter if it had been updated by the true gradient. Tsuruoka et al. do consider data sparsity, i.e.
for each instance, only the parameters related to the current instance are updated. But they too do
not take the model sparsity into account.
Coordinate Descent (CD) and Blockwise Coordinate Descent (BCD) are popular methods for `1 regularized problem. In each coordinate descent iteration, it solves an one-dimensional quadratic
approximation of the objective function, which has a closed-form solution. It requires the second
partial derivative with respect to the coordinate. But as discussed by Sokolovska et al., the exact
second derivative in CRF problem is intractable. So they instead use an approximation of the second
derivative, which can be computed efficiently by the same inference oracle queried for the gradient
evaluation. However, pure CD is very expensive because it requires to call the inference oracle for
the instances related to the current coordinate in each coordinate update. BCD alleviates this problem by grouping the parameters with the same x feature into a block. Then each block update only
6
needs to call the inference oracle once for the instances related to the current x feature. However,
it cannot alleviate the large number of inference oracle calls unless the data is very sparse such that
every instance appears only in very few blocks.
Proximal Newton method has proven successful on problems of `1 -regularized logistic regression
[13] and Sparse Invariance Covariance Estimation [5], where the Hessian-vector product can be
cheaply re-evaluated for each update of coordinate. However, the Hessian-vector product for CI
function like CRF requires the query of the inference oracle no matter how many coordinates are
updated at a time [17], which then makes the coordinate update on quadratic approximation as expensive as coordinate update in the original problem. Our proximal quasi-Newton method avoids
such problem by replacing Hessian with a low-rank matrix from BFGS update.
6
Numerical Experiments
We compare our approach, Prox-QN, with four other methods, Proximal Gradient (Prox-GD), OWLQN [23], SGD [21] and BCD [16]. For OWL-QN, we directly use the OWL-QN optimizer developed by Andrew et al.1 , where we set the memory size as m = 10, which is the same as that in
Prox-QN. For SGD, we implement the algorithm proposed by Tsuruoka et al. [21], and use cumulative `1 penalty with learning rate ?k = ?0 /(1 + k/N ), where k is the SGD iteration and N is
the number of samples. For BCD, we follow Sokolovska et al. [16] but with three modifications.
First, we add a line search procedure in each block update since we found it is required for convergence. Secondly, we apply shrinking strategy as discussed in Section 2.3. Thirdly, when the second
derivative for some coordinate is less than 10?10 , we set it to be 10?10 because otherwise the lack
of `2 -regularization in our problem setting will lead to a very large new iterate.
We evaluate the performance of Prox-QN method on two problems, sequence labeling and hierarchical classification. In particular, we plot the relative objective difference (f (wt )?f (w? ))/f (w? )
and the number of non-zero parameters (on a log scale) against time in seconds. More experiment
results, for example, the testing accuracy and the performance for different ??s, are in Appendix
5. All the experiments are executed on 2.8GHz Intel Xeon E5-2680 v2 Ivy Bridge processor with
1/4TB memory and Linux OS.
6.1
Sequence Labeling
In sequence labeling problems, each instance (x, y) = {(xt , yt )}t=1,2...,T is a sequence of T pairs
of observations and the corresponding labels. Here we consider the optical character recognition
(OCR) problem, which aims to recognize the handwriting words. The dataset 2 was preprocessed by
Taskar et al. [19] and was originally collected by Kassel [20], and contains 6877 words (instances).
We randomly divide the dataset into two part: training part with 6216 words and testing part with 661
words. The character label set Y consists of 26 English letters and the observations are characters
which are represented by images of 16 by 8 binary pixels as shown in Figure 1(a). We use degree
2 pixels as the raw features, which means all pixel pairs are considered. Therefore, the number of
raw features is J = 128 ? 127/2 + 128 + 1, including a bias. For degree 2 features, xtj = 1
only when both pixels are 1 and otherwise xtj = 0, where xtj is the j-th raw feature of xi . For
the feature functions, we use unigram feature functions 1(yt = y, xtj = 1) and bigram feature
functions 1(yt = y, yt+1 = y 0 ) with their associated weights, ?y,j and ?y,y0 , respectively. So
2
w = {?, ?} for ? ? R|Y |?J and ? ? R|Y |?|Y | and the total number of parameters, d = |Y | +
|Y | ? J = 215, 358.
the above feature functions, the potential
function can be specified as,
n Using
o
PT
PT ?1
P?w (y, x) = exp h?, t=1 (eyt xTt )i + h?, t=1 (eyt eTyt+1 )i ,where h?, ?i is the sum of elementwise product and ey ? R|Y | is an unit vector with 1 at y-th entry and 0 at other entries. The gradient
and the inference oracle are given in Appendix 4.1.
In our experiment, ? is set as 100, which leads to a relative high testing accuracy and an optimal
solution with a relative small number of non-zero parameters (see Appendix 5.2). The learning rate
?0 for SGD is tuned to be 2 ? 10?4 for best performance. In BCD, the unigram parameters are
grouped into J blocks according to the x features while the bigram parameters are grouped into one
block. Our proximal quasi-Newton method can be seen to be much faster than the other methods.
1
2
http://research.microsoft.com/en-us/downloads/b1eb1016-1738-4bd5-83a9-370c9d498a03/
http://www.seas.upenn.edu/ taskar/ocr/
7
Sequence?Labelling?nnz?100
Sequence?Labelling?100
BCD
OWL?QN
Prox?GD
Prox?QN
SGD
5
?2
10
nnz
Relative?objective?difference
10
?4
10
BCD
OWL?QN
Prox?GD
Prox?QN
SGD
?6
10
?8
10
(a) Graphical model of OCR
4
10
0
500
1000
time(s)
3
10
1500
0
(b) Relative Objective Difference
500
1000
time(s)
1500
(c) Non-zero Parameters
Figure 1: Sequence Labeling Problem
6.2
Hierarchical Classification
In hierarchical classification problems, we have a label taxonomy, where the classes are grouped
into a tree as shown in Figure 2(a). Here y ? Y is one of the leaf nodes. If we have totally K
classes (number of nodes) and J raw features, then the number of parameters is d = K ? J. Let
W ? RK?J denote the weights. The feature function corresponding to Wk,j is fk,j (y, x) = 1[k ?
Path(y)]xj , wherenk ? Path(y) means
o class k is an ancestor of y or y itself. The potential function is
P
T
?
PW (y, x) = exp
w x where wT is the weight vector of k-th class, i.e. the k-th row
k
k?Path(y)
k
of W . The gradient and the inference oracle are given in Appendix 4.2.
The dataset comes from Task1 of the dry-run dataset of LSHTC13 . It has 4,463 samples, each with
J=51,033 raw features. The hierarchical tree has 2,388 classes which includes 1,139 leaf labels.
Thus, the number of the parameters d =121,866,804. The feature values are scaled by svm-scale
program in the LIBSVM package. We set ? = 1 to achieve a relative high testing accuracy and
high sparsity of the optimal solution. The SGD initial learning rate is tuned to be ?0 = 10 for best
performance. In BCD, parameters are grouped into J blocks according to the raw features.
Hierarchicial?Classification?nnz?1
BCD
OWL?QN
Prox?GD
Prox?QN
SGD
0
10
BCD
OWL?QN
Prox?GD
Prox?QN
SGD
5
10
?2
10
nnz
Relative?objective?difference
Hierarchical?Classification?1
?4
10
4
10
?6
10
2000
(a) Label Taxonomy
4000
6000
time(s)
8000
500
10000
(b) Relative Objective Difference
1000 1500 2000 2500 3000 3500
time(s)
(c) Non-zero Parameters
Figure 2: Hierarchical Classification Problem
As both Figure 1(b),1(c) and Figure 2(b),2(c) show, Prox-QN achieves much faster convergence and
moreover obtains a sparse model in much less time.
Acknowledgement
This research was supported by NSF grants CCF-1320746 and CCF-1117055. P.R. acknowledges
the support of ARO via W911NF-12-1-0390 and NSF via IIS-1149803, IIS-1320894, IIS-1447574,
and DMS-1264033. K.Z. acknowledges the support of the National Initiative for Modeling and
Simulation fellowship
3
http://lshtc.iit.demokritos.gr/node/1
8
References
[1] I. E.H. Yen, C.-J. Hsieh, P. Ravikumar, and I. S. Dhillon. Constant Nullspace Strong Convexity and Fast
Convergence of Proximal Methods under High-Dimensional Settings. In NIPS 2014.
[2] X. Tang and K. Scheinberg. Efficiently Using Second Order Information in Large l1 Regularization Problems. arXiv:1303.6935, 2013.
[3] J. D. Lee, Y. Sun, and M. A. Saunders. Proximal newton-type methods for minimizing composite functions. In NIPS 2012.
[4] M. Schmidt, E. Van Den Berg, M.P. Friedlander, and K. Murphy. Optimizing costly functions with simple
constraints: A limited-memory projected Quasi-Newton algorithm. In Int. Conf. Artif. Intell. Stat., 2009.
[5] C.-J. Hsieh, M. A. Sustik, I. S. Dhillon, and P. Ravikumar. Sparse inverse covariance estimation using
quadratic approximation. In NIPS 2011.
[6] S. Boyd and L. Vandenberghe. Convex Optimization, Cambridge Univ. Press, Cambridge, U.K., 2003.
[7] P.-W. Wang and C.-J. Lin. Iteration Complexity of Feasible Descent Methods for Convex Optimization.
Technical report, Department of Computer Science, National Taiwan University, Taipei, Taiwan, 2013.
[8] G.-X. Yuan, K.-W. Chang, C.-J. Hsieh, and C.-J. Lin. A comparison of optimization methods and software for large-scale l1-regularized linear classification. Journal of Machine Learning Research (JMLR),
11:3183-3234, 2010.
[9] A. Agarwal, S. Negahban, and M. Wainwright. Fast Global Convergence Rates of Gradient Methods for
High-Dimensional Statistical Recovery. In NIPS 2010.
[10] K. Hou, Z. Zhou, A. M.-S. So, and Z.-Q. Luo. On the linear convergence of the proximal gradient method
for trace norm regularization. In NIPS 2014.
[11] L. Xiao and T. Zhang. A proximal-gradient homotopy method for the l1-regularized least-squares problem. In ICML 2012.
[12] P. Tseng and S. Yun. A coordinate gradient descent method for nonsmooth separable minimization, Math.
Prog. B. 117, 2009.
[13] G.-X. Yuan, C.-H. Ho, and C.-J. Lin. An improved GLMNET for l1-regularized logistic regression,
JMLR, 13:1999-2030, 2012.
[14] R.-E. Fan, K.-W. Chang, C.-J. Hsieh, X.-R. Wang, and C.-J. Lin. LIBLINEAR: A library for large linear
classification, JMLR, 9:1871-1874, 2008.
[15] A. J Hoffman. On approximate solutions of systems of linear inequalities. Journal of Research of the
National Bureau of Standards, 1952.
[16] N. Sokolovska, T. Lavergne, O. Cappe, and F. Yvon. Efficient Learning of Sparse Conditional Random
Fields for Supervised Sequence Labelling. arXiv:0909.1308, 2009.
[17] Y. Tsuboi, Y. Unno, H. Kashima, and N. Okazaki. Fast Newton-CG Method for Batch Learning of Conditional Random Fields, Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence,
2011.
[18] J. Nocedal and S. J. Wright. Numerical Optimization. Springer Series in Operations Research. Springer,
New York, NY, USA, 2nd edition, 2006.
[19] B. Taskar, C. Guestrin, and D. Koller. Max-margin markov networks. In NIPS 2003.
[20] R. Kassel. A Comparison of Approaches to On-line Handwritten Character Recognition. PhD thesis, MIT
Spoken Language Systems Group, 1995.
[21] Y. Tsuruoka, J. Tsujii, and S. Ananiadou. Stochastic gradient descent training for l1- regularized log-linear
models with cumulative penalty. In Proceedings of the Joint Conference of the 47th Annual Meeting of
the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP, pages
477-485, Suntec, Singapore, 2009.
[22] J. Yu, S.V.N. Vishwanathan, S. Gunter, and N. N. Schraudolph. A Quasi-Newton approach to nonsmooth
convex optimization problems in machine learning, JMLR, 11:1-57, 2010.
[23] G. Andrew and J. Gao. Scalable training of `1 -regularized log-linear models. In ICML 2007.
[24] J.E. Dennis and J.J. More. A characterization of superlinear convergence and its application to QuasiNewton methods. Math. Comp., 28(126):549560, 1974.
[25] K. Scheinberg and X. Tang. Practical Inexact Proximal Quasi-Newton Method with Global Complexity
Analysis. COR@L Technical Report at Lehigh University. arXiv:1311.6547, 2013
[26] J. D. Lee, Y. Sun, and M. A. Saunders. Proximal Newton-type methods for minimizing composite functions. arXiv:1206.1623, 2012
[27] M. J. Wainwright and M. I. Jordan. Graphical models, exponential families, and variational inference.
Technical Report 649, Dept. Statistics, Univ. California, Berkeley. 2003
9
| 5384 |@word pw:5 bigram:2 norm:1 owlqn:2 nd:1 termination:2 simulation:1 tried:1 covariance:2 hsieh:4 sgd:11 liblinear:1 initial:1 contains:1 series:1 tuned:2 task1:1 current:8 com:1 luo:1 si:1 attracted:2 bd:1 written:1 hou:1 numerical:2 partition:1 plot:1 update:24 progressively:2 intelligence:1 leaf:2 accordingly:1 beginning:2 short:1 caveat:2 iterates:3 math:2 node:3 characterization:1 firstly:1 zhang:1 constructed:2 kvk2:1 become:1 initiative:1 yuan:2 prove:3 consists:1 combine:1 inter:1 sacrifice:1 upenn:1 indeed:1 behavior:2 surge:1 totally:1 becomes:2 underlying:3 moreover:3 developed:1 spoken:1 guarantee:2 pseudo:1 berkeley:1 every:1 exactly:1 scaled:1 unit:2 grant:1 ice:1 positive:1 engineering:1 understood:1 path:3 abuse:1 downloads:1 acl:1 suggests:1 limited:3 bi:2 practical:2 testing:4 block:8 definite:1 implement:1 procedure:6 nnz:6 composite:3 projection:3 lshtc:1 word:4 boyd:1 onto:1 superlinear:11 selection:2 close:2 cannot:1 optimize:1 www:1 yt:6 crfs:3 timeconsuming:1 latest:1 convex:7 recovery:1 pure:1 m2:2 estimator:11 rule:2 dominate:1 vandenberghe:1 coordinate:22 traditionally:1 updated:8 qq:1 pt:2 exact:2 programming:1 us:2 expensive:10 particularly:5 recognition:2 cooling:1 taskar:3 subproblem:1 solved:2 wang:2 pradeepr:1 sun:2 convexity:15 complexity:9 dynamic:1 solving:4 f2:1 basis:1 joint:2 iit:1 represented:1 univ:2 fast:8 query:1 artificial:1 labeling:9 saunders:2 whose:1 heuristic:1 kai:1 solve:4 widely:1 say:1 otherwise:4 ability:1 statistic:1 itself:1 afnlp:1 a9:1 sequence:17 differentiable:1 advantage:2 propose:1 aro:1 product:4 shrunken:1 till:1 alleviates:1 achieve:1 ivy:1 exploiting:1 convergence:27 sea:1 converges:1 andrew:3 stat:1 received:1 b0:6 solves:1 strong:15 implemented:1 c:1 involves:1 come:3 qd:2 direction:5 shrunk:1 stochastic:2 owl:11 require:1 f1:1 alleviate:1 homotopy:1 secondly:2 summation:1 hold:4 stt:5 considered:2 wright:1 exp:6 optimizer:1 achieves:1 estimation:3 label:9 utexas:2 bridge:1 grouped:4 hoffman:1 minimization:1 gunter:1 brought:1 mit:1 super:2 aim:2 rather:2 modified:1 pn:1 ej:1 zhong:1 zhou:1 rank:2 likelihood:1 cg:1 inference:11 dependent:2 stopping:3 typically:4 pqn:2 bt:23 entire:1 relation:1 ancestor:1 koller:1 quasi:19 interested:2 provably:1 pixel:4 arg:3 classification:11 overall:1 issue:1 art:2 raised:1 initialize:1 field:5 once:2 construct:1 equal:1 sampling:1 represents:2 yu:2 icml:2 minimized:1 report:3 nonsmooth:2 simplify:1 employ:2 few:1 randomly:1 recognize:1 kwt:3 national:3 intell:1 xtj:4 murphy:1 phase:2 eyt:2 maintain:2 microsoft:1 interest:4 fd:1 evaluation:8 alignment:1 analyzed:1 pradeep:1 tj:3 partial:5 minw:1 orthogonal:2 unless:1 indexed:3 conduct:2 tree:3 divide:1 re:2 yvon:1 minimal:1 instance:16 column:3 soft:1 xeon:1 facet:2 modeling:1 w911nf:1 assignment:1 subset:3 entry:3 successful:1 zhongkai:1 conducted:1 too:1 gr:1 dependency:1 kn:1 proximal:27 considerably:1 chooses:1 mles:5 st:12 gd:5 international:1 randomized:2 negahban:1 lee:2 linux:1 thesis:1 aaai:1 satisfied:6 conf:1 derivative:5 aggressive:6 account:4 prox:21 bfgs:14 potential:2 wk:3 includes:1 int:1 matter:1 satisfy:4 caused:1 depends:1 closed:2 analyze:1 portion:1 start:2 yen:2 contribution:4 square:1 accuracy:5 conducting:1 efficiently:4 dry:1 raw:6 handwritten:1 produced:1 comp:1 processor:1 definition:2 inexact:2 against:1 nonetheless:1 dm:1 proof:5 associated:2 handwriting:1 dataset:5 proved:2 ianyen:1 popular:3 dimensionality:1 carefully:5 back:4 cappe:1 focusing:1 appears:1 originally:1 dt:8 supervised:1 follow:1 wherein:1 improved:1 evaluated:2 shrink:2 strongly:3 tsuruoka:4 just:4 until:1 working:18 dennis:1 replacing:1 o:1 lack:1 continuity:1 logistic:2 tsujii:1 artif:1 usa:1 effect:1 contain:1 true:1 ccf:2 regularization:4 aggressively:1 symmetric:2 dhillon:3 iteratively:1 criterion:5 generalized:1 yun:1 crf:9 tt:5 l1:5 cnsc:11 image:1 wise:1 consideration:2 variational:1 recently:2 empirically:1 exponentially:1 thirdly:3 discussed:4 elementwise:1 onedimensional:1 refer:1 significant:1 cambridge:2 queried:1 rd:7 fk:5 language:2 dj:5 had:1 add:1 recent:5 showed:1 optimizing:1 irrelevant:1 termed:2 store:1 claimed:1 inequality:1 binary:1 continue:1 kwk1:2 meeting:1 inverted:1 seen:2 guestrin:1 fortunately:1 employed:3 ey:1 converge:1 determine:1 redundant:2 sokolovska:3 ii:4 multiple:1 reduces:2 smooth:2 technical:3 faster:4 adapt:1 schraudolph:1 lin:4 ravikumar:3 mle:1 prediction:1 variant:5 involving:1 converging:1 regression:2 scalable:1 expectation:1 arxiv:4 iteration:23 agarwal:1 fellowship:1 zw:2 strict:1 deficient:1 facilitates:1 member:2 leveraging:1 incorporates:1 seem:2 jordan:1 call:4 leverage:3 intermediate:3 iterate:7 xj:1 fit:1 reduce:3 inner:18 intensive:5 texas:1 penalty:6 algebraic:1 hessian:9 proceed:1 jj:3 york:1 dramatically:1 detailed:1 amount:1 reduced:1 http:3 outperform:1 exist:1 ananiadou:1 nsf:2 singapore:1 arising:1 group:1 key:8 iter:3 four:1 threshold:1 prevent:1 preprocessed:1 clean:1 libsvm:1 backward:2 nocedal:1 year:2 sum:1 run:2 package:2 letter:1 inverse:1 extends:1 prog:1 family:1 appendix:8 guaranteed:1 convergent:1 fan:2 quadratic:5 oracle:12 annual:1 adapted:1 constraint:1 vishwanathan:1 software:2 bcd:10 speed:1 min:5 separable:1 optical:1 department:2 structured:1 according:4 character:4 y0:1 modification:1 s1:1 subtler:1 ltt:1 happens:1 den:1 restricted:6 computationally:6 scheinberg:3 discus:2 loose:1 fail:1 end:3 sustik:1 cor:1 adopted:1 operation:1 demokritos:1 apply:2 hierarchical:8 v2:1 ocr:3 kashima:1 schmidt:1 batch:1 ho:1 slower:1 original:1 bureau:1 include:1 graphical:2 bd5:1 maintaining:1 newton:26 exploit:1 taipei:1 k1:3 rqt:1 build:1 kassel:2 classical:1 objective:9 quic:1 strategy:13 costly:1 diagonal:1 said:1 gradient:32 subspace:1 outer:2 nondifferentiable:1 evaluate:2 argue:1 collected:1 tseng:1 reason:1 provable:1 taiwan:2 assuming:1 length:1 index:5 minimizing:3 balance:1 executed:1 taxonomy:3 blockwise:2 trace:1 negative:1 implementation:6 twenty:1 observation:3 markov:1 descent:13 orthant:3 varied:1 introduced:2 pair:5 cleaned:1 required:1 specified:1 california:1 nip:6 address:1 able:1 proceeds:1 usually:3 pattern:2 below:2 sparsity:19 summarize:2 tb:1 program:1 including:4 memory:8 max:3 wainwright:2 natural:1 regularized:25 library:1 acknowledges:2 epoch:9 literature:1 acknowledgement:1 friedlander:1 xtt:1 asymptotic:7 relative:8 loss:5 sublinear:1 proven:1 cdn:1 degree:2 s0:2 xiao:1 inexactness:1 cd:3 austin:1 row:4 supported:1 english:1 bias:1 allow:1 institute:1 correspondingly:1 fifth:1 sparse:7 ghz:1 van:1 dimension:2 cumulative:3 avoids:1 qn:27 computes:2 doesn:1 author:1 kz:2 projected:2 forward:2 approximate:1 compact:1 obtains:1 global:4 active:10 xi:2 don:1 search:7 continuous:2 table:1 ignoring:1 e5:1 constructing:1 diag:1 linearly:1 whole:2 edition:1 complementary:2 weed:1 intel:1 en:1 ny:1 shrinking:15 lehigh:1 exponential:3 suntec:1 jmlr:4 nullspace:5 ian:1 tang:3 formula:2 theorem:6 rk:1 xt:1 unigram:2 showing:1 symbol:1 svm:1 tsuboi:1 grouping:1 intractable:1 ci:7 phd:1 labelling:3 margin:1 suited:2 lt:3 backtracking:1 gao:2 cheaply:1 glmnet:2 inderjit:2 chang:2 springer:2 quasinewton:4 inexactly:1 relies:1 satisfies:4 conditional:7 goal:1 formulated:1 twofold:1 lipschitz:4 absence:1 considerable:1 feasible:1 typical:2 specifically:1 determined:1 wt:17 called:3 total:2 invariance:1 m3:2 select:1 berg:1 support:2 armijo:2 dept:1 |
4,843 | 5,385 | Discriminative Metric Learning by
Neighborhood Gerrymandering
Shubhendu Trivedi, David McAllester, Gregory Shakhnarovich
Toyota Technological Institute
Chicago, IL - 60637
{shubhendu,mcallester,greg}@ttic.edu
Abstract
We formulate the problem of metric learning for k nearest neighbor classification
as a large margin structured prediction problem, with a latent variable representing
the choice of neighbors and the task loss directly corresponding to classification
error. We describe an efficient algorithm for exact loss augmented inference, and
a fast gradient descent algorithm for learning in this model. The objective drives
the metric to establish neighborhood boundaries that benefit the true class labels
for the training points. Our approach, reminiscent of gerrymandering (redrawing
of political boundaries to provide advantage to certain parties), is more direct in
its handling of optimizing classification accuracy than those previously proposed.
In experiments on a variety of data sets our method is shown to achieve excellent
results compared to current state of the art in metric learning.
1
Introduction
Nearest neighbor classifiers are among the oldest and the most widely used tools in machine learning. Although nearest neighor rules are often successful, their performance tends to be limited by
two factors: the computational cost of searching for nearest neighbors and the choice of the metric
(distance measure) defining ?nearest?. The cost of searching for neighbors can be reduced with efficient indexing, e.g., [1, 4, 2] or learning compact representations, e.g., [13, 19, 16, 9]. We will not
address this issue here. Here we focus on the choice of the metric. The metric is often taken to be
Euclidean, Manhattan or ?2 distance. However, it is well known that in many cases these choices are
suboptimal in that they do not exploit statistical regularities that can be leveraged from labeled data.
This paper focuses on supervised metric learning. In particular, we present a method of learning a
metric so as to optimize the accuracy of the resulting nearest neighbor classifier.
Existing works on metric learning formulate learning as an optimization task with various constraints
driven by considerations of computational feasibility and reasonable, but often vaguely justified
principles [23, 8, 7, 22, 21, 14, 11, 18]. A fundamental intuition is shared by most of the work
in this area: an ideal distance for prediction is distance in the label space. Of course, that can not
be measured, since prediction of a test example?s label is what we want to use the similarities to
begin with. Instead, one could learn a similarity measure with the goal for it to be a good proxy
for the label similarity. Since the performance of kNN prediction often is the real motivation for
similarity learning, the constraints typically involve ?pulling? good neighbors (from the correct class
for a given point) closer while ?pushing? the bad neighbors farther away. The exact formulation of
?good? and ?bad? varies but is defined as a combination of proximity and agreement between labels.
We give a formulation that facilitates a more direct attempt to optimize for the kNN accuracy as
compared to previous work as far as we are aware. We discuss existing methods in more detail in
section 2, where we also place our work in context.
1
In the kNN prediction problem, given a point and a chosen metric, there is an implicit hidden
variable: the choice of k ?neighbors?. The inference of the predicted label from these k examples is
trivial, by simple majority vote among the associated labels. Given a query point, there can possibly
exist a very large number of choices of k points that might correspond to zero loss: any set of k
points with the majority of correct class will do. We would like a metric to ?prefer? one of these
?good? example sets over any set of k neighbors which would vote for a wrong class. Note that to
win, it is not necessary for the right class to account for all the k neighbors ? it just needs to get
more votes than any other class. As the number of classes and the value of k grow, so does the space
of available good (and bad) example sets.
These considerations motivate our approach to metric learning. It is akin to the common, albeit
negatively viewed, practice of gerrymandering in drawing up borders of election districts so as to
provide advantages to desired political parties, e.g., by concentrating voters from that party or by
spreading voters of opposing parties. In our case, the ?districts? are the cells in the Voronoi diagram
defined by the Mahalanobis metric, the ?parties? are the class labels voted for by the neighbors
falling in each cell, and the ?desired winner? is the true label of the training points associated with
the cell. This intuition is why we refer to our method as neighborhood gerrymandering in the title.
Technically, we write kNN prediction as an inference problem with a structured latent variable being
the choice of k neighbors. Thus learning involves minimizing a sum of a structural latent hinge loss
and a regularizer [3]. Computing structural latent hinge loss involves loss-adjusted inference ? one
must compute loss-adjusted values of both the output value (the label) and the latent items (the set
of nearest neighbors). The loss augmented inference corresponds to a choice of worst k neighbors
in the sense that while having a high average similarity they also correspond to a high loss (?worst
offending set of k neighbors?). Given the inherent combinatorial considerations, the key to such a
model is efficient inference and loss augmented inference. We give an efficient algorithm for exact
inference. We also design an optimization algorithm based on stochastic gradient descent on the
surrogate loss. Our approach achieves kNN accuracy higher than state of the art for most of the data
sets we tested on, including some methods specialized for the relevant input domains.
Although the experiments reported here are restricted to learning a Mahalanobis distance in an explicit feature space, the formulation allows for nonlinear similarity measures, such as those defined
by nonlinear kernels, provided computing the gradients of similarities with respect to metric parameters is feasible. Our formulation can also naturally handle a user-defined loss matrix on labels.
2
Related Work and Discussion
There is a large body of work on similarity learning done with the stated goal of improving kNN
performance. In much of the recent work, the objective can be written as a combination of some
sort of regularizer on the parameters of similarity, with loss reflecting the desired ?purity? of the
neighbors under learned similarity. Optimization then balances violation of these constraints with
regularization. The main contrast between this body of work and our approach here is in the form
of the loss.
A well known family of methods of this type is based on the Large Margin Nearest Neighbor
(LMNN) algorithm [22] . In LMNN, the constraints for each training point involve a set of predefined ?target neighbors? from correct class, and ?impostors? from other classes. The set of target
neighbors here plays a similar role to our ?best correct set of k neighbors? (h? in Section 4). However the set of target neighbors are chosen at the onset based on the euclidean distance (in absence
of a priori knowledge). Moreover as the metric is optimized, the set of ?target neighbors? is not dynamically updated. There is no reason to believe that the original choice of neighbors based on the
euclidean distance is optimal while the metric is updated. Also h? represents the closest neighbors
that have zero loss but they are not necessarily of the same class. In LMNN the target neighbors are
forced to be of the same class. In doing so it does not fully leverage the power of the kNN objective.
The role of imposters is somewhat similar to the role of the ?worst offending set of k neighbors? in
our method (b
h in Section 4). See Figure 2 for an illustration. Extensions of LMNN [21, 11] allow
for non-linear metrics, but retain the same general flavor of constraints. There is another extension
to LMNN that is more aligned to our work [20], in that they lift the constraint of having a static set
of neighbors chosen based on the euclidean distance and instead learn the neighborhood.
2
d
d
b
a
h
b
a
h
e
e
c
c
f
f
x
x
i
g
g
j
i
j
Figure 1: Illustration of objectives of LMNN (left) and our structured approach (right) for k = 3.
The point x of class blue is the query point. In LMNN, the target points are the nearest neighbors of
the same class, which are points a, b and c (the circle centered at x has radius equal to the farthest
of the target points i.e. point b). The LMNN objective will push all the points of the wrong class
that lie inside this circle out (points e, f, h, i, andj), while pulling in the target points to enforce the
margin. For our structured approach (right), the circle around x has radius equal to the distance of
the farthest of the three nearest neighbors irrespective of class. Our objective only needs to ensure
zero loss which is achieved by pushing in point a of the correct class (blue) while pushing out the
point having the incorrect class (point f ). Note that two points of the incorrect class lie inside the
circle (e, andf ), both being of class red. However f is pushed out and not e since it is farther from
x. Also see section 2.
The above family of methods may be contrasted with methods of the flavor as proposed in [23].
Here ?good? neighbors are defined as all similarly labeled points and each class is mapped into a
ball of a fixed radius, but no separation is enforced between the classes. The kNN objective does
not require that similarly labeled points be clustered together and consequently such methods try to
optimize a much harder objective for learning the metric.
In Neighborhood Component Analysis (NCA) [8], the piecewise-constant error of the kNN rule
is replaced by a soft version. This leads to a non-convex objective that is optimized via gradient
descent. This is similar to our method in the sense that it also attempts to directly optimize for
the choice of the nearest neighbor at the price of losing convexity. This issue of non-convexity
was partly remedied in [7], by optimization of a similar stochastic rule while attempting to collapse
each class to one point. While this makes the optimization convex, collapsing classes to distinct
points is unrealistic in practice. Another recent extension of NCA [18] generalizes the stochastic
classification idea to kNN classification with k > 1.
In Metric Learning to Rank (MLR)[14], the constraints involve all the points: the goal is to push
all the correct matches in front of all the incorrect ones. This again is not the same as requiring
correct classification. In addition to global optimization constraints on the rankings (such as mean
average precision for target class), the authors allow localized evaluation criteria such as Precision
at k, which can be used as a surrogate for classification accuracy for binary classification, but is a
poor surrogate for multi-way classification. Direct use of kNN accuracy in optimization objective is
briefly mentioned in [14], but not pursued due to the difficulty in loss-augmented inference. This is
because the interleaving technique of [10] that is used to perform inference with other losses based
inherently on contingency tables, fails for the multiclass case (since the number of data interleavings
could be exponential). We take a very different approach to loss augmented inference, using targeted
inference and the classification loss matrix, and can easily extend it to arbitrary number of classes.
A similar approach is taking in [15], where the constraints are derived from triplets of points formed
by a sample, correct and incorrect neighbors. Again, these are assumed to be set statically as an
input to the algorithm, and the optimization focuses on the distance ordering (ranking) rather than
accuracy of classification.
3
Problem setup
We are given N training examples X = {x1 , . . . , xN }, represented by a ?native? feature map,
xi ? Rd , and their class labels y = [y1 , . . . , yN ]T , with yi ? [R], where [R] stands for the set
3
{1, . . . , R}. We are also given the loss matrix ? with ?(r, r0 ) being the loss incurred by predicting
r0 when the correct class is r. We assume ?(r, r) = 0, and ?(r, r0 ), ?(r, r0 ) ? 0.
We are interested in Mahalanobis metrics
T
DW (x, xi ) = (x ? xi ) W (x ? xi ) ,
(1)
parameterized by positive semidefinite d ? d matrices W. Let h ? X be a set of examples in X.
For a given W we define the distance score of h w.r.t. a point x as
X
SW (x, h) = ?
DW (x, xj )
(2)
xj ?h
Hence, the set of k nearest neighbors of x in X is
hW (x) = argmax SW (x, h).
(3)
|h|=k
For the remainder we will assume that k is known and fixed. From any set h of k examples from X,
we can predict the label of x by (simple) majority vote:
yb (h) = majority{yj : xj ? h},
with ties resolved by a heuristic, e.g., according to 1NN vote. In particular, the kNN classifier
predicts yb(hW (x)). Due to this deterministic dependence between yb and h, we can define the
classification loss incured by a voting classifier when using the set h as
?(y, h) = ? (y, yb(h)) .
4
(4)
Learning and inference
P
One might want to learn W to minimize training loss i ? (yi , hW (xi )). However, this fails due
to the intractable nature of classification loss ?. We will follow the usual remedy: define a tractable
surrogate loss.
Here we note that in our formulation, the output of the prediction is a structured object hW , for
which we eventually report the deterministically computed yb. Structured prediction problems usually involve loss which is a generalization of the hinge loss; intuitively, it penalizes the gap between
score of the correct structured output and the score of the ?worst offending? incorrect output (the
one with the highest score and highest ?).
However, in our case there is no single correct output h, since in general many choices of h would
lead to correct yb and zero classification loss: any h in which the majority votes for the right class.
Ideally, we want SW to prefer at least one of these correct hs over all incorrect hs. This intuition
leads to the following surrogate loss definition:
L(x, y, W) = max [SW (x, h) + ?(y, h)]
h
?
max
h:?(y,h)=0
SW (x, h).
(5)
(6)
This is a bit different in spirit from the notion of margin sometimes encountered in ranking problems
where we want all the correct answers to be placed ahead of all the wrong ones. Here, we only care
to put one correct answer on top; it does not matter which one, hence the max in (6).
5
Structured Formulation
Although we have motivated this choice of L by intuitive arguments, it turns out that our problem is
an instance of a familiar type of problems: latent structured prediction [24], and thus our choice of
loss can be shown to form an upper bound on the empirical task loss ?.
First, we note that the score SW can be written as
*
+
X
T
SW (x, h) = W, ?
(x ? xj )(x ? xj )
,
xj ?h
4
(7)
where h?, ?i stands for the Frobenius inner product. Defining the feature map
X
?(x, h) , ?
(x ? xj )(x ? xj )T ,
(8)
xj ?h
we get a more compact expression hW, ?(x, h)i for (7).
Furthermore, we can encode the deterministic dependence between y and h by a ?compatibility?
function A(y, h) = 0 if y = yb(h) and A(y, h) = ?? otherwise. This allows us to write the joint
inference of y and (hidden) h performed by kNN classifier as
ybW (x), b
hW (x) = argmax [A(y, h) + hW, ?(x, h)i] .
(9)
h,y
This is the familiar form of inference in a latent structured model [24, 6] with latent variable h. So,
despite our model?s somewhat unusual property that the latent h completely determines the inferred
y, we can show the equivalence to the ?normal? latent structured prediction.
5.1
Learning by gradient descent
We define the objective in learning W as
min kWk2F + C
X
W
L (xi , yi , W) ,
(10)
i
where k ? k2F stands for Frobenius norm of a matrix.1 The regularizer is convex, but as in other
latent structured models, the loss L is non-convex due to the subtraction of the max in (6). To
optimize (10), one can use the convex-concave procedure (CCCP) [25] which has been proposed
specifically for latent SVM learning [24]. However, CCCP tends to be slow on large problems.
Furthermore, its use is complicated here due to the requirement that W be positive semidefinite
(PSD). This means that the inner loop of CCCP includes solving a semidefinite program, making
the algorithm slower still. Instead, we opt for a simpler choice, often faster in practice: stochastic
gradient descent (SGD), described in Algorithm 1.
Algorithm 1: Stochastic gradient descent
Input: labeled data set (X, Y ), regularization parameter C, learning rate ?(?)
initialize W(0) = 0
for t = 0, . . ., while not converged do
sample i ? [N ]
b
hi = argmaxh [SW(t) (xi , h) + ?(yi , h)]
h?i = argmaxh:?(yi ,h)=0 SW(t) (xi , h)
#
"
?SW (xi , h?i )
?SW (xi , b
hi )
?
?W =
(t)
?W
?W
W
W(t+1) = (1 ? ?(t))W(t) ? C?W
project W(t+1) to PSD cone
The SGD algorithm requires solving two inference problems (b
h and h? ), and computing the gradient
2
of SW which we address below.
5.1.1
Targeted inference of h?i
Here we are concerned with finding the highest-scoring h constrained to be compatible with a given
target class y. We give an O(N log N ) algorithm in Algorithm 2. Proof of its correctness and
complexity analysis is in Appendix.
1
We discuss other choices of regularizer in Section 7.
We note that both inference problems over h are done in leave one out settings, i.e., we impose an additional
constraint i ?
/ h under the argmax, not listed in the algorithm explicitly.
2
5
Algorithm 2: Targeted inference
Input: x, W, target class y, ? , Jties forbiddenK
Output: argmaxh:by(h)=y SW (x)
Let n? = d k+? (R?1)
e
// min.
R
h := ?
for j = 1, . . . , n? do
h := h ? argmin DW (x, xi )
required number of neighbors from y
xi : yi =y,i?h
/
for l = n? + 1, . . . , k do
define #(r) , |{i : xi ? h, yi = r}| // count selected neighbors from class r
h := h ?
argmin
DW (x, xi )
xi : yi =y, or #(yi )<#(y)??, i?h
/
return h
The intuition behind Algorithm 2 is as follows. For a given combination of R (number of classes)
and k (number of neighbors), the minimum number of neighbors from the target class y required to
allow (although not guarantee) zero loss, is n? (see Proposition 1 in the App. The algorithm first
includes n? highest scoring neighbors from the target class. The remaining k ? n? neighbors are
picked by a greedy procedure that selects the highest scoring neighbors (which might or might not
be from the target class) while making sure that no non-target class ends up in a majority.
When using Alg. 2 to find an element in H ? , we forbid ties, i.e. set ? = 1.
5.1.2
Loss augmented inference b
hi
Calculating the max term in (5) is known as loss augmented inference. We note that
n
o
0
0
0
0
hW,
?(x,
h
)i
+
?(y,
h
)
=
max
max
hW,
?(x,
h
)i
max
+
?(y,
y
)
0
0
0
?
0
h
y
h ?H (y )
(11)
= hW,?(x,h? (x,y0 ))i
which immediately leads to Algorithm 3, relying on Algorithm 2. The intuition: perform targeted
inference for each class (as if that were the target class), and the choose the set of neighbors for the
class for which the loss-augmented score is the highest. In this case, in each call to Alg. 2 we set
? = 0, i.e., we allow ties, to make sure the argmax is over all possible h?s.
Algorithm 3: Loss augmented inference
Input: x, W,target class y
Output: argmaxh [SW (x, h) + ?(y, h)]
for r ? {1, . . . , R} do
h(r) := h? (x, W, r, 1)
Let Value (r) := SW (x, h(r) ), + ?(y, r)
Let r? = argmaxr Value (r)
?
return h(r )
5.1.3
// using Alg. 2
Gradient update
Finally, we need to compute the gradient of the distance score. From (7), we have
X
?SW (x, h)
= ?(x, h) = ?
(x ? xj )(x ? xj )T .
?W
(12)
xj ?h
Thus, the update in Alg 1 has a simple interpretation, illustrated in Fig 2 on the right. For every
xi ? h? \ b
h, it ?pulls? xi closer to x. For every xi ? b
h \ h? , it ?pushes? it farther from x; these push
and pull refer to increase/decrease of Mahalanobis distance under the updated W. Any other xi ,
including any xi ? h? ? b
h, has no influence on the update. This is a difference of our approach from
6
LMNN, MLR etc. This is illustrated in Figure 2. In particular h? corresponds to points a, c and e,
whereas b
h corresponds to points c, e and f . Thus point a is pulled while point f is pushed.
Since the update does not necessarily preserve W as a PSD matrix, we enforce it by projecting W
onto the PSD cone, by zeroing negative eigenvalues. Note that since we update (or ?downdate?)
W each time by matrix of rank at most 2k, the eigendecomposition can be accomplished more
efficiently than the na??ve O(d3 ) approach, e.g., as in [17].
Using first order methods, and in particular gradient methods for optimization of non-convex functions, has been common across machine learning, for instance in training deep neural networks.
Despite lack (to our knowledge) of satisfactory guarantees of convergence, these methods are often
successful in practice; we will show in the next section that this is true here as well.
One might wonder if this method is valid for our objective that is not differentiable; we discuss this
briefly before describing experiments. A given x imposes a Voronoi-type partition of the space of
W into a finite number of cells; each cell is associated with a particular combination of b
h(x) and
h? (x) under the values of W in that cell. The score SW is differentiable (actually linear) on the
interior of the cell, but may be non-differentiable (though continuous) on the boundaries. Since the
boundaries between a finite number of cells form a set of measure zero, we see that the score is
differentiable almost everywhere.
6
Experiments
We compare the error of kNN classifiers using metrics learned with our approach to that with other
learned metrics. For this evaluation we replicate the protocol in [11], using the seven data sets in
Table 1. For all data sets, we report error of kNN classifier for a range of values of k; for each
k, we test the metric learned for that k. Competition to our method includes Euclidean Distance,
LMNN [22], NCA, [8], ITML [5], MLR [14] and GB-LMNN [11]. The latter learns non-linear
metrics rather than Mahalanobis.
For each of the competing methods, we used the code provided by the authors. In each case we tuned
the parameters of each method, including ours, in the same cross-validation protocol. We omit a few
other methods that were consistently shown in literature to be dominated by the ones we compare
to, such as ?2 distance, MLCC, M-LMNN. We also could not include ?2 -LMNN since code for it is
not available; however published results for k = 3 [11] indicate that our method would win against
?2 -LMNN as well.
Isolet and USPS have a standard training/test partition, for the other five data sets, we report the mean
and standard errors of 5-fold cross validation (results for all methods are on the same folds). We
experimented with different methods for initializing our method (given the non-convex objective),
including the euclidean distance, all zeros etc. and found the euclidean initialization to be always
worse. We initialize each fold with either the diagonal matrix learned by ReliefF [12] (which gives a
scaled euclidean distance) or all zeros depending on whether the scaled euclidean distance obtained
using ReliefF was better than unscaled euclidean distance. In each experiment, x are scaled by mean
and standard deviation of the training portion.3 The value of C is tuned on on a 75%/25% split of
the training portion. Results using different scaling methods are attached in the appendix.
Our SGD algorithm stops when the running average of the surrogate loss over most recent epoch no
longer descreases substantially, or after max. number of iterations. We use learning rate ?(t) = 1/t.
The results show that our method dominates other competitors, including non-linear metric learning
methods, and in some cases achieves results significantly better than those of the competition.
7
Conclusion
We propose a formulation of the metric learning for kNN classifier as a structured prediction problem, with discrete latent variables representing the selection of k neighbors. We give efficient algorithms for exact inference in this model, including loss-augmented inference, and devise a stochastic
gradient algorithm for learning. This approach allows us to learn a Mahalanobis metric with an objective which is a more direct proxy for the stated goal (improvement of classification by kNN rule)
3
For Isolet we also reduce dimensionality to 172 by PCA computed on the training portion.
7
k=3
DSLR
Amazon
Webcam
Caltech
800
157
10
800
958
10
800
295
10
800
1123
10
75.20 ?3.0
24.17 ?4.5
21.65 ?4.8
36.93 ?2.6
19.07 ?4.9
31.90 ?4.9
17.18 ?4.7
k=7
DSLR
60.13 ?1.9
26.72 ?2.1
26.72 ?2.1
24.01 ?1.8
33.83 ?3.3
30.27 ?1.3
21.34 ?2.5
?2.5
80.5 ?4.6
46.93 ?3.9
46.11 ?3.9
46.76 ?3.4
48.78 ?4.5
46.66 ?1.8
43.37 ?2.4
62.21
29.23
29.12
23.17
31.42
29.22
22.44
letters
76.45 ?6.2
25.44 ?4.3
25.44 ?4.3
33.73 ?5.5
22.32 ?2.5
36.94 ?2.6
21.61 ?5.9
k = 11
DSLR
5.89 ?0.4
4.09 ?0.1
2.86 ?0.2
15.54 ?6.8
6.52 ?0.8
6.04 ?2.8
3.05 ?0.1
73.87 ?2.8
23.64 ?3.4
23.64 ?3.4
36.25 ?13.1
22.28 ?3.1
40.06 ?6.0
22.28 ?4.9
64.61 ?4.2
30.12 ?2.9
30.07 ?3.0
24.32 ?3.8
30.48 ?1.4
30.69 ?2.9
24.11 ?3.2
Dataset
Isolet
USPS
letters
d
N
C
Euclidean
LMNN
GB-LMNN
MLR
ITML
NCA
ours
170
7797
26
8.66
4.43
4.13
6.61
7.89
6.16
4.87
256
9298
10
6.18
5.48
5.48
8.27
5.78
5.23
5.18
16
20000
26
4.79 ?0.2
3.26 ?0.1
2.92 ?0.1
14.25 ?5.8
4.97 ?0.2
4.71 ?2.2
2.32 ?0.1
Dataset
Isolet
USPS
letters
Euclidean
LMNN
GB-LMNN
MLR
ITML
NCA
ours
7.44
3.78
3.54
5.64
7.57
6.09
4.61
6.08
4.9
4.9
8.27
5.68
5.83
4.9
5.40 ?0.3
3.58 ?0.2
2.66 ?0.1
19.92 ?6.4
5.37 ?0.5
5.28 ?2.5
2.54 ?0.1
Dataset
Isolet
USPS
Euclidean
LMNN
GB-LMNN
MLR
ITML
NCA
ours
8.02
3.72
3.98
5.71
7.77
5.90
4.11
6.88
4.78
4.78
11.11
6.63
5.73
4.98
Amazon
?2.2
?2.0
?2.1
?2.1
?1.9
?2.7
?1.3
Amazon
56.27
15.59
13.56
23.05
13.22
16.27
10.85
?2.2
?1.9
?2.8
?4.6
?1.5
?3.1
Webcam
57.29
14.58
12.45
18.98
10.85
22.03
11.19
?6.3
?2.2
?4.6
?2.9
?3.1
?6.5
?3.3
Webcam
59.66
13.90
13.90
17.97
11.86
26.44
11.19
?5.5
?2.2
?1.0
?4.1
?5.6
?6.3
?4.4
Caltech
80.76
46.75
46.17
46.85
51.74
45.50
41.61
?3.7
?2.9
?2.8
?4.1
?2.8
?3.0
?2.6
Caltech
81.39
49.06
49.15
44.97
50.76
46.48
40.76
?4.2
?2.3
?2.8
?2.6
?1.9
?4.0
?1.8
Table 1: kNN error,for k=3, 7 and 11. Features were scaled by z-scoring. Mean and standard
deviation are shown for data sets on which 5-fold partition was used. Best performing methods are
shown in bold. Note that the only non-linear metric learning method in the above is GB-LMNN.
than previously proposed similarity learning methods. Our learning algorithm is simple yet efficient,
converging on all the data sets we have experimented upon in reasonable time as compared to the
competing methods.
Our choice of Frobenius regularizer is motivated by desire to control model complexity without
biasing towards a particular form of the matrix. We have experimented with alternative regularizers,
both the trace norm of W and the shrinkage towards Euclidean distance, kW ? Ik2F , but found both
to be inferior to kWk2F . We suspect that often the optimal W corresponds to a highly anisotropic
scaling of data dimensions, and thus bias towards I may be unhealthy.
The results in this paper are restricted to Mahalanobis metric, which is an appealing choice for a
number of reasons. In particular, learning such metrics is equivalent to learning linear embedding
of the data, allowing very efficient methods for metric search. Still, one can consider non-linear
embeddings x ? ?(x; w) and define the distance D in terms of the embeddings, for example, as
D(x, xi ) = k?(x) ? ?(xi )k or as ??(x)T ?(xi ). Learning S in the latter form can be seen as
learning a kernel with discriminative objective of improving kNN performance. Such a model would
be more expressive, but also more challenging to optimize. We are investigating this direction.
Acknowledgments
This work was partly supported by NSF award IIS-1409837.
8
References
[1] S. Arya, D. M. Mount, N. S. Netanyahu, R. Silverman, and A. Y. Wu. An optimal algorithm
for approximate nearest neighbor searching fixed dimensions. J. ACM, 45(6):891?923, 1998.
[2] A. Beygelzimer, S. Kakade, and J. Langford. Cover trees for nearest neighbor. In ICML, pages
97?104. ACM, 2006.
[3] B. E. Boser, I. M. Guyon, and V. N. Vapnik. A training algorithm for optimal margin classifiers.
In COLT, pages 144?152. ACM Press, 1992.
[4] M. Datar, N. Immorlica, P. Indyk, and V. S. Mirrokni. Locality-sensitive hashing scheme based
on p-stable distributions. In SoCG, pages 253?262. ACM, 2004.
[5] J. V. Davis, B. Kulis, J. Prateek, S. Suvrit, and D. Inderjeet. Information-theoretic metric
learning. pages 209?216, 2007.
[6] P. F. Felzenszwalb, R. B. Girshick, D. McAllester, and D. Ramanan. Object detection with
discriminatively trained part-based models. IEEE T. PAMI, 32(9):1627?1645, 2010.
[7] A. Globerson and S. Roweis. Metric learning by collapsing classes. In Y. Weiss, B. Sch?olkopf,
and J. Platt, editors, NIPS, pages 451?458, Cambridge, MA, 2006. MIT Press.
[8] J. Goldberger, S. Roweis, G. Hinton, and R. Salakhutdinov. Neighbourhood components analysis. In NIPS, 2004.
[9] Y. Gong and S. Lazebnik. Iterative quantization: A procrustean approach to learning binary
codes. In CVPR, pages 817?824. IEEE, 2011.
[10] T. Joachims. A support vector method for multivariate performance measures. In ICML, pages
377?384. ACM Press, 2005.
[11] D. Kedem, S. Tyree, K. Weinberger, F. Sha, and G. Lanckriet. Non-linear metric learning. In
NIPS, pages 2582?2590, 2012.
[12] K. Kira and A. Rendell. The feature selection problem: Traditional methods and a new algorithm). AAAI, 2:129?134, 1992.
[13] B. Kulis and T. Darrell. Learning to hash with binary reconstructive embeddings. NIPS,
22:1042?1050, 2009.
[14] B. McFee and G. Lanckriet. Metric learning to rank. In ICML, 2010.
[15] M. Norouzi, D. Fleet, and R. Salakhutdinov. Hamming distance metric learning. In NIPS,
pages 1070?1078, 2012.
[16] M. Norouzi and D. J. Fleet. Minimal loss hashing for compact binary codes. ICML, 1:2, 2011.
[17] P. Stange. On the efficient update of the singular value decomposition. PAMM, 8(1):10827?
10828, 2008.
[18] D. Tarlow, K. Swersky, I. Sutskever, and R. S. Zemel. Stochastic k-neighborhood selection for
supervised and unsupervised learning. ICML, 28:199?207, 2013.
[19] J. Wang, S. Kumar, and S.-F. Chang. Sequential projection learning for hashing with compact
codes. In ICML, 2010.
[20] J. Wang, A. Woznica, and A. Kalousis. Learning neighborhoods for metric learning. In ECMLPKDD, 2012.
[21] K. Q. Weinberger and L. K. Saul. Fast solvers and efficient implementations for distance metric
learning. In ICML, pages 1160?1167. ACM, 2008.
[22] K. Q. Weinberger and L. K. Saul. Distance metric learning for large margin nearest neighbor
classification. JMLR, 10:207?244, 2009.
[23] E. P. Xing, A. Y. Ng, M. I. Jordan, and S. Russell. Distance metric learning, with application
to clustering with side-information. In NIPS, pages 505?512. MIT Press, 2002.
[24] C.-N. J. Yu and T. Joachims. Learning structural svms with latent variables. In ICML, pages
1169?1176. ACM, 2009.
[25] A. L. Yuille, A. Rangarajan, and A. Yuille. The concave-convex procedure (cccp). NIPS,
2:1033?1040, 2002.
9
| 5385 |@word h:2 kulis:2 version:1 briefly:2 norm:2 replicate:1 decomposition:1 sgd:3 harder:1 offending:3 score:9 tuned:2 ours:4 imposter:1 existing:2 current:1 beygelzimer:1 goldberger:1 yet:1 reminiscent:1 must:1 written:2 chicago:1 partition:3 update:6 hash:1 pursued:1 selected:1 greedy:1 item:1 oldest:1 farther:3 tarlow:1 district:2 simpler:1 five:1 direct:4 incorrect:6 inside:2 multi:1 salakhutdinov:2 lmnn:21 relying:1 election:1 solver:1 begin:1 provided:2 moreover:1 project:1 what:1 prateek:1 argmin:2 substantially:1 finding:1 guarantee:2 every:2 voting:1 concave:2 tie:3 classifier:9 wrong:3 scaled:4 control:1 ramanan:1 farthest:2 omit:1 yn:1 platt:1 positive:2 before:1 tends:2 despite:2 mount:1 datar:1 pami:1 might:5 voter:2 initialization:1 dynamically:1 equivalence:1 challenging:1 limited:1 collapse:1 range:1 nca:6 acknowledgment:1 globerson:1 yj:1 practice:4 impostor:1 silverman:1 procedure:3 mcfee:1 area:1 empirical:1 significantly:1 projection:1 get:2 onto:1 interior:1 selection:3 put:1 context:1 influence:1 optimize:6 equivalent:1 map:2 deterministic:2 convex:8 formulate:2 amazon:3 immediately:1 rule:4 isolet:5 pull:2 dw:4 embedding:1 searching:3 handle:1 notion:1 updated:3 target:17 play:1 user:1 exact:4 losing:1 relieff:2 agreement:1 lanckriet:2 element:1 native:1 labeled:4 predicts:1 role:3 kedem:1 initializing:1 wang:2 worst:4 ordering:1 decrease:1 technological:1 highest:6 russell:1 mentioned:1 intuition:5 convexity:2 complexity:2 ideally:1 motivate:1 trained:1 shakhnarovich:1 solving:2 technically:1 negatively:1 upon:1 yuille:2 completely:1 usps:4 interleavings:1 easily:1 resolved:1 joint:1 various:1 represented:1 regularizer:5 forced:1 fast:2 describe:1 distinct:1 reconstructive:1 query:2 zemel:1 lift:1 neighborhood:7 heuristic:1 widely:1 cvpr:1 drawing:1 otherwise:1 knn:19 indyk:1 advantage:2 eigenvalue:1 differentiable:4 propose:1 product:1 remainder:1 relevant:1 aligned:1 loop:1 achieve:1 roweis:2 intuitive:1 frobenius:3 competition:2 olkopf:1 sutskever:1 convergence:1 regularity:1 requirement:1 darrell:1 rangarajan:1 leave:1 object:2 depending:1 gong:1 measured:1 nearest:15 predicted:1 involves:2 indicate:1 direction:1 radius:3 correct:15 stochastic:7 centered:1 mcallester:3 require:1 clustered:1 generalization:1 opt:1 proposition:1 adjusted:2 extension:3 proximity:1 around:1 normal:1 predict:1 achieves:2 label:13 spreading:1 combinatorial:1 title:1 sensitive:1 correctness:1 tool:1 argmaxr:1 mit:2 always:1 rather:2 shrinkage:1 encode:1 derived:1 focus:3 joachim:2 improvement:1 consistently:1 rank:3 political:2 contrast:1 sense:2 inference:25 voronoi:2 nn:1 unhealthy:1 typically:1 hidden:2 interested:1 selects:1 compatibility:1 issue:2 classification:16 among:2 colt:1 priori:1 art:2 constrained:1 initialize:2 equal:2 aware:1 having:3 ng:1 represents:1 kw:1 yu:1 k2f:1 icml:8 unsupervised:1 report:3 piecewise:1 inherent:1 few:1 preserve:1 ve:1 familiar:2 replaced:1 argmax:4 opposing:1 attempt:2 psd:4 detection:1 highly:1 evaluation:2 violation:1 semidefinite:3 behind:1 regularizers:1 predefined:1 closer:2 necessary:1 tree:1 euclidean:14 penalizes:1 desired:3 circle:4 girshick:1 minimal:1 instance:2 soft:1 cover:1 cost:2 deviation:2 wonder:1 successful:2 front:1 itml:4 reported:1 answer:2 varies:1 gregory:1 fundamental:1 forbid:1 retain:1 together:1 na:1 again:2 aaai:1 leveraged:1 possibly:1 choose:1 collapsing:2 worse:1 return:2 account:1 bold:1 includes:3 matter:1 explicitly:1 ranking:3 onset:1 performed:1 try:1 picked:1 doing:1 red:1 portion:3 sort:1 xing:1 complicated:1 minimize:1 il:1 voted:1 greg:1 accuracy:7 formed:1 efficiently:1 correspond:2 norouzi:2 drive:1 published:1 app:1 converged:1 dslr:3 definition:1 against:1 competitor:1 naturally:1 associated:3 proof:1 static:1 hamming:1 stop:1 dataset:3 concentrating:1 knowledge:2 dimensionality:1 actually:1 reflecting:1 higher:1 hashing:3 supervised:2 follow:1 wei:1 yb:7 formulation:7 done:2 though:1 furthermore:2 just:1 implicit:1 langford:1 expressive:1 nonlinear:2 lack:1 pulling:2 believe:1 requiring:1 true:3 remedy:1 regularization:2 hence:2 satisfactory:1 illustrated:2 mahalanobis:7 inferior:1 davis:1 criterion:1 procrustean:1 theoretic:1 rendell:1 lazebnik:1 consideration:3 common:2 specialized:1 mlr:6 winner:1 attached:1 anisotropic:1 extend:1 interpretation:1 refer:2 cambridge:1 rd:1 similarly:2 zeroing:1 stable:1 similarity:11 longer:1 etc:2 closest:1 multivariate:1 recent:3 optimizing:1 driven:1 certain:1 suvrit:1 binary:4 yi:9 accomplished:1 scoring:4 devise:1 seen:1 minimum:1 additional:1 somewhat:2 care:1 impose:1 caltech:3 purity:1 r0:4 subtraction:1 downdate:1 ii:1 match:1 faster:1 cross:2 cccp:4 award:1 feasibility:1 prediction:11 converging:1 metric:41 iteration:1 kernel:2 sometimes:1 achieved:1 cell:8 justified:1 addition:1 want:4 whereas:1 diagram:1 grow:1 argmaxh:4 singular:1 sch:1 sure:2 suspect:1 facilitates:1 spirit:1 jordan:1 call:1 structural:3 leverage:1 ideal:1 split:1 embeddings:3 concerned:1 variety:1 xj:12 competing:2 suboptimal:1 inner:2 idea:1 reduce:1 multiclass:1 fleet:2 whether:1 motivated:2 expression:1 pca:1 gb:5 akin:1 deep:1 involve:4 listed:1 svms:1 reduced:1 exist:1 nsf:1 blue:2 kira:1 write:2 discrete:1 woznica:1 key:1 falling:1 d3:1 vaguely:1 sum:1 cone:2 enforced:1 parameterized:1 everywhere:1 letter:3 swersky:1 place:1 family:2 reasonable:2 almost:1 wu:1 guyon:1 separation:1 prefer:2 appendix:2 scaling:2 pushed:2 bit:1 bound:1 hi:3 fold:4 encountered:1 ahead:1 constraint:10 dominated:1 argument:1 min:2 kumar:1 attempting:1 performing:1 statically:1 structured:13 according:1 combination:4 ball:1 poor:1 kalousis:1 andj:1 across:1 y0:1 appealing:1 kakade:1 making:2 intuitively:1 restricted:2 indexing:1 projecting:1 ik2f:1 taken:1 socg:1 previously:2 discus:3 eventually:1 turn:1 count:1 describing:1 tractable:1 end:1 unusual:1 available:2 generalizes:1 away:1 enforce:2 neighbourhood:1 alternative:1 weinberger:3 slower:1 original:1 top:1 remaining:1 ensure:1 include:1 running:1 ecmlpkdd:1 clustering:1 hinge:3 sw:17 pushing:3 calculating:1 exploit:1 establish:1 webcam:3 objective:15 sha:1 dependence:2 usual:1 diagonal:1 surrogate:6 mirrokni:1 traditional:1 gradient:12 win:2 distance:25 mapped:1 remedied:1 majority:6 seven:1 trivial:1 reason:2 code:5 illustration:2 minimizing:1 balance:1 setup:1 trace:1 stated:2 negative:1 design:1 implementation:1 perform:2 allowing:1 upper:1 arya:1 finite:2 descent:6 defining:2 hinton:1 y1:1 arbitrary:1 ttic:1 inferred:1 david:1 required:2 optimized:2 learned:5 boser:1 nip:7 address:2 usually:1 below:1 biasing:1 program:1 including:6 max:9 power:1 unrealistic:1 difficulty:1 predicting:1 representing:2 scheme:1 irrespective:1 epoch:1 literature:1 manhattan:1 loss:41 fully:1 discriminatively:1 localized:1 validation:2 eigendecomposition:1 contingency:1 incurred:1 proxy:2 imposes:1 principle:1 editor:1 tyree:1 netanyahu:1 unscaled:1 course:1 compatible:1 placed:1 supported:1 bias:1 allow:4 pulled:1 side:1 institute:1 neighbor:46 saul:2 taking:1 felzenszwalb:1 benefit:1 boundary:4 dimension:2 xn:1 stand:3 valid:1 author:2 party:5 far:1 approximate:1 compact:4 global:1 investigating:1 assumed:1 discriminative:2 xi:23 continuous:1 latent:14 search:1 triplet:1 iterative:1 why:1 table:3 learn:4 nature:1 inherently:1 improving:2 alg:4 excellent:1 necessarily:2 domain:1 protocol:2 main:1 motivation:1 border:1 kwk2f:2 body:2 augmented:10 x1:1 fig:1 slow:1 precision:2 fails:2 explicit:1 deterministically:1 exponential:1 lie:2 jmlr:1 toyota:1 learns:1 interleaving:1 hw:10 bad:3 experimented:3 svm:1 dominates:1 intractable:1 quantization:1 albeit:1 vapnik:1 sequential:1 push:4 margin:6 trivedi:1 gap:1 flavor:2 locality:1 desire:1 chang:1 corresponds:4 determines:1 acm:7 ma:1 goal:4 viewed:1 targeted:4 consequently:1 towards:3 shared:1 absence:1 feasible:1 price:1 specifically:1 contrasted:1 partly:2 vote:6 immorlica:1 support:1 latter:2 andf:1 tested:1 handling:1 |
4,844 | 5,386 | Fundamental Limits of Online and Distributed
Algorithms for Statistical Learning and Estimation
Ohad Shamir
Weizmann Institute of Science
[email protected]
Abstract
Many machine learning approaches are characterized by information constraints
on how they interact with the training data. These include memory and sequential
access constraints (e.g. fast first-order methods to solve stochastic optimization
problems); communication constraints (e.g. distributed learning); partial access
to the underlying data (e.g. missing features and multi-armed bandits) and more.
However, currently we have little understanding how such information constraints
fundamentally affect our performance, independent of the learning problem semantics. For example, are there learning problems where any algorithm which
has small memory footprint (or can use any bounded number of bits from each
example, or has certain communication constraints) will perform worse than what
is possible without such constraints? In this paper, we describe how a single set
of results implies positive answers to the above, for several different settings.
1
Introduction
Information constraints play a key role in machine learning. Of course, the main constraint is the
availability of only a finite data set to learn from. However, many current problems in machine
learning can be characterized as learning with additional information constraints, arising from the
manner in which the learner may interact with the data. Some examples include:
? Communication constraints in distributed learning: There has been much recent work on learning
when the training data is distributed among several machines. Since the machines may work
in parallel, this potentially allows significant computational speed-ups and the ability to cope
with large datasets. On the flip side, communication rates between machines is typically much
slower than their processing speeds, and a major challenge is to perform these learning tasks with
minimal communication.
? Memory constraints: The standard implementation of many common learning tasks requires
memory which is super-linear in the data dimension. For example, principal component analysis
(PCA) requires us to estimate eigenvectors of the data covariance matrix, whose size is quadratic
in the data dimension and can be prohibitive for high-dimensional data. Another example is kernel
learning, which requires manipulation of the Gram matrix, whose size is quadratic in the number
of data points. There has been considerable effort in developing and analyzing algorithms for
such problems with reduced memory footprint (e.g. [20, 7, 27, 24]).
? Online learning constraints: The need for fast and scalable learning algorithms has popularised
the use of online algorithms, which work by sequentially going over the training data, and incrementally updating a (usually small) state vector. Well-known special cases include gradient
descent and mirror descent algorithms. The requirement of sequentially passing over the data
can be seen as a type of information constraint, whereas the small state these algorithms often
maintain can be seen as another type of memory constraint.
1
? Partial-information constraints: A common situation in machine learning is when the available
data is corrupted, sanitized (e.g. due to privacy constraints), has missing features, or is otherwise
partially accessible. There has also been considerable interest in online learning with partial
information, where the learner only gets partial feedback on his performance. This has been
used to model various problems in web advertising, routing and multiclass learning. Perhaps
the most well-known case is the multi-armed bandits problem with many other variants being
developed, such as contextual bandits, combinatorial bandits, and more general models such as
partial monitoring [10, 11].
Although these examples come from very different domains, they all share the common feature
of information constraints on how the learning algorithm can interact with the training data. In
some specific cases (most notably, multi-armed bandits, and also in the context of certain distributed
protocols, e.g. [6, 29]) we can even formalize the price we pay for these constraints, in terms of
degraded sample complexity or regret guarantees. However, we currently lack a general informationtheoretic framework, which directly quantifies how such constraints can impact performance. For
example, are there cases where any online algorithm, which goes over the data one-by-one, must
have a worse sample complexity than (say) empirical risk minimization? Are there situations where
a small memory footprint provably degrades the learning performance? Can one quantify how a
constraint of getting only a few bits from each example affects our ability to learn?
In this paper, we make a first step in developing such a framework. We consider a general class of
learning processes, characterized only by information-theoretic constraints on how they may interact
with the data (and independent of any specific problem semantics). As special cases, these include
online algorithms with memory constraints, certain types of distributed algorithms, as well as online
learning with partial information. We identify cases where any such algorithm must perform worse
than what can be attained without such information constraints. The tools developed allows us to
establish several results for specific learning problems:
? We prove a new and generic
p regret lower bound for partial-information online learning with expert
advice, of the form ?( (d/b)T ), where T is the number of rounds, d is the dimension of the
loss/reward vector, and b is the number of bits b extracted from each loss vector. It is optimal
up to log-factors (without further assumptions), and holds no matter what these b bits are ? a
single coordinate (as in multi-armed bandits), some information on several coordinates (as in
semi-bandit feedback), a linear projection (as in bandit linear optimization), some feedback signal
from a restricted set (as in partial monitoring) etc. Interestingly, it holds even if the online learner
is allowed to adaptively choose which bits of the loss vector it can retain at each round. The lower
bound quantifies directly how information constraints in online learning degrade the attainable
regret, independent of the problem semantics.
? We prove that for some learning and estimation problems - in particular, sparse PCA and sparse
covariance estimation in Rd - no online algorithm can attain statistically optimal performance (in
? 2 ) memory. To the best of our knowledge, this is
terms of sample complexity) with less than ?(d
the first formal example of a memory/sample complexity trade-off in a statistical learning setting.
? We show that for similar types of problems, there are cases where no distributed algorithm (which
is based on a non-interactive or serial protocol on i.i.d. data) can attain optimal performance with
? 2 ) communication per machine. To the best of our knowledge, this is the first formal
less than ?(d
example of a communication/sample complexity trade-off, in the regime where the communication
budget is larger than the data dimension, and the examples at each machine come from the same
underlying distribution.
? We demonstrate the existence of (synthetic) stochastic optimization problems where any algorithm which uses memory linear in the dimension (e.g. stochastic gradient descent or mirror
descent) cannot be statistically optimal.
Related Work
In stochastic optimization, there has been much work on lower bounds for sequential algorithms
(e.g. [22, 1, 23]). However, these results all hold in an oracle model, where data is assumed to be
made available in a specific form (such as a stochastic gradient estimate). As already pointed out in
2
[22], this does not directly translate to the more common setting, where we are given a dataset and
wish to run a simple sequential optimization procedure.
In the context of distributed learning and statistical estimation, information-theoretic lower bounds
were recently shown in the pioneering work [29], which identifies cases where communication constraints affect statistical performance. These results differ from ours (in the context of distributed
learning) in two important ways. First, they pertain to parametric estimation in Rd , where the communication budget per machine is much smaller than what is needed to even specify the answer
with constant accuracy (O(d) bits). In contrast, our results pertain to simpler detection problems,
where the answer requires only O(log(d)) bits, yet lead to non-trivial lower bounds even when the
budget size is much larger (in some cases, much larger than d). The second difference is that their
work focuses on distributed algorithms, while we address a more general class of algorithms, which
includes other information-constrained settings. Strong lower bounds in the context of distributed
learning have also been shown in [6], but they do not apply to a regime where examples across machines come from the same distribution, and where the communication budget is much larger than
what is needed to specify the output.
There are well-known lower bounds for multi-armed bandit problems and other online learning with
partial-information settings. However, they crucially depend on the semantics of the information
feedback considered. For example, the standard multi-armed bandit lower bound [5] pertain to a
setting where we can view a single coordinate of the loss vector, but doesn?t apply as-is when we
can view more than one coordinate (e.g. [4, 25]), get side-information (e.g. [19]), receive a linear
or non-linear projection (as in bandit linear and convex optimization), or receive a different type of
partial feedback (e.g. partial monitoring [11]). In contrast, our results are generic and can directly
apply to any such setting.
Memory and communication constraints have been extensively studied within theoretical computer
science (e.g. [3, 21]). Unfortunately, almost all these results pertain to data which was either adversarially generated, ordered (in streaming algorithms) or split (in distributed algorithms), and do
not apply to statistical learning tasks, where the data is drawn i.i.d. from an underlying distribution.
[28, 15] do consider i.i.d. data, but focus on problems such as detecting graph connectivity and
counting distinct elements, and not learning problems such as those considered here. Also, there are
works on provably memory-efficient algorithms for statistical problems (e.g. [20, 7, 17, 13]), but
these do not consider lower bounds or provable trade-offs.
Finally, there has been a line of works on hypothesis testing and statistical estimation with finite
memory (see [18] and references therein). However, the limitations shown in these works apply
when the required precision exceeds the amount of memory available. Due to finite sample effects,
this regime is usually relevant only when the data size is exponential in the memory size. In contrast,
we do not rely on finite precision considerations.
2
Information-Constrained Protocols
We begin with a few words about notation. We use bold-face letters (e.g. x) to denote vectors, and
let ej ? Rd denote j-th standard basis vector. When convenient, we use the standard asymptotic
?
notation O(?), ?(?), ?(?) to hide constants, and an additional ? sign (e.g. O(?))
to also hide logfactors. log(?) refers to the natural logarithm, and log2 (?) to the base-2 logarithm.
Our main object of study is the following generic class of information-constrained algorithms:
Definition 1 ((b, n, m) Protocol). Given access to a sequence of mn i.i.d. instances (vectors in Rd ),
an algorithm is a (b, n, m) protocol if it has the following form, for some functions ft returning an
output of at most b bits, and some function f :
? For t = 1, . . . , m
? Let X t be a batch of n i.i.d. instances
? Compute message W t = ft (X t , W 1 , W 2 , . . . W t?1 )
? Return W = f (W 1 , . . . , W m )
3
Note that the functions {ft }m
t=1 , f are completely arbitrary, may depend on m and can also be
randomized. The crucial assumption is that the outputs W t are constrained to be only b bits.
As the definition above may appear quite abstract, let us consider a few specific examples:
? b-memory online protocols: Consider any algorithm which goes over examples one-by-one, and
incrementally updates a state vector W t of bounded size b. We note that a majority of online
learning and stochastic optimization algorithms have bounded memory. For example, for linear
predictors, most gradient-based algorithms maintain a state whose size is proportional to the size
of the parameter vector that is being optimized. Such algorithms correspond to (b, n, m) protocols
where W t is the state vector after round t, with an update function ft depending only on W t?1 ,
and f depends only on W m . n = 1 corresponds to algorithms which use one example at a time,
whereas n > 1 corresponds to algorithms using mini-batches.
? Non-interactive and serial distributed algorithms: There are m machines and each machine receives an independent sample X t of size n. It then sends a message W t = ft (X t ) (which here
depends only on X t ). A centralized server then combines the messages to compute an output
f (W 1 . . . W m ). This includes for instance divide-and-conquer style algorithms proposed for distributed stochastic optimization (e.g. [30]). A serial variant of the above is when there are m
machines, and one-by-one, each machine t broadcasts some information W t to the other machines, which depends on X t as well as previous messages sent by machines 1, 2, . . . , (t ? 1).
? Online learning with partial information: Suppose we sequentially receive d-dimensional loss
vectors, and from each of these we can extract and use only b bits of information, where b d.
This includes most types of bandit problems [10].
In our work, we contrast the performance attainable by any algorithm corresponding to such a protocol, to constraint-free protocols which are allowed to interact with the data in any manner.
3
Basic Results
Our results are based on a simple ?hide-and-seek? statistical estimation problem, for which we show
a strong gap between the performance of information-constrained protocols and constraint-free protocols. It is parameterized by a dimension d, bias ?, and sample size mn, and defined as follows:
Definition 2 (Hide-and-seek Problem). Consider the set of product distributions {Prj (?)}dj=1 over
{?1, 1}d defined via Ex?Prj (?) [xi ] = 2? 1i=j for all coordinates i = 1, . . . d. Given an i.i.d. sample
of mn instances generated from Prj (?), where j is unknown, detect j.
In words, Prj (?) corresponds to picking all coordinates other than j to be ?1 uniformly
at random,
and independently picking coordinate j to be +1 with a higher probability 21 + ? . The goal is to
detect the biased coordinate j based on a sample.
First, we note that without information constraints, it is easy to detect the biased coordinate with
O(log(d)/?2 ) instances. This is formalized in the following theorem, which is an immediate consequence of Hoeffding?s inequality and a union bound:
Theorem 1. Consider the hide-and-seek problem defined earlier. Given mn samples, if J? is the
coordinate with the highest empirical average, then Prj (J? = j) ? 1 ? 2d exp ? 21 mn?2 .
We now show that for this hide-and-seek problem, there is a large regime where detecting j is
information-theoretically possible (by Thm. 1), but any information-constrained protocol will fail to
do so with high probability. We first show this for (b, 1, m) protocols (i.e. protocols which process
one instance at a time, such as bounded-memory online algorithms, and distributed algorithms where
each machine holds a single instance):
Theorem 2. Consider the hide-and-seek problem on d > 1 coordinates, with some bias ? ? 1/4
and sample size m. Then for any estimate J? of the biased coordinate returned by any (b, 1, m)
protocol, there exists some coordinate j such that
r
3
?2 b
?
Prj (J = j) ? + 21 m
.
d
d
4
The theorem implies that any algorithm corresponding to (b, 1, m) protocols requires sample size
m ? ?((d/b)/?2 ) to reliably detect some j. When b is polynomially smaller than d (e.g. a constant),
we get an exponential gap compared to constraint-free protocols, which only require O(log(d)/?2 )
instances.
Moreover, Thm. 2 is tight up to log-factors: Consider a b-memory online algorithm, which splits
the d coordinates into O(d/b) segments of O(b) coordinates each, and sequentially goes over the
2
?
segments, each time using O(1/?
) independent instances to determine if one of the coordinates in
each segment is biased by ? (assuming ? is not exponentially smaller than b, this can be done with
O(b) memory by maintaining the empirical average of each coordinate). This will allow to detect
2
?
the biased coordinate, using O((d/b)/?
) instances.
We now turn to provide an analogous result for general (b, n, m) protocols (where n is possibly
greater than 1). However, it is a bit weaker in terms of the dependence on the bias parameter1 :
Theorem 3. Consider the hide-and-seek problem on d > 1 coordinates, with some bias ? ? 1/4n
and sample size mn. Then for any estimate J? of the biased coordinate returned by any (b, n, m)
protocol, there exists some coordinate j such that
s
3
10?b 2
?
Prj (J = j) ? + 5 mn min
,? .
d
d
The
theorem
n implies
othat any (b, n, m) protocol will require a sample size mn which is at least
(d/b) 1
? max
in order to detect the biased coordinate. This is larger than the O(log(d)/?2 )
? , ?2
instances required by constraint-free protocols whenever ? > b log(d)/d, and establishes trade-offs
between sample complexity and information complexities such as memory and communication.
Due to lack of space, all our proofs appear in the supplementary material. However, the technical
details may obfuscate the high-level intuition, which we now turn to explain.
From an information-theoretic viewpoint, our results are based on analyzing the mutual information
between j and W t in a graphical model as illustrated in figure 1. In this model, the unknown message
j (i.e. the identity of the biased coordinate) is correlated with one of d independent binary-valued
random vectors (one for each coordinate across the data instances X t ). All these random vectors
are noisy, and the mutual information in bits between Xjt and j can be shown to be on the order of
n?2 . Without information constraints, it follows that given m instantiations of X t , the total amount
of information conveyed on j by the data is ?(mn?2 ), and if this quantity is larger than log(d), then
there is enough information to uniquely identify j. Note that no stronger bound can be established
with standard statistical lower-bound techniques, since these do not consider information constraints
internal to the algorithm used.
Indeed, in our information-constrained setting there is an added complication, since the output W t
can only contain b bits. If b d, then W t cannot convey all the information on X1t , . . . , Xdt .
Moreover, it will likely convey only little information if it doesn?t already ?know? j. For example,
W t may provide a little bit of information on all d coordinates, but then the amount of information
conveyed on each (and in particular, the random variable Xjt which is correlated with j) will be
very small. Alternatively, W t may provide accurate information on O(b) coordinates, but since
the relevant coordinate j is not known, it is likely to ?miss? it. The proof therefore relies on the
following components:
? No matter what, a (b, n, m) protocol cannot provide more than b/d bits of information (in expectation) on Xjt , unless it already ?knows? j.
? Even if the mutual information between W t and Xjt is only b/d, and the mutual information between Xjt and j is n?2 , standard information-theoretic tools such as the data processing inequality
only implies that the mutual information between W t and j is bounded by min{n?2 , b/d}. We
essentially prove a stronger information contraction bound, which is the product of the two terms
1
The proof of Thm. 2 also applies in the case n > 1, but the dependence on n is exponential - see the proof
for details.
5
?1?
Figure 1:
Illustration of the relationship between j, the coordinates
1, 2, . . . , j, . . . , d of the sample X t , and
the message W t . The coordinates are independent of each other, and most of them
just output ?1 uniformly at random. Only
Xjt has a slightly different distribution and
hence contains some information on j.
?2?
?
?
??
???
?
???
O(?2 b/d) when n = 1, and O(n?b/d) for general n. At a technical level, this is achieved by
considering the relative entropy between the distributions of W t with and without a biased coordinate j, relating it to the ?2 -divergence between these distributions (using relatively recent
analytic results on Csisz?ar f-divergences [16], [26]), and performing algebraic manipulations to
upper bound it by ?2 times the mutual information between W t and Xjt , which is on average b/d
as discussed earlier. This eventually leads to the m?2 b/d term in Thm. 2, as well as Thm. 3 using
somewhat different calculations.
4
4.1
Applications
Online Learning with Partial Information
Consider the setting of learning with expert advice, defined as a game over T rounds, where each
round t a loss vector `t ? [0, 1]d is chosen, and the learner (without knowing `t ) needs to pick an
action it from a fixed set {1, . . . , d}, after which the learner suffers loss `t,it . The goal of the learner
PT
PT
is to minimize the regret with respect to any fixed action i, t=1 `t,it ? t=1 `t,i . We are interested
in variants where the learner only gets some partial information on `t . For example, in multi-armed
bandits, the learner can only view `t,it . The following theorem is a simple corollary of Thm. 2:
Theorem 4. Suppose d > 3. For any
there
over loss
hP(b, 1, T ) protocol,
i is an i.i.d.
n distribution
o
p
PT
T
d
vectors `t ? [0, 1] for which minj E
(d/b)/T , where
t=1 `t,jt ?
t=1 `t,j ? c min T,
c > 0 is a numerical constant.
As a result, we get that for any algorithm with any partial information feedback model (where b
bitspare extracted from each d-dimensional loss vector), it is impossible to get regret lower than
assumptions on the feedback model, the
?( (d/b)T ) for sufficiently large T . Without further
p
bound is optimal up to log-factors, as shown by O( (d/b)T ) upper bounds for linear or coordinate
measurements (where b is the number of measurements or coordinates seen2 ) [2, 19, 25]. However,
the lower bound extends beyond these specific settings, and include cases such as arbitrary nonlinear measurements of the loss vector, or receiving feedback signals of bounded size (although
some setting-specific lower bounds may be stronger). It also simplifies previous lower bounds,
tailored to specific types of partial information feedback, or relying on careful reductions to multiarmed bandits (e.g. [12, 25]). Interestingly, the bound holds even if the algorithm is allowed to
examine each loss vector `t and adaptively choose which b bits of information it wishes to retain.
4.2
Stochastic Optimization
We now turn to consider an example from stochastic optimization, where our goal is to approximately minimize F (h) = EZ [f (h; Z)] given access to m i.i.d. instantiations of Z, whose distribution is unknown. This setting has received much attention in recent years, and can be used to
model many statistical learning problems. In this section, we demonstrate a stochastic optimization
problem where information-constrained protocols provably pay a performance price compared to
non-constrained algorithms. We emphasize that it is a simple toy problem, and not meant to represent anything realistic. We present it for two reasons: First, it illustrates another type of situation
2
Strictly speaking, if the losses are continuous-valued, these require arbitrary-precision measurements, but
in any practical implementation we can assume the losses and measurements are discrete.
6
where information-constrained protocols may fail (in particular, problems involving matrices). Second, the intuition of the construction is also used in the more realistic problem of sparse PCA and
covariance estimation, considered in the next section.
Specifically, suppose we wish to solve min(w,v) F (w, v) = EZ [f ((w, v); Z)], where
f ((w, v); Z) = w> Zv , Z ? [?1, +1]d?d
Pd
Pd
and w, v range over all vectors in the simplex (i.e. wi , vi ? 0 and i=1 wi =
i=1 vi = 1).
A minimizer of F (w, v) is (ei? , ej ? ), where (i? , j ? ) are indices of the matrix entry with minimal mean. Moreover, by a standard concentration of measure argument, given m i.i.d. instan? J)
? =
tiations Z 1 , . P
. . , Z m from any distribution over Z, then the solution (eI?, eJ?), where (I,
m
1
t
arg mini,j m t=1 Zi,j are the indices of the entry with empirically smallest mean, satisfies
p
F (eI?, eJ?) ? minw,v F (w, v) + O
log(d)/m with high probability.
? J)
? as above requires us to track d2 empirical means, which may be exHowever, computing (I,
pensive when d is large. If instead we constrain ourselves to (b, 1, m) protocols where b = O(d)
(e.g. any sort of stochastic gradient method optimization algorithm, whose memory
pis linear in the
number of parameters), then we claim that we have
a
lower
bound
of
?(min{1,
d/m}) on the
p
expected error, which is much higher than the O( log(d)/m) upper bound for constraint-free protocols. This claim is a straightforward consequence of Thm. 2: We consider distributions where
Z ? {?1, +1}d?d with probability 1, each of the d2 entries
p is chosen independently, and E[Z] is
zero except some coordinate (i? p
, j ? ) where it equals O( d/m). For such distributions, getting optimization error smaller than O( d/m) reduces to detecting (i? , j ? ), and this inp
turn reduces to the
hide-and-seek problem defined earlier, over d2 coordinates and a bias ? = O( d/m). However,
Thm. 2 shows that no (b, 1, m) protocol (where b = O(d)) will succeed if md?2 d2 , which
indeed happens if ? is small enough.
Similar kind of gaps can be shown using Thm. 3 for general (b, n, m) protocols, which apply to any
special case such as non-interactive distributed learning.
4.3
Sparse PCA, Sparse Covariance Estimation, and Detecting Correlations
The sparse PCA problem ([31]) is a standard and well-known statistical estimation problem, defined
as follows: We are given an i.i.d. sample of vectors x ? Rd , and we assume that there is some
direction, corresponding to some sparse vector v (of cardinality at most k), such that the variance
E[(v> x)2 ] along that direction is larger than at any other direction. Our goal is to find that direction.
We will focus here on the simplest possible form of this problem, where the maximizing direction v
is assumed to be 2-sparse, i.e. there are only 2 non-zero coordinates vi , vj . In that case, E[(v> x)2 ] =
v12 E[x21 ] + v22 E[x22 ] + 2v1 v2 E[xi xj ]. Following previous work (e.g. [8]), we even assume that
E[x2i ] = 1 for all i, in which case the sparse PCA problem reduces to detecting a coordinate pair
(i? , j ? ), i? < j ? for which xi? , xj ? are maximally correlated. A special case is a simple and natural
sparse covariance estimation problem [9], where we assume that all covariates are uncorrelated
(E[xi xj ] = 0) except for a unique correlated pair (i? , j ? ) which we need to detect.
This setting bears a resemblance to the example seen in the context of stochastic optimization in section 4.2: We have a d ? d stochastic matrix xx> , and we need to detect an off-diagonal biased entry
at location (i? , j ? ). Unfortunately, these stochastic matrices are rank-1, and do not have independent
entries as in the example considered in section 4.2. Instead, we use a more delicate construction,
relying on distributions supported on sparse vectors. The intuition is that then each instantiation of
xx> is sparse, and the situation can be reduced to a variant of our hide-and-seek problem where only
a few coordinates are non-zero at a time. The theorem below establishes performance gaps between
constraint-free protocols (in particular, a simple plug-in estimator), and any (b, n, m) protocol for a
specific choice of n, or any b-memory online protocol (See Sec. 2).
Theorem 5. Consider the class of 2-sparse PCA (or covariance estimation) problems in d ? 9
dimensions as described above, and all distributions such that E[x2i ] = 1 for all i, and:
1. For a unique pair of distinct coordinates (i? , j ? ), it holds that E[xi? xj ? ] = ? > 0, whereas
E[xi xj ] = 0 for all distinct coordinate pairs (i, j) 6= (i? , j ? ).
7
2. For any i < j, if xg
i xj is the
empirical average
of xi xj over m i.i.d. instances, then
?
2
Pr |xg
i xj ? E[xi xj ]| ? 2 ? 2 exp ?m? /6 .
Then the following holds:
? ?
? J)
? = arg maxi<j xg
? ?
? Let (I,
i xj . Then for any distribution as above, Pr((I, J) = (i , j )) ?
2
2
1 ? d exp(?m? /6). In particular, when the bias ? equals ?(1/d log(d)),
m
? ?
2
?
?
Pr((I, J) = (i , j )) ? 1 ? d exp ??
.
d2 log2 (d)
? J)
? of (i? , j ? ) returned by any b-memory online protocol using m instances,
? For any estimate (I,
m
c) protocol, there exists a distribution with bias ? = ?(1/d log(d))
or any (b, d(d ? 1), b d(d?1)
as above such that
r
m
? J)
? = (i? , j ? ) ? O 1 +
Pr (I,
.
d2
d4 /b
The theorem implies that in the regime where b d2 / log2 (d), we can choose any m such that
2
d4
2
? ?
b m d log (d), and get that the chances of the protocol detecting (i , j ) are arbitrarily
? ?
small, even though the empirical average reveals (i , j ) with arbitrarily high probability. Thus, in
this sparse PCA / covariance estimation setting, any online algorithm with sub-quadratic memory
cannot be statistically optimal for all sample sizes. The same holds for any (b, n, m) protocol in an
appropriate regime of (n, m), such as distributed algorithms as discussed earlier.
To the best of our knowledge, this is the first result which explicitly shows that memory constraints
can incur a statistical cost for a standard estimation problem. It is interesting that sparse PCA was
also shown recently to be affected by computational constraints on the algorithm?s runtime ([8]).
The proof appears in the supplementary material. Besides using a somewhat different hide-andseek construction as mentioned earlier, it also relies on the simple but powerful observation that
any b-memory online protocol is also a (b, ?, bm/?c) protocol for arbitrary ?. Therefore, we only
need to prove the theorem for (b, ?, bm/?c) for some ? (chosen to equal d(d ? 1) in our case) to
automatically get the same result for b-memory protocols.
5
Discussion and Open Questions
In this paper, we investigated cases where a generic type of information-constrained algorithm has
strictly inferior statistical performance compared to constraint-free algorithms. As special cases,
we demonstrated such gaps for memory-constrained and communication-constrained algorithms
(e.g. in the context of sparse PCA and covariance estimation), as well as online learning with
partial information and stochastic optimization. These results are based on explicitly considering
the information-theoretic structure of the problem, and depend only on the number of bits extracted
from each data batch.
Several questions remain open. One question is whether Thm. 3 can be improved. We conjecture
this is true, and that the bound should actually depend on mn?2 b/d rather than mn min{?b/d, ?2 }.
This would allow, for instance, to show the same type of performance gaps for (b, 1, m) protocols
and (b, n, m) protocols. A second open question is whether there are convex stochastic optimization
problems, for which online or distributed algorithms are provably inferior to constraint-free algorithms (the example discussed in section 4.2 refers to an easily-solvable yet non-convex problem). A
third open question is whether our results for distributed algorithms can be extended to more interactive protocols, where the different machines can communicate over several rounds. There is a rich
literature on the subject within theoretical computer science, but it is not clear how to ?import? these
results to a statistical setting based on i.i.d. data. A fourth open question is whether the performance
gap that we demonstrated for sparse-PCA / covariance estimation can be extended to a ?natural?
distribution (e.g. Gaussian), as our result uses a tailored distribution, which has a sufficiently controlled tail behavior but is ?spiky? and not sub-Gaussian uniformly in the dimension. More generally,
it would be interesting to extend the results to other learning problems and information constraints.
Acknowledgements: This research is supported by the Intel ICRI-CI Institute, Israel Science Foundation grant 425/13, and an FP7 Marie Curie CIG grant. We thank John Duchi, Yevgeny Seldin and
Yuchen Zhang for helpful comments.
8
References
[1] A. Agarwal, P. Bartlett, P. Ravikumar, and M. Wainwright. Information-theoretic lower bounds on
the oracle complexity of stochastic convex optimization. Information Theory, IEEE Transactions on,
58(5):3235?3249, 2012.
[2] A. Agarwal, O. Dekel, and L. Xiao. Optimal algorithms for online convex optimization with multi-point
bandit feedback. In COLT, 2010.
[3] Noga Alon, Yossi Matias, and Mario Szegedy. The space complexity of approximating the frequency
moments. In STOC, 1996.
[4] J.-Y. Audibert, S. Bubeck, and G. Lugosi. Minimax policies for combinatorial prediction games. In
COLT, 2011.
[5] P. Auer, N. Cesa-Bianchi, Y. Freund, and R. Schapire. The nonstochastic multiarmed bandit problem.
SIAM Journal on Computing, 32(1):48?77, 2002.
[6] M. Balcan, A. Blum, S. Fine, and Y. Mansour. Distributed learning, communication complexity and
privacy. In COLT, 2012.
[7] A. Balsubramani, S. Dasgupta, and Y. Freund. The fast convergence of incremental pca. In NIPS, 2013.
[8] A. Berthet and P. Rigollet. Complexity theoretic lower bounds for sparse principal component detection.
In COLT, 2013.
[9] J. Bien and R. Tibshirani. Sparse estimation of a covariance matrix. Biometrika, 98(4):807?820, 2011.
[10] S. Bubeck and N. Cesa-Bianchi. Regret analysis of stochastic and nonstochastic multi-armed bandit
problems. Foundations and Trends in Machine Learning, 5(1):1?122, 2012.
[11] N. Cesa-Bianchi and L. Gabor. Prediction, learning, and games. Cambridge University Press, 2006.
[12] N. Cesa-Bianchi, S. Shalev-Shwartz, and O. Shamir. Efficient learning with partially observed attributes.
The Journal of Machine Learning Research, 12:2857?2878, 2011.
[13] S. Chien, K. Ligett, and A. McGregor. Space-efficient estimation of robust statistics and distribution
testing. In ICS, 2010.
[14] T. Cover and J. Thomas. Elements of information theory. John Wiley & Sons, 2006.
[15] M. Crouch, A. McGregor, and D. Woodruff. Stochastic streams: Sample complexity vs. space complexity.
In MASSIVE, 2013.
[16] S. S. Dragomir. Upper and lower bounds for Csisz?ar?s f-divergence in terms of the Kullback-Leibler distance and applications. In Inequalities for Csisz?ar f-Divergence in Information Theory. RGMIA Monographs, 2000.
[17] S. Guha and A. McGregor. Space-efficient sampling. In AISTATS, 2007.
[18] L. Kontorovich. Statistical estimation with bounded memory. Statistics and Computing, 22(5):1155?
1164, 2012.
[19] S. Mannor and O. Shamir. From bandits to experts: On the value of side-observations. In NIPS, 2011.
[20] I. Mitliagkas, C. Caramanis, and P. Jain. Memory limited, streaming pca. In NIPS, 2013.
[21] S. Muthukrishnan. Data streams: Algorithms and applications. Now Publishers Inc, 2005.
[22] A. Nemirovsky and D. Yudin. Problem Complexity and Method Efficiency in Optimization. WileyInterscience, 1983.
[23] M. Raginsky and A. Rakhlin. Information-based complexity, feedback and dynamics in convex programming. Information Theory, IEEE Transactions on, 57(10):7036?7056, 2011.
[24] A. Rahimi and B. Recht. Random features for large-scale kernel machines. In NIPS, 2007.
[25] Y. Seldin, P. Bartlett, K. Crammer, and Y. Abbasi-Yadkori. Prediction with limited advice and multiarmed
bandits with paid observations. In ICML, 2014.
[26] I. Taneja and P. Kumar. Relative information of type s, Csisz?ar?s f-divergence, and information inequalities. Inf. Sci., 166(1-4):105?125, 2004.
[27] C. Williams and M. Seeger. Using the nystr?om method to speed up kernel machines. In NIPS, 2001.
[28] D. Woodruff. The average-case complexity of counting distinct elements. In ICDT, 2009.
[29] Y. Zhang, J. Duchi, M. Jordan, and M. Wainwright. Information-theoretic lower bounds for distributed
statistical estimation with communication constraints. In NIPS, 2013.
[30] Y. Zhang, J. Duchi, and M. Wainwright. Communication-efficient algorithms for statistical optimization.
In NIPS, 2012.
[31] H. Zou, T. Hastie, and R. Tibshirani. Sparse principal component analysis. Journal of computational and
graphical statistics, 15(2):265?286, 2006.
9
| 5386 |@word stronger:3 dekel:1 open:5 d2:7 seek:8 crucially:1 nemirovsky:1 covariance:10 contraction:1 attainable:2 pick:1 paid:1 nystr:1 moment:1 reduction:1 contains:1 woodruff:2 ours:1 interestingly:2 current:1 contextual:1 parameter1:1 yet:2 must:2 import:1 john:2 numerical:1 realistic:2 analytic:1 ligett:1 update:2 v:1 prohibitive:1 detecting:6 mannor:1 complication:1 location:1 simpler:1 zhang:3 along:1 popularised:1 prove:4 combine:1 privacy:2 manner:2 theoretically:1 notably:1 expected:1 indeed:2 behavior:1 examine:1 multi:9 relying:2 automatically:1 little:3 armed:8 considering:2 cardinality:1 begin:1 xx:2 underlying:3 bounded:7 notation:2 moreover:3 what:6 israel:1 kind:1 developed:2 guarantee:1 interactive:4 runtime:1 returning:1 biometrika:1 grant:2 appear:2 positive:1 limit:1 consequence:2 analyzing:2 approximately:1 lugosi:1 therein:1 studied:1 limited:2 range:1 statistically:3 weizmann:2 practical:1 unique:2 testing:2 union:1 regret:6 footprint:3 procedure:1 empirical:6 attain:2 gabor:1 projection:2 ups:1 word:2 convenient:1 refers:2 inp:1 get:8 cannot:4 pertain:4 context:6 risk:1 impossible:1 demonstrated:2 missing:2 maximizing:1 go:3 attention:1 straightforward:1 independently:2 convex:6 williams:1 formalized:1 estimator:1 his:1 coordinate:39 analogous:1 shamir:4 play:1 suppose:3 pt:3 construction:3 massive:1 programming:1 us:2 hypothesis:1 element:3 trend:1 updating:1 pensive:1 observed:1 role:1 ft:5 trade:4 highest:1 mentioned:1 intuition:3 pd:2 monograph:1 complexity:16 covariates:1 reward:1 dynamic:1 depend:4 tight:1 segment:3 incur:1 efficiency:1 learner:8 basis:1 completely:1 easily:1 v22:1 various:1 caramanis:1 muthukrishnan:1 distinct:4 fast:3 describe:1 jain:1 shalev:1 whose:5 quite:1 larger:7 solve:2 supplementary:2 say:1 valued:2 otherwise:1 ability:2 statistic:3 noisy:1 online:25 sequence:1 cig:1 product:2 relevant:2 translate:1 icdt:1 csisz:4 getting:2 x1t:1 convergence:1 requirement:1 prj:7 incremental:1 object:1 depending:1 alon:1 ac:1 received:1 strong:2 implies:5 come:3 quantify:1 differ:1 direction:5 attribute:1 stochastic:19 routing:1 material:2 require:3 strictly:2 hold:8 sufficiently:2 considered:4 ic:1 exp:4 claim:2 major:1 smallest:1 estimation:20 combinatorial:2 currently:2 establishes:2 tool:2 minimization:1 offs:2 gaussian:2 super:1 rather:1 ej:4 corollary:1 focus:3 rank:1 contrast:4 seeger:1 detect:8 helpful:1 streaming:2 typically:1 bandit:19 going:1 interested:1 semantics:4 provably:4 arg:2 among:1 colt:4 constrained:13 special:5 crouch:1 mutual:6 equal:3 sampling:1 adversarially:1 icml:1 simplex:1 fundamentally:1 few:4 divergence:5 ourselves:1 maintain:2 delicate:1 detection:2 interest:1 message:6 centralized:1 x22:1 accurate:1 partial:17 ohad:2 minw:1 unless:1 divide:1 logarithm:2 yuchen:1 theoretical:2 minimal:2 instance:15 earlier:5 ar:4 cover:1 cost:1 entry:5 predictor:1 guha:1 answer:3 corrupted:1 synthetic:1 adaptively:2 recht:1 fundamental:1 randomized:1 siam:1 accessible:1 retain:2 off:3 receiving:1 picking:2 kontorovich:1 connectivity:1 abbasi:1 cesa:4 choose:3 broadcast:1 hoeffding:1 possibly:1 worse:3 expert:3 style:1 return:1 toy:1 szegedy:1 bold:1 sec:1 availability:1 includes:3 matter:2 inc:1 explicitly:2 audibert:1 depends:3 vi:3 stream:2 view:3 mario:1 sort:1 parallel:1 curie:1 minimize:2 il:1 om:1 degraded:1 accuracy:1 variance:1 correspond:1 identify:2 advertising:1 monitoring:3 explain:1 minj:1 suffers:1 whenever:1 definition:3 matias:1 frequency:1 proof:5 dataset:1 knowledge:3 formalize:1 actually:1 auer:1 appears:1 attained:1 higher:2 specify:2 maximally:1 improved:1 done:1 though:1 just:1 spiky:1 correlation:1 receives:1 web:1 ei:3 nonlinear:1 lack:2 incrementally:2 perhaps:1 resemblance:1 icri:1 effect:1 contain:1 true:1 hence:1 leibler:1 illustrated:1 round:6 game:3 uniquely:1 inferior:2 anything:1 d4:2 theoretic:8 demonstrate:2 duchi:3 dragomir:1 balcan:1 consideration:1 recently:2 common:4 rigollet:1 empirically:1 exponentially:1 discussed:3 tail:1 extend:1 relating:1 significant:1 measurement:5 multiarmed:3 cambridge:1 rd:5 hp:1 pointed:1 dj:1 access:4 etc:1 base:1 recent:3 hide:11 inf:1 manipulation:2 certain:3 server:1 inequality:4 binary:1 arbitrarily:2 seen:3 additional:2 greater:1 somewhat:2 determine:1 signal:2 semi:1 reduces:3 rahimi:1 exceeds:1 technical:2 characterized:3 calculation:1 plug:1 serial:3 ravikumar:1 controlled:1 impact:1 prediction:3 scalable:1 variant:4 basic:1 xjt:7 essentially:1 expectation:1 involving:1 kernel:3 tailored:2 represent:1 agarwal:2 achieved:1 receive:3 whereas:3 fine:1 sends:1 crucial:1 publisher:1 biased:10 noga:1 comment:1 subject:1 sent:1 jordan:1 counting:2 split:2 easy:1 enough:2 affect:3 xj:10 zi:1 nonstochastic:2 hastie:1 simplifies:1 knowing:1 multiclass:1 whether:4 pca:13 bartlett:2 effort:1 returned:3 algebraic:1 passing:1 speaking:1 action:2 generally:1 clear:1 eigenvectors:1 amount:3 extensively:1 simplest:1 reduced:2 schapire:1 sign:1 arising:1 per:2 track:1 tibshirani:2 discrete:1 dasgupta:1 affected:1 zv:1 key:1 blum:1 drawn:1 marie:1 v1:1 graph:1 year:1 raginsky:1 run:1 letter:1 parameterized:1 powerful:1 communicate:1 fourth:1 extends:1 almost:1 v12:1 bit:17 bound:27 pay:2 quadratic:3 oracle:2 constraint:41 constrain:1 speed:3 argument:1 min:6 kumar:1 performing:1 relatively:1 conjecture:1 developing:2 smaller:4 across:2 slightly:1 remain:1 son:1 wi:2 happens:1 restricted:1 pr:4 turn:4 eventually:1 fail:2 needed:2 know:2 flip:1 yossi:1 fp7:1 available:3 apply:6 balsubramani:1 v2:1 generic:4 appropriate:1 batch:3 yadkori:1 slower:1 existence:1 thomas:1 include:5 x21:1 graphical:2 log2:3 maintaining:1 xdt:1 establish:1 conquer:1 approximating:1 already:3 quantity:1 added:1 question:6 degrades:1 parametric:1 dependence:2 concentration:1 md:1 diagonal:1 gradient:5 distance:1 thank:1 sci:1 majority:1 degrade:1 trivial:1 reason:1 provable:1 assuming:1 besides:1 index:2 relationship:1 mini:2 illustration:1 unfortunately:2 potentially:1 stoc:1 implementation:2 reliably:1 policy:1 unknown:3 perform:3 bianchi:4 upper:4 observation:3 datasets:1 finite:4 descent:4 immediate:1 situation:4 extended:2 communication:17 mansour:1 arbitrary:4 thm:10 pair:4 required:2 optimized:1 mcgregor:3 established:1 nip:7 address:1 beyond:1 usually:2 below:1 regime:6 bien:1 challenge:1 pioneering:1 max:1 memory:32 wainwright:3 natural:3 rely:1 solvable:1 mn:11 minimax:1 x2i:2 identifies:1 xg:3 extract:1 understanding:1 literature:1 acknowledgement:1 asymptotic:1 relative:2 freund:2 loss:13 bear:1 interesting:2 limitation:1 proportional:1 foundation:2 conveyed:2 xiao:1 viewpoint:1 uncorrelated:1 share:1 pi:1 obfuscate:1 course:1 supported:2 free:8 side:3 formal:2 bias:7 allow:2 institute:2 weaker:1 face:1 sparse:20 distributed:21 feedback:11 dimension:8 gram:1 yudin:1 rich:1 doesn:2 berthet:1 made:1 bm:2 polynomially:1 cope:1 transaction:2 emphasize:1 informationtheoretic:1 kullback:1 chien:1 sequentially:4 instantiation:3 reveals:1 assumed:2 xi:8 shwartz:1 alternatively:1 continuous:1 quantifies:2 learn:2 robust:1 correlated:4 interact:5 investigated:1 zou:1 domain:1 protocol:42 vj:1 aistats:1 main:2 yevgeny:1 allowed:3 convey:2 advice:3 intel:1 wiley:1 precision:3 sub:2 wish:3 exponential:3 third:1 theorem:12 specific:9 jt:1 maxi:1 rakhlin:1 exists:3 sequential:3 ci:1 mirror:2 mitliagkas:1 budget:4 illustrates:1 gap:7 entropy:1 likely:2 bubeck:2 seldin:2 ez:2 ordered:1 partially:2 applies:1 corresponds:3 minimizer:1 satisfies:1 relies:2 extracted:3 chance:1 succeed:1 goal:4 identity:1 careful:1 instan:1 price:2 considerable:2 specifically:1 except:2 uniformly:3 miss:1 principal:3 total:1 internal:1 crammer:1 meant:1 wileyinterscience:1 ex:1 |
4,845 | 5,387 | Optimal rates for k-NN density and mode estimation
Sanjoy Dasgupta
University of California, San Diego, CSE
[email protected]
Samory Kpotufe ?
Princeton University, ORFE
[email protected]
Abstract
We present two related contributions of independent interest: (1) high-probability
finite sample rates for k-NN density estimation, and (2) practical mode estimators
? based on k-NN ? which attain minimax-optimal rates under surprisingly general
distributional conditions.
1
Introduction
We prove finite sample bounds for k-nearest neighbor (k-NN) density estimation, and subsequently
apply these bounds to the related problem of mode estimation. These two main results, while related,
are interesting on their own.
First, k-NN density estimation [1] is one of the better known and simplest density estimation procedures. The estimate fk (x) of an unknown density f (see Definition 1 of Section 3) is a simple
n
functional of the distance rk (x) from x to its k-th nearest neighbor in a sample X[n] , {Xi }i=1 .
As such it is intimately related to other functionals of rk (x), e.g. the degree of vertices x in k-NN
graphs and their variants used in modeling communities and in clustering applications (see e.g. [2]).
While this procedure has been known for a long time, its convergence properties are still not fully
understood. The bulk of research in the area has concentrated on establishing its asymptotic convergence, while its finite sample properties have received little attention in comparison. Our finite
sample bounds are concisely derived once the proper tools are identified. The bounds hold with high
probability, under general conditions on the unknown density f . This generality proves quite useful
as shown in our subsequent application to the problem of mode estimation.
The basic problem of estimating the modes (local maxima) of an unknown density f has also been
studied for a while (see e.g. [3] for an early take on the problem). It arises in various unsupervised
problems where modes are used as a measure of typicality of a sample X. In particular, in modern
applications, mode estimation is often used in clustering, with the modes representing cluster centers
(see e.g. [4, 5] and general applications of the popular mean-shift procedure).
While there exists a rich literature on mode estimation, the bulk of theoretical work concerns estimators of a single mode (highest maximum of f ), and often concentrates on procedures that are
hard to implement in practice. Given the generality of our first result on k-NN density estimation,
we can prove that some simple implementable procedures yield optimal estimates of the modes of
an unknown density f , under surprisingly general conditions on f .
Our results are overviewed in the following section, along with an overview of the rich literature on
k-NN density estimation and mode estimation. This is followed by our theoretical setup in Section 3;
our rates for k-NN density estimation are detailed in Section 4, while the results on mode estimation
are given in Section 5.
?
Much of this work was conducted when this author was at TTI-Chicago.
1
2
2.1
Overview of results and related Work
Rates for k-NN density estimates
The k-NN density estimator dates back perhaps to the early work of [1] where it is shown to be
consistent when the unknown density f is continuous on Rd . While one of the best known and
simplest procedure for density estimation, it has proved more cumbersome to analyze than its smooth
counterpart, the kernel density estimator.
More general consistency results such as [6, 7] have been established since its introduction. In
particular [6] shows that, for f Lipschitz in a neighborhood of a point x, where f (x) > 0, and
k = k(n) satisfying k ? ? and k/n2/(2+d) ? 0, the estimator is asymptotically normal, i.e.
?
D
k(fk (x) ? f (x))/f (x) ?? N (0, 1). The recent work of [8], concerning generalized weighted
variants of k-NN, shows that asymptotic normality holds under the weaker restriction k/n4/(4+d) ?
0 if f is twice differentiable at x.
Asymptotic normality as stated above yields
? some insight into the rate of convergence of fk : we
can expect that |fk (x) ? f (x)| . f (x)/ k under the stated conditions on k. In fact, [8] shows
that such a result can be obtained in expectation for n = n(x) sufficiently large. In particular,
their conditions on k allows for a setting of k ? n4/(4+d) (not allowed under the above conditions)
2
yielding a minimax-optimal l2 risk E |fk (x) ? f (x)| . f (x)2 /k = O(n?4/(4+d) ).
While consistency results and bounds on expected error are now well understood, we still don?t have
a clear understanding of the conditions under which high probability bounds on |fk (x) ? f (x)| are
possible. This is particularly important given the inherent instability of nearest neighbors estimates
which are based on order-statistics rather than the more stable average statistics at the core of kerneldensity estimates. The recent result of [9] provides an initial answer: they obtain a high-probability
bound uniformly over x taking value in the sample X[n] , however under conditions not allowing for
optimal settings of k (where f is assumed Lipschitz).
The bounds in the present paper hold with high-probability, simultaneously for all x in the support
of f . Rather than requiring smoothness conditions on f , we simply give the bounds in terms of the
modulus of continuity of f at any x, i.e. how much f can change in a neighborhood of x. This
allows for a useful degree of flexibility in applying these bounds. In particular, optimal bounds
under various degrees of smoothness of f at x easily follow. More importantly, for our application
to mode estimation, the bounds allow us to handle |fk (x) ? f (x)| at different x ? Rd with varying
smoothness in f . As a result we can derive minimax-optimal mode estimation rates for practical
procedures under surprisingly weak assumptions.
2.2
Mode estimation
There is an extensive literature on mode estimation and we unfortunately can only overview some
of the relevant work. Most of the literature covers the case of a unimodal distribution, or one where
there is a single maximizer x0 of f .
Early work on estimating the (single) mode of a distribution focused primarily on understanding the consistency and rates achievable by various approaches, with much less emphasis on the
ease of implementation of these approaches. The common approaches consist of estimating x0 as
x
? , arg supx?Rd fn (x) where fn is an estimate of f , usually a kernel density estimate. Various
work such as [3, 10, 11] establish consistency properties of the approach and achievable rates under
various Euclidean settings and regularity assumptions on the distribution F. More recent work such
as [12, 13] address the problem of optimal choice of bandwidth and kernel to adaptively achieve
the minimax risk for mode estimation. Essentially, under smoothness ? (e.g. f is ? times differentiable), the minimax risk (inf x? supf Ef k?
x ? x0 k) is of the form n?(??1)/(2?+d) , as independently
established in [14] and [15].
As noticed early in [16], the estimator arg supx?Rd fn (x), while yielding much insight into the
problem, is hard to implement in practice. Hence, other work, apparently starting with [16, 14]
have looked into so-called recursive estimators of the (single) mode which are practical and easy
to update as the sample size increases. These approaches can be viewed as some form of gradient2
ascent of fn with carefully chosen step sizes. The later versions of [14] are shown to be minimaxoptimal. Another line of work is that of so-called direct mode estimators which estimate the mode
from practical statistics of the data [17, 18]. In particular, [18] shows that the simple and practical
estimator arg maxx?X[n] fn (x), where fn is a kernel-density estimator, is a consistent estimator of
the mode. We show in the present paper that arg maxx?X[n] fk (x), where fk is a k-NN density
estimator, is not only consistent, but converges at a minimax-optimal rate under surprisingly mild
distributional conditions.
The more general problem of estimating all modes of distribution has received comparatively little
attention. The best known practical approach for this problem is the mean-shift procedure and its
variants [19, 4, 20, 21], quite related to recursive-mode-estimators, as they essentially consist of
gradient ascent of fn starting from every sample point, where fn is required to be appropriately
smooth to ascend (e.g. a smooth kernel estimate). While mean-shift is popular in practice, it has
proved quite difficult to analyze. A recent result of [22] comes close to establishing the consistency
of mean-shift, as it establishes the convergence of the procedure to the right gradient lines (essentially the ascent path to the mode) if it is seeded from fixed starting points rather than the random
samples themselves. It remains unclear however whether mean-shift produces only true modes,
given the inherent variability in estimating f from sample. This question was recently addressed by
[23] which proposes a hypothesis test to detect false modes based on confidence intervals around
Hessians estimated at the modes returned by any procedure.
Interestingly, while a k-NN density estimate fk is far from smooth, in fact not even continuous, we
show a simple practical procedure that identifies any mode of the unknown density f under mild
conditions: we mainly require that f is well approximated by a quadratic in a neighborhood of
each mode. Our finite sample rates (on k?
x ? x0 k, for an estimate x
? of any mode x0 ) are of the
form O(k ?1/4 ), hold with high-probability and are minimax-optimal for an appropriate choice of
k = ?(n4/(4+d) ).
If in addition f is Lipschitz or more generally H?older-continuous (in principle uniform continuity
of f is enough), all the modes returned above a level set ? of fk can be optimally assigned to
n??
separate modes of the unknown f . Since ? ????? 0, the procedure consistently prunes false modes.
This feature is made intrinsic to the procedure by borrowing from insights of [9, 24] on identifying
false clusters by inspecting levels sets of fn . These last works concern the related area of level set
estimation, and do not study mode estimation rates.
As alluded to so far, our results are given in terms of local assumptions on modes rather than
global distributional conditions. We show that any mode that is sufficiently salient (this is locally
parametrized) w.r.t. the finite sample size n, is optimally estimated, while false modes are pruned
away. In particular our results allow for f having a countably infinite number of modes.
3
Preliminaries
n
Throughout the analysis, we assume access to a sample X[n] = {Xi }i=1 drawn i.i.d. from an
absolutely continuous distribution F over Rd , with Lebesgue-density function f . We let X denote
the support of the density function f .
The k-NN density estimate at a point x is defined as follows.
Definition 1 (k-NN density estimate). For every x ? Rd , let rk (x) denote the distance from x to its
k-th nearest neighbor in X[n] . The density estimate is given as:
fk (x) ,
k
,
n ? vd ? rk (x)d
where vd denotes the volume of the unit sphere in Rd .
All balls considered in the analysis are closed Euclidean balls of Rd .
3
4
k-NN density estimation rates
In this section we bound the error in estimating f (x) as fk (x) at every x ? X . The main results of
the section are Lemmas 3 and 4. These lemmas are easily obtained given the right tools: uniform
concentration bounds on the empirical mass of balls in Rd , using relative Vapnik-Chervonenkis
bounds, i.e. Bernstein?s type bounds rather than Chernoff type bounds (see e.g. Theorem 5.1 of
[25]). We next state a form of these bounds for completion.
Lemma 1. Let G be a class of functions from X to {0, 1} with VC dimension d < ?, and P a
probability distribution on X . Let E denote expectation with respect to P. Suppose n points are
drawn independently at random from P; let En denote expectation with respect to this sample. Then
for any ? > 0, with probability at least 1 ? ?, the following holds for all g ? G:
p
p
p
p
? min(?n En g, ?n2 + ?n Eg) ? Eg ? En g ? min(?n2 + ?n En g, ?n Eg),
p
where ?n = (4/n)(d ln 2n + ln(8/?)).
These sort of relative VC bounds allows for a tighter relation (than Chernoff type bounds) between
empirical and true mass of sets (En g and Eg) in those situations where these quantities are small,
?
above. This is particularly useful since the balls we have to deal
i.e. of the order of ?n2 = O(1/n)
with are those containing approximately k points, and hence of (small) mass approximately k/n.
A direct result of the above lemma is the following lemma of [26]. This next lemma essentially
reworks Lemma 1 above into a form we can use more directly. We re-use C?,n below throughout
the analysis.
?
Lemma 2 ([26]). Pick 0 < ? < 1. Let C?,n , 16 log(2/?) d log n. Assume k ? d log n. With
probability at least 1 ? ?, for every ball B ? Rd we have,
?
d log n
F(B) ? C?,n
=? Fn (B) > 0,
? n
k
k
k
F(B) ? + C?,n
=? Fn (B) ? , and
n
n ?
n
k
k
k
F(B) ? ? C?,n
=? Fn (B) < .
n
n
n
The main idea in bounding fk (x) is to bound the random term rk (x) in terms of f (x) using Lemma
2 above. We can deduce from the lemma that if a ball B(x, r) centered has mass roughly k/n, then
its empirical mass is likely to be of the order k/n; hence rk (x) is likely to be close to the radius r
of B(x, r). Now if f does not vary too much in B(x, r), then we can express the mass of B(x, r) in
terms of f (x), and thus get our desired bound on rk (x) and fk (x) in terms of f (x).
Our results are given in terms of how f varies in a neighborhood of x, captured as follows.
n
o
Definition 2. For x ? Rd and > 0, define r?(, x) , sup r : supkx?x0 k?r f (x0 ) ? f (x) ? ,
n
o
and r?(, x) , sup r : supkx?x0 k?r f (x) ? f (x0 ) ? .
The continuity parameters r?(, x) and r?(, x) (related to the modulus of continuity of f at x) are easily bounded under smoothness assumptions on f at x. Our high-probability bounds on the estimates
fk (x) in terms of f (x) and the continuity parameters are given as follows.
2
Lemma 3 (Upper-bound on fk ). Suppose k ? 4C?,n
. Then, with probability at least 1 ? ?, for all
x ? Rd and all > 0,
C?,n
fk (x) < 1 + 2 ?
(f (x) + ) ,
k
provided k satisfies vd ? r?(, x)d ? (f (x) + ) ?
k
n
4
?
? C?,n
k
n .
Lemma 4 (Lower-bound on fk ). Then, with probability at least 1 ? ?, for all x ? Rd and all > 0,
C?,n
fk (x) ? 1 ? ?
(f (x) ? ) ,
k
provided k satisfies vd ? r?(, x)d ? (f (x) ? ) ?
k
n
?
+ C?,n
k
n .
The proof of these results are concise applications of Lemma 2 above. They are given in the appendix
(long version). The trick is in showing that, under the conditions on k, there exists an r ? (k/(n ?
f (x)))1/d which is at most r?(, x) or r?(, x) as appropriate; hence, f does not vary much on B(x, r)
so we must have
k
F (B(x, r)) ? volume (B(x, r)) ? f (x) = vd ? rd ? f (x) ? .
n
?
Using Lemma 2 we get rk (x) ? r; plug this value into fk (x) to obtain fk (x) ? (1 + 1/ k)f (x).
Lemmas 3 and 4 allow a great deal of flexibility as we will soon see with their application to mode
estimation. In particular we can consider various smoothness conditions simultaneously at different
x for different biases .
Suppose for instance that f is locally H?older at x, i.e. ?r, L, ? > 0 s.t. for all x0 ?
0 ?
B(x, r), |f (x) ? f (x0 )| ? L
?kx ? x k . Then for small , both r?(, x) and r?(, x) are at least
1/?
(/L) ; pick = O(f (x)/
? k) for n sufficiently large, then by both lemmas
? we have, w.h.p.,
|fk (x) ? f (x)| ? O(f (x)/ k) provided k = ?(log2 n) and satisfies vd (1/L k)d/? f (x) ? Ck/n
for some constant C. This allows for a setting of k = ? n2?/(2?+d) for a minimax-optimal rate
of |fk (x) ? f (x)| = O n??/(2?+d .
The ability to consider various biases would prove particularly helpful in the next section on
mode estimation where we have to consider different approximations in different parts of space with
varying smoothness in f . In particular, at a mode x, we will essentially have ? = 2 (f is twice
differentiable) while elsewhere on X we might not have much smoothness in f .
5
Mode estimation
We start with the following definition of modes.
Definition 3. We denote the set of modes of f by M ? {x : ?r > 0, ?x0 ? B(x, r), f (x0 ) < f (x)} .
We need the following assumption at modes.
Assumption 1. f is twice differentiable in a neighborhood of every x ? M. We denote the gradient
and Hessian of f by ?f and ?2 f . Furthermore, ?2 f (x) is negative definite at all x ? M.
Assumption 1 excludes modes at the boundary of the support of f (where f cannot be continuously
differentiable). We note that most work on the subject consider only interior modes as we are
doing here. Modes on the boundary can however be handled under additional boundary smoothness
assumptions to ensure that f puts sufficient mass on any ball around such modes. This however only
complicates the analysis, while the main insights remain the same as for interior modes.
An implication of Assumption 1 is that for all x ? M, ?f is continuous in a neighborhood of x,
with ?f (x) = 0. Together with ?2 f (x) ? 0 (i.e. negative definite), f is well-approximated by a
quadratic in a neighborhood of a mode x ? M. This is stated in the following lemma.
Lemma 5. Let f satisfy Assumption 1. Consider any x ? M. Then there exists a neighborhood
B(x, r), r > 0, and constants C?x , C?x > 0 such that, for all x0 ? B(x, r), we have
2
2
C?x kx0 ? xk ? f (x) ? f (x0 ) ? C?x kx0 ? xk .
(1)
We can therefore parametrize a mode x ? M locally as follows:
Definition 4 (Critical radius rx around mode x). For every mode x ? M, there exists rx > 0, such
that B(x, rx ) is contained in a set Ax , satisfying the following conditions:
(i) Ax is a connected component of a level set X ? , {x0 ? X : f (x0 ) > ?} for some ? > 0.
2
2
(ii) ?C?x , C?x > 0, ?x0 ? Ax , C?x kx0 ? xk ? f (x) ? f (x0 ) ? C?x kx0 ? xk . (So Ax ? M = {x}.)
5
Return arg maxx?X[n] fk (x).
Figure 1: Estimate the mode of a unimodal density f from X[n] .
Figure 2: The analysis argues over different regions (depicted) around a mode x.
Finally, we assume that every hill in f corresponds to a mode in M:
Assumption 2. Each connected component of any level set X ? , ? > 0, contains a mode in M.
5.1
Single mode
We start with the simple but common assumption that |M| = 1. This case has been extensively
studied to get a handle on the inherent difficulty of mode estimation. The usual procedures in the
statistical literature are known to be minimax-optimal but are not practical: they invariably return the
maximizer of some density estimator (usually a kernel estimate) over the entire space Rd . Instead
we analyze the practical procedure of Figure 1 where we pick the maximizer of fk out of the finite
sample X[n] . The rates of Theorem 1 are optimal (O(n?1/(4+d) )) for a setting of k = O(n4/(4+d) ).
Theorem 1. Let ? > 0. Assume f has a single mode x0 and satisfies Assumptions 1, 2. There exists
Nx0 ,? such that the following holds for n ? Nx0 ,? . Let C?x0 , C?x0 be as in Definition 4. Suppose k
satisfies
!4d/(4+d)
!2
s
4/(4+d)
24C?,n f (x0 )
1 C?,n
(2d+4)/(4+d) vd
f
(x
)
?
k
?
n
.
(2)
0
2 C?x0
4
C?x0 rx20
Let x be the mode returned in the procedure of Figure 1. With probability at least 1 ? 2? we have
s
C?,n
1
f (x0 ) ? 1/4 .
kx ? x0 k ? 5
k
C?x
0
Proof. Let rx0 be the critical radius of Definition 4. Let rn (x0 ) ? inf r : B(x0 , r) ? X[n] 6= ? .
Let 0 < ? < 1 to be later specified, and assume the event that rn (x0 ) ? ?2 rx0 . We will bound the
probability of this event once the proper setting of ? becomes clear.
Consider r? satisfying rx0 ? r? ? 2rn (x0 )/? (see Figure 2). We will first upper bound fk for any x
outside B(x0 , r?), then lower-bound fk for x ? B(x0 , rn (x0 )).
Recall Ax0 from Definition 4. By equation (1) we have
sup
f (x) ? f (x0 ) ? C?x0 (?
r/2)2 , F? .
(3)
x?Ax0 \B(x0 ,?
r /2)
The above allows us to apply Lemma 3 as follows. First note that for any x ? X \B(x0 , r?/2), f (x) ?
F? since Ax0 is a level set of the unimodal f , i.e. supx?A
/ x0 f (x) ? inf x?Ax0 f (x). Therefore, for
. ?
any x ? X \ B(x0 , r?) let = F ? f (x). By equation (3) the modulus of continuity r?(, x) is at least
6
Initialize: Mn ? ?.
For ? = maxx?X[n] fn (x) down to 0:
?
? Let ? , ? ? C?,n / k.
n om
be the CCs of G (? ? ? ? ?) disjoint from Mn .
? Let A?i
i=1
n
om
? Mn ? Mn ? xi , arg maxx?A?i ?X ? fn (x)
.
[n]
i=1
Return the estimated modes Mn .
Figure 3: Estimate the modes of a multimodal f from X[n] . The parameter ? serves to prune.
r?/2. Therefore, if k satisfies
?
k
k
2
?
vd ? (?
r/2) ? f (x0 ) ? Cx0 (?
r/2) ? ? C?,n
,
n
n
we have with probability at least 1 ? ?
C?,n
sup
fk (x) < 1 + 2 ?
f (x0 ) ? C?x0 (?
r/2)2 .
k
x?X \B(x0 ,?
r)
d
(4)
(5)
Now we turn to x ? B(x0 , rn (x0 )). We have again by equation (1) that inf x?B(x,? r?) f (x) ?
f (x0 ) ? C?x0 (? r?)2 , F? . Therefore, for x ? B(x0 , rn (x0 )) let = f (x) ? F? , we have r?(, x) ?
? r? ? rn (x0 ) ? ? r?/2. It follows that, if k satisfies
?
k
k
d
,
(6)
vd ? ((? /2)?
r) ? f (x0 ) ? C?x0 (? r?)2 ? + C?,n
n
n
we have by Lemma 4 that, with probability at least 1 ? ? (under the same event used in Lemma 3)
C?,n
?
inf
fk (x) ? 1 ?
f (x0 ) ? C?x0 (? r?)2 .
(7)
x?B(x,rn (x0 ))
k
Next, with a bit of algebra, we can pick ? and r? so that the l.h.s. of (5) is less
? than the l.h.s.
2
2
?
?
?
of equation (7). It suffices to pick ? = Cx0 /8Cx0 and r? ? 24f (x0 )C?,n /Cx0 k. Given these
settings, equations (4) and (6) are satisfied whenever k satisfies equation (2) of the lemma statement.
It follows that, with probability at least 1 ? ?, inf x?B(x,rn (x0 )) fk (x) > supx?X \B(x0 ,?r) fk (x).
Therefore, ther
empirical mode chosen by the procedureis in B(x0 , r?). We are free to choose r? as
?
small as max
24f (x0 )C?,n / C?x0 k , 2rn (x0 )/? .
We?ve assumed so far the event that rn (x0 ) ? ?2 rx0 . We bound the probability of this event as
q
?
follows. Let r , 24f (x0 )C?,n /C?x0 k. Under the above setting of ? , the Theorem?s assumptions
?
d
on k imply that r ? rx0 , and that vd ? ((? /2)r) ? f (x0 ) ? C?x0 ((? /2)r)2 ? nk + C?,n nk . Again,
?
by equation (1), this implies that F(B(x0 , (? /2)r)) ? nk + C?,n nk . By Lemma, 2, with probability
at least 1 ? ?, Fn (B(x0 , (? /2)r)) ? k/n and therefore rn (x0 ) ? (? /2)r ? (? /2)rx0 . It now
becomes clear that we can just pick r? = r.
5.2
Multiple modes
In this section we turn to the problem of estimating the modes of a more general density f with an
unknown number of modes.
The algorithm of Figure 3 operates on the following set of nested graphs G(?). These are subgraphs
of a mutual k-NN graph on the sample X[n] , where vertices are connected if they are in each other?s
nearest neighbor sets. The connected components (CCs) of these graphs G(?) are known to be good
estimates of the CCs of corresponding level sets of the unknown density f [9, 26, 27].
7
Definition
Given ? ? R, let G(?) denote the graph with vertices in
5 (k-NN level set G(?)).
?
X[n] , x ? X[n] : fn (x) ? ? , and where vertices x, x0 are connected by an edge when and only
?
when kx ? x0 k ? ? ? min {rk (x), rk (x0 )}, for some ? ? 2.
We will show that for a given n, any sufficiently salient mode is optimally recovered; furthermore,
if f is uniformly continuous on Rd , then the procedure returns no false mode above a level ?n ? 0.
5.2.1
Optimal Recovery for Any Mode
The guarantees of this section would be given in terms of salient modes as defined below. Essentially
a mode x0 is salient if it is separated from other modes by a sufficiently wide and deep valley.
We define saliency in a way similar to [9], but simpler: we only require a wide valley since the
smoothness of f at the mode (as expressed in equation 1) takes care of the depth.
We start with a notion of separation between sets inspired from [26].
Definition 6 (r-separation). A, A0 ? X are r-separated if there exists a (separating) set S ? Rd
such that: every path from A to A0 crosses S, and supx?S+B(0,r) f (x) < inf x?A?A0 f (x).
Our notion of mode saliency follows: for a mode x, we require the critical set Ax of Definition 4 to
be well separated from all components at the level where it appears.
Definition 7 (r-salient Modes). A mode x of f is said to be r-salient for r > 0 if the following
holds. There exist Ax as in Definition 4 (with the corresponding rx , C?x and C?x ), which is a CC of
say X ?x , {x ? X : f (x) ? ?x }. Ax is r-separated from X ?x \ Ax .
The next theorem again yields the optimal rates O(n?1/(4+d) ) for k = O(n4/(4+d) ).
Theorem 2 (Recovery of salient modes). Assume f satisfies Assumptions 1, 2. Suppose
? =
n??
2
?(n) ????? 0. Let x0 be an r-salient mode for some r > 0. Assume k = ? C?,n . Then
there exist N = N (x0 , {?
(n)}) depending on x0 and ?(n) such that the following holds for n ? N .
?
?
Let Ax0 , Cx0 , Cx0 be as in Definition 4, and let ?x0 , inf x?Ax0 f (x). Let ? > 0. Suppose k further
satisfies
!4d/(4+d)
!2
s
4/(4+d)
24C?,n f (x0 )
1 C?,n
(2d+4)/(4+d) vd
.
?
n
?
k
?
x
0
2 C?x0
4
C?x0 min rx20 /4, (r/?)2
Let Mn be the modes returned by the procedure of Figure 3. With probability at least 1 ? 2?, there
exists x ? Mn such that
s
C?,n
1
kx ? x0 k ? 5
f (x0 ) ? 1/4 .
?
k
C x0
5.2.2
Pruning guarantees
The proof of the main theorem of this section is based on Lemma 7.4 of [24].
Theorem 3. Let ? , supx f (x) and r() , supx?Rd max {?
r(, x), r?(, x)}. Assume f satisfies
1/d
Assumption 2. Suppose r(?
) = ? (k/n) , which is feasible whenever f is uniformly continuous
on Rd . In particular, if f is H?older continuous, i.e.
?x, x0 ? Rd ,
?
|f (x) ? f (x0 )| ? L kx ? x0 k , for some L > 0, 0 < ? ? 1,
?/d
since r(?
) ? (?
/L)1/? . Define
)
? !
? 2
k
k
2
+ C?,n
?0 = max 2?
, 8 C?,n ,
.
k
n
n
vd r(?
)d
then we can just let ? = ? (k/n)
(
2
Assume k ? 9C?,n
. The following holds with probability at least 1 ? ?. Pick any ? ? 2?0 , and
?
let ?f = inf x?X ? f (x). All estimated modes in Mn ? X[n]
can be assigned to distinct modes in
[n]
M ? X ?f .
8
References
[1] Don O Loftsgaarden, Charles P Quesenberry, et al. A nonparametric estimate of a multivariate density
function. The Annals of Mathematical Statistics, 36(3):1049?1051, 1965.
[2] M. Maier, M. Hein, and U. von Luxburg. Optimal construction of k-nearest neighbor graphs for identifying noisy clusters. Theoretical Computer Science, 410:1749?1764, 2009.
[3] Emanuel Parzen et al. On estimation of a probability density function and mode. Annals of mathematical
statistics, 33(3):1065?1076, 1962.
[4] Yizong Cheng. Mean shift, mode seeking, and clustering. Pattern Analysis and Machine Intelligence,
IEEE Transactions on, 17(8):790?799, 1995.
[5] Fr?ed?eric Chazal, Leonidas J Guibas, Steve Y Oudot, and Primoz Skraba. Persistence-based clustering in
riemannian manifolds. Journal of the ACM (JACM), 60(6):41, 2013.
[6] David S Moore and James W Yackel. Large sample properties of nearest neighbor density function
estimators. Technical report, DTIC Document, 1976.
[7] L.P. Devroye and T.J. Wagner. The strong uniform consistency of nearest neighbor density estimates. The
Annals of Statistics, 5:536?540, 1977.
[8] G?erard Biau, Fr?ed?eric Chazal, David Cohen-Steiner, Luc Devroye, Carlos Rodriguez, et al. A weighted
k-nearest neighbor density estimate for geometric inference. Electronic Journal of Statistics, 5:204?237,
2011.
[9] S. Kpotufe and U. von Luxburg. Pruning nearest neighbor cluster trees. In International Conference on
Machine Learning, 2011.
[10] Herman Chernoff. Estimation of the mode. Annals of the Institute of Statistical Mathematics, 16(1):31?
41, 1964.
[11] William F Eddy et al. Optimum kernel estimators of the mode. The Annals of Statistics, 8(4):870?882,
1980.
[12] Birgit Grund, Peter Hall, et al. On the minimisation of lp error in mode estimation. The Annals of
Statistics, 23(6):2264?2284, 1995.
[13] Jussi Klemel?a. Adaptive estimation of the mode of a multivariate density. Journal of Nonparametric
Statistics, 17(1):83?105, 2005.
[14] Aleksandr Borisovich Tsybakov. Recursive estimation of the mode of a multivariate distribution. Problemy Peredachi Informatsii, 26(1):38?45, 1990.
[15] David L Donoho and Richard C Liu. Geometrizing rates of convergence, iii. The Annals of Statistics,
pages 668?701, 1991.
[16] Luc Devroye. Recursive estimation of the mode of a multivariate density. Canadian Journal of Statistics,
7(2):159?167, 1979.
[17] Ulf Grenander et al. Some direct estimates of the mode. The Annals of Mathematical Statistics, 36(1):131?
138, 1965.
[18] Christophe Abraham, G?erard Biau, and Beno??t Cadre. On the asymptotic properties of a simple estimate
of the mode. ESAIM: Probability and Statistics, 8:1?11, 2004.
[19] Keinosuke Fukunaga and Larry Hostetler. The estimation of the gradient of a density function, with
applications in pattern recognition. Information Theory, IEEE Transactions on, 21(1):32?40, 1975.
[20] Dorin Comaniciu and Peter Meer. Mean shift: A robust approach toward feature space analysis. Pattern
Analysis and Machine Intelligence, IEEE Transactions on, 24(5):603?619, 2002.
[21] Jia Li, Surajit Ray, and Bruce G Lindsay. A nonparametric statistical approach to clustering via mode
identification. Journal of Machine Learning Research, 8(8), 2007.
[22] Ery Arias-Castro, David Mason, and Bruno Pelletier. On the estimation of the gradient lines of a density
and the consistency of the mean-shift algorithm. Unpublished Manuscript, 2013.
[23] Christopher Genovese, Marco Perone-Pacifico, Isabella Verdinelli, and Larry Wasserman. Nonparametric
inference for density modes. arXiv preprint arXiv:1312.7567, 2013.
[24] K. Chaudhuri, S. Dasgupta, S. Kpotufe, and U. von Luxburg. Consistent procedures for cluster tree
estimation and pruning. Arxiv, 2014.
[25] O. Bousquet, S. Boucheron, and G. Lugosi. Introduction to statistical learning theory. Lecture Notes in
Artificial Intelligence, 3176:169?207, 2004.
[26] K. Chaudhuri and S. Dasgupta. Rates for convergence for the cluster tree. In Advances in Neural Information Processing Systems, 2010.
[27] S. Balakrishnan, S. Narayanan, A. Rinaldo, A. Singh, and L. Wasserman. Cluster trees on manifolds. In
Advances in Neural Information Processing Systems, pages 2679?2687, 2013.
9
| 5387 |@word mild:2 version:2 achievable:2 eng:1 pick:7 concise:1 initial:1 liu:1 contains:1 chervonenkis:1 document:1 interestingly:1 kx0:4 steiner:1 recovered:1 must:1 fn:16 chicago:1 subsequent:1 update:1 intelligence:3 xk:4 core:1 provides:1 cse:1 simpler:1 mathematical:3 along:1 direct:3 prove:3 ray:1 x0:88 ascend:1 expected:1 roughly:1 themselves:1 inspired:1 little:2 becomes:2 provided:3 estimating:7 bounded:1 mass:7 guarantee:2 every:8 birgit:1 unit:1 understood:2 local:2 aleksandr:1 establishing:2 path:2 approximately:2 lugosi:1 might:1 twice:3 emphasis:1 studied:2 ease:1 practical:9 practice:3 recursive:4 implement:2 definite:2 cadre:1 procedure:20 area:2 empirical:4 maxx:5 attain:1 isabella:1 persistence:1 confidence:1 get:3 cannot:1 close:2 interior:2 valley:2 put:1 risk:3 applying:1 instability:1 restriction:1 center:1 attention:2 starting:3 independently:2 typicality:1 focused:1 identifying:2 recovery:2 wasserman:2 subgraphs:1 estimator:16 insight:4 importantly:1 meer:1 handle:2 notion:2 beno:1 annals:8 diego:1 suppose:7 construction:1 lindsay:1 hypothesis:1 trick:1 satisfying:3 particularly:3 approximated:2 recognition:1 distributional:3 preprint:1 region:1 connected:5 highest:1 grund:1 singh:1 algebra:1 eric:2 easily:3 multimodal:1 various:7 separated:4 distinct:1 artificial:1 neighborhood:8 outside:1 quite:3 skraba:1 say:1 ability:1 statistic:14 noisy:1 differentiable:5 grenander:1 fr:2 relevant:1 date:1 flexibility:2 achieve:1 chaudhuri:2 convergence:6 cluster:7 regularity:1 optimum:1 produce:1 tti:1 converges:1 derive:1 depending:1 completion:1 nearest:10 received:2 strong:1 come:1 implies:1 concentrate:1 radius:3 subsequently:1 vc:2 centered:1 larry:2 require:3 suffices:1 preliminary:1 tighter:1 inspecting:1 hold:9 marco:1 sufficiently:5 around:4 considered:1 normal:1 guibas:1 great:1 hall:1 vary:2 early:4 dorin:1 estimation:36 establishes:1 tool:2 weighted:2 rather:5 ck:1 varying:2 minimisation:1 derived:1 ax:8 consistently:1 mainly:1 detect:1 problemy:1 helpful:1 inference:2 nn:19 entire:1 a0:3 borrowing:1 relation:1 arg:6 proposes:1 initialize:1 mutual:1 once:2 having:1 chernoff:3 unsupervised:1 genovese:1 report:1 inherent:3 primarily:1 richard:1 modern:1 simultaneously:2 ve:1 geometrizing:1 lebesgue:1 william:1 invariably:1 interest:1 yielding:2 implication:1 edge:1 tree:4 euclidean:2 re:1 desired:1 hein:1 theoretical:3 complicates:1 instance:1 modeling:1 cover:1 ax0:6 vertex:4 uniform:3 conducted:1 too:1 optimally:3 answer:1 supx:7 varies:1 adaptively:1 density:42 international:1 klemel:1 together:1 continuously:1 parzen:1 again:3 von:3 satisfied:1 containing:1 choose:1 return:4 li:1 satisfy:1 leonidas:1 later:2 closed:1 analyze:3 apparently:1 sup:4 start:3 sort:1 doing:1 carlos:1 keinosuke:1 ery:1 jia:1 bruce:1 contribution:1 om:2 oudot:1 maier:1 yield:3 saliency:2 biau:2 ulf:1 weak:1 identification:1 rx:4 cc:4 chazal:2 cumbersome:1 whenever:2 ed:2 definition:15 james:1 proof:3 riemannian:1 emanuel:1 proved:2 popular:2 recall:1 eddy:1 carefully:1 back:1 appears:1 manuscript:1 steve:1 follow:1 generality:2 furthermore:2 just:2 hostetler:1 christopher:1 maximizer:3 rodriguez:1 continuity:6 mode:98 perhaps:1 modulus:3 requiring:1 true:2 counterpart:1 hence:4 seeded:1 assigned:2 boucheron:1 moore:1 eg:4 deal:2 comaniciu:1 generalized:1 yizong:1 hill:1 quesenberry:1 argues:1 cx0:6 ef:1 recently:1 charles:1 common:2 functional:1 overview:3 cohen:1 volume:2 smoothness:10 rd:20 fk:32 consistency:7 mathematics:1 bruno:1 stable:1 access:1 deduce:1 multivariate:4 own:1 orfe:1 recent:4 inf:9 christophe:1 captured:1 additional:1 care:1 prune:2 borisovich:1 ii:1 multiple:1 unimodal:3 smooth:4 technical:1 plug:1 cross:1 long:2 sphere:1 concerning:1 variant:3 basic:1 essentially:6 expectation:3 arxiv:3 kernel:7 addition:1 addressed:1 interval:1 appropriately:1 ascent:3 subject:1 balakrishnan:1 bernstein:1 iii:1 easy:1 enough:1 canadian:1 bandwidth:1 identified:1 idea:1 shift:8 whether:1 handled:1 peter:2 returned:4 hessian:2 deep:1 useful:3 generally:1 detailed:1 clear:3 nonparametric:4 tsybakov:1 locally:3 extensively:1 concentrated:1 narayanan:1 simplest:2 exist:2 rx0:6 estimated:4 disjoint:1 bulk:2 dasgupta:4 express:1 salient:8 drawn:2 graph:6 asymptotically:1 excludes:1 luxburg:3 throughout:2 electronic:1 separation:2 appendix:1 bit:1 bound:29 followed:1 cheng:1 quadratic:2 informatsii:1 bousquet:1 min:4 fukunaga:1 pruned:1 pelletier:1 ball:7 remain:1 intimately:1 lp:1 n4:5 castro:1 jussi:1 ln:2 alluded:1 equation:8 remains:1 turn:2 serf:1 parametrize:1 apply:2 away:1 appropriate:2 denotes:1 clustering:5 ensure:1 log2:1 pacifico:1 prof:1 establish:1 comparatively:1 seeking:1 noticed:1 question:1 quantity:1 looked:1 concentration:1 usual:1 unclear:1 said:1 gradient:5 distance:2 separate:1 separating:1 parametrized:1 vd:11 manifold:2 toward:1 devroye:3 setup:1 unfortunately:1 difficult:1 minimaxoptimal:1 statement:1 stated:3 negative:2 implementation:1 proper:2 unknown:9 kpotufe:3 allowing:1 upper:2 finite:7 implementable:1 situation:1 variability:1 rn:12 ucsd:1 community:1 david:4 unpublished:1 required:1 specified:1 extensive:1 california:1 concisely:1 established:2 address:1 usually:2 below:2 pattern:3 herman:1 max:3 critical:3 event:5 difficulty:1 mn:8 minimax:9 representing:1 normality:2 older:3 esaim:1 imply:1 identifies:1 literature:5 l2:1 understanding:2 geometric:1 asymptotic:4 relative:2 fully:1 expect:1 lecture:1 interesting:1 degree:3 sufficient:1 consistent:4 principle:1 elsewhere:1 surprisingly:4 last:1 soon:1 free:1 bias:2 weaker:1 allow:3 institute:1 neighbor:10 wide:2 taking:1 wagner:1 peredachi:1 boundary:3 dimension:1 depth:1 rich:2 author:1 made:1 adaptive:1 san:1 far:3 transaction:3 functionals:1 pruning:3 countably:1 global:1 assumed:2 xi:3 don:2 continuous:8 loftsgaarden:1 robust:1 main:5 abraham:1 bounding:1 n2:5 allowed:1 en:5 samory:2 rk:10 theorem:8 down:1 showing:1 mason:1 concern:2 exists:7 consist:2 intrinsic:1 false:5 vapnik:1 perone:1 aria:1 kx:5 nk:4 dtic:1 supf:1 depicted:1 simply:1 likely:2 jacm:1 surajit:1 erard:2 rinaldo:1 expressed:1 contained:1 corresponds:1 nested:1 satisfies:11 acm:1 viewed:1 donoho:1 lipschitz:3 luc:2 feasible:1 hard:2 change:1 infinite:1 uniformly:3 operates:1 lemma:24 called:2 sanjoy:1 verdinelli:1 support:3 overviewed:1 arises:1 absolutely:1 princeton:2 |
4,846 | 5,388 | Learning on graphs using Orthonormal
Representation is Statistically Consistent
Chiranjib Bhattacharyya
Department of CSA
Indian Institute of Science
Bangalore, 560012, INDIA
[email protected]
Rakesh S
Department of Electrical Engineering
Indian Institute of Science
Bangalore, 560012, INDIA
[email protected]
Abstract
Existing research [4] suggests that embedding graphs on a unit sphere can be beneficial in learning labels on the vertices of a graph. However the choice of optimal
embedding remains an open issue. Orthonormal representation of graphs, a class
of embeddings over the unit sphere, was introduced by Lov?asz [2]. In this paper,
we show that there exists orthonormal representations which are statistically consistent over a large class of graphs, including power law and random graphs. This
result is achieved by extending the notion of consistency designed in the inductive
setting to graph transduction. As part of the analysis, we explicitly derive relationships between the Rademacher complexity measure and structural properties
of graphs, such as the chromatic number. We further show the fraction of vertices
of a graph G, on n nodes, that need to be labelled for the learning algorithm to be
14
consistent, also known as labelled sample complexity, is ? ?(G)
where ?(G)
n
is the famous Lov?asz ? function of the graph. This, for the first time, relates labelled sample complexity to graph connectivity properties, such as the density of
graphs. In the multiview setting, whenever individual views are expressed by a
graph, it is a well known heuristic that a convex combination of Laplacians [7]
tend to improve accuracy. The analysis presented here easily extends to Multiple graph transduction, and helps develop a sound statistical understanding of the
heuristic, previously unavailable.
1
Introduction
In this paper we study the problem of graph transduction on a simple, undirected graph G = (V, E),
with vertex set V = [n] and edge set E ? V ?V . We consider individual vertices to be labelled with
binary values, ?1. Without loss of generality we assume that the first f n vertices are labelled, i.e.,
the set of labelled vertices is given by S = [f n], where f ? (0, 1). Let S? = V \S be the unlabelled
vertex set, and let yS and yS? be the labels corresponding to subgraphs S and S? respectively.
? ? Rn , such that erS0-1
Given G and yS , the goal of graph transduction is to learn predictions y
y] =
? [?
P
? = sgn(?
?j , y
y) is small. To aid further discussion we introduce some notations.
? 1 yj 6= y
j?S
Notation Let S n?1 = {u ? Rn |kuk2 = 1} denote a (n ? 1) dimensional sphere. Let Dn , Sn
and S+
n denote a set of n ? n diagonal, square symmetric and square symmetric positive semidefinite matrices respectively. Let Rn+ be a non-negative orthant. Let 1n ? Rn denote a vector
of all 1?s. Let [n] := {1, . . . , n}. For any M ? Sn , let ?1 (M) ? . . . ? ?n (M) denote the
eigenvalues and Mi denote the ith row of M, ?i ? [n]. We denote the adjacency matrix of a
graph G by A. Let di denote the degree of vertex i ? [n], di := A>
i 1n . Let D ? Dn , where
1
1
1
Dii = di , ?i ? [n]. We refer I ? D? 2 AD? 2 as the Laplacian, where I denotes the identity matrix.
? = 1n 1n > ? I ? A. For
? denote the complement graph of G, with the adjacency matrix A
Let G
n
K ? S+
and
y
?
{?1}
,
the
dual
formulation
of
Support
vector
machine
(SVM) is given by
n
P
n
n
P
?(K, y) = max??Rn+ g(?, K, y) =
?i ? 21
?i ?j yi yj Kij . Let Y = Y? = {?1}, Yb ? R
i=1
i,j=1
be the label, prediction and soft-prediction spaces over V . Given a graph G and labels y ? Y n
P
b
on V , let cut(A, y) :=
yi 6=yj Aij . We use ` : Y ? Y ? R+ to denote any loss function.
0-1
b let ` (a, b) = 1[ab < 0], `hinge (a, b) = (1 ? ab)+ 1 and
In particular, for a ? Y, b ? Y,
ramp
`
(a, b) = min(1, (1 ? ab)+ ) denote the 0-1, hinge and ramp loss respectively. The notations
O, o, ?, ? will denote standard measures defined in asymptotic analysis [14].
Motivation Regularization framework is a widely used tool for learning labels on the vertices of
a graph [23, 4]
1 X
?
minn
`(yi , y?i ) + ??
y> K?1 y
(1)
? ?Y |S|
y
i?S
where K is a kernel matrix and ? > 0 is an appropriately chosen regularization parameter. It was
? ? satisfies the following generalization bound
shown in [4] that the optimal y
tr (K) p
p
? + c2
y] + ??
y> K?1 y
ES erS0-1
y? ] ? c1 inf n erV [?
? [?
? ?Y
y
?|S|
1/p
P
Pn
(?)
1
(?)
?i ), H ? V 2 ; trp (K) = n1 i=1 Kpii
where erH [?
y] := |H|
, p > 0 and
i?H ` (yi , y
c1 , c2 are dependent on `. [4] argued that for good generalization, trp (K) should be a constant,
which motivated them to normalize the diagonal entries of K. It is important to note that the set of
normalized kernels is quite big and the above presented analysis gives little insight in choosing the
optimal kernel from such a set.
The important problem of consistency erS? ? 0 as n ? ?, to be
formally defined in Section 3 of graph transduction algorithms was introduced in [5]. [5] showed
that the formulation (1), when
q used
with a laplacian dependent kernel, achieves a generalization
q
3
error of ES [erS? [?
y? ]] = O
nf , where q is the number of pure components . Though [5]?s
algorithm is consistent for a small number of pure components, they achieve the above convergence
rate by choosing ? dependent on true labels of the unlabeled nodes, which is not practical [6].
In this paper, we formalize the notion of consistency of graph transduction algorithms and derive
novel graph-dependent statistical estimates for the following formulation.
X
X
1
?C (K, yS ) = min
minn ?> K? + C
` y?i , yi + C
` y?j , y?j
(2)
?
?
??R
? j ?Y,j?S
y
+ 2
?
i?S
P
j?S
P
where y?k = i?S Kik yi ?i + j?S? Kjk y?j ?j , ?k ? V . If all the labels are observed then [22]
showed that the above formulation is equivalent to (1). We note that the normalization step considered by [4] is equivalent to finding an embedding of a graph on a sphere. Thus, we study orthonormal representations of graphs [2], which define a rich class of graph embeddings on an unit
sphere. We show that the formulation (2) working with orthonormal representations of graphs is
statistically consistent over a large class of graphs, including random and power law graphs. In the
sequel, we apply Rademacher complexity to orthonormal representations of graphs and derive novel
graph-dependent transductive error bound. We also extend our analysis to study multiple graph
transduction. More specifically, we make the following contributions.
Contributions The main contribution of this paper is that we show there exists orthonormal representations of graphs that are statistically consistent on a large class of graph families Gc . For a
special orthonormal representation?LS labelling, we show consistency on Erd?os R?enyi random
graphs. Given a graph G ? Gc , with a constant fraction of nodes labelled f = O(1), we derive
1
(a)+ = max(a, 0).
? , when implicit from the context.
We drop the argument y
3
Pure component is a connected subgraph, where all the nodes in the subgraph have the same label.
2
2
an error convergence rate of erS0-1
? = O
?(G)
n
14
, with high probability; where ?(G) is the Lov?asz
pq
? function of the graph G. Existing work [5] showed an expected convergence rate of O
n ,
however q is dependent on the true labels of the unlabelled nodes. Hence their bound cannot be
computed explicitly [6]. We also apply Rademacher complexity measure to the function class associated with orthonormal representations and derive a tight bound relating to ?(G), the chromatic
number of the graph G. We show that the Laplacian inverse [4] has O(1) complexity on graphs with
1
high connectivity, whereas LS labelling exhibits a complexity of ?(n 4 ). Experiments demonstrate
superior performance of LS labelling on several real world datasets. We derive novel transductive
error bound, relating to graph structural measures. Using our analysis, we show that observing labels
41
fraction of the nodes is sufficient to achieve consistency. We also propose an effiof ? ?(G)
n
cient Multiple Kernel Learning (MKL) based algorithm, with generalization guarantees for multiple
graph transduction. Experiments demonstrate improved performance in combining multiple graphs.
2
Preliminaries
Orthonormal Representation: [2] introduced the idea of orthonormal representations for the problem of embedding a graph on a unit sphere. More formally, an orthonormal representation of a
simple, undirected graph G = (V, E) with V = [n], is a matrix U = [u1 , . . . , un ] ? Rd?n such
that uTi uj = 0 whenever (i, j) ?
/ E and ui ? S d?1 ?i ? [n].
Let Lab(G)denote the set of all possible orthonormal
representations of the graph G given by
Lab(G) := U|U is an Orthonormal Representation . [1] recently introduced the notion of graph
embedding to Machine Learning community and showed connections to graph kernel matrices. Con/ E}.
sider the set of graph kernels K(G) := {K ? S+
n |Kii = 1, ?i ? [n]; Kij = 0, ?(i, j) ?
[1] showed that for every valid kernel K ? K(G), there exists an orthonormal representation
U ? Lab(G); and it is easy to see the other way, K = U> U ? K(G). Thus, the two sets,
Lab(G) and K(G) are equivalent. Orthonormal representation is also associated
with an interesting
quantity, the Lov?asz number [2], defined as: ?(G) = 2 minK?K(G) ?(K, 1n ) [1]. ? function is a
fundamental tool for combinatorial optimization and approximation algorithms for graphs.
? ?? G
? ? ?(G);
Lov?asz Sandwich
Theorem:
[2]
Given
an
undirected
graph
G
=
(V,
E),
I
G
? is the independent number of the complement graph G.
?
where I G
3
Statistical Consistency of Graph Transduction Algorithms
In this section, we formalize the notion of consistency of graph transduction algorithms. Given
a graph Gn = (Vn , En ) of n nodes, with labels of subgraph Sn ? Vn observable, let erS??n :=
y] denote the minimal unlabelled node set error. Consistency is a measure of the
inf y? ?Y? n erS?n [?
? are the predictions made
quality of the learning algorithm A, comparing erS?n [?
y] to er?S?n , where y
by A. A related notion of loss consistency has been extensively studied in literature [3, 12], which
only show that the difference erS?n [?
y] ? erSn [?
y] ? 0 as n ? ? [6]. This does not confirm the
optimality of A, that is erS?n [?
y] ? erS??n . Hence, a notion stronger than loss consistency is needed.
Let Gn belong to a graph family G, ?n. Let ?f be the uniform distribution over the random draw of
the labelled subgraph Sn ? Vn , such that |Sn | = f n, f ? (0, 1). As discussed earlier, we want the
`-regret, RSn [A] = erS?n [?
y] ? erS??n to be small. Since, the labelled nodes are drawn randomly, there
is a small probability that one gets an unrepresentative subgraph Sn . However, for large n, we want
`-regret to be close to zero with high probability4 . In other words, for every finite and fixed n, we
want to have an estimate on the `-regret, which decreases as n increases. We define the following
notion of consistency of graph transduction algorithms to capture this requirement
Definition 1. Let G be a graph family and f ? (0, 1) be fixed. Let V = {(vi , yi , Ei )}?
i=1 be an
infinite sequence of labelled node set, where yi ? Y and Ei is the edge information of node vi
with the previously observed nodes v1 , . . . , vi?1 , ?i ? 2. Let Vn be the first n nodes in V, and let
4
If G is not deterministic (e.g., Erd?os R?eyni), then there is small probability that one gets an unrepresentative
graph, in which case we want the `-regret to be close to zero with high probability over Gn ? G.
3
Gn ? G be the graph defined by (Vn , E1 , . . . , En ). Let Sn ? Vn , and let yn , ySn be the labels of
? is
Vn , Sn respectively. A learning algorithm A when given Gn and ySn returns soft-predictions y
said to be `-consistent w.r.t G if, when the labelled subgraph Sn are random drawn from ?f , the
`-regret converges in probability to zero, i.e., ? > 0
PrSn ??f [RSn [A] ? ] ? 0
as
n??
In Section 6 we show that the kernel learning style algorithm (2) working with orthonormal representations is consistent on a large class of graph families. To the best of our knowledge, we are
not aware of any literature which provide an explicit empirical error convergence rate and prove
consistency of the graph transduction algorithm considered. Before we prove our main result, we
gather useful tools?a) complexity measure, which reacts to the structural properties of the graph
(Section 4); b) generalization analysis to bound erS? (Section 5). In the interest of space, we defer
most of the proofs to the supplementary material5 .
4
Graph Complexity Measures
In this section we apply Rademacher complexity to orthonormal representations of graphs, and relate
to the chromatic number. In particular, we study LS labelling, whose class complexity can be shown
to be greater than that of the Laplacian inverse on a large class of graphs.
Let (2) be solved for K ? K(G), and let U ? Lab(G) be the orthonormal representation corresponding to K (Section 2). Then by Representer?s theorem, the classifier learnt by (2) is of the
form h = U?, ? ? Rn . We define Rademacher complexity of the function class associated with
orthonormal representations
Definition 2(Rademacher Complexity).
Given a graph G = (V, E), with V = [n]; let U ? Lab(G)
? U = h|h = U?, ? ? Rn be the function class associated with U. For p ? (0, 1/2], let
and H
? = (?1 , . . . , ?n ) be a vector of i.i.d. random variables such that ?i ? {+1, ?1, 0} w.p. p, p and
? U is given by
1 ? 2p respectively. The Rademacher complexity of the graph G defined by U, H
h
i
n
P
? U , p) = 1 E? suph?H?
R(H
?i hh, ui i
n
U
i=1
The above definition is motivated from [9, 3]. This is an empirical complexity measure, suited for
the transductive settings. We derive the following novel tight Rademacher bound
Theorem
4.1. Let G = (V, E)
= [n], U ? Lab(G) and
be a simple, undirected graph with
? V
p ? 1/n, 1/2 . Let HU = h h = U?, ? ? Rn , k?k2 ? tC n , C > 0, t ? [0, 1] and let
K = U> U ? K(G) be the graph-kernel corresponding
to U. The Rademacher
complexity
of graph
p
?
?
G defined by U is given by R(HU , p) = c0 tC p?1 (K), where 1/2 2 ? c0 ? 2 is a constant.
The above result provides a lower bound for the Rademacher complexity for any unit sphere graph
embedding. While upper-bounds maybe available [9, 3] but, to the best of our knowledge, this is the
first attempt at establishing lower bounds. The use of orthonormal representations allow us to relate
class complexity measure to graph-structural properties.
p
Corollary 4.2. For C, t, p = O(1), R(HU , p) = O( ?(G)). (Suppl.)
Such connections between learning theory complexity measures and graph properties was previously
unavailable [9,p3]. Corollary 4.2 suggests that there exists graph regularizers with class complexity
as large as O( ?(G)), which motivate us to find substantially better regularizers. In particular, we
investigate LS labelling [16]; given a graph G, LS labelling KLS ? K(G) is defined as
KLS =
A
+ I, ? ? |?n (A)|
?
(3)
LS labellinghas high Rademacher complexity on a large class of graphs, in particular
Corollary 4.3. For a random graph G(n, q), q ? [0, 1), where each edge is present independently
w.p. q; for C, t, q = O(1) the Rademacher complexity of the function class associated with LS
1
labelling (3) is ?(n 4 ), with high probability. (Suppl.)
5
mllab.csa.iisc.ernet.in/rakeshs/nips14/suppl.pdf
4
For the limiting case of complete graphs, we can show that Laplacian inverse [4], the most widely
used graph regularizer has O(1) complexity (Claim 2, Suppl.), thus indicating that it may be suboptimal for graphs with high connectivity. Experimental results illustrate our observation.
We derive a class complexity measure for unit sphere graph embeddings, which indicates the richness of the function class, and helps the learning algorithm to choose an effective embedding.
5
Generalization Error Bound
In the previous section, we applied Rademacher complexity to orthonormal representations. In
this section we derive novel graph-dependent generalization error bounds, which will be used in
Section 6. Following a similar proof technique as in [3], we propose the following error bound?
Theorem 5.1. Given a graph G = (V, E), V = [n] with y ? Y n being the unknown binary labels
? U = {h|h = U?, ? ?
over V ; let U ? Lab(G), and K ? K(G) be the corresponding kernel. Let H
Rn , k?k? ? C}, C > 0. Let ` be any loss function, bounded in [0, B] and L-Lipschitz in its
second argument. For f ? (0, 1/2]6 , let labels of subgraph S ? V be observable, |S| = nf . Let
? U , with probability ? 1 ? ? over S ? ?f
S? = V \S. For any ? > 0 and h ? H
s
r
2?1 (K)
1
c1 B
1
y] ? erS [?
y] + LC
erS? [?
+
log
(4)
f (1 ? f ) 1 ? f nf
?
? = U> h and c1 > 0 is a constant. (Suppl.)
where y
Discussion Note that from [2], ?1 (K) ? ?(G) and ?(G) is in-turn bounded by the maximum
degree of the graph [21]. Thus, if ?
L, B, f = O(1), then for sparse, degree bounded graphs; for
the choice of parameter C = ?(1/ n), the slack term and the complexity term goes to zero as n
increases. Thus, making the bound useful. Examples include tree, cycle, path, star and d-regular
(with d = O(1)). Such connections relating generalization error to graph properties was not available before. We exploit this novel connection to analyze graph transduction algorithms in Section 6.
Also, in Section 7, we extend the above result to the problem of multiple graph transduction.
5.1
Max-margin Orthonormal Representation
To analyze erS0-1 relating to graph structural measure, the ? function, we study the maximum margin
induced by any orthonormal representation, in an oracle setting.
We study a fully ?labelled graph? G = (V, E, y), where y ? Y n are the binary labels on the vertices
V . Given any U ? Lab(G), the maximum margin classifier is computed by solving ?(K, y) =
g(?? , K, y) where K = U> U ? K(G). It is interesting to note that knowing all the labels, the
max-margin orthonormal representation can be computed by solving an SDP. More formally
Definition 3. Given aS
labelled graph G = (V, E, y), where V = [n] and y ? Y n are the binary
?
? U , where H
? U = {h|h = U?, ? ? Rn }. Let K ? K(G) be
labels on V , let H = U?Lab(G) H
the kernel corresponding to U ? Lab(G). The max-margin orthonormal representation associated
with the kernel function is given by Kmm = argminK?K(G) ?(K, y).
By definition, Kmm induces the largest margin amongst other orthonormal representations, and
hence is optimal. The optimal margin has interesting connections to the Lov?asz ? function ?
Theorem 5.2. Given a labelled graph G = (V, E, y), with V = [n] and y ? Y n being the binary
labels on vertices. Let Kmm be as in Definition 3, then ?(Kmm , y) = ?(G)/2. (Suppl.)
Thus, knowing all the labels, computing Kmm is equivalent to solving the ? function. However,
in the transductive setting, Kmm cannot be computed. Alternatively, we explore LS labelling (3),
which gives a constant factor approximation to the optimal margin on a large class of graphs.
Definition 4. A class of labelled graphs G = {G = (V, E, y)} is said to be a Labelled SVM-? graph
family, if there exist a constant ? > 1 such that ?G ? G, ?(KLS , y) ? ??(Kmm , y).
6
We can generalize our result for f ? (0, 1), but for the simplicity of the proof we assume f ? (0, 1/2].
This is also true in practice, where the number of labelled examples is usually very small.
5
Algorithm 1
Input: U, yS and C > 0.
? S?? by solving ?C (K, yS ) (2) for `hinge and K = U> U.
Get ?? , y
? = U> hS , where hS = UY?? ; Y ? Dn , Y = yi , if i ? S, otherwise y?i? .
Return: y
Such class of graphs are interesting, because one can get a constant factor approximation to the
optimal margin, without the knowledge of the
? true labels e.g., Mixture of random graphs: G =
(V, E, y), with y> 1n = 0, cut(A, y) ? c n, for c > 1 being a constant and the subgraphs
corresponding to the two classes form G(n/2, 1/2) random graphs (Claim 3, Suppl.).
We relate the maximum geometric margin induced by orthonormal representations to the ? function
of the graph. This allows us to derive novel graph dependent learning theory estimates.
6
Consistency of Orthonormal Representation of Graphs
Aggregating results from Section 4 and 5, we show that Algorithm 1 working with orthonormal
representations of graphs is consistent on a large class of graph families. For every finite and fixed
n, we derive an estimate on erS0-1
?n .
? be the predictions
Theorem 6.1. For the setting as in Definition 1, let f ? (0, 1/2] be fixed. Let y
14
?2 (Gn )(1?f )
?
learnt by Algorithm 1 with inputs Un ? Lab(Gn ), ySn and C = 23 n2 f ? G?
. Then ?Un ?
( n)
Lab(Gn ), ?Gn such that with probability atleast 1 ? n1 over Sn ? ?f
s
!
41
?(G
)
1
log n
n
0-1
y] = O
+
erS?n [?
f 3 (1 ? f )n
1?f
nf
Proof. Let Kn ? K(Gn ) be the max-margin kernel associated with Gn (Definition 3), and let
Un ? Lab(G) be the corresponding orthonormal representation. Since `ramp is an upper bound on
`0-1 , we concentrate on bounding erSramp
[?
y]. Note that for any C > 0
?n
C|Sn | ? erSramp
[?
y] ? C|Sn | ? erShinge
[?
y] ? ?C (Kn , ySn )
n
n
?(Gn )
2
The last inequality follows from Theorem 5.2. Note that for ramp loss L = B = 1; using Theorem 5.1 for ? = n1 , it follows that with probability atleast 1 ? n1 over the random draw of Sn ? ?f ,
s
s
?(Gn )
2?1 (Kn )
c1
log n
ramp
erS?n [?
y] ?
+C
+
(5)
2Cnf
f (1 ? f ) 1 ? f
nf
? ?C (Kn , yn ) ? ?(Kn , yn ) =
? n ) [2] and optimizing RHS for C, we get C ? =
where c1 = O(1). Using ?1 (Kn ) ? ?(G
2
41
? (Gn )(1?f )
? n = n [2] proves the claim.
. Plugging back C ? and using ?(Gn )? G
?n)
23 n2 f ?(G
pq
[5] showed that ES erS?n = O
n . However, as noted in Section 1, the quantity q is dependent
on yS?n , and hence their bounds cannot be computed explicitly [6].
We assume that the graph does not contain duplicate nodes with opposite labels, erS??n = 0. Thus,
consistency follows from the fact that ?(G) ? n and for large families of graphs it is O(nc ) where
0 ? c < 1. This theorem points to the fact that if f = O(1), then by Definition 1, Algorithm 1 is
`0-1 - consistency over such class of graph families. Examples include
Power-law graphs:
Graphs where the degree sequence follows a power law distribution. We show
? = O(?n) for naturally occurring power law graphs (Claim 4, Suppl.). Thus, working
that ?(G)
? , makes Algorithm 1 consistent.
with the complement graph G
6
?
Random graphs: For G(n, q) graphs, q = O(1); with high probability ?(G(n, q)) = ?( n) [13].
Note that choosing Kn for various graph families is difficult. Alternatively, for Labelled SVM-?
graph family (Definition 4), if Lov?asz ? function is sub-linear, then for the choice of LS labelling,
Algorithm 1 is `0-1 consistent. Examples include Mixture of random graphs (Section 5.1). Furthermore, we analyze the fraction of labelled nodes to be observed, such that Algorithm 1 is consistent.
Corollary 6.2 (Labelled Sample Complexity). Given a graph family Gc , such that ?(Gn ) =
1/3??
n)
O(nc ), ?Gn ? Gc where 0 ? c < 1. For C = C ? as in Theorem 6.1; 21 ?(G
, ?>0
n
fraction of labelled nodes is sufficient for Algorithm 1 to be `0-1 -consistent w.r.t. Gc .
The proof directly follows from Theorem 6.1. As a consequence of the above result, we can argue that for sparse graphs (?(G) is large) one would need a larger fraction of nodes labelled, but
for denser graphs (?(G) is small) even a smaller fraction of nodes being labelled suffices. Such
connections relating sample complexity and graph properties was not available before.
To end this section, we discuss on the possible extensions to the inductive setting (Claim 5, Suppl.)?
we can show that that the uniform convergence of erS? to erS in the transductive setting (for f = 1/2)
is a necessary and sufficient condition for the uniform convergence of erS to the generalization error.
Thus, the results presented here can be extended to the supervised setting. Furthermore, combining
Theorem 5.1 with the results of [9], we can also extend our results to the semi-supervised setting.
7
Multiple Graph Transduction
Many real world problems can be posed as learning on multiple graphs [19, ?]. Existing algorithms
for single graph transduction [10, 15] cannot be trivially extended to the new setting. It is a well
known heuristic that taking a convex combination of Laplacian improves classification performance
[7], however the underlying principle is not well understood. We propose an efficient MKL style
algorithm with generalization guarantees. Formally, the problem of multiple graph transduction is?
Problem 1. Given G = {G(1) , . . . , G(m) } a set of simple, undirected graphs G(k) = V, E (k) ,
defined on a common vertex set V = [n]. Without loss of generality we assume that the first f n
vertices are labelled, i.e., the set of labelled vertices is given by S = [f n], where f ? (0, 1). Let
S? = V \S be the unlabelled node set. Let yS , yS? be the labels of S, S? respectively. Given G and
labels yS , the goal is to accurately predict the labels of yS? .
Let K = {K(1) , . . . , K(m) } be the set of kernels corresponding to graphs G; K(k) ? K(G(k) ), ?k ?
[m]. We propose the following MKL style formulation for multiple graph transduction
X
m
(k)
?
]
(6)
?C (K, yS ) =
min
min
max
g
?,
?
K
,
[y
,
y
?
k
S
S
m
n
?
? ??R ,k?k? ?C
??R+ ,k?k1 =1 y?j ?Y,?j?
S
+
k=1
Extending our analysis from Section 5, we propose the following error bound
Theorem 7.1. For the setting as in Problem 1, let f ? (0, 1/2]7 and K =
? S?? be the solution to ?C (K, yS )
{K(1) , . . . , K(m) }, K(k) ? K(G(k) ), ?k ? [m]. Let ?? , ? ? , y
m
P ? (k) ? ?
? ? Dn , Y
? ii = yi if i ? S, otherwise y?? . Then, for any
? =
(6). Let y
?k K Y? , where Y
i
i=1
? > 0, with probability ? 1 ? ? over the choice of S ? V such that |S| = nf
s
r
?
??)
?(K,
y)
2?(G
c1
1
1
0-1
erS? [?
y] ?
+C
+
log
Cnf
f (1 ? f ) 1 ? f nf
?
?
where c1 = O(1), ?(K,
y) = mink?[m] ?(K(k) , y) and G? is the union of graphs G8 . (Suppl.)
The above result gives us the ability for the first time to analyze generalization performance of multi?
ple graph transduction algorithms. The expression ?(K,
y) suggests that combining multiple graphs
should improve performance over considering individual graphs separately. Similar to Section 6,
7
8
As in Theorem 5.1, we can generalize our results for f ? (0, 1).
G? = (V, E ? ), where (i, j) ? E ? if edge (i, j) is present in atleast one of the graphs G(k) ? G, k ? [m].
7
(l)
we can show that if one of the graph families G (l) , l ? [m] of G obey ?(Gn ) = O(nc ), 0 ? c < 1;
(l)
Gn ? G (l) , then there exists orthonormal representations K, such that the MKL style algorithm
optimizing for (6) is `0-1 -consistent over G (Claim 6, Suppl.). We can also show that combining
graphs improves labelled sample complexity (Claim 7, Suppl.). This is a first attempt in developing
a statistical understanding for the problem of multiple graph transduction.
8
Experimental results
Table 1: Superior performance of LS labelling.
We conduct two sets of experiments9 .
LS-lab Un-Lap N-Lap KS-Lap
Superior performance of LS labelling: We Dataset
?
?
AuralSonar
76.5
68.1
66.7
69.2
use two datasets?similarity matrices from
?
54.1
52.9
53.3
[11] and RBF kernel10 as similarity matrices for Yeast-SW-5-7 ? 60.4
61.2
60.5
64.3
the UCI datasets? [8]. We built an unweighted Yeast-SW-5-12? 78.6
Yeast-SW-7-12
76.5
64.0
59.5
63.1
graph by thresholding the similarity matrices
?
73.1
68.3
68.6
68.5
about the mean. Let L = D ? A. For the reg- Diabetes
?
73.3
69.3
71.2
71.8
ularized formulation (1), with 10% of labelled Fourclass
nodes observable, we test four types of kernel
matrices?LS labelling(LS-lab), (?1 I + L)?1 (Un-Lap), (?2 I + D?1/2 LD?1/2 )?1 (N-Lap) and
K-Scaling (KS-Lap) [4]. We choose the parameters ?, ?1 and ?2 by cross validation. Table 1
summarizes the results. Each entry is accuracy in % w.r.t. 0-1 loss, and the results were averaged
over 100 iterations. Since we are thresholding by mean, the graphs have high connectivity. Thus,
from Corollary 4.3, the function class associated with LS labellingis rich and expressive, and hence
it outperforms previously proposed regularizers.
Graph transduction across Multiple-views: Table 2: Multiple Graphs Transduction.
Learning on mutli-view data has been of recent Each entry is accuracy in %.
interest [18]. Following a similar line of attack, we
pose the problem of classification on multi-view Graph 1vs2 1vs3 1vs4 2vs3 2vs4 3vs4
data as multiple graph transduction. We investigate Aud 62.8 64.8 68.3 59.3 50.8 61.5
68.9 65.6 68.9 69.1 70.3 75.1
the recently launched Google dataset [17], which Vis
68.7 59.2 64.8 64.6 60.9 65.4
contains multiple views of video game YouTube Txt
videos, consisting of 13 feature types of auditory Unn 69.7 60.3 52.7 62.7 67.4 62.5
(Aud), visual (Vis) and textual (Txt) description. Maj 72.7 75.2 80.5 65.4 62.6 77.4
80.6 83.6 86.0 90.9 75.3 91.8
Each video is labelled one of 30 classes. For each Int
of the views we construct similarity matrices using MV 98.9 93.4 95.6 97.7 87.7 98.8
cosine distance and threshold about the mean to obtain
unweighted graphs. We considered 20% of the data to be labelled. We show results on pair-wise
classification for the first four classes. As a natural way of combining graphs, we compared our
algorithm (6) (MV) with union (Unn), intersection (Int) and majority (Maj)11 of graphs. We used
LS labelling as the graph-kernel and (2) was used to solve single graph transduction. Table 2
summarizes the results, averaged over 20 iterations. We also state top accuracy in each of the views
for comparison. As expected from our analysis in Theorem 7.1, we observe that combining multiple
graphs significantly improves classification accuracy.
9
Conclusion
For the problem of graph transduction, we show that there exists orthonormal representations that
are consistent over a large class of graphs. We also note that the Laplacian inverse regularizer
is suboptimal on graphs with high connectivity, and alternatively show that LS labellingis not only
consistent, but also exhibits high Rademacher complexity on a large class of graphs. Using our analysis, we also develop a sound statistical understanding of the improved classification performance
in combining multiple graphs.
9
10
11
Relevant resources at: mllab.csa.iisc.ernet.in/rakeshs/nips14
?kxi ?xj k2
The (i, j)th entry of an RBF kernel is given by exp
. We set ? to the mean distance.
2? 2
Majority graph is a graph where an edge (i, j) is present, if a majority of the graphs have the edge (i, j).
8
References
[1]
V. Jethava, A. Martinsson, C. Bhattacharyya, and D. P. Dubhashi The Lov?asz ? function, SVMs and
finding large dense subgraphs. Neural Information Processing Systems , pages 1169?1177, 2012.
[2]
L. Lov?asz. On the shannon capacity of a graph. IEEE Transactions on Information Theory, 25(1):1?7,
1979.
[3]
R. El-Yaniv and D. Pechyony. Transductive Rademacher complexity and its applications. In Learning
Theory, pages 151?171. Springer, 2007.
[4]
R. Ando, and T. Zhang. Learning on graph with Laplacian regularization. Neural Information Processing
Systems , 2007.
[5]
R. Johnson and T. Zhang. On the effectiveness of Laplacian normalization for graph semi-supervised
learning. Journal of Machine Learning Research, 8(4), 2007.
[6]
R. El-Yaniv and D. Pechyony. Transductive Rademacher complexity and its applications. Journal of
Machine Learning Research, 35(1):193, 2009
[7]
A. Argyriou, M. Herbster, and M. Pontil. Combining graph Laplacians for semi-supervised learning.
Neural Information Processing Systems , 2005.
[8]
A. Asuncion, and D. Newman. UCI machine learning repository. 2000.
[9]
P. L. Bartlett and S. Mendelson. Rademacher and gaussian complexities: Risk bounds and structural
results. Journal of Machine Learning Research, 3:463?482, 2003.
[10] A. Blum and S. Chawla. Learning from labeled and unlabeled data using graph mincuts. International
Conference on Machine Learning, pages 19?26. Morgan Kaufmann Publishers Inc., 2001.
[11] Y. Chen, M. Gupta, and B. Recht. Learning kernels from indefinite similarities. International Conference
on Machine Learning, pages 145?152. ACM, 2009.
[12] C. Cortes, M. Mohri, D. Pechyony and A. Rastogi. Stability Analysis and Learning Bounds for Transductive Regression Algorithms. arXiv preprint arXiv:0904.0814, 2009.
[13] A. Coja-Oghlan. The Lov?asz number of random graphs. Combinatorics, Probability and Computing, 14
(04):439?465, 2005.
[14] T. H. Cormen, C. E. Leiserson, R. L. Rivest, and C. Stein. Introduction to Algorithms. MIT Press, 3rd
edition, 2009.
[15] M. Szummer and T. Jaakkola Partially labeled classification with Markov random walks. Neural Information Processing Systems , 14:945?952, 2002.
[16] C. J. Luz and A. Schrijver. A convex quadratic characterization of the Lov?asz theta number. SIAM Journal
on Discrete Mathematics, 19(2):382?387, 2005.
[17] O. Madani, M. Georg, and D. A. Ross. On using nearly-independent feature families for high precision
and confidence. Machine Learning, 92:457?477, 2013.
[18] W. Tang, Z. Lu, and I. S. Dhillon. Clustering with multiple graphs. International Conference on Data
Mining, pages 1016?1021. IEEE, 2009.
[19] K. Tsuda, H. Shin, and B. Sch?olkopf. Fast protein classification with multiple networks. Bioinformatics,
21(suppl 2):ii59?ii65, 2005.
[20] V. N. Vapnik and A. J. Chervonenkis. On the uniform convergence of relative frequencies of events to
their probabilities Theory of Probability & Its Applications, 16(2):264?280. SIAM, 1971.
[21] D. J. A. Welsh and M. B. Powell. An upper bound for the chromatic number of a graph and its application
to timetabling problems. The Computer Journal, 10(1):85?86, 1967.
[22] T. Zhang and R. Ando. Analysis of spectral kernel design based semi-supervised learning. Neural Information Processing Systems , 18:1601, 2006.
[23] D. Zhou, O. Bousquet, T. N. Lal, J. Weston, and B. Sch?olkopf. Learning with local and global consistency.
Neural Information Processing Systems , 16(16):321?328, 2008.
9
| 5388 |@word h:2 repository:1 stronger:1 c0:2 open:1 hu:3 tr:1 ld:1 contains:1 chervonenkis:1 bhattacharyya:2 outperforms:1 existing:3 com:1 comparing:1 gmail:1 designed:1 drop:1 ith:1 nips14:2 provides:1 characterization:1 node:20 attack:1 zhang:3 dn:4 c2:2 prove:2 introduce:1 lov:11 expected:2 sdp:1 multi:2 little:1 considering:1 iisc:3 notation:3 bounded:3 underlying:1 rivest:1 substantially:1 finding:2 guarantee:2 every:3 nf:7 classifier:2 k2:2 unit:6 yn:3 positive:1 before:3 engineering:1 aggregating:1 understood:1 local:1 consequence:1 establishing:1 path:1 studied:1 k:2 suggests:3 statistically:4 averaged:2 uy:1 practical:1 yj:3 practice:1 regret:5 union:2 shin:1 powell:1 pontil:1 empirical:2 significantly:1 word:1 sider:1 regular:1 confidence:1 protein:1 get:5 cannot:4 unlabeled:2 close:2 context:1 risk:1 equivalent:4 deterministic:1 go:1 l:18 convex:3 independently:1 simplicity:1 pure:3 subgraphs:3 insight:1 orthonormal:33 embedding:7 stability:1 notion:7 limiting:1 diabetes:1 cut:2 labeled:2 observed:3 preprint:1 electrical:1 capture:1 solved:1 connected:1 richness:1 cycle:1 decrease:1 complexity:33 ui:2 motivate:1 tight:2 solving:4 easily:1 various:1 regularizer:2 enyi:1 fast:1 effective:1 newman:1 choosing:3 quite:1 heuristic:3 widely:2 supplementary:1 whose:1 larger:1 ramp:5 otherwise:2 denser:1 posed:1 ability:1 jethava:1 solve:1 transductive:8 kik:1 prsn:1 sequence:2 eigenvalue:1 propose:5 uci:2 combining:8 relevant:1 argmink:1 subgraph:7 achieve:2 description:1 normalize:1 olkopf:2 convergence:7 yaniv:2 requirement:1 extending:2 rademacher:17 converges:1 help:2 derive:11 develop:2 illustrate:1 pose:1 fourclass:1 concentrate:1 aud:2 sgn:1 dii:1 adjacency:2 argued:1 kii:1 suffices:1 generalization:11 preliminary:1 extension:1 considered:3 exp:1 predict:1 claim:7 g8:1 achieves:1 label:24 combinatorial:1 ross:1 largest:1 tool:3 mit:1 gaussian:1 pn:1 zhou:1 chromatic:4 jaakkola:1 corollary:5 indicates:1 dependent:9 el:2 issue:1 dual:1 classification:7 rsn:2 special:1 ernet:3 aware:1 construct:1 nearly:1 representer:1 bangalore:2 duplicate:1 randomly:1 individual:3 madani:1 consisting:1 welsh:1 n1:4 sandwich:1 ab:3 attempt:2 ando:2 interest:2 investigate:2 mining:1 leiserson:1 mixture:2 semidefinite:1 regularizers:3 kjk:1 edge:6 necessary:1 tree:1 conduct:1 walk:1 tsuda:1 minimal:1 kij:2 soft:2 earlier:1 gn:19 unn:2 vertex:14 entry:4 uniform:4 johnson:1 kn:7 learnt:2 kxi:1 recht:1 density:1 fundamental:1 herbster:1 international:3 siam:2 sequel:1 connectivity:5 choose:2 style:4 return:2 star:1 int:2 inc:1 combinatorics:1 explicitly:3 mv:2 ad:1 vi:5 view:7 lab:16 vs2:1 observing:1 analyze:4 asuncion:1 defer:1 contribution:3 square:2 accuracy:5 kaufmann:1 rastogi:1 generalize:2 famous:1 accurately:1 lu:1 pechyony:3 whenever:2 definition:11 vs3:2 frequency:1 naturally:1 associated:8 mi:1 di:3 con:1 proof:5 auditory:1 dataset:2 knowledge:3 improves:3 formalize:2 oghlan:1 back:1 supervised:5 improved:2 erd:2 yb:1 formulation:7 luz:1 though:1 generality:2 furthermore:2 implicit:1 working:4 auralsonar:1 ei:2 expressive:1 o:2 mkl:4 google:1 quality:1 yeast:3 normalized:1 true:4 contain:1 inductive:2 regularization:3 hence:5 symmetric:2 dhillon:1 game:1 noted:1 mutli:1 cosine:1 pdf:1 multiview:1 complete:1 demonstrate:2 wise:1 novel:7 recently:2 superior:3 common:1 extend:3 belong:1 discussed:1 relating:5 martinsson:1 refer:1 rd:2 consistency:16 trivially:1 mathematics:1 pq:2 similarity:5 showed:6 recent:1 optimizing:2 inf:2 inequality:1 binary:5 yi:10 morgan:1 greater:1 semi:4 relates:1 multiple:20 sound:2 ii:1 unlabelled:4 cross:1 sphere:8 e1:1 y:13 plugging:1 laplacian:9 prediction:6 regression:1 txt:2 arxiv:2 iteration:2 kernel:20 normalization:2 suppl:13 achieved:1 c1:8 whereas:1 want:4 separately:1 publisher:1 appropriately:1 launched:1 sch:2 asz:11 induced:2 tend:1 undirected:5 effectiveness:1 structural:6 embeddings:3 easy:1 reacts:1 erv:1 xj:1 suboptimal:2 opposite:1 idea:1 knowing:2 motivated:2 expression:1 bartlett:1 cnf:2 useful:2 maybe:1 stein:1 extensively:1 induces:1 svms:1 exist:1 discrete:1 georg:1 four:2 indefinite:1 threshold:1 blum:1 drawn:2 v1:1 graph:163 fraction:7 inverse:4 extends:1 family:13 uti:1 vn:7 p3:1 draw:2 summarizes:2 scaling:1 bound:20 quadratic:1 oracle:1 bousquet:1 u1:1 argument:2 min:4 optimality:1 department:2 developing:1 combination:2 mllab:2 cormen:1 beneficial:1 smaller:1 across:1 making:1 kmm:7 chiranjib:1 resource:1 remains:1 previously:4 turn:1 slack:1 discus:1 hh:1 needed:1 end:1 available:3 apply:3 obey:1 observe:1 spectral:1 chawla:1 ers0:5 denotes:1 top:1 include:3 clustering:1 hinge:3 sw:3 exploit:1 k1:1 uj:1 prof:1 dubhashi:1 quantity:2 diagonal:2 said:2 exhibit:2 amongst:1 distance:2 capacity:1 majority:3 argue:1 minn:2 relationship:1 nc:3 difficult:1 relate:3 negative:1 mink:2 design:1 unknown:1 coja:1 upper:3 observation:1 datasets:3 markov:1 finite:2 orthant:1 extended:2 rn:10 gc:5 community:1 introduced:4 complement:3 pair:1 connection:6 lal:1 textual:1 usually:1 laplacians:2 built:1 including:2 max:7 video:3 power:5 event:1 natural:1 improve:2 theta:1 maj:2 sn:13 understanding:3 literature:2 geometric:1 asymptotic:1 law:5 relative:1 loss:9 fully:1 interesting:4 suph:1 erh:1 validation:1 degree:4 gather:1 sufficient:3 consistent:16 principle:1 thresholding:2 atleast:3 row:1 mohri:1 last:1 aij:1 allow:1 india:2 institute:2 taking:1 sparse:2 world:2 valid:1 rich:2 unweighted:2 made:1 ple:1 transaction:1 observable:3 confirm:1 global:1 alternatively:3 un:6 table:4 learn:1 unavailable:2 csa:4 ysn:4 main:2 dense:1 rh:1 motivation:1 big:1 bounding:1 edition:1 n2:2 cient:1 en:2 transduction:25 aid:1 lc:1 precision:1 sub:1 explicit:1 tang:1 theorem:15 kuk2:1 er:22 svm:3 gupta:1 cortes:1 exists:6 mendelson:1 vapnik:1 labelling:13 occurring:1 margin:11 chen:1 suited:1 tc:2 intersection:1 lap:6 explore:1 visual:1 expressed:1 partially:1 trp:2 springer:1 satisfies:1 kls:3 acm:1 weston:1 chiru:1 goal:2 identity:1 rbf:2 labelled:29 unrepresentative:2 lipschitz:1 youtube:1 specifically:1 infinite:1 vs4:3 mincuts:1 e:3 experimental:2 schrijver:1 rakesh:1 shannon:1 indicating:1 formally:4 support:1 szummer:1 ularized:1 bioinformatics:1 indian:2 reg:1 argyriou:1 |
4,847 | 5,389 | Optimal prior-dependent neural population codes
under shared input noise
?
Agnieszka Grabska-Barwinska
Gatsby Computational Neuroscience Unit
University College London
[email protected]
Jonathan W. Pillow
Princeton Neuroscience Institute
Department of Psychology
Princeton University
[email protected]
Abstract
The brain uses population codes to form distributed, noise-tolerant representations of sensory and motor variables. Recent work has examined the theoretical
optimality of such codes in order to gain insight into the principles governing
population codes found in the brain. However, the majority of the population
coding literature considers either conditionally independent neurons or neurons
with noise governed by a stimulus-independent covariance matrix. Here we analyze population coding under a simple alternative model in which latent ?input
noise? corrupts the stimulus before it is encoded by the population. This provides
a convenient and tractable description for irreducible uncertainty that cannot be
overcome by adding neurons, and induces stimulus-dependent correlations that
mimic certain aspects of the correlations observed in real populations. We examine prior-dependent, Bayesian optimal coding in such populations using exact
analyses of cases in which the posterior is approximately Gaussian. These analyses extend previous results on independent Poisson population codes and yield
an analytic expression for squared loss and a tight upper bound for mutual information. We show that, for homogeneous populations that tile the input domain,
optimal tuning curve width depends on the prior, the loss function, the resource
constraint, and the amount of input noise. This framework provides a practical
testbed for examining issues of optimality, noise, correlation, and coding fidelity
in realistic neural populations.
1
Introduction
A substantial body of work has examined the optimality of neural population codes [1?19]. However, the classical literature has focused mostly on codes with independent Poisson noise, and on
Fisher information-based analyses of unbiased decoding. Real neurons, by contrast, exhibit dependencies beyond those induced by the stimulus (i.e., ?noise correlations?), and Fisher information
does not accurately quantify information when performance is close to threshold [7, 15, 18], or
when biased decoding is optimal. Moreover, the classical population codes with independent Poisson noise predict unreasonably good performance with even a small number of neurons. A variety
of studies have shown that the information extracted from independently recorded neurons (across
trials or even animals) outperforms the animal itself [20, 21]. For example, a population of only 220
Poisson neurons with tuning width of 60 deg (full width at half height) and tuning amplitude of 10
spikes can match the human orientation discrimination threshold of ? 1 deg. (See Supplement S1
for derivation.) Note that even fewer neurons would be required if peak spike counts were higher.
The mismatch between this predicted efficiency and animals? actual behaviour has been attributed
to the presence of information-limiting correlations between neurons [22, 23]. However, deviation
1
stimulus prior
0
stimulus
stimulus
posterior
p(stimulus)
spike count
p(stimulus)
population response
Poisson
noise
spike count
tuning curves
input noise
preferred stimulus
likelihood
stimulus
+
Figure 1: Bayesian formulation of neural population coding with input noise.
from independence renders most analytical treatments infeasible, necessitating the use of numerical
methods (Monte Carlo simulations) for quantifying the performance of such codes [7, 15].
Here we examine a family of population codes for which the posterior is Gaussian, which makes it
feasible to perform a variety of analytical treatments. In particular, when tuning curves are Gaussian
and ?tile? the input domain, we obtain codes for which the likelihood is proportional to a Gaussian
[2, 16]. Combined with a Gaussian stimulus prior, this results in a Gaussian posterior whose variance
depends only on the total spike count. This allows us to derive tractable expressions for neurometric
functions such as mean squared error (MSE) and mutual information (MI), and to analyze optimality
without resorting to Fisher information, which can be inaccurate for short time windows or small
spike counts [7, 15, 18]. Secondly, we extend this framework to incorporate shared ?input noise? in
the stimulus variable of interest (See Fig. 1). This form of noise differs from many existing models,
which assume noise to arise from shared connectivity, but with no direct relationship to the stimulus
coding [5, 15, 18, 24] (although see [16, 25] for related approaches).
This paper is organised as follows. In Sec. 2, we describe an idealized Poisson population code with
tractable posteriors, and review classical results based on Fisher Information. In Sec. 3, we provide
a Bayesian treatment of these codes, deriving expressions for mean squared error (MSE) and mutual
information (MI). In Sec. 4, we extend these analyses to a population with input noise. Finally, in
Sec. 5 we examine the patterns of pairwise dependencies introduced by input noise.
2
Independent Poisson population codes
Consider an idealized population of Poisson neurons that encode a scalar stimulus s with Gaussianshaped tuning curves. Under this (classical) model, the response vector r = (r1 , . . . rN )> is conditionally Poisson distributed:
ri |s ? Poiss(fi (s)),
p(r|s) =
N
Y
ri
fi (s)
1
,
ri ! fi (s) e
(Poisson encoding) (1)
i=1
where tuning curves fi (s) take the form
?
fi (s) = ? A exp
1
2
?
?
s i )2 ,
2 (s
t
?
(tuning curves) (2)
with equally-spaced preferred stimuli s = ( s 1 , . . . s N ), tuning width t , amplitude A, and time
window for counting spikes ? . We assume that the tuning curves ?tile?, i.e., sum to a constant over
the relevant stimulus range:
N
X
?
?
(tiling property) (3)
fi (s) =
i=1
We
determine by integrating
the summed tuning curves (eq. 3) over the stimulus space, giving
p
R can
PN
ds i=1 fi (s) = N A 2? t = S , with solution:
(expected total spike count) (4)
p
where
= S/N is the spacing between tuning curve centers, and a = 2?A? is a constant that
we will refer to as the ?effective amplitude?, since it depends on true tuning curve amplitude and
= a t/
2
the time window for integrating spikes. Note, that tiling holds almost perfectly if tuning curves are
broad compared to their spacing (e.g. t > ). However, our results hold on average for a much
broader range of t . (See Supplementary Figs S2 and S3 for a numerical analysis.)
P
Let R =
ri denote the total spike count from the entire population. Due to tiling, R is a Poisson
1 R
random variable with rate , regardless of the stimulus: p(R|s) = R!
e .
For simplicity, we will consider stimuli drawn from a zero-mean Gaussian prior with variance
Q
s ? N (0,
2
s ),
p(s) =
p 1
2?
s
e
s2
2
2 s
.
2
s:
(stimulus prior) (5)
Since i e fi (s) = e due to the tiling assumption, the likelihood (eq. 1 as a function of s) and
posterior both take Gaussian forms:
Y
?
p(r|s) /
fi (s)ri / N s R1 r >s, R1 t2
(likelihood) (6)
i
p(s|r) = N
? r >s?
2 ?
t
,
,
R+? R+?
(posterior) (7)
where ? = t2 / s2 denotes the ratio of the tuning curve variance to prior variance. The maximum of
?
the likelihood (eq. 6) is the so-called center-of-mass estimator estimator, R1 r >s, while the mean of
the posteror (eq. 7) is biased toward zero by an amount that depends on ?. Note that the posterior
variance does not depend on which neurons emitted spikes, only the total spike count R, a fact that
will be important for our analyses below.
2.1
Capacity constraints for defining optimality
Defining optimality for a population code requires some form of constraint on the capacity of the
neural population, since clearly we can achieve arbitrarily narrow posteriors if we allow arbitrarily
large total spike count R. In the following, we will consider two different biologically plausible
constraints:
? A space constraint, in which we constrain only the number of neurons. This means that
increasing the tuning width t will increase the expected population spike count (see
eq. 4), since more neurons will respond as tuning curves grow wider.
? An energy constraint, in which we fix while allowing t and amplitude A to vary. Here,
we can make tuning curves wider but must reduce the amplitude so that total expected spike
count remains fixed.
We will show that the optimal tuning depends strongly on which kind of constraint we apply.
2.2
Analyses based on Fisher Information
The Fisher information provides a popular, tractable metric for quantifying the efficiency of a neural
@2
code, given by E[ @s
2 log p(r|s)], where expectation is taken with respect to encoding distribution
p(r|s). For our idealized Poisson population, the total Fisher information is:
N
N
?
? (s s? )2 ?
X
X
fi0 (s)2
(s s i )2
a
i
IF (s) =
=
A
exp
=
= 2 , (Fisher info) (8)
4
2
f
(s)
2
i
t
t
t
t
i=1
i=1
which we can derive, as before, using the tiling property (eq. 3). (See also Supplemental Sec. S2).
The first of the two expressions at right reflects IF for the space constraint, where varies implicitly
as we vary t . The second expresses IF under the energy constraint, where is constant so that a
varies implicitly with t . For both constraints, IF increases with increasing a and decreasing t [5].
Fisher information provides a well-known bound on the variance of an unbiased estimator s?(r)
known as the Cram?er-Rao (CR) bound, namely var(?
s|s) 1/IF (s). Since FI is constant over s in
our idealized setting, this leads to a bound on the mean squared error ([7, 12]):
?
2
?
?
1
t
MSE , E (?
s(r) s)2 p(r,s) E
=
= t,
(9)
IF (s) p(s)
a
3
effects of prior stdev
effects of time window (ms)
3
MSE
0
10 50
0
10
CR bound
200
10
CR bound
1
10
MSE
energy
constraint
= 25
400
10
2
10
2
1
10
16
8
4
2
1
=3
space
constraint
10
0
10
10
10
10
0
2
4
tuning width
6
8
10
0
2
4
tuning width
6
8
Figure 2: Mean squared error as a function of the tuning width t , under space constraint (top row)
and energy constraint (bottom row), for spacing
= 1 and amplitude A = 20 sp/s. and Top
left: MSE for different prior widths s (with A=2,? = 200ms), showing that optimal t increases
with larger prior variance. Cram?er-Rao bound (gray solid) is minimized at t = 0, whereas bound
(eq. 12, gray dashed) accurately captures shape and location of the minimum. Top right: Similar
curves for different time windows ? for counting spikes (with s =32), showing that optimal t increases for lower spike counts. p
Bottom row: Similar traces under energy constraint (where A scales
inversely with t so that = 2?? A t is constant). Although the CR bound grossly understates
the true MSE for small counting windows (right), the optimal tuning is maximally narrow in this
configuration, consistent with the CR curve.
which is simply the inverse of Fisher Information (eq. 8).
Fisher information also provides a (quasi) lower bound on the mutual information, since an efficient
estimator (i.e., one that achieves the CR bound) has entropy upper-bounded by that of a Gaussian
with variance 1/IF (see [3]). In our setting this leads to the lower bound:
?
?
?
?
2 a
2
1
1
MI(s, r) , H(s) H(s|r)
log
=
log
(10)
s
s
2 .
2
2
t
t
Note that neither of these FI-based bounds apply exactly to the Bayesian setting we consider here,
since Bayesian estimators are generally biased, and are inefficient in the regime of low spike counts
[7]. We examine them here nonetheless (gray traces in Figs. 2 and 3) due to their prominence in the
prior literature ([5, 12, 14]), and to emphasize their limitations for characterizing optimal codes.
2.3
Exact Bayesian analyses
In our idealized population, the total spike count R is a Poisson random variable with mean , which
allows us to compute the MSE and MI by taking expectations w.r.t. this distribution.
Mean Squared Error (MSE)
The mean squared error, which equals the average posterior variance (eq. 7), can be computed
analytically for this model:
? 2
? R
1 ?
2
X
t
t
MSE = E
=
e = t2 e
(?) ? (?,
),
(11)
R + ? p(R)
R + ? R!
R=0
Rz a 1 t
1
where ? = t2 / s2 and ? (a, z) = z a (a)
t
e dt is the holomorphic extension of the lower
0
incomplete gamma function [26] (see SI for derivation). When the tuning curve is narrower than the
prior (i.e., t2 ? s2 ), we can obtain a relatively tight lower bound:
MSE
2
t
1
e
4
+(
2
s
2
t )e
.
(12)
effects of prior stdev
MI (bits)
space
constraint
6
effects of time window (ms)
6
FI-based bound
4
16
8
4
2
1
MI (bits)
0
energy
constraint
4
= 32
2
= 400
0
6
6
4
4
2
2
0
= 25
2
0
2
4
tuning width
6
8
2
4
tuning width
6
8
Figure 3: Mutual information as a function of tuning width t , directly analogous to plots in Fig. 2.
Note the problems with the lower bound on MI derived from Fisher information (top, gray traces)
and the close match of the derived bound (eq. 14, dashed gray traces). The effects are similar
to Fig. 2, except that MI-optimal tuning widths are slightly smaller (upper left and right) than for
MSE-optimal codes. For both loss functions, optimal width is minimal under an energy constraint.
Figure 2 shows the MSE (and derived bound) as a function of the tuning width t over the range
where tiling approximately holds. Note the high accuracy of the approximate formula (12, dashed
gray traces) and that the FI-based bound does not actually lower-bound the MSE in the case of
narrow priors (darker traces).
For the space-constrained setting (top row, obtained by substituting = a t / in eqs. 11 and
12), we observe substantial discrepancies between the true MSE and FI-based analysis. While FI
suggests that optimal tuning width is near zero (down to the limits of tiling), analyses reveal that the
optimal t grows with prior variance (left) and decreasing time window (right). These observations
agree well with the existing literature (e.g. [15, 16]). However, if we restrict the average population
firing rate (energy constraint, bottom plots), the optimal tuning curves once again approach zero. In
this case, FI provides correct intuitions and better approximation of the true MSE.
Mutual Information (MI)
For a tiling population and Gaussian prior, mutual information between the stimulus and response
is:
h ?
?i
2
MI(s, r) = 12 E log 1 + R s2
,
(13)
t
P (R)
which has no closed-form solution, but can
p be calculated efficiently with a discrete sum over R from
0 to some large integer (e.g., R = + n
to capture n standard deviations above the mean). We
can derive an upper bound using the Taylor expansion to log while preserving the exact zeroth order
term:
?
?
?
?
2
2
a t/
MI(s, r) ? 1 2e log 1 + ( 1 e ) s2 = 1 e 2
log 1 + 1 e aa t / ts
(14)
t
Once again, we investigate the efficiency of population coding for neurons, now in terms of the
maximal MI. Figure 3 shows MI as a function of the neural tuning width t . We observe a similar
effect as for the MSE: the optimal tuning widths are now different from zero,but only for the space
constraint. The energy constraint, as well as implications from FI indicate optimum near t =0.
5
3
Poisson population coding with input noise
We can obtain a more general family of correlated population codes by considering ?input noise?,
where the stimulus s is corrupted by an additive noise n (see Fig. 1):
s ? N (0,
2
s)
2
n)
(prior) (15)
n ? N (0,
ri |s, n ? Poiss(fi (s + n))
(input noise) (16)
(population response) (17)
The use of Gaussians allows us to marginalise over n analytically, resulting in a Gaussian form for
the likelihood and Gaussian posterior:
p(r|s) / N s R1 r >s, R1 t2 + n2
?
?
r >s
( t2 + R n2 )
p(s|r) = N
2 / 2 + R( 2 / 2 + 1) , 2 + R( 2 +
t
t
s
n
s
n
?
2
s
2
s)
?
(likelihood) (18)
(posterior) (19)
Note that even in the limit of large spike counts, the posterior variance is non-zero, converging to
2 2
2
2
n s /( n + s ).
3.1
Population coding characteristics: FI, MSE, & MI
Fisher information for a population with input noise can be using the fact that the likelihood (eq. 18)
is Gaussian: Eq. (18):
? 2
?
e
d log p(r|s)
R
=
E
=
(1 + ?) ? (1 + ?,
) (20)
IF (s) , E
2
2
2
2
ds
+
R
t
n
n
p(r|s)
p(R)
where ? = t2 / n2 and ? (?, ?) once again denotes holomorphic extension of lower incomplete
gamma function. Note that for n = 0, this reduces to (eq. 8).
It is straightforward to employ the results from Sec. 2.3 for the exact Bayes analyses of a Gaussian
posterior (19):
?
?
?
2
2
2 2
1
R
t +R n
2
2
s n
MSE = s E 2
= s? E
+ 2+ 2 E
2 + 2)
s
n
+
R(
?
+
R
?
+
R p(R)
t
n
s
p(R)
p(R)
2
?
?
= ? (?) ? (?,
) + 2 +n 2 (1 + ?) ? (1 + ?,
) s2 e ,
(21)
s
?
?
MI = 12 E log 1 +
R s2
2
t +R
n
2
n
?
(22)
,
p(R)
where ? = t2 /( s2 + n2 ). Although we could not determine closed-form analytical expressions
for MI, it can be computed efficiently by summing over a range of integers [0, . . . Rmax ] for which
P (R) has sufficient support. Note this is still a much faster procedure than estimating these values
from Monte Carlo simulations.
3.2
Optimal tuning width under input noise
Fig. 4 shows the optimal tuning width under the space constraint: the value of t minimizing MSE
(left) or maximising MI (right) as a function of the prior width s , for selected time windows of
integration ? . Blue traces show results for a Poisson population, while green traces correspond to a
population with input noise ( n = 1).
For both MSE and MI loss functions, optimal tuning width decreases for narrower priors. However,
under input noise (green traces), the optimal tuning width saturates at the value that depends on
the available number of spikes. As the prior grows wider, the growth of the optimal tuning width
depends strongly on the choice of loss function: optimal t grows approximately logarithmically
with s for minimizing MSE (left), but it grows much slower for maximizing MI (right). Note that
for realistic prior widths (i.e. for s > n ), the effects of input noise on optimal tuning width are far
more substantial under MI than under MSE.
6
mutual information
MSE
optimal TC width
8
8
6
Poisson
noise only
4
2
w/ input noise
=
0
20
6
00
=1
4
= 50
2
= 25
0
0.1
1
prior stdev
0
10
0.1
1
prior stdev
10
Figure 4: Optimal tuning width t (under space constraint only) as a function of prior width s ,
for classic Poisson populations (blue) and populations with input-noise (green, n2 = 1). Different
traces correspond to different time windows of integration, for
= 1 and A = 20 sp/s. As n
increases, the optimal tuning width increases under MI, and under MSE when s < n (traces not
shown). For MSE, predictions of the Poisson and input-noise model converge for priors s > n .
We have not shown plots for energy-constrained population codes because the optimal tuning width
sits at the minimum of the range over which tiling can be said to hold, regardless of prior width,
input noise level, time window, or choice of loss function. This can be seen easily in the expressions
for MI (eqs. 13 and 22), in which each term in the expectation is a decreasing function of t for
all R > 0. This suggests that, contrary to some recent arguments (e.g., [15, 16]), narrow tuning (at
least down to the limit of tiling) really is best if the brain has a fixed energetic budget for spiking, as
opposed to a mere constraint on the number of neurons.
4
Correlations induced by input noise
Input noise alters the mean, variance, and pairwise correlations of population responses in a systematic manner that we can compute directly (see Supplement for derivations). In Fig. 5 we show
the effects input noise with standard deviation n = 0.5 , for neurons with the tuning amplitude
of A = 10. The tuning curve (mean response) becomes slightly flatter (A), while the variance increases, especially at the flanks (B). Fig. 5C shows correlations between the two neurons with tuning
curves and variance are shown in panels A-B: one pair with the same preferred orientation at zero
(red) and a second with a 2 degree difference in preferred orientation (blue). From these plots, it is
clear that the correlation structure depends on both the tuning as well as the stimulus. Thus, in order
to describe such correlations one needs to consider the entire stimulus range, not simply the average
correlation marginalized over stimuli.
Figure 5D shows the pairwise correlations across an entire population of 21 neurons given a stimulus
at s = 0. Although we assumed Gaussian tuning curves here, one obtain similar plots for arbitrary
unimodal tuning curves (see Supplement), which should make it feasible to test our predictions
in real data. However, the time scale of the input noise and basic neural computations is about
10 ms. At such short spike count windows, available number of spikes is low, and so are correlations
induced by input noise. With other sources of second order statistics, such as common input gains
(e.g. by contrast or adaptation), these correlations might be too subtle to recover [23].
5
Discussion
We derived exact expressions for mean squared error and mutual information in a Bayesian analysis
of: (1) an idealized Poisson population coding model; and (2) a correlated, conditionally Poisson
population coding model with shared input noise. These expressions allowed us to examine the
optimal tuning curve width under both loss functions, under two kinds of resource constraints. We
have confirmed that optimal t diverges from predictions based on Fisher information, if the overall
spike count allowed is allowed to grow with tuning width (i.e., because more neurons respond to
the stimulus when tuning curves become broader). We referred to this as a ?space constraint? to
differentiate it from an ?energy constraint?, in which tuning curve amplitude scales down with tuning
7
B
mean
variance
(sp / s)2
sp / s
A
5
5
stimulus s
stimulus s
C
D
r
r
preferred stim
correlation
5
stimulus s
preferred stim
Figure 5: Response statistics of neural population with input noise, for standard deviation n = 0.5.
?
?
(A) Expected spike responses of two neurons: s 1 = 0 (red) and s 2 = 2 (blue). The common
noise effectively smooths blurs the tuning curves with a Gaussian kernel of width n . (B) Variance
of neuron 1, its tuning curve replotted in black for reference. Input noise has largest influence on
variance at the steepest parts of the tuning curve. (C) Cross-correlation of the neuron 1 with two
?
others: one sharing the same preference (red), and one with s = 2 (blue). Note that correlation
of two identically tuned neurons is largest at the steepest part of the tuning curve. (D) Spike count
correlations for entire population of 21 neurons given a fixed stimulus s = 0, illustrating that the
pattern of correlations is signal dependent.
width so that average total spike count is invariant to tuning width. In this latter scenario, predictions
from Fisher information are no longer inaccurate, and we find that optimal tuning width should be
narrow (down to the limit at which the tiling assumption applies), regardless of the duration, prior
width, or input noise level.
We also derived explicit predictions for the dependencies (i.e., response correlations) induced by
the input noise. These depend on the shape (and scale) of tuning responses, and on the amount of
noise ( n ). However, for a reasonable assumption that noise distribution is much narrower than the
width of the prior (and tuning curves), under which the mean firing rate changes little, we can derive
predictions for the covariances directly from the measured tuning curves. An important direction
for future work will be to examine the detailed structure of correlations measured in large populations. We feel that the input noise model ? which describes exactly those correlations that are most
harmful for decoding ? has the potential to shed light on the factors affecting the coding capacity
in optimal neural populations [23].
Finally, if we return to our example from the Introduction to see how the number of neurons necessary to reach the human discrimination threshold of s=1 degree changes in the presence of input
noise. As n approaches s, the number of neurons required goes rapidly to infinity (See Supplementary Fig. S1).
Acknowledgments
This work was supported by the McKnight Foundation (JP), NSF CAREER Award IIS-1150186
(JP), NIMH grant MH099611 (JP) and the Gatsby Charitable Foundation (AGB).
References
[1] HS Seung and H. Sompolinsky. Simple models for reading neuronal population codes. Proceedings of
the National Academy of Sciences, 90(22):10749?10753, 1993.
[2] R. S. Zemel, P. Dayan, and A. Pouget. Probabilistic interpretation of population codes. Neural Comput,
10(2):403?430, Feb 1998.
8
[3] Nicolas Brunel and Jean-Pierre Nadal. Mutual information, fisher information, and population coding.
Neural Computation, 10(7):1731?1757, 1998.
[4] Kechen Zhang and Terrence J. Sejnowski. Neuronal tuning: To sharpen or broaden? Neural Computation,
11(1):75?84, 1999.
[5] A. Pouget, S. Deneve, J. Ducom, and P. E. Latham. Narrow versus wide tuning curves: What?s best for a
population code? Neural Computation, 11(1):85?90, 1999.
[6] Alexandre Pouget, Sophie Deneve, and Peter E Latham. The relevance of fisher information for theories
of cortical computation and attention. Visual attention and cortical circuits, pages 265?284, 2001.
[7] M. Bethge, D. Rotermund, and K. Pawelzik. Optimal short-term population coding: When fisher information fails. Neural computation, 14(10):2317?2351, 2002.
[8] W. J. Ma, J. M. Beck, P. E. Latham, and A. Pouget. Bayesian inference with probabilistic population
codes. Nature Neuroscience, 9:1432?1438, 2006.
[9] Marcelo A Montemurro and Stefano Panzeri. Optimal tuning widths in population coding of periodic
variables. Neural computation, 18(7):1555?1576, 2006.
[10] P. Seri`es, A. A. Stocker, and E. P. Simoncelli. Is the homunculus ?aware? of sensory adaptation? Neural
Computation, 21(12):3271?3304, Dec 2009.
[11] R. Haefner and M. Bethge. Evaluating neuronal codes for inference using fisher information. Neural
Information Processing Systems, 2010.
[12] D. Ganguli and E. P. Simoncelli. Implicit encoding of prior probabilities in optimal neural populations.
In J. Lafferty, C. Williams, R. Zemel, J. Shawe-Taylor, and A. Culotta, editors, Adv. Neural Information
Processing Systems, volume 23, Cambridge, MA, May 2010. MIT Press.
[13] Xue-Xin Wei and Alan Stocker. Efficient coding provides a direct link between prior and likelihood in perceptual bayesian inference. In P. Bartlett, F.C.N. Pereira, C.J.C. Burges, L. Bottou, and K.Q. Weinberger,
editors, Advances in Neural Information Processing Systems 25, pages 1313?1321, 2012.
[14] Z Wang, A Stocker, and A Lee. Optimal neural tuning curves for arbitrary stimulus distributions: Discrimax, infomax and minimum lp loss. In Advances in Neural Information Processing Systems 25, pages
2177?2185, 2012.
[15] P. Berens, A.S. Ecker, S. Gerwinn, A.S. Tolias, and M. Bethge. Reassessing optimal neural population
codes with neurometric functions. Proceedings of the National Academy of Sciences, 108(11):4423, 2011.
[16] Steve Yaeli and Ron Meir. Error-based analysis of optimal tuning functions explains phenomena observed
in sensory neurons. Frontiers in computational neuroscience, 4, 2010.
[17] Jeffrey M Beck, Peter E Latham, and Alexandre Pouget. Marginalization in neural circuits with divisive
normalization. J Neurosci, 31(43):15310?15319, Oct 2011.
[18] Stuart Yarrow, Edward Challis, and Peggy Seri`es. Fisher and shannon information in finite neural populations. Neural Computation, 24(7):1740?1780, 2012.
[19] D Ganguli and E P Simoncelli. Efficient sensory encoding and Bayesian inference with heterogeneous
neural populations. Neural Computation, 26(10):2103?2134, Oct 2014. Published online: 24 July 2014.
[20] E. Zohary, M. N. Shadlen, and W. T. Newsome. Correlated neuronal discharge rate and its implications
for psychophysical performance. Nature, 370(6485):140?143, Jul 1994.
[21] Keiji Miura, Zachary Mainen, and Naoshige Uchida. Odor representations in olfactory cortex: distributed
rate coding and decorrelated population activity. Neuron, 74(6):1087?1098, 2012.
[22] Jakob H. Macke, Manfred Opper, and Matthias Bethge. Common input explains higher-order correlations
and entropy in a simple model of neural population activity. Phys. Rev. Lett., 106(20):208102, May 2011.
[23] Rub?en Moreno-Bote, Jeffrey Beck, Ingmar Kanitscheider, Xaq Pitkow, Peter Latham, and Alexandre
Pouget. Information-limiting correlations. Nat Neurosci, 17(10):1410?1417, Oct 2014.
[24] K. Josic, E. Shea-Brown, B. Doiron, and J. de la Rocha. Stimulus-dependent correlations and population
codes. Neural Comput, 21(10):2774?2804, Oct 2009.
[25] G. Dehaene, J. Beck, and A. Pouget. Optimal population codes with limited input information have finite
tuning-curve widths. In CoSyNe, Salt Lake City, Utah, February 2013.
[26] NIST Digital Library of Mathematical Functions. http://dlmf.nist.gov/, Release 1.0.9 of 2014-08-29.
[27] David C Burr and Sally-Ann Wijesundra. Orientation discrimination depends on spatial frequency. Vision
Research, 31(7):1449?1452, 1991.
[28] Russell L De Valois, E William Yund, and Norva Hepler. The orientation and direction selectivity of cells
in macaque visual cortex. Vision research, 22(5):531?544, 1982.
9
| 5389 |@word h:1 trial:1 illustrating:1 simulation:2 covariance:2 prominence:1 solid:1 configuration:1 valois:1 mainen:1 tuned:1 outperforms:1 existing:2 si:1 must:1 numerical:2 additive:1 realistic:2 blur:1 shape:2 analytic:1 motor:1 moreno:1 plot:5 discrimination:3 half:1 fewer:1 selected:1 steepest:2 short:3 manfred:1 provides:7 location:1 sits:1 preference:1 ron:1 zhang:1 height:1 mathematical:1 direct:2 become:1 burr:1 olfactory:1 manner:1 pairwise:3 expected:4 montemurro:1 examine:6 brain:3 decreasing:3 gov:1 actual:1 little:1 window:12 zohary:1 considering:1 increasing:2 becomes:1 pawelzik:1 estimating:1 moreover:1 bounded:1 panel:1 mass:1 circuit:2 what:1 grabska:1 kind:2 rmax:1 nadal:1 supplemental:1 growth:1 shed:1 exactly:2 uk:1 unit:1 grant:1 before:2 limit:4 encoding:4 firing:2 approximately:3 might:1 zeroth:1 black:1 reassessing:1 examined:2 suggests:2 limited:1 range:6 challis:1 practical:1 acknowledgment:1 differs:1 procedure:1 holomorphic:2 convenient:1 integrating:2 cram:2 cannot:1 close:2 influence:1 ecker:1 center:2 maximizing:1 straightforward:1 regardless:3 go:1 independently:1 duration:1 focused:1 attention:2 williams:1 simplicity:1 pitkow:1 pouget:7 insight:1 estimator:5 deriving:1 rocha:1 population:63 classic:1 analogous:1 limiting:2 feel:1 discharge:1 exact:5 homogeneous:1 us:1 logarithmically:1 observed:2 bottom:3 wang:1 capture:2 culotta:1 adv:1 sompolinsky:1 decrease:1 russell:1 substantial:3 intuition:1 nimh:1 seung:1 depend:2 tight:2 efficiency:3 easily:1 stdev:4 derivation:3 describe:2 london:1 monte:2 effective:1 sejnowski:1 zemel:2 seri:2 whose:1 encoded:1 supplementary:2 plausible:1 larger:1 jean:1 statistic:2 josic:1 naoshige:1 itself:1 online:1 differentiate:1 analytical:3 matthias:1 ucl:1 maximal:1 adaptation:2 relevant:1 rapidly:1 achieve:1 academy:2 fi0:1 description:1 optimum:1 r1:6 diverges:1 wider:3 derive:4 ac:1 measured:2 eq:15 edward:1 predicted:1 indicate:1 quantify:1 direction:2 correct:1 human:2 explains:2 behaviour:1 fix:1 really:1 secondly:1 extension:2 frontier:1 hold:4 exp:2 panzeri:1 predict:1 substituting:1 vary:2 achieves:1 largest:2 mh099611:1 city:1 reflects:1 mit:1 clearly:1 gaussian:16 pn:1 cr:6 poi:2 broader:2 encode:1 derived:5 release:1 likelihood:9 contrast:2 inference:4 ganguli:2 dependent:5 dayan:1 inaccurate:2 entire:4 yund:1 quasi:1 corrupts:1 issue:1 fidelity:1 orientation:5 overall:1 animal:3 constrained:2 summed:1 integration:2 mutual:10 spatial:1 equal:1 once:3 aware:1 broad:1 stuart:1 mimic:1 minimized:1 t2:9 stimulus:34 discrepancy:1 others:1 employ:1 irreducible:1 future:1 gamma:2 national:2 beck:4 jeffrey:2 hepler:1 william:1 interest:1 investigate:1 light:1 stocker:3 implication:2 necessary:1 incomplete:2 taylor:2 harmful:1 theoretical:1 minimal:1 rao:2 newsome:1 miura:1 deviation:4 examining:1 too:1 dependency:3 varies:2 corrupted:1 periodic:1 xue:1 combined:1 peak:1 lee:1 systematic:1 probabilistic:2 terrence:1 decoding:3 infomax:1 bethge:4 connectivity:1 squared:8 again:3 recorded:1 opposed:1 tile:3 cosyne:1 inefficient:1 macke:1 return:1 potential:1 de:2 coding:17 sec:6 flatter:1 depends:9 idealized:6 closed:2 analyze:2 red:3 bayes:1 recover:1 jul:1 marcelo:1 accuracy:1 variance:17 characteristic:1 efficiently:2 yield:1 spaced:1 correspond:2 bayesian:10 accurately:2 mere:1 carlo:2 confirmed:1 published:1 reach:1 phys:1 sharing:1 decorrelated:1 grossly:1 energy:11 nonetheless:1 frequency:1 attributed:1 mi:22 gain:2 treatment:3 popular:1 subtle:1 amplitude:9 actually:1 alexandre:3 steve:1 higher:2 dt:1 response:10 maximally:1 wei:1 formulation:1 strongly:2 governing:1 implicit:1 correlation:25 d:2 gray:6 reveal:1 grows:4 utah:1 effect:8 brown:1 unbiased:2 true:4 analytically:2 conditionally:3 width:42 m:4 bote:1 necessitating:1 latham:5 stefano:1 fi:19 common:3 spiking:1 salt:1 jp:3 volume:1 extend:3 interpretation:1 refer:1 cambridge:1 tuning:69 resorting:1 sharpen:1 shawe:1 longer:1 cortex:2 feb:1 posterior:14 recent:2 scenario:1 selectivity:1 certain:1 gerwinn:1 arbitrarily:2 preserving:1 minimum:3 seen:1 xaq:1 determine:2 converge:1 dashed:3 signal:1 ii:1 full:1 unimodal:1 simoncelli:3 reduces:1 july:1 smooth:1 barwinska:1 match:2 faster:1 alan:1 cross:1 equally:1 award:1 converging:1 prediction:6 basic:1 heterogeneous:1 vision:2 metric:1 poisson:20 expectation:3 kernel:1 normalization:1 dec:1 cell:1 whereas:1 affecting:1 spacing:3 keiji:1 grow:2 source:1 ingmar:1 biased:3 induced:4 dehaene:1 contrary:1 lafferty:1 emitted:1 integer:2 near:2 presence:2 counting:3 identically:1 variety:2 independence:1 marginalization:1 psychology:1 perfectly:1 restrict:1 reduce:1 expression:8 bartlett:1 energetic:1 render:1 peter:3 generally:1 clear:1 detailed:1 amount:3 induces:1 http:1 homunculus:1 meir:1 nsf:1 s3:1 alters:1 neuroscience:4 blue:5 discrete:1 express:1 threshold:3 drawn:1 neither:1 deneve:2 sum:2 inverse:1 yaeli:1 uncertainty:1 respond:2 family:2 almost:1 reasonable:1 lake:1 rub:1 rotermund:1 bit:2 bound:21 activity:2 constraint:27 infinity:1 constrain:1 ri:6 uchida:1 aspect:1 argument:1 optimality:6 relatively:1 department:1 mcknight:1 across:2 slightly:2 smaller:1 describes:1 lp:1 rev:1 biologically:1 s1:2 invariant:1 taken:1 resource:2 agree:1 remains:1 count:19 tractable:4 tiling:11 available:2 gaussians:1 apply:2 observe:2 pierre:1 alternative:1 weinberger:1 odor:1 slower:1 rz:1 broaden:1 unreasonably:1 denotes:2 top:5 kanitscheider:1 marginalized:1 giving:1 especially:1 february:1 classical:4 psychophysical:1 spike:27 said:1 exhibit:1 link:1 capacity:3 majority:1 considers:1 toward:1 stim:2 neurometric:2 maximising:1 code:28 relationship:1 ratio:1 minimizing:2 mostly:1 info:1 trace:11 peggy:1 perform:1 allowing:1 upper:4 neuron:28 observation:1 finite:2 nist:2 t:1 defining:2 saturates:1 rn:1 jakob:1 arbitrary:2 introduced:1 david:1 namely:1 required:2 pair:1 marginalise:1 narrow:6 testbed:1 macaque:1 gaussianshaped:1 beyond:1 below:1 pattern:2 mismatch:1 regime:1 reading:1 replotted:1 green:3 discrimax:1 ducom:1 inversely:1 library:1 review:1 prior:31 literature:4 loss:8 limitation:1 proportional:1 organised:1 var:1 versus:1 digital:1 foundation:2 degree:2 sufficient:1 consistent:1 shadlen:1 principle:1 editor:2 charitable:1 row:4 supported:1 infeasible:1 allow:1 burges:1 institute:1 wide:1 characterizing:1 taking:1 distributed:3 overcome:1 lett:1 curve:34 calculated:1 cortical:2 pillow:2 kechen:1 evaluating:1 sensory:4 zachary:1 opper:1 agnieszka:2 far:1 approximate:1 emphasize:1 preferred:6 implicitly:2 deg:2 tolerant:1 summing:1 assumed:1 tolias:1 latent:1 nature:2 nicolas:1 career:1 flank:1 expansion:1 mse:26 bottou:1 berens:1 domain:2 sp:4 neurosci:2 s2:11 noise:46 arise:1 n2:5 allowed:3 body:1 neuronal:4 fig:10 referred:1 en:1 gatsby:3 darker:1 fails:1 pereira:1 explicit:1 doiron:1 comput:2 governed:1 perceptual:1 formula:1 down:4 showing:2 er:2 adding:1 effectively:1 shea:1 supplement:3 nat:1 budget:1 entropy:2 tc:1 simply:2 visual:2 scalar:1 sally:1 applies:1 brunel:1 aa:1 extracted:1 ma:2 haefner:1 oct:4 narrower:3 quantifying:2 ann:1 shared:4 fisher:20 feasible:2 change:2 except:1 sophie:1 total:9 called:1 e:2 xin:1 divisive:1 shannon:1 la:1 college:1 support:1 latter:1 jonathan:1 relevance:1 incorporate:1 princeton:3 phenomenon:1 correlated:3 |
4,848 | 539 | Nonlinear Pattern Separation in Single Hippocampal
Neurons with Active Dendritic Membrane
Anthony M. Zador t
t
Brenda J. Claiborne?
Depts. of Psychology and Cellular
& Molecular Physiology
Yale University
New Haven, CT 06511
[email protected]
Thomas H. Brown
t
?Division of Life Sciences
University of Texas
San Antonio, TX 78285
ABSTRACT
The dendritic trees of cortical pyramidal neurons seem ideally suited to
perfonn local processing on inputs. To explore some of the implications
of this complexity for the computational power of neurons, we simulated
a realistic biophysical model of a hippocampal pyramidal cell in which a
"cold spot"-a high density patch of inhibitory Ca-dependent K channels
and a colocalized patch of Ca channels-was present at a dendritic
branch point. The cold spot induced a non monotonic relationship between the strength of the synaptic input and the probability of neuronal
fIring. This effect could also be interpreted as an analog stochastic XOR.
1 INTRODUCTION
Cortical neurons consist of a highly branched dendritic tree that is electrically coupled to
the soma. In a typical hippocampal pyramidal cell, over 10,000 excitatory synaptic inputs
are distributed across the tree (Brown and Zador, 1990). Synaptic activity results in current
flow through a transient conductance increase at the point of synaptic contact with the
membrane. Since the primary means of rapid intraneuronal signalling is electrical, information flow can be characterized in tenns of the electrical circuit defIned by the synapses,
the dendritic tree, and the soma.
Over a dozen nonlinear membrane channels have been described in hippocampal pyramidal neurons (Brown and Zador, 1990). There is experimental evidence for a heterogeneous
distribution of some of these channels in the dendritic tree (e.g. Jones et al .? 1989). In the
absence of these dendritic channels, the input-output function can sometimes be reasonably
approximated by a modifIed sigmoidal model. Here we report that introducing a cold spot
51
52
Zador, Claiborne, and Brown
at the junction of two dendritic branches can result in a fundamentally different, nonmonotonic input-output function.
2 MODEL
The biophysical details of the circuit class defined by dendritic trees have been well characterized (reviewed in RaIl, 1977; Jack et al., 1983). The fundamental circuit consists of a
linear and a nonlinear component The linear component can be approximated by a set of
electrical compartments coupled in series (Fig. 1C), each consisting of a resistor and capacitor in parallel (Fig. 1B). The nonlinear component consists of a set of nonlinear resistors
in parallel with the capacitance.
The model is summarized in Fig. 1A. Briefly, simulations were performed on a
3000-compartment anatomical reconstruction of a region CAl hippocampal neuron (Claiborne et aI., 1992; Brown et al., 1992). All dendritic membrane was passive, except at the
cold spot (Fig. 1A). At the soma, fast K and Na channels (cf. Hodgkin-Huxley, 1952) generated action potentials in response to stimuli. The parameters for these channels were
modified from Lytton and Sejnowski (1991; cf. Borg-Graham, 1991).
A
Synapti_c......;+
input-~
Cold spot
Fast somatic
andNa
c
I
r-------------------------------~
I
,
,
"
t
,
" ___________________ .. _____________ ~ I
<
I
)
Radial and longitudinal Ca+2 diffusion
Fig. 1 The model. (A) The 3000-compartrnent electrical model used in these simulations was obtained from a 3-dimensional reconstruction of a hippocampal region CAl pyramidal neuron (Clai?
borne et al, 1992). Each synaptic pathway (A-D) consisted of an adjustable number of synapses
arrayed along the single branch indicated (see text). Random background activity was generated with
a spatially uniform distribution of synapses firing according to Poisson statistics. The neuronal mem?
brane was completely passive (linear), except at the indicated cold spot and at the soma. (B) In the
nonlinear circuit associated with a patch a neuronal membrane containing active channels, each chan?
nel is described by a voltage-dependent conductance in series with its an ionic battery (see text). In
the present model the channels were spatially localized, so no single patch contained all of the non?
linearities depicted in this hypothetical illustration. (Cl. A dendritic segment is illustrated in which
both electrical and ca2+ dynamics were modelled. Ca +buffering, and both radial and longitudinal
Ca2+ diffusion were simulated.
Nonlinear Panern Separation in Single Hippocampal Neurons
We distinguished four synaptic pathways A-D (see Fig. lA). Each pathway consisted of a
population of synapses activated synchronously. The synapses were of the fast AMPA type
(see Brown et. al., 1992). In addition. random background synaptic activity distributed uniformly across the dendritic tree fIred according to Poisson statistics.
The cold spot consisted of a high density of a Ca-activated K channel. the AHP current
(Lancaster and Nicoll. 1987; Lancaster et. aI., 1991) colocalized with a low density patch
ofN-type Ca channels (Lytton and Sejnowski, 1991; cf. Borg-Graham, 1991). Upon localized depolarization in the region of the cold ~t. influx of Ca2+ through the Ca channel resulted in a transient increase in the local rCa +]. The model included Ca2+ buffering, and
both radial and longitudinal diffusion in the region of the cold spot. The increased [Ca2+]
activated the inhibitory AHP current. The interplay between the direct excitatory effect of
synaptic input, and its inhibitory effect via the AHP channels formed the functional basis
of the cold spot.
3 RESULTS
3.1 DYNAMIC BEHAVIOR
Representative behavior of the model is illustrated in Fig. 2. The somatic potential is plotted as a function of time in a series of simulations in which the number of activated synapses in pathway AlB was increased from 0 to about 100. For the fIrst 100 msec of each
simulation, background synaptic activity generated a noisy baseline. At t = 100 msec, the
indicated number of synapses fired synchronously five times at 100 Hz. Since the background activity was noisy, the outcome of the each simulation was a random process.
The key effect of the cold spot was to impose a limit on the maximum stimulus amplitude
that caused firing, resulting in a window of stimulus strengths that triggered an action potential. In the absence of the cold spot a greater synaptic stimulus invariably increased the
likelihood that a spike fIred. This limit resulted from the relative magnitude of the AHP
Sample Soma
tic Voltage Tra celll
60
0
~
.
I
0
>-
-60
<:>
Fig. 2 Sample runs. The membrane voltage at the soma is plotted as a f'wtction of time and synaptic
stimulus intensity. At t = 100 msec, a synaptic stimulus consisting of 5 pulses was activitated. The
noisy baseline resulted from random synaptic input. A single action potential resulted for input intensities within a range determined by the kinetics of the cold spot
53
54
Zador, Claiborne, and Brown
current "threshold" to the threshold for somatic spiking. The AHP current required a relatively high level of activity for its activation. This AHP current "threshold" reflected the
sigmoidal voltage dependence of N-type Ca current activation (V1I2 = -28 mV), since only
as the dendritic voltage approached V1I2 did dendritic [Ca2+] rise enough to activate the
AHP current. Because V1I2 was much higher than the threshold for somatic spiking (about
-55 mV under current clamp), there was a window of stimulus strengths sufficient to trigger
a somatic action potential but insufficient to activate the AHP current Only within this
window of between about 20 and 60 synapses (Fig. 2) did an action potential occur.
3.2 LOCAL NON-MONOTONIC RESPONSE FUNCTION
Because the background activity was random, the outcome of each simulation (e.g. Fig. 2)
represented a sample of a random process. This random process can be used to defme many
different random variables. One variable of interest is whether a spike fired in response to
a stimulus. Although this measure ignores the dynamic nature of neuronal activity, it was
still relatively informative because in these simulations no more than one spike fired per
experiment
Fig. 3A shows the dependence of firing probability on stimulus strength. It was obtained
by averaging over a population of simulations of the type illustrated in Fig. 2. In the absence of AHP current (dotted line), the fIring probability was a sigmoidal function of activity. In its presence, the firing probability was a smooth nonmonotonic function of the
activity (solid line). The firing probability was maximum at about 35 synapses, and occurred only in the range between about 10 and 80 synapses. The statistics illustrated in Fig.
3A quantify the nonmonotonicity that is implied by the single sample shown in Fig. 2.
Spikes required the somatic Hodgkin-Huxley-like Na and K channels. To a first approximation, the effect of these channels was to convert a continuous variable-the somatic voltage-into a discrete variable-the presence or absence of a spike. Although this
approximation ignores the complex interactions between the soma and the cold spot, it is
useful for a qualitative analysis. The nonmonotonic dependence of somatic activity on syn-
A
B
1.0
.....>..
-
Cold SpolCold Spol+
..... 0 .6
;0
III
,ll
0
'"'
Po.
DO
0.6
0.4
,
?
~
C
.~
I
I
I
I
r;: 0.2
0.0 '--__' - - _........_ ........--.i::o.........~"__._.......J
o 20 40 60 80 100 120
Number of active synpases
..
-56
Cold SpotCold Spot+
> -56
E
cu -60
".-
tIO
CI
0p
..
.- .'-
"
.-
.. --"
"
-62
-64
JI:
CI
cu
Po. -66
o
20 40 60 60 100 120
Number of active synpases
Fig. 3 Nonmonotonic input-output relation. (A) Each point represents the probability that at least
one spike was fIred at the indicated activity level. In the absence of a cold spot, the fIring probability increased sharply and monotonically as the number of synapses in pathway C/D increased (dotted Une). In contrast, the fIring probability reached amaximumforpathwayA/B and then decreased
(solid line). (B) Each point represents the peak somatic voltage for a single simulation at the indicated activity level in the presence (pathway AlB; solid line) and absence (pathway C/D,? dotted
Une) of a cold spot Because each point represents the outcome of a single simulation, in contrast
to the average used in (A), the points reflect the variance due to the random background activity.
Nonlinear Pattern Separation in Single Hippocampal Neurons
aptic activity was preserved even when active channels at the soma were eliminated (Fig.
3B). This result emphasizes that the critical nonlinearity was the cold spot itself.
3.3 NONLINEAR PATTERN SEPARATION
So far. we have treated the output as a function of a scalar-the total activity in pathway
AlB (or CID). In Fig. 3 for example. the total activity was defmed as the sum of the activities in pathway A and B. The spatial organization of the afferents onto 2 pairs of branches-A & B and C & D (Fig. I)-suggested considering the output as a function of the
activity in the separate elements of each pair.
The effect of the cold spot can be viewed in terms of the dependence of fIring as a function
of separate activity in pathways A and B (Fig. 4). Each fIlled circle indicates that the neuron
fIred for the indicated input intensity of pathways A and B. while a small dot indicates that
it did not fire. As suggested by (Fig. 3). the fIring probability was highest when the total
activity in the two pathways was at some intennediate level. The neuron did not fIre when
the total activity in the two pathways was too large or too small. In the absence of the cold
spot, only a minimum activity level was required.
In our model the probability of fIring was a continuous function of the inputs. In the presence of the dendritic cold spot, the corners of this function suggested the logical operation
XOR. The probability of fIring was high if only one input was activated and low if both or
neither was activated.
4 DISCUSSION
4.1 ASSUMPTIONS
Neuronal morphology in the present model was based on a precise reconstruction of a region CAl pyramidal neuron. The main additional assumptions involved the kinetics and
distribution of the four membrane channels. and the dynamics of Ca2+in the neighborhood
of influx. The forms assumed for these mechanisms were biophysically plausible. and the
kinetic parameters were based on estimates from a collection of experimental studies (listed
in Lytton and Sejnowski. 1991; Zador et aI .? 1990). Variation within the range of uncertainty of these parameters did not alter the main conclusions. The chief untested assumption of this model was the existence of cold spots. Although there is experimental evidence
...... ....... ..
_
......
.
.....
_..... ?... .
t --..... .. .. .
- . e .??. . . . . . .
~
~
==' =::::~ ? ????
p...
~
.. :. iq:I:I!'! :
iiiilli Iii iI fifl iii
~::~:~
:~ ~~~:.:
? ????
...... :::::::;Z!Zi
::::: :::::
Input A
-+
Fig.4 Nonlinear pattern separation Neuronal fIring is represented as a joint nmction of two input
pathways (AlB). Filled circles indicate that the neuron fIred for the indicated stimulus parameters.
Some indication of the stochastic nature of this function. resulting fonn the noisy background, is given by the density of interdigitation of points and circles.
55
56
Zador, Claiborne, and Brown
supporting the presence of both Ca and AHP channels in the dendrites. there is at present
no direct evidence regarding their colocalization.
4.2 COMPUTATIONS IN SINGLE NEURONS
4.2.1 Neurons and Processing Elements
The limitations of the McCulloch and Pitts (1943) PE as a neuron model have long been
recognized. Their threshold PEt in which the output is the weighted sum of the inputs
passed through a threshold, is static, deterministic and treats all inputs equivalently. This
model ignores at least three key complexities of neurons: temporal, spatial and stochastic.
In subsequent years, augmented models have attempted to capture aspects of these complexities. For example, the leaky integrator (Caianiello, 1961; Hopfield, 1984) incorporates the temporal dynamics implied by the linear RC component of the circuit element
pictured in Fig. IB. We have demonstrated that the input-output function of a realistic neuron model can have qualitatively different behavior from that of a single processing element(pE).
4.2.2 Interactions Within The Dendritic Tree
The early work ofRall (1964) stressed the spatial complexity of even linear dendritic models. He noted that input from different synapses cannot be considered to arrive at a single
point, the soma. Koch et al. (1982) extended this observation by exploring the nonlinear
interactions between synaptic inputs to different regions of the dendritic tree. They emphasized that these interactions can be local in the sense that they effect subpopulations of synapses and suggested that the entire dendritic tree can be considered in terms of electrically
isolated subunits. They proposed a specific role for these subunits in computing a vetoan analog AND-NOT---that might underlie directional selectivity in retinal ganglion cells.
The veto was achieved through inhibitory inputs.
The underlying neuron models of Koch et al. (1982) and Rall (1964) were time-varying but
linear, so it is not surprising that the resulting nonlinearities were monotonic. Much steeper
nonlinearities were achieved by Shepherd and Brayton (1987) in a model that assumed excitable spines with fast Hodgkin-Huxley K and Na channels. These channels alone could
implement the digital logic operations AND and OR. With the addition of extrinsic inhibitory inputs, they showed that a neuron could implement a full complement of digital logic
operations, and concluded that a dendritic tree could in principle implement arbitrarily
complex logic operations.
The emphasis of the present model differs from that of both the purely linear and of the digital approaches, although it shares their emphasis on the locality of dendritic computation.
Because the cold spot involved strongly nonlinear channels, it implemented a non mono tonic response function, in contrast to strictly linear dendritic models. At the same time, the
present model retained the essentially analog nature of intraneuronal signalling, in contrast
to the digital dendritic models. This analog mode seems better suited to processing large
numbers of noisy inputs because it preserves the uncertainties rather than making an immediate decision. Focussing on the analog nature of the response eliminated the requirement
for operating within the digital range of channel dynamics.
The NMDA receptor-gated channel can give rise to an analog AND with a weaker voltagedependence than that induced by fast Na and K channels. Mel (1992) described a model in
which synapses mediating increases to both the NMDA and AMPA conductances were distributed across the dendritic tree of a cortical neuron. When the synaptic activity was dis-
Nonlinear Pattern Separation in Single Hippocampal Neurons
tributed in appropriately sized clusters, the resulting neuronal response function was
reminiscent of that of a sigma-pi unit With suitable preprocessing of inputs. the neuron
could perform complex pattern discrimination.
A unique feature of the present model is that functional inhibition arose from purely excitatory inputs. This mechanism underlying this inhibition -the AHP current-was intrinsic
to the membrane. In both the Koch et ale (1982) and Brayton and Shepherd (1987) models.
the veto or NOT operation was achieved through extrinsic synaptic inhibition. This requires
additional neuronal circuitry. In the case of a dedicated sensory system like the directionally selective retinal granule cell. it is not unreasonable to imagine that the requisite neuronal circuitry is hardwired. In the limiting case of the digital model, the requisite circuitry
would involve a separate inhibitory interneuron for each NOT-gate.
4.2.3 Adaptive Dendritic Computation
What algorithms can harness the computational potential of the dendritic tree? Adaptive
dendritic computation is a very new subject. Brown et ale (1991, 1992) developed a model
in which synapses distributed across the dendritic tree showed interesting forms of spatial
self-organization. Synaptic plasticity was governed by a local biophysically-motivated
Hebb rule (Zador et al' 1990). When temporally correlated but spatially uncorrelated inputs were presented to the neuron, spatial clusters of strengthened synapses emerged within
the dendritic tree. The neuron converted a temporal correlation into a spatial correlation.
J
The computational role of clusters of strengthened synapses within the dendritic tree becomes important in the presence of nonlinear membrane. If the dendrites are purely passive. then saturation ensures that the current injected per synapse actually decreases as the
clustering increases. If purely regenerative nonlinearities are present (Brayton and Shepherd. 1987; Mel. 1992), then the response increases. The cold spot extends the range of
local dendritic computations.
What might control the formation and distribution of the cold spot itself? Cold spots might
arise from the fortuitous colocalization of Ca and K AHP channels. Another possibility is
that some specific biophysical mechanism creates cold spots in a use-dependent manner.
Candidate mechanisms might involve local changes in second messengers such as [Ca2+]
or longitudinal potential gradients (if. Poo, 1985). Bell (1992) has shown that this second
mechanism can induce computationally interesting distributions of membrane channels.
4.3 WHY STUDY SINGLE NEURONS?
We have illustrated an important functional difference between a single neuron and aPE.
A neuron with cold spots can perform extensive local processing in the dendritic tree, giving rise to a complex mapping between input and output. A neuron may perhaps be likened
to a "micronet" of simpler PEs. since any mapping can be approximated by a sufficiently
complex network of sigmoidal units (Cybenko, 1989). This raises the objection that since
micronets represent just a subset of all neural networks, there may be little to be gained by
studying the properties of the special case of neurons.
The intuitive justification for studying single neurons is that they represent a large but highly constrained subset that may have very special properties. Knowledge of the properties
general to all sufficiently complex PE networks may provide little insight into the properties specific to single neurons. These properties may have implications for the behavior of
circuits of neurons. It is not unreasonable to suppose that adaptive mechanisms in biological circuits will utilize the specific strengths of single neurons.
57
58
Zador, Claiborne, and Brown
Acknowledgments
We thank Michael Hines for providing NEURON-MODL assisting with new membrane
mechanisms. This research was supported by grants from the Office of Naval Research,
the Defense Advanced Research Projects Agency, and the Air Force Office of Scientific
Research.
References
Bell, T. (1992) Neural in/ormation processing systems 4 (in press).
Borg-Graham, L.J. (1991) In H. Wheal and J. Chad (Eds.) Cellular and Molecular Neurobiology: A
Practical Approach. New York: Oxford University Press.
Brown, T.H. and Zador, AM. (1990). In G. Shepherd (Ed.) The synaptic organization of the brain
(Vol. 3, pp. 346-388). New York: Oxford University Press.
Brown, T.H., Mainen, Z.F., Zador, A.M. and Claiborne, B.l (1991) Neural inJormationprocessing
systems 3: 39-45.
Brown, T .H., Zador, A.M., Mainen, Z.F., and Claiborne, B.J. (1992). In: Single neuron computation.
Eds. T. McKenna, J. Davis, and S.F. Zornetzer. Academic Press (inpress).
Caianiello, E.R. (1961) J. Thear. BioI. 1: 209-235.
Claiborne, BJ., Zador, A.M., Mainen, Z.F., and Brown, T.H. (1992). In: Single neuron computation.
Eds. T. McKenna, J. Davis, and S.F. Zornetzer. Academic Press (in press).
Cybenko, G. (1989) Math. Control, Signals Syst. 2: 303-314.
Hines, M. (1989). Int. J. Biomed. Comp, 24: 55-68.
Hodgkin, A.L. and Huxley, A.F. (1952) J. Physiol.117: 500-544.
Hopfield. JJ. (1984)Proc. Natl. Acad. Sci. USA 81: 3088-3092.
Jack, J. Noble, A. and Tsien, R.W. (1975) Electrical current flow in excitable membranes. London:
Oxford Press.
Jones, O.T., Kunze, D.L. and Angelides, KJ. (1989) Science. 244:1189-1193.
Koch, C., Poggio, T. and Torre, V. (1982) Proc. R. Soc. London B. 298: 227-264.
Lancaster, B. and Nicoll, R.A. (1987) J. Physiol. 389: 187-203.
Lancaster, B., Perkel, 0.1, and Nicoll, R.A. (1991) J. Neurosci. 11:23-30.
Lytton, W.W. and Sejnowski, T.l (1991) J. Neurophys. 66: 1059-1079.
McCulloch, W.S. and Pitts, W. (1943) Bull. Math. Biophys. 5: 115-137.
Mel, B. (1992) Neural Computation (in press).
Poo, M-m. (1985) Ann. Rev. Neurosci. 8: 369-406.
Rall, W. (1977) In: Handbook ofphysiology. Eds. E. Kandel and S. Geiger. Washington D.C.: American Physiological Society, pp. 39-97.
Rall, W. (1964) In: Neural theory and modeling. Ed. R.F. Reiss. Stanford Univ. Press, pp. 73-79.
Shepherd, G.M. and Brayton, R.K. (1987) Neuroscience 21: 151-166.
Zador, A., Koch, C. and Brown, T.H. (1990) Proc. Natl. Acad. Sci. USA 87: 6718-6722.
| 539 |@word cu:2 briefly:1 seems:1 pulse:1 simulation:10 fonn:1 fortuitous:1 solid:3 series:3 mainen:3 longitudinal:4 current:14 neurophys:1 surprising:1 activation:2 reminiscent:1 physiol:2 subsequent:1 realistic:2 informative:1 plasticity:1 arrayed:1 discrimination:1 alone:1 signalling:2 une:2 math:2 sigmoidal:4 simpler:1 five:1 rc:1 along:1 direct:2 borg:3 qualitative:1 consists:2 pathway:14 manner:1 spine:1 rapid:1 behavior:4 morphology:1 integrator:1 brain:1 perkel:1 rall:3 little:2 window:3 considering:1 becomes:1 project:1 linearity:1 underlying:2 circuit:7 intennediate:1 mcculloch:2 what:2 tic:1 interpreted:1 depolarization:1 developed:1 perfonn:1 temporal:3 hypothetical:1 control:2 unit:2 underlie:1 grant:1 angelides:1 local:8 treat:1 likened:1 limit:2 acad:2 receptor:1 oxford:3 tributed:1 firing:14 might:4 emphasis:2 range:5 unique:1 acknowledgment:1 practical:1 implement:3 differs:1 spot:27 cold:29 bell:2 physiology:1 radial:3 subpopulation:1 induce:1 onto:1 cannot:1 cal:3 deterministic:1 demonstrated:1 poo:2 zador:15 rule:1 insight:1 population:2 variation:1 justification:1 limiting:1 imagine:1 trigger:1 suppose:1 element:4 approximated:3 role:2 electrical:6 capture:1 region:6 ensures:1 ormation:1 decrease:1 highest:1 agency:1 complexity:4 ideally:1 battery:1 dynamic:6 caianiello:2 raise:1 segment:1 purely:4 upon:1 division:1 creates:1 completely:1 basis:1 po:2 joint:1 hopfield:2 represented:2 tx:1 univ:1 fast:5 activate:2 sejnowski:4 london:2 approached:1 formation:1 nonmonotonic:4 lancaster:4 outcome:3 neighborhood:1 emerged:1 stanford:1 plausible:1 statistic:3 noisy:5 itself:2 directionally:1 interplay:1 triggered:1 indication:1 biophysical:3 reconstruction:3 clamp:1 interaction:4 fired:8 intuitive:1 cluster:3 requirement:1 lytton:4 iq:1 soc:1 brayton:4 implemented:1 indicate:1 quantify:1 untested:1 torre:1 stochastic:3 transient:2 cybenko:2 dendritic:32 biological:1 exploring:1 kinetics:2 strictly:1 koch:5 considered:2 sufficiently:2 mapping:2 bj:1 pitt:2 circuitry:3 early:1 proc:3 weighted:1 modified:2 rather:1 arose:1 varying:1 voltage:7 office:2 naval:1 likelihood:1 indicates:2 contrast:4 baseline:2 sense:1 am:1 dependent:3 entire:1 relation:1 selective:1 biomed:1 spatial:6 special:2 constrained:1 washington:1 eliminated:2 buffering:2 represents:3 jones:2 noble:1 alter:1 report:1 stimulus:10 fundamentally:1 haven:1 preserve:1 resulted:4 consisting:2 fire:2 conductance:3 invariably:1 interest:1 organization:3 highly:2 possibility:1 activated:6 natl:2 implication:2 poggio:1 tree:17 filled:2 circle:3 plotted:2 isolated:1 increased:5 modeling:1 bull:1 introducing:1 subset:2 uniform:1 too:2 density:4 fundamental:1 peak:1 michael:1 na:4 reflect:1 containing:1 borne:1 corner:1 american:1 syst:1 potential:8 nonlinearities:3 converted:1 retinal:2 summarized:1 int:1 modl:1 tra:1 caused:1 mv:2 afferent:1 performed:1 steeper:1 reached:1 parallel:2 air:1 formed:1 compartment:2 xor:2 variance:1 directional:1 modelled:1 biophysically:2 emphasizes:1 ionic:1 comp:1 synapsis:17 messenger:1 synaptic:18 ed:6 pp:3 involved:2 associated:1 static:1 logical:1 knowledge:1 nmda:2 amplitude:1 syn:1 actually:1 higher:1 reflected:1 response:7 harness:1 synapse:1 strongly:1 just:1 correlation:2 chad:1 nonlinear:14 mode:1 perhaps:1 indicated:7 alb:4 scientific:1 usa:2 effect:7 brown:15 consisted:3 spatially:3 illustrated:5 ll:1 self:1 defmed:1 davis:2 noted:1 mel:3 hippocampal:9 dedicated:1 passive:3 jack:2 functional:3 spiking:2 ji:1 analog:6 occurred:1 he:1 ai:3 nonlinearity:1 dot:1 operating:1 inhibition:3 chan:1 showed:2 selectivity:1 tenns:1 arbitrarily:1 life:1 minimum:1 greater:1 additional:2 impose:1 recognized:1 focussing:1 monotonically:1 signal:1 ale:2 branch:4 ii:1 full:1 assisting:1 smooth:1 characterized:2 academic:2 long:1 molecular:2 regenerative:1 heterogeneous:1 essentially:1 poisson:2 sometimes:1 represent:2 achieved:3 cell:4 panern:1 background:7 addition:2 preserved:1 decreased:1 objection:1 pyramidal:6 concluded:1 appropriately:1 shepherd:5 induced:2 hz:1 subject:1 ape:1 veto:2 flow:3 capacitor:1 seem:1 incorporates:1 presence:6 iii:3 enough:1 psychology:1 zi:1 regarding:1 texas:1 whether:1 motivated:1 defense:1 passed:1 york:2 jj:1 action:5 antonio:1 useful:1 listed:1 involve:2 nel:1 inhibitory:6 dotted:3 neuroscience:1 extrinsic:2 per:2 anatomical:1 discrete:1 vol:1 key:2 four:2 soma:9 threshold:6 mono:1 neither:1 branched:1 diffusion:3 utilize:1 convert:1 sum:2 colocalized:2 run:1 year:1 uncertainty:2 ca2:8 hodgkin:4 injected:1 arrive:1 extends:1 separation:6 patch:5 geiger:1 decision:1 graham:3 ct:1 yale:2 activity:24 strength:5 occur:1 huxley:4 sharply:1 influx:2 aspect:1 relatively:2 according:2 electrically:2 membrane:12 across:4 rev:1 making:1 rca:1 computationally:1 nicoll:3 mechanism:7 ahp:12 studying:2 junction:1 operation:5 unreasonable:2 distinguished:1 gate:1 existence:1 thomas:1 clustering:1 cf:3 giving:1 granule:1 society:1 contact:1 implied:2 capacitance:1 spike:6 primary:1 dependence:4 cid:1 gradient:1 separate:3 thank:1 simulated:2 sci:2 cellular:2 pet:1 retained:1 relationship:1 illustration:1 insufficient:1 providing:1 equivalently:1 mediating:1 sigma:1 rise:3 adjustable:1 gated:1 perform:2 neuron:37 observation:1 supporting:1 immediate:1 subunit:2 extended:1 tonic:1 precise:1 neurobiology:1 synchronously:2 somatic:9 intensity:3 complement:1 defme:1 required:3 pair:2 extensive:1 suggested:4 pattern:6 saturation:1 power:1 critical:1 suitable:1 treated:1 force:1 hardwired:1 advanced:1 pictured:1 temporally:1 excitable:2 coupled:2 brenda:1 kj:1 text:2 nonmonotonicity:1 relative:1 interesting:2 limitation:1 localized:2 digital:6 sufficient:1 principle:1 uncorrelated:1 share:1 pi:1 excitatory:3 supported:1 dis:1 weaker:1 leaky:1 distributed:4 cortical:3 ignores:3 sensory:1 collection:1 qualitatively:1 san:1 preprocessing:1 adaptive:3 far:1 claiborne:9 logic:3 active:5 mem:1 handbook:1 assumed:2 zornetzer:2 continuous:2 chief:1 reviewed:1 why:1 channel:26 reasonably:1 nature:4 ca:10 dendrite:2 cl:1 ampa:2 anthony:1 complex:6 did:5 main:2 neurosci:2 arise:1 neuronal:9 fig:22 representative:1 augmented:1 strengthened:2 hebb:1 msec:3 resistor:2 kandel:1 candidate:1 governed:1 rail:1 pe:4 ib:1 dozen:1 ofn:1 specific:4 emphasized:1 mckenna:2 physiological:1 evidence:3 consist:1 intrinsic:1 gained:1 ci:2 magnitude:1 tio:1 biophys:1 depts:1 interneuron:1 suited:2 locality:1 depicted:1 tsien:1 explore:1 ganglion:1 contained:1 scalar:1 monotonic:3 kinetic:1 hines:2 bioi:1 viewed:1 sized:1 ann:1 absence:7 change:1 included:1 typical:1 except:2 uniformly:1 determined:1 averaging:1 total:4 experimental:3 la:1 attempted:1 stressed:1 requisite:2 reiss:1 correlated:1 |
4,849 | 5,390 | Optimal Neural Codes for Control and Estimation
Alex Susemihl1 , Manfred Opper
Methods of Artificial Intelligence
Technische Universit?at Berlin
1
Current affiliation: Google
Ron Meir
Department of Electrical Engineering
Technion - Haifa
Abstract
Agents acting in the natural world aim at selecting appropriate actions based on
noisy and partial sensory observations. Many behaviors leading to decision making and action selection in a closed loop setting are naturally phrased within a
control theoretic framework. Within the framework of optimal Control Theory,
one is usually given a cost function which is minimized by selecting a control
law based on the observations. While in standard control settings the sensors are
assumed fixed, biological systems often gain from the extra flexibility of optimizing the sensors themselves. However, this sensory adaptation is geared towards
control rather than perception, as is often assumed. In this work we show that sensory adaptation for control differs from sensory adaptation for perception, even for
simple control setups. This implies, consistently with recent experimental results,
that when studying sensory adaptation, it is essential to account for the task being
performed.
1
Introduction
Biological systems face the difficult task of devising effective control strategies based on partial information communicated between sensors and actuators across multiple distributed networks. While
the theory of Optimal Control (OC) has become widely used as a framework for studying motor control, the standard framework of OC neglects many essential attributes of biological control [1, 2, 3].
The classic formulation of closed loop OC considers a dynamical system (plant) observed through
sensors which transmit their output to a controller, which in turn selects a control law that drives
actuators to steer the plant. This standard view, however, ignores the fact that sensors, controllers
and actuators are often distributed across multiple sub-systems, and disregards the communication
channels between these sub-systems. While the importance of jointly considering control and communication within a unified framework was already clear to the pioneers of the field of Cybernetics
(e.g., Wiener and Ashby), it is only in recent years that increasing effort is being devoted to the
formulation of a rigorous systems-theoretic framework for control and communication (e.g., [4]).
Since the ultimate objective of an agent is to select appropriate actions, it is clear that sensation and
communication must subserve effective control, and should be gauged by their contribution to action
selection. In fact, given the communication constraints that plague biological systems (and many
current distributed systems, e.g., cellular networks, sensor arrays, power grids, etc.), a major concern
of a control design is the optimization of sensory information gathering and communication (consistently with theories of active perception). For example, recent theoretical work demonstrated a sharp
communication bandwidth threshold below which control (or even stabilization) cannot be achieved
(for a summary of such results see [4]). Moreover, when informational constraints exists within a
control setting, even simple (linear and Gaussian) problems become nonlinear and intractable, as
exemplified in the famous Witsenhausen counter-example [5].
The inter-dependence between sensation, communication and control is often overlooked both in
control theory and in computational neuroscience, where one assumes that the overall solution to
the control problem consists of first estimating the state of the controlled system (without reference
1
to the control task), followed by constructing a controller based on the estimated state. This idea,
referred to as the separation principle in Control Theory, while optimal in certain restricted settings
(e.g., Linear Quadratic Gaussian (LQG) control) is, in general, sub-optimal [6]. Unfortunately, it is
in general very difficult to provide optimal solutions in cases where separation fails. A special case
of the separation principle, referred to as Certainty Equivalence (CE), occurs when the controller
treats the estimated state as the true state, and forms a controller assuming full state information. It
is generally overlooked, however, that although the optimal control policy does not depend directly
on the observation model at hand, the expected future costs do depend on the specifics of that model
[7]. In this sense, even when CE holds, costs still arise from uncertain estimates of the state and one
can optimise the sensory observation model to minimise these costs, leading to sensory adaptation.
At first glance, it might seem that the observation model that will minimise the expected future cost
will be the observation model that minimises the estimation error. We will show, however, that this
is not generally the case.
A great deal of the work in computational neuroscience has dealt independently with the problem
of sensory adaptation and control, while, as stated above, these two issues are part and parcel of
the same problem. In fact, it is becoming increasingly clear that biological sensory adaptation is
task-dependent [8, 9]. For example, [9] demonstrates that task-dependent sensory adaptation takes
place in purely motor tasks, explaining after-effect phenomena seen in experiments. In [10], the
authors show that specific changes occur in sensory regions, implying sensory plasticity in motor
learning. In this work we consider a simple setting for control based on spike time sensory coding,
and study the optimal coding of sensory information required in order to perform a well-defined
motor task. We show that even if CE holds, the optimal encoder strategy, minimising the control cost,
differs from the optimal encoder required for state estimation. This result demonstrates, consistently
with experiments, that neural encoding must be tailored to the task at hand. In other words, when
analyzing sensory neural data, one must pay careful care to the task being performed. Interestingly,
work within the distributed control community dealing with optimal assignment and selection of
sensors, leads to similar conclusions and to specific schemes for sensory adaptation.
The interplay between information theory and optimal control is a central pillar of modern control
theory, and we believe it must be accounted for in the computational neuroscience community.
Though statistical estimation theory has become central in neural coding issues, often through the
Cram?er-Rao bound, there have been few studies bridging the gap between partially observed control
and neural coding. We hope to narrow this gap by presenting a simple example where control
and estimation yield different conclusions. The remainder of the paper is organised as follows:
In section 1.1 we introduce the notation and concepts; In section 2 we derive expressions for the
cost-to-go of a linear-quadratic control system observed through spikes from a dense populations of
neurons; in section 3 we present the results and compare optimal codes for control and estimation
with point-process filtering, Kalman filtering and LQG control; in section 4 we discuss the results
and their implications.
1.1
Optimal Codes for Estimation and Control
We will deal throughout this paper with a dynamic system with state Xt , observed through noisy
sensory observations Zt , whose conditional distribution can be parametrised by a set of parameters
?, e.g., the widths and locations of the tuning curves of a population of neurons or the noise properties of the observation process. The conditional distribution is then given by P? (Zt |Xt = x).
Zt could stand for a diffusion process dependent on Xt (denoted Yt ) or a set of doubly-stochastic
Poisson processes dependent on Xt (denoted Ntm ). In that sense, the optimal Bayesian encoder for
an estimation problem, based on the Mean Squared Error (MSE) criterion, can be written as
2
??e = argmin E z E Xt Xt ? X?t (Zt ) Zt = z ,
?
? t (Zt ) = E [Xt |Zt ] is the posterior mean, computable, in the linear Gaussian case, by the
where X
Kalman filter. We will throughout this paper consider the MMSE in the equilibrium, that is, the
error in estimating Xt from long sequences of observations Z[0,t] . Similarly, considering a control
problem with a cost given by
Z T
C(X 0 , U 0 ) =
c(Xs , Us , s)ds + cT (XT ),
0
2
where X t = {Xs |s ? [t, T ]}, U t = {Us |s ? [t, T ]}, and so forth. We can define
??c = argmin E z min [E X t [C(X 0 , U 0 )|Z t = z]] .
Ut
?
The certainty equivalence principle states that given a control policy ? ? : X ? U which minimises
the cost C,
? ? = argmin C(X 0 , ?(X 0 )),
?
the optimal control policy for the partially observed problem given by noisy observations Z 0 of X 0
is given by
?CE (Z t ) = ? ? (E [X 0 |Z t ]) .
Note that we have used the notation ?(X 0 ) = {?(Xs ), s ? [0, T ]}.
2
Stochastic Optimal Control
In stochastic optimal control we seek to minimize the expected future cost incurred by a system
with respect to a control variable applied to that system. We will consider linear stochastic systems
governed by the SDE
dXt = (AXt + BUt ) dt + D1/2 dWt ,
(1a)
with a cost given by
Z T
C(X t , U t , t) =
Xs> QXs + Us> RUs ds + XT> QT XT .
(1b)
t
From Bellman?s optimality principle or variational analysis [11], it is well known that the optimal
control is given by Ut? = ?R?1 B > St Xt , where St is the solution of the Riccati equation
?S? t = Q + ASt + St A> ? St B > R?1 BSt ,
(2)
with boundary condition ST = QT . The expected future cost at time t and state x under the optimal
control is then given by
Z T
1 >
J(x, t) = min E [C(X t , U t , t)|Xt = x] = x St x +
Tr (DSs ) ds.
Ut
2
t
This is usually called the optimal cost-to-go. However, the system?s state is not always directly
accessible and we are often left with noisy observations of it. For a class of systems e.g. LQG
control, CE holds and the optimal control policy for the indirectly observed control problem is
simply the optimal control policy for the original control problem applied to the Bayesian estimate of
the system?s state. In that sense, if the CE were to hold for the system above observed through noisy
observations Yt of the state at time t, the optimal control would be given simply by the observationdependent control Ut? = ?R?1 B > St E [Xt |Yt ] [7].
Though CE, when applicable, gives us a simple way to determine the optimal control, when considering neural systems we are often interested in finding the optimal encoder, or the optimal observation model for a given system. That is equivalent to finding the optimal tuning function for a
given neuron model. Since CE neatly separates the estimation and control steps, it would be tempting to assume the optimal codes obtained for an estimation problem would also be optimal for an
associated control problem. We will show here that this is not the case.
As an illustration, let us consider the case of LQG with incomplete state information. One could,
for example, take the observations to be a secondary process Yt , which itself is a solution to
dYt = F Xt dt + G1/2 dVt ,
the optimal cost-to-go would then be given by [11]
J(y, t) = min E C(X t , U t , t)Y[0,t] = y
(3)
Ut
=?t> St ?t + Tr (Kt St ) +
Z
T
Z
Tr (DSs ) ds +
t
t
3
T
Tr Ss BR?1 B > Ss Ks ds,
where we have defined Y[0,t] = {Ys , s ? [0, t]}, ?t = E[Xt |Y[0,t] ] and Kt = cov[Xt |Y[0,t] ]. We
give a demonstration of these results in the SI, but for a thorough review see [11]. Note that through
the last term in equation (3) the cost-to-go now depends on the parameters of the Yt process. More
precisely, the variance of the distribution of Xs given Yt , for s > t obeys the ODE
K? t = AKt + Kt A> + D ? Kt F > G?1 F Kt .
(4)
One could then choose the matrices F and G in such a way as to minimise the contribution of
the rightmost term in equation (3). Note that in the LQG case this is not particularly interesting,
as the conclusion is simply that we should strive to make Kt as small as possible, by making the
term F > G?1 F as large as possible. This translates to choosing an observation process with very
strong steering from the unobserved process (large F ) and a very small noise (small G). One case
that provides some more interesting situations is if we consider a two-dimensional system, where
we are restricted to a noise covariance with constant determinant. That means the hypervolume
spanned by the eigenvectors of the covariance matrix is constant. We will compare this case with
the Poisson-coded case below.
2.1
LQG Control with Dense Gauss-Poisson Codes
Let us now consider the case of the system given by equation (1a), but instead of observing the
system directly we observe a set of doubly-stochastic Poisson processes {Ntm } with rates given by
1
>
?m (x) = ? exp ? (x ? ?m ) P ? (x ? ?m ) .
(5)
2
To clarify, the process Ntm is a counting process which counts how many spikes the neuron m
has fired up to time t. In that sense, the differential of the counting process dNtm will give the
spike train process, a sum of Dirac delta functions placed at the times of spikes fired by neuron
m. Here P ? denotes the pseudo-inverse of P , which is used to allow for tuning functions that
do not depend on certain coordinates of the stimulus x. Furthermore, we will assume that the
tuning centre ?m are such that the probability of observing a spike of any neuron at a given time
? = P ?m (x) is independent of the specific value of the world state x. This can be a consequence
?
m
of either a dense packing of the tuning centres ?m along a given dimension of x, or of an absolute
insensitivity to that aspect of x through a null element in the diagonal of P ? . This is often called
the dense coding hypothesis [12]. It can be readily be shown that the filtering distribution is given
by P (Xt |{N[0,t) }) = N (?t , ?t ), where the mean and covariance are solutions to the stochastic
differential equations (see [13])
X
?1 ?
d?t = (A?t + BUt ) dt +
?t I + P ? ?t
P (?m ? ?t ) dNtm ,
(6a)
m
?1
d?t = A?t + ?t A> + D dt ? ?t P ? ?t I + P ? ?t
dNt ,
(6b)
m
m
where we have defined ?t = E[Xt |{N[0,t]
}] and ?t = cov[Xt |{N[0,t]
}]. Note that we have also
P
m
m
m
defined N[0,t] = {Ns |s ? [0, t]}, the history of the process Ns up to time t, and Nt = m Ntm .
Using Lemma 7.1 from [11] provides a simple connection between the cost function and the solution
of the associated Ricatti equation for a stochastic process. We have
Z T
>
C(X t , U t , t) = XT> QT XT +
Xs QXs + Us> RUs ds
=Xt> St Xt
Z
+
t
T
(Us + R?1 B > Ss Xs )> R(Us + R?1 B > Ss Xs )ds
t
Z
+
T
Z
Tr(DSs )ds +
t
T
dWs> D>/2 Ss Xs ds
t
Z
+
T
Xs> Ss D1/2 dWs .
t
We can average over P (X t , N t |{N[0,t) }) to obtain the expected future cost. That gives us
"Z
# Z
T
T
>
?1 >
>
?1 >
?t St ?t +Tr(?t St )+E
(Us + R B Ss Xs ) R(Us + R B Ss Xs )ds{N[0,t) } +
Tr(DSs )ds
t
t
4
m
We can evaluate the average over P (X t , {N m
t }|{N[0,t) }) in two steps, by first averaging over the
m
Gaussian densities P (Xs |{N[0,s] }) and then over P ({N[0,s] }|{N[0,t) }). The average gives
Z
E
t
T
h
i
(Us + R?1 B > Ss ?s )> R(Us + R?1 B > Ss ?s ) + Tr Ss BR?1 B > Ss ?s ({N[0,s] }) ds{N[0,t) } ,
where ?s and ?s are the mean and variance associated with the distribution P (Xs |{N[0,s) }). Note
that choosing Us = ?R?1 B > Ss ?s will minimise the expression above, consistently with CE. The
optimal cost-to-go is therefore given by
J({N[0,t) }, t) =?>
t St ?t + Tr(?t St )
Z
Z T
Tr (DSs ) ds +
+
T
Tr Ss BR?1 B > Ss E ?s ({N[0,s] })|{N[0,t) } ds
t
t
(7)
Note that the only term in the cost-to-go function that depends on the parameters of the encoders is
the rightmost term and it depends on it only through the average over future paths of the filtering
variance ?s . The average of the future covariance matrix is precisely the MMSE for the filtering
problem conditioned on the belief state at time t [13]. We can therefore analyse the quality of an
encoder for a control task by looking at the values of the term on the right for different encoding
parameters. Furthermore,
since the
dynamics of ?t given by equation (6b) is Markovian, we can
write the average E ?s |{N[0,t) } as E [?s |?t ]. We will define then the function f (?, t) which
gives us the uncertainty-related expected future cost for the control problem as
Z T
f (?, t) =
Tr Ss BR?1 B > Ss E [?s |?t = ?] ds.
(8)
t
2.2
Mutual Information
Many results in information theory are formulated in terms of the mutual information of the communication channel P? (Y |X). For example, the maximum cost reduction achievable with R bits of
information about an unobserved variable X has been shown to be a function of the rate-distortion
function with the cost as the distortion function [14]. More recently there has also been a lot of
interest in the so-called I-MMSE relations, which provide connections between the mutual information of a channel and the minimal mean squared error of the Bayes estimator derived from the
same channel [15, 16]. The mutual information for the cases we are considering is not particularly
complex, as all distributions are Gaussians. Let us denote by ?0t the covariance of of the unobserved
process Xt conditioned on some initial Gaussian distribution P0 = N (?0 , ?0 ) at time 0. We can
then consider the Mutual Information between the stimulus at time t, Xt , and the observations up to
time t, Y[0,t] or N[0,t] . For the LQG/Kalman case we have simply
Z
I(Xt ; Y[0,t] |P0 ) = dx dyP (x, y) [log P (x|y) ? log P (x)] = log |?0t | ? log |?t |,
where ?t is a solution of equation (4). For the Dense Gauss-Poisson code, we can also write
Z
I(Xt ; Nt |P0 ) = dx dn P (x, n) [log P (x|n) ? log P (x)] = log |?0t | ? E N[0,t] log |?t (N[0,t] )| ,
where ?t (N[0,t] ) is a solution to the stochastic differential equation (6b) for the given value of N[0,t] .
3
Optimal Neural Codes for Estimation and Control
What could be the reasons for an optimal code for an estimation problem to be sub-optimal for a
control problem? We present examples that show two possible reasons for different optimal coding
strategies in estimation and control. First, one should note that control problems are often defined
over a finite time horizon. One set of classical experiments involves reaching for a target under
time constraints [3]. If we take the maximal firing rate of the neurons (?) to be constant while
varying the width of the tuning functions, this will lead the number of observed spikes to be inversely
proportional to the precision of those spikes, forcing a trade-off between the number of observations
5
and their quality. This trade-off can be tilted to either side in the case of control depending on the
information available at the start of the problem. If we are given complete information on the system
state at the initial time 0, the encoder needs fewer spikes to reliably estimate the system?s state
throughout the duration of the control experiment, and the optimal encoder will be tilted towards
a lower number of spikes with higher precision. Conversely, if at the beginning of the experiment
we have very little information about the system?s state, reflected in a very broad distribution, the
encoder will be forced towards lower precision spikes with higher frequency. These results are
discussed in section 3.1.
Secondly, one should note that the optimal encoder for estimation does not take into account the
differential weighting of different dimensions of the system?s state. When considering a multidimensional estimation problem, the optimal encoder will generally allocate all its resources equally
between the dimensions of the system?s state. In the framework presented we can think of the dimensions as the singular vectors of the tuning matrix P and the resources allocated to it are the singular
values. In this sense, we will consider a set of coding strategies defined by matrices P of constant
determinant in section 3.2. This constrains the overall firing rate of the population of neurons to be
constant, and we can then consider how the population will best allocate its observations between
these dimensions. Clearly, if we have an anisotropic control problem, which places a higher importance in controlling one dimension, the optimal encoder for the control problem will be expected to
allocate more resources to that dimension. This is indeed shown to be the case for the Poisson codes
considered, as well as for a simple LQG problem when we constrain the noise covariance to have
the same structure.
We do not mean our analysis to be exhaustive as to the factors leading to different optimal codes
in estimation and control settings, as the general problem is intractable, and indeed, is not even
separable. We intend this to be a proof of concept showing two cases in which the analogy between
control and estimation breaks down.
3.1
The Trade-off Between Precision and Frequency of Observations
In this section we consider populations of neurons with tuning functions as given by equation (5)
with tuning centers ?m distributed along a one- dimensional line. In the case of the OrnsteinUhlenbeck process these will be simply one-dimensional values ?m whereas in the case of the
stochastic oscillator, we will consider tuning centres of the form ?m = (?m , 0)> , filling only the
first dimension of the stimulus space. Note
? that in both cases the (dense) population firing rate
? = P ?m (x) will be given by ?
? = 2?p?/|??|, where ?? is the separation between neigh?
m
bouring tuning centres ?m .
The Ornstein-Uhlenbeck (OU) process controlled by a process Ut is given by the SDE
dXt = (bUt ? ?Xt )dt + D1/2 dWt .
Equation (7) can then be solved by simulating the dynamics of ?s . This has been considered extensively in [13] and we refer to the results therein. Specifically, it has been found that the dynamics of
the average can be approximated in a mean-field approach yielding surprisingly good results. The
evolution of the average posterior variance is given by the average of equation (6b), which involves
nonlinear averages over the covariances. These are intractable, but a simple mean-field approach
yields the approximate equation for the evolution of the average h?s i = E [?s |?0 ]
d h?s i
>
? h?s i P ? h?s i I + P ? h?s i ?1 .
= A h?s i + h?s i A> + D ? ?
ds
The alternative is to simulate the stochastic dynamics of ?t for a large number of samples and
compute numerical averages. These results can be directly employed to evaluate the optimal costto-go in the control problem f (?, t).
Alternatively, we can look at a system with more complex dynamics, and we take as an example the
stochastic damped harmonic oscillator given by the system of equations
X? t = Vt , dVt = bUt ? ?Vt ? ? 2 Xt dt + ? 1/2 dWt .
(9)
Furthermore, we assume that the tuning functions only depend on the position of the oscillator,
therefore not giving us any information about the velocity. The controller in turn seeks to keep the
6
0.40
0.35
0.30
0.12
M M SE
M M SE
0.18
0.16
0.14
0.10
0.08
0.25
0.20
0.06
0.15
0.04
0.10
0.02
0.05
0.007
1.35
0.006
1.30
1.25
0 ,t0 )
0.003
1.20
f(
0 ,0)
f(
0.005
0.004
b)
1.10
1.15
0.002
1.05
0.001
1.00
0.95
0.0
0.000
0
1
2
3
4
5
0.2
0.4
0.6
p
0.8
1.0
1.2
Figure 1: The trade-off between the precision and the frequency of spikes is illustrated for the OU process
(a) and the stochastic oscillator (b). In both figures, the initial condition has a very uncertain estimate of the
system?s state, biasing the optimal tuning width towards higher values. This forces the encoder to amass the
maximum number of observations within the duration of the control experiment. Parameters for figure (a) were:
T = 2, ? = 1.0, ? = 0.6, b = 0.2, ? = 0.1, ?? = 0.05, Q = 0.1, QT = 0.001, R = 0.1. Parameters for
figure (b) were T = 5, ? = 0.4, ? = 0.8, ? = 0.4, r = 0.4, q = 0.4, QT = 0, ? = 0.5, ?? = 0.1.
oscillator close to the origin while steering only the velocity. This can be achieved by the choice of
matrices A = (0, 1; ?? 2 , ??), B = (0, 0; 0, b), D = (0, 0; 0, ? 2 ), R = (0, 0; 0, r), Q = (q, 0; 0, 0)
and P = (p2 , 0; 0, 0).
In figure 1 we provide the uncertainty-dependent costs for LQG control, for the Poisson observed
control, as well as the MMSE for the Poisson filtering problem and for a Kalman-Bucy filter with the
same noise covariance matrix P . This illustrates nicely the difference between Kalman filtering and
the Gauss-Poisson filtering considered here. The Kalman filter MSE has a simple, monotonically
increasing dependence on the noise covariance, and one should simply strive to design sensors with
the highest possible precision (p = 0) to minimise the MMSE and control costs. The Poisson
case leads to optimal performance at a non-zero value of p. Importantly the optimal values of p
for estimation and control differ. Furthermore, in view of section 2.2, we also plotted the mutual
information between the process Xt and the observation process Nt , to illustrate that informationbased arguments would lead to the same optimal encoder as MMSE-based arguments.
3.2
Allocating Observation Resources in Anisotropic Control Problems
A second factor that could lead to different optimal encoders in estimation and control is the structure of the cost function C. Specifically, if the cost functions depends more strongly on a certain
coordinate of the system?s state, uncertainty in that particular coordinate will have a higher impact
on expected future costs than uncertainty in other coordinates. We will here consider two simple
linear control systems observed by a population of neurons restricted to a certain firing rate. This
can be thought of as a metabolic constraint, since the regeneration of membrane potential necessary
for action potential generation is one of the most significant metabolic expenditures for neurons
[17]. This will lead to a trade-off, where an increase in precision in one coordinate will result in a
decrease in precision in the other coordinate.
We consider a population of neurons whose tuning functions cover a two-dimensional space. Taking
a two-dimensional isotropic OU system with state Xt = (X1,t , X2,t )> where both dimensions are
uncoupled, we can consider a population with tuning centres ?m = (?1m , ?2m )> densely covering
the stimulus space. To consider a smoother class of stochastic systems we will also consider a
two-dimensional stochastic oscillator with state Xt = (X1,t , V1,t , X2,t , V2,t )> , where again, both
dimensions are uncoupled, and the tuning centres of the form ?m = (?1m , 0, ?2m , 0)> , covering
densely the position space, but not the velocity space.
Since we are interested in the case of limited resources, we will restrict ourselves to populations with a tuning matrix P yielding a constant population firing rate. We can parametrise
these simply as POU (?) = p2 Diag(tan(?), cotan(?)), for the OU case and POsc (?) =
7
estimation
kalman filter
mean field
stochastic
LQG control
0.80
1.2
0.75
1.0
0.8
0.70
0.6
0.65
0.4
0.60
0.2
0.300
0.295
0.290
0.285
0.280
0.275
0.270
0.265
0.260
0.255
0.0
0.15
0.14
0.13
0.12
0.11
0.10
0.09
0.08
0.07
0.06
0.0
f (?0 , t0 )
f (?0 , t0 )
MMSE
0.85
0.2
0.4
0.6
0.8
?
1.0
1.2
1.4
Poisson MMSE
Kalman MMSE
Mean Field f
Stochastic f
LQG f
1.4
MMSE
0.90
1.6
b)
0.2
0.4
0.6
0.8
?
1.0
1.2
1.4
1.6
Figure 2: The differential allocation of resources in control and estimation for the OU process (left) and the
stochastic oscillator (right). Even though the estimation MMSE leads to a symmetric optimal encoder both in
the Poisson and in the Kalman filtering problem, the optimal encoders for the control problem are asymmetric,
allocating more resources to the first coordinate of the stimulus.
p2 Diag(tan(?), 0, cotan(?), 0) for the stochastic oscillator, where ? ? (0, ?/2). Note that this
? = 2?p?/(??)2 , independent of the specifics of the matrix P .
will yield the firing rate ?
We can then compare the performance of all observers with the same firing rate in both control and
estimation tasks. As mentioned, we are interested in control problems where the cost functions are
anisotropic, that is, one dimension of the system?s state vector contributes more heavily to the cost
function. To study this case we consider cost functions of the type
2
2
2
2
c(Xt , Ut ) = Q1 X1,t
+ Q2 X2,t
+ R1 U1,t
+ R2 U2,t
.
This again, can be readily cast into the formalism introduced above, with a suitable choice of matrices Q and R for both the OU process as for the stochastic oscillator. We will also consider the case
where the first dimension of Xt contributes more strongly to the state costs (i.e., Q1 > Q2 ).
The filtering error can be obtained from the formalism developed in [13] in the case of Poisson
observations and directly from the Kalman-Bucy equations in the case of Kalman filtering [18]. For
LQG control, one can simply solve the control problem for the system mentioned using the standard
methods (see e.g. [11]). The Poisson-coded version of the control problem can be solved using
either direct simulation of the dynamics of ?s or by a mean-field approach which has been shown
to yield excellent results for the system at hand. These results are summarised in figure 2, with
similar notation to that in figure 1. Note the extreme example of the stochastic oscillator, where the
optimal encoder is concentrating all the resources in one dimension, essentially ignoring the second
dimension.
4
Conclusion and Discussion
We have here shown that the optimal encoding strategies for a partially observed control problem is
not the same as the optimal encoding strategy for the associated state estimation problem. Note that
this is a natural consequence of considering noise covariances with a constant determinant in the
case of Kalman filtering and LQG control, but it is by no means trivial in the case of Poisson-coded
processes. For a class of stochastic processes for which the certainty equivalence principle holds we
have provided an exact expression for the optimal cost-to-go and have shown that minimising this
cost provides us with an encoder that in fact minimises the incurred cost in the control problem.
Optimality arguments are central to many parts of computational neuroscience, but it seems that
partial observability and the importance of combining adaptive state estimation and control have
rarely been considered in this literature, although supported by recent experiments. We believe the
present work, while treating only a small subset of the formalisms used in neuroscience, provides a
first insight into the differences between estimation and control. Much emphasis has been placed on
tracing the parallels between the two (see [19, 20], for example), but one must not forget to take into
account the differences as well.
8
References
[1] Jun Izawa and Reza Shadmehr. On-line processing of uncertain information in visuomotor
control. The Journal of neuroscience : the official journal of the Society for Neuroscience,
28(44):11360?8, October 2008.
[2] Emanuel Todorov and Michael I Jordan. Optimal feedback control as a theory of motor coordination. Nature neuroscience, 5(11):1226?35, November 2002.
[3] Peter W Battaglia and Paul R Schrater. Humans trade off viewing time and movement duration
to improve visuomotor accuracy in a fast reaching task. The Journal of neuroscience : the
official journal of the Society for Neuroscience, 27(26):6984?94, June 2007.
[4] Boris Rostislavovich Andrievsky, Aleksei Serafimovich Matveev, and Aleksandr L?vovich
Fradkov. Control and estimation under information constraints: Toward a unified theory of
control, computation and communications. Automation and Remote Control, 71(4):572?633,
2010.
[5] Hans S Witsenhausen. A counterexample in stochastic optimum control. SIAM Journal on
Control, 6(1):131?147, 1968.
[6] Edison Tse and Yaakov Bar-Shalom. An actively adaptive control for linear systems with
random parameters via the dual control approach. Automatic Control, IEEE Transactions on,
18(2):109?117, 1973.
[7] Yaakov Bar-Shalom and Edison Tse. Dual Effect, Certainty Equivalence, and Separation in
Stochastic Control. IEEE Transactions on Automatic Control, (5), 1974.
[8] D. Huber, D. A. Gutnisky, S. Peron, D. H. O?Connor, J. S. Wiegert, L. Tian, T. G. Oertner,
L. L. Looger, and K. Svoboda. Multiple dynamic representations in the motor cortex during
sensorimotor learning. Nature, 484(7395):473?478, Apr 2012. n2123 (unprinted).
[9] AA Mattar, Mohammad Darainy, David J Ostry, et al. Motor learning and its sensory effects:
time course of perceptual change and its presence with gradual introduction of load. J Neurophysiol, 109(3):782?791, 2013.
[10] S. Vahdat, M. Darainy, T.E. Milner, and D.J. Ostry. Functionally specific changes in restingstate sensorimotor networks after motor learning. J Neurosci, 31(47):16907?16915, 2011.
? om. Introdution to Stochastic Control Theory. Courier Dover Publications, Mine[11] Karl J. Astr?
ola, NY, 1st edition, 2006.
[12] Steve Yaeli and Ron Meir. Error-based analysis of optimal tuning functions explains phenomena observed in sensory neurons. Frontiers in computational neuroscience, 4(October):16,
2010.
[13] Alex Susemihl, Ron Meir, and Manfred Opper. Dynamic state estimation based on Poisson
spike trains - towards a theory of optimal encoding. Journal of Statistical Mechanics: Theory
and Experiment, 2013(03):P03009, March 2013.
[14] Fumio Kanaya and Kenji Nakagawa. On the practical implication of mutual information for
statistical decisionmaking. IEEE transactions on information theory, 37(4):1151?1156, 1991.
[15] N Merhav. Optimum estimation via gradients of partition functions and information measures:
a statistical-mechanical perspective. Information Theory, IEEE Transactions on, 57(6):3887?
3898, 2011.
[16] Dongning Guo, Shlomo Shamai, and Sergio Verd?u. Mutual information and minimum meansquare error in gaussian channels. Information Theory, IEEE Transactions on, 51(4):1261?
1282, 2005.
[17] David Attwell and Simon B Laughlin. An energy budget for signaling in the grey matter of the
brain. Journal of Cerebral Blood Flow & Metabolism, 21(10):1133?1145, 2001.
[18] R. S. Bucy. Nonlinear filtering theory. Automatic Control, IEEE Transactions, 10(2):198,
1965.
[19] Rudolph Emil Kalman. A new approach to linear filtering and prediction problems. Journal of
basic Engineering, 82(1):35?45, 1960.
[20] Emanuel Todorov. General duality between optimal control and estimation. 2008 47th IEEE
Conference on Decision and Control, (5):4286?4292, 2008.
9
| 5390 |@word determinant:3 version:1 achievable:1 seems:1 pillar:1 grey:1 seek:2 simulation:1 gradual:1 covariance:10 p0:3 meansquare:1 q1:2 tr:12 reduction:1 initial:3 selecting:2 interestingly:1 mmse:11 rightmost:2 current:2 nt:3 si:1 dx:2 must:5 written:1 pioneer:1 readily:2 tilted:2 numerical:1 partition:1 shlomo:1 plasticity:1 lqg:13 motor:8 shamai:1 treating:1 implying:1 intelligence:1 fewer:1 devising:1 metabolism:1 isotropic:1 beginning:1 dover:1 manfred:2 provides:4 ron:3 location:1 along:2 dn:1 direct:1 become:3 differential:5 consists:1 doubly:2 introduce:1 inter:1 huber:1 indeed:2 expected:8 behavior:1 themselves:1 mechanic:1 brain:1 bellman:1 informational:1 little:1 considering:6 increasing:2 provided:1 estimating:2 moreover:1 notation:3 null:1 what:1 sde:2 argmin:3 q2:2 developed:1 unified:2 finding:2 unobserved:3 certainty:4 thorough:1 pseudo:1 multidimensional:1 axt:1 universit:1 demonstrates:2 hypervolume:1 control:107 bst:1 engineering:2 treat:1 consequence:2 vahdat:1 encoding:5 aleksandr:1 analyzing:1 path:1 becoming:1 firing:7 might:1 emphasis:1 therein:1 k:1 equivalence:4 conversely:1 limited:1 tian:1 ola:1 obeys:1 practical:1 differs:2 communicated:1 signaling:1 thought:1 courier:1 word:1 cram:1 cannot:1 close:1 selection:3 parametrise:1 ast:1 equivalent:1 demonstrated:1 yt:6 center:1 go:8 independently:1 duration:3 estimator:1 insight:1 array:1 importantly:1 spanned:1 classic:1 population:11 coordinate:7 transmit:1 target:1 controlling:1 tan:2 heavily:1 exact:1 svoboda:1 milner:1 verd:1 hypothesis:1 origin:1 element:1 velocity:3 approximated:1 particularly:2 asymmetric:1 observed:12 electrical:1 solved:2 informationbased:1 region:1 remote:1 counter:1 trade:6 highest:1 decrease:1 movement:1 mentioned:2 constrains:1 mine:1 dynamic:9 depend:4 purely:1 neurophysiol:1 packing:1 looger:1 train:2 forced:1 fast:1 effective:2 artificial:1 visuomotor:2 choosing:2 exhaustive:1 whose:2 widely:1 solve:1 distortion:2 s:17 regeneration:1 encoder:16 cov:2 g1:1 think:1 jointly:1 noisy:5 itself:1 analyse:1 rudolph:1 interplay:1 sequence:1 emil:1 maximal:1 adaptation:9 remainder:1 loop:2 combining:1 riccati:1 flexibility:1 fired:2 insensitivity:1 forth:1 dirac:1 optimum:2 r1:1 decisionmaking:1 boris:1 derive:1 depending:1 illustrate:1 minimises:3 qt:5 p2:3 strong:1 kenji:1 involves:2 implies:1 differ:1 sensation:2 attribute:1 filter:4 stochastic:24 stabilization:1 human:1 viewing:1 explains:1 biological:5 secondly:1 frontier:1 clarify:1 hold:5 ds:5 pou:1 considered:4 exp:1 great:1 equilibrium:1 posc:1 major:1 battaglia:1 estimation:30 applicable:1 coordination:1 hope:1 clearly:1 sensor:8 gaussian:6 always:1 aim:1 rather:1 reaching:2 varying:1 publication:1 derived:1 june:1 consistently:4 rigorous:1 sense:5 dependent:5 relation:1 selects:1 interested:3 overall:2 issue:2 dual:2 denoted:2 special:1 mutual:8 field:6 oertner:1 nicely:1 broad:1 neigh:1 look:1 filling:1 future:9 minimized:1 stimulus:5 few:1 modern:1 densely:2 ourselves:1 dvt:2 interest:1 wiegert:1 expenditure:1 extreme:1 yielding:2 parametrised:1 devoted:1 damped:1 implication:2 kt:6 allocating:2 partial:3 necessary:1 incomplete:1 haifa:1 plotted:1 theoretical:1 minimal:1 uncertain:3 ostry:2 formalism:3 tse:2 steer:1 rao:1 markovian:1 cover:1 assignment:1 cost:34 technische:1 subset:1 costto:1 technion:1 encoders:3 bucy:3 st:15 density:1 siam:1 accessible:1 off:6 michael:1 kanaya:1 squared:2 central:3 again:2 choose:1 strive:2 leading:3 actively:1 account:3 potential:2 coding:7 automation:1 dnt:1 matter:1 depends:4 ornstein:1 performed:2 view:2 lot:1 closed:2 break:1 observing:2 observer:1 start:1 bayes:1 parallel:1 simon:1 contribution:2 minimize:1 om:1 accuracy:1 wiener:1 variance:4 yield:4 dealt:1 famous:1 bayesian:2 drive:1 cybernetics:1 history:1 qxs:2 energy:1 sensorimotor:2 frequency:3 naturally:1 associated:4 proof:1 gain:1 emanuel:2 concentrating:1 ut:7 ou:6 steve:1 higher:5 dt:6 reflected:1 formulation:2 though:3 strongly:2 furthermore:4 d:16 hand:3 nonlinear:3 google:1 glance:1 quality:2 believe:2 effect:3 concept:2 true:1 evolution:2 symmetric:1 illustrated:1 deal:2 parcel:1 during:1 width:3 covering:2 oc:3 criterion:1 presenting:1 theoretic:2 complete:1 mattar:1 mohammad:1 variational:1 harmonic:1 recently:1 ornsteinuhlenbeck:1 reza:1 cerebral:1 anisotropic:3 discussed:1 schrater:1 functionally:1 refer:1 significant:1 connor:1 counterexample:1 restingstate:1 subserve:1 tuning:18 grid:1 automatic:3 similarly:1 neatly:1 centre:6 astr:1 geared:1 han:1 cortex:1 etc:1 sergio:1 posterior:2 recent:4 perspective:1 optimizing:1 shalom:2 forcing:1 certain:4 affiliation:1 vt:2 seen:1 yaakov:2 minimum:1 care:1 steering:2 employed:1 determine:1 ntm:4 tempting:1 monotonically:1 smoother:1 multiple:3 full:1 minimising:2 long:1 equally:1 y:1 coded:3 controlled:2 impact:1 prediction:1 basic:1 controller:6 essentially:1 poisson:16 tailored:1 uhlenbeck:1 achieved:2 whereas:1 ode:1 singular:2 allocated:1 extra:1 flow:1 seem:1 jordan:1 counting:2 presence:1 todorov:2 bandwidth:1 restrict:1 observability:1 idea:1 computable:1 br:4 translates:1 minimise:5 t0:3 expression:3 allocate:3 ultimate:1 bridging:1 effort:1 peter:1 action:5 generally:3 clear:3 eigenvectors:1 se:2 dws:2 extensively:1 meir:3 neuroscience:11 estimated:2 delta:1 summarised:1 write:2 threshold:1 blood:1 ce:9 diffusion:1 v1:1 year:1 sum:1 inverse:1 yaeli:1 uncertainty:4 place:2 throughout:3 separation:5 decision:2 bit:1 bound:1 ct:1 ashby:1 followed:1 pay:1 quadratic:2 occur:1 constraint:5 precisely:2 alex:2 constrain:1 x2:3 phrased:1 aspect:1 simulate:1 argument:3 min:3 optimality:2 u1:1 separable:1 department:1 march:1 membrane:1 across:2 increasingly:1 making:2 restricted:3 gathering:1 equation:15 resource:8 turn:2 discus:1 count:1 studying:2 available:1 gaussians:1 actuator:3 observe:1 v2:1 appropriate:2 indirectly:1 simulating:1 dwt:3 matveev:1 alternative:1 original:1 assumes:1 denotes:1 neglect:1 giving:1 classical:1 society:2 objective:1 intend:1 already:1 occurs:1 spike:13 strategy:6 dependence:2 diagonal:1 gradient:1 separate:1 berlin:1 considers:1 cellular:1 trivial:1 reason:2 toward:1 assuming:1 ru:2 code:10 kalman:13 illustration:1 demonstration:1 setup:1 difficult:2 unfortunately:1 october:2 merhav:1 stated:1 design:2 dyp:1 zt:7 policy:5 reliably:1 perform:1 observation:23 neuron:13 finite:1 november:1 situation:1 communication:10 looking:1 sharp:1 community:2 overlooked:2 introduced:1 david:2 cast:1 required:2 mechanical:1 connection:2 plague:1 narrow:1 uncoupled:2 bar:2 usually:2 perception:3 dynamical:1 below:2 exemplified:1 biasing:1 optimise:1 belief:1 power:1 suitable:1 natural:2 force:1 scheme:1 improve:1 inversely:1 jun:1 review:1 literature:1 law:2 plant:2 dxt:2 gauged:1 interesting:2 generation:1 organised:1 filtering:14 proportional:1 analogy:1 allocation:1 incurred:2 agent:2 principle:5 metabolic:2 karl:1 course:1 summary:1 accounted:1 placed:2 last:1 surprisingly:1 supported:1 side:1 allow:1 laughlin:1 explaining:1 face:1 taking:1 absolute:1 tracing:1 distributed:5 curve:1 opper:2 boundary:1 world:2 stand:1 dimension:14 feedback:1 sensory:20 ignores:1 author:1 adaptive:2 transaction:6 approximate:1 keep:1 dealing:1 active:1 assumed:2 edison:2 alternatively:1 channel:5 nature:2 ignoring:1 contributes:2 mse:2 excellent:1 complex:2 constructing:1 gutnisky:1 diag:2 official:2 apr:1 dense:6 neurosci:1 noise:7 arise:1 paul:1 edition:1 x1:3 referred:2 ny:1 akt:1 n:2 precision:8 sub:4 fails:1 position:2 governed:1 perceptual:1 ricatti:1 weighting:1 down:1 load:1 specific:6 xt:35 showing:1 er:1 r2:1 x:14 concern:1 essential:2 exists:1 intractable:3 importance:3 conditioned:2 illustrates:1 budget:1 horizon:1 gap:2 forget:1 simply:8 peron:1 partially:3 u2:1 aa:1 conditional:2 formulated:1 careful:1 towards:5 oscillator:10 change:3 specifically:2 nakagawa:1 shadmehr:1 acting:1 averaging:1 susemihl:1 lemma:1 attwell:1 called:3 dyt:1 secondary:1 duality:1 experimental:1 disregard:1 gauss:3 rarely:1 select:1 guo:1 evaluate:2 d1:3 phenomenon:2 |
4,850 | 5,391 | The Large Margin Mechanism
for Differentially Private Maximization
Kamalika Chaudhuri
UC San Diego
La Jolla, CA
[email protected]
Daniel Hsu
Columbia University
New York, NY
[email protected]
Shuang Song
UC San Diego
La Jolla, CA
[email protected]
Abstract
A basic problem in the design of privacy-preserving algorithms is the private maximization problem: the goal is to pick an item from a universe that (approximately)
maximizes a data-dependent function, all under the constraint of differential privacy. This problem has been used as a sub-routine in many privacy-preserving
algorithms for statistics and machine learning.
Previous algorithms for this problem are either range-dependent?i.e., their utility
diminishes with the size of the universe?or only apply to very restricted function
classes. This work provides the first general purpose, range-independent algorithm for private maximization that guarantees approximate differential privacy.
Its applicability is demonstrated on two fundamental tasks in data mining and machine learning.
1
Introduction
Differential privacy [17] is a cryptographically motivated definition of privacy that has recently
gained significant attention in the data mining and machine learning communities. An algorithm
for processing sensitive data enforces differential privacy by ensuring that the likelihood of any
outcome does not change by much when a single individual?s private data changes. Privacy is
typically guaranteed by adding noise either to the sensitive data, or to the output of an algorithm
that processes the sensitive data. For many machine learning tasks, this leads to a corresponding
degradation in accuracy or utility. Thus a central challenge in differentially private learning is to
design algorithms with better tradeoffs between privacy and utility for a wide variety of statistics
and machine learning tasks.
In this paper, we study the private maximization problem, a fundamental problem that arises while
designing privacy-preserving algorithms for a number of statistical and machine learning applications. We are given a sensitive dataset D ? X n comprised of records from n individuals. We are
also given a data-dependent objective function f : U ? X n ? R, where U is a universe of K items
to choose from, and f (i, ?) is (1/n)-Lipschitz for all i ? U. That is, |f (i, D0 ) ? f (i, D00 )| ? 1/n for
all i and for any D0 , D00 ? X n differing in just one individual?s entry. Always selecting an item that
exactly maximizes f (?, D) is generally non-private, so the goal is to select, in a differentially private
manner, an item i ? U with as high an objective f (i, D) as possible. This is a very general algorithmic problem that arises in many applications, include private PAC learning [25] (choosing the most
accurate classifier), private decision tree induction [21] (choosing the most informative split), private
frequent itemset mining [5] (choosing the most frequent itemset), private validation [12] (choosing
the best tuning parameter), and private multiple hypothesis testing [32] (choosing the most likely
hypothesis).
The most common algorithms for this problem are the exponential mechanism [28], and a computationally efficient alternative from [5], which we call the max-of-Laplaces mechanism. These
1
algorithms are general?they do not require any additional conditions on f to succeed?and hence
have been widely applied. However, a major limitation of both algorithms is that their utility suffers from an explicit range-dependence: the utility deteriorates with increasing universe size. The
range-dependence persists even when there is a single clear maximizer of f (?, D), or a few near
maximizers, and even when the maximizer remains the same after changing the entries of a large
number of individuals in the data. Getting around range-dependence has therefore been a goal for
designing algorithms for this problem.
This problem has also been addressed by recent algorithms of [31, 3], who provide algorithms that
are range-independent and satisfy approximate differential privacy, a relaxed version of differential
privacy. However, none of these algorithms is general; they explicitly fail unless additional special
conditions on f hold. For example, the algorithm from [31] provides a range-independent result only
when there is a single clear maximizer i? such that f (i? , D) is greater than the second highest value
by some margin; the algorithm from [3] also has restrictive conditions that limit its applicability (see
Section 2.2). Thus, a challenge is to develop a private maximization algorithm that is both rangeindependent and free of additional conditions; this is necessary to ensure that an algorithm is widely
applicable and provides good utility when the universe size is large.
In this work, we provide the first such general purpose range-independent private maximization
algorithm. Our algorithm is based on two key insights. The first is that private maximization is
easier when there is a small set of near maximizing items j ? U for which f (j, D) is close to the
maximum value maxi?U f (i, D). A plausible algorithm based on this insight is to first find a set
of near maximizers, and then run the exponential mechanism on this set. However, finding this
set directly in a differentially private manner is very challenging. Our second insight is that only
the number ` of near maximizers needs to be found in a differentially private manner?a task that
is considerably easier. Provided there is a margin between the maximum value and the (` + 1)-th
maximum value of f (i, D), running the exponential mechanism on the items with the top ` values
of f (i, D) results in approximate differential privacy as well as good utility.
Our algorithm, which we call the large margin mechanism, automatically exploits large margins
when they exist to simultaneously (i) satisfy approximate differential privacy (Theorem 2), as well as
(ii) provide a utility guarantee that depends (logarithmically) only on the number of near maximizers,
rather than the universe size (Theorem 3). We complement our algorithm with a lower bound,
showing that the utility of any approximate differentially private algorithm must deteriorate with
the number of near maximizers (Theorem 1). A consequence of our lower bound is that rangeindependence cannot be achieved with pure differential privacy (Proposition 1), which justifies our
relaxation to approximate differential privacy.
Finally, we show the applicability of our algorithm to two problems from data mining and machine
learning: frequent itemset mining and private PAC learning. For the first problem, an application
of our method gives the first algorithm for frequent itemset mining that simultaneously guarantees
approximate differential privacy and utility independent of the itemset universe size. For the second
problem, our algorithm achieves tight sample complexity bounds for private PAC learning analogous
to the shell bounds of [26] for non-private learning.
2
Background
This section reviews differential privacy and introduces the private maximization problem.
2.1
Definitions of Differential Privacy and Private Maximization
For the rest of the paper, we consider randomized algorithms A : X n ? ?(S) that take as input
datasets D ? X n comprised of records from n individuals, and output values in a range S. Two
datasets D, D0 ? X n are said to be neighbors if they differ in a single individual?s entry. A function
? : X n ? R is L-Lipschitz if |?(D) ? ?(D0 )| ? L for all neighbors D, D0 ? X n .
The following definitions of (approximate) differential privacy are from [17] and [20].
Definition 1 (Differential Privacy). A randomized algorithm A : X n ? ?(S) is said to be (?, ?)approximate differentially private if, for all neighbors D, D0 ? X n and all S ? S,
Pr(A(D) ? S) ? e? Pr(A(D0 ) ? S) + ?.
2
The algorithm A is ?-differentially private if it is (?, 0)-approximate differentially private.
Smaller values of the privacy parameters ? > 0 and ? ? [0, 1] imply stronger guarantees of privacy.
Definition 2 (Private Maximization). In the private maximization problem, a sensitive dataset
D ? X n comprised of records from n individuals is given as input; there is also a universe
U := {1, . . . , K} of K items, and a function f : U ? X n ? R such that f (i, ?) is (1/n)-Lipschitz
for all i ? U. The goal is to return an item i ? U such that f (i, D) is as large as possible while
satisfying (approximate) differential privacy.
Always returning the exact maximizer of f (?, D) is non-private, as changing a single individuals?
private values can potentially change the maximizer. Our goal is to design a randomized algorithm
that outputs an approximate maximizer with high probability. (We loosely refer to the expected
f (?, D) value of the chosen item as the utility of the algorithm.)
Note that this problem is different from private release of the maximum value of f (?, D); a solution for the latter is easily obtained by adding Laplace noise with standard deviation O(1/(?n)) to
maxi?U f (i, D) [17]. Privately returning a nearly maximizing item itself is much more challenging.
Private maximization is a core problem in the design of differentially private algorithms, and arises
in numerous statistical and machine learning tasks. The examples of frequent itemset mining and
PAC learning are discussed in Sections 4.1 and 4.2.
2.2
Previous Algorithms for Private Maximization
The standard algorithm for private maximization is the exponential mechanism [28]. Given a privacy
parameter ? > 0, the exponential mechanism randomly draws an item i ? U with probability
pi ? en?f (i,D)/2 ; this guarantees ?-differential privacy. While the exponential mechanism is widely
used because of its generality, a major limitation is its range-dependence?i.e., its utility diminishes
with the universe size K. To be more precise, consider the following example where X := U = [K]
and
1
f (i, D) := |{j ? [n] : Dj ? i}|
(1)
n
(where Dj is the j-th entry in the dataset D). When D = (1, 1, . . . , 1), there is a clear maximizer
i? = 1, which only changes when the entries of at least n/2 individuals in D change. It stands
to reason that any algorithm should report i = 1 in this case with high probability. However, the
exponential mechanism outputs i = 1 only with probability en?/2 /(K ? 1 + en?/2 ), which is small
unless n = ?(log(K)/?). This implies that the utility of the exponential mechanism deteriorates
with K.
Another general purpose algorithm is the max-of-Laplaces mechanism from [5]. Unfortunately, this
algorithm is also range-dependent. Indeed, our first observation is that all ?-differentially private
algorithms that succeed on a wide class of private maximization problems share this same drawback.
Proposition 1 (Lower bound for differential privacy). Let A be any ?-differentially private algorithm for private maximization, ? ? (0, 1), and n ? 2. There exists a domain X , a function
f : U ? X n ? R such that f (i, ?) is (1/n)-Lipschitz for all i ? U, and a dataset D ? X n such that:
!
log K?1
1
2
Pr f (A(D), D) > max f (i, D) ?
< .
i?U
?n
2
We remark that results similar to Proposition 1 have appeared in [23, 2, 10, 11, 7]; we simply reframe those results here in the context of private maximization.
Proposition 1 implies that in order to remove range-dependence, we need to relax the privacy notion.
We consider a relaxation of the privacy constraint to (?, ?)-approximate differential privacy with
? > 0.
The approximate differentially private algorithm from [31] applies in the case where there is a single
clear maximizer whose value is much larger than that of the rest. This algorithm adds Laplace noise
with standard deviation O(1/(?n)) to the difference between the largest and the second-largest values of f (?, D), and outputs the maximizer if this noisy difference is larger than O(log(1/?)/(?n));
3
otherwise, it outputs Fail. Although this solution has high utility for the example in (1) with
D = (1, 1, . . . , 1), it fails even when there is a single additional item j ? U with f (j, D) close to
the maximum value; for instance, D = (2, 2, . . . , 2).
[3] provides an approximate differentially private algorithm that applies when f satisfies a condition
called `-bounded growth. This condition entails the following: first, for any i ? U, adding a single
individual to any dataset D can either keep f (i, D) constant, or increase it by 1/n; and second,
f (i, D) can only increase in this case for at most ` items i ? U. The utility of this algorithm depends
only on log `, rather than log K. In contrast, our algorithm does not require the first condition.
Furthermore, to ensure that our algorithm only depends on log `, it suffices that there only be ?`
near maximizers, which is substantially less restrictive than the `-bounded growth condition.
As mentioned earlier, we avoid range-dependence with an algorithm that finds and optimizes over
near maximizers of f (?, D). We next specify what we mean by near maximizers using a notion of
margin.
3
The Large Margin Mechanism
We now our new algorithm for private maximization, called the large margin mechanism, along with
its privacy and utility guarantees.
3.1
Margins
We first introduce the notion of margin on which our algorithm is based. Given an instance of the
private maximization problem and a positive integer ` ? N, let f (`) (D) denote the `-th highest value
of f (?, D). We adopt the convention that f (K+1) (D) = ??.
Condition 1 ((`, ?)-margin condition). For any ` ? N and ? > 0, we say a dataset D ? X n satisfies
the (`, ?)-margin condition if
f (`+1) (D) < f (1) (D) ? ?
(i.e., there are at most ` items within ? of the top item according to f (?, D)).1
By convention, every dataset satisfies the (K, ?)-margin condition. Intuitively, a (`, ?)-margin condition with a relatively large ? implies that there are ?` near maximizers, so the private maximization
problem is easier when D satisfies an (`, ?)-margin condition with small `.
How large should ? be for a given `? The following lower bound suggests that in order to have
n = O(log(`)/?), we need ? to be roughly log(`)/(?n).
Theorem 1 (Lower bound for approximate differential privacy). Fix any ? ? (0, 1), ` > 1, and ? ?
[0, (1 ? exp(??))/(2(` ? 1))]; and assume n ? 2. Let A be any (?, ?)-approximate differentially
private algorithm, and ? := min{1/2, log((` ? 1)/2)/(n?)}. There exists a domain X , a function
f : U ? X n ? R such that f (i, ?) is (1/n)-Lipschitz for all i ? U, and a dataset D ? X n such that:
1. D satisfies the (`, ?)-margin condition.
1
2. Pr f (A(D), D) > f (1) (D) ? ? < .
2
A consequence of Theorem 1 is that complete range-independence for all (1/n)-Lipschitz functions f is not possible, even with approximate differential privacy. For instance, if D satisfies an
(`, ?(log(`)/(?n)))-margin condition only when ` = ?(K), then n must be ?(log(K)/?) in order
for an approximate differentially private algorithm to be useful.
3.2
Algorithm
The lower bound in Theorem 1 suggests the following algorithm. First, privately determine a pair
(`, ?), with ` is as small as possible and ? = ?(log(`)/(?n)), such that D satisfies the (`, ?)-margin
1
Our notion of margins here is different from the usual notion of margins from statistical learning that
underlies linear prediction methods like support vector machines and boosting. In fact, our notion is more
closely related to the shell decomposition bounds of [26], which we discuss in Section 4.2.
4
Algorithm 1 The large margin mechanism LMM(?, ?, D)
input Privacy parameters ? > 0 and ? ? (0, 1), database D ? X n .
output Item I ? U.
1: For each r = 1, 2, . . . , K, let
6
ln(3r/?)
1
1
r
(r)
t :=
1+
=O
,
+
log
n
?
n n?
?
3
6
3
12
3r(r + 1)
1
r
3
1
(r)
(r)
T :=
ln
+
ln +
ln
+t =O
+
log
.
n? 2?
n? ?
n?
?
n n?
?
2: Draw Z ? Lap(3/?).
3: Let m := f (1) (D) + Z/n. {Estimate of max value.}
iid
Draw G ? Lap(6/?) and Z1 , Z2 , . . . , ZK?1 ? Lap(12/?).
Let ` := 1. {Adaptively determine value ` such that D satisfies (`, t(`) )-margin condition.}
while ` < K do
if m ? f (`+1) (D) > (Z` + G)/n + T (`) then
Break out of while-loop with current value of `.
else
Let ` := ` + 1.
end if
end while
Let U` be the set of ` items in U with highest f (i, D) value (ties broken arbitrarily).
Draw I ? p where pi ? 1{i ? U` } exp(n?f (i, D)/6). {Exponential mechanism on top `
items.}
15: return I.
4:
5:
6:
7:
8:
9:
10:
11:
12:
13:
14:
condition. Then, run the exponential mechanism on the set U` ? U of items with the ` highest
f (?, D) values. This sounds rather natural and simple, but a knee-jerk reaction to this approach is
that the set U` itself depends on the sensitive dataset D, and it may have high sensitivity in the sense
that membership of many items in U` can change when a single individual?s private value is changed.
Thus differentially private computation of U` appears challenging.
It turns out we do not need to guarantee the privacy of the set U` , but rather just of a valid (`, ?)
pair. This is essentially because when D satisfies the (`, ?)-margin condition, the probability that
the exponential mechanism picks an item i that occurs in U` when the sensitive dataset is D but not
in U` when the sensitive dataset is its neighbor D0 is very small.
Moreover, we can find such a valid (`, ?) pair using a differentially private search procedure based
on the sparse vector technique [22]. Combining these ideas gives a general (and adaptive) algorithm whose loss of utility due to privacy is only O(log(`/?)/?n) when the dataset satisfies a
(`, O(log(`/?)/(?n))-margin condition. We call this general algorithm the large margin mechanism (Algorithm 1), or LMM for short.
3.3
Privacy and Utility Guarantees
We first show that LMM satisfies approximate differential privacy.
Theorem 2 (Privacy guarantee). LMM(?, ?, ?) satisfies (?, ?)-approximate differential privacy.
The proof of Theorem 2 is in Appendix A.1. The following theorem, proved in Appendix A.2,
provides a guarantee on the utility of LMM.
Theorem 3 (Utility guarantee). Pick any ? ? (0, 1). Suppose D ? X n satisfies the (`? , ? ? )-margin
condition with
?
21
3
ln + T (` ) .
?? =
n? ?
Then with probability at least 1 ? ?, I := LMM(?, ?, D) satisfies
f (I, D) ? f (1) (D) ?
5
6 ln(2`? /?)
.
n?
?
(Above, T (` ) is as defined in Algorithm 1.)
Remark 1. Fix some ?, ? ? (0, 1). Theorem 3 states that if the dataset D satisfies the (`? , ? ? )margin condition, for some positive integer `? and ? ? = C log(`? /?)/(n?) for some universal
constant C > 0, then the value f (I, D) of the item I returned by LMM is within O(log(`? )/(n?))
of the maximum, with high probability. There is no explicit dependence on the cardinality K of the
universe U.
4
Illustrative Applications
We now describe applications of LMM to problems from data mining and machine learning.
4.1
Private Frequent Itemset Mining
Frequent Itemset Mining (FIM) is the following popular data mining problem: given the purchase
lists of users (say, for an online grocery store), the goal is to find the sets of items that are purchased together most often. The work of [5] provides the first differentially private algorithms
for FIM. However, as these algorithms rely on the exponential mechanism and the max-of-Laplaces
mechanism, their utilities degrade with the total number of possible itemsets. Subsequent algorithms
exploit other properties of itemsets or avoid directly finding the most frequent itemset [34, 27, 15, 8].
Let I be the set of items that can be purchased, and let B be the maximum length of an user?s
purchase list. Let U ? 2I be the family of itemsets of interest. For simplicity, we let U := Ir ?
i.e., all itemsets of size r?and consider the problem of picking the itemset with the (approximately)
highest frequency. This is a private maximization problem where D is the users? lists of purchased
items, and f (i, D) is the fraction of users who purchase an itemset i ? U. Let fmax be the highest
frequency of an itemset in D. Let L be the total number of itemsets with non-zero frequency, so
L ? n Br , which is |I|r whenever B |I|. Applying LMM gives the following guarantee.
Corollary 1. Suppose we use LMM(?, ?, ?) on the FIM problem above. Then there exists a constant
C > 0 such that the following holds. If fmax ? C ? log(L/?)/(n?), then with probability ? 1 ? ?,
the frequency of the itemset ILMM output by LMM is
log(L/?)
f (ILMM , D) ? fmax ? O
.
n?
In contrast, the itemset IEM returned by the exponential mechanism is only guaranteed to satisfy
r log(|I|/?)
f (IEM , D) ? fmax ? O
,
n?
which is significantly worse than Corollary 1 whenever L |I|r (as is typically the case). Second,
to ensure differential privacy by running the exponential mechanism, one needs a priori knowledge
of the set U (and thus the universe of items I) independently of the observed data; otherwise the
process will not be end-to-end differentially private. In contrast, our algorithm does not need to
know I in order to provide end-to-end differential privacy. Finally, unlike [31], our algorithm does
not require a gap between the top two itemset frequencies.
4.2
Private PAC Learning
We now consider private PAC learning with a finite hypothesis class H with bounded VC dimension
d [25]. Here, the dataset D consists of n labeled training examples drawn iid from a fixed distribution. The error err(h) of a hypothesis h ? H is the probability that it misclassifies a random
example drawn from the same distribution. The goal is to return a hypothesis h ? H with error
as low as possible. A standard procedure that has been well-studied in the literature simply returns
? ? H of the empirical error err(h,
the minimizer h
c
D) computed on the training data D, but this
does not guarantee (approximate) differential privacy. The work of [25] instead uses the exponential
mechanism to select a hypothesis hEM ? H. With probability ? 1 ? ?0 ,
!
r
d log(n/?0 ) log |H| + log(1/?0 )
+
.
(2)
err(hEM ) ? min err(h) + O
h?H
n
?n
6
The dependence on log |H| is improved to d log |?| by [7] when the data entries come from a finite set ?. The subsequent work of [4] introduces the notion of representation dimension, and
shows how it relates to differentially private learning in the discrete and finite case, and [3] provides improved convergence bounds with approximate differential privacy that exploit the structure
of some specific hypothesis classes. For the case of infinite hypothesis classes and continuous data
distributions, [10] shows that distribution-free private PAC learning is not generally possible, but
distribution-dependent learning can be achieved under certain conditions.
We provide a sample complexity bound of a rather different character compared to previous work.
Our bound only relies on uniform convergence properties of H, and can be significantly tighter
than the bounds from [25] when the number of hypotheses with error close to minh?H err(h) is
small. Indeed, the bounds are a private analogue of the shell bounds of [26], which characterize
the structure of the hypothesis class as a function of the properties of a decomposition based on
hypotheses? error rates. In many situation, these bounds are significantly tighter than those that do
not involve the error distributions.
p
Following [26], we divide the hypothesis class H into R = O( n/(d log n)) shells; the t-th shell
H(t) is defined by
(
)
r
d log(n/?0 )
0
H(t) := h ? H : err(h) ? min
err(h ) + C0 t
.
h0 ?H
n
Above, C0 > 0 is the constant from uniform convergence bounds?i.e., C0 is the smallest
c > 0 such
p
that for all h ? H, with probability ? 1 ? ?0 , we have |err(h,
c
D) ? err(h)| ? c d log(n/?0 )/n.
Observe that H(t + 1) ? H(t); and moreover,
p with probability ? 1 ? ?0 , all h ? H(t) have
0
err(h,
c
D) ? minh0 ?H err(h ) + C0 ? (t + 1) d log(n/?0 )/n.
Let t? (n) as the smallest integer t ? N such that
?
log(|H(t + 1)|) + log(1/?)
C0 ? dn log n
?
t
C
where C > 0 is the constant from Remark 1. Then, with probability ? 1 ? ?0 , the dataset D with
f = 1? err
c satisfies the (`, ?)-margin condition, with ` = |H(t? (n)+1)| and ? = C log(|H(t? (n)+
1)|/?)/(?n). Therefore, we have the following guarantee for applying LMM to this problem.
Corollary 2. Suppose we use LMM(?, ?, ?) on the learning problem above (with U = H and f =
1 ? err).
c Then, with probability ? 1 ? ?0 ? ?, the hypothesis hLMM returned by LMM satisfies
!
r
d log(n/?0 ) log(|H(t? (n) + 1)|/?)
err(hLMM ) ? min err(h) + O
+
.
h?H
n
?n
The dependence on log |H| from (2) is replaced here by log(|H(t? (n) + 1)|/?), which can be vastly
smaller, as discussed in [26].
5
Additional Related Work
There has been a large amount of work on differential privacy for a wide range of statistical and machine learning tasks over the last decade [6, 30, 13, 21, 33, 24, 1]; for overviews, see [18] and [29]. In
particular, algorithms for the private maximization problem (and variants) have been used as subroutines in many applications; examples include PAC learning [25], principle component analysis [14],
performance validation [12], and multiple hypothesis testing [32].
A separation between pure and approximate differential privacy has been shown in several previous
works [19, 31, 3]. The first approximate differentially private algorithm that achieves a separation is
the Propose-Test-Release (PTR) framework [19]. Given a function, PTR determines an upper bound
on its local sensitivity at the input dataset through a search procedure; noise proportional to this
upper bound is then added to the actual function value. We note that the PTR framework does not
directly apply to our setting as the sensitivity is not generally defined for a discrete universe.
In the context of private PAC learning, the work of [3] gives the first separation between pure and
approximate differential privacy. In addition to using the algorithm from [31], they devise two
7
additional algorithmic techniques: a concave maximization procedure for learning intervals, and an
algorithm for the private maximization problem under the `-bounded growth condition discussed
in Section 2.2. The first algorithm is specific to their problem and does not appear to apply to
general private maximization problems. The second algorithm has a sample complexity bound of
n = O(log(`)/?) when the function f satisfies the `-bounded growth condition.
Lower bounds for approximate differential privacy have been shown by [7, 16, 11, 9], and the proof
of our Theorem 1 borrows some techniques from [11].
6
Conclusion and Future Work
In this paper, we have presented the first general and range-independent algorithm for approximate
differentially private maximization. The algorithm automatically adapts to the available large margin
properties of the sensitive dataset, and reverts to worst-case guarantees when such properties are
lacking. We have illustrated the applicability of the algorithm in two fundamental problems from
data mining and machine learning; in future work, we plan to study other applications where rangeindependence is a substantial boon.
Acknowledgments. We thank an anonymous reviewer for suggesting the simpler variant of LMM
based on the exponential mechanism. (The original version of LMM used a max of truncated exponentials mechanism, which gives the same guarantees up to constant factors.) This work was
supported in part by the NIH under U54 HL108460 and the NSF under IIS 1253942.
References
[1] Raef Bassily, Adam Smith, and Abhradeep Thakurta. Private empirical risk minimization,
revisited. arXiv:1405.7085, 2014.
[2] Amos Beimel, Shiva Prasad Kasiviswanathan, and Kobbi Nissim. Bounds on the sample complexity for private learning and private data release. In Theory of Cryptography, pages 437?
454. Springer, 2010.
[3] Amos Beimel, Kobbi Nissim, and Uri Stemmer. Private learning and sanitization: Pure vs.
approximate differential privacy. In RANDOM, 2013.
[4] Amos Beimel, Kobbi Nissim, and Uri Stemmer. Characterizing the sample complexity of
private learners. In ITCS, pages 97?110, 2013.
[5] Raghav Bhaskar, Srivatsan Laxman, Adam Smith, and Abhradeep Thakurta. Discovering frequent patterns in sensitive data. In KDD, 2010.
[6] A. Blum, C. Dwork, F. McSherry, and K. Nissim. Practical privacy: the SuLQ framework. In
PODS, 2005.
[7] Avrim Blum, Katrina Ligett, and Aaron Roth. A learning theory approach to noninteractive
database privacy. Journal of the ACM, 60(2):12, 2013.
[8] Luca Bonomi and Li Xiong. Mining frequent patterns with differential privacy. Proceedings
of the VLDB Endowment, 6(12):1422?1427, 2013.
[9] Mark Bun, Jonathan Ullman, and Salil Vadhan. Fingerprinting codes and the price of approximate differential privacy. In STOC, 2014.
[10] Kamalika Chaudhuri and Daniel Hsu. Sample complexity bounds for differentially private
learning. In COLT, 2011.
[11] Kamalika Chaudhuri and Daniel Hsu. Convergence rates for differentially private statistical
estimation. In ICML, 2012.
[12] Kamalika Chaudhuri and Staal A Vinterbo. A stability-based validation procedure for differentially private machine learning. In Advances in Neural Information Processing Systems, pages
2652?2660, 2013.
[13] Kamalika Chaudhuri, Claire Monteleoni, and Anand D. Sarwate. Differentially private empirical risk minimization. Journal of Machine Learning Research, 12:1069?1109, 2011.
8
[14] Kamalika Chaudhuri, Anand D. Sarwate, and Kaushik Sinha. Near-optimal differentially private principal components. In Advances in Neural Information Processing Systems, pages
998?1006, 2012.
[15] Rui Chen, Noman Mohammed, Benjamin CM Fung, Bipin C Desai, and Li Xiong. Publishing
set-valued data via differential privacy. In VLDB, 2011.
[16] Anindya De. Lower bounds in differential privacy. In Ronald Cramer, editor, Theory of
Cryptography, volume 7194 of Lecture Notes in Computer Science, pages 321?338. SpringerVerlag, 2012.
[17] C. Dwork, F. McSherry, K. Nissim, and A. Smith. Calibrating noise to sensitivity in private
data analysis. In Theory of Cryptography, 2006.
[18] Cynthia Dwork. Differential privacy: A survey of results. In Theory and Applications of
Models of Computation, pages 1?19. Springer, 2008.
[19] Cynthia Dwork and Jing Lei. Differential privacy and robust statistics. In Proceedings of the
41st annual ACM symposium on Theory of computing, pages 371?380. ACM, 2009.
[20] Cynthia Dwork, Krishnaram Kenthapadi, Frank McSherry, Ilya Mironov, and Moni Naor. Our
data, ourselves: Privacy via distributed noise generation. In EURO-CRYPT. 2006.
[21] A. Friedman and A. Schuster. Data mining with differential privacy. In KDD, 2010.
[22] Moritz Hardt and Guy N Rothblum. A multiplicative weights mechanism for privacypreserving data analysis. In FOCS, 2010.
[23] Moritz Hardt and Kunal Talwar. On the geometry of differential privacy. In Proceedings of the
42nd ACM symposium on Theory of computing, pages 705?714. ACM, 2010.
[24] Prateek Jain, Pravesh Kothari, and Abhradeep Thakurta. Differentially private online learning.
In COLT, 2012.
[25] Shiva Prasad Kasiviswanathan, Homin K Lee, Kobbi Nissim, Sofya Raskhodnikova, and Adam
Smith. What can we learn privately? SIAM Journal on Computing, 40(3):793?826, 2011.
[26] John Langford and David McAllester. Computable shell decomposition bounds. J. Mach.
Learn. Res., 5:529?547, 2004.
[27] Ninghui Li, Wahbeh Qardaji, Dong Su, and Jianneng Cao. Privbasis: frequent itemset mining
with differential privacy. In VLDB, 2012.
[28] Frank McSherry and Kunal Talwar. Mechanism design via differential privacy. In FOCS, 2007.
[29] A.D. Sarwate and K. Chaudhuri. Signal processing and machine learning with differential
privacy: Algorithms and challenges for continuous data. Signal Processing Magazine, IEEE,
30(5):86?94, Sept 2013. ISSN 1053-5888. doi: 10.1109/MSP.2013.2259911.
[30] Adam Smith. Privacy-preserving statistical estimation with optimal convergence rates. In
STOC, 2011.
[31] Adam Smith and Abhradeep Thakurta. Differentially private feature selection via stability
arguments, and the robustness of the lasso. In COLT, 2013.
[32] Caroline Uhler, Aleksandra B. Slavkovic, and Stephen E. Fienberg. Privacy-preserving data
sharing for genome-wide association studies. arXiv:1205.0739, 2012.
[33] Larry Wasserman and Shuheng Zhou. A statistical framework for differential privacy. Journal
of the American Statistical Association, 105(489):375?389, 2010.
[34] Chen Zeng, Jeffrey F Naughton, and Jin-Yi Cai. On differentially private frequent itemset
mining. In VLDB, 2012.
9
| 5391 |@word private:80 version:2 stronger:1 nd:1 c0:5 bun:1 vldb:4 prasad:2 eng:1 decomposition:3 pick:3 selecting:1 daniel:3 reaction:1 err:15 current:1 z2:1 must:2 john:1 ronald:1 subsequent:2 informative:1 kdd:2 remove:1 ligett:1 v:1 discovering:1 item:26 smith:6 core:1 short:1 record:3 provides:7 boosting:1 revisited:1 kasiviswanathan:2 simpler:1 along:1 dn:1 differential:44 symposium:2 focs:2 consists:1 naor:1 introduce:1 manner:3 privacy:66 deteriorate:1 shuheng:1 indeed:2 expected:1 roughly:1 automatically:2 actual:1 cardinality:1 increasing:1 provided:1 bounded:5 moreover:2 maximizes:2 what:2 prateek:1 cm:1 substantially:1 differing:1 finding:2 kenthapadi:1 guarantee:16 every:1 concave:1 growth:4 tie:1 exactly:1 returning:2 classifier:1 appear:1 laxman:1 positive:2 persists:1 local:1 limit:1 consequence:2 mach:1 approximately:2 rothblum:1 itemset:17 studied:1 suggests:2 challenging:3 range:16 acknowledgment:1 practical:1 enforces:1 testing:2 procedure:5 universal:1 empirical:3 significantly:3 cannot:1 close:3 selection:1 bonomi:1 context:2 applying:2 risk:2 raskhodnikova:1 demonstrated:1 reviewer:1 maximizing:2 roth:1 attention:1 independently:1 pod:1 survey:1 simplicity:1 knee:1 pure:4 wasserman:1 mironov:1 insight:3 stability:2 notion:7 beimel:3 laplace:5 analogous:1 diego:2 suppose:3 user:4 exact:1 magazine:1 us:1 designing:2 hypothesis:14 kunal:2 logarithmically:1 satisfying:1 database:2 labeled:1 observed:1 itemsets:5 worst:1 desai:1 highest:6 mentioned:1 substantial:1 benjamin:1 broken:1 complexity:6 salil:1 tight:1 learner:1 easily:1 jain:1 describe:1 doi:1 outcome:1 choosing:5 h0:1 whose:2 widely:3 plausible:1 larger:2 say:2 relax:1 otherwise:2 katrina:1 valued:1 statistic:3 raef:1 itself:2 noisy:1 online:2 cai:1 propose:1 frequent:12 cao:1 loop:1 combining:1 fmax:4 chaudhuri:7 adapts:1 differentially:31 getting:1 convergence:5 jing:1 adam:5 develop:1 c:2 implies:3 come:1 convention:2 differ:1 drawback:1 closely:1 slavkovic:1 vc:1 mcallester:1 larry:1 require:3 suffices:1 fix:2 anonymous:1 d00:2 proposition:4 tighter:2 hold:2 around:1 cramer:1 exp:2 algorithmic:2 major:2 achieves:2 adopt:1 smallest:2 purpose:3 estimation:2 diminishes:2 applicable:1 pravesh:1 thakurta:4 sensitive:10 largest:2 djhsu:1 amos:3 minimization:2 always:2 rather:5 avoid:2 zhou:1 corollary:3 release:3 likelihood:1 contrast:3 sense:1 dependent:5 membership:1 typically:2 sulq:1 subroutine:1 colt:3 priori:1 grocery:1 misclassifies:1 special:1 plan:1 uc:2 icml:1 nearly:1 purchase:3 future:2 report:1 few:1 krishnaram:1 randomly:1 simultaneously:2 individual:11 replaced:1 geometry:1 ourselves:1 jeffrey:1 iem:2 friedman:1 uhler:1 interest:1 mining:16 dwork:5 introduces:2 mcsherry:4 cryptographically:1 accurate:1 necessary:1 unless:2 tree:1 loosely:1 divide:1 re:1 sinha:1 instance:3 earlier:1 maximization:26 applicability:4 deviation:2 entry:6 uniform:2 comprised:3 shuang:1 hem:2 characterize:1 reframe:1 considerably:1 adaptively:1 st:1 fundamental:3 randomized:3 sensitivity:4 siam:1 lee:1 dong:1 picking:1 together:1 ilya:1 vastly:1 central:1 choose:1 guy:1 worse:1 american:1 kobbi:4 return:4 ullman:1 li:3 suggesting:1 de:1 satisfy:3 explicitly:1 depends:4 multiplicative:1 break:1 ir:1 accuracy:1 who:2 sofya:1 itcs:1 iid:2 none:1 caroline:1 suffers:1 monteleoni:1 sharing:1 whenever:2 definition:5 crypt:1 frequency:5 proof:2 hsu:3 dataset:17 proved:1 popular:1 hardt:2 knowledge:1 routine:1 appears:1 specify:1 improved:2 generality:1 furthermore:1 just:2 langford:1 su:1 zeng:1 maximizer:9 lei:1 calibrating:1 hence:1 moritz:2 illustrated:1 kaushik:1 illustrative:1 ptr:3 complete:1 recently:1 nih:1 common:1 overview:1 volume:1 sarwate:3 discussed:3 association:2 significant:1 refer:1 tuning:1 dj:2 entail:1 moni:1 add:1 recent:1 jolla:2 optimizes:1 store:1 certain:1 mohammed:1 arbitrarily:1 yi:1 devise:1 preserving:5 additional:6 relaxed:1 greater:1 determine:2 lmm:16 signal:2 ii:2 relates:1 multiple:2 sound:1 stephen:1 d0:8 luca:1 ensuring:1 prediction:1 underlies:1 basic:1 variant:2 essentially:1 arxiv:2 achieved:2 abhradeep:4 background:1 addition:1 addressed:1 interval:1 else:1 rest:2 unlike:1 privacypreserving:1 anand:2 bhaskar:1 call:3 integer:3 vadhan:1 near:11 anindya:1 split:1 variety:1 independence:1 jerk:1 shiva:2 lasso:1 idea:1 tradeoff:1 br:1 computable:1 motivated:1 utility:21 fim:3 song:1 returned:3 york:1 remark:3 generally:3 useful:1 clear:4 involve:1 amount:1 exist:1 nsf:1 deteriorates:2 aleksandra:1 discrete:2 key:1 blum:2 drawn:2 changing:2 relaxation:2 fraction:1 naughton:1 run:2 talwar:2 family:1 separation:3 draw:4 decision:1 appendix:2 bound:25 guaranteed:2 annual:1 constraint:2 argument:1 min:4 relatively:1 fung:1 according:1 smaller:2 character:1 intuitively:1 restricted:1 pr:4 fienberg:1 computationally:1 ln:6 remains:1 discus:1 turn:1 mechanism:28 fail:2 know:1 msp:1 end:6 available:1 apply:3 observe:1 xiong:2 alternative:1 srivatsan:1 robustness:1 original:1 top:4 running:2 ensure:3 include:2 publishing:1 exploit:3 restrictive:2 purchased:3 objective:2 added:1 occurs:1 dependence:9 usual:1 said:2 thank:1 degrade:1 bipin:1 nissim:6 reason:1 induction:1 raghav:1 length:1 code:1 issn:1 unfortunately:1 potentially:1 stoc:2 frank:2 design:5 upper:2 observation:1 kothari:1 datasets:2 finite:3 minh:1 jin:1 truncated:1 situation:1 precise:1 ucsd:2 fingerprinting:1 community:1 david:1 complement:1 pair:3 z1:1 pattern:2 appeared:1 challenge:3 reverts:1 max:6 analogue:1 natural:1 rely:1 imply:1 numerous:1 columbia:2 sept:1 review:1 literature:1 lacking:1 loss:1 lecture:1 generation:1 limitation:2 proportional:1 borrows:1 validation:3 boon:1 principle:1 editor:1 pi:2 share:1 endowment:1 claire:1 changed:1 supported:1 last:1 free:2 wide:4 neighbor:4 stemmer:2 characterizing:1 sparse:1 distributed:1 dimension:2 stand:1 valid:2 genome:1 adaptive:1 san:2 u54:1 approximate:30 vinterbo:1 keep:1 search:2 continuous:2 decade:1 learn:2 zk:1 robust:1 ca:2 ninghui:1 sanitization:1 domain:2 universe:12 privately:3 noise:6 cryptography:3 euro:1 en:3 bassily:1 ny:1 sub:1 fails:1 explicit:2 exponential:17 theorem:12 specific:2 pac:9 showing:1 cynthia:3 maxi:2 list:3 maximizers:9 exists:3 avrim:1 adding:3 kamalika:7 gained:1 justifies:1 uri:2 margin:29 rui:1 gap:1 easier:3 chen:2 homin:1 lap:3 simply:2 likely:1 applies:2 springer:2 minimizer:1 satisfies:18 relies:1 determines:1 acm:5 shell:6 succeed:2 goal:7 lipschitz:6 price:1 change:6 springerverlag:1 infinite:1 degradation:1 principal:1 called:2 total:2 la:2 aaron:1 select:2 support:1 mark:1 latter:1 arises:3 jonathan:1 schuster:1 |
4,851 | 5,392 | Extremal Mechanisms for Local Differential Privacy
Peter Kairouz1
Sewoong Oh2
Pramod Viswanath1
1
Department of Electrical & Computer Engineering
2
Department of Industrial & Enterprise Systems Engineering
University of Illinois Urbana-Champaign
Urbana, IL 61801, USA
{kairouz2,swoh,pramodv}@illinois.edu
Abstract
Local differential privacy has recently surfaced as a strong measure of privacy
in contexts where personal information remains private even from data analysts.
Working in a setting where the data providers and data analysts want to maximize
the utility of statistical inferences performed on the released data, we study the
fundamental tradeoff between local differential privacy and information theoretic
utility functions. We introduce a family of extremal privatization mechanisms,
which we call staircase mechanisms, and prove that it contains the optimal privatization mechanism that maximizes utility. We further show that for all information
theoretic utility functions studied in this paper, maximizing utility is equivalent
to solving a linear program, the outcome of which is the optimal staircase mechanism. However, solving this linear program can be computationally expensive
since it has a number of variables that is exponential in the data size. To account
for this, we show that two simple staircase mechanisms, the binary and randomized response mechanisms, are universally optimal in the high and low privacy
regimes, respectively, and well approximate the intermediate regime.
1
Introduction
In statistical analyses involving data from individuals, there is an increasing tension between the
need to share the data and the need to protect sensitive information about the individuals. For
example, users of social networking sites are increasingly cautious about their privacy, but still find
it inevitable to agree to share their personal information in order to benefit from customized services
such as recommendations and personalized search [1, 2]. There is a certain utility in sharing data for
both data providers and data analysts, but at the same time, individuals want plausible deniability
when it comes to sensitive information.
For such systems, there is a natural core optimization problem to be solved. Assuming both the
data providers and analysts want to maximize the utility of the released data, how can they do so
while preserving the privacy of participating individuals? The formulation and study of an optimal
framework addressing this tradeoff is the focus of this paper.
Local differential privacy. The need for data privacy appears in two different contexts: the local
privacy context, as in when individuals disclose their personal information (e.g., voluntarily on
social network sites), and the global privacy context, as in when institutions release databases of
information of several people or answer queries on such databases (e.g., US Government releases
census data, companies like Netflix release proprietary data for others to test state of the art data
analytics). In both contexts, privacy is achieved by randomizing the data before releasing it. We
study the setting of local privacy, in which data providers do not trust the data collector (analyst).
Local privacy dates back to Warner [29], who proposed the randomized response method to provide
plausible deniability for individuals responding to sensitive surveys.
1
A natural notion of privacy protection is making inference of information beyond what is released
hard. Differential privacy has been proposed in the global privacy context to formally capture this
notion of privacy [11, 13, 12]. In a nutshell, differential privacy ensures that an adversary should
not be able to reliably infer whether or not a particular individual is participating in the database
query, even with unbounded computational power and access to every entry in the database except
for that particular individual?s data. Recently, the notion of differential privacy has been extended
to the local privacy context [10]. Formally, consider a setting where there are n data providers each
owning a data Xi defined on an input alphabet X . In this paper, we shall deal, almost exclusively,
with finite alphabets. The Xi ?s are independently sampled from some distribution P? parameterized
by ? ? {0, 1}. A statistical privatization mechanism Qi is a conditional distribution that maps
Xi ? X stochastically to Yi ? Y, where Y is an output alphabet possibly larger than X . The
Yi ?s are referred to as the privatized (sanitized) views of Xi ?s. In a non-interactive setting where
the individuals do not communicate with each other and the Xi ?s are independent and identically
distributed, the same privatization mechanism Q is used by all individuals. For a non-negative ?, we
follow the definition of [10] and say that a mechanism Q is ?-locally differentially private if
sup
S??(Y),x,x0 ?X
Q(S|Xi = x)
? e? ,
Q(S|Xi = x0 )
(1)
where ?(Y) denotes an appropriate ?-field on Y.
Information theoretic utilities for statistical analyses. The data analyst is interested in the statistics of the data as opposed to individual samples. Naturally, the utility should also be measured in
terms of the distribution rather than sample quantities. Concretely, consider a client-server setting,
where each client with data Xi sends a privatized version of the data Yi , via an ?-locally differentially private privatization mechanism Q. Given the privatized views {Yi }ni=1 , the data analyst wants
to make inferences based on the induced marginal distribution
Z
M? (S) ?
Q(S|x)dP? (x) ,
(2)
for S ? ?(Y) and ? ? {0, 1}. The power to discriminate data generated from P0 to data generated
from P1 depends on the ?distance? between the marginals M0 and M1 . To measure the ability of
such statistical discrimination, our choice of utility of a particular privatization mechanism Q is an
information theoretic quantity called Csisz?ar?s f -divergence defined as
Z
dM0
dM1 ,
(3)
Df (M0 ||M1 ) = f
dM1
for some convex function f such that f (1) = 0. The Kullback-Leibler (KL) divergence
Dkl (M0 ||M1 ) is a special case with f (x) = x log x, and so is the total variation kM0 ? M1 kTV
with f (x) = (1/2)|x ? 1|. Such f -divergences capture the quality of statistical inference, such as
minimax rates of statistical estimation or error exponents in hypothesis testing [28]. As a motivating
example, suppose a data analyst wants to test whether the data is generated from P0 or P1 based on
privatized views Y1 , . . . , Yn . According to Chernoff-Stein?s lemma, for a bounded type I error probability, the best type II error probability scales as e?n Dkl (M0 ||M1 ) . Naturally, we are interested in
finding a privatization mechanism Q that minimizes the probability of error by solving the following
constraint maximization problem
maximize
Q?D?
Dkl (M0 ||M1 ) ,
(4)
where D? is the set of all ?-locally differentially private mechanisms satisfying (1). Motivated by
such applications in statistical inference, our goal is to provide a general framework for finding
optimal privatization mechanisms that maximize the f -divergence between the induced marginals
under local differential privacy.
Contributions. We study the fundamental tradeoff between local differential privacy and f divergence utility functions. The privacy-utility tradeoff is posed as a constrained maximization
problem: maximize f -divergence utility functions subject to local differential privacy constraints.
This maximization problem is (a) nonlinear: f -divergences are convex in Q; (b) non-standard: we
are maximizing instead of minimizing a convex function; and (c) infinite dimensional: the space
of all differentially private mechanisms is uncountable. We show, in Theorem 2.1, that for all f divergences, any ?, and any pair of distributions P0 and P1 , a finite family of extremal mechanisms
2
(a subset of the corner points of the space of privatization mechanisms), which we call staircase
mechanisms, contains the optimal privatization mechanism. We further prove, in Theorem 2.2, that
solving the original problem is equivalent to solving a linear program, the outcome of which is the
optimal staircase mechanism. However, solving this linear program can be computationally expensive since it has 2|X | variables. To account for this, we show that two simple staircase mechanisms
(the binary and randomized response mechanisms) are optimal in the high and low privacy regimes,
respectively, and well approximate the intermediate regime. This contributes an important progress
in the differential privacy area, where the privatization mechanisms have been few and almost no exact optimality results are known. As an application, we show that the effective sample size reduces
from n to ?2 n under local differential privacy in the context of hypothesis testing.
Related work. Our work is closely related to the recent work of [10] where an upper bound on
Dkl (M0 ||M1 ) was derived under the same local differential privacy setting. Precisely, Duchi et. al.
proved that the KL-divergence maximization problem in (4) is at most 4(e? ? 1)2 kP1 ? P2 k2T V .
This bound was further used to provide a minimax bound on statistical estimation using information
theoretic converse techniques such as Fano?s and Le Cam?s inequalities.
In a similar spirit, we are also interested in maximizing information theoretic quantities of the
marginals under local differential privacy. We generalize the results of [10], and provide stronger
results in the sense that we (a) consider a broader class of information theoretic utilities; (b) provide explicit constructions of the optimal mechanisms; and (c) recover the existing result of [10,
Theorem 1] (with a stronger condition on ?).
While there is a vast literature on differential privacy, exact optimality results are only known for a
few cases. The typical recipe is to propose a differentially private mechanism inspired by [11, 13,
26, 20], and then establish its near-optimality by comparing the achievable utility to a converse, for
example in principal component analysis [8, 5, 19, 24], linear queries [21, 18], logistic regression [7]
and histogram release [25]. In this paper, we take a different route and solve the utility maximization
problem exactly.
Optimal differentially private mechanisms are known only in a few cases. Ghosh et. al. showed
that the geometric noise adding mechanism is optimal (under a Bayesian setting) for monotone
utility functions under count queries (sensitivity one) [17]. This was generalized by Geng et. al.
(for a worst-case input setting) who proposed a family of mechanisms and proved its optimality
for monotone utility functions under queries with arbitrary sensitivity [14, 16, 15]. The family of
optimal mechanisms was called staircase mechanisms because for any y and any neighboring x and
x0 , the ratio of Q(y|x) to Q(y|x0 ) takes one of three possible values e? , e?? , or 1. Since the optimal
mechanisms we develop also have an identical property, we retain the same nomenclature.
2
Main results
In this section, we give a formal definition for staircase mechanisms and show that they are the
optimal solutions to maximization problems of the form (5). Using the structure of staircase mechanisms, we propose a combinatorial representation for this family of mechanisms. This allows us
to reduce the nonlinear program of (5) to a linear program with 2|X | variables. Potentially, for
any instance of the problem, one can solve this linear program to obtain the optimal privatization
mechanism, albeit with significant computational challenges since the number of variables scales
exponentially in the alphabet size. To address this, we prove that two simple staircase mechanisms,
which we call the binary mechanism and the randomized response mechanism, are optimal in high
and low privacy regimes, respectively. We also show how our results can be used to derive upper
bounds on f -divergences under privacy. Finally, we give a concrete example illustrating the exact
tradeoff between privacy and statistical inference in the context of hypothesis testing.
2.1
Optimality of staircase mechanisms
Consider a random variable X ? X generated according to P? , ? ? {0, 1}. The distribution of the
privatized output Y , whenever X is distributed according to P? , is represented by M? and given by
(2). We are interested in characterizing the optimal solution of
maximize
Q?D?
Df (M0 ||M1 ) ,
3
(5)
where D? is the set of all ?-differentially private mechanisms satisfying, for all x, x0 ? X and y ? Y,
Q(y|x)
(6)
0 ? ln
? ?.
Q(y|x0 )
This includes maximization over information theoretic quantities of interest in statistical estimation
and hypothesis testing such as total variation, KL-divergence, and ?2 -divergence [28]. In general
this is a complicated nonlinear program: we are maximizing a convex function in Q; further, the
dimension of Q might be unbounded: the optimal privatization mechanism Q? might produce an
infinite output alphabet Y. The following theorem proves that one never needs an output alphabet
larger than the input alphabet in order to achieve the maximum divergence, and provides a combinatorial representation of the optimal solution.
Theorem 2.1. For any ?, any pair of distributions P0 and P1 , and any f -divergence, there exists an
optimal mechanism Q? maximizing the f -divergence in (5) over all ?-locally differentially private
mechanisms, such that
Q? (y|x)
(7)
ln
? {0, ?} ,
Q? (y|x0 )
for all y ? Y, x, x0 ? X and the output alphabet size is at most equal to the input alphabet size:
|Y| ? |X |.
The optimal solution is an extremal mechanism, since the absolute value of the log-likelihood ratios
can only take one of the two extremal values (see Figure 1). We refer to such a mechanism as a
staircase mechanism, and define the family of staircase mechanisms as
S? ? {Q | satisfying (7)} .
This family includes all the optimal mechanisms (for all choices of ? ? 0, P0 , P1 and f ), and since
(7) implies (6), staircase mechanisms are locally differentially private.
e?
3+e?
e?
1+e?
y=1
1
3+e?
1
1+e?
y=1
2
3
2
4
x=1
2
3
4
5
x=1
2
3
4
Figure 1: Examples of staircase mechanisms: the binary and randomized response mechanisms.
For global differential privacy, we can generalize the definition of staircase mechanisms to hold
for all neighboring database queries x, x0 (or equivalently within some sensitivity), and recover all
known existing optimal mechanisms. Precisely, the geometric mechanism shown to be optimal in
[17], and the mechanisms shown to be optimal in [14, 16] (also called staircase mechanisms) are
special cases of the staircase mechanisms defined above. We believe that the characterization of
these extremal mechanisms and the analysis techniques developed in this paper can be of independent interest to researchers interested in optimal mechanisms for global privacy and more general
utilities.
Combinatorial representation of the staircase mechanisms. Now that we know staircase mechanisms are optimal, we can try to combinatorially search for the best staircase mechanism for any
fixed ?, P0 , P1 , and f . To this end, we give a simple representation of all staircase mechanisms,
exploiting the fact that they are scaled copies of a finite number of patterns.
Let Q ? R|X |?|Y| be a staircase mechanism and k = |X | denote the input alphabet size. Then,
using the definition of staircase mechanisms, Q(y|x)/Q(y|x0 ) ? {e?? , 1, e? } and each column
Q(y|?) must be proportional to one of the canonical staircase patterns. For example, when k = 3,
4
k
there are 2k = 8 canonical patterns. Define a staircase pattern matrix S (k) ? {1, e? }k?(2 ) taking
values either 1 or e? , such that the i-th column of S (k) has a staircase pattern corresponding to the
binary representation of i ? 1 ? {0, . . . , 2k ? 1}. We order the columns of S (k) in this fashion for
notational convenience. For example,
#
"
1 1 1 1 e? e? e? e?
(3)
?
?
?
?
1 1 e e .
S = 1 1 e e
1 e? 1 e? 1 e? 1 e?
For all values of k, there are exactly 2k such patterns, and any column of Q(y|?) is a scaled version
of one of the columns of S (k) . Using this ?pattern? matrix, we will show that we can represent (an
equivalence class of) any staircase mechanism Q as
Q = S (k) ? ,
2k ?2k
(8)
(k)
where ? ? R
is a diagonal matrix representing the scaling of the columns of S . We can
now formulate the problem of maximizing the divergence between the induced marginals as a linear
program and prove that it is equivalent the original nonlinear program.
Theorem 2.2. For any ?, any pair of distributions P0 and P1 , and any f -divergence, the nonlinear
program of (5) and the following linear program have the same optimal value
k
maximize
??R2k ?2k
2
X
(k)
?(Si )?ii
(9)
i=1
(k)
S ?1 = 1 ,
? is a diagonal matrix ,
??0,
P
P
(k)
(k)
(k) P
(k)
(k)
where ?(Si ) = ( x?X P1 (x)Sxi )f ( x?X P0 (x)Sxi / x?X P1 (x)Sxi ) and Si is the i-th
k
P2
(k)
column of S (k) , such that Df (M0 ||M1 ) =
i=1 ?(Si )?ii . The solutions of (5) and (9) are
related by (8).
subject to
The infinite dimensional nonlinear program of (5) is now reduced to a finite dimensional linear
program. The first constraint ensures that we get a valid probability transition matrix Q = S (k) ?
with a row sum of one. One could potentially solve this LP with 2k variables but its computational
complexity scales exponentially in the alphabet size k = |X |. For practical values of k this might
not always be possible. However, in the following section, we give a precise description for the
optimal mechanisms in the high privacy and low privacy regimes.
In order to understand the above theorem, observe that both the f -divergences and the differential
privacy constraints are invariant under permutation (or relabelling) of the columns of a privatization
mechanism Q. For example, the KL-divergence Dkl (M0 ||M1 ) does not change if we permute the
columns of Q. Similarly, both the f -divergences and the differential privacy constraints are invariant
under merging/splitting of outputs with the same pattern. To be specific, consider a privatization
mechanism Q and suppose there exist two outputs y and y 0 that have the same pattern, i.e. Q(y|?) =
C Q(y 0 |?) for some positive constant C. Then, we can consider a new mechanism Q0 by merging the
two columns corresponding to y and y 0 . Let y 00 denote this new output. It follows that Q0 satisfies
the differential privacy constraints and the resulting f -divergence is also preserved. Precisely, using
the fact that Q(y|?) = C Q(y 0 |?), it follows that
P
P
(Q(y|x) + Q(y 0 |x))P0 (x)
(1 + C) x Q(y|x)P0 (x)
M00 (y 00 )
M0 (y)
M0 (y 0 )
x
P
P
=
=
=
=
,
0
M10 (y 00 )
(1 + C) x Q(y|x)P1 (x)
M1 (y)
M1 (y 0 )
x (Q(y|x) + Q(y |x))P1 (x)
and thus the corresponding f -divergence is invariant:
M (y)
M (y 0 )
M 0 (y 00 )
0
0
0
f
M1 (y) + f
M1 (y 0 ) = f
M10 (y 00 ) .
0
M1 (y)
M1 (y )
M10 (y 00 )
We can naturally define equivalence classes for staircase mechanisms that are equivalent up to a
permutation of columns and merging/splitting of columns with the same pattern:
[Q] = {Q0 ? S? | exists a sequence of permutations and merge/split of columns from Q0 to Q} . (10)
5
To represent an equivalence class, we use a mechanism in [Q] that is ordered and merged to match
the patterns of the pattern matrix S (k) . For any staircase mechanism Q, there exists a possibly
different staircase mechanism Q0 ? [Q] such that Q0 = S (k) ? for some diagonal matrix ? with
nonnegative entries. Therefore, to solve optimization problems of the form (5), we can restrict our
attention to such representatives of equivalent classes. Further, for privatization mechanisms of the
form Q = S (k) ?, the f -divergences take the form given in (9), a simple linear function of ?.
2.2
Optimal mechanisms in high and low privacy regimes
For a given P0 and P1 , the binary mechanism is defined as a staircase mechanism with only two
outputs y ? {0, 1} satisfying (see Figure 1)
e?
e?
if P0 (x) ? P1 (x) ,
if P0 (x) < P1 (x) ,
1+e?
1+e?
Q(0|x) =
Q(1|x)
=
(11)
1
1
if
P
(x)
<
P
(x)
.
if P0 (x) ? P1 (x) .
?
?
0
1
1+e
1+e
Although this mechanism is extremely simple, perhaps surprisingly, we will establish that this is the
optimal mechanism when high level of privacy is required. Intuitively, the output is very noisy in the
high privacy regime, and we are better off sending just one bit of information that tells you whether
your data is more likely to have come from P0 or P1 .
Theorem 2.3. For any pair of distributions P0 and P1 , there exists a positive ?? that depends on P0
and P1 such that for any f -divergences and any positive ? ? ?? , the binary mechanism maximizes
the f -divergence between the induced marginals over all ?-local differentially private mechanisms.
This implies that in the high privacy regime, which is a typical setting studied in much of differential
privacy literature, the binary mechanism is a universally optimal solution for all f -divergences in (5).
In particular this threshold ?? is universal, in that it does not depend on the particular choice of which
f -divergence we are maximizing. This is established by proving a very strong statistical dominance
using Blackwell?s celebrated result on comparisons of statistical experiments [4]. In a nutshell, we
prove that for sufficiently small ?, the output of any ?-locally differentially private mechanism can be
simulated from the output of the binary mechanism. Hence, the binary mechanism dominates over
all other mechanisms and at the same time achieves the maximum divergence. A similar idea has
been used previously in [27] to exactly characterize how much privacy degrades under composition.
The optimality of binary mechanisms is not just for high privacy regimes. The next theorem shows
that it is the optimal solution of (5) for all ?, when the objective function is the total variation
Df (M0 ||M1 ) = kM0 ? M1 kTV .
Theorem 2.4. For any pair of distributions P0 and P1 , and any ? ? 0, the binary mechanism maximizes total variation between the induced marginals M0 and M1 among all ?-local differentially
private mechanisms.
When maximizing the KL-divergence between the induced marginals, we show that the binary
mechanism still achieves a good performance for all ? ? C where C ? ?? now does not depend on
P0 and P1 . For the special case of KL-divergence, let OPT denote the maximum value of (5) and
BIN denote the KL-divergence when the binary mechanism is used. The next theorem shows that
BIN
?
2(e?
1
OPT .
+ 1)2
Theorem 2.5. For any ? and for any pair of distributions P0 and P1 , the binary mechanism is an
1/(2(e? + 1)2 ) approximation of the maximum KL-divergence between the induced marginals M0
and M1 among all ?-locally differentially private mechanisms.
Note that 2(e? + 1)2 ? 32 for ? ? 1, and ? ? 1 is a common regime of interest in differential
privacy. Therefore, we can always use the simple binary mechanism in this regime and the resulting
divergence is at most a constant factor away from the optimal one.
The randomized response mechanism is defined as a staircase mechanism with the same set of
outputs as the input, Y = X , satisfying (see Figure 1)
(
e?
if y = x ,
|X |?1+e?
Q(y|x) =
1
if y 6= x .
|X |?1+e?
6
It is a randomization over the same alphabet where we are more likely to give an honest response.
We view it as a multiple choice generalization of the randomized response proposed by Warner [29],
assuming equal privacy level for all choices. We establish that this is the optimal mechanism when
low level of privacy is required. Intuitively, the noise is small in the low privacy regime, and we
want to send as much information about our current data as allowed, but no more. For a special case
of maximizing KL-divergence, we show that the randomized response mechanism is the optimal
solution of (5) in the low privacy regime (? ? ?? ).
Theorem 2.6. There exists a positive ?? that depends on P0 and P1 such that for any P0 and P1 , and
all ? ? ?? , the randomized response mechanism maximizes the KL-divergence between the induced
marginals over all ?-locally differentially private mechanisms.
2.3
Lower bounds in differential privacy
In this section, we provide converse results on the fundamental limit of differentially private mechanisms. These results follow from our main theorems and are of independent interest in other applications where lower bounds in statistical analysis are studied [3, 21, 6, 9]. For example, a bound
similar to (12) was used to provide converse results on the sample complexity for statistical estimation with differentially private data in [10].
Corollary 2.7. For any ? ? 0, let Q be any conditional distribution that guarantees ?-local differential privacy. Then, for any pair of distributions P0 and P1 , and any positive ? > 0, there exists a
positive ?? that depends on P0 , P1 , and ? such that for any ? ? ?? , the induced marginals M0 and
M1 satisfy the bound
Dkl M0 ||M1 + Dkl M1 ||M0
?
2(1 + ?)(e? ? 1)2
P0 ? P1
2 .
?
TV
(e + 1)
(12)
This follows from Theorem 2.3 and the fact that under the binary mechanism, Dkl M0 ||M1 =
P0 ? P1
2 (e? ? 1)2 /(e? + 1) + O(?3 ) . Compared to [10, Theorem 1], we recover their bound
TV
of 4(e? ? 1)2 kP0 ? P1 k2TV with a smaller constant. We want to note that Duchi et al.?s bound holds
for all values of ? and uses different techniques. However no achievable mechanism is provided. We
instead provide an explicit mechanism that is optimal in high privacy regime.
Similarly, in the high privacy regime, we can show the following converse result.
Corollary 2.8. For any ? ? 0, let Q be any conditional distribution that guarantees ?-local differential privacy. Then, for any pair of distributions P0 and P1 , and any positive ? > 0, there exists a
positive ?? that depends on P0 , P1 , and ? such that for any ? ? ?? , the induced marginals M0 and
M1 satisfy the bound
Dkl M0 ||M1 + Dkl M1 ||M0
? Dkl (P0 ||P1 ) ? (1 ? ?)G(P0 , P1 )e?? .
where G(P0 , P1 ) =
P
x?X (1
? P0 (x)) log(P1 (x)/P0 (x)).
This follows directly from Theorem 2.6 and the fact that under the randomized response mechanism,
Dkl (M0 ||M1 ) = Dkl (P0 ||P1 ) ? G(P0 , P1 )e?? + O(e?2? ) .
Similarly for total variation, we can get the following converse result. This follows from Theorem
2.4 and explicitly computing the total variation achieved by the binary mechanism.
Corollary 2.9. For any ? ? 0, let Q be any conditional distribution that guarantees ?-local differential privacy. Then,
marginals M0 and M1
for any
pair of distributions P0 and
P1 , the induced
satisfy the bound
M0 ? M1
TV ? ((e? ? 1)/(e? + 1))
P0 ? P1
TV , and equality is achieved
by the binary mechanism.
2.4
Connections to hypothesis testing
Under the data collection scenario, there are n individuals each with data Xi sampled from a distribution P? for a fixed ? ? {0, 1}. Let Q be a non-interactive privatization mechanism guaranteeing
?-local differential privacy. The privatized views {Yi }ni=1 , are independently distributed according
to one of the induced marginals M0 or M1 defined in (2).
7
Given the privatized views {Yi }ni=1 , the data analyst wants to test whether they were generated from
M0 or M1 . Let the null hypothesis be H0 : Yi ?s are generated from M0 , and the alternative hypothesis H1 : Yi ?s are generated from M1 . For a choice of rejection region R ? Y n , the probability
of false alarm (type I error) is ? = M0n (R) and the probability of miss detection (type II error) is
? = M1n (Y n \ R). Let ? ? = minR?Y n ,?<?? ? denote the minimum type II error achievable while
keeping type I error rate at most ?? . According to Chernoff-Stein lemma, we know that
?
1
log ? ? = ?Dkl (M0 ||M1 ) .
n?? n
lim
Suppose the analyst knows P0 , P1 , and Q. Then, in order to achieve optimal asymptotic error rate,
one would want to maximize the KL-divergence between the induced marginals over all ?-locally
differentially private mechanisms Q. Theorems 2.3 and 2.6 provide an explicit construction of the
optimal mechanisms in high and low privacy regimes. Further, our converse results in Section 2.3
provides a fundamental limit on the achievable error rates under differential privacy. Precisely,
with data collected from an ?-locally differentially privatization mechanism, one cannot achieve an
asymptotic type II error smaller than
?
1
(1 + ?)(e? ? 1)2
(1 + ?)(e? ? 1)2
2
log ? ? ? ?
kP
?
P
k
?
?
Dkl (P0 ||P1 ) ,
0
1
TV
n?? n
(e? + 1)
2(e? + 1)
lim
whenever ? ? ?? , where ?? is dictated by Theorem 2.3. In the equation above, the second inequality
follows from Pinsker?s inequality. Since (e? ? 1)2 = O(?2 ) for small ?, the effective sample size is
now reduced from n to ?2 n. This is the price of privacy. In the low privacy regime where ? ? ?? ,
for ?? dictated by Theorem 2.6, one cannot achieve an asymptotic type II error smaller than
?
1
log ? ? ? ?Dkl (P0 ||P1 ) + (1 ? ?)G(P0 , P1 )e?? .
n?? n
lim
3
Discussion
In this paper, we have considered f -divergence utility functions and assumed a setting where individuals cannot collaborate (communicate with each other) before releasing their data. It turns out that
the optimality results presented in Section 2 are general and hold for a large class of convex utility
function [22]. In addition, the techniques developed in this work can be generalized to find optimal
privatization mechanisms in a setting where different individuals can collaborate interactively and
each individual can be an analyst [23].
Binary hypothesis testing is a canonical statistical inference problem with a wide range of applications. However, there are a number of nontrivial and interesting extensions to our work. Firstly,
in some scenarios the Xi ?s could be correlated (e.g., when different individuals observe different functions of the same random variable). In this case, the data analyst is interested in inferring whether the data was generated from P0n or P1n , where P?n is one of two possible joint priors on X1 , ..., Xn . This is a challenging problem because knowing Xi reveals information about
Xj , j 6= i. Therefore, the utility maximization problems for different individuals are coupled in
this setting. Secondly, in some cases the data analyst need not have access to P0 and P1 , but
rather two classes of prior distribution P?0 and P?1 for ?0 ? ?0 and ?1 ? ?1 . Such problems
are studied under the rubric of universal hypothesis testing and robust hypothesis testing. One
possible direction is to select the privatization mechanism that maximizes the worst case utility:
Q? = arg maxQ?D? min?0 ??0 ,?1 ??1 Df (M?0 ||M?1 ), where M?? is the induced marginal under
P?? . Finally, the more general problem of private m-ary hypothesis testing is also an interesting but
challenging one. In this setting, the Xi ?s can follow one of m distributions P0 , P1 , ..., Pm?1 , and
therefore the Yi ?s can follow one of m distributions M0 , M1 , ..., Mm?1 . The
Putility can be defined
as the average f -divergence between any two distributions: 1/(m(m ? 1)) i6=j Df (Mi ||Mj ), or
the worst case one: mini6=j Df (Mi ||Mj ).
References
[1] Alessandro Acquisti. Privacy in electronic commerce and the economics of immediate gratification. In
Proceedings of the 5th ACM conference on Electronic commerce, pages 21?29. ACM, 2004.
8
[2] Alessandro Acquisti and Jens Grossklags. What can behavioral economics teach us about privacy. Digital
Privacy, page 329, 2007.
[3] A. Beimel, K. Nissim, and E. Omri. Distributed private data analysis: Simultaneously solving how and
what. In Advances in Cryptology?CRYPTO 2008, pages 451?468. Springer, 2008.
[4] D. Blackwell. Equivalent comparisons of experiments. The Annals of Mathematical Statistics, 24(2):265?
272, 1953.
[5] J. Blocki, A. Blum, A. Datta, and O. Sheffet. The johnson-lindenstrauss transform itself preserves differential privacy. In Foundations of Computer Science, 2012 IEEE 53rd Annual Symposium on, pages
410?419. IEEE, 2012.
[6] K. Chaudhuri and D. Hsu. Convergence rates for differentially private statistical estimation. arXiv preprint
arXiv:1206.6395, 2012.
[7] K. Chaudhuri and C. Monteleoni. Privacy-preserving logistic regression. In NIPS, volume 8, pages
289?296, 2008.
[8] K. Chaudhuri, A. D. Sarwate, and K. Sinha. Near-optimal differentially private principal components. In
NIPS, pages 998?1006, 2012.
[9] A. De. Lower bounds in differential privacy. In Theory of Cryptography, pages 321?338. Springer, 2012.
[10] John C Duchi, Michael I Jordan, and Martin J Wainwright. Local privacy and statistical minimax rates. In
Foundations of Computer Science, 2013 IEEE 54th Annual Symposium on, pages 429?438. IEEE, 2013.
[11] C. Dwork. Differential privacy. In Automata, languages and programming, pages 1?12. Springer, 2006.
[12] C. Dwork and J. Lei. Differential privacy and robust statistics. In Proceedings of the 41st annual ACM
symposium on Theory of computing, pages 371?380. ACM, 2009.
[13] C. Dwork, F. McSherry, K. Nissim, and A. Smith. Calibrating noise to sensitivity in private data analysis.
In Theory of Cryptography, pages 265?284. Springer, 2006.
[14] Q. Geng and P. Viswanath.
arXiv:1212.1186, 2012.
The optimal mechanism in differential privacy.
arXiv preprint
[15] Q. Geng and P. Viswanath. The optimal mechanism in differential privacy: Multidimensional setting.
arXiv preprint arXiv:1312.0655, 2013.
[16] Q. Geng and P. Viswanath.
arXiv:1305.1330, 2013.
The optimal mechanism in (,?)-differential privacy.
arXiv preprint
[17] A. Ghosh, T. Roughgarden, and M. Sundararajan. Universally utility-maximizing privacy mechanisms.
SIAM Journal on Computing, 41(6):1673?1693, 2012.
[18] M. Hardt, K. Ligett, and F. McSherry. A simple and practical algorithm for differentially private data
release. In NIPS, pages 2348?2356, 2012.
[19] M. Hardt and A. Roth. Beating randomized response on incoherent matrices. In Proceedings of the 44th
symposium on Theory of Computing, pages 1255?1268. ACM, 2012.
[20] M. Hardt and G. N. Rothblum. A multiplicative weights mechanism for privacy-preserving data analysis.
In Foundations of Computer Science, 2010 51st Annual IEEE Symposium on, pages 61?70. IEEE, 2010.
[21] M. Hardt and K. Talwar. On the geometry of differential privacy. In Proceedings of the 42nd ACM
symposium on Theory of computing, pages 705?714. ACM, 2010.
[22] P. Kairouz, S. Oh, and P. Viswanath. Extremal mechanisms for local differential privacy. arXiv preprint
arXiv:1407.1338, 2014.
[23] P. Kairouz, S. Oh, and P. Viswanath. Optimality of non-interactive randomized response. arXiv preprint
arXiv:1407.1546, 2014.
[24] M. Kapralov and K. Talwar. On differentially private low rank approximation. In SODA, volume 5,
page 1. SIAM, 2013.
[25] J. Lei. Differentially private m-estimators. In NIPS, pages 361?369, 2011.
[26] F. McSherry and K. Talwar. Mechanism design via differential privacy. In Foundations of Computer
Science, 2007. 48th Annual IEEE Symposium on, pages 94?103. IEEE, 2007.
[27] S. Oh and P. Viswanath.
arXiv:1311.0776, 2013.
The composition theorem for differential privacy.
arXiv preprint
[28] A. B. Tsybakov and V. Zaiats. Introduction to nonparametric estimation, volume 11. Springer, 2009.
[29] Stanley L Warner. Randomized response: A survey technique for eliminating evasive answer bias. Journal
of the American Statistical Association, 60(309):63?69, 1965.
9
| 5392 |@word private:26 version:2 illustrating:1 achievable:4 stronger:2 eliminating:1 nd:1 sheffet:1 p0:43 celebrated:1 contains:2 exclusively:1 ktv:2 existing:2 current:1 comparing:1 protection:1 si:4 must:1 john:1 ligett:1 discrimination:1 smith:1 core:1 institution:1 provides:2 characterization:1 kairouz:2 firstly:1 unbounded:2 relabelling:1 mathematical:1 enterprise:1 differential:38 symposium:7 prove:5 behavioral:1 privacy:82 introduce:1 x0:10 p1:44 warner:3 inspired:1 company:1 kp0:1 increasing:1 provided:1 bounded:1 maximizes:5 null:1 what:3 minimizes:1 developed:2 finding:2 ghosh:2 guarantee:3 every:1 multidimensional:1 nutshell:2 pramod:1 interactive:3 exactly:3 scaled:2 converse:7 yn:1 before:2 service:1 engineering:2 local:22 positive:8 limit:2 merge:1 rothblum:1 might:3 studied:4 k2t:1 equivalence:3 challenging:2 analytics:1 range:1 practical:2 commerce:2 testing:9 acquisti:2 area:1 universal:2 evasive:1 get:2 convenience:1 cannot:3 context:9 equivalent:6 map:1 roth:1 maximizing:10 send:1 attention:1 economics:2 independently:2 convex:5 survey:2 formulate:1 automaton:1 splitting:2 privatized:7 estimator:1 oh:3 proving:1 notion:3 variation:6 beimel:1 annals:1 construction:2 suppose:3 user:1 exact:3 programming:1 us:1 hypothesis:11 expensive:2 satisfying:5 viswanath:6 database:5 disclose:1 preprint:7 electrical:1 solved:1 capture:2 worst:3 km0:2 ensures:2 region:1 voluntarily:1 alessandro:2 complexity:2 viswanath1:1 cam:1 m0n:1 personal:3 pinsker:1 pramodv:1 solving:7 depend:2 joint:1 represented:1 alphabet:12 effective:2 kp:1 query:6 tell:1 outcome:2 h0:1 larger:2 plausible:2 posed:1 say:1 solve:4 ability:1 statistic:3 transform:1 noisy:1 itself:1 sequence:1 propose:2 neighboring:2 date:1 chaudhuri:3 achieve:4 description:1 participating:2 csisz:1 cautious:1 differentially:23 recipe:1 exploiting:1 convergence:1 produce:1 guaranteeing:1 derive:1 develop:1 cryptology:1 measured:1 blocki:1 progress:1 p2:2 strong:2 come:2 implies:2 direction:1 closely:1 merged:1 mini6:1 deniability:2 bin:2 government:1 generalization:1 randomization:1 opt:2 secondly:1 extension:1 hold:3 mm:1 sufficiently:1 considered:1 m0:29 achieves:2 released:3 estimation:6 combinatorial:3 extremal:7 sensitive:3 combinatorially:1 always:2 rather:2 broader:1 corollary:3 release:5 focus:1 derived:1 notational:1 rank:1 likelihood:1 industrial:1 sense:1 inference:7 interested:6 arg:1 among:2 exponent:1 art:1 special:4 constrained:1 marginal:2 field:1 equal:2 never:1 chernoff:2 identical:1 inevitable:1 geng:4 others:1 few:3 kp1:1 simultaneously:1 divergence:38 preserve:1 individual:17 geometry:1 detection:1 interest:4 dwork:3 mcsherry:3 sinha:1 instance:1 column:13 ar:1 maximization:8 minr:1 addressing:1 entry:2 subset:1 johnson:1 motivating:1 characterize:1 answer:2 randomizing:1 p0n:1 st:2 m10:3 fundamental:4 randomized:13 sensitivity:4 siam:2 retain:1 off:1 michael:1 concrete:1 interactively:1 opposed:1 possibly:2 stochastically:1 corner:1 american:1 account:2 de:1 includes:2 satisfy:3 explicitly:1 depends:5 performed:1 view:6 try:1 h1:1 multiplicative:1 sup:1 kapralov:1 netflix:1 recover:3 complicated:1 contribution:1 il:1 ni:3 who:2 oh2:1 generalize:2 bayesian:1 provider:5 researcher:1 ary:1 r2k:1 networking:1 monteleoni:1 sharing:1 whenever:2 definition:4 m00:1 naturally:3 mi:2 sampled:2 hsu:1 proved:2 hardt:4 lim:3 stanley:1 back:1 appears:1 follow:4 tension:1 response:14 formulation:1 just:2 working:1 trust:1 nonlinear:6 logistic:2 quality:1 perhaps:1 lei:2 believe:1 usa:1 calibrating:1 staircase:33 hence:1 equality:1 q0:6 leibler:1 deal:1 generalized:2 theoretic:8 duchi:3 recently:2 common:1 exponentially:2 volume:3 sarwate:1 association:1 m1:35 marginals:14 sundararajan:1 significant:1 refer:1 composition:2 rd:1 swoh:1 collaborate:2 pm:1 fano:1 similarly:3 i6:1 illinois:2 language:1 access:2 recent:1 showed:1 dictated:2 scenario:2 route:1 certain:1 server:1 inequality:3 binary:20 yi:9 jens:1 preserving:3 minimum:1 maximize:8 ii:7 multiple:1 infer:1 reduces:1 champaign:1 match:1 dkl:16 qi:1 involving:1 regression:2 df:7 arxiv:14 histogram:1 represent:2 achieved:3 preserved:1 addition:1 want:9 sends:1 releasing:2 induced:14 subject:2 spirit:1 jordan:1 call:3 near:2 intermediate:2 split:1 identically:1 xj:1 restrict:1 reduce:1 idea:1 knowing:1 tradeoff:5 honest:1 whether:5 motivated:1 utility:24 peter:1 nomenclature:1 proprietary:1 nonparametric:1 stein:2 tsybakov:1 locally:10 reduced:2 exist:1 canonical:3 shall:1 dominance:1 threshold:1 blum:1 sxi:3 vast:1 monotone:2 sum:1 talwar:3 parameterized:1 you:1 communicate:2 soda:1 family:7 almost:2 electronic:2 scaling:1 bit:1 bound:13 nonnegative:1 annual:5 nontrivial:1 roughgarden:1 constraint:6 precisely:4 your:1 personalized:1 optimality:8 extremely:1 min:1 martin:1 department:2 tv:5 according:5 smaller:3 increasingly:1 lp:1 making:1 intuitively:2 invariant:3 census:1 computationally:2 ln:2 agree:1 remains:1 previously:1 equation:1 count:1 mechanism:122 turn:1 crypto:1 know:3 end:1 sending:1 rubric:1 observe:2 away:1 appropriate:1 alternative:1 original:2 responding:1 denotes:1 uncountable:1 prof:1 establish:3 objective:1 quantity:4 degrades:1 diagonal:3 dp:1 distance:1 simulated:1 nissim:2 collected:1 analyst:13 assuming:2 ratio:2 minimizing:1 equivalently:1 potentially:2 teach:1 m1n:1 negative:1 design:1 reliably:1 upper:2 urbana:2 finite:4 immediate:1 extended:1 precise:1 y1:1 arbitrary:1 datta:1 pair:9 required:2 kl:11 blackwell:2 connection:1 protect:1 established:1 maxq:1 nip:4 address:1 beyond:1 adversary:1 able:1 pattern:12 beating:1 privatization:20 regime:18 challenge:1 program:14 wainwright:1 power:2 natural:2 client:2 customized:1 minimax:3 representing:1 incoherent:1 coupled:1 prior:2 literature:2 geometric:2 asymptotic:3 permutation:3 interesting:2 proportional:1 digital:1 foundation:4 sewoong:1 share:2 row:1 surprisingly:1 copy:1 keeping:1 formal:1 bias:1 understand:1 wide:1 characterizing:1 taking:1 absolute:1 benefit:1 distributed:4 dimension:1 xn:1 valid:1 transition:1 lindenstrauss:1 concretely:1 collection:1 universally:3 social:2 approximate:2 kullback:1 global:4 reveals:1 assumed:1 xi:12 search:2 mj:2 robust:2 zaiats:1 contributes:1 permute:1 main:2 noise:3 alarm:1 allowed:1 collector:1 p1n:1 cryptography:2 x1:1 site:2 referred:1 representative:1 owning:1 fashion:1 inferring:1 explicit:3 exponential:1 omri:1 theorem:22 specific:1 dominates:1 exists:7 albeit:1 adding:1 merging:3 false:1 dm0:1 rejection:1 likely:2 ordered:1 recommendation:1 springer:5 satisfies:1 acm:7 conditional:4 goal:1 price:1 hard:1 change:1 infinite:3 except:1 typical:2 miss:1 lemma:2 principal:2 called:3 total:6 discriminate:1 formally:2 select:1 people:1 correlated:1 |
4,852 | 5,393 | Reputation-based Worker Filtering in Crowdsourcing
Srikanth Jagabathula1 Lakshminarayanan Subramanian2,3 Ashwin Venkataraman2,3
1
Department of IOMS, NYU Stern School of Business
Department of Computer Science, New York University
3
CTED, New York University Abu Dhabi
[email protected] {lakshmi,ashwin}@cs.nyu.edu
2
Abstract
In this paper, we study the problem of aggregating noisy labels from crowd workers to infer the underlying true labels of binary tasks. Unlike most prior work
which has examined this problem under the random worker paradigm, we consider
a much broader class of adversarial workers with no specific assumptions on their
labeling strategy. Our key contribution is the design of a computationally efficient
reputation algorithm to identify and filter out these adversarial workers in crowdsourcing systems. Our algorithm uses the concept of optimal semi-matchings
in conjunction with worker penalties based on label disagreements, to assign a
reputation score for every worker. We provide strong theoretical guarantees for
deterministic adversarial strategies as well as the extreme case of sophisticated
adversaries where we analyze the worst-case behavior of our algorithm. Finally,
we show that our reputation algorithm can significantly improve the accuracy of
existing label aggregation algorithms in real-world crowdsourcing datasets.
1
Introduction
The growing popularity of online crowdsourcing services (e.g. Amazon Mechanical Turk, CrowdFlower etc.) has made it easy to collect low-cost labels from the crowd to generate training datasets
for machine learning applications. However, these applications remain vulnerable to noisy labels
introduced either unintentionally by unreliable workers or intentionally by spammers and malicious
workers [10, 11]. Recovering the underlying true labels in the face of noisy input in online crowdsourcing environments is challenging due to three key reasons: (a) Workers are often anonymous
and transient and can provide random or even malicious labels (b) The reliabilities or reputations of
the workers are often unknown (c) Each task may receive labels from only a (small) subset of the
workers.
Several existing approaches aim to address the above challenges under the following standard setup.
There is a set T of binary tasks, each with a true label in { 1, 1}. A set of workers W are asked
to label the tasks, and the assignment of the tasks to the workers can be represented by a bipartite
graph with the workers on one side, tasks on the other side, and an edge connecting each worker
to the set of tasks she is assigned. We term this the worker-task assignment graph. Workers are
assumed to generate labels according to a probabilistic model - given a task t, a worker w provides
the true label with probability pw . Note that every worker is assumed to label each task independent
of other tasks. The goal then is to infer the underlying true labels of the tasks by aggregating the
labels provided by the workers. Prior works based on the above model can be broadly classified into
two categories: machine-learning based and linear-algebra based. The machine-learning approaches
are typically based on variants of the EM algorithm [3, 16, 24, 14]. These algorithms perform well
in most scenarios, but they lack any theoretical guarantees. More recently, linear-algebra based
algorithms [9, 6, 2] have been proposed, which provide guarantees on the error in estimating the
true labels of the tasks (under appropriate assumptions), and have also been shown to perform well
on various real-world datasets. While existing work focuses on workers making random errors,
recent work and anecdotal evidence have shown that worker labeling strategies that are common in
practice do not fit the standard random model [19]. Specific examples include vote pollution attacks
1
on Digg [18], malicious behavior in social media [22, 12] and low-precision worker populations in
crowdsourcing experiments [4].
In this paper, we aim to go beyond the standard random model and study the problem of inferring
the true labels of tasks under a much broader class of adversarial worker strategies with no specific
assumptions on their labeling pattern. For instance, deterministic labeling, where the workers always
give the same label, cannot be captured by the standard random model. Also, malicious workers can
employ arbitrary labeling patterns to degrade the accuracy of the inferred labels. Our goal is to
accurately infer the true labels of the tasks without restricting workers? strategies.
Main results. Our main contribution is the design of a reputation algorithm to identify and filter out
adversarial workers in online crowdsourcing systems. Specifically, we propose 2 computationally
efficient algorithms to compute worker reputations using only the labels provided by the workers
(see Algorithms 1 and 2), which are robust to manipulation by adversaries. We compute worker
reputations by assigning penalties to a worker for each task she is assigned. The assigned penalty is
higher for tasks on which there is ?a lot? of disagreement with the other workers. The penalties are
then aggregated in a ?load-balanced? manner using the concept of optimal semi-matchings [7]. The
reputation algorithm is designed to be used in conjunction with any of the existing label aggregation
algorithms that are designed for the standard random worker model: workers with low reputations1
are filtered out and the aggregation algorithm is used on the remaining labels. As a result, our
algorithm can be used to boost the performance of existing label aggregation algorithms.
We demonstrate the effectiveness of our algorithm using a combination of strong theoretical guarantees and empirical results on real-world datasets. Our analysis considers three scenarios. First,
we consider the standard setting in which workers are not adversarial and provide labels according
to the random model. In this setting, we show that when the worker-task assignment graph is (l, r)regular, the reputation scores are proportional to the reliabilities of the workers (see Theorem 1), so
that only unreliable workers are filtered out. As a result, our reputation scores are consistent with
worker reliabilities in the absence of adversarial workers. The analysis becomes significantly complicated for more general graphs (a fact observed in prior works; see [2]); hence, we demonstrate
improvements using simulations and experiments on real world datasets. Second, we evaluate the
performance of our algorithm in the presence of workers who use deterministic labeling strategies
(always label 1 or 1). For these strategies, when the worker-task assignment graph is (l, r)-regular,
we show (see Theorem 2) that the adversarial workers receive lower reputations than their ?honest?
counterparts, provided honest workers have ?high enough? reliabilities ? the exact bound depends
on the prevalence of tasks with true label 1, the fraction of adversarial workers and the average
reliability of the honest workers.
Third, we consider the case of sophisticated adversaries, i.e. worst-case adversarial workers whose
goal is to maximize the number of tasks they affect (i.e. cause to get incorrect labels). Under
this assumption, we provide bounds on the ?damage? they can do: We prove that irrespective of
the label aggregation algorithm (as long as it is agnostic to worker/task identities), there is a nontrivial minimum fraction of tasks whose true label is incorrectly inferred. This bound depends on
the graph structure of the honest worker labeling pattern (see Theorem 3 for details). Our result
is valid across different labeling patterns and a large class of label aggregation algorithms, and
hence provides fundamental limits on the damage achievable by adversaries. Further, we propose a
label aggregation algorithm utilizing the worker reputations computed in Algorithm 2 and prove the
existence of an upper bound on the worst-case accuracy in inferring the true labels (see Theorem 4).
Finally, using several publicly available crowdsourcing datasets (see Section 4), we show that our
reputation algorithm: (a) can help in enhancing the accuracy of state-of-the-art label aggregation
algorithms (b) is able to detect workers in these datasets who exhibit certain ?non-random? strategies.
Additional Related Work: In addition to the references cited above, there have been works which
use gold standard tasks, i.e. tasks whose true label is already known [17, 5, 11] to correct for worker
bias. [8] proposed a way of quantifying worker quality by transforming the observed labels into soft
posterior labels based on the estimated confusion matrix [3]. The authors in [13] propose an empirical Bayesian algorithm to eliminate workers who label randomly without looking at the particular
task (called spammers), and estimate the consensus labels from the remaining workers. Both these
1
As will become evident later, reputations are measures of how adversarial a worker is and are different
from reliabilities of workers.
2
works use the estimated parameters to define ?good workers? whereas we compute the reputation
scores using only the labels provided by the workers. The authors in [20] focus on detecting specific
kinds of spammers and then replace their labels with new workers. We consider all types of adversarial workers, not just spammers and don?t assume access to a pool of workers who can be asked
to label the tasks.
2
Model and reputation algorithms
Notation. Consider a set of tasks T having true labels in {1, 1}. Let yj denotes the true label of a
task tj 2 T and suppose that the tasks are sampled from a population in which the prevalence of the
positive tasks is 2 [0, 1], so that there is a fraction of tasks with true label 1. A set of workers
W provide binary labels to the tasks in T . We let G denote the bipartite worker-task assignment
graph where an edge between worker wi and task tj indicates that wi has labeled tj . Further, let
wi (tj ) denote the label provided by worker wi to task tj , where we set wi (tj ) = 0 if worker wi did
not label task tj . For a task tj , let Wj ? W denote the set of workers who labeled tj and likewise,
for a worker wi , let Ti denote the set of tasks the worker has labeled. Denote by d+
j (resp. dj ) the
|W |?|T |
number of workers labeling task tj as 1 (resp. 1). Finally, let L 2 {1, 0, 1}
denote the
matrix representing the labels assigned by the workers to the tasks, i.e. Lij = wi (tj ). Given L, the
goal is to infer the true labels yj of the tasks.
Worker model. We consider the setting in which workers may be honest or adversarial. That is,
W = H [ A with H \ A = ;. Honest workers are assumed to provide labels according to a
probabilistic model: for task tj with true label yj , worker wi provides label yj with probability pi
and yj with probability 1 pi . Note that the parameter pi doesn?t depend on the particular task that
the worker is labeling, so an honest worker labels each task independently. It is standard to define the
reliability of an honest worker as ?i = 2pi 1, so that we have ?i 2 [ 1, 1]. Further, we assume that
the honest workers are sampled from a population with average reliability ? > 0. Adversaries, on
the other hand, may adopt any arbitrary (deterministic or randomized) labeling strategy that cannot
be described using the above probabilistic process. For instance, the adversary could always label
all tasks as 1, irrespective of the true label. Another example is when the adversary decides her
labels based on existing labels cast by other workers (assuming the adversary has access to such
information). Note however, that adversarial workers need not always provide the incorrect labels.
Essentially, the presence of such workers breaks the assumptions of the model and can adversely
impact the performance of aggregation algorithms. Hence, our focus in this paper is on designing
algorithms to identify and filter out such adversarial workers. Once this is achieved, we can use
existing state-of-the-art label aggregation algorithms on the remaining labels to infer the true labels
of the tasks.
To identify these adversarial workers, we propose an algorithm for computing ?reputation? or ?trust?
scores for each worker. More concretely, we assign penalties (in a suitable way) to every worker and
higher the penalty, worse the reputation of the worker. First note that any task that has all 1 labels
(or 1 labels) does not provide us with any information about the reliabilities of the workers who
labeled the task. Hence, we focus on the tasks that have both 1 and 1 labels and we call this set
the conflict set Tcs . Further, since we have no ?side? information on the identities of workers, any
reputation score computation must be based solely on the labels provided by the workers.
We start with the following basic idea to compute reputation scores: a worker is penalized for every
?conflict? s/he is involved in (a task in the conflict set the worker has labeled on). This idea is
motivated by the fact that in an ideal scenario, where all honest workers have a reliability ?i = 1,
a conflict indicates the presence of an adversary and the reputation score aims to capture a measure
of the number of conflicts each worker is involved in: the higher the number of conflicts, the worse
the reputation score. However, a straightforward aggregation of penalties for each worker may overpenalize (honest) workers who label several tasks.
In order to overcome the issue of over-penalizing (honest) workers, we propose two techniques:
(a) soft and (b) hard assignment of penalties. In the soft assignment of penalties (Algorithm 1),
we assign a penalty of 1/d+
j to all workers who label 1 on task tj and 1/dj to all workers who
label 1 on task tj . Then, for each worker, we aggregate the penalties across all assigned tasks by
taking the average. The above assignment of penalties implicitly rewards agreements by making
the penalty inversely proportional to the number of other workers that agree with a worker. Further,
taking the average normalizes for the number of tasks labeled by the worker. Since we expect the
3
honest workers to agree with the majority more often than not, we expect this technique to assign
lower penalties to honest workers when compared to adversaries. The soft assignment of penalties
can be shown to perform quite well in identifying low reliability and adversarial workers (refer
to Theorems 1 and 2). However, it may still be subject to manipulation by more ?sophisticated?
adversaries who can adapt and modify their labeling strategy to target certain tasks and to inflate
the penalty of specific honest workers. In fact for such worst-case adversaries, we can show that
(Theorem 3) given any honest worker labeling pattern, there exists a lower bound on the number of
tasks whose true label cannot be inferred correctly, by any label aggregation algorithm.
To address the case of these sophisticated adversaries, we propose a hard penalty assignment scheme
(Algorithm 2) where the key idea is not to distribute the penalty evenly across all workers; but to
only choose two workers to penalize per conflict task: one ?representative? worker among those
who labeled 1 and another ?representative? worker among those who labeled 1. While choosing
such workers, the goal is to pick these representative workers in a load-balanced manner to ?spread?
the penalty across all workers, so that it is not concentrated on one/few workers. The final penalty of
each worker is the sum of the accrued penalties across all the (conflict) tasks assigned to the worker.
Intuitively, such hard assignment of penalties will penalize workers with higher degrees and many
conflicts (who are potential ?worst-case? adversaries), limiting their impact.
We use the concept of optimal semi-matchings [7] on bipartite graphs to distribute penalties in a
load balanced manner, which we briefly discuss here. For a bipartite graph B = (U, V, E), a semimatching in B is a set of edges M ? E such that each vertex in V is incident to exactly one edge in
M (note that vertices in U could be incident to multiple edges in M ). A semi-matching generalizes
the notion of matchings on bipartite graphs. To define an optimal semi-matching, we introduce a cost
function for a semi-matching - for each u 2 U , let degM (u) denote the number of edges in M that
PdegM (u)
M (u)+1)
are incident to u and let costM (u) be defined as costM (u)
i = degM (u)(deg
.
2
P= i=1
An optimal semi-matching then, is one which minimizes u2U costM (u). This notion of cost is
motivated by the load balancing problem for scheduling tasks on machines (refer to [7] for further
details). Intuitively, an optimal semi-matching fairly matches the V -vertices across the U -vertices
so that the maximum ?load? on any U -vertex is minimized.
Algorithm 1 SOFT PENALTY
Algorithm 2 HARD PENALTY
1: Input: W , T and L
1: Input: W , T and L
2: For every task tj 2 Tcs , assign penalty sij 2: Create a bipartite graph B cs as follows:
to each worker wi 2 Wj as follows:
sij = d1+ if Lij = 1
j
sij =
if Lij =
1
pen(wi ) =
P
1
dj
3: Output: Penalty of worker wi
3
tj 2Ti \ Tcs
sij
|Ti \ Tcs |
(i) Each worker wi 2 W is represented by
a node on the left (ii) Each task tj 2 Tcs is
represented by two nodes on the right t+
j and
tj (iii) Add the edge (wi , t+
)
if
L
=
1 or
ij
j
edge (wi , tj ) if Lij = 1.
3: Compute an optimal semi-matching OSM on
B cs and let di ( degree of worker wi in OSM
4: Output: Penalty of worker wi pen(wi ) = di
Theoretical Results
Soft penalty. We focus on (l, r)-regular worker-task assignment graphs in which every worker
is assigned l tasks and every object is labeled by r workers. The performance of our reputation
algorithms depend on the reliabilities of the workers as well as the true labels of the tasks. Hence,
we consider the following probabilistic model: for a given (l, r)-regular worker-task assignment
graph G, the reliabilities of the workers and the true labels of tasks are sampled independently (from
distributions described in Section 2). We then analyze the performance of our algorithms as the
task degree r (and hence number of workers |W |) goes to infinity. Specifically, we establish the
following results (the proofs of all theorems are in the supplementary material).
Theorem 1. Suppose there are no adversarial workers, i.e A = ; and that the worker-task assignment graph G is (l, r)-regular. Then, with high probability as r ! 1, for any pair of workers wi
and wi0 , ?i < ?i0 =) pen(wi ) > pen(wi0 ), i.e. higher reliability workers are assigned lower
penalties by Algorithm 1.
4
The probability in the above theorem is according to the model described above. Note that the theorem justifies our definition of the reputation scores by establishing their consistency with worker
reliabilities in the absence of adversarial workers. Next, consider the setting in which adversarial
workers adopt the following uniform strategy: label 1 on all assigned tasks (the 1 case is symmetric).
Theorem 2. Suppose that the worker-task assignment graph G is (l, r)-regular. Let the probability
of an arbitrary worker being honest be q and suppose each adversary adopts the uniform strategy
in which she labels 1 on all the tasks assigned to her. Denote an arbitrary honest worker as hi and
any adversary as a. Then, with high probability as r ! 1, we have
1. If
=
1
2
and ?i = 1, then pen(hi ) < pen(a) if and only if q >
2. If
=
1
2
and q >
1
1+? ,
1
1+?
then pen(hi ) < pen(a) if and only if
?i
(2
q)(1
(2
q q 2 ?2 ) q 2 ?2
q)q + q 2 ?2
The above theorem establishes that when adversaries adopt the uniform strategy, the soft-penalty
algorithm assigns lower penalties to honest workers whose reliability excess a threshold, as long as
the fraction of honest workers is ?large enough?. Although not stated, the result above immediately
extends (with a modified lower bound for ?i ) to the case when > 1/2, which corresponds to
adversaries adopting smart strategies by labeling based on the prevalence of positive tasks.
Sophisticated adversaries. Numerous real-world incidents show evidence of malicious worker behavior in online systems [18, 22, 12]. Moreover, attacks on the training process of ML models
have also been studied [15, 1]. Recent work [21] has also shown the impact of powerful adversarial attacks by administrators of crowdturfing (malicious crowdsourcing) sites. Motivated by these
examples, we consider sophisticated adversaries:
Definition 1. Sophisticated adversaries are computationally unbounded and colluding. Further,
they have knowledge of the labels provided by the honest workers and their goal is to maximize the
number of tasks whose true label is incorrectly identified.
We now raise the following question: In the presence of sophisticated adversaries, does there exist
a fundamental limit on the number of tasks whose true label can be correctly identified, irrespective
of the aggregation algorithm employed to aggregate the worker labels?
In order to answer the above question precisely, we introduce some notation. Let n = |W | and
m
m = |T |. Then, we represent any label aggregation algorithm as a decision rule R : L ! {1, 1} ,
which maps the observed labeling matrix L to a set of output labels for each task. Because of the
absence of any auxiliary information about the workers or the tasks, the class of decision rules, say
C, is invariant to permutations of the identities of workers and/or tasks. More precisely, C denotes
the class of decision rules that satisfy R(P LQ) = R(L)Q, for any n ? n permutation matrix P and
m ? m permutation matrix Q. We say that a task is affected if the decision rule outputs the incorrect
label for the task. We define the quality of a decision rule R(?) as the worst-case number of affected
tasks over all possible true labelings and adversary strategies with a fixed honest worker labeling
pattern. Fixing the honest worker labeling pattern allows isolation of the effect of the adversary
strategy on the accuracy of the decision rule. Considering the worst-case over all possible true
labels makes the metric robust to ground-truth assignments, which are typically application specific.
Next to formally define the quality, let BH denote the honest worker-task assignment graph and
y = (y1 , y2 , . . . , ym ) denote the vector of true labels for the tasks. Note that since the number of
affected tasks also depends on the actual honest worker labels, we further assume that all honest
workers have reliability ?i = 1, i.e they always label correctly. This assumption allows us to
attribute any mis-identification of true labels to the presence of adversaries because, otherwise, in
the absence of any adversaries, the true labels of all the tasks can be trivially identified. Finally, let
Sk denote the strategy space of k adversaries, where each strategy 2 Sk specifies the k ? m label
matrix provided by the adversaries. Since we do not restrict the adversary strategy in any way, it
k?m
follows that Sk = { 1, 0, 1}
. The quality of a decision rule is then defined as
n
o
y,
A?(R, BH , k) =
max
t
2
T
:
R
=
6
y
)
,
j
j
t
j
m
2Sk ,y2{1, 1}
5
where Rty, 2 {1, 1} is the label output by the decision rule R for task t when the true label vector
is y and the adversary strategy is . Note that our notation A?(R, BH , k) makes the dependence of
the quality measure on the honest worker-task assignment graph BH and the number of adversaries k
explicit. We answer the question raised above in the affirmative, i.e. there does exist a fundamental
limit on identification. In the theorem below, PreIm(T 0 ) is the set of honest workers who label
atleast one task in T 0 .
Theorem 3. Suppose that k = |A| and ?h = 1 for all honest workers h 2 H. Then, given
any honest worker-task assignment graph BH , there exists an adversary strategy ? 2 Sk that is
independent of any decision rule R 2 C such that
L ? max m A?(R, ? , y) 8R 2 C, where
y2{ 1,1}
1
max
|T 0 | ,
2 T 0 ?T : |PreIm(T 0 )|?k
and A?(R, ? , y) denotes the number of affected tasks under adversary strategy ? , decision rule
R, and true label vector y (with the assumption that max over an empty set is zero).
L=
We describe the main idea of the proof which proceeds in two steps: (i) we provide an explicit
construction of an adversary strategy ? that depends on BH and y, and (ii) we show the existence of
another true labeling y
? such that R outputs exactly the same labels in both scenarios. The adversary
labeling strategy we construct uses the idea of indistinguishability, which captures the fact that by
carefully choosing their labels, the adversaries can render themselves indistinguishable from honest
workers. Specifically, in the simple case when there is only one honest worker, the adversary simply
labels the opposite of the honest worker on all assigned tasks, so that each task has two labels of
opposite parity. It can be argued that since there is no other information, it is impossible for any
decision rule R 2 C to distinguish the honest worker from the adversary and hence identify the
true label of any task (better than a random guess). We extend this to the general case, where the
adversary ?targets? atmost k honest workers and derives a strategy based on the subgraph of BH
restricted to the targeted workers. The resultant strategy can be shown to result in incorrect labels
for atleast L tasks for some true labeling of the tasks.
Hard penalty. We now analyze the performance of the hard penalty-based reputation algorithm in
the presence of sophisticated adversarial workers. For the purposes of the theorem, we consider a
natural extension of our reputation algorithm to also perform label aggregation (see figure 1).
Theorem 4. Suppose that k = |A| and ?i = 1 for each honest worker, i.e an honest worker always
provides the correct label. Further, let d1
d2
???
d|H| denote the degrees of the honest
workers in the optimal semi-matching on BH . For any true labeling of the tasks and under the
penalty-based label aggregation algorithm (with the convention that di = 0 for i > |H|) :
Pk 1
1. There exists an adversary strategy ? such that the number of tasks affected
i=1 di .
2. No adversary strategy can affect more than U tasks where
Pk
(a) U = i=1 di , when atmost one adversary provides correct labels
P2k
(b) U = i=1 di , in the general case
A few remarks are in order. First, it can be shown [7] that for optimal semi-matchings, the degree
sequence is unique and therefore, the bounds in the theorem above are uniquely defined given BH .
Also, the assumption that ?i = 1 is required for analytical tractability, proving theoretical guarantees in crowd-sourced settings (even without adversaries) for general graph structures is notoriously
hard [2]. Note that the result of Theorem 4 provides both a lower and upper bound for the number
of tasks that can be affected by k adversaries when using the penalty-based aggregation algorithm.
The characterization we obtain is reasonably tight for the case when atmost 1 adversary provides
correct labels; in this case the gap between the upper and lower bounds is dk , which can be ?small?
for k large enough. However, our characterization is loose in the general case when adversaries can
P2k
label arbitrarily; here the gap is i=k di which we attribute to our proof technique and conjecture
Pk
that the upper bound of i=1 di also applies in the more general case.
4
Experiments
In this section, we evaluate the performance of our reputation algorithms on both synthetic and real
datasets. We consider the following popular label aggregation algorithms: (a) simple majority vot6
Random
MV
EM
KOS
KOS +
PRECISION
BEST
Malicious
Uniform
Low
High
Low
High
Low
High
9.9
-1.9
-4.3
-3.9
81.7
7.9
6.3
13.1
7.3
82.1
16.8
-1.6
-8.3
-8.3
92.5
15.6
-49.4
-98.7
-69.6
59.4
26.0
-1.2
-6.5
-6.0
80.8
15.0
-9.1
12.9
10.7
62.4
MV- SOFT MV- HARD MV- SOFT KOS MV- SOFT MV- HARD
PENALTY- BASED AGGREGATION
wt ( worker that task t is mapped
to in OSM in Algorithm 2
Output y(t) = 1 if dwt+ < dwt
y(t) = 1 if dwt+ > dwt and
y(t) = 0 otherwise
(here y refers to the label of the task
and dw refers to the degree of worker
w in OSM)
Figure 1: Left: Percentage decrease in incorrect tasks on synthetic data (negative indicates increase in incorrect
tasks). We implemented both SOFT and HARD and report the best outcome. Also reported is the precision
when removing 15 workers with the highest penalties. The columns specify the three types of adversaries
and High/Low indicates the degree bias of the adversaries. The probability that a worker is honest q was
set to 0.7 and the prevalence of positive tasks was set to 0.5. The numbers reported are an average over
100 experimental runs. The last row lists the combination with the best accuracy in each case. Right: The
penalty-based label aggregation algorithm.
ing MV (b) the EM algorithm [3] (c) the BP-based KOS algorithm [9] and (d) KOS +, a normalized
version of KOS that is amenable for arbitrary graphs (KOS is designed for random regular graphs),
and compare their accuracy in inferring the true labels of the tasks, when implemented in conjunction with our reputation algorithms. We implemented iterative versions of Algorithms 1(SOFT) and
2(HARD), where in each iteration we filtered out the worker with the highest penalty and recomputed
penalties for the remaining workers.
Synthetic Dataset. We analyzed the performance of our soft penalty-based reputation algorithm on
(l, r)-regular graphs in section 3. In many practical scenarios, however, the worker-task assignment
graph forms organically where the workers decide which tasks to label on. To consider this case, we
simulated a setup of 100 workers with a power-law distribution for worker degrees to generate the
bipartite worker-task assignment graph. We assume that an honest worker always labels correctly
(the results are qualitatively similar when honest workers make errors with small probability) and
consider three notions of adversaries: (a) random - who label each task 1 or 1 with prob. 1/2
(b) malicious - who always label incorrectly and (c) uniform - who label 1 on all tasks. Also,
we consider both cases - one where the adversaries are biased to have high degrees and the other
where they have low degrees. Further, we arbitrarily decided to remove 15% of the workers with the
highest penalties and we define precision as the percentage of workers filtered who were adversarial.
Figure 1 shows the performance improvement of the different benchmarks in the presence of our
reputation algorithm.
We make a few observations. First, we are successful in identifying random adversaries as well as
low-degree malicious and uniform adversaries (precision > 80%). This shows that our reputation
algorithms also perform well when worker-task assignment graphs are non-regular, complementing
the theoretical results (Theorems 1 and 2) for regular graphs. Second, our filtering algorithm can
result in significant reduction (upto 26%) in the fraction of incorrect tasks. In fact, in 5 out of 6 cases,
the best performing algorithm incorporates our reputation algorithm. Note that since 15 workers
are filtered out, labels from fewer workers are used to infer the true labels of the tasks. Despite
using fewer labels, we improve performance because the high precision of our algorithms results in
mostly adversaries being filtered out. Third, we note that when the adversaries are malicious and
have high degrees, the removal of 15 workers degrades the performance of the KOS (and KOS +) and
EM algorithms. We attribute this to the fact that while KOS and EM are able to automatically invert
the malicious labels, we discard these labels which hurts performance, since the adversaries have
high degrees. Finally, note that the SOFT (HARD) penalty algorithm tends to perform better when
adversaries are biased towards low (high) degrees, and this insight can be used to aid the choice of
the reputation algorithm to be employed in different scenarios.
Real Datasets. Next, we evaluated our algorithm on some standard datasets: (a) TREC2 : a collection
of topic-document pairs labeled as relevant or non-relevant by workers on AMT. We consider two
versions: stage2 and task2. (b) NLP [17]: consists of annotations by AMT workers for different
NLP tasks (1) rte - which provides binary judgments for textual entailment, i.e. whether one
2
http://sites.google.com/site/treccrowd/home
7
Dataset
rte
temp
bluebird
stage2
task2
MV
EM
KOS +
KOS
Base
Soft
Hard
Base
Soft
Hard
Base
Soft
Hard
Base
Soft
Hard
91.9
93.9
75.9
74.1
64.3
92.1(8)
93.9
75.9
74.1
64.3
92.5(3)
94.3(5)
75.9
81.4(3)
68.4(10)
92.7
94.1
89.8
64.7
66.8
92.7
94.1
89.8
65.3(6)
66.8
93.3(9)
94.1
89.8
78.9(2)
67.3(9)
49.7
56.9
72.2
74.5
57.4
88.8(9)
69.2(4)
75.9(3)
74.5
57.4
91.6(10)
93.7(3)
72.2
75.2(3)
66.7(10)
91.3
93.9
72.2
75.5
59.3
92.7(8)
94.3(7)
75.9(3)
76.6(2)
59.4(4)
92.8(10)
94.3(1)
72.2
77.2(3)
67.9(9)
82.5
81.6 81.7
84.7
62.1 73.2
79.9
78.4 79.8
aggregate 80.0 80.0
80.9
Table 1: Percentage accuracy of benchmark algorithms when combined with our reputation algorithms. For
each benchmark, the best performing combination is shown in bold. The number in the parentheses represents the number of workers filtered by our reputation algorithm (an absence indicates that no performance
improvement was achieved while removing upto 10 workers with the highest penalties).
sentence can be inferred from another (2) temp - which provides binary judgments for temporal
ordering of events. (c) bluebird [23] contains judgments differentiating two kinds of birds in
an image. Table 1 reports the best accuracy achieved when upto 10 workers are filtered using our
iterative reputation algorithms.
The main conclusion we draw is that our reputation algorithms are able to boost the performance
of state-of-the-art aggregation algorithms by a significant amount across the datasets: the average
improvement for MV and KOS + is 2.5%, EM is 3% and for KOS is almost 18%, when using the hard
penalty-based reputation algorithm. Second, we can elevate the performance of simpler algorithms
such as KOS and MV to the levels of the more general (and in some cases, complicated) EM algorithm. The KOS algorithm for instance, is not only easier to implement, but also has tight theoretical
guarantees when the underlying assignment graph is sparse random regular and further is robust to
different initializations [9]. The modified version KOS + can be used in graphs where worker degrees are skewed, but we are still able to enhance its accuracy. Our results provide evidence for the
fact that existing random models are inadequate in capturing the behavior of workers in real-world
datasets, necessitating the need for a more general approach. Third, note that the hard penalty-based
algorithm outperforms the soft version across the datasets. Since the hard penalty algorithm works
well when adversaries have higher degrees (a fact noticed in the simulation results above), this suggests the presence of high-degree adversarial workers in theses datasets. Finally, our algorithm was
successful in identifying the following types of ?adversaries?: (1) uniform workers who always label 1 or 1 (in temp, task2, stage2), (2) malicious workers who mostly label incorrectly (in
bluebird, stage2) and (3) random workers who label each task independent of its true label
(such workers were identified as ?spammers? in [13]). Therefore, our algorithm is able to identify a
broad set of adversary strategies in addition to those detected by existing techniques.
5
Conclusions
This paper analyzes the problem of inferring true labels of tasks in crowdsourcing systems against
a broad class of adversarial labeling strategies. The main contribution is the design of a reputationbased worker filtering algorithm that uses a combination of disagreement-based penalties and optimal semi-matchings to identify adversarial workers. We show that our reputation scores are consistent with the existing notion of worker reliabilities and further can identify adversaries that employ
deterministic labeling strategies. Empirically, we show that our algorithm can be applied in real
crowd-sourced datasets to enhance the accuracy of existing label aggregation algorithms. Further,
we analyze the scenario of worst-case adversaries and establish lower bounds on the minimum
?damage? achievable by the adversaries.
Acknowledgments
We thank the anonymous reviewers for their valuable feedback. Ashwin Venkataraman was supported by the Center for Technology and Economic Development (CTED).
References
[1] B. Biggio, B. Nelson, and P. Laskov. Poisoning attacks against support vector machines. arXiv preprint
arXiv:1206.6389, 2012.
8
[2] N. Dalvi, A. Dasgupta, R. Kumar, and V. Rastogi. Aggregating crowdsourced binary ratings. In Proceedings of the 22nd international conference on World Wide Web, pages 285?294. International World Wide
Web Conferences Steering Committee, 2013.
[3] A. P. Dawid and A. M. Skene. Maximum likelihood estimation of observer error-rates using the em
algorithm. Applied Statistics, pages 20?28, 1979.
[4] G. Demartini, D. E. Difallah, and P. Cudr?e-Mauroux. Zencrowd: leveraging probabilistic reasoning
and crowdsourcing techniques for large-scale entity linking. In Proceedings of the 21st international
conference on World Wide Web, pages 469?478. ACM, 2012.
[5] J. S. Downs, M. B. Holbrook, S. Sheng, and L. F. Cranor. Are your participants gaming the system?:
screening mechanical turk workers. In Proceedings of the 28th international conference on Human factors
in computing systems, pages 2399?2402. ACM, 2010.
[6] A. Ghosh, S. Kale, and P. McAfee. Who moderates the moderators?: crowdsourcing abuse detection
in user-generated content. In Proceedings of the 12th ACM conference on Electronic commerce, pages
167?176. ACM, 2011.
[7] N. J. Harvey, R. E. Ladner, L. Lov?asz, and T. Tamir. Semi-matchings for bipartite graphs and load
balancing. In Algorithms and data structures, pages 294?306. Springer, 2003.
[8] P. G. Ipeirotis, F. Provost, and J. Wang. Quality management on amazon mechanical turk. In Proceedings
of the ACM SIGKDD workshop on human computation, pages 64?67. ACM, 2010.
[9] D. R. Karger, S. Oh, and D. Shah. Iterative learning for reliable crowdsourcing systems. Neural Information Processing Systems, 2011.
[10] A. Kittur, E. H. Chi, and B. Suh. Crowdsourcing user studies with mechanical turk. In Proceedings of the
SIGCHI conference on human factors in computing systems, pages 453?456. ACM, 2008.
[11] J. Le, A. Edmonds, V. Hester, and L. Biewald. Ensuring quality in crowdsourced search relevance evaluation: The effects of training question distribution.
[12] K. Lee, P. Tamilarasan, and J. Caverlee. Crowdturfers, campaigns, and social media: tracking and revealing crowdsourced manipulation of social media. In 7th international AAAI conference on weblogs and
social media (ICWSM), Cambridge, 2013.
[13] V. C. Raykar and S. Yu. Eliminating spammers and ranking annotators for crowdsourced labeling tasks.
The Journal of Machine Learning Research, 13:491?518, 2012.
[14] V. C. Raykar, S. Yu, L. H. Zhao, G. H. Valadez, C. Florin, L. Bogoni, and L. Moy. Learning from crowds.
The Journal of Machine Learning Research, 99:1297?1322, 2010.
[15] B. I. Rubinstein, B. Nelson, L. Huang, A. D. Joseph, S.-h. Lau, S. Rao, N. Taft, and J. Tygar. Antidote:
understanding and defending against poisoning of anomaly detectors. In Proceedings of the 9th ACM
SIGCOMM conference on Internet measurement conference, pages 1?14. ACM, 2009.
[16] P. Smyth, U. Fayyad, M. Burl, P. Perona, and P. Baldi. Inferring ground truth from subjective labelling of
venus images. Advances in neural information processing systems, pages 1085?1092, 1995.
[17] R. Snow, B. O?Connor, D. Jurafsky, and A. Y. Ng. Cheap and fast?but is it good?: evaluating non-expert
annotations for natural language tasks. In Proceedings of the conference on empirical methods in natural
language processing, pages 254?263. Association for Computational Linguistics, 2008.
[18] N. Tran, B. Min, J. Li, and L. Subramanian. Sybil-resilient online content voting. In Proceedings of
the 6th USENIX symposium on Networked systems design and implementation, pages 15?28. USENIX
Association, 2009.
[19] J. Vuurens, A. P. de Vries, and C. Eickhoff. How much spam can you take? an analysis of crowdsourcing
results to increase accuracy. In Proc. ACM SIGIR Workshop on Crowdsourcing for Information Retrieval
(CIR11), pages 21?26, 2011.
[20] J. B. Vuurens and A. P. de Vries. Obtaining high-quality relevance judgments using crowdsourcing.
Internet Computing, IEEE, 16(5):20?27, 2012.
[21] G. Wang, T. Wang, H. Zheng, and B. Y. Zhao. Man vs. machine: Practical adversarial detection of
malicious crowdsourcing workers. In 23rd USENIX Security Symposium, USENIX Association, CA, 2014.
[22] G. Wang, C. Wilson, X. Zhao, Y. Zhu, M. Mohanlal, H. Zheng, and B. Y. Zhao. Serf and turf: crowdturfing
for fun and profit. In Proceedings of the 21st international conference on World Wide Web, pages 679?688.
ACM, 2012.
[23] P. Welinder, S. Branson, S. Belongie, and P. Perona. The multidimensional wisdom of crowds. Advances
in Neural Information Processing Systems, 23:2424?2432, 2010.
[24] J. Whitehill, P. Ruvolo, T. Wu, J. Bergsma, and J. Movellan. Whose vote should count more: Optimal
integration of labels from labelers of unknown expertise. Advances in Neural Information Processing
Systems, 22(2035-2043):7?13, 2009.
9
| 5393 |@word version:5 briefly:1 pw:1 achievable:2 eliminating:1 nd:1 d2:1 simulation:2 pick:1 profit:1 reduction:1 contains:1 score:11 karger:1 document:1 outperforms:1 existing:11 subjective:1 com:1 assigning:1 must:1 cheap:1 remove:1 designed:3 v:1 fewer:2 guess:1 complementing:1 ruvolo:1 filtered:8 provides:9 detecting:1 node:2 characterization:2 attack:4 simpler:1 unbounded:1 become:1 symposium:2 incorrect:7 prove:2 consists:1 dalvi:1 baldi:1 introduce:2 manner:3 lov:1 behavior:4 themselves:1 growing:1 chi:1 automatically:1 actual:1 considering:1 becomes:1 provided:8 estimating:1 underlying:4 notation:3 moreover:1 medium:4 agnostic:1 kind:2 minimizes:1 affirmative:1 ghosh:1 guarantee:6 temporal:1 every:7 multidimensional:1 ti:3 voting:1 fun:1 exactly:2 indistinguishability:1 positive:3 service:1 aggregating:3 modify:1 tends:1 limit:3 despite:1 establishing:1 solely:1 abuse:1 bird:1 initialization:1 studied:1 examined:1 collect:1 challenging:1 suggests:1 jurafsky:1 branson:1 campaign:1 decided:1 unique:1 practical:2 acknowledgment:1 yj:5 commerce:1 practice:1 implement:1 rte:2 movellan:1 prevalence:4 empirical:3 significantly:2 revealing:1 matching:7 regular:11 refers:2 cted:2 get:1 cannot:3 scheduling:1 bh:9 impossible:1 deterministic:5 map:1 stage2:4 reviewer:1 center:1 go:2 straightforward:1 kale:1 independently:2 sigir:1 amazon:2 identifying:3 assigns:1 immediately:1 rule:11 insight:1 utilizing:1 oh:1 dw:1 population:3 proving:1 notion:4 hurt:1 limiting:1 resp:2 target:2 suppose:6 construction:1 user:2 exact:1 rty:1 anomaly:1 us:3 designing:1 smyth:1 agreement:1 dawid:1 labeled:10 observed:3 preprint:1 wang:4 capture:2 worst:8 wj:2 venkataraman:1 ordering:1 decrease:1 highest:4 valuable:1 balanced:3 environment:1 transforming:1 reward:1 asked:2 depend:2 raise:1 tight:2 algebra:2 smart:1 bipartite:8 matchings:7 represented:3 various:1 fast:1 describe:1 detected:1 rubinstein:1 labeling:24 aggregate:3 choosing:2 crowd:6 sourced:2 outcome:1 whose:8 quite:1 supplementary:1 say:2 otherwise:2 statistic:1 noisy:3 final:1 online:5 sequence:1 analytical:1 caverlee:1 propose:6 tran:1 relevant:2 networked:1 subgraph:1 gold:1 empty:1 object:1 help:1 fixing:1 unintentionally:1 ij:1 school:1 inflate:1 strong:2 recovering:1 c:3 auxiliary:1 implemented:3 convention:1 snow:1 correct:4 attribute:3 filter:3 human:3 transient:1 material:1 argued:1 resilient:1 taft:1 assign:5 anonymous:2 extension:1 weblogs:1 ground:2 adopt:3 crowdflower:1 purpose:1 estimation:1 proc:1 label:127 create:1 establishes:1 anecdotal:1 always:9 aim:3 modified:2 broader:2 wilson:1 conjunction:3 srikanth:1 focus:5 u2u:1 she:3 improvement:4 indicates:5 likelihood:1 adversarial:27 sigkdd:1 detect:1 i0:1 typically:2 eliminate:1 her:2 perona:2 labelings:1 issue:1 among:2 development:1 art:3 raised:1 fairly:1 tygar:1 integration:1 once:1 construct:1 having:1 ng:1 represents:1 broad:2 yu:2 minimized:1 report:2 employ:2 few:3 randomly:1 detection:2 screening:1 zheng:2 evaluation:1 analyzed:1 extreme:1 tj:19 amenable:1 edge:8 worker:206 hester:1 theoretical:7 instance:3 column:1 soft:19 rao:1 assignment:23 cost:3 tractability:1 vertex:5 subset:1 uniform:7 welinder:1 successful:2 inadequate:1 reported:2 answer:2 elevate:1 synthetic:3 combined:1 st:2 accrued:1 international:6 cited:1 fundamental:3 randomized:1 lee:1 probabilistic:5 pool:1 enhance:2 connecting:1 ym:1 p2k:2 thesis:1 aaai:1 management:1 choose:1 huang:1 administrator:1 worse:2 adversely:1 expert:1 zhao:4 valadez:1 li:1 distribute:2 potential:1 de:2 ioms:1 bold:1 lakshminarayanan:1 satisfy:1 mv:10 depends:4 ranking:1 later:1 break:1 lot:1 observer:1 analyze:4 start:1 aggregation:22 crowdsourced:4 complicated:2 participant:1 annotation:2 contribution:3 publicly:1 accuracy:12 who:22 likewise:1 judgment:4 identify:8 rastogi:1 wisdom:1 bayesian:1 identification:2 accurately:1 notoriously:1 expertise:1 classified:1 moderator:1 detector:1 definition:2 against:3 turk:4 intentionally:1 involved:2 resultant:1 proof:3 di:8 mi:1 sampled:3 dataset:2 popular:1 knowledge:1 sophisticated:9 carefully:1 higher:6 specify:1 entailment:1 evaluated:1 just:1 hand:1 sheng:1 web:4 trust:1 gaming:1 lack:1 google:1 quality:8 organically:1 effect:2 concept:3 true:39 y2:3 counterpart:1 normalized:1 hence:7 assigned:11 wi0:2 burl:1 symmetric:1 indistinguishable:1 skewed:1 raykar:2 uniquely:1 biggio:1 evident:1 demonstrate:2 confusion:1 necessitating:1 reasoning:1 image:2 recently:1 common:1 empirically:1 turf:1 extend:1 he:1 linking:1 association:3 refer:2 significant:2 measurement:1 bluebird:3 cambridge:1 ashwin:3 connor:1 rd:1 consistency:1 trivially:1 language:2 dj:3 reliability:18 access:2 etc:1 add:1 base:4 labelers:1 posterior:1 bergsma:1 recent:2 moderate:1 discard:1 scenario:7 manipulation:3 certain:2 harvey:1 binary:6 arbitrarily:2 captured:1 minimum:2 additional:1 analyzes:1 steering:1 employed:2 aggregated:1 paradigm:1 maximize:2 semi:13 ii:2 multiple:1 infer:6 ing:1 match:1 adapt:1 long:2 retrieval:1 sigcomm:1 parenthesis:1 impact:3 ensuring:1 variant:1 basic:1 ko:17 enhancing:1 essentially:1 metric:1 arxiv:2 iteration:1 represent:1 adopting:1 achieved:3 invert:1 penalize:2 receive:2 addition:2 whereas:1 malicious:13 biased:2 unlike:1 asz:1 subject:1 incorporates:1 leveraging:1 effectiveness:1 call:1 presence:8 ideal:1 iii:1 easy:1 enough:3 mcafee:1 affect:2 fit:1 isolation:1 florin:1 identified:4 restrict:1 opposite:2 economic:1 idea:5 venus:1 honest:41 whether:1 motivated:3 penalty:49 render:1 moy:1 spammer:6 york:2 cause:1 remark:1 amount:1 concentrated:1 category:1 generate:3 specifies:1 http:1 exist:2 percentage:3 estimated:2 popularity:1 correctly:4 per:1 broadly:1 edmonds:1 dasgupta:1 affected:6 abu:1 vuurens:2 key:3 recomputed:1 threshold:1 penalizing:1 graph:29 fraction:5 sum:1 run:1 prob:1 powerful:1 you:1 extends:1 almost:1 decide:1 electronic:1 eickhoff:1 wu:1 home:1 draw:1 decision:11 capturing:1 bound:11 hi:3 internet:2 laskov:1 distinguish:1 nontrivial:1 infinity:1 precisely:2 your:1 bp:1 fayyad:1 min:1 kumar:1 performing:2 poisoning:2 conjecture:1 skene:1 department:2 according:4 combination:4 remain:1 across:8 em:9 wi:20 joseph:1 temp:3 making:2 intuitively:2 invariant:1 sij:4 restricted:1 lau:1 computationally:3 agree:2 discus:1 loose:1 committee:1 defending:1 count:1 serf:1 available:1 generalizes:1 appropriate:1 disagreement:3 upto:3 sigchi:1 dwt:4 shah:1 existence:2 denotes:3 remaining:4 include:1 nlp:2 linguistics:1 establish:2 pollution:1 already:1 question:4 noticed:1 strategy:31 damage:3 dependence:1 degrades:1 exhibit:1 thank:1 mapped:1 simulated:1 entity:1 majority:2 degrade:1 evenly:1 topic:1 nelson:2 considers:1 consensus:1 reason:1 assuming:1 setup:2 mostly:2 whitehill:1 stated:1 negative:1 design:4 implementation:1 stern:2 unknown:2 perform:6 upper:4 ladner:1 observation:1 datasets:15 degm:2 benchmark:3 incorrectly:4 looking:1 y1:1 provost:1 arbitrary:5 usenix:4 inferred:4 rating:1 introduced:1 cast:1 mechanical:4 pair:2 required:1 sentence:1 security:1 conflict:9 textual:1 boost:2 address:2 beyond:1 adversary:62 able:5 below:1 pattern:7 proceeds:1 challenge:1 max:4 reliable:1 power:1 suitable:1 event:1 business:1 natural:3 subramanian:1 ipeirotis:1 task2:3 zhu:1 representing:1 scheme:1 improve:2 technology:1 inversely:1 numerous:1 irrespective:3 lij:4 prior:3 understanding:1 removal:1 law:1 expect:2 permutation:3 filtering:3 proportional:2 annotator:1 degree:17 incident:4 consistent:2 pi:4 balancing:2 atleast:2 normalizes:1 row:1 penalized:1 supported:1 parity:1 atmost:3 last:1 side:3 bias:2 wide:4 face:1 taking:2 differentiating:1 sparse:1 overcome:1 feedback:1 evaluating:1 world:10 valid:1 tamir:1 doesn:1 antidote:1 author:2 made:1 concretely:1 adopts:1 qualitatively:1 collection:1 spam:1 social:4 excess:1 implicitly:1 unreliable:2 deg:1 ml:1 decides:1 assumed:3 belongie:1 don:1 suh:1 pen:8 iterative:3 reputation:40 sk:5 search:1 table:2 reasonably:1 robust:3 ca:1 obtaining:1 did:1 pk:3 main:5 spread:1 osm:4 representative:3 site:3 aid:1 precision:6 inferring:5 explicit:2 lq:1 digg:1 third:3 theorem:19 removing:2 down:1 load:6 specific:6 nyu:3 dk:1 list:1 evidence:3 derives:1 exists:3 workshop:2 restricting:1 labelling:1 justifies:1 vries:2 gap:2 easier:1 tc:5 simply:1 bogoni:1 tracking:1 vulnerable:1 applies:1 springer:1 corresponds:1 truth:2 amt:2 acm:11 goal:6 identity:3 targeted:1 quantifying:1 towards:1 replace:1 absence:5 content:2 hard:19 man:1 specifically:3 colluding:1 wt:1 called:1 experimental:1 vote:2 formally:1 support:1 icwsm:1 relevance:2 evaluate:2 d1:2 crowdsourcing:18 |
4,853 | 5,394 | Feedback Detection for Live Predictors
Stefan Wager, Nick Chamandy, Omkar Muralidharan, and Amir Najmi
[email protected], {chamandy, omuralidharan, amir}@google.com
Stanford University and Google, Inc.
Abstract
A predictor that is deployed in a live production system may perturb the features
it uses to make predictions. Such a feedback loop can occur, for example, when a
model that predicts a certain type of behavior ends up causing the behavior it predicts, thus creating a self-fulfilling prophecy. In this paper we analyze predictor
feedback detection as a causal inference problem, and introduce a local randomization scheme that can be used to detect non-linear feedback in real-world problems. We conduct a pilot study for our proposed methodology using a predictive
system currently deployed as a part of a search engine.
1
Introduction
When statistical predictors are deployed in a live production environment, feedback loops can become a concern. Predictive models are usually tuned using training data that has not been influenced
by the predictor itself; thus, most real-world predictors cannot account for the effect they themselves
have on their environment. Consider the following caricatured example: A search engine wants to
train a simple classifier that predicts whether a search result is ?newsy? or not, meaning that the
search result is relevant for people who want to read the news. This classifier is trained on historical data, and learns that high click-through rate (CTR) has a positive association with ?newsiness.?
Problems may arise if the search engine deploys the classifier, and starts featuring search results that
are predicted to be newsy for some queries: promoting the search result may lead to a higher CTR,
which in turn leads to higher newsiness predictions, which makes the result be featured even more.
If we knew beforehand all the channels through which predictor feedback can occur, then detecting
feedback would not be too difficult. For example, in the context of the above example, if we knew
that feedback could only occur through some changes to the search result page that were directly
triggered by our model, then we could estimate feedback by running small experiments where we
turn off these triggering rules. However, in large industrial systems where networks of classifiers all
feed into each other, we can no longer hope to understand a priori all the ways in which feedback
may occur. We need a method that lets us detect feedback from sources we might not have even
known to exist.
This paper proposes a simple method for detecting feedback loops from unknown sources in live
systems. Our method relies on artificially inserting a small amount of noise into the predictions made
by a model, and then measuring the effect of this noise on future predictions made by the model. If
future model predictions change when we add artificial noise, then our system has feedback.
1
To understand how random noise can enable us to detect feedback, suppose that we have a model
with predictions y? in which tomorrow?s prediction y?(t+1) has a linear feedback dependence on today?s prediction y?(t) : if we increase y?(t) by , then y?(t+1) increases by
for some 2 R. Intuitively, we should be able to fit this slope by perturbing y?(t) with a small amount of noise
? ? N 0, ?2 and then regressing the new y?(t+1) against the noise; the reason least squares should
work here is that the noise ? is independent of all other variables by construction. The main contribution of this paper is to turn this simple estimation idea into a general procedure that can be used
to detect feedback in realistic problems where the feedback has non-linearities and jumps.
Counterfactuals and Causal Inference Feedback detection is a problem in causal inference. A
model suffers from feedback if the predictions it makes today affect the predictions it will make tomorrow. We are thus interested in discovering a causal relationship between today?s and tomorrow?s
predictions; simply detecting a correlation is not enough. The distinction between causal and associational inference is acute in the case of feedback: today?s and tomorrow?s predictions are almost
always strongly correlated, but this correlation by no means implies any causal relationship.
In order to discover causal relationships between consecutive predictions, we need to use some form
of randomized experimentation. In our case, we add a small amount of random noise to our predictions. Because the noise is fully artificial, we can reasonably ask counterfactual questions of the
type: ?How would tomorrow?s predictions have changed if we added more/less noise to the predictions today?? The noise acts as an independent instrument that lets us detect feedback. We frame
our analysis in terms of a potential outcomes model that asks how the world would have changed
had we altered a treatment; in our case, the treatment is the random noise we add to our predictions.
This formalism, often called the Rubin causal model [1], is regularly used for understanding causal
inference [2, 3, 4]. Causal models are useful for studying the behavior of live predictive systems on
the internet, as shown by, e.g., the recent work of Bottou et al. [5] and Chan et al. [6].
Outline and Contributions In order to define a rigorous feedback detection procedure, we need
to have a precise notion of what we mean by feedback. Our first contribution is thus to provide such
a model by defining statistical feedback in terms of a potential outcomes model (Section 2). Given
this feedback model, we propose a local noising scheme that can be used to fit feedback functions
with non-linearities and jumps (Section 4). Before presenting general version of our approach, however, we begin by discussing the linear case in Section 3 to elucidate the mathematics of feedback
detection: as we will show, the problem of linear feedback detection using local perturbations reduces to linear regression. Finally, in Section 5 we conduct a pilot study based on a predictive model
currently deployed as a part of a search engine.
2
Feedback Detection for Statistical Predictors
(t)
Suppose that we have a model that makes predictions y?i in time periods t = 1, 2, ... for examples
i = 1, ..., n. The predictive model itself is taken as given; our goal is to understand feedback effects
(t)
(t+1)
between consecutive pairs of predictions y?i and y?i
. We define statistical feedback in terms
(t+1)
(t)
of counterfactual reasoning: we want to know what would have happened to y?i
had y?i been
different. We use potential outcomes notation [e.g., 7] to distinguish between counterfactuals: let
(t+1) (t)
(t)
y?i
[yi ] be the predictions our model would have made at time t + 1 if we had published yi as
(t+1) (t)
(t)
our time-t prediction. In practice we only get to observe y?i
[yi ] for a single yi ; all other values
(t+1) (t)
(t+1)
of y?i
[yi ] are counterfactual. We also consider y
?i
[?], the prediction our model would have
made at time t + 1 if the model never made any of its predictions public and so did not have the
chance to affect its environment. With this notation, we define feedback as
(t)
(t+1)
feedbacki = y?i
2
(t)
[?
yi ]
(t+1)
y?i
[?],
(1)
i.e., the difference between the predictions our model actually made and the predictions it would
have made had it not had the chance to affect its environment by broadcasting predictions in the
past. Thus, statistical feedback is a difference in potential outcomes.
An additive feedback model In order to get a handle on feedback as defined above, we assume
(t+1) (t)
(t+1)
(t)
that feedback enters the model additively: y?i
[yi ] = y
?i
[?] + f (yi ), where f is a feedback
(t)
function, and yi is the prediction published at time t. In other words, we assume that the predictions
made by our model at time t + 1 are the sum of the prediction the model would have made if there
were no feedback, plus a feedback term that only depends on the previous prediction made by the
model. Our goal is to estimate the feedback function f .
(t)
(t+1)
Artificial noising for feedback detection The relationship between y?i and y?i
can be influenced by many things, such as trends, mean reversion, random fluctuations, as well as feedback. In
order to isolate the effect of feedback, we need to add some noise to the system to create a situation
that resembles a randomized experiment. Ideally, we might hope to sometimes turn our predic(t)
tive system off in order to get estimates of y?i [?]. However, predictive models are often deeply
integrated into large software systems, and it may not be clear what the correct system behavior
would be if we turned the predictor off. To side-step this concern, we randomize our system by
(t)
adding artificial noise to predictions: at time t, instead of deploying the prediction y?i , we deploy
(t)
(t)
(t)
(t) iid
y?i = y?i + ?i , where ?i ? N is artificial noise drawn from some distribution N . Because the
(t)
noise ?i is independent from everything else, it puts us in a randomized experimental setup that
(t+1)
(t)
allows us to detect feedback as a causal effect. If the time t + 1 prediction y?i
is affected by ?i ,
(t)
(t+1)
then our system must have feedback because the only way ?i can influence y?i
is through the
interaction between our model predictions and the surrounding environment at time t.
(t)
Local average treatment effect In practice, we want the noise ?i to be small enough that it does
not disturb the regular operation of the predictive model too much. Thus, our experimental setup
(t)
allows us to measure feedback as a local average treatment effect [4], where the artificial noise ?i
acts as a continuous treatment. Provided our additive model holds, we can then piece together these
local treatment effects into a single global feedback function f .
3
Linear Feedback
We begin with an analysis of linear feedback problems; the linear setup allows us to convey the main
insights with less technical overhead. We discuss the non-linear case in Section 4. Suppose that we
have some natural process x(1) , x(2) , ... and a predictive model of the form y? = w ? x. (Suppose for
notational convenience that x includes the constant, and the intercept term is folded into w.) For our
purposes, w is fixed and known; for example, w may have been set by training on historical data.
At some point, we ship a system that starts broadcasting the predictions y? = w ? x, and there is a
concern that the act of broadcasting the y? may perturb the underlying x(t) process. Our goal is to
(t+1) (t)
(t+1) (t)
detect any such feedback. Following earlier notation we write y?i
[?
yi ] = w ? x i
[?
yi ] for the
(t+1)
(t+1)
time t + 1 variables perturbed by feedback, and y?i
[?] = w ? xi
[?] for the counterparts we
would have observed without any feedback.
(t)
(t+1)
(t)
In this setup, any effect of y?i on xi
[?
yi ] is feedback. A simple way to constrain this relationship
(t+1) (t)
(t+1)
(t)
(t+1) (t)
is using a linear model xi
[?
yi ] = x i
[?] + y
?i . In other words, we assume that xi
[?
yi ]
(t)
is perturbed by an amount that scales linearly with y?i . Given this simple model, we find that:
(t+1)
y?i
(t)
[?
yi ]
(t+1)
= y?i
3
[?]
(t)
+ w ? y?i ,
(2)
and so f (y) =
y with
= w ? ; f is the feedback function we want to fit.
(t+1)
We cannot work with (2) directly, because y?i
[?] is not observed. In order to get around this
(t)
(t)
(t)
problem, we add artificial noise to our predictions: at time t, we publish predictions y?i = y?i +?i
(t)
instead of the raw predictions y?i . As argued in Section 2, this method lets us detect feedback
(t+1)
(t)
because y?i
can only depend on ?i through a feedback mechanism, and so any relationship
(t+1)
(t)
between y?i
and ?i must be a symptom of feedback.
(t)
(t+1)
A Simple Regression Approach With the linear feedback model (2), the effect of ?i on y?i
is
(t+1) (t)
(t+1) (t)
(t)
(t)
y?i
[?
yi +?i ] = y
?i
[?
yi ] + ?i . This relationship suggests that we should be able to recover
(t+1)
(t)
by regressing y?i
against the added noise ?i . The following result confirms this intuition.
(t)
Theorem 1. Suppose that (2) holds, and that we add noise ?i to our time t predictions. If we
estimate using linear least squares
h
i
h
i1
0
(t+1) (t)
d y?(t+1) [?yi(t) +?i(t) ], ? (t)
?
Cov
Var y?i
[?
yi ]
p ?
i
i
?=
A , (3)
h
i
, then n ?
) N @0,
2
d ? (t)
?
Var
i
h
i
(t)
where ?2 = Var ?i and n is the number of examples to which we applied our predictor.
Theorem 1 gives us a baseline understanding for the difficulty of the feedback detection problem:
the precision of our feedback estimates scales as the ratio of the artificial noise ?2 to natural noise
(t+1) (t)
Var[?
yi
[?
yi ]]. Note that the proof of Theorem 1 assumes that we only used predictions from a
(t+1) (t)
single time period t + 1 to fit feedback, and that the raw predictions y?i
[?
yi ] are all independent.
If we relax these assumptions we get a regression problem with correlated errors, and need to be
more careful with technical conditions.
(t+1)
(t)
Efficiency and Conditioning The simple regression model (3) treats the term y?i
[?
yi ] as noise.
(t)
(t+1) (t)
This is quite wasteful: if we know y?i we usually have a fairly good idea of what y?i
[?
yi ] should
be, and not using this information needlessly inflates the noise. Suppose that we knew the function1
h
i
(t+1) (t)
(t)
?(y) := E y?i
[?
yi ] y
?i = y .
(4)
Then, we could write our feedback model as
?
? ?
(t+1) (t)
(t)
(t+1) (t)
(t)
y?i
[?
yi +?i ] = ? y
?i
+ y?i
[?
yi ]
?
??
(t)
? y?i
+
(t)
?i ,
(5)
(t)
where ?(?
yi ) is a known offset. Extracting this offset improves the precision of our estimate for ?.
Theorem 2. Under the conditions of Theorem 1 suppose that the function ? from (4) is known and
(t+1)
(t)
that the y?i
are all independent of each other conditional on y?i . Then, given the information
available at time t, the estimate
h
?
?
i
d y?(t+1) [?yi(t) +?i(t) ] ? y?(t) , ? (t)
Cov
i
i
i
?? =
h
i
has asymptotic distribution
(6)
d ? (t)
Var
i
h
h
ii 1
0
(t+1) (t)
(t)
?
?
E
Var
y
?
[?
y
]
y
?
i
p
i
i
A.
(7)
n ??
) N @0,
2
?
1
In practice we do not know ?, but we can estimate it; see Section 4.
4
(t)
(t+1)
Moreover, if the variance of ?i = y?i
linear unbiased estimator of .
(t)
[?
yi ]
(t)
(t)
?(?
yi ) does not depend on y?i , then ?? is the best
Theorem 2 extends the general result from above that the precision with which we can estimate
feedback scales as the ratio of artificial noise to natural noise. The reason why ?? is more efficient
than ? is that we managed to condition away some of the natural noise, and reduced the variance of
our estimate for by
h ?
?i
h
i
h
h
ii
(t)
(t+1) (t)
(t+1) (t)
(t)
Var ? y?i
= Var y?i
[?
yi ]
E Var y?i
[?
yi ] y
?i
.
(8)
In other words, the variance reduction we get from ?? directly matches the amount of variability we
can explain away by conditioning. The estimator (6) is not practical as stated, because it requires
knowledge of the unknown function ? and is restricted to the case of linear feedback. In the next
section, we generalize this estimator into one that does not require prior knowledge of ? and can
handle non-linear feedback.
4
Fitting Non-Linear Feedback
Suppose now that we have the same setup as in the previous section, except that now feedback has
(t+1) (t)
(t+1)
(t)
a non-linear dependence on the prediction: y?i
[?
yi ] = y
?i
[?] + f (?
yi ) for some arbitrary
function f . For example, in the case of a linear predictive model y? = w ? x, this kind of feedback
(t+1) (t)
(t+1)
(t)
could arise if we have feature feedback xi
[?
yi ] = x i
[?] + f(x) (?
yi ); the feedback function
(t)
then becomes f (?) = w ? f(x) (?). When we add noise ?i to the above predictions, we only affect
the feedback term f (?):
?
?
?
?
(t+1) (t)
(t+1) (t)
(t)
(t)
(t)
(t)
y?i
[?
yi +?i ]
y?i
[?
yi ] = f y
?i + ?i
f y?i .
(9)
(t)
Thus, by adding artificial noise ?i , we are able to cancel out the nuisance terms, and isolate the
feedback function f that we want to estimate. We cannot use (9) in practice, though, as we can only
(t+1) (t)
(t+1) (t)
(t)
observe one of y?i
[?
yi +?i ] or y
?i
[?
yi ] in reality; the other one is counterfactual. We can get
(t)
around this problem by conditioning on y?i as in Section 3. Let
h
i
(t+1) (t)
(t)
(t)
? (y) = E y?i
[?
yi +?i ] y
?i = y
(10)
h
i
(t+1)
(t)
= t (y) + 'N ? f (y) , where t (y) = E y?i
[?] y
?i = y
is a term that captures trend effects that are not due to feedback. The ? denotes convolution:
h ?
?
i
(t)
(t)
(t)
(t)
'N ? f (y) = E f y?i + ?i
y?i = y with ?i ? N.
(11)
Using the conditional mean function ? we can write our expression of interest as
?
?
?
?
?
?
(t+1) (t)
(t)
(t)
(t)
(t)
(t)
(t)
y?i
[?
yi +?i ]
? y?i
= f y?i + ?i
'N ? f y?i
+ ?i ,
(12)
?
?
(t)
(t+1)
(t)
where ?i := y?i
[?]
t y?i . If we have a good idea of what ? is, the left-hand side can be
(t+1)
(t)
(t)
(t)
(t)
measured, as it only depends on y?i
[?
yi +?i ] and y
?i . Meanwhile, conditional on y?i , the first
(t)
(t)
(t)
two terms on the right-hand side only depend on ?i , while ?i is independent of ?i and mean(t)
zero. The upshot is that we can treat (12) as a regression problem where ?i is noise. In practice,
(t+1) (t)
(t)
(t)
we estimate ? from an auxiliary problem where we regress y?i
[?
yi +?i ] against y
?i .
5
A Pragmatic Approach There are many possible approaches to solving the non-parametric system of equations (12) for f [e.g., 8, Chapter 5]. Here, we take a pragmatic approach, and constrain
ourselves to solutions of the form ?
?(y) = ?? ? b? (y) and f?(y) = ?f ? bf (y), where b? : R ! Rp?
pf
and bf : R ! R are predetermined basis expansions. This approach transforms our problem
into an ordinary least-squares problem, and works well in terms of producing reasonable feedback
estimates in real-world problems (see Section 5). If this relation in fact holds for some values ?
and f , the result below shows that we can recover f by least-squares.
Theorem 3. Suppose that ? and f are defined as above, and that we have an unbiased estimator
?? of ? with variance V? = Var[ ?? ]. Then, if we fit f by least squares using (12) as described in
Appendix A, the resulting estimate ?f is unbiased and has variance
h i ?
? 1
?
? 1
Var ?f = Xf| Xf
Xf| VY + X? V? X?| Xf Xf| Xf
,
(13)
where the design matrices X? and Xf are defined as
0
1
0
1
..
..
.
.
B ?
B ?
?C
?
?
?C
B
B
(t) C
(t)
(t)
(t) C
|
X? = Bb|? y?i C and Xf = Bb|f y?i + ?i
('N ? bf ) y?i C
@
A
@
A
..
..
.
.
i
h
(t+1) (t)
(t)
and VY is a diagonal matrix with (VY )ii = Var y?i
[?
yi ] y
?i .
(14)
In the case where our spline model is misspecified, we can obtain a similar result using methods
due to Huber [9] and White [10]. In practice, we can treat ?? as known since fitting ?(?) is usually
easier than fitting f (?): estimating ?(?) is just a smoothing problem whereas estimating f (?) requires
(t)
fitting differences. If we also treat the errors ?i in (12) as roughly homoscedatic, (13) reduces to
h
h
ii
h i E Var y?i(t+1) [?yi(t) ] y?i(t)
?
?
?
?
(t)
(t)
(t)
Var ?f ?
,
where
s
=
b
y
?
+
?
'
?
b
y
?
. (15)
i
f
N
f
i
i
i
n E [ksi k22 ]
This simplified form again shows that the precision of our estimate of f (?) scales roughly as the ratio
(t)
of the variance of the artificial noise ?i to the variance of the natural noise.
Our Method in Practice For convenience, we summarize the steps needed to implement our
(t)
(t) iid
method here: (1) At time t, compute model predictions y?i and draw noise terms ?i ? N for
(t)
(t)
(t)
some noise distribution N . Deploy predictions y?i = y?i +??i ?in the live system. (2) Fit a non(t+1)
(t)
(t)
(t)
parametric least-squares regression of y?i
[?
yi +?i ] ? ? y
?i
to learn the function ? (y) :=
h
i
(t+1) (t)
(t)
(t)
E y?i
[?
yi +?i ] y
?i = y . We use the R formula notation, where a ? g(b) means that we
want to learn a function g(b) that predicts a. (3) Set up the non-parametric least-squares regression
problem
?
?
?
?
?
?
(t+1) (t)
(t)
(t)
(t)
(t)
(t)
y?i
[?
yi +?i ]
? y?i
? f y?i + ?i
'N ? f y?i ,
(16)
(t)
where the goal is to learn f . Here, 'N is the density of ?i , and ? denotes convolution. In Appendix
A we show how to carry out these steps using standard R libraries.
(t)
The resulting function f (y) is our estimate of feedback: If we make a prediction y?i at time t, then
(t)
(t)
(t)
our time t + 1 prediction will be boosted by f (?
yi ). The above equation only depends on y?i , ?i ,
6
(t+1)
(t)
(t)
and y?i
[?
yi +?i ], which are all quantities that can be observed in the context of an experiment
with noised predictions. Note that as we only fit f using the differences in (16), the intercept of f
is not identifiable. We fix the intercept (rather arbitrarily) by setting the average fitted feedback over
all training examples to 0; we do not include an intercept term in the basis bf .
Choice of Noising Distribution Adding noise to deployed predictions often has a cost that may
depend on the shape of the noise distribution N . A good choice of N should reflect this cost. For
example, if the practical cost of adding noise only depends on the largest amount of noise we ever
(t)
add, then it may be a good idea to draw ?i uniformly at random from {?"} for some " > 0. In our
(t)
experiments, we draw noise from a Gaussian distribution ?i ? N (0, ?2 ).
5
A Pilot Study
The original motivation for this research was to develop a methodology for detecting feedback in
real-world systems. Here, we present results from a pilot study, where we added signal to historical
data that we believe should emulate actual feedback. The reason for monitoring feedback on this
system is that our system was about to be more closely integrated with other predictive systems, and
there was a concern that the integration could induce bad feedback loops. Having a reliable method
for detecting feedback would provide us with an early warning system during the integration.
The predictive model in question is a logistic regression classifier. We added feedback to historical
(t)
(t)
data collected from log files according to half a dozen rules of the form ?if ai is high and y?i > 0,
(t+1)
(t)
then increase ai
by a random amount?; here y?i is the time-t prediction deployed by our system
(t)
(in log-odds space) and ai is some feature with a positive coefficient. These feedback generation
rules do not obey the additive assumption. Thus our model is misspecified in the sense that there
(t)
is no function f such that a current prediction y?i increased the log-odds of the next prediction by
(t)
f (?
yi ), and so this example can be taken as a stretch case for our method.
Our dataset had on the order of 100,000 data points, half of which were used for fitting the model
itself and half of which were used for feedback simulation. We generated data for 5 simulated time
periods, adding noise with ? = 0.1 at each step, and fit feedback using a spline basis discussed in
Appendix B. The ?true feedback? curve was obtained by fitting a spline regression to the additive
(t+1)
feedback model by looking at the unobservable y?i
[?]; we used a df = 5 natural spline with
knots evenly spread out on [ 9, 3] in log-odds space plus a jump at 0.
For our classifier of interest, we have fairly strong reasons to believe that the feedback function may
have a jump at zero, but probably shouldn?t have any other big jumps. Assuming that we know a
priori where to look for jumps does not seem to be too big a problem for the practical applications
we have considered. Results for feedback detection are shown in Figure 1. Although the fit is not
perfect, we appear to have successfully detected the shape of feedback. The error bars for estimated
feedback were obtained using a non-parametric bootstrap [11] for which we resampled pairs of
(current, next) predictions.
This simulation suggests that our method can be used to accurately detect feedback on scales that
may affect real-world systems. Knowing that we can detect feedback is reassuring from an engineering point of view. On a practical level, the feedback curve shown in Figure 1 may not be too
big a concern yet: the average feedback is well within the noise level of the classifier. But in largescale systems the ways in which a model interacts with its environment is always changing, and it
is entirely plausible that some innocuous-looking change in the future would increase the amount
of feedback. Our methodology provides us with a way to continuously monitor how feedback is
affected by changes to the system, and can alert us to changes that cause problems. In Appendix B,
we show some simulations with a wider range of effect sizes.
7
0.4
0.2
0.0
0.1
Feedback
0.3
True Feedback
Estimated Feedback
0.2
0.4
0.6
0.8
Prediction
Figure 1: Simulation aiming to replicate realistic feedback in a real-world classifier. The red solid
line is our feedback estimate; the black dashed line is the best additive approximation to the true
feedback. The x-axis shows predictions in probability space; the y axis shows feedback in logodds space. The error bars indicate pointwise confidence intervals obtained using a non-parametric
bootstrap with B = 10 replicates, and stretch 1 SE in each direction. Further experiments are
provided in Appendix B.
6
Conclusion
In this paper, we proposed a randomization scheme that can be used to detect feedback in real-world
predictive systems. Our method involves adding noise to the predictions made by the system; this
noise puts us in a randomized experimental setup that lets us measure feedback as a causal effect.
In general, the scale of the artificial noise required to detect feedback is smaller than the scale of
the natural predictor noise; thus, we can deploy our feedback detection method without disturbing
our system of interest too much. The method does not require us to make hypotheses about the
mechanism through which feedback may propagate, and so it can be used to continuously monitor
predictive systems and alert us if any changes to the system lead to an increase in feedback.
Related Work The interaction between models and the systems they attempt to describe has been
extensively studied across many fields. Models can have different kinds of feedback effects on their
environments. At one extreme of the spectrum, models can become self-fulfilling prophecies: for
example, models that predict economic growth may in fact cause economic growth by instilling
market confidence [12, 13]. At the other end, models may distort the phenomena they seek to
describe and therefore become invalid. A classical example of this is a concern that any metric used
to regulate financial risk may become invalid as soon as it is widely used, because actors in the
financial market may attempt to game the metric to avoid regulation [14]. However, much of the
work on model feedback in fields like finance, education, or macro-economic theory has focused on
negative results: there is an emphasis on understanding when feedback can happen and promoting
awareness about how feedback can interact with policy decisions, but there does not appear to be
much focus on actually fitting feedback. One notable exception is a paper by Akaike [15], who
showed how to fit cross-component feedback in a system with many components; however, he did
not add artificial noise to the system, and so was unable to detect feedback of a single component on
itself.
Acknowledgments The authors are grateful to Alex Blocker, Randall Lewis, and Brad Efron for
helpful suggestions and interesting conversations. S. W. is supported by a B. C. and E. J. Eaves
Stanford Graduate Fellowship.
8
References
[1] Paul W Holland. Statistics and causal inference. Journal of the American Statistical Association,
81(396):945?960, 1986.
[2] Joshua D Angrist, Guido W Imbens, and Donald B Rubin. Identification of causal effects using instrumental variables. Journal of the American Statistical Association, 91(434):444?455, 1996.
[3] Bradley Efron and David Feldman. Compliance as an explanatory variable in clinical trials. Journal of
the American Statistical Association, 86(413):9?17, 1991.
[4] Guido W Imbens and Joshua D Angrist. Identification and estimation of local average treatment effects.
Econometrica, 62(2):467?475, 1994.
[5] L?eon Bottou, Jonas Peters, Joaquin Qui?nonero-Candela, Denis X Charles, D Max Chickering, Elon Portugaly, Dipankar Ray, Patrice Simard, and Ed Snelson. Counterfactual reasoning and learning systems:
The example of computational advertising. Journal of Machine Learning Research, 14:3207?3260, 2013.
[6] David Chan, Rong Ge, Ori Gershony, Tim Hesterberg, and Diane Lambert. Evaluating online ad campaigns in a pipeline: Causal models at scale. In Proceedings of the 16th ACM SIGKDD International
Conference on Knowledge Discovery and Data Mining, pages 7?16. ACM, 2010.
[7] Donald B Rubin. Causal inference using potential outcomes. Journal of the American Statistical Association, 100(469):322?331, 2005.
[8] Trevor Hastie, Robert Tibshirani, and Jerome Friedman. The Elements of Statistical Learning. Springer
New York, second edition, 2009.
[9] Peter J Huber. The behavior of maximum likelihood estimates under nonstandard conditions. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, pages 221?233,
1967.
[10] Halbert White. A heteroskedasticity-consistent covariance matrix estimator and a direct test for heteroskedasticity. Econometrica: Journal of the Econometric Society, 48(4):817?838, 1980.
[11] Bradley Efron and Robert Tibshirani. An Introduction to the Bootstrap. CRC press, 1993.
[12] Robert K Merton. The self-fulfilling prophecy. The Antioch Review, 8(2):193?210, 1948.
[13] Fabrizio Ferraro, Jeffrey Pfeffer, and Robert I Sutton. Economics language and assumptions: How theories can become self-fulfilling. Academy of Management Review, 30(1):8?24, 2005.
[14] J?on Dan?elsson. The emperor has no clothes: Limits to risk modelling. Journal of Banking & Finance,
26(7):1273?1296, 2002.
[15] Hirotugu Akaike. On the use of a linear model for the identification of feedback systems. Annals of the
Institute of Statistical Mathematics, 20(1):425?439, 1968.
[16] Theodoros Evgeniou, Massimiliano Pontil, and Tomaso Poggio. Regularization networks and support
vector machines. Advances in Computational Mathematics, 13(1):1?50, 2000.
[17] Federico Girosi, Michael Jones, and Tomaso Poggio. Regularization theory and neural networks architectures. Neural Computation, 7(2):219?269, 1995.
[18] Peter J Green and Bernard W Silverman. Nonparametric Regression and Generalized Linear Models: A
Roughness Penalty Approach. Chapman & Hall London, 1994.
[19] Trevor Hastie and Robert Tibshirani. Generalized Additive Models. CRC Press, 1990.
[20] Grace Wahba. Spline Models for Observational Data. Siam, 1990.
[21] Stefanie Biedermann, Holger Dette, and David C Woods. Optimal design for additive partially nonlinear
models. Biometrika, 98(2):449?458, 2011.
[22] Werner G M?uller. Optimal design for local fitting.
55(3):389?397, 1996.
Journal of statistical planning and inference,
[23] William J Studden and D J VanArman. Admissible designs for polynomial spline regression. The Annals
of Mathematical Statistics, 40(5):1557?1569, 1969.
[24] Erich Leo Lehmann and George Casella. Theory of Point Estimation. Springer, second edition, 1998.
9
| 5394 |@word trial:1 version:1 polynomial:1 instrumental:1 replicate:1 bf:4 additively:1 confirms:1 simulation:4 propagate:1 seek:1 covariance:1 asks:1 solid:1 carry:1 reduction:1 tuned:1 past:1 bradley:2 current:2 com:1 yet:1 must:2 realistic:2 additive:7 happen:1 predetermined:1 shape:2 girosi:1 half:3 discovering:1 amir:2 detecting:5 provides:1 denis:1 theodoros:1 mathematical:2 alert:2 reversion:1 become:5 symposium:1 direct:1 tomorrow:5 jonas:1 overhead:1 fitting:8 ray:1 dan:1 introduce:1 huber:2 market:2 roughly:2 themselves:1 planning:1 tomaso:2 behavior:5 actual:1 pf:1 becomes:1 begin:2 discover:1 linearity:2 notation:4 provided:2 estimating:2 underlying:1 moreover:1 what:5 kind:2 clothes:1 warning:1 berkeley:1 act:3 growth:2 finance:2 biometrika:1 classifier:8 hirotugu:1 appear:2 producing:1 positive:2 before:1 engineering:1 local:8 treat:4 limit:1 aiming:1 sutton:1 fluctuation:1 might:2 plus:2 black:1 emphasis:1 resembles:1 studied:1 suggests:2 innocuous:1 campaign:1 range:1 graduate:1 practical:4 acknowledgment:1 practice:7 implement:1 silverman:1 bootstrap:3 procedure:2 pontil:1 logodds:1 featured:1 word:3 induce:1 regular:1 confidence:2 donald:2 get:7 cannot:3 convenience:2 noising:3 put:2 context:2 live:6 influence:1 intercept:4 risk:2 economics:1 focused:1 rule:3 insight:1 estimator:5 financial:2 handle:2 notion:1 annals:2 construction:1 suppose:9 today:5 elucidate:1 deploy:3 guido:2 us:1 akaike:2 hypothesis:1 trend:2 element:1 predicts:4 pfeffer:1 observed:3 enters:1 capture:1 noised:1 news:1 deeply:1 intuition:1 environment:7 deploys:1 ideally:1 econometrica:2 trained:1 depend:4 solving:1 grateful:1 predictive:13 efficiency:1 basis:3 chapter:1 emulate:1 surrounding:1 train:1 leo:1 massimiliano:1 describe:2 london:1 query:1 artificial:13 detected:1 outcome:5 quite:1 stanford:3 plausible:1 widely:1 relax:1 federico:1 cov:2 statistic:3 itself:4 patrice:1 online:1 triggered:1 propose:1 interaction:2 macro:1 inserting:1 causing:1 loop:4 relevant:1 turned:1 nonero:1 academy:1 disturb:1 perfect:1 muralidharan:1 wider:1 tim:1 develop:1 measured:1 strong:1 auxiliary:1 predicted:1 involves:1 implies:1 indicate:1 direction:1 closely:1 correct:1 enable:1 observational:1 public:1 everything:1 education:1 crc:2 argued:1 require:2 fix:1 randomization:2 roughness:1 rong:1 stretch:2 hold:3 around:2 considered:1 hall:1 predict:1 consecutive:2 early:1 purpose:1 estimation:3 currently:2 largest:1 create:1 successfully:1 stefan:1 hope:2 uller:1 always:2 gaussian:1 rather:1 avoid:1 boosted:1 focus:1 notational:1 modelling:1 likelihood:1 industrial:1 sigkdd:1 rigorous:1 baseline:1 detect:13 sense:1 helpful:1 inference:8 hesterberg:1 integrated:2 explanatory:1 relation:1 i1:1 interested:1 unobservable:1 priori:2 proposes:1 smoothing:1 integration:2 fairly:2 field:2 never:1 having:1 evgeniou:1 chapman:1 holger:1 prophecy:3 cancel:1 look:1 jones:1 future:3 spline:6 inflates:1 ourselves:1 jeffrey:1 william:1 attempt:2 friedman:1 detection:11 interest:3 mining:1 regressing:2 replicates:1 extreme:1 wager:1 beforehand:1 poggio:2 conduct:2 causal:16 halbert:1 fitted:1 increased:1 formalism:1 earlier:1 predic:1 measuring:1 werner:1 portugaly:1 ordinary:1 cost:3 predictor:11 too:5 perturbed:2 density:1 international:1 randomized:4 siam:1 off:3 michael:1 together:1 continuously:2 ctr:2 again:1 reflect:1 management:1 creating:1 american:4 simard:1 account:1 potential:5 includes:1 coefficient:1 inc:1 notable:1 depends:4 ad:1 piece:1 view:1 ori:1 candela:1 analyze:1 counterfactuals:2 red:1 start:2 recover:2 slope:1 contribution:3 square:7 variance:7 who:2 generalize:1 raw:2 identification:3 lambert:1 accurately:1 iid:2 knot:1 monitoring:1 advertising:1 published:2 nonstandard:1 explain:1 influenced:2 suffers:1 deploying:1 casella:1 ed:1 trevor:2 distort:1 against:3 regress:1 proof:1 pilot:4 dataset:1 treatment:7 merton:1 ask:1 counterfactual:5 knowledge:3 efron:3 improves:1 conversation:1 actually:2 feed:1 dette:1 higher:2 methodology:3 though:1 strongly:1 symptom:1 just:1 correlation:2 jerome:1 hand:2 joaquin:1 nonlinear:1 google:2 logistic:1 believe:2 effect:16 k22:1 unbiased:3 managed:1 counterpart:1 true:3 regularization:2 ferraro:1 read:1 white:2 during:1 self:4 nuisance:1 game:1 generalized:2 presenting:1 outline:1 reasoning:2 meaning:1 snelson:1 charles:1 misspecified:2 perturbing:1 conditioning:3 function1:1 association:5 discussed:1 he:1 feldman:1 ai:3 erich:1 mathematics:3 language:1 had:6 actor:1 longer:1 acute:1 heteroskedasticity:2 add:9 recent:1 chan:2 showed:1 ship:1 certain:1 arbitrarily:1 discussing:1 yi:53 joshua:2 george:1 period:3 signal:1 ii:4 dashed:1 reduces:2 technical:2 match:1 xf:8 cross:1 clinical:1 prediction:56 regression:11 metric:2 df:1 publish:1 sometimes:1 whereas:1 want:7 fellowship:1 interval:1 else:1 source:2 file:1 probably:1 isolate:2 compliance:1 thing:1 regularly:1 seem:1 odds:3 extracting:1 enough:2 affect:5 fit:10 hastie:2 architecture:1 click:1 triggering:1 economic:3 idea:4 wahba:1 knowing:1 whether:1 expression:1 angrist:2 penalty:1 peter:3 york:1 cause:2 useful:1 clear:1 se:1 amount:8 transforms:1 nonparametric:1 extensively:1 reduced:1 exist:1 vy:3 happened:1 estimated:2 fabrizio:1 tibshirani:3 write:3 affected:2 monitor:2 drawn:1 wasteful:1 changing:1 econometric:1 blocker:1 sum:1 shouldn:1 wood:1 lehmann:1 extends:1 almost:1 reasonable:1 draw:3 decision:1 appendix:5 qui:1 banking:1 entirely:1 internet:1 resampled:1 distinguish:1 identifiable:1 swager:1 occur:4 constrain:2 alex:1 software:1 according:1 eaves:1 smaller:1 across:1 randall:1 imbens:2 intuitively:1 restricted:1 fulfilling:4 taken:2 pipeline:1 equation:2 turn:4 discus:1 mechanism:2 needed:1 know:4 ge:1 instrument:1 end:2 studying:1 available:1 operation:1 experimentation:1 promoting:2 observe:2 obey:1 away:2 regulate:1 rp:1 original:1 assumes:1 running:1 denotes:2 include:1 eon:1 perturb:2 classical:1 society:1 question:2 added:4 quantity:1 randomize:1 parametric:5 dependence:2 diagonal:1 interacts:1 needlessly:1 grace:1 unable:1 simulated:1 evenly:1 collected:1 reason:4 assuming:1 pointwise:1 relationship:7 ratio:3 difficult:1 setup:6 regulation:1 robert:5 stated:1 negative:1 design:4 policy:1 unknown:2 convolution:2 defining:1 situation:1 variability:1 precise:1 ever:1 frame:1 looking:2 perturbation:1 arbitrary:1 tive:1 david:3 pair:2 required:1 nick:1 engine:4 distinction:1 able:3 bar:2 usually:3 below:1 summarize:1 reliable:1 max:1 green:1 natural:7 difficulty:1 largescale:1 scheme:3 altered:1 library:1 axis:2 stefanie:1 prior:1 understanding:3 upshot:1 discovery:1 review:2 asymptotic:1 fully:1 generation:1 suggestion:1 interesting:1 var:14 awareness:1 consistent:1 rubin:3 production:2 featuring:1 changed:2 supported:1 soon:1 side:3 understand:3 institute:1 fifth:1 feedback:118 curve:2 world:8 evaluating:1 author:1 made:11 jump:6 disturbing:1 simplified:1 historical:4 bb:2 global:1 knew:3 xi:5 spectrum:1 search:9 continuous:1 why:1 reality:1 channel:1 reasonably:1 learn:3 interact:1 expansion:1 diane:1 bottou:2 artificially:1 meanwhile:1 did:2 elon:1 main:2 spread:1 linearly:1 motivation:1 noise:47 arise:2 big:3 paul:1 edition:2 convey:1 deployed:6 precision:4 chickering:1 learns:1 admissible:1 dozen:1 theorem:7 formula:1 bad:1 offset:2 concern:6 adding:6 associational:1 ksi:1 easier:1 caricatured:1 broadcasting:3 simply:1 dipankar:1 brad:1 partially:1 holland:1 springer:2 chance:2 relies:1 lewis:1 acm:2 reassuring:1 conditional:3 goal:4 careful:1 invalid:2 change:6 folded:1 except:1 uniformly:1 called:1 bernard:1 experimental:3 exception:1 pragmatic:2 people:1 support:1 phenomenon:1 correlated:2 |
4,854 | 5,395 | DFacTo: Distributed Factorization of Tensors
S. V. N. Vishwanathan
Statistics and Computer Science
Purdue University
West Lafayette IN 47907
[email protected]
Joon Hee Choi
Electrical and Computer Engineering
Purdue University
West Lafayette IN 47907
[email protected]
Abstract
We present a technique for significantly speeding up Alternating Least Squares
(ALS) and Gradient Descent (GD), two widely used algorithms for tensor factorization. By exploiting properties of the Khatri-Rao product, we show how to
efficiently address a computationally challenging sub-step of both algorithms. Our
algorithm, DFacTo, only requires two sparse matrix-vector products and is easy
to parallelize. DFacTo is not only scalable but also on average 4 to 10 times faster
than competing algorithms on a variety of datasets. For instance, DFacTo only
takes 480 seconds on 4 machines to perform one iteration of the ALS algorithm
and 1,143 seconds to perform one iteration of the GD algorithm on a 6.5 million
? 2.5 million ? 1.5 million dimensional tensor with 1.2 billion non-zero entries.
1
Introduction
Tensor data appears naturally in a number of applications [1, 2]. For instance, consider a social
network evolving over time. One can form a users ? users ? time tensor which contains snapshots
of interactions between members of the social network [3]. As another example consider an online
store such as Amazon.com where users routinely review various products. One can form a users ?
items ? words tensor from the review text [4]. Similarly a tensor can be formed by considering the
various contexts in which a user has interacted with an item [5]. Finally, consider data collected by
the Never Ending Language Learner from the Read the Web project which contains triples of noun
phrases and the context in which they occur, such as, (?George Harrison?, ?plays?, ?guitars?) [6].
While matrix factorization and matrix completion have become standard tools that are routinely used
by practitioners, unfortunately, the same cannot be said about tensor factorization. The reasons are
not very hard to see: There are two popular algorithms for tensor factorization namely Alternating
Least Squares (ALS) (Appendix B), and Gradient Descent (GD) (Appendix C). The key step in
both algorithms is to multiply a matricized tensor and a Khatri-Rao product of two matrices (line 4
of Algorithm 2 and line 4 of Algorithm 3). However, this process leads to a computationallychallenging, intermediate data explosion problem. This problem is exacerbated when the dimensions
of tensor we need to factorize are very large (of the order of hundreds of thousands or millions), or
when sparse tensors contain millions to billions of non-zero entries. For instance, a tensor we formed
using review text from Amazon.com has dimensions of 6.5 million ? 2.5 million ? 1.5 million and
contains approximately 1.2 billion non-zero entries.
Some studies have identified this intermediate data explosion problem and have suggested ways of
addressing it. First, the Tensor Toolbox [7] uses the method of reducing indices of the tensor for
sparse datasets and entrywise multiplication of vectors and matrices for dense datasets. However, it
is not clear how to store data or how to distribute the tensor factorization computation to multiple
machines (see Appendix D). That is, there is a lack of distributable algorithms in existing studies.
Another possible strategy to solve the data explosion problem is to use GigaTensor [8]. Unfortunately, while GigaTensor does address the problem of parallel computation, it is relatively slow. To
1
summarize, existing algorithms for tensor factorization such as the excellent Tensor Toolbox of [7],
or the Map-Reduce based GigaTensor algorithm of [8] often do not scale to large problems.
In this paper, we introduce an efficient, scalable and distributed algorithm, DFacTo, that addresses
the data explosion problem. Since most large-scale real datasets are sparse, we will focus exclusively on sparse tensors. This is well justified because previous studies have shown that designing
specialized algorithms for sparse tensors can yield significant speedups [7]. We show that DFacTo
can be applied to both ALS and GD, and naturally lends itself to a distributed implementation.
Therefore, it can be applied to massive real datasets which cannot be stored and manipulated on a
single machine. For ALS, DFacTo is on average around 5 times faster than GigaTensor and around
10 times faster than the Tensor Toolbox on a variety of datasets. In the case of GD, DFacTo is
on average around 4 times faster than CP-OPT [9] from the Tensor Toolbox. On the Amazon.com
review dataset, DFacTo only takes 480 seconds on 4 machines to perform one iteration of ALS and
1,143 seconds to perform one iteration of GD.
As with any algorithm, there is a trade-off: DFacTo uses 3 times more memory than the Tensor
Toolbox, since it needs to store 3 flattened matrices as opposed to a single tensor. However, in
return, our algorithm only requires two sparse matrix-vector multiplications, making DFacTo easy to
implement using any standard sparse linear algebra library. Therefore, there are two merits of using
our algorithm: 1) computations are distributed in a natural way; and 2) only standard operations are
required.
2
Notation and Preliminaries
Our notation is standard, and closely follows [2]. Also see [1]. Lower case letters such as x denote
scalars, bold lower case letters such as x denote vectors, bold upper case letters such as X represent
matrices, and calligraphic letters such as X denote three-dimensional tensors.
The i-th element of a vector x is written as xi . In a similar vein, the (i, j)-th entry of a matrix
X is denoted as xi,j and the (i, j, k)-th entry of a tensor X is written as xi,j,k . Furthermore, xi,:
(resp. x:,i ) denotes the i-th row (resp. column) of X. We will use X?,: (resp. X:,? ) to denote the
sub-matrix of X which contains the rows (resp. columns) indexed by the set ?. For instance, if
? = {2, 4}, then X?,: is a matrix which contains the second and fourth rows of X. Extending the
above notation to tensors, we will write Xi,:,: , X:,j,: and X:,:,k to respectively denote the horizontal,
lateral and frontal slices of a third-order tensor X. The column, row, and tube fibers of X are given
by x:,j,k , xi,:,k , and xi,j,: respectively.
Sometimes a matrix or tensor may not be fully observed. We will use ?X or ?X respectively to
denote the set of indices corresponding to the observed (or equivalently non-zero) entries in a matrix
X
X or a tensor X. Extending this notation, ?X
i,: (resp. ?:,j ) denotes the set of column (resp. row)
indices corresponding to the observed entries in the i-th row (resp. j-th column) of X. We define
X
X
?X
i,:,: , ?:,j,: , and ?:,:,k analogously as the set of indices corresponding to the observed entries of the
i-th horizontal, j-th lateral, or k-th frontal slices of X. Also, nnzr(X) (resp. nnzc(X)) denotes the
number of rows (resp. columns) of X which contain at least one non-zero element.
X> denotes the transpose, X? denotes the Moore-Penrose pseudo-inverse, and kXk (resp. kXk)
denotes the Frobenius norm of a matrix X (resp. tensor X) [10]. Given a matrix A ? Rn?m , the
linear operator vec(A) yields a vector x ? Rnm , which is obtained by stacking the columns of A.
On the other hand, given a vector x ? Rnm , the operator unvec(n,m) (x) yields a matrix A ? Rn?m .
A ? B denotes the Kronecker product, A B the Khatri-Rao product, and A ? B the Hadamard
product of matrices A and B. The outer product of vectors a and b is written as a ? b (see e.g.,
[11]). Definitions of these standard matrix products can be found in Appendix A.
2.1
Flattening Tensors
Just like the vec(?) operator flattens a matrix, a tensor X may also be unfolded or flattened into a
matrix in three ways namely by stacking the horizontal, lateral, and frontal slices. We use Xn to
denote the n-mode flattening of a third-order tensor X ? RI?J?K ; X1 is of size I ? JK, X2 is of
size J ? KI, and X3 is of size K ? IJ. The following relationships hold between the entries of X
2
and its unfolded versions (see Appendix A.1 for an illustrative example):
xi,j,k = x1i,j+(k?1)J = x2j,k+(i?1)K = x3k,i+(j?1)I .
(1)
We can view X1 as consisting of K stacked frontal slices of X, each of size I ? J. Similarly, X2
consists of I slices of size J ? K and X3 is made up of J slices of size K ? I. If we use Xn,m to
denote the m-th slice in the n-mode flattening of X, then observe that the following holds:
x1i,j+(k?1)J = x1,k
i,j ,
x2j,k+(i?1)K = x2,i
j,k ,
x3k,i+(j?1)I = x3,j
k,i .
(2)
One can state a relationship between the rows and columns of various flattenings of a tensor, which
will be used to derive our distributed tensor factorization algorithm in Section 3. The proof of the
below lemma is in Appendix A.2.
0
Lemma 1 Let (n, n0 ) ? {(2, 1), (3, 2), (1, 3)}, and let Xn and Xn be the n and n0 -mode flattening
0
respectively of a tensor X. Moreover, let Xn,m be the m-th slice in Xn , and xnm,: be the m-th row
0
0
of Xn . Then, vec(Xn,m ) = xnm,: .
3
DFacTo
Recall that the main challenge of implementing ALS or GD for solving tensor factorization lies in
multiplying a matricized tensor and a Khatri-Rao product of two matrices: X1 (C B)1 . If B is
of size J ? R and C is of size K ? R, explicitly forming (C B) requires O(JKR) memory and
is infeasible when J and K are large. This is called the intermediate data explosion problem in the
literature [8]. The lemma below will be used to derive our efficient algorithm, which avoids this
problem. Although the proof can be inferred using results in [2], we give an elementary proof for
completeness.
Lemma 2 The r-th column of X1 (C B) can be computed as
i>
h
>
1
X (C B) :,r = unvec(K,I) X2 b:,r
c:,r
(3)
Proof We need to show that
i>
h
>
1
X (C B) :,r = unvec(K,I) X2 b:,r
c:,r
?
? > 2,1
b:,r X c:,r
?
?
..
=?
?.
.
2,I
c:,r
b>
:,r X
1
2,i
c:,r . Using (13)
Or equivalently it suffices to show that X (C B) i,r = b>
:,r X
2,i
>
2,i
vec b>
c:,r = c>
.
:,r X
:,r ? b:,r vec X
2,i
c:,r is a scalar. Moreover, using Lemma 1 we can write vec X
Observe that b>
:,r X
This allows us to rewrite the above equation as
>
2,i
b>
c:,r = x1i,: (c:,r ? b:,r ) = X1 (C B) i,r ,
:,r X
(4)
2,i
= x1i,: .
which completes the proof.
Unfortunately, a naive computation of X1 (C B) :,r by using (3) does not solve the intermediate
>
data explosion problem. This is because X2 b:,r produces a KI dimensional vector, which is
then reshaped by the unvec(K,I) (?) operator into a K ? I matrix. However, as the next lemma
>
asserts, only a small number of entries of X2 b:,r are non-zero.
For convenience, let a vector produced by (X2 )> b:,r be v:,r and a matrix produced by
>
unvec(K,I) (v:,r ) be Mr .
1
We mainly concentrate on the update to A since the updates to B and C are analogous.
3
Lemma 3 The number of non-zeros in v:,r is at most nnzr((X2 )> ) and nnzc(X2 ).
Proof Multiplying an all-zero row in (X2 )> and b:,r produces zero. Therefore, the number of nonzeros in v:,r is equal to the number of rows in (X2 )> that contain at least one non-zero element.
Also, by definition, nnzr((X2 )> ) is equal to nnzc(X2 ).
As a consequence of the above lemma, we only need to explicitly compute the non-zero entries of
>
v:,r . However, the problem of reshaping v:,r via the unvec(K,I) (?) operator still remains. The
next lemma shows how to overcome this difficulty.
Lemma 4 The location of the non-zero entries of Mr depends on (X2 )> and is independent of b:,r .
Proof The product of the (k + (i ? 1)K)-th row of (X2 )> and b:,r is the (k + (i ? 1)K)-th element
>
of v:,r . And, this element is the (i, k)-th entry of Mr by definition of unvec(K,I) (?) . Therefore,
if all the entries in the (k + (i ? 1)K)-th row of (X2 )> are zero, then the (i, k)-th entry of Mr is
zero regardless of b:,r . Consequently, the location of the non-zero entries of Mr is independent of
b:,r , and is only determined by (X2 )> .
Given X one can compute (X2 )> to know the locations of the non-zero entries of Mr . In other
words, we can infer the non-zero pattern and therefore preallocate memory for Mr . We will show
>
below how this allows us to perform the unvec(K,I) (?) operation for free.
Recall the Compressed Sparse Row (CSR) Format, which stores a sparse matrix as three arrays
namely values, columns, and rows. Here, values represents the non-zero values of the matrix; while
columns stores the column indices of the non-zero values. Also, rows stores the indices of the
columns array where each row starts. For example, if a sparse matrix Mr is
1 0 2
Mr =
,
0 3 4
then the CSR of Mr is
value(Mr ) = [ 1
2
3
4 ]
r
col(M ) = [ 0
2
1
2 ]
r
2
4 ].
row(M ) = [ 0
Different matrices with the same sparsity pattern can be represented by simply changing the entries
of the value array. For our particular case, what this means is that we can pre-compute col(Mr ) and
row(Mr ) and pre-allocate value(Mr ). By writing the non-zero entries of v:,r into value(Mr ) we
can ?reshape? v:,r into Mr .
? 2 )> . Then, Algorithm 1 shows the
Let the matrix with all-zero rows in (X2 )> removed be (X
1
? 2 )> , B, C, and
DFacTo algorithm for computing N := X (C B). Here, the input values are (X
r
2
>
? ) and b:,r directly
M preallocated in CSR format. By storing the results of the product of (X
into value(Mr ), we can obtain Mr because Mr was preallocated in the CSR format. Then, the
product of Mr and c:,r yields the r-th column of N. We obtain the output N by repeating these two
sparse matrix-vector products R times.
Algorithm 1: DFacTo algorithm for Tensor Factorization
? 2 )> , B, C, value(Mr ) col(Mr ), row(Mr )
1 Input: (X
2 Output: N
3 while r=1, 2,. . . , R do
? 2 )> b:,r
4
value(Mr ) ? (X
r
5
n:,r ? M c:,r
6 end
It is immediately obvious that using the above lemmas to compute
N requires no extra memory
other than storing Mr , which contains at most nnzc(X2 ) ? ?X non-zero entries. Therefore, we
4
completely avoid the intermediate data explosion problem. Moreover, the same subroutine can be
used for both ALS and GD (see Appendix E for detailed pseudo-code).
3.1
Distributed Memory Implementation
Our algorithm is easy to parallelize using a master-slave architecture of MPI(Message Passing Interface). At every iteration, the master transmits A, B, and C to the slaves. The slaves hold a fraction
of the rows of X2 using which a fraction of the rows of N is computed. By performing a synchronization step, the slaves can exchange rows of N. In ALS, this N is used to compute A which is
transmitted back to the master. Then, the master updates A, and the iteration proceeds. In GD, the
slaves transmit N back to the master, which computes ?A. Then, the master computes the step size
by a line search algorithm, updates A, and the iteration proceeds.
3.2
Complexity Analysis
A naive computation of N requires JK + ?X R flops; forming C
B requires JKR flops
1
?X R flops. Our algorithm
and performing the matrix-matrix
multiplication
X
(C
B)
requires
requires only nnzc(X2 ) + ?X R flops; ?X R flops for computing v:,r and nnzc(X2 )R flops
for computing Mr c:,r . Note that, typically, nnzc(X2 ) both JK and ?X (see Table 1). In terms
of memory, the naive algorithm requires O(JKR) extra memory, while our algorithm only requires
nnzc(X2 ) extra space to store Mr .
4
Related Work
Two papers that are most closely related to our work are the GigaTensor algorithm proposed by [8]
and the Sparse Tensor Toolbox of [7]. As discussed above, both algorithms attack the problem of
computing N
efficiently. In order to
computes two intermediate
matrices
compute n:,r , GigaTensor
>
>
1
1
N1 := X ? 1I (c:,r ? 1J )
and N2 := bin X ? 1I (1K ? b:,r ) . Next, N3 :=
N1 ? N
by computing N3 1JK . As reported in [8], GigaTensor
2 is computed, and n:,ris obtained
uses 2 ?X extra storage and 5 ?X flops to compute one column of N. The Sparse Tensor Toolbox
stores a tensor as a vector of non-zero values and a matrix of corresponding indices. Entries of B
and C are replicated appropriately to create intermediate vectors. A Hadamard product is computed
between the non-zero entries of the matrix and intermediate
and a selected set of
entries
vectors,
are summed to form columns of N. The algorithm uses 2 ?X extra storage and 5 ?X flops to
compute one column of N. See Appendix D for a detailed illustrative example which shows all the
intermediate calculations performed by our algorithm as well as the algorithm of [8] and [7].
Also, [9] suggests the gradient-based optimization of CANDECOMP/PARAFAC (CP) using the
same method as [7] to compute X1 (C B). [9] refers to this gradient-based optimization algorithm
as CPOPT and the ALS algorithm of CP using the method of [7] as CPALS. Following [9], we use
these names, CPALS and CPOPT.
5
Experimental Evaluation
Our experiments are designed to study the scaling behavior of DFacTo on both publicly available
real-world datasets as well as synthetically generated data. We contrast the performance of DFacTo
(ALS) with GigaTensor [8] as well as with CPALS [7], while the performance of DFacTo (GD) is
compared with CPOPT [9]. We also present results to show the scaling behavior of DFacTo when
data is distributed across multiple machines.
Datasets See Table 1 for a summary of the real-world datasets we used in our experiments. The
NELL-1 and NELL-2 datasets are from [8] and consists of (noun phrase 1, context, noun phrase 2)
triples from the ?Read the Web? project [6]. NELL-2 is a version of NELL-1, which is obtained by
removing entries whose values are below a threshold.
5
The Yelp Phoenix dataset is from the Yelp Data Challenge 2 , while Cellartracker, Ratebeer, Beeradvocate and Amazon.com are from the Stanford Network Analysis Project (SNAP) home page. All
these datasets consist of product or business reviews. We converted them into a users ? items ?
words tensor by first splitting the text into words, removing stop words, using Porter stemming [12],
and then removing user-item pairs which did not have any words associated with them. In addition,
for the Amazon.com dataset we filtered words that appeard less than 5 times or in fewer than 5
documents. Note that the number of dimensions as well as the number of non-zero entries reported
in Table 1 differ from those reported in [4] because of our pre-processing.
?
X
Dataset
I
J
K
? nnzc(X1 ) nnzc(X2 ) nnzc(X3 )
Yelp
45.97K
11.54K
84.52K
9.85M
4.32M
6.11M
229.83K
Cellartracker 36.54K 412.36K 163.46K
25.02M
19.23M
5.88M
1.32M
NELL-2
12.09K
9.18K
28.82K
76.88M
16.56M
21.48M
337.37K
Beeradvocate 33.37K
66.06K 204.08K
78.77M
18.98M
12.05M
1.57M
Ratebeer
29.07K 110.30K 294.04K
77.13M
22.40M
7.84M
2.85M
NELL-1
2.90M
2.14M
25.50M 143.68M
113.30M
119.13M
17.37M
Amazon
6.64M
2.44M
1.68M
1.22B
525.25M
389.64M
29.91M
Table 1: Summary statistics of the datasets used in our experiments.
We also generated the following two kinds of synthetic data for our experiments:
? the number of non-zero entries in the tensor is held fixed but we vary I, J, and K.
? the dimensions I, J, and K are held fixed but the number of non-zeros entries varies.
To simulate power law behavior, both the above datasets were generated using the following preferential attachment model [13]: the probability that a non-zero entry is added at index (i, j, k) is given
by pi ? pj ? pk , where pi (resp. pj and pk ) is proportional to the number of non-zero entries at index
i (resp. j and k).
Implementation and Hardware All experiments were conducted on a computing cluster where
each node has two 2.1 GHz 12-core AMD 6172 processors with 48 GB physical memory per node.
Our algorithms are implemented in C++ using the Eigen library3 and compiled with the Intel Compiler. We downloaded Version 2.5 of the Tensor Toolbox, which is implemented in MATLAB4 .
Since open source code for GigaTensor is not freely available, we developed our own version in
C++ following the description in [8]. Also, we used MPICH25 in order to distribute the tensor factorization computation to multiple machines. All our codes are available for download under an
open source license from http://www.joonheechoi.com/research.
Scaling on Real-World Datasets Both CPALS and our implementation of GigaTensor are uniprocessor codes. Therefore, for this experiment we restricted ourselves to datasets which can fit on a
single machine. When initialized with the same starting point, DFacTo and its competing algorithms
will converge to the same solution. Therefore, we only compare the CPU time per iteration of the
different algorithms. The results are summarized in Table 2. On many datasets DFacTo (ALS) is
around 5 times faster than GigaTensor and 10 times faster than CPALS; the differences are more
pronounced on the larger datasets. Also, DFacTo (GD) is around 4 times faster than CPOPT.
The differences in performance between DFacTo (ALS) and CPALS and between DFacTo (GD)
and CPOPT can partially be explained by the fact that DFacTo (ALS, GD) is implemented in C++
while CPALS and CPOPT use MATLAB. However, it must be borne in mind that both MATLAB
and our implementation use an optimized BLAS library to perform their computationally intensive
numerical linear algebra operations.
Compared to the Map-Reduce version implemented in Java and used for the experiments reported
in [8], our C++ implementation of GigaTensor is significantly faster and more optimized. As per [8],
2
https://www.yelp.com/dataset challenge/dataset
http://eigen.tuxfamily.org
4
http://www.sandia.gov/?tgkolda/TensorToolbox/
5
http://www.mpich.org/static/downloads/
3
6
Dataset
Yelp Phoenix
Cellartracker
NELL-2
Beeradvocate
Ratebeer
NELL-1
DFacTo (ALS)
9.52
23.89
32.59
43.84
44.20
322.45
GigaTensor
26.82
80.65
186.30
224.29
240.80
772.24
CPALS
46.52
118.25
376.10
364.98
396.63
-
DFacTo (GD)
13.57
35.82
80.79
94.85
87.36
742.67
CPOPT
45.9
130.32
386.25
481.06
349.18
-
Table 2: Times per iteration (in seconds) of DFacTo (ALS), GigaTensor, CPALS, DFacTo (GD), and
CPOPT on datasets which can fit in a single machine (R=10).
Machines
1
2
4
8
16
32
DFacTo (ALS)
NELL-1
Amazon
Iter.
CPU
Iter.
CPU
322.45 322.45
205.07 167.29
141.02 101.58 480.21 376.71
86.09
62.19 292.34 204.41
81.24
46.25 179.23
98.07
90.31
34.54 142.69
54.60
DFacTo (GD)
NELL-1
Amazon
Iter.
CPU
Iter.
CPU
742.67 104.23
492.38
55.11
322.65
28.55 1143.7 127.57
232.41
16.24 727.79
62.61
178.92
9.70 560.47
28.61
209.39
7.45 471.91
15.78
Table 3: Total Time and CPU time per iteration (in seconds) as a function of number of machines
for the NELL-1 and Amazon datasets (R=10).
the Java implementation took approximately 10,000 seconds per iteration to handle a tensor with
around 109 non-zero entries, when using 35 machines. In contrast, the C++ version was able to
handle one iteration of the ALS algorithm on the NELL-1 dataset on a single machine in 772 seconds. However, because DFacto (ALS) uses a better algorithm, it is able to handsomely outperform
GigaTensor and only takes 322 seconds per iteration.
Also, the execution time of DFacTo (GD) is longer than that of DFacTo (ALS) because DFacTo
(GD) spends more time on the line search algorithm to obtain an appropriate step size.
Scaling across Machines Our goal is to study scaling behavior of the time per iteration as datasets
are distributed across different machines. Towards this end we worked with two datasets. NELL-1
is a moderate-size dataset which our algorithm can handle on a single machine, while Amazon is a
large dataset which does not fit on a single machine. Table 3 shows that the iteration time decreases
as the number of machines increases on the NELL-1 and Amazon datasets. While the decrease in
iteration time is not completely linear, the computation time excluding both synchronization and
line search time decreases linearly. The Y-axis in Figure 1 indicates T4 /Tn where Tn is the single
iteration time with n machines on the Amazon dataset.
(a) DFacTo(ALS)
(b) DFacTo(GD)
Figure 1: The scalability of DFacTo with respect to the number of machines on the Amazon dataset
7
Synthetic Data Experiments We perform two experiments with synthetically generated tensor
data. In the first experiment we fix the number of non-zero entries to be 106 and let I = J = K
and vary the dimensions of the tensor. For the second experiment we fix the dimensions and let
I = J = K and the number of non-zero entries is set to be 2I. The scaling behavior of the three
algorithms on these two datasets is summarized in Table 4. Since we used a preferential attachment
model to generate the datasets, the non-zero indices exhibit a power law behavior. Consequently,
the number of columns with non-zero elements (nnzc(?)) for X1 , X2 and X3 is very close to the
total number of non-zero entries in the tensor. Therefore, as predicted by theory, DFacTo (ALS, GD)
does not enjoy significant speedups when compared to GigaTensor, CPALS and CPOPT. However,
it must be noted that DFacto (ALS) is faster than either GigaTensor or CPALS in all but one case and
DFacTo (GD) is faster than CPOPT in all cases. We attribute this to better memory locality which
arises as a consequence of reusing the memory for N as discussed in Section 3.
I=J =K
104
105
106
107
104
105
106
107
Non-zeros
106
106
106
106
2 ? 104
2 ? 105
2 ? 106
2 ? 107
DFacTo (ALS)
1.14
2.72
7.26
41.64
0.05
0.92
12.06
144.48
GigaTensor
2.80
6.71
11.86
38.19
0.09
1.61
22.08
251.89
CPALS
5.10
6.11
16.54
175.57
0.52
1.50
15.84
214.37
DFacTo (GD)
2.32
5.87
16.51
121.30
0.09
1.81
21.74
275.19
CPOPT
5.21
11.70
29.13
202.71
0.57
2.98
26.04
324.2
Table 4: Time per iteration (in seconds) on synthetic datasets (non-zeros = 106 or 2I, R=10)
Rank Variation Experiments Table 5 shows the time per iteration on various ranks (R) with the
NELL-2 dataset. We see that the computation time of our algorithm increases lineraly in R like the
time complexity analyzed in Section 3.2.
R
NELL-2
5
15.84
10
31.92
20
58.71
50
141.43
100
298.89
200
574.63
500
1498.68
Table 5: Time per iteration (in seconds) on various R
6
Discussion and Conclusion
We presented a technique for significantly speeding up the Alternating Least Squares (ALS) and the
Gradient Descent (GD) algorithm for tensor factorization by exploiting properties of the Khatri-Rao
product. Not only is our algorithm, DFacto, computationally attractive, but it is also more memory
efficient compared to existing algorithms. Furthermore, we presented a strategy for distributing the
computations across multiple machines.
We hope that the availability of a scalable tensor factorization algorithm will enable practitioners
to work on more challenging tensor datasets, and therefore lead to advances in the analysis and
understanding of tensor data. Towards this end we intend to make our code freely available for
download under a permissive open source license.
Although we mainly focused on tensor factorization using ALS and GD, it is worth noting that one
can extend the basic ideas behind DFacTo to other related problems such as joint matrix completion
and tensor factorization. We present such a model in Appendix F. In fact, we believe that this joint
matrix completion and tensor factorization model by itself is somewhat new and interesting in its
own right, despite its resemblance to other joint models including tensor factorization such as [14].
In our joint model, we are given a user ? item ratings matrix Y, and some side information such as
a user ? item ? words tensor X. Preliminary experimental results suggest that jointly factorizing
Y and X outperforms vanilla matrix completion. Please see Appendix F for details of the algorithm
and some experimental results.
8
References
[1] Age Smilde, Rasmus Bro, and Paul Geladi. Multi-way Analysis with Applications in the Chemical Sciences. John Wiley and Sons, Ltd, 2004.
[2] Tamara G. Kolda and Brett W. Bader. Tensor decompositions and applications. SIAM Review,
51(3):455?500, 2009.
[3] Jure Leskovec, Jon M. Kleinberg, and Christos Faloutsos. Graphs over time: densification
laws, shrinking diameters and possible explanations. In KDD, pages 177?187, 2005.
[4] J. McAuley and J. Leskovec. Hidden Factors and Hidden Topics: Understanding Rating Dimensions with Review Text. In Proceedings of the 7th ACM Conference on Recommender
Systems, pages 165?172, 2013.
[5] Alexandros Karatzoglou, Xavier Amatriain, Linas Baltrunas, and Nuria Oliver. Multiverse
recommendation: N-dimensional tensor factorization for context-aware collaborative filtering.
In Proceeedings of the 4th ACM Conference on Recommender Systems (RecSys), 2010.
[6] A. Carlson, J. Betteridge, B. Kisiel, B. Settles, E.R. Hruschka Jr., and T.M. Mitchell. Toward
an architecture for never-ending language learning. In In Proceedings of the Conference on
Artificial Intelligence (AAAI), 2010.
[7] Brett W. Bader and Tamara G. Kolda. Efficient matlab computations with sparse and factored
tensors. SIAM Journal on Scientific Computing, 30(1):205?231, 2007.
[8] U. Kang, Evangelos E. Papalexakis, Abhay Harpale, and Christos Faloutsos. Gigatensor: scaling tensor analysis up by 100 times - algorithms and discoveries. In Conference on Knowledge
Discovery and Data Mining, pages 316?324, 2012.
[9] Evrim Acar, Daniel M. Dunlavy, and Tamara G. Kolda. A scalable optimization approach
for fitting canonical tensor decompositions. Journal of Chemometrics, 25(2):67?86, February
2011.
[10] R. A. Horn and C. R. Johnson. Matrix analysis. Cambridge Univ Press, 1990.
[11] Dennis S. Bernstein. Matrix Mathematics. Princeton University Press, 2005.
[12] M. Porter. An algorithm for suffix stripping. Program, 14(3):130?137, 1980.
[13] A. Barabasi and R. Albert. Emergence of scaling in random networks. Science, 286:509?512,
1999.
[14] Evrim Acar, Tamara G. Kolda, and Daniel M. Dunlavy. All-at-once optimization for coupled matrix and tensor factorizations. In MLG?11: Proceedings of Mining and Learning with
Graphs, August 2011.
9
| 5395 |@word version:6 norm:1 open:3 decomposition:2 mcauley:1 contains:6 exclusively:1 daniel:2 document:1 outperforms:1 existing:3 com:7 nell:16 written:3 must:2 john:1 stemming:1 numerical:1 kdd:1 acar:2 designed:1 update:4 n0:2 intelligence:1 selected:1 fewer:1 item:6 core:1 filtered:1 alexandros:1 completeness:1 geladi:1 node:2 location:3 attack:1 org:2 become:1 xnm:2 consists:2 fitting:1 introduce:1 behavior:6 multi:1 nuria:1 unfolded:2 cpu:6 gov:1 considering:1 project:3 brett:2 notation:4 moreover:3 what:1 kind:1 spends:1 developed:1 pseudo:2 every:1 matlab4:1 dunlavy:2 enjoy:1 harpale:1 engineering:1 papalexakis:1 yelp:5 consequence:2 despite:1 parallelize:2 approximately:2 baltrunas:1 downloads:1 suggests:1 challenging:2 factorization:19 lafayette:2 horn:1 implement:1 x3:5 evolving:1 significantly:3 java:2 word:8 pre:3 refers:1 suggest:1 cannot:2 convenience:1 close:1 operator:5 hee:1 storage:2 context:4 writing:1 www:4 map:2 regardless:1 starting:1 focused:1 amazon:13 splitting:1 immediately:1 factored:1 array:3 handle:3 variation:1 analogous:1 transmit:1 resp:13 kolda:4 play:1 user:9 massive:1 us:5 designing:1 element:6 jk:4 vein:1 observed:4 electrical:1 thousand:1 trade:1 removed:1 decrease:3 complexity:2 solving:1 rewrite:1 algebra:2 learner:1 completely:2 joint:4 routinely:2 various:5 fiber:1 represented:1 stacked:1 univ:1 artificial:1 whose:1 widely:1 solve:2 stanford:1 snap:1 larger:1 compressed:1 bro:1 statistic:2 reshaped:1 jointly:1 itself:2 distributable:1 emergence:1 online:1 took:1 interaction:1 product:17 hadamard:2 description:1 frobenius:1 asserts:1 pronounced:1 scalability:1 chemometrics:1 exploiting:2 billion:3 interacted:1 cluster:1 extending:2 produce:2 derive:2 completion:4 stat:1 ij:1 exacerbated:1 implemented:4 predicted:1 differ:1 concentrate:1 closely:2 attribute:1 bader:2 enable:1 karatzoglou:1 settle:1 implementing:1 bin:1 exchange:1 suffices:1 fix:2 preliminary:2 opt:1 elementary:1 hold:3 around:6 vary:2 barabasi:1 create:1 tool:1 hope:1 evangelos:1 avoid:1 parafac:1 focus:1 rank:2 indicates:1 mainly:2 contrast:2 suffix:1 typically:1 hidden:2 subroutine:1 denoted:1 noun:3 summed:1 equal:2 aware:1 never:2 once:1 represents:1 jon:1 manipulated:1 consisting:1 ourselves:1 n1:2 message:1 mining:2 multiply:1 evaluation:1 analyzed:1 behind:1 held:2 oliver:1 explosion:7 preferential:2 indexed:1 initialized:1 leskovec:2 instance:4 column:18 rao:5 phrase:3 stacking:2 addressing:1 entry:33 hundred:1 conducted:1 johnson:1 stored:1 reported:4 stripping:1 varies:1 synthetic:3 gd:24 siam:2 off:1 analogously:1 aaai:1 tube:1 opposed:1 borne:1 return:1 reusing:1 distribute:2 converted:1 bold:2 summarized:2 availability:1 preallocated:2 explicitly:2 depends:1 performed:1 view:1 compiler:1 start:1 parallel:1 collaborative:1 square:3 formed:2 publicly:1 efficiently:2 yield:4 matricized:2 produced:2 multiplying:2 worth:1 processor:1 uniprocessor:1 khatri:5 definition:3 mlg:1 vishy:1 tamara:4 obvious:1 naturally:2 transmits:1 permissive:1 associated:1 proof:7 static:1 stop:1 dataset:13 popular:1 mitchell:1 recall:2 knowledge:1 back:2 appears:1 entrywise:1 furthermore:2 just:1 hand:1 horizontal:3 web:2 dennis:1 lack:1 porter:2 mode:3 resemblance:1 scientific:1 believe:1 name:1 contain:3 xavier:1 chemical:1 alternating:3 read:2 moore:1 attractive:1 please:1 illustrative:2 noted:1 mpi:1 tn:2 cp:3 interface:1 dfacto:44 specialized:1 physical:1 phoenix:2 million:8 discussed:2 blas:1 extend:1 significant:2 cambridge:1 vec:6 vanilla:1 mathematics:1 similarly:2 language:2 longer:1 compiled:1 own:2 moderate:1 store:8 calligraphic:1 transmitted:1 george:1 somewhat:1 mr:27 freely:2 converge:1 gigatensor:19 multiple:4 nonzeros:1 infer:1 faster:10 evrim:2 calculation:1 reshaping:1 scalable:4 basic:1 albert:1 iteration:20 sometimes:1 represent:1 justified:1 addition:1 harrison:1 completes:1 source:3 appropriately:1 extra:5 member:1 practitioner:2 noting:1 synthetically:2 intermediate:9 bernstein:1 easy:3 variety:2 fit:3 architecture:2 competing:2 identified:1 reduce:2 idea:1 intensive:1 allocate:1 gb:1 distributing:1 ltd:1 passing:1 matlab:3 clear:1 detailed:2 repeating:1 hardware:1 diameter:1 http:5 generate:1 outperform:1 canonical:1 per:11 write:2 key:1 iter:4 rnm:2 threshold:1 license:2 changing:1 linas:1 pj:2 graph:2 fraction:2 inverse:1 letter:4 fourth:1 master:6 home:1 appendix:10 scaling:8 ki:2 occur:1 vishwanathan:1 kronecker:1 worked:1 ri:1 x2:28 n3:2 kleinberg:1 simulate:1 performing:2 relatively:1 format:3 speedup:2 jr:1 across:4 son:1 amatriain:1 making:1 explained:1 restricted:1 computationally:3 equation:1 remains:1 know:1 merit:1 mind:1 end:3 sandia:1 available:4 operation:3 observe:2 appropriate:1 reshape:1 hruschka:1 faloutsos:2 eigen:2 denotes:7 carlson:1 february:1 tensor:66 intend:1 added:1 flattens:1 strategy:2 said:1 exhibit:1 gradient:5 lends:1 lateral:3 outer:1 recsys:1 amd:1 topic:1 collected:1 reason:1 toward:1 code:5 index:10 relationship:2 rasmus:1 equivalently:2 unfortunately:3 proceeedings:1 abhay:1 implementation:7 perform:7 upper:1 recommender:2 snapshot:1 datasets:26 purdue:4 descent:3 kisiel:1 flop:8 excluding:1 multiverse:1 rn:2 august:1 download:2 csr:4 inferred:1 rating:2 namely:3 required:1 toolbox:8 pair:1 optimized:2 smilde:1 kang:1 address:3 able:2 suggested:1 proceeds:2 below:4 pattern:2 jure:1 candecomp:1 sparsity:1 summarize:1 challenge:3 program:1 including:1 memory:11 explanation:1 power:2 natural:1 difficulty:1 business:1 library:2 attachment:2 axis:1 naive:3 coupled:1 speeding:2 text:4 review:7 literature:1 understanding:2 discovery:2 multiplication:3 law:3 synchronization:2 fully:1 interesting:1 proportional:1 filtering:1 triple:2 age:1 downloaded:1 storing:2 pi:2 row:24 summary:2 transpose:1 free:1 infeasible:1 side:1 sparse:15 distributed:8 slice:8 overcome:1 dimension:7 xn:8 ending:2 avoids:1 world:3 computes:3 ghz:1 made:1 replicated:1 beeradvocate:3 social:2 factorize:1 xi:8 factorizing:1 search:3 table:12 jkr:3 excellent:1 flattening:4 did:1 dense:1 main:1 joon:1 pk:2 linearly:1 paul:1 n2:1 x1:10 west:2 intel:1 slow:1 ratebeer:3 wiley:1 christos:2 sub:2 shrinking:1 slave:5 x1i:4 lie:1 col:3 third:2 removing:3 choi:1 densification:1 guitar:1 betteridge:1 consist:1 flattened:2 execution:1 t4:1 locality:1 simply:1 forming:2 penrose:1 kxk:2 partially:1 scalar:2 recommendation:1 acm:2 goal:1 tuxfamily:1 consequently:2 towards:2 hard:1 determined:1 reducing:1 lemma:11 x2j:2 called:1 total:2 experimental:3 arises:1 frontal:4 princeton:1 |
4,855 | 5,396 | Distributed Power-law Graph Computing:
Theoretical and Empirical Analysis
Ling Yan
Dept. of Comp. Sci. and Eng.
Shanghai Jiao Tong University
800 Dongchuan Road
Shanghai 200240, China
[email protected]
Cong Xie
Dept. of Comp. Sci. and Eng.
Shanghai Jiao Tong University
800 Dongchuan Road
Shanghai 200240, China
[email protected]
Wu-Jun Li
National Key Lab. for Novel Software Tech.
Dept. of Comp. Sci. and Tech.
Nanjing University
Nanjing 210023, China
[email protected]
Zhihua Zhang
Dept. of Comp. Sci. and Eng.
Shanghai Jiao Tong University
800 Dongchuan Road
Shanghai 200240, China
[email protected]
Abstract
With the emergence of big graphs in a variety of real applications like social
networks, machine learning based on distributed graph-computing (DGC) frameworks has attracted much attention from big data machine learning community.
In DGC frameworks, the graph partitioning (GP) strategy plays a key role to affect the performance, including the workload balance and communication cost.
Typically, the degree distributions of natural graphs from real applications follow
skewed power laws, which makes GP a challenging task. Recently, many methods
have been proposed to solve the GP problem. However, the existing GP methods
cannot achieve satisfactory performance for applications with power-law graphs.
In this paper, we propose a novel vertex-cut method, called degree-based hashing (DBH), for GP. DBH makes effective use of the skewed degree distributions
for GP. We theoretically prove that DBH can achieve lower communication cost
than existing methods and can simultaneously guarantee good workload balance.
Furthermore, empirical results on several large power-law graphs also show that
DBH can outperform the state of the art.
1
Introduction
Recent years have witnessed the emergence of big graphs in a large variety of real applications,
such as the web and social network services. Furthermore, many machine learning and data mining
algorithms can also be modeled with graphs [13]. Hence, machine learning based on distributed
graph-computing (DGC) frameworks has attracted much attention from big data machine learning
community [13, 15, 14, 6, 11, 7]. To perform distributed (parallel) graph-computing on clusters with
several machines (servers), one has to partition the whole graph across the machines in a cluster.
Graph partitioning (GP) can dramatically affect the performance of DGC frameworks in terms of
workload balance and communication cost. Hence, the GP strategy typically plays a key role in
DGC frameworks. The ideal GP method should minimize the cross-machine communication cost,
and simultaneously keep the workload in every machine approximately balanced.
1
Existing GP methods can be divided into two main categories: edge-cut and vertex-cut methods.
Edge-cut tries to evenly assign the vertices to machines by cutting the edges. In contrast, vertex-cut
tries to evenly assign the edges to machines by cutting the vertices. Figure 1 illustrates the edgecut and vertex-cut partitioning results of an example graph. In Figure 1 (a), the edges (A,C) and
(A,E) are cut, and the two machines store the vertex sets {A,B,D} and {C,E}, respectively. In
Figure 1 (b), the vertex A is cut, and the two machines store the edge sets {(A,B), (A,D), (B,D)}
and {(A,C), (A,E), (C,E)}, respectively. In edge-cut, both machines of a cut edge should maintain
a ghost (local replica) of the vertex and the edge data. In vertex-cut, all the machines associated
with a cut vertex should maintain a mirror (local replica) of the vertex. The ghosts and mirrors are
shown in shaded vertices in Figure 1. In edge-cut, the workload of a machine is determined by
the number of vertices located in that machine, and the communication cost of the whole graph is
determined by the number of edges spanning different machines. In vertex-cut, the workload of a
machine is determined by the number of edges located in that machine, and the communication cost
of the whole graph is determined by the number of mirrors of the vertices.
(a) Edge-Cut
(b) Vertex-Cut
Figure 1: Two strategies for graph partitioning. Shaded vertices are ghosts and mirrors, respectively.
Most traditional DGC frameworks, such as GraphLab [13] and Pregel [15], use edge-cut methods [9, 18, 19, 20] for GP. Very recently, the authors of PowerGraph [6] find that the vertex-cut
methods can achieve better performance than edge-cut methods, especially for power-law graphs. Hence, vertex-cut has attracted more and more attention from DGC research community. For
example, PowerGraph [6] adopts a random vertex-cut method and two greedy variants for GP.
GraphBuilder [8] provides some heuristics, such as the grid-based constrained solution, to improve
the random vertex-cut method.
Large natural graphs usually follow skewed degree distributions like power-law distributions, which
makes GP challenging. Different vertex-cut methods can result in different performance for powerlaw graphs. For example, Figure 2 (a) shows a toy power-law graph with only one vertex having
much higher degree than the others. Figure 2 (b) shows a partitioning strategy by cutting the vertices
{E, F, A, C, D}, and Figure 2 (c) shows a partitioning strategy by cutting the vertices {A, E}. We
can find that the partitioning strategy in Figure 2 (c) is better than that in Figure 2 (b) because the
number of mirrors in Figure 2 (c) is smaller which means less communication cost. The intuition
underlying this example is that cutting higher-degree vertices can result in fewer mirror vertices.
Hence, the power-law degree distribution can be used to facilitate GP. Unfortunately, existing vertexcut methods, including those in PowerGraph and GraphBuilder, make rarely use of the power-law
degree distribution for GP. Hence, they cannot achieve satisfactory performance in natural powerlaw graphs. PowerLyra [4] tries to combine both edge-cut and vertex-cut together by using the
power-law degree distribution. However, it is lack of theoretical guarantee.
(b)
(a)
(c)
Sample
Bad partitioning
Good partitioning
Figure 2: Partition a sample graph with vertex-cut.
2
In this paper, we propose a novel vertex-cut GP method, called degree-based hashing (DBH), for
distributed power-law graph computing. The main contributions of DBH are briefly outlined as
follows:
? DBH can effectively exploit the power-law degree distributions in natural graphs for vertexcut GP.
? Theoretical bounds on the communication cost and workload balance for DBH can be derived, which show that DBH can achieve lower communication cost than existing methods
and can simultaneously guarantee good workload balance.
? DBH can be implemented as an execution engine for PowerGraph [6], and hence all
PowerGraph applications can be seamlessly supported by DBH.
? Empirical results on several large real graphs and synthetic graphs show that DBH can
outperform the state-of-the-art methods.
2
Problem Formulation
Let G = (V, E) denote a graph, where V = {v1 , v2 , . . . , vn } is the set of vertices and E ? V ? V
is the set of edges in G. Let |V | denote the cardinality of the set V , and hence |V | = n. vi and vj are
called neighbors if (vi , vj ) ? E. The degree of vi is denoted as di , which measures the number of
neighbors of vi . Please note that we only need to consider the GP task for undirected graphs because
the workload mainly depends on the number of edges no matter directed or undirected graphs the
computation is based on. Even if the computation is based on directed graphs, we can also use the
undirected counterparts of the directed graphs to get the partitioning results.
Assume we have a cluster of p machines. Vertex-cut GP is to assign each edge with the two corresponding vertices to one of the p machines in the cluster. The assignment of an edge is unique, while
vertices may have replicas across different machines. For DGC frameworks based on vertex-cut GP,
the workload (amount of computation) of a machine is roughly linear in the number of edges located
in that machine, and the replicas of the vertices incur communication for synchronization. So the
goal of vertex-cut GP is to minimize the number of replicas and simultaneously balance the number
of edges on each machine.
Let M (e) ? {1, . . . , p} be the machine edge e ? E is assigned to, and A(v) ? {1, . . . , p} be
the span of vertex v over different machines. Hence, |A(v)| is the number of replicas of v among
different machines. Similar to PowerGraph [6], one of the replicas of a vertex is chosen as the master
and the others are treated as the mirrors of the master. We let M aster(v) denote the machine in
which the master of v is located. Hence, the goal of vertex-cut GP can be formulated as follows:
n
min
A
1X
|A(vi )|
n i=1
s.t. max |{e ? E | M (e) = m}| < ?
m
|E|
n
, and max |{v ? V | M aster(v) = m}| < ? ,
m
p
p
where m ? {1, . . . , p} denotes a machine, ? ? 1 and ? ? 1 are imbalance factors. We den
P
p
fine n1
|A(vi )| as replication factor, |E|
max |{e ? E | M (e) = m}| as edge-imbalance, and
p
n
i=1
m
max |{v ? V | M aster(v) = m}| as vertex-imbalance. To get a good balance of workload, ?
m
and ? should be as small as possible.
The degrees of natural graphs usually follow skewed power-law distributions [3, 1]:
Pr(d) ? d?? ,
where Pr(d) is the probability that a vertex has degree d and the power parameter ? is a positive
constant. The lower the ? is, the more skewed a graph will be. This power-law degree distribution makes GP challenging [6]. Although vertex-cut methods can achieve better performance than
edge-cut methods for power-law graphs [6], existing vertex-cut methods, such as random method in
PowerGraph and grid-based method in GraphBuilder [8], cannot make effective use of the powerlaw distribution to achieve satisfactory performance.
3
3
Degree-Based Hashing for GP
In this section, we propose a novel vertex-cut method, called degree-based hashing (DBH), to effectively exploit the power-law distribution for GP.
3.1
Hashing Model
We refer to a certain machine by its index idx, and the idxth machine is denoted as Pidx . We first define two kinds of hash functions: vertex-hash function idx = vertex hash(v) which hashes vertex
v to the machine Pidx , and edge-hash function idx = edge hash(e) or idx = edge hash(vi , vj )
which hashes edge e = (vi , vj ) to the machine Pidx .
Our hashing model includes two main components:
? Master-vertex assignment: The master replica of vi is uniquely assigned to one of the
p machines with equal probability for each machine by some randomized hash function
vertex hash(vi ).
? Edge assignment: Each edge e = (vi , vj ) is assigned to one of the p machines by some
hash function edge hash(vi , vj ).
It is easy to find that the above hashing model is a vertex-cut GP method. The master-vertex assignment can be easily implemented, which can also be expected to achieve a low vertex-imbalance
score. On the contrary, the edge assignment is much more complicated. Different edge-hash functions can achieve different replication factors and different edge-imbalance scores. Please note that
replication factor reflects communication cost, and edge-imbalance reflects workload-imbalance.
Hence, the key of our hashing model lies in the edge-hash function edge hash(vi , vj ).
3.2
Degree-Based Hashing
From the example in Figure 2, we observe that in power-law graphs the replication factor, which is
defined as the total number of replicas divided by the total number of vertices, will be smaller if we
cut vertices with relatively higher degrees. Based on this intuition, we define the edge hash(vi , vj )
as follows:
vertex hash(vi ) if di < dj ,
edge hash(vi , vj ) =
(1)
vertex hash(vj ) otherwise.
It means that we use the vertex-hash function to define the edge-hash function. Furthermore, the
edge-hash function value of an edge is determined by the degrees of the two associated vertices.
More specifically, the edge-hash function value of an edge is defined by the vertex-hash function
value of the associated vertex with a smaller degree. Hence, our method is called degree-based
hashing (DBH). DBH can effectively capture the intuition that cutting vertices with higher degrees
will get better performance.
Our DBH method for vertex-cut GP is briefly summarized in Algorithm 1, where [n] = {1, . . . , n}.
Algorithm 1 Degree-based hashing (DBH) for vertex-cut GP
Input: The set of edges E; the set of vertices V ; the number of machines p.
Output: The assignment M (e) ? [p] for each edge e.
1: Initialization: count the degree di for each i ? [n] in parallel
2: for all e = (vi , vj ) ? E do
3:
Hash each edge in parallel:
4:
if di < dj then
5:
M (e) ? vertex hash(vi )
6:
else
7:
M (e) ? vertex hash(vj )
8:
end if
9: end for
4
4
Theoretical Analysis
In this section, we present theoretical analysis for our DBH method. For comparison, the random vertex-cut method (called Random) of PowerGraph [6] and the grid-based constrained solution (called Grid) of GraphBuilder [8] are adopted as baselines. Our analysis is based on randomization. Moreover, we assume that the graph is undirected and there are no duplicated edges in the
graph. We mainly study the performance in terms of replication factor, edge-imbalance and verteximbalance defined in Section 2. Due to space limitation, we put the proofs of all theoretical results
into the supplementary material.
4.1
Partitioning Degree-fixed Graphs
Firstly, we assume that the degree sequence {di }ni=1 is fixed. Then we can get the following expected
replication factor produced by different methods.
Random assigns each edge evenly to the p machines via a randomized hash function. The result can
be directly got from PowerGraph [6].
Lemma 1. Assume that we have a sequence of n vertices {vi }ni=1 and the corresponding degree
sequence D = {di }ni=1 . A simple randomized vertex-cut on p machines has the expected replication
factor:
" n
#
n
1X
pX
1 di
E
|A(vi )|D =
1? 1?
.
n i=1
n i=1
p
?
By using the Grid hash function, each vertex
? has p rather than p candidate machines compared to
Random. Thus we simply replace p with p to get the following corollary.
Corollary 1. By using Grid for hashing, the expected replication factor on p machines is:
" n
# ? X
n
p
1X
1 di
E
|A(vi )|D =
1? 1? ?
.
n i=1
n i=1
p
Using DBH method in Section 3.2, we obtain the following result by fixing the sequence {hi }ni=1 ,
where hi is defined as the number of vi ?s adjacent edges which are hashed by the neighbors of vi
according to the edge-hash function defined in (1).
Theorem 1. Assume that we have a sequence of n vertices {vi }ni=1 and the corresponding degree
sequence D = {di }ni=1 . For each vi , di ? hi adjacent edges of it are hashed by vi itself. Define
H = {hi }ni=1 . Our DBH method on p machines has the expected replication factor:
#
" n
n
n
pX
1 hi +1
1 di
1X
pX
|A(vi )|H, D =
1? 1?
1? 1?
,
E
?
n i=1
n i=1
p
n i=1
p
where hi ? di ? 1 for any vi .
This theorem says that our DBH method has smaller expected replication factor than Random of
PowerGraph [6].
Next we turn to the analysis of the balance constraints. We still fix the degree sequence and have the
following result for our DBH method.
Theorem 2. Our DBH method on p machines with the sequences {vi }ni=1 , {di }ni=1 and {hi }ni=1
defined in Theorem 1 has the edge-imbalance:
n
P
P
hi
+
max
(di ? hi )
p
max |{e ? E | M (e) = m}|
j?[p] vi ?Pj
i=1
m
=
.
|E|/p
2|E|/p
Although the master vertices are evenly assigned to each machine, we want to show how the randomized assignment is close to the perfect balance. This problem is well studied in the model of
uniformly throwing n balls into p bins when n p(ln p)3 [17].
5
Lemma 2. The maximum number of master vertices for each machine is bounded as follows:
Pr[M axLoad > ka ] = o(1)
if a > 1,
Pr[M axLoad > ka ] = 1 ? o(1) if 0 < a < 1.
r
ln ln p
n
Here M axLoad = max |{v ? V | M aster(v) = m}|, and ka = p + 2npln p 1 ? 2a
ln p .
m
4.2
Partitioning Power-law Graphs
Now we change the sequence of fixed degrees into a sequence of random samples generated from
the power-law distribution. As a result, upper-bounds can be provided for the above three methods,
which are Random, Grid and DBH.
Theorem 3. Let the minimal degree be dmin and each d ? {di }ni=1 be sampled from a power-law
degree distribution with parameter ? ? (2, 3). The expected replication factor of Random on p
machines can be approximately bounded by:
" n
#
pX
1 ??
1 di
?p 1? 1?
,
1? 1?
ED
n i=1
p
p
? = dmin ?
where ?
??1
??2 .
This theorem says that when the degree sequence is under power-law distribution, the upper bound
of the expected replication factor increases as ? decreases. This implies that Random yields a worse
partitioning when the power-law graph is more skewed.
?
Like Corollary 1, we replace p with p to get the similar result for Grid.
Corollary 2. By using Grid method, the expected replication factor on p machines can be approximately bounded by:
"? n
#
pX
1 d i
1 ??
?
1? 1? ?
? p 1? 1? ?
,
ED
n i=1
p
p
? = dmin ?
where ?
Note that
?
??1
??2 .
p 1? 1?
?1
p
??
??
? p 1 ? 1 ? p1
. So Corollary 2 tells us that Grid can reduce
the replication factor but it is not motivated by the skewness of the degree distribution.
Theorem 4. Assume each edge is hashed by our DBH method and hi ? di ? 1 for any vi . The
expected replication factor of DBH on p machines can be approximately bounded by:
" n
#
pX
1 hi +1
1 ??0
EH,D
1? 1?
?p 1? 1?
,
n i=1
p
p
where ??0 = dmin ?
??1
??2
? dmin ?
??1
2??3
+ 12 .
Note that
1 ??
1 ??0
<p 1? 1?
.
p 1? 1?
p
p
??1
Therefore, our DBH method can expectedly reduce the replication factor. The term 2??3
increases
as ? decreases, which means our DBH reduces more replication factor when the power-law graph
is more skewed. Note that Grid and our DBH method actually use two different ways to reduce the
replication factor. Grid reduces more replication factor when p grows. These two approaches can
be combined to obtain further improvement, which is not the focus of this paper.
Finally, we show that our DBH methd also guarantees good edge-balance (workload balance) under
power-law distributions.
6
Theorem 5. Assume each edge is hashed by the DBH method with dmin , {vi }ni=1 , {di }ni=1 and
{hi }ni=1 defined
above. The vertices are evenly assigned. By taking the constant 2|E|/p =
n
P
ED
di = nED [d] /p, there exists ? (0, 1) such that the expected edge-imbalance of DBH
i=1
on p machines can be bounded w.h.p (with high probability). That is,
?
?
n
X
X
h
2|E|
i
EH,D ?
+ max
(di ? hi )? ? (1 + )
.
p
p
j?[p]
i=1
vi ?Pj
Note that any that satisfies 1/ n/p could work for this theorem, which results in a tighter
bound for large n. Therefore, together with Theorem 4, this theorem shows that our DBH method
can reduce the replication factor and simultaneously guarantee good workload balance.
5
Empirical Evaluation
In this section, empirical evaluation on real and synthetic graphs is used to verify the effectiveness
of our DBH method. The cluster for experiment contains 64 machines connected via 1 GB Ethernet.
Each machine has 24 Intel Xeon cores and 96GB of RAM.
5.1
Datasets
The graph datasets used in our experiments include both synthetic and real-world power-law graphs.
Each synthetic power-law graph is generated by a combination of two synthetic directed graphs. The
in-degree and out-degree of the two directed graphs are sampled from the power-law degree distributions with different power parameters ? and ?, respectively. Such a collection of synthetic graphs is
separated into two subsets: one subset with parameter ? ? ? which is shown in Table 1(a), and the
other subset with parameter ? < ? which is shown in Table 1(b). The real-world graphs are shown
in Table 1(c). Some of the real-world graphs are the same as those in the experiment of PowerGraph.
And some additional real-world graphs are from the UF Sparse Matrices Collection [5].
Table 1: Datasets
(a) Synthetic graphs: ? ? ? (b) Synthetic graphs: ? < ?
Alias
S1
S2
S3
S4
S5
S6
S7
S8
S9
5.2
?
2.2
2.2
2.2
2.2
2.1
2.1
2.1
2.0
2.0
?
2.2
2.1
2.0
1.9
2.1
2.0
1.9
2.0
1.9
|E|
71,334,974
88,305,754
134,881,233
273,569,812
103,838,645
164,602,848
280,516,909
208,555,632
310,763,862
Alias
S10
S11
S12
S13
S14
S15
?
2.1
2.0
2.0
1.9
1.9
1.9
|E|
88,617,300
135,998,503
145,307,486
280,090,594
289,002,621
327,718,498
?
2.2
2.2
2.1
2.2
2.1
2.0
(c) Real-world graphs
Alias
Tw
Arab
Wiki
LJ
WG
Graph
Twitter [10]
Arabic-2005 [5]
Wiki [2]
LiveJournal [16]
WebGoogle [12]
|V |
42M
22M
5.7M
5.4M
0.9M
|E|
1.47B
0.6B
130M
79M
5.1M
Baselines and Evaluation Metric
In our experiment, we adopt the Random of PowerGraph [6] and the Grid of GraphBuilder [8]1
as baselines for empirical comparison. The method Hybrid of PowerLyra [4] is not adopted for
comparison because it combines both edge-cut and vertex-cut which is not a pure vertex-cut method.
One important metric is the replication factor, which reflects the communication cost. To test the
speedup for real applications, we use the total execution time for PageRank which is forced to take
100 iterations. The speedup is defined as: speedup = 100% ? (?Alg ? ?DBH )/?Alg , where ?Alg is
the execution time of PageRank with the method Alg. Here, Alg can be Random or Grid. Because
all the methods can achieve good workload balance in our experiments, we do not report it here.
1
GraphLab 2.2 released in July 2013 has used PowerGraph as its engine, and the Grid GP method has been
adopted by GraphLab 2.2 to replace the original Random GP method. Detailed information can be found at:
http://graphlab.org/projects/index.html
7
5.3
Results
Figure 3 shows the replication factor on two subsets of synthetic graphs. We can find that our DBH
method achieves much lower replication factor than Random and Grid. The replication factor of
DBH is reduced by up to 80% compared to Random and 60% compared to Grid.
30
30
Random
Grid
DBH
Random
Grid
DBH
25
Replication Factor
Replication Factor
25
20
15
10
5
20
15
10
5
0
S1
S2
S3
S4
S5
S6
S7
S8
1+10?12
S9
0
S10
S11
S12
S13
S14
1+10?12
S15
(a) Replication Factor
(b) Replication Factor
Figure 3: Experiments on two subsets of synthetic graphs. The X-axis denotes different datasets in Table 1(a)
and Table 1(b). The number of machines is 48.
18
70
Random
Grid
DBH
16
60.6%
Figure 4 (a) shows the replication factor on the real-world graphs. We can also find that DBH
achieves the best performance. Figure 4 (b) shows that the relative speedup of DBH is up to 60%
over Random and 25% over Grid on the PageRank computation.
Random
Grid
60
10
0
WG
LJ
Wiki
Arab
1+10?12
Tw
0
(a) Replication Factor
WG
LJ
Wiki
25%
13.3%
4.28%
8.42%
20
2
6.06%
30
6
4
23.6%
40
26.5%
8
21.2%
10
31.5%
50
12
Speedup(%)
Replication Factor
14
Arab
1+10?12
Tw
(b) Execution Speedup
Figure 4: Experiments on real-world graphs. The number of machines is 48.
Figure 5 shows the replication factor and execution time for PageRank on Twitter graph when the
number of machines ranges from 8 to 64. We can find our DBH achieves the best performance for
all cases.
20
2000
Random
Grid
DBH
18
Random
Grid
DBH
1800
1600
Execution Time (Sec)
Replication Factor
16
14
12
10
8
6
4
1400
1200
1000
800
600
400
2
200
0
8
16
24
Number of Machines
48
64
1+10?12
(a) Replication Factor
8
16
24
Number of Machines
48
64
1+10?12
(b) Execution Time
Figure 5: Experiments on Twitter graph. The number of machines ranges from 8 to 64.
6
Conclusion
In this paper, we have proposed a new vertex-cut graph partitioning method called degree-based
hashing (DBH) for distributed graph-computing frameworks. DBH can effectively exploit the
power-law degree distributions in natural graphs to achieve promising performance. Both theoretical and empirical results show that DBH can outperform the state-of-the-art methods. In our
future work, we will apply DBH to more big data machine learning tasks.
7
Acknowledgements
This work is supported by the NSFC (No. 61100125, No. 61472182), the 863 Program of China
(No. 2012AA011003), and the Fundamental Research Funds for the Central Universities.
8
References
[1] Lada A Adamic and Bernardo A Huberman. Zipf?s law and the internet. Glottometrics, 3(1):143?150,
2002.
[2] Paolo Boldi and Sebastiano Vigna. The webgraph framework I: compression techniques. In Proceedings
of the 13th international conference on World Wide Web (WWW), 2004.
[3] Andrei Broder, Ravi Kumar, Farzin Maghoul, Prabhakar Raghavan, Sridhar Rajagopalan, Raymie Stata,
Andrew Tomkins, and Janet Wiener. Graph structure in the web. Computer networks, 33(1):309?320,
2000.
[4] Rong Chen, Jiaxin Shi, Yanzhe Chen, Haibing Guan, and Haibo Chen. Powerlyra: Differentiated graph
computation and partitioning on skewed graphs. Technical Report IPADSTR-2013-001, Institute of Parallel and Distributed Systems, Shanghai Jiao Tong University, 2013.
[5] Timothy A Davis and Yifan Hu. The University of Florida sparse matrix collection. ACM Transactions
on Mathematical Software, 38(1):1, 2011.
[6] Joseph E Gonzalez, Yucheng Low, Haijie Gu, Danny Bickson, and Carlos Guestrin. Powergraph: Distributed graph-parallel computation on natural graphs. In Proceedings of the 10th USENIX Symposium
on Operating Systems Design and Implementation (OSDI), 2012.
[7] Joseph E. Gonzalez, Reynold S. Xin, Ankur Dave, Daniel Crankshaw, Michael J. Franklin, and Ion Stoica.
GraphX: Graph processing in a distributed dataflow framework. In Proceedings of the 11th USENIX
Symposium on Operating Systems Design and Implementation (OSDI), 2014.
[8] Nilesh Jain, Guangdeng Liao, and Theodore L Willke. Graphbuilder: scalable graph etl framework. In
Proceedings of the First International Workshop on Graph Data Management Experiences and Systems,
2013.
[9] George Karypis and Vipin Kumar. Multilevel graph partitioning schemes. In Proceedings of the International Conference on Parallel Processing (ICPP), 1995.
[10] Haewoon Kwak, Changhyun Lee, Hosung Park, and Sue Moon. What is twitter, a social network or a
news media. In Proceedings of the 19th international conference on World Wide Web (WWW), 2010.
[11] Aapo Kyrola, Guy E. Blelloch, and Carlos Guestrin. Graphchi: Large-scale graph computation on just a
PC. In Proceedings of the 10th USENIX Symposium on Operating Systems Design and Implementation
(OSDI), 2012.
[12] Jure Leskovec. Stanford large network dataset collection. URL http://snap. stanford. edu/data/index.
html, 2011.
[13] Yucheng Low, Joseph Gonzalez, Aapo Kyrola, Danny Bickson, Carlos Guestrin, and Joseph M. Hellerstein. GraphLab: A new framework for parallel machine learning. In Proceedings of the Conference on
Uncertainty in Artificial Intelligence (UAI), 2010.
[14] Yucheng Low, Joseph Gonzalez, Aapo Kyrola, Danny Bickson, Carlos Guestrin, and Joseph M. Hellerstein. Distributed graphlab: A framework for machine learning in the cloud. In Proceedings of the
International Conference on Very Large Data Bases (VLDB), 2012.
[15] Grzegorz Malewicz, Matthew H Austern, Aart JC Bik, James C Dehnert, Ilan Horn, Naty Leiser, and
Grzegorz Czajkowski. Pregel: a system for large-scale graph processing. In Proceedings of the ACM
SIGMOD International Conference on Management of Data (SIGMOD), 2010.
[16] Alan Mislove, Massimiliano Marcon, Krishna P Gummadi, Peter Druschel, and Bobby Bhattacharjee.
Measurement and analysis of online social networks. In Proceedings of the 7th ACM SIGCOMM conference on Internet Measurement, 2007.
[17] Martin Raab and Angelika Steger. balls into binsa simple and tight analysis. In Randomization and
Approximation Techniques in Computer Science, pages 159?170. Springer, 1998.
[18] Isabelle Stanton and Gabriel Kliot. Streaming graph partitioning for large distributed graphs. In Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining
(KDD), 2012.
[19] Charalampos Tsourakakis, Christos Gkantsidis, Bozidar Radunovic, and Milan Vojnovic. Fennel: Streaming graph partitioning for massive scale graphs. In Proceedings of the 7th ACM International Conference
on Web Search and Data Mining (WSDM), 2014.
[20] Lu Wang, Yanghua Xiao, Bin Shao, and Haixun Wang. How to partition a billion-node graph. In Proceedings of the International Conference on Data Engineering (ICDE), 2014.
9
| 5396 |@word arabic:1 briefly:2 compression:1 vldb:1 hu:1 eng:3 contains:1 score:2 daniel:1 franklin:1 existing:6 ka:3 com:1 gmail:1 boldi:1 attracted:3 danny:3 partition:3 kdd:1 fund:1 bickson:3 hash:30 greedy:1 fewer:1 intelligence:1 core:1 provides:1 node:1 firstly:1 org:1 zhang:2 mathematical:1 symposium:3 replication:34 prove:1 combine:2 theoretically:1 expected:11 roughly:1 p1:1 wsdm:1 cardinality:1 etl:1 provided:1 project:1 underlying:1 moreover:1 bounded:5 medium:1 what:1 kind:1 skewness:1 bhattacharjee:1 guarantee:5 every:1 bernardo:1 partitioning:18 positive:1 nju:1 service:1 local:2 engineering:1 nsfc:1 approximately:4 steger:1 initialization:1 china:5 studied:1 ankur:1 theodore:1 challenging:3 shaded:2 range:2 karypis:1 directed:5 unique:1 horn:1 empirical:7 yan:1 got:1 road:3 get:6 nanjing:2 cannot:3 close:1 s9:2 janet:1 put:1 www:2 shi:1 attention:3 assigns:1 pure:1 powerlaw:3 s6:2 play:2 massive:1 located:4 cut:45 role:2 cloud:1 wang:2 capture:1 cong:1 connected:1 news:1 decrease:2 balanced:1 intuition:3 aster:4 angelika:1 tight:1 incur:1 gu:1 graphx:1 shao:1 workload:15 easily:1 druschel:1 haijie:1 jiao:4 separated:1 forced:1 effective:2 jain:1 massimiliano:1 artificial:1 tell:1 heuristic:1 supplementary:1 solve:1 stanford:2 say:2 snap:1 otherwise:1 wg:3 gp:30 emergence:2 itself:1 online:1 czajkowski:1 sequence:11 propose:3 achieve:11 sebastiano:1 milan:1 billion:1 cluster:5 prabhakar:1 perfect:1 andrew:1 fixing:1 implemented:2 c:1 implies:1 ethernet:1 raghavan:1 material:1 bin:2 multilevel:1 assign:3 fix:1 randomization:2 blelloch:1 tighter:1 rong:1 matthew:1 achieves:3 adopt:1 released:1 s12:2 reflects:3 rather:1 corollary:5 derived:1 focus:1 improvement:1 kwak:1 kyrola:3 mainly:2 seamlessly:1 tech:2 contrast:1 sigkdd:1 baseline:3 osdi:3 twitter:4 streaming:2 typically:2 lj:3 among:1 html:2 denoted:2 art:3 constrained:2 equal:1 having:1 park:1 future:1 others:2 report:2 simultaneously:5 national:1 maintain:2 n1:1 mining:3 stata:1 evaluation:3 pc:1 edge:60 experience:1 bobby:1 theoretical:7 minimal:1 leskovec:1 witnessed:1 xeon:1 assignment:7 cost:11 vertex:79 subset:5 synthetic:10 combined:1 fundamental:1 randomized:4 international:9 broder:1 lee:1 nilesh:1 michael:1 together:2 central:1 management:2 guy:1 worse:1 li:1 toy:1 s13:2 ilan:1 summarized:1 sec:1 includes:1 matter:1 jc:1 vi:33 depends:1 stoica:1 try:3 lab:1 s15:2 carlos:4 parallel:7 complicated:1 contribution:1 minimize:2 ni:14 wiener:1 moon:1 yield:1 produced:1 lu:1 lada:1 comp:4 dave:1 ed:3 james:1 associated:3 di:20 proof:1 sampled:2 dataset:1 dataflow:1 duplicated:1 knowledge:1 arab:3 actually:1 hashing:13 higher:4 xie:1 follow:3 leiser:1 formulation:1 furthermore:3 just:1 web:5 adamic:1 lack:1 grows:1 facilitate:1 verify:1 counterpart:1 hence:11 assigned:5 satisfactory:3 adjacent:2 skewed:8 uniquely:1 please:2 davis:1 vipin:1 novel:4 recently:2 shanghai:7 s8:2 refer:1 s5:2 measurement:2 austern:1 isabelle:1 zipf:1 grid:24 powergraph:14 outlined:1 dj:2 hashed:4 operating:3 pregel:2 base:1 recent:1 store:2 certain:1 server:1 s11:2 reynold:1 guestrin:4 krishna:1 additional:1 george:1 july:1 reduces:2 alan:1 graphchi:1 technical:1 cross:1 divided:2 gummadi:1 sigcomm:1 variant:1 scalable:1 aapo:3 liao:1 dgc:8 metric:2 sue:1 iteration:1 ion:1 want:1 fine:1 else:1 undirected:4 contrary:1 effectiveness:1 bik:1 ideal:1 easy:1 variety:2 affect:2 s14:2 reduce:4 cn:3 motivated:1 gb:2 url:1 s7:2 peter:1 dramatically:1 gabriel:1 detailed:1 amount:1 s4:2 liwujun:1 category:1 reduced:1 http:2 wiki:4 outperform:3 s3:2 paolo:1 key:4 pj:2 ravi:1 replica:9 v1:1 ram:1 graph:81 icde:1 year:1 master:8 uncertainty:1 wu:1 vn:1 gonzalez:4 bound:4 hi:13 internet:2 expectedly:1 constraint:1 throwing:1 s10:2 software:2 min:1 span:1 kumar:2 relatively:1 px:6 ned:1 uf:1 speedup:6 martin:1 according:1 ball:2 combination:1 across:2 smaller:4 tw:3 joseph:6 s1:2 den:1 pr:4 ln:4 turn:1 count:1 end:2 dehnert:1 adopted:3 apply:1 observe:1 v2:1 differentiated:1 hellerstein:2 florida:1 original:1 denotes:2 include:1 tomkins:1 exploit:3 sigmod:2 especially:1 webgraph:1 malewicz:1 strategy:6 traditional:1 sci:4 vigna:1 evenly:5 idx:4 sjtu:2 spanning:1 modeled:1 index:3 balance:13 willke:1 unfortunately:1 design:3 implementation:3 tsourakakis:1 perform:1 imbalance:10 upper:2 dmin:6 datasets:4 communication:12 dongchuan:3 community:3 usenix:3 grzegorz:2 engine:2 jure:1 yucheng:3 usually:2 ghost:3 rajagopalan:1 program:1 pagerank:4 including:2 max:8 power:30 natural:7 treated:1 eh:2 hybrid:1 scheme:1 improve:1 stanton:1 alias:3 axis:1 jun:1 acknowledgement:1 discovery:1 zh:1 relative:1 law:29 synchronization:1 limitation:1 degree:40 xiao:1 supported:2 institute:1 neighbor:3 wide:2 taking:1 sparse:2 distributed:11 world:9 author:1 adopts:1 collection:4 social:4 transaction:1 cutting:6 keep:1 mislove:1 graphlab:6 uai:1 yifan:1 search:1 table:6 promising:1 alg:5 vj:12 main:3 big:5 ling:1 whole:3 s2:2 sridhar:1 intel:1 andrei:1 tong:4 christos:1 lie:1 candidate:1 guan:1 theorem:11 bad:1 exists:1 livejournal:1 workshop:1 effectively:4 mirror:7 execution:7 illustrates:1 chen:3 timothy:1 simply:1 zhihua:1 radunovic:1 springer:1 satisfies:1 acm:5 vojnovic:1 goal:2 formulated:1 gkantsidis:1 replace:3 change:1 determined:5 specifically:1 uniformly:1 huberman:1 lemma:2 called:8 total:3 xin:1 rarely:1 dept:4 |
4,856 | 5,397 | Scalable Nonlinear Learning with
Adaptive Polynomial Expansions
Alina Beygelzimer
Yahoo! Labs
[email protected]
Alekh Agarwal
Microsoft Research
[email protected]
Daniel Hsu
Columbia University
[email protected]
John Langford
Microsoft Research
[email protected]
Matus Telgarsky?
Rutgers University
[email protected]
Abstract
Can we effectively learn a nonlinear representation in time comparable to linear
learning? We describe a new algorithm that explicitly and adaptively expands
higher-order interaction features over base linear representations. The algorithm
is designed for extreme computational efficiency, and an extensive experimental
study shows that its computation/prediction tradeoff ability compares very favorably against strong baselines.
1
Introduction
When faced with large datasets, it is commonly observed that using all the data with a simpler
algorithm is superior to using a small fraction of the data with a more computationally intense but
possibly more effective algorithm. The question becomes: What is the most sophisticated algorithm
that can be executed given a computational constraint?
At the largest scales, Na?ve Bayes approaches offer a simple, easily distributed single-pass algorithm. A more computationally difficult, but commonly better-performing approach is large scale
linear regression, which has been effectively parallelized in several ways on real-world large scale
datasets [1, 2]. Is there a modestly more computationally difficult approach that allows us to commonly achieve superior statistical performance?
The approach developed here starts with a fast parallelized online learning algorithm for linear models, and explicitly and adaptively adds higher-order interaction features over the course of training,
using the learned weights as a guide. The resulting space of polynomial functions increases the
approximation power over the base linear representation at a modest increase in computational cost.
Several natural folklore baselines exist. For example, it is common to enrich feature spaces with ngrams or low-order interactions. These approaches are naturally computationally appealing, because
these nonlinear features can be computed on-the-fly avoiding I/O bottlenecks. With I/O bottlenecked
datasets, this can sometimes even be done so efficiently that the additional computational complexity
is negligible, so improving over this baseline is quite challenging.
The design of our algorithm is heavily influenced by considerations for computational efficiency, as
discussed further in Section 2. Several alternative designs are plausible but fail to provide adequate
computation/prediction tradeoffs or even outperform the aforementioned folklore baselines. An
extensive experimental study in Section 3 compares efficient implementations of these baselines with
?
This work was performed while MT was visiting Microsoft Research, NYC.
1
Relative error vs time tradeoff
1.0
linear
quadratic
cubic
apple(0.125)
apple(0.25)
apple(0.5)
apple(0.75)
apple(1.0)
relative error
0.8
0.6
0.4
0.2
0.0
?0.2
0
10
1
10
relative time
2
10
Figure 1: Computation/prediction tradeoff points using non-adaptive polynomial expansions and
adaptive polynomial expansions (apple). The markers are positioned at the coordinate-wise median of (relative error, relative time) over 30 datasets, with bars extending to 25th and 75th
percentiles. See Section 3 for definition of relative error and relative time used here.
the proposed mechanism and gives strong evidence of the latter?s dominant computation/prediction
tradeoff ability (see Figure 1 for an illustrative summary).
Although it is notoriously difficult to analyze nonlinear algorithms, it turns out that two aspects
of this algorithm are amenable to analysis. First, we prove a regret bound showing that we can
effectively compete with a growing feature set. Second, we exhibit simple problems where this
algorithm is effective, and discuss a worst-case consistent variant. We point the reader to the full
version [3] for more details.
Related work. This work considers methods for enabling nonlinear learning directly in a highlyscalable learning algorithm. Starting with a fast algorithm is desirable because it more naturally
allows one to improve statistical power by spending more computational resources until a computational budget is exhausted. In contrast, many existing techniques start with a (comparably) slow
method (e.g., kernel SVM [4], batch PCA [5], batch least-squares regression [5]), and speed it up by
sacrificing statistical power, often just to allow the algorithm to run at all on massive data sets.
A standard alternative to explicit polynomial expansions is to employ polynomial kernels with the
kernel trick [6]. While kernel methods generally have computation scaling at least quadratically
with the number of training examples, a number of approximations schemes have been developed to
enable a better tradeoff. The Nystr?m method (and related techniques) can be used to approximate
the kernel matrix while permitting faster training [4]. However, these methods still suffer from the
drawback that the model size after n examples is typically O(n). As a result, even single pass online
implementations [7] typically suffer from O(n2 ) training and O(n) testing time complexity.
Another class of approximation schemes for kernel methods involves random embeddings into a
high (but finite) dimensional Euclidean space such that the standard inner product there approximates the kernel function [8?11]. Recently, such schemes have been developed for polynomial kernels [9?11] with computational scaling roughly linear in the polynomial degree. However, for many
sparse, high-dimensional datasets (such as text data), the embedding of [10] creates dense, high dimensional examples, which leads to a substantial increase in computational complexity. Moreover,
neither of the embeddings from [9, 10] exhibits good statistical performance unless combined with
dense linear dimension reduction [11], which again results in dense vector computations. Such feature construction schemes are also typically unsupervised, while the method proposed here makes
use of label information.
Among methods proposed for efficiently learning polynomial functions [12?16], all but [13] are
batch algorithms. The method of [13] uses online optimization together with an adaptive rule for
creating interaction features. A variant of this is discussed in Section 2 and is used in the experimental study in Section 3 as a baseline.
2
Algorithm 1 Adaptive Polynomial Expansion (apple)
input Initial features S1 = {x1 , . . . , xd }, expansion sizes (sk ), epoch schedule (?k ), stepsizes (?t ).
1: Initial weights w 1 := 0, initial epoch k := 1, parent set P1 := ?.
2: for t = 1, 2, . . . : do
3:
Receive stochastic gradient g t .
4:
Update weights: wt+1 := wt ? ?t [g t ]Sk ,
where [?]Sk denotes restriction to monomials in the feature set Sk .
5:
if t = ?k then
6:
Let Mk ? Sk be the top sk monomials m(x) ? Sk such that m(x) ?
/ Pk , ordered from
highest-to-lowest by the weight magnitude in wt+1 .
7:
Expand feature set: Sk+1 := Sk ? {xi ? m(x) : i ? [d], m(x) ? Mk }, and
Pk+1 := Pk ? {m(x) : m(x) ? Mk }.
8:
k := k + 1.
9:
end if
10: end for
2
Adaptive polynomial expansions
This section describes our new learning algorithm, apple.
2.1
Algorithm description
The pseudocode is given in Algorithm 1. The algorithm proceeds as stochastic gradient descent over
the current feature set to update a weight vector. At specified times ?k , the feature set Sk is expanded
to Sk+1 by taking the top monomials in the current feature set, ordered by weight magnitude in the
current weight vector, and creating interaction features between these monomials and x. Care is
exercised to not repeatedly pick the same monomial for creating higher order monomial by tracking
a parent set Pk , the set of all monomials for which higher degree terms have been expanded. We
provide more intuition for our choice of this feature growing heuristic in Section 2.3.
There are two benefits to this staged process. Computationally, the stages allow us to amortize the
cost of the adding of monomials?which is implemented as an expensive dense operation?over
several other (possibly sparse) operations. Statistically, using stages guarantees that the monomials
added in the previous stage have an opportunity to have their corresponding parameters converge.
We have found it empirically effective to set sk := average k[g t ]S1 k0 , and to update the feature set
at a constant number of equally-spaced times over the entire course of learning. In this case, the
number of updates (plus one) bounds the maximum degree of any monomial in the final feature set.
2.2
Shifting comparators and a regret bound for regularized objectives
Standard regret bounds compare the cumulative loss of an online learner to the cumulative loss of a
single predictor (comparator) from a fixed comparison class. Shifting regret is a more general notion
of regret, where the learner is compared to a sequence of comparators u1 , u2 , . . . , uT .
Existing shifting regret bounds can be used to loosely justify the use of online gradient descent
PT
methods over
feature spaces [17]. These bounds are roughly of the form t=1 ft (wt ) ?
p expanding
P
ft (ut ) . T t<T kut ? ut+1 k, where ut is allowed to use the same features available to wt ,
and ft is the convex cost function in step t. This suggests a relatively high cost for a substantial
total change in the comparator, and thus in the feature space. Given a budget, one could either do a
liberal expansion a small number of times, or opt for including a small number of carefully chosen
monomials more frequently. We have found that the computational cost of carefully picking a small
number of high quality monomials is often quite high. With computational considerations at the
forefront, we will prefer a more liberal but infrequent expansion. This also effectively exposes the
learning algorithm to a large number of nonlinearities quickly, allowing their parameters to jointly
converge between the stages.
It is natural to ask if better guarantees are possible under some structure on the learning problem.
Here, we consider the stochastic setting (rather than the harsher adversarial setting of [17]), and
3
further assume that our objective takes the form
f (w) := E[`(hw, xyi)] + ?kwk2 /2,
(1)
where the expectation is under the (unknown) data generating distribution D over (x, y) ? S ? R,
and ` is some convex loss function on which suitable restrictions will be placed. Here S is such
that S1 ? S2 ? . . . ? S, based on the largest degree monomials we intend to expand. We assume
that in round t, we observe a stochastic gradient of the objective f , which is typically done by first
sampling (xt , yt ) ? D and then evaluating the gradient of the regularized objective on this sample.
This setting has some interesting structural implications over the general setting of online learning
with shifting comparators. First, the fixed objective f gives us a more direct way of tracking the
change in comparator through f (ut ) ? f (ut+1 ), which might often be milder than kut ? ut+1 k.
In particular, if ut = arg minu?Sk f (u) in epoch k, for a nested subspace sequence Sk , then we
immediately obtain f (ut+1 ) ? f (ut ). Second, the strong convexity of the regularized objective
enables the possibility of faster O(1/T ) rates than prior work [17]. Indeed, in this setting, we obtain
the following stronger result. We use the shorthand Et [?] to denote the conditional expectation at
time t, conditioning over the data from rounds 1, . . . , t ? 1.
Theorem 1. Let a distribution over (x, y), twice differentiable convex loss ` with ` ? 0 and
max{`0 , `00 } ? 1, and a regularization parameter ? > 0 be given. Recall the definition (1) of
the objective f . Let (wt , g t )t?1 be as specified by apple with step size ?t := 1/(?(t + 1)), where
Et ([g t ]S(t) ) = [?f (wt )]S(t) and S(t) is the support set corresponding to epoch kt at time t in
apple. Then for any comparator sequence (ut )?
t=1 satisfying ut ? S(t) , for any fixed T ? 1,
!
PT
2
(X + ?)(X + ?D)2
1
t=1 (t + 2)f (ut )
E f (wT +1 ) ?
,
?
PT
T +1
2?2
t=1 (t + 2)
where X ? maxt kxt yt k and D ? maxt max{kwt k, kut k}.
Quite remarkably, the result exhibits no dependence on the cumulative shifting of the comparators
unlike existing bounds [17]. This is the first result of this sort amongst shifting bounds to the best of
our knowledge, and the only one that yields 1/T rates of convergence even with strong convexity. Of
course, we limit ourselves to the stochastic setting, and prove expected regret guarantees on the final
PT
predictor wT as opposed to a bound on t=1 f (wt )/T . A curious distinction is our comparator,
which is a weighted average of f (ut ) as opposed to the more standard uniform average. Recalling
that f (ut+1 ) ? f (ut ) in our setting, this is a strictly harder benchmark than an unweighted average
and overemphasizes the later comparator terms which are based on larger support sets. Indeed, this
is a nice compromise between competing against uT , which is the hardest yardstick, and u1 , which
is what a standard non-shifting analysis compares to. Indeed our improvement can be partially
attributed to the stability of the averaged f values as opposed to just f (uT ) (more details in [3]).
Overall, this result demonstrates that in our setting, while there is generally a cost to be paid for
shifting the comparator too much, it can still be effectively controlled in favorable cases. One
problem for future work is to establish these fast 1/T rates also with high probability.
Note that the regret bound offers no guidance on how or when to select new monomials to add.
2.3
Feature expansion heuristics
Previous work on learning sparse polynomials [13] suggests that it is possible to anticipate the utility
of interaction features before even evaluating them. For instance, one of the algorithms from [13]
orders monomials m(x) by an estimate of E[r(x)2 m(x)2 ]/E[m(x)2 ], where r(x) = E[y|x]? f?(x)
is the residual of the current predictor f? (for least-squares prediction of the label y). Such an index
is shown to be related to the potential error reduction by polynomials with m(x) as a factor. We call
this the SSM heuristic (after the authors of [13], though it differs from their original algorithm).
Another plausible heuristic, which we use in Algorithm 1, simply orders the monomials in Sk by
their weight magnitude in the current weight vector. We can justify this weight heuristic in the
following Q
simple example. Suppose a target function E[y|x] is just a single monomial in x, say,
m(x) := i?M xi for some M ? [d], and that x has a product distribution over {0, 1}d with 0 <
E[xi ] =: p ? 1/2 for all i ? [d]. Suppose we repeatedly perform 1-sparse regression with the current
4
feature set Sk , and pick the top weight magnitude monomial for inclusion in the parent set Pk+1 . It
is easy to show that the weight on a degree ` sub-monomial of m(x) in this regression is p|M |?` , and
the weight is strictly smaller for any term which is not a proper sub-monomial of m(x). Thus we
repeatedly pick the largest available sub-monomial of m(x) and expand it, eventually discovering
m(x). After k stages of the algorithm, we have at most kd features in our regression here, and
hence we find m(x) with a total of d|M | variables in our regression, as opposed to d|M | which
typical feature selection approaches would need. This intuition can be extended more generally to
scenarios where we do not necessarily do a sparse regression and beyond product distributions, but
we find that even this simplest example illustrates the basic motivations underlying our choice?we
want to parsimoniously expand on top of a base feature set, while still making progress towards a
good polynomial for our data.
2.4
Fall-back risk-consistency
Neither the SSM heuristic nor the weight heuristic is rigorously analyzed (in any generality). Despite
this, the basic algorithm apple can be easily modified to guarantee a form of risk consistency,
regardless of which feature expansion heuristic is used. Consider the following variant of the support
update rule in the algorithm apple. Given the current feature budget sk , we add sk ? 1 monomials
ordered by weight magnitudes as in Step 7. We also pick a monomial m(x) of the smallest degree
such that m(x) ?
/ Pk . Intuitively, this ensures that all degree 1 terms are in Pk after d stages,
all degree 2 terms are in Pk after k = O(d2 ) stages and so on. In general, it is easily seen that
k = O(d`?1 ) ensures that all degree ` ? 1 monomials are in Pk and hence all degree ` monomials
are in Sk . For ease of exposition, let us assume that sk is set to be a constant s independent of k.
Then the total number of monomials in Pk when k = O(d`?1 ) is O(sd`?1 ), which means the total
number of features in Sk is O(sd` ).
Suppose we were interested in competing with all ?-sparse polynomials of degree `. The most direct
approach would be to consider the explicit enumeration of all monomials of degree up to `, and then
perform `1 -regularized regression [18] or a greedy variable selection method such as OMP [19] as
means of enforcing sparsity. This ensures consistent estimation with n = O(? log d` ) = O(?` log d)
examples. In contrast, we might need n = O(?(` log d + log s)) examples in the worst case using
this fall back rule, a minor overhead at best. However, in favorable cases, we stand to gain a lot when
the heuristic succeeds in finding good monomials rapidly. Since this is really an empirical question,
we will address it with our empirical evaluation.
3
Experimental study
We now describe of our empirical evaluation of apple.
3.1
Implementation, experimental setup, and performance metrics
In order to assess the effectiveness of our algorithm, it is critical to build on top of an efficient
learning framework that can handle large, high-dimensional datasets. To this end, we implemented
apple in the Vowpal Wabbit (henceforth VW) open source machine learning software1 . VW is a
good framework for us, since it also natively supports quadratic and cubic expansions on top of the
base features. These expansions are done dynamically at run-time, rather than being stored and read
from disk in the expanded form for computational considerations. To deal with these dynamically
enumerated features, VW uses hashing to associate features with indices, mapping each feature to a
b-bit index, where b is a parameter. The core learning algorithm is an online algorithm as assumed
in apple, but uses refinements of the basic stochastic gradient descent update (e.g., [20?23]).
We implemented apple such that the total number of epochs was always 6 (meaning 5 rounds of
adding new features). At the end of each epoch, the non-parent monomials with largest magnitude
weights were marked as parents. Recall that the number of parents is modulated at s? for some
? > 0, with s being the average number of non-zero features per example in the dataset so far. We
will present experimental results with different choices of ?, and we found ? = 1 to be a reliable
1
Please see https://github.com/JohnLangford/vowpal_wabbit and the associated git
repository, where -stage_poly and related command line options execute apple.
5
30
linear
quadratic
cubic
apple
apple-best
ssm
ssm-best
25
20
15
number of datasets (cumulative)
number of datasets (cumulative)
30
10
5
0
?1.5
?1.0
?0.5
0.0
0.5
1.0
1.5
relative error
25
20
linear
quadratic
cubic
apple
apple-best
ssm
ssm-best
15
10
5
0
1
10
100
relative time
(a)
(b)
Figure 2: Dataset CDFs across all 30 datasets: (a) relative test error, (b) relative training time (log
scale). {apple, ssm} refer to the ? = 1 default; {apple, ssm}-best picks best ? per dataset.
default. Upon seeing an example, the features are enumerated on-the-fly by recursively expanding
the marked parents, taking products with base monomials. These operations are done in a way to
respect the sparsity (in terms of base features) of examples which many of our datasets exhibit.
Since the benefits of nonlinear learning over linear learning themselves are very dataset dependent,
and furthermore can vary greatly for different heuristics based on the problem at hand, we found it
important to experiment with a large testbed consisting of a diverse collection of medium and largescale datasets. To this end, we compiled a collection of 30 publicly available datasets, across a number of KDDCup challenges, UCI repository and other common resources (detailed in the appendix).
For all the datasets, we tuned the learning rate for each learning algorithm based on the progressive
validation error (which is typically a reliable bound on test error) [24]. The number of bits in hashing
was set to 18 for all algorithms, apart from cubic polynomials, where using 24 bits for hashing was
found to be important for good statistical performance. For each dataset, we performed a random
split with 80% of the data used for training and the remainder for testing. For all datasets, we used
squared-loss to train, and 0-1/squared-loss for evaluation in classification/regression problems. We
also experimented with `1 and `2 regularization, but these did not help much. The remaining settings
were left to their VW defaults.
For aggregating performance across 30 diverse datasets, it was important to use error and running
time measures on a scale independent of the dataset. Let `, q and c refer to the test errors of linear,
quadratic and cubic baselines respectively (with lin, quad, and cubic used to denote the baseline
algorithms themselves). For an algorithm alg, we compute the relative (test) error:
rel err(alg) =
err(alg) ? min(`, q, c)
,
max(`, q, c) ? min(`, q, c)
(2)
where min(`, q, c) is the smallest error among the three baselines on the dataset, and max(`, q, c)
is similarly defined. We also define the relative (training) time as the ratio to running time of lin:
rel time(alg) = time(alg)/time(lin). With these definitions, the aggregated plots of relative
errors and relative times for the various baselines and our methods are shown in Figure 2. For each
method, the plots show a cumulative distribution function (CDF) across datasets: an entry (a, b)
on the left plot indicates that the relative error for b datasets was at most a. The plots include the
baselines lin, quad, cubic, as well as a variant of apple (called ssm) that replaces the weight
heuristic with the SSM heuristic, as described in Section 2.3. For apple and ssm, the plot shows
the results with the fixed setting of ? = 1, as well as the best setting chosen per dataset from
? ? {0.125, 0.25, 0.5, 0.75, 1} (referred to as apple-best and ssm-best).
3.2
Results
In this section, we present some aggregate results. Detailed results with full plots and tables are
presented in the appendix. In the Figure 2(a), the relative error for all of lin, quad and cubic is
6
10
8
linear
quadratic
cubic
apple
apple-best
number of datasets (cumulative)
number of datasets (cumulative)
12
6
4
2
0
?1.5
?1.0
?0.5
0.0
0.5
1.0
1.5
relative error
(a)
12
10
linear
quadratic
cubic
apple
apple-best
8
6
4
2
0
1
10
100
relative time
(b)
Figure 3: Dataset CDFs across 13 datasets where time(quad) ? 2time(lin): (a) relative test error,
(b) relative training time (log scale).
always to the right of 0 (due to the definition of rel err). In this plot, a curve enclosing a larger area
indicates, in some sense, that one method uniformly dominates another. Since apple uniformly
dominates ssm statistically (with only slightly longer running times), we restrict the remainder of
our study to comparing apple to the baselines lin, quad and cubic. We found that on 12 of the
30 datasets, the relative error was negative, meaning that apple beats all the baselines. A relative
error of 0.5 indicates that we cover at least half the gap between min(`, q, c) and max(`, q, c). We
find that we are below 0.5 on 27 out of 30 datasets for apple-best, and 26 out of the 30 datasets for
the setting ? = 1. This is particularly striking since the error min(`, q, c) is attained by cubic on
a majority of the datasets (17 out of 30), where the relative error of cubic is 0. Hence, statistically
apple often outperforms even cubic, while typically using a much smaller number of features. To
support this claim, we include in the appendix a plot of the average number of features per example
generated by each method, for all datasets. Overall, we find the statistical performance of apple
from Figure 2 to be quite encouraging across this large collection of diverse datasets.
The running time performance of apple is also extremely good. Figure 2(b) shows that the running
time of apple is within a factor of 10 of lin for almost all datasets, which is quite impressive
considering that we generate a potentially much larger number of features. The gap between lin
and apple is particularly small for several large datasets, where the examples are sparse and highdimensional. In these cases, all algorithms are typically I/O-bottlenecked, which is the same for all
algorithms due to the dynamic feature expansions used. It is easily seen that the statistically efficient
baseline of cubic is typically computationally infeasible, with the relative time often being as large
as 102 and 105 on the biggest dataset. Overall, the statistical performance of apple is competitive
with and often better than min(`, q, c), and offers a nice intermediate in computational complexity.
A surprise in Figure 2(b) is that quad appears to computationally outperform apple for a relatively
large number of datasets, at least in aggregate. This is due to the extremely efficient implementation
of quad in VW: on 17 of 30 datasets, the running time of quad is less than twice that of lin. While
we often statistically outperform quad on many of these smaller datasets, we are primarily interested
in the larger datasets where the relative cost of nonlinear expansions (as in quad) is high.
In Figure 3, we restrict attention to the 13 datasets where time(quad)/time(lin) ? 2. On these
larger datasets, our statistical performance seems to dominate all the baselines (at least in terms
of the CDFs, more on individual datasets will be said later). In terms of computational time, we
see that we are often much better than quad, and cubic is essentially infeasible on most of these
datasets. This demonstrates our key intuition that such adaptively chosen monomials are key to
effective nonlinear learning in large, high-dimensional datasets.
We also experimented with picky algorithms of the sort mentioned in Section 2.2. We tried the
original algorithm from [13], which tests a candidate monomial before adding it to the feature set Sk ,
rather than just testing candidate parent monomials for inclusion in Pk ; and also a picky algorithm
based on our weight heuristic. Both algorithms were extremely computationally expensive, even
when implemented using VW as a base: the explicit testing for inclusion in Sk (on a per-example
7
5
Relative error, ordered by average nonzero features per example 103 Relative time, ordered by average nonzero features per example
linear
quadratic
cubic
apple(0.125)
apple(0.25)
apple(0.5)
apple(0.75)
apple(1.0)
4
3
2
2
10
1
10
1
0
0
rcv1
nomao
year
20news
slice
cup98
10
rcv1
nomao
(a)
year
20news
slice
cup98
(b)
Figure 4: Comparison of different methods on the top 6 datasets by non-zero features per example:
(a) relative test errors, (b) relative training times.
Test AUC
Training time (in s)
lin
0.81664
1282
lin + apple
0.81712
2727
bigram
0.81757
2755
bigram + apple
0.81796
7378
Table 1: Test error and training times for different methods in a large-scale distributed setting. For
{lin, bigram} + apple, we used ? = 0.25.
basis) caused too much overhead. We ruled out other baselines such as polynomial kernels for
similar computational reasons.
To provide more intuition, we also show individual results for the top 6 datasets with the highest
average number of non-zero features per example?a key factor determining the computational cost
of all approaches. In Figure 4, we show the performance of the lin, quad, cubic baselines, as well
as apple with 5 different parameter settings in terms of relative error (Figure 4(a)) and relative time
(Figure 4(b)). The results are overall quite positive. We see that on 3 of the datasets, we improve
upon all the baselines statistically, and even on other 3 the performance is quite close to the best of
the baselines with the exception of the cup98 dataset. In terms of running time, we find cubic to be
extremely expensive in all the cases. We are typically faster than quad, and in the few cases where
we take longer, we also obtain a statistical improvement for the slight increase in computational
cost. In conclusion, on larger datasets, the performance of our method is quite desirable.
Finally, we also implemented a parallel version of our algorithm, building on the repeated averaging
approach [2, 25], using the built-in AllReduce communication mechanism of VW, and ran an experiment using an internal advertising dataset consisting of approximately 690M training examples,
with roughly 318 non-zero features per example. The task is the prediction of click/no-click
events. The data was stored in a large Hadoop cluster, split over 100 partitions. We implemented the
lin baseline, using 5 passes of online learning with repeated averaging on this dataset, but could
not run full quad or cubic baselines due to the prohibitive computational cost. As an intermediate,
we generated bigram features, which only doubles the number of non-zero features per example.
We parallelized apple as follows. In the first pass over the data, each one of the 100 nodes locally
selects the promising features over 6 epochs, as in our single-machine setting. We then take the
union of all the parents locally found across all nodes, and freeze that to be the parent set for the rest
of training. The remaining 4 passes are now done with this fixed feature set, repeatedly averaging
local weights. We then ran apple, on top of both lin as well as bigram as the base features to
obtain maximally expressive features. The test error was measured in terms of the area under ROC
curve (AUC), since this is a highly imbalanced dataset. The error and time results, reported in Table 1, show that using nonlinear features does lead to non-trivial improvements in AUC, albeit at
an increased computational cost. Once again, this should be put in perspective with the full quad
baseline, which did not finish in over a day on this dataset.
Acknowledgements: We thank Leon Bottou, Rob Schapire and Dean Foster for helpful discussions.
8
References
[1] I. Mukherjee, K. Canini, R. Frongillo, and Y. Singer. Parallel boosting with momentum. In Proceedings
of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery
in Databases, 2013.
[2] A. Agarwal, O. Chapelle, M. Dud?k, and J. Langford. A reliable effective terascale linear learning system.
Journal of Machine Learning Research, 15(Mar):1111?1133, 2014.
[3] A. Agarwal, A. Beygelzimer, D. Hsu, J. Langford, and M. Telgarsky. Scalable nonlinear learning with
adaptive polynomial expansions. 2014. arXiv:1410.0440 [cs.LG].
[4] C. Williams and M. Seeger. Using the Nystr?m method to speed up kernel machines. In Advances in
Neural Information Processing Systems 13, 2001.
[5] M. W. Mahoney. Randomized algorithms for matrices and data. Foundations and Trends in Machine
Learning, 3(2):123?224, 2011.
[6] B. Sch?lkopf and A. Smola. Learning with Kernels. MIT Press, Cambridge, MA, 2002.
[7] A. Bordes, S. Ertekin, J. Weston, and L. Bottou. Fast kernel classifiers with online and active learning.
Journal of Machine Learning Research, 6:1579?1619, 2005.
[8] A. Rahimi and B. Recht. Random features for large-scale kernel machines. In Advances in Neural
Information Processing Systems 20, 2008.
[9] P. Kar and H. Karnick. Random feature maps for dot product kernels. In AISTATS, 2012.
[10] N. Pham and R. Pagh. Fast and scalable polynomial kernels via explicit feature maps. In Proceedings of
the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2013.
[11] R. Hamid, A. Gittens, Y. Xiao, and D. Decoste. Compact random feature maps. In ICML, 2014.
[12] A. G. Ivakhnenko. Polynomial theory of complex systems. Systems, Man and Cybernetics, IEEE Transactions on, SMC-1(4):364?378, 1971.
[13] T. D. Sanger, R. S. Sutton, and C. J. Matheus. Iterative construction of sparse polynomial approximations.
In Advances in Neural Information Processing Systems 4, 1992.
[14] A. T. Kalai, A. Samorodnitsky, and S.-H. Teng. Learning and smoothed analysis. In FOCS, 2009.
[15] A. Andoni, R. Panigrahy, G. Valiant, and L. Zhang. Learning sparse polynomial functions. In SODA,
2014.
[16] A. G. Dimakis, A. Klivans, M. Kocaoglu, and K. Shanmugam. A smoothed analysis for learning sparse
polynomials. CoRR, abs/1402.3902, 2014.
[17] M. Zinkevich. Online convex programming and generalized infinitesimal gradient ascent. In ICML, 2003.
[18] R. Tibshirani. Regression shrinkage and selection via the lasso. J. Royal. Statist. Soc B., 58(1):267?288,
1996.
[19] J. A. Tropp and A. C. Gilbert. Signal recovery from random measurements via orthogonal matching
pursuit. IEEE Transactions on Information Theory, 53(12):4655?4666, December 2007.
[20] J. Duchi, E. Hazan, and Y. Singer. Adaptive subgradient methods for online learning and stochastic
optimization. The Journal of Machine Learning Research, 12:2121?2159, 2011.
[21] H. B. McMahan and M. J. Streeter. Adaptive bound optimization for online convex optimization. In
COLT, pages 244?256, 2010.
[22] N. Karampatziakis and J. Langford. Online importance weight aware updates. In UAI, pages 392?399,
2011.
[23] S. Ross, P. Mineiro, and J. Langford. Normalized online learning. In UAI, 2013.
[24] A. Blum, A. Kalai, and J. Langford. Beating the hold-out: Bounds for k-fold and progressive crossvalidation. In COLT, 1999.
[25] K. Hall, S. Gilpin, and G. Mann. Mapreduce/bigtable for distributed optimization. In Workshop on
Learning on Cores, Clusters, and Clouds, 2010.
[26] S. Bubeck. Theory of convex optimization for machine learning.
[math.OC].
2014.
arXiv:1405.4980
[27] O. Shamir and T. Zhang. Stochastic gradient descent for non-smooth optimization: Convergence results
and optimal averaging schemes. In ICML, 2013.
9
| 5397 |@word repository:2 version:2 polynomial:23 stronger:1 seems:1 bigram:5 disk:1 open:1 d2:1 tried:1 git:1 pick:5 paid:1 nystr:2 harder:1 recursively:1 reduction:2 nomao:2 initial:3 daniel:1 tuned:1 outperforms:1 existing:3 err:3 current:7 com:4 comparing:1 beygelzimer:2 john:1 partition:1 enables:1 designed:1 plot:8 update:7 v:1 greedy:1 discovering:1 half:1 prohibitive:1 core:2 boosting:1 node:2 math:1 liberal:2 simpler:1 ssm:13 zhang:2 direct:2 focs:1 prove:2 shorthand:1 overhead:2 expected:1 indeed:3 roughly:3 p1:1 frequently:1 growing:2 nor:1 themselves:2 encouraging:1 enumeration:1 quad:16 considering:1 decoste:1 becomes:1 moreover:1 underlying:1 medium:1 lowest:1 what:2 dimakis:1 developed:3 finding:1 guarantee:4 expands:1 xd:1 demonstrates:2 classifier:1 before:2 negligible:1 positive:1 aggregating:1 local:1 sd:2 limit:1 despite:1 sutton:1 approximately:1 might:2 plus:1 twice:2 dynamically:2 suggests:2 challenging:1 ease:1 smc:1 forefront:1 cdfs:3 statistically:6 averaged:1 testing:4 union:1 regret:8 practice:1 differs:1 area:2 empirical:3 matching:1 seeing:1 close:1 selection:3 put:1 risk:2 restriction:2 zinkevich:1 dean:1 map:3 yt:2 vowpal:1 gilbert:1 williams:1 regardless:1 starting:1 attention:1 convex:6 mtelgars:1 recovery:1 immediately:1 rule:3 dominate:1 embedding:1 stability:1 notion:1 coordinate:1 handle:1 construction:2 pt:4 heavily:1 massive:1 infrequent:1 suppose:3 target:1 us:3 programming:1 shamir:1 trick:1 associate:1 trend:1 expensive:3 satisfying:1 particularly:2 mukherjee:1 database:1 observed:1 ft:3 cloud:1 fly:2 worst:2 ensures:3 alekha:1 news:2 highest:2 ran:2 substantial:2 intuition:4 mentioned:1 convexity:2 complexity:4 rigorously:1 dynamic:1 compromise:1 creates:1 upon:2 efficiency:2 learner:2 beygel:1 basis:1 easily:4 k0:1 various:1 train:1 fast:5 describe:2 effective:5 aggregate:2 vowpal_wabbit:1 quite:8 heuristic:13 larger:6 plausible:2 say:1 ability:2 jointly:1 final:2 online:14 sequence:3 differentiable:1 kxt:1 wabbit:1 interaction:6 product:5 remainder:2 uci:1 rapidly:1 achieve:1 description:1 crossvalidation:1 parent:10 convergence:2 cluster:2 extending:1 double:1 generating:1 telgarsky:2 help:1 measured:1 minor:1 progress:1 strong:4 soc:1 implemented:6 c:3 involves:1 drawback:1 stochastic:8 enable:1 mann:1 really:1 hamid:1 opt:1 anticipate:1 enumerated:2 strictly:2 pham:1 hold:1 hall:1 minu:1 mapping:1 matus:1 claim:1 vary:1 smallest:2 favorable:2 estimation:1 label:2 expose:1 ross:1 exercised:1 largest:4 djhsu:1 weighted:1 mit:1 always:2 modified:1 rather:3 kalai:2 frongillo:1 shrinkage:1 stepsizes:1 command:1 improvement:3 karampatziakis:1 indicates:3 greatly:1 contrast:2 sigkdd:1 adversarial:1 baseline:22 sense:1 seeger:1 helpful:1 milder:1 dependent:1 typically:9 entire:1 expand:4 interested:2 selects:1 arg:1 aforementioned:1 among:2 overall:4 classification:1 colt:2 yahoo:2 enrich:1 once:1 aware:1 sampling:1 progressive:2 hardest:1 unsupervised:1 comparators:4 icml:3 future:1 employ:1 primarily:1 few:1 ve:1 kwt:1 individual:2 parsimoniously:1 ourselves:1 consisting:2 microsoft:5 recalling:1 ab:1 possibility:1 highly:1 mining:1 evaluation:3 mahoney:1 analyzed:1 extreme:1 amenable:1 implication:1 kt:1 intense:1 modest:1 unless:1 orthogonal:1 euclidean:1 loosely:1 ruled:1 sacrificing:1 guidance:1 mk:3 instance:1 increased:1 cover:1 cost:11 entry:1 monomials:23 uniform:1 predictor:3 too:2 stored:2 reported:1 combined:1 adaptively:3 recht:1 international:1 randomized:1 kut:3 pagh:1 picking:1 together:1 quickly:1 na:1 again:2 squared:2 opposed:4 possibly:2 henceforth:1 creating:3 potential:1 nonlinearities:1 inc:1 explicitly:2 caused:1 performed:2 later:2 lot:1 lab:1 analyze:1 hazan:1 start:2 bayes:1 sort:2 option:1 competitive:1 parallel:2 ass:1 square:2 publicly:1 efficiently:2 spaced:1 yield:1 lkopf:1 comparably:1 xyi:1 advertising:1 notoriously:1 apple:52 cybernetics:1 influenced:1 definition:4 infinitesimal:1 against:2 naturally:2 associated:1 attributed:1 hsu:2 gain:1 dataset:15 ask:1 recall:2 knowledge:3 ut:18 schedule:1 positioned:1 sophisticated:1 carefully:2 back:2 appears:1 higher:4 hashing:3 attained:1 day:1 maximally:1 done:5 though:1 execute:1 generality:1 furthermore:1 just:4 stage:7 mar:1 smola:1 langford:6 until:1 hand:1 tropp:1 expressive:1 nonlinear:10 marker:1 quality:1 building:1 normalized:1 regularization:2 hence:3 read:1 dud:1 nonzero:2 deal:1 round:3 please:1 auc:3 illustrative:1 percentile:1 oc:1 generalized:1 duchi:1 spending:1 wise:1 consideration:3 meaning:2 recently:1 superior:2 common:2 pseudocode:1 mt:1 empirically:1 conditioning:1 discussed:2 slight:1 approximates:1 kwk2:1 refer:2 measurement:1 freeze:1 cambridge:1 nyc:1 consistency:2 similarly:1 inclusion:3 dot:1 chapelle:1 harsher:1 alekh:1 compiled:1 longer:2 impressive:1 base:8 add:3 dominant:1 imbalanced:1 perspective:1 apart:1 scenario:1 kar:1 seen:2 additional:1 care:1 omp:1 parallelized:3 converge:2 aggregated:1 signal:1 full:4 desirable:2 rahimi:1 smooth:1 faster:3 offer:3 lin:17 permitting:1 equally:1 controlled:1 prediction:6 scalable:3 regression:10 variant:4 basic:3 essentially:1 expectation:2 rutgers:1 metric:1 arxiv:2 sometimes:1 kernel:15 agarwal:3 receive:1 remarkably:1 want:1 ertekin:1 jcl:1 median:1 source:1 sch:1 rest:1 unlike:1 pass:2 ascent:1 johnlangford:1 december:1 allreduce:1 effectiveness:1 call:1 structural:1 curious:1 vw:7 intermediate:2 split:2 embeddings:2 easy:1 finish:1 competing:2 restrict:2 click:2 inner:1 lasso:1 tradeoff:6 bottleneck:1 pca:1 utility:1 suffer:2 repeatedly:4 adequate:1 generally:3 detailed:2 picky:2 bottlenecked:2 locally:2 statist:1 simplest:1 http:1 generate:1 outperform:3 exist:1 schapire:1 per:11 tibshirani:1 diverse:3 key:3 blum:1 alina:1 neither:2 subgradient:1 fraction:1 year:2 compete:1 run:3 striking:1 soda:1 almost:1 reader:1 prefer:1 scaling:2 appendix:3 comparable:1 bit:3 bound:13 fold:1 quadratic:8 replaces:1 constraint:1 aspect:1 speed:2 u1:2 min:6 extremely:4 leon:1 performing:1 expanded:3 rcv1:2 kocaoglu:1 relatively:2 klivans:1 kd:1 describes:1 smaller:3 across:7 slightly:1 appealing:1 gittens:1 rob:1 making:1 s1:3 intuitively:1 computationally:8 resource:2 turn:1 discus:1 fail:1 mechanism:2 eventually:1 singer:2 end:5 staged:1 available:3 operation:3 pursuit:1 observe:1 alternative:2 batch:3 original:2 denotes:1 top:9 remaining:2 running:7 include:2 opportunity:1 sanger:1 folklore:2 build:1 establish:1 objective:7 intend:1 question:2 added:1 dependence:1 modestly:1 visiting:1 exhibit:4 gradient:8 amongst:1 subspace:1 said:1 thank:1 majority:1 considers:1 trivial:1 reason:1 enforcing:1 panigrahy:1 index:3 ratio:1 difficult:3 executed:1 setup:1 lg:1 potentially:1 favorably:1 negative:1 design:2 implementation:4 proper:1 enclosing:1 unknown:1 perform:2 allowing:1 datasets:41 benchmark:1 enabling:1 finite:1 descent:4 beat:1 canini:1 matheus:1 extended:1 communication:1 ucsd:1 smoothed:2 specified:2 extensive:2 learned:1 quadratically:1 distinction:1 testbed:1 address:1 beyond:1 bar:1 proceeds:1 below:1 beating:1 sparsity:2 challenge:1 ivakhnenko:1 built:1 royal:1 including:1 max:5 reliable:3 shifting:8 power:3 suitable:1 critical:1 natural:2 event:1 regularized:4 largescale:1 residual:1 scheme:5 improve:2 github:1 columbia:2 faced:1 text:1 epoch:7 prior:1 nice:2 acknowledgement:1 discovery:2 determining:1 relative:32 mapreduce:1 loss:6 interesting:1 validation:1 foundation:1 degree:12 consistent:2 xiao:1 foster:1 principle:1 terascale:1 bordes:1 maxt:2 course:3 summary:1 placed:1 infeasible:2 monomial:10 guide:1 allow:2 fall:2 taking:2 sparse:10 shanmugam:1 distributed:3 slice:2 benefit:2 curve:2 dimension:1 default:3 world:1 cumulative:8 evaluating:2 unweighted:1 stand:1 author:1 commonly:3 adaptive:9 refinement:1 collection:3 karnick:1 far:1 transaction:2 approximate:1 compact:1 active:1 uai:2 assumed:1 xi:3 kddcup:1 iterative:1 mineiro:1 sk:23 ngrams:1 table:3 streeter:1 promising:1 learn:1 expanding:2 hadoop:1 improving:1 alg:5 expansion:16 bottou:2 necessarily:1 european:1 complex:1 did:2 pk:11 dense:4 aistats:1 s2:1 motivation:1 n2:1 allowed:1 repeated:2 x1:1 referred:1 biggest:1 roc:1 cubic:21 amortize:1 slow:1 sub:3 natively:1 momentum:1 explicit:4 candidate:2 mcmahan:1 hw:1 theorem:1 xt:1 showing:1 experimented:2 svm:1 evidence:1 dominates:2 workshop:1 rel:3 adding:3 effectively:5 importance:1 albeit:1 andoni:1 valiant:1 corr:1 magnitude:6 budget:3 exhausted:1 illustrates:1 gap:2 surprise:1 simply:1 bubeck:1 ordered:5 tracking:2 partially:1 u2:1 nested:1 acm:1 cdf:1 ma:1 weston:1 comparator:7 conditional:1 marked:2 exposition:1 towards:1 man:1 samorodnitsky:1 change:2 typical:1 uniformly:2 wt:10 justify:2 averaging:4 total:5 called:1 pas:3 teng:1 experimental:6 succeeds:1 exception:1 select:1 highdimensional:1 gilpin:1 internal:1 support:5 latter:1 modulated:1 yardstick:1 avoiding:1 |
4,857 | 5,398 | Orbit Regularization
Andr?e F. T. Martins?
Instituto de Telecomunicac?o? es
Instituto Superior T?ecnico
1049?001 Lisboa, Portugal
[email protected]
Renato Negrinho
Instituto de Telecomunicac?o? es
Instituto Superior T?ecnico
1049?001 Lisboa, Portugal
[email protected]
Abstract
We propose a general framework for regularization based on group-induced majorization. In this framework, a group is defined to act on the parameter space and
an orbit is fixed; to control complexity, the model parameters are confined to the
convex hull of this orbit (the orbitope). We recover several well-known regularizers as particular cases, and reveal a connection between the hyperoctahedral group
and the recently proposed sorted `1 -norm. We derive the properties a group must
satisfy for being amenable to optimization with conditional and projected gradient algorithms. Finally, we suggest a continuation strategy for orbit exploration,
presenting simulation results for the symmetric and hyperoctahedral groups.
1
Introduction
The main motivation behind current sparse estimation methods and regularized empirical risk minimization is the principle of parsimony, which states that simple explanations should be preferred
over complex ones. Traditionally, this has been done by defining a function ? : V ? R that evaluates the complexity of a model w ? V and trades off this quantity with a data-dependent term. The
penalty function ? is often designed to be a convex surrogate of an otherwise non-tractable quantity,
a strategy which has led to important achievements in sparse regression [1], compressed sensing
[2], and matrix completion [3], allowing to successfully recover parameters from highly incomplete
information. Prior knowledge about the structure of the variables and the intended sparsity pattern,
when available, can be taken into account when designing ? via sparsity-inducing norms [4]. Performance bounds under different regimes have been established theoretically [5, 6], contributing to
a better understanding of the success and failure modes of these techniques.
In this paper, we introduce a new way to characterize the complexity of a model via the concept
of group-induced majorization. Rather than regarding complexity in an absolute manner via ?, we
define it relative to a prototype model v ? V , by requiring that the estimated model w satisfies
w G v,
(1)
where G is an ordering relation on V induced by a group G. This idea is rooted in majorization
theory, a well-established field [7, 8] which, to the best of our knowledge, has never been applied to
machine learning. We therefore review these concepts in ?2, where we show that this formulation
subsumes several well-known regularizers and motivates new ones. Then, in ?3, we introduce two
important properties of groups that serve as building blocks for the rest of the paper: the notions
of matching function and region cones. In ?4, we apply these tools to the permutation and signed
permutation groups, unveiling connections with the recent sorted `1 -norm [9] as a byproduct. In ?5
we turn to algorithmic considerations, pinpointing the group-specific operations that make a group
amenable to optimization with conditional and projected gradient algorithms.
?
Also at Priberam Labs, Alameda D. Afonso Henriques, 41 - 2? , 1000?123, Lisboa, Portugal.
1
Figure 1: Examples of orbitopes for the orthogonal group O(d) (left) and the hyperoctahedral
group P? (right). Shown are also the corresponding region cones, which in the case of O(d) degenerates into a ray.
A key aspect of our framework is a decoupling in which the group G captures the invariances of
the regularizer, while the data-dependent term is optimized in the group orbitopes. In ?6, we build
on this intuition to propose a simple continuation algorithm for orbit exploration. Finally, ?7 shows
some simulation results, and we conclude in ?8.
2
2.1
Orbitopes and Majorization
Vector Spaces and Groups
Let V be a vector space with an inner product h?, ?i. We will be mostly concerned with the case where
V = Rd , i.e., the d-dimensional real Euclidean space, but some of the concepts introduced here
generalize to arbitrary Hilbert spaces. A group is a set G endowed with an operation ? : G ? G ? G
satisfying closure (g ? h ? G, ?g, h ? G), associativity ((f ? g) ? h = f ? (g ? h), ?f, g, h ? G),
existence of identity (?1G ? G such that 1G ? g = g ? 1G = g, ?g ? G), and existence of inverses
(each g ? G has an inverse g ?1 ? G such that g ? g ?1 = g ?1 ? g = 1G ). Throughout, we use
boldface letters u, v, w, . . . for vectors, and g, h, . . . for group elements. We also omit the group
operation symbol, writing gh instead of g ? h.
2.2
Group Actions, Orbits, and Orbitopes
A (left) group action of G on V [10] is a function ? : G ? V ? V satisfying ?(g, ?(h, v)) =
?(g ? h, v) and ?(1G , v) = v for all g, h ? G and v ? V . When the action is clear from the context,
we omit the letter ?, writing simply gv for the action of the group element g on v, instead of ?(g, v).
In this paper, we always assume our actions are linear, i.e., g(c1 v 1 + c2 v 2 ) = c1 gv 1 + c2 gv 2 for
scalars c1 and c2 and vectors v 1 and v 2 . In some cases, we also assume they are norm-preserving,
i.e., kgvk = kvk for any g ? G and v ? V . When V = Rd , we may regard the groups underlying
these actions as subgroups of the general linear group GL(d) and of the orthogonal group O(d),
respectively. GL(d) is the set of d-by-d invertible matrices, and O(d) the set of d-by-d orthogonal
matrices {U ? Rd?d | U > U = U U > = Id }, where Id denotes the d-dimensional identity matrix.
A group action defines an equivalence relation on V , namely w ? v iff there is g ? G such that
w = gv. The orbit of a vector v ? V under the action of G is the set Gv := {gv | g ? G}, i.e., the
vectors that result from acting on v with some element of G. Its convex hull is called the orbitope:
OG (v) := conv(Gv).
(2)
Fig. 1 (left) illustrates this concept for the orthogonal group in R2 . An important concept associated
with group actions and orbitopes is that of G-majorization [7]:
Definition 1 Let v, w ? V . We say that w is G-majorized by v, denoted w G v, if w ? OG (v).
Proposition 2 If the group action is linear, then G is reflexive and transitive, i.e., it is a pre-order.
Proof: See supplemental material.
Group majorization plays an important role in the area of multivariate inequalities in statistics [11].
In this paper, we use this concept for representing model complexity, as described next.
2.3
Orbit Regularization
We formulate our learning problem as follows:
minimize L(w) s.t. w G v,
(3)
where L : V ? R is a loss function, G is a given group, and v ? V is a seed vector. This
formulation subsumes several well-known cases, outlined below.
2
? `2 -regularization. If G := O(d) is the orthogonal group acting by multiplication, we recover `2
regularization. Indeed, we have Gv = {U v ? Rd | U ? O(d)} = {w ? Rd | kwk2 = kvk2 }, for
any seed v ? Rd . That is, the orbitope OG (v) = conv(Gv) becomes the `2 -ball with radius kvk2 .
The only property of the seed that matters in this case is its `2 -norm.
? Permutahedron. Let P be the symmetric group (also called the permutation group), which can
be represented as the set of d-by-d permutation matrices. Given v ? Rd , the orbitope induced by v
under P is the convex hull of all the permutations of v, which can be equivalently described as the
vectors that are transformations of v through a doubly stochastic matrix:
OP (v) = conv{P v | P ? P} = {M v | M 1 = 1, M > 1 = 1, M ? 0}.
(4)
This set is called the permutahedron [12]. We will revisit this case in ?4.
? Signed permutahedron. Let P? be the hyperoctahedral group (also called the signed permutation group), i.e., the d-by-d matrices with entries in {0, ?1} such that the sum of the absolute values
in each row and column is 1. The action of P? on Rd permutes the entries of vectors and arbitrarily
switches signs. Given v ? Rd , the orbitope induced by v under P? is:
OP? (v) = conv{Diag(s)P v | P ? P, s ? {?1}d },
(5)
where Diag(s) denotes a diagonal matrix formed by the entries of s. We call this set the signed
permutahedron; it is depicted in Fig. 1 and will also be revisited in ?4.
? `1 and `? -regularization. As a particular case of the signed permutahedron, we recover `1 and
`? balls by choosing seeds of the form v = ?e1 (a scaled canonical basis vector) and v = ?1 (a
constant vector), respectively, where ? is a scalar. In the first case, we obtain the `1 -ball, OG (v) =
? conv({e1 , . . . , ed }) and in the second case, we get the `? -ball OG (v) = ? conv({?1}d ).
? Symmetric matrices with majorized eigenvalues. Let G := O(d) be again the orthogonal
group, but now acting by conjugation on the vector space of d-by-d symmetric matrices, V = Sd .
Given a seed v ? A ? Sd , its orbit is Gv = {U AU > | U ? O(d)} = {U Diag(?(A))U > | U ?
O(d)}, where ?(A) denotes a vector containing the eigenvalues of A in decreasing order (so we
may assume without loss of generality that the seed is diagonal). The orbitope OG (v) becomes:
OG (v) := {B ? Sd | ?(B) P ?(A)},
(6)
which is the set of matrices whose eigenvalues are in the permutahedron OP (?(A)) (see example
above). This is called the Schur-Horn orbitope in the literature [8].
? Squared matrices with majorized singular values. Let G := O(d) ? O(d) act on Rd?d (the
space of squared matrices, not necessarily symmetric) as gU,V A := U AV > . Given a seed v ? A,
its orbit is Gv = {U AV > | U, V ? O(d)} = {U Diag(?(A))V > | U, V ? O(d)}, where ?(A)
contains the singular values of A in decreasing order (so we may assume without loss of generality
that the seed is diagonal and non-negative). The orbitope OG (v) becomes:
OG (v) := {B ? Rd?d | ?(B) P ?(A)},
(7)
which is the set of matrices whose singular values are in the permutahedron OP (?(A)).
? Spectral and nuclear norm regularization. The previous case subsumes spectral and nuclear
norm balls: indeed, for a seed A = ?Id , the orbitope becomes the convex hull of orthogonal matrices, which is the spectral norm ball {A ? Rd?d | kAk2 := ?1 (A) ? ?}; while for a seed
A = ? Diag(e1 ), the orbitope becomes the convex hull ofPrank-1 matrices with norm bounded by
?, which is the nuclear norm ball {A ? Rd?d | kAk? := i ?i ? ?}. This norm has been widely
used for low-rank matrix factorization and matrix completion [3].
Besides these examples, other regularization strategies, such as non-overlapping `2,1 and `?,1 norms
[13, 4] can be obtained by considering products of the groups above. We omit details for space.
2.4
Relation with Atomic Norms
Atomic norms have been recently proposed as a toolbox for structured sparsity [6]. Let A ? V
be a centrally symmetric set of atoms, i.e., v ? A iff ?v ? A. The atomic norm induced by A
is defined as kwkA := inf{t > 0 | w ? t conv(A)}. The corresponding atomic ball is the set
{w | kwkA ? 1} = conv(A). Not surprisingly, orbitopes are often atomic norm balls.
3
Proposition 3 (Atomic norms) If G is a subgroup of the general linear group GL(d) and satisfies
?v ? Gv, then the set OG (v) is the ball of an atomic norm.
Proof: Under the given assumption, the set Gv is centrally symmetric, i.e., it satisfies w ? Gv iff
?w ? Gv (indeed, the left hand side implies that w = gv for some g ? G, and ?v ? Gv implies
that ?v = hv for some h ? G, therefore, ?w = ?gh?1 (?v) = gh?1 v ? Gv). As shown by
Chandrasekaran et al. [6], this guarantees that k.kGv satisfies the axioms of a norm.
Corollary 4 For any choice of seed, the signed permutahedron OP? (v) and the orbitope formed
by the squared matrices with majorized singular values are both atomic norm balls. If d is even and
d/2
v is of the form v = (v + , ?v + ), with v + ? R+ , then the permutahedron OP (v) and the orbitope
formed by the symmetric matrices with eigenvalues majorized by ?(v) are both atomic norm balls.
3
Matching Function and Region Cones
We now construct a unifying perspective that highlights the role of the group G. Two key concepts
that play a crucial role in our analysis are that of matching function and region cone. In the sequel,
these will work as building blocks for important algorithmic and geometric characterizations.
Definition 5 (Matching function) The matching function of G, mG : V ? V ? R, is defined as:
mG (u, v) := sup{hu, wi | w ? Gv}.
(8)
Intuitively, mG (u, v) ?aligns? the orbits of u and v before taking the inner product. Note also that
mG (u, v) = sup{hu, wi | w ? OG (v)}, since we may equivalently maximize the linear objective
over OG (v), which is the convex hull of Gv. We therefore have the following
Proposition 6 (Duality) Fix v ? V , and define the indicator function of the orbitope, IOG (v) (w) =
0 if w ? OG (v), and ?? otherwise. The Fenchel dual of IOG (v) is mG (., v). As a consequence,
letting L? : V ? R is the Fenchel dual of the loss L, the dual problem of Eq. 3 is:
maximize ? L? (?u) ? mG (u, v) w.r.t. u ? V.
(9)
Note that if k.kGv is a norm (e.g., if the conditions of Prop. 3 are satisfied), then the statement above
means that mG (., v) = k.k?Gv is its dual norm. We will revisit this dual formulation in ?4.
The following properties have been established in [14, 15].
Proposition 7 For any u, v ? V , we have: (i) mG (c1 u, c2 v) = c1 c2 mG (u, v) for c1 , c2 ? 0;
(ii) mG (g1 u, g2 v) = mG (u, v) for g1 , g2 ? G; (iii) mG (u, v) = mG (v, u). Furthermore, the
following three statements are equivalent: (i) w G v, (ii) f (w) ? f (v) for all G-invariant convex
functions f : V ? R, (iii) mG (u, w) ? mG (u, v) for all u ? V .
In the sequel, we always assume that G is a subgroup of the orthogonal group O(d). This implies
that the orbitope OG (v) is compact for any v ? V (and therefore the sup in Eq. 8 can be replaced by
a max), and that kgvk = kvk for any v ? V . Another important concept is that of the normal cone
of a point w ? V with respect to the orbitope OG (v), denoted as NGv (w) and defined as follows:
NGv (w) := {u ? V | hu, w0 ? wi ? 0 ?w0 G v}.
(10)
Normal cones plays an important role in convex analysis [16]. The particular case of the normal
cone at the seed v (illustrated in Fig. 1) is of great importance, as will be seen below.
Definition 8 (Region cone) Given v ? V , the region cone at v is KG (v) := NGv (v). It is the set
of points that are ?maximally aligned? with v in terms of the matching function:
KG (v) = {u ? V | mG (u, v) = hu, vi}.
(11)
4
Permutahedra and Sorted `1 -Norms
In this section, we focus on the permutahedra introduced in ?2. Below, given a vector w ? Rd , we
denote by w(k) its kth order statistic, i.e., we will ?sort? w so that w(1) ? w(2) ? . . . ? w(d) . We
also consider the order statistics of the magnitudes |w|(k) by sorting the absolute values.
4
4.1
Signed Permutahedron
We start by defining the ?sorted `1 -norm,? proposed by Bogdan et al. [9] in their recent SLOPE
method as a means to control the false discovery rate, and studied by Zeng and Figueiredo [17].
Definition 9 (Sorted `1 -norm) Let v, w ? Rd , with v1 ? v2 ? . . . ? vd ? 0 and v1 > 0. The
Pd
sorted `1 -norm of w (weighted by v) is defined as: kwkSLOPE,v := j=1 vj |w|(j) .
In [9] it is shown that k.kSLOPE,v satisfies the axioms of a norm. The rationale is that larger components of w are penalized more than smaller ones, in a way controlled by the prescribed v. For
v = 1, we recover the standard `1 -norm, while the `? -norm corresponds toP
v = e1 . Another special case is the OSCAR regularizer [18, 19], kwkOSCAR,?1 ,?2 := ?1 kwk1 + ?2 i<j max{|wi |, |wj |},
corresponding to a linearly spaced v, vj = (?1 + ?2 (d ? j)) for j = 1, . . . , d. The next proposition
reveals a connection between SLOPE and the atomic norm induced by the signed permutahedron.
Proposition 10 Let v ? Rd+ be as in Def. 9. The sorted `1 -norm weighted by v and the atomic norm
induced by the P ? -orbitope seeded at v are dual to each other: k.k?P? v = k.kSLOPE,v .
Proof: From Prop. 6, we have kwk?P? v = mP? (w, v). Let P be a signed permutation matrix s.t.
? := P w has its components sorted by decreasing magnitude, |w|
w
? 1 ? . . . ? |w|
? d . From Prop. 7,
? v) = h|w|,
? vi = kwkSLOPE,v .
we have mP? (w, v) = m(w,
The next proposition [7, 14] provides a characterization of the P? -orbitope in terms of inequalities
about the cumulative distribution of the order statistics.
Proposition 11 (Submajorization ordering) The orbitope OP? (v) can be characterized as:
n
o
P
P
OP? (v) = w ? Rd
|w|
?
|v|
,
?i
=
1,
.
.
.
,
d
.
(12)
(j)
(j)
j?i
j?i
Prop. 11 leads to a precise characterization
Pof the atomic
P norm kwkP? v , and therefore of the dual
norm of SLOPE: kwkP? v = maxi=1,...,d j?i |w|(j) / j?i |v|(j) .
4.2
Permutahedron
The unsigned counterpart of Prop. 11 goes back to Hardy et al. [20].
Proposition 12 (Majorization ordering) The P-orbitope seeded at v can be characterized as:
n
o
P
P
OP (v) = w ? Rd 1> w = 1> v ?
w
?
v
,
?i
=
1,
.
.
.
,
d
?
1
.
(13)
j?i (j)
j?i (j)
As seen in Corollary 4, if d is even and v = (v + , ?v + ), with v ? 0, then kwkPv qualifies as a
Pd
norm (we need to confine to the linear subspace V := {w ? Rd |
w = 0}). From Prop. 12,
Pj=1 j P
we have that this norm can be written as: kwkPv = maxi=1,...,d?1 j?i w(j) / j?i v(j) .
Proposition 13 Assume the conditions above hold and that v1 ? v2 ? . . . ? vd/2 ? 0 and v1 > 0.
Pd/2
The dual norm of k.kPv is kwk?Pv = j=1 vj (w(j) ? w(d?j+1) ).
Proof: Similar to the proof of Prop. 11.
5
Conditional and Projected Gradient Algorithms
Two important classes of algorithms in sparse modeling are the conditional gradient method [21, 22]
and the proximal gradient method [23, 24]. Under Ivanov regularization as in Eq. 3, the latter reduces
to the projected gradient method. In this section, we show that both algorithms are a good fit for
solving Eq. 3 for arbitrary groups, as long as the two building blocks mentioned in ?3 are available:
(i) a procedure for evaluating the matching function (necessary for conditional gradient methods)
and (ii) a procedure for projecting onto the region cone (necessary for projected gradient).
5
1: Initialize w 1 = 0
2: for t = 1, 2, . . . do
3:
Choose a stepsize ?t
4:
a = wt ? ?t ?L(wt )
5:
wt+1 = arg minwG v kw ? ak
6: end for
1: Initialize w 1 = 0
2: for t = 1, 2, . . . do
3:
ut = arg maxuG v h??L(wt ), ui
4:
?t = 2/(t + 2)
5:
wt+1 = (1 ? ?t )wt + ?t ut
6: end for
Figure 2: Conditional gradient (left) and projected gradient (right) algorithms.
5.1
Conditional Gradient
The conditional gradient method is shown in Fig. 2 (left). We assume that a procedure is available
for computing the gradient of the loss. The relevant part is the maximization in line 3, which
corresponds precisely to an evaluation of the matching function m(s, v), with s = ??L(wt )
(cf. Eq. 8). Fortunately, this step is efficient for a variety of cases:
Permutations. If G = P, the matching function can be evaluated in time O(d log d) with a simple sort operation. Without losing generality, we assume the seed v is sorted in descending order
(otherwise, pre-sort it before the main loop starts). Then, each time we need to evaluate m(s, v),
we compute a permutation P such that P s is also sorted. The minimizer in line 3 will equal P ?1 v.
Signed permutations. If G = P? , a similar procedure with the same O(d log d) runtime also
works, except that now we sort the absolute values, and set the signs of P ?1 v to match those of s.
Symmetric matrices with majorized eigenvalues. Let A = UA ?(A)UA> ? Sd and B =
UB ?(B)UB> ? Sd , where the eigenvalues ?(A) and ?(B) are sorted in decreasing order.
In this case, the matching function becomes mG (A, B) = maxV ?O(d) trace(A> V BV > ) =
h?(A), ?(B)i due to von Neumann?s trace inequality [25], the maximizer being V = UA UB> .
Therefore, we need only to make an eigen-decomposition and set B 0 = UA ?(B)UA> .
Squared matrices with majorized singular values. Let A = UA ?(A)VA> ? Rd?d and
B = UB ?(B)VB> ? Rd?d , where the singular values are sorted. We have mG (A, B) =
maxU,V ?O(d) trace(A> U BV > ) = h?(A), ?(B)i also from von Neumann?s inequality [25]. To
evaluate the matching function, we need only to make an SVD and set B 0 = UA ?(B)VA> .
5.2
Projected Gradient
The projected gradient algorithm is illustrated in Fig. 2 (right); the relevant part is line 5, which
involves a projection onto the orbitope OG (v). This projection may be hard to compute directly,
since the orbitope may lack a concise half-space representation. However, we next transform this
problem into a projection onto the region cone KG (v) (the proof is in the supplemental material).
Proposition 14 Assume G is a subgroup of O(d). Let g ? G be such that ha, gvi = mG (a, v).
Then, the solution of the problem in line 5 is w? = a ? ?KG (gv) (a ? gv).
Thus, all is necessary is computing the arg-max associated with the matching function, and a black
box that projects onto the region cone KG (v). Again, this step is efficient in several cases:
Permutations. If G = P, the region cone of a point v is the set of points w satisfying vi > vj ?
wi ? wj , for all i, j ? 1, . . . , d. Projecting onto this cone is a well-studied problem in isotonic
regression [26, 27], with existing O(d) algorithms.
Signed permutations. If G = P? , this problem is precisely the evaluation of the proximity operator of the sorted `1 -norm, also solvable in O(d) runtime with a stack-based algorithm [9].
6
Continuation Algorithm
Finally, we present a general continuation procedure for exploring regularization paths when L
is a convex loss function (not necessarily differentiable) and the seed v is not prescribed. The
6
Require: Factor > 0, interpolation parameter ? ? [0, 1]
1: Initialize seed v 0 randomly and set kv 0 k =
2: Set t = 0
3: repeat
4:
Solve wt = arg minwG vt L(w)
5:
Pick v 0t ? Gv t ? KG (wt )
6:
Set next seed v t+1 = (1 + )(?v 0t + (1 ? ?)wt )
7:
t?t+1
8: until kw t kGvt < 1.
b ? {w1 , w2 , . . .}
9: Use cross-validation to choose the best w
Figure 3: Left: Continuation algorithm. Right: Reachable region WG for the hyperoctahedral group, with a
reconstruction loss L(w) = kw ? ak2 . Only points v s.t. ??L(v) = a ? v ? KG (v) belong to this set.
Different initializations of v 0 lead to different paths along WG , all ending in a.
procedure?outlined in Fig. 3?solves instances of Eq. 3 for a sequence of seeds v 1 , v 2 , . . ., using
a simple heuristic for choosing the next seed given the previous one and the current solution.
The basic principle behind this procedure is the same as in other homotopy continuation methods
[28, 29, 30, 31]: we start with very strong regularization (using a small norm ball), and then gradually
weaken the regularization (increasing the ball) while ?tracking? the solution. The process stops
when the solution is found to be in the interior of the ball (the condition in line 8), which means the
regularization constraint is no longer active. The main difference with respect to classical homotopy
methods is that we do not just scale the ball (in our case, the G-orbitope); we also generate new
seeds that shape the ball along the way. To do so, we adopt a simple heuristic (line 6) to make
the seed move toward the current solution wt before scaling the orbitope. This procedure depends
on the initialization (see Fig. 3 for an illustration), which drives the search into different regions.
Reasoning in terms of groups, line 4 makes us move inside the orbits, while line 6 is an heuristic
to jump to a nearby orbit. For any choice of > 0 and ? ? [0, 1], the algorithm is convergent and
produces a strictly decreasing sequence L(w1 ) > L(w2 ) > ? ? ? before it terminates (a proof is
provided as supplementary material). We expect that, eventually, a seed v will be generated that is
b Although it may not be obvious at first sight why would it be desirable
close to the true model w.
b we provide a simple result below (Prop. 15) that sheds some light on this matter, by
that v ? w,
characterizing the set of points in V that are ?reachable? by optimizing Eq. 3.
From the optimality conditions of convex programming [32, p. 257], we have that w? is a solution
of the optimization problem in Eq. 3 if and only if 0 ? ?L(w? ) + NGv (w? ), where ?L(w) denotes
the subdifferential of L at w, and NGv (w) is the normal cone to OG (v) at w, defined in ?3. For
certain seeds v ? V , it may happen that the optimal solution w? of Eq. 3 is the seed itself. Let WG
be the set of seeds with this property:
WG := {v ? V | L(v) ? L(w), ?w G v} = {v ? V | 0 ? ?L(v) + KG (v)},
(14)
where KG (v) is the region cone and the right hand side follows from the optimality conditions. We
next show that this set is all we need to care about.
c
Proposition
15 Consider ?the set of points that are solutions of Eq. 3 for some seed v ? V , WG :=
?
c
w ? V ?v ? V : w ? arg minwG v L(w) . We have WG = WG .
cG . For the reverse direction, suppose that w? ? W
cG ,
Proof: Obviously, v ? WG ? v ? W
?
?
in which case there is some v ? V such that w G v and L(w ) ? L(w) for any w G v.
Since G is a pre-order, it must hold in particular that L(w? ) ? L(w) for any w G w? G v.
Therefore, we also have that w? ? arg minwG w? L(w), i.e., w? ? WG .
7
Simulation Results
We describe the results of numerical experiments when regularizing with the permutahedron (symmetric group) and the signed permutahedron (hyperoctahedral group). All problems were solved
b ? Rd
using the conditional gradient algorithm, as described in ?5. We generated the true model w
7
Figure 4: Learning curves for the permutahedron and signed permutahedron regularizers with a perfect seed.
Shown are averages and standard deviations over 10 trials. The baselines are `1 (three leftmost plots, resp. with
k = 150, 250, 400), and `2 (last plot, with k = 500).
Figure 5: Mean squared errors in the training set (left) and the test set (right) along the regularization path.
For the permutahedra regularizers, this path was traced with the continuation algorithm. The baseline is `1
regularization. The horizontal lines in the right plot show the solutions found with validation in a held-out set.
by sampling the entries from a uniform distribution in [0, 1] and subtracted the mean, keeping k ? d
b was normalized to have unit `2 -norm. Then, we sampled a random nnonzeros; after which w
by-d matrix X with i.i.d. Gaussian entries and variance ? 2 = 1/d, and simulated measurements
b + n, where n ? N (0, ?n2 ) is Gaussian noise. We set d = 500 and ?n = 0.3?.
y = Xw
For the first set of experiments (Fig. 4), we set k ? {150, 250, 400, 500} and varied the number
of measurements n. To assess the advantage of knowing the true parameters up to a group transb up to a constant
formation, we used for the orbitope regularizers a seed in the orbit of the true w,
factor (this constant, and the regularization constants for `1 and `2 , were all chosen with validation in a held-out set). As expected, this information was beneficial, and no significant difference
was observed between the permutahedron and the signed permutahedron. For the second set of
experiments (Fig. 5), where the aim is to assess the performance of the continuation method, no
information about the true model was given. Here, we fixed n = 250 and k = 300 and ran the
continuation algorithm with = 0.1 and ? = 0.0, for 5 different initializations of v 0 . We observe
that this procedure was effective at exploring the orbits, eventually finding a slightly better model
than the one found with `1 and `2 regularizers.
8
Conclusions and Future Work
In this paper, we proposed a group-based regularization scheme using the notion of orbitopes. Simple choices of groups recover commonly used regularizers such as `1 , `2 , `? , spectral and nuclear
matrix norms; as well as some new ones, such as the permutahedron and signed permutahedron.
As a byproduct, we revealed a connection between the permutahedra and the recently proposed
sorted `1 -norm. We derived procedures for learning with these orbit regularizers via conditional and
projected gradient algorithms, and a continuation strategy for orbit exploration.
There are several avenues for future research. For example, certain classes of groups, such as reflection groups [33], have additional properties that may be exploited algorithmically. Our work should
be regarded as a first step toward group-based regularization?we believe that the regularizers studied here are just the tip of the iceberg. Groups and their representations are well studied in other
disciplines [10], and chances are high that this framework can lead to new regularizers that are a
good fit to specific machine learning problems.
Acknowledgments
We thank all reviewers for their valuable comments. This work was partially supported by FCT
grants PTDC/EEI-SII/2312/2012 and PEst-OE/EEI/LA0008/2011, and by the EU/FEDER programme, QREN/POR Lisboa (Portugal), under the Intelligo project (contract 2012/24803).
8
References
[1] R. Tibshirani. Regression Shrinkage and Selection via the Lasso. Journal of the Royal Statistical Society
B., pages 267?288, 1996.
[2] D. Donoho. Compressed Sensing. IEEE Transactions on Information Theory, 52(4):1289?1306, 2006.
[3] E. Cand`es and B. Recht. Exact Matrix Completion via Convex Optimization. Foundations of Computational Mathematics, 9(6):717?772, 2009.
[4] F. Bach, R. Jenatton, J. Mairal, and G. Obozinski. Convex optimization with sparsity-inducing norms. In
Optimization for Machine Learning. MIT Press, 2011.
[5] S. Negahban, P. Ravikumar, M. Wainwright, and B. Yu. A Unified Framework for High-Dimensional
Analysis of M-estimators with Decomposable Regularizers. In Neural Information Processing Systems,
pages 1348?1356, 2009.
[6] V. Chandrasekaran, B. Recht, P. Parrilo, and A. Willsky. The Convex Geometry of Linear Inverse Problems. Foundations of Computational Mathematics, 12(6):805?849, 2012.
[7] A. Marshall, I. Olkin, and B. Arnold. Inequalities: Theory of Majorization and Its Applications. Springer,
2010.
[8] R. Sanyal, F. Sottile, and B. Sturmfels. Orbitopes. Technical report, arXiv:0911.5436, 2009.
[9] M. Bogdan, E. Berg, W. Su, and E. Cand`es. Statistical estimation and testing via the ordered `1 norm.
Technical report, arXiv:1310.1969, 2013.
[10] J. Serre and L. Scott. Linear Representations of Finite Groups, volume 42. Springer, 1977.
[11] Y. Tong. Probability Inequalities in Multivariate Distributions, volume 5. Academic Press New York,
1980.
[12] G. Ziegler. Lectures on Polytopes, volume 152. Springer, 1995.
[13] M. Yuan and Y. Lin. Model selection and estimation in regression with grouped variables. Journal of the
Royal Statistical Society Series B (Statistical Methodology), 68(1):49, 2006.
[14] M. Eaton. On group induced orderings, monotone functions, and convolution theorems. Lecture NotesMonograph Series, pages 13?25, 1984.
[15] A. Giovagnoli and H. Wynn. G-Majorization with Applications to Matrix Orderings. Linear algebra and
its applications, 67:111?135, 1985.
[16] R.T. Rockafellar. Convex Analysis. Princeton University Press, 1970.
[17] X. Zeng and M. A. T. Figueiredo. Decreasing weighted sorted `1 regularization. Technical report,
arXiv:1404.3184, 2014.
[18] H. Bondell and B. Reich. Simultaneous Regression Shrinkage, Variable Selection, and Supervised Clustering of Predictors with OSCAR. Biometrics, 64(1):115?123, 2008.
[19] L. Zhong and J. Kwok. Efficient Sparse Modeling with Automatic Feature Grouping. IEEE Transactions
on Neural Networks and Learning Systems, 23(9):1436?1447, 2012.
[20] G. Hardy, J. Littlewood, and G. P?olya. Inequalities. Cambridge University Press, 1952.
[21] M. Frank and P. Wolfe. An Algorithm for Quadratic Programming. Naval research logistics quarterly, 3
(1-2):95?110, 1956.
[22] M. Jaggi. Revisiting Frank-Wolfe: Projection-free Sparse Convex Optimization. In Proc. of the International Conference on Machine Learning, pages 427?435, 2013.
[23] S.J. Wright, R. Nowak, and M. A. T. Figueiredo. Sparse reconstruction by separable approximation. IEEE
Transactions on Signal Processing, 57(7):2479?2493, 2009.
[24] A. Beck and M. Teboulle. A fast iterative shrinkage-thresholding algorithm for linear inverse problems.
SIAM Journal on Imaging Sciences, 2(1):183?202, 2009.
[25] L. Mirsky. A trace inequality of john von neumann. Monatshefte f?ur Mathematik, 79(4):303?306, 1975.
[26] P. Pardalos and G. Xue. Algorithms for a Class of Isotonic Regression Problems. Algorithmica, 23(3):
211?222, 1999.
[27] R. Luss, S. Rosset, and M. Shahar. Decomposing Isotonic Regression for Efficiently Solving Large
Problems. In Neural Information Processing Systems, pages 1513?1521, 2010.
[28] M.R. Osborne, B. Presnell, and B.A. Turlach. A new approach to variable selection in least squares
problems. IMA Journal of Numerical Analysis, 20:389?403, 2000.
[29] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani. Least angle regression. The Annals of statistics, 32:
407?499, 2004.
[30] M. A. T. Figueiredo, R. Nowak, and S. Wright. Gradient projection for sparse reconstruction: Application
to compressed sensing and other inverse problems. IEEE Journal of Selected Topics in Signal Processing,
1(4):586?597, 2007.
[31] E. Hale, W. Yin, and Y. Zhang. Fixed-point continuation for l1-minimization: Methodology and convergence. SIAM Journal on Optimization, 19:1107?1130, 2008.
[32] D.P. Bertsekas, A. Nedic, and A.E. Ozdaglar. Convex analysis and optimization. Athena Scientific, 2003.
[33] A. Steerneman. g-majorization, group-induced cone orderings, and reflection groups. Linear Algebra and
its Applications, 127:107?119, 1990.
[34] J.J. Moreau. Fonctions convexes duales et points proximaux dans un espace hilbertien. CR de l?Acad?emie
des Sciences de Paris S?erie A Mathematics, 255:2897?2899, 1962.
9
| 5398 |@word trial:1 norm:45 turlach:1 hu:4 closure:1 simulation:3 decomposition:1 pick:1 concise:1 contains:1 kpv:1 series:2 hardy:2 existing:1 current:3 com:1 olkin:1 gmail:1 must:2 written:1 john:1 numerical:2 happen:1 shape:1 gv:24 designed:1 plot:3 maxv:1 half:1 selected:1 characterization:3 provides:1 revisited:1 zhang:1 along:3 c2:6 kvk2:2 sii:1 yuan:1 doubly:1 ray:1 inside:1 manner:1 introduce:2 theoretically:1 expected:1 indeed:3 cand:2 olya:1 decreasing:6 ivanov:1 considering:1 ua:7 conv:8 becomes:6 pof:1 underlying:1 bounded:1 project:2 increasing:1 provided:1 kg:9 parsimony:1 supplemental:2 unified:1 finding:1 transformation:1 guarantee:1 act:2 proximaux:1 shed:1 runtime:2 scaled:1 control:2 unit:1 grant:1 omit:3 ozdaglar:1 bertsekas:1 before:4 sd:5 instituto:4 consequence:1 acad:1 ak:1 id:3 path:4 interpolation:1 signed:15 black:1 au:1 studied:4 initialization:3 equivalence:1 mirsky:1 negrinho:2 factorization:1 horn:1 acknowledgment:1 testing:1 atomic:12 block:3 procedure:10 area:1 empirical:1 axiom:2 matching:12 projection:5 pre:3 suggest:1 get:1 onto:5 interior:1 close:1 operator:1 selection:4 unsigned:1 risk:1 context:1 writing:2 descending:1 isotonic:3 equivalent:1 reviewer:1 go:1 convex:17 formulate:1 decomposable:1 permutes:1 estimator:1 regarded:1 nuclear:4 notion:2 traditionally:1 resp:1 pt:1 play:3 suppose:1 annals:1 exact:1 losing:1 programming:2 designing:1 element:3 wolfe:2 satisfying:3 observed:1 role:4 solved:1 capture:1 hv:1 revisiting:1 region:13 wj:2 oe:1 ordering:6 trade:1 eu:1 valuable:1 ran:1 mentioned:1 intuition:1 pd:3 complexity:5 ui:1 solving:2 algebra:2 serve:1 basis:1 gu:1 represented:1 regularizer:2 notesmonograph:1 describe:1 effective:1 fast:1 formation:1 choosing:2 whose:2 heuristic:3 widely:1 larger:1 solve:1 say:1 supplementary:1 otherwise:3 compressed:3 wg:9 majorized:7 statistic:5 g1:2 hilbertien:1 transform:1 itself:1 obviously:1 sequence:2 eigenvalue:6 mg:19 differentiable:1 advantage:1 propose:2 reconstruction:3 product:3 dans:1 aligned:1 relevant:2 loop:1 iff:3 degenerate:1 inducing:2 kv:1 achievement:1 convergence:1 neumann:3 produce:1 perfect:1 bogdan:2 derive:1 completion:3 op:9 eq:10 strong:1 solves:1 involves:1 implies:3 alameda:1 direction:1 radius:1 hull:6 stochastic:1 exploration:3 material:3 pardalos:1 require:1 ptdc:1 fix:1 homotopy:2 proposition:12 exploring:2 strictly:1 hold:2 proximity:1 confine:1 wright:2 normal:4 great:1 seed:27 algorithmic:2 eaton:1 maxu:2 adopt:1 estimation:3 proc:1 ziegler:1 grouped:1 successfully:1 tool:1 weighted:3 minimization:2 mit:1 always:2 sight:1 gaussian:2 aim:1 rather:1 cr:1 shrinkage:3 zhong:1 og:17 corollary:2 derived:1 focus:1 naval:1 rank:1 cg:2 baseline:2 dependent:2 associativity:1 relation:3 arg:6 dual:8 denoted:2 special:1 initialize:3 ak2:1 field:1 construct:1 never:1 equal:1 atom:1 sampling:1 kw:3 yu:1 future:2 espace:1 report:3 randomly:1 ima:1 beck:1 replaced:1 intended:1 geometry:1 algorithmica:1 highly:1 evaluation:2 kvk:2 light:1 behind:2 regularizers:11 held:2 amenable:2 nowak:2 byproduct:2 necessary:3 minw:4 orthogonal:8 biometrics:1 incomplete:1 euclidean:1 orbit:17 weaken:1 fenchel:2 column:1 modeling:2 instance:1 teboulle:1 marshall:1 maximization:1 reflexive:1 deviation:1 entry:5 uniform:1 predictor:1 characterize:1 proximal:1 xue:1 rosset:1 recht:2 international:1 negahban:1 siam:2 sequel:2 contract:1 off:1 invertible:1 tip:1 discipline:1 w1:2 again:2 squared:5 satisfied:1 von:3 containing:1 choose:2 por:1 qualifies:1 account:1 parrilo:1 de:5 subsumes:3 rockafellar:1 matter:2 satisfy:1 mp:2 eei:2 vi:3 depends:1 lab:1 kwk:2 sup:3 wynn:1 start:3 recover:6 sort:4 slope:3 majorization:10 ass:2 square:1 atm:1 minimize:1 formed:3 variance:1 efficiently:1 spaced:1 generalize:1 lu:1 drive:1 simultaneous:1 afonso:1 ed:1 aligns:1 definition:4 evaluates:1 telecomunicac:2 failure:1 obvious:1 associated:2 proof:8 stop:1 sampled:1 knowledge:2 ut:2 efron:1 hilbert:1 jenatton:1 back:1 supervised:1 methodology:2 maximally:1 formulation:3 done:1 evaluated:1 box:1 generality:3 furthermore:1 just:2 until:1 hand:2 horizontal:1 zeng:2 su:1 overlapping:1 maximizer:1 lack:1 defines:1 mode:1 reveal:1 scientific:1 believe:1 building:3 serre:1 normalized:1 concept:8 requiring:1 counterpart:1 true:5 regularization:19 seeded:2 symmetric:10 illustrated:2 rooted:1 kak:1 leftmost:1 presenting:1 ecnico:2 l1:1 gh:3 reflection:2 reasoning:1 consideration:1 recently:3 superior:2 iceberg:1 volume:3 belong:1 kwk2:1 measurement:2 significant:1 cambridge:1 fonctions:1 rd:22 automatic:1 outlined:2 mathematics:3 portugal:4 reachable:2 permutahedron:20 reich:1 longer:1 jaggi:1 multivariate:2 recent:2 perspective:1 optimizing:1 inf:1 reverse:1 certain:2 inequality:8 shahar:1 success:1 arbitrarily:1 kwk1:1 vt:1 exploited:1 preserving:1 seen:2 fortunately:1 care:1 additional:1 maximize:2 signal:2 ii:3 lisboa:4 desirable:1 reduces:1 technical:3 match:1 characterized:2 academic:1 cross:1 long:1 bach:1 lin:1 e1:4 ravikumar:1 iog:2 controlled:1 va:2 regression:8 basic:1 arxiv:3 confined:1 pinpointing:1 c1:6 subdifferential:1 singular:6 crucial:1 duales:1 w2:2 rest:1 comment:1 induced:10 schur:1 call:1 revealed:1 iii:2 concerned:1 switch:1 variety:1 fit:2 hastie:1 lasso:1 inner:2 regarding:1 prototype:1 idea:1 knowing:1 avenue:1 feder:1 presnell:1 penalty:1 york:1 action:11 clear:1 sturmfels:1 continuation:11 generate:1 andr:1 canonical:1 revisit:2 sign:2 estimated:1 algorithmically:1 tibshirani:2 unveiling:1 group:57 key:2 traced:1 pj:1 v1:4 imaging:1 monotone:1 cone:17 sum:1 inverse:5 letter:2 oscar:2 angle:1 throughout:1 chandrasekaran:2 scaling:1 vb:1 renato:2 bound:1 def:1 conjugation:1 centrally:2 convergent:1 quadratic:1 bv:2 precisely:2 constraint:1 kwkp:2 nearby:1 aspect:1 prescribed:2 optimality:2 separable:1 martin:1 fct:1 structured:1 convexes:1 ball:17 erie:1 smaller:1 terminates:1 beneficial:1 slightly:1 ur:1 wi:5 intuitively:1 invariant:1 projecting:2 gradually:1 taken:1 bondell:1 mathematik:1 turn:1 eventually:2 letting:1 tractable:1 end:2 available:3 operation:4 endowed:1 decomposing:1 apply:1 observe:1 kwok:1 v2:2 spectral:4 quarterly:1 stepsize:1 subtracted:1 pest:1 eigen:1 existence:2 denotes:4 top:1 cf:1 clustering:1 unifying:1 xw:1 build:1 classical:1 society:2 objective:1 move:2 quantity:2 strategy:4 kak2:1 diagonal:3 surrogate:1 gradient:18 kth:1 subspace:1 thank:1 simulated:1 vd:2 athena:1 w0:2 topic:1 toward:2 boldface:1 willsky:1 besides:1 illustration:1 equivalently:2 mostly:1 statement:2 frank:2 trace:4 negative:1 motivates:1 allowing:1 av:2 convolution:1 finite:1 logistics:1 defining:2 precise:1 varied:1 stack:1 arbitrary:2 introduced:2 namely:1 paris:1 toolbox:1 connection:4 optimized:1 established:3 subgroup:4 polytopes:1 below:4 pattern:1 scott:1 regime:1 sparsity:4 emie:1 max:3 royal:2 explanation:1 wainwright:1 regularized:1 indicator:1 solvable:1 nedic:1 representing:1 scheme:1 transitive:1 prior:1 understanding:1 review:1 literature:1 geometric:1 multiplication:1 contributing:1 relative:1 discovery:1 loss:7 expect:1 permutation:12 highlight:1 rationale:1 lecture:2 validation:3 foundation:2 principle:2 thresholding:1 row:1 penalized:1 gl:3 surprisingly:1 repeat:1 last:1 figueiredo:4 keeping:1 henriques:1 side:2 free:1 supported:1 arnold:1 johnstone:1 taking:1 characterizing:1 absolute:4 sparse:7 moreau:1 regard:1 curve:1 evaluating:1 cumulative:1 ending:1 commonly:1 jump:1 projected:9 programme:1 transaction:3 compact:1 preferred:1 active:1 reveals:1 mairal:1 conclude:1 search:1 iterative:1 un:1 why:1 decoupling:1 complex:1 necessarily:2 diag:5 vj:4 main:3 linearly:1 motivation:1 noise:1 n2:1 osborne:1 fig:9 tong:1 pv:1 theorem:1 specific:2 hale:1 sensing:3 symbol:1 r2:1 maxi:2 littlewood:1 grouping:1 false:1 importance:1 magnitude:2 illustrates:1 sorting:1 depicted:1 led:1 yin:1 simply:1 ordered:1 tracking:1 g2:2 scalar:2 partially:1 springer:3 corresponds:2 minimizer:1 satisfies:5 chance:1 prop:8 obozinski:1 conditional:10 sorted:15 identity:2 donoho:1 hard:1 except:1 acting:3 wt:11 called:5 invariance:1 e:4 duality:1 svd:1 berg:1 kgv:2 latter:1 ub:4 evaluate:2 princeton:1 regularizing:1 |
4,858 | 5,399 | Covariance shrinkage for autocorrelated data
Daniel Bartz
Department of Computer Science
TU Berlin, Berlin, Germany
[email protected]
?
Klaus-Robert Muller
TU Berlin, Berlin, Germany
Korea University, Korea, Seoul
[email protected]
Abstract
The accurate estimation of covariance matrices is essential for many signal processing and machine learning algorithms. In high dimensional settings the sample
covariance is known to perform poorly, hence regularization strategies such as
analytic shrinkage of Ledoit/Wolf are applied. In the standard setting, i.i.d. data
is assumed, however, in practice, time series typically exhibit strong autocorrelation structure, which introduces a pronounced estimation bias. Recent work by
Sancetta has extended the shrinkage framework beyond i.i.d. data. We contribute
in this work by showing that the Sancetta estimator, while being consistent in the
high-dimensional limit, suffers from a high bias in finite sample sizes. We propose
an alternative estimator, which is (1) unbiased, (2) less sensitive to hyperparameter choice and (3) yields superior performance in simulations on toy data and on a
real world data set from an EEG-based Brain-Computer-Interfacing experiment.
1
Introduction and Motivation
Covariance matrices are a key ingredient in many algorithms in signal processing, machine learning
and statistics. The standard estimator, the sample covariance matrix S, has appealing properties in
the limit of large sample sizes n: its entries are unbiased and consistent [HTF08]. On the other hand,
for sample sizes of the order of the dimensionality p or even smaller, its entries have a high variance
and the spectrum has a large systematic error. In particular, large eigenvalues are overestimated and
small eigenvalues underestimated, the condition number is large and the matrix difficult to invert
[MP67, ER05, BS10]. One way to counteract this issue is to shrink S towards a biased estimator T
(the shrinkage target) with lower variance [Ste56],
Csh := (1 ? ?)S + ?T,
the default choice being T = p?1 trace(S)I, the identity multiplied by the average eigenvalue.
For the optimal shrinkage intensity ?? , a reduction of the expected mean squared error is guaranteed [LW04]. Model selection for ? can be done by cross-validation (CV) with the known drawbacks: for (i) problems with many hyperparameters, (ii) very high-dimensional data sets, or (iii)
online settings which need fast responses, CV can become unfeasible and a faster model selection
method is required. A popular alternative to CV is Ledoit and Wolf?s analytic shrinkage procedure [LW04] and more recent variants [CWEH10, BM13]. Analytic shrinkage directly estimates the
shrinkage intensity which minimizes the expected mean squared error of the convex combination
with a negligible computational cost, especially for applications which rely on expensive matrix
inversions or eigendecompositions in high dimensions.
All of the above algorithms assume i.i.d. data. Real world time series, however, are often non-i.i.d. as
they possess pronounced autocorrelation (AC). This makes covariance estimation in high dimensions
even harder: the data dependence lowers the effective sample size available for constructing the
estimator [TZ84]. Thus, stronger regularization ? will be needed. In Figure 1 the simple case of an
autoregressive model serves as an example for an arbitrary generative model with autocorrelation.
1
40
0.8
0.6
0.4
0.2
0
0
20
40
60
80
100
time lag
40
population
sample
30
30
variance
no AC (AR?coeff. = 0)
low AC (AR?coeff. = 0.7)
high AC (AR?coeff. = 0.95)
eigenvalue
autocorrelation (AC)
1
20
10
0
150
20
10
160
170
180
190
200
0
150
#eigenvalue
160
170
180
190
200
#sample eigendirection
Figure 1: Dependency of the eigendecomposition on autocorrelation. p = 200, n = 250.
The Figure shows, for three levels of autocorrelation (left), the population and sample eigenvalues
(middle): with increasing autocorrelation the sample eigenvalues become more biased. This bias
is an optimistic measure for the quality of the covariance estimator: it neglects that population and
sample eigenbasis also differ [LW12]. Comparing sample eigenvalues to the population variance in
the sample eigenbasis, the bias is even larger (right).
In practice, violations of the i.i.d. assumption are often ignored [LG11, SBMK13, GLL+ 14], although Sancetta proposed a consistent shrinkage estimator under autocorrelation [San08]. In this
paper, we contribute by showing in theory, simulations and on real world data, that (i) ignoring autocorrelations for shrinkage leads to large estimation errors and (ii) for finite samples Sancetta?s estimator is still substantially biased and highly sensitive to the number of incorporated time lags. We
propose a new bias-corrected estimator which (iii) outperforms standard shrinkage and Sancetta?s
method under the presence of autocorrelation and (iv) is robust to the choice of the lag parameter.
2
Shrinkage for autocorrelated data
Ledoit and Wolf derived a formula for the optimal shrinkage intensity [LW04, SS05]:
P
ij Var Sij
?
h
? =P
2 i .
E
S
?
T
ij
ij
ij
(1)
? is obtained by replacing expectations with sample estimates:
The analytic shrinkage estimator ?
n
n
2
X
1X
d Sij = 1
x
x
x
x
?
(2)
Var Sij
?? Var
it jt
is js
n2 s=1
n t=1
h
2 i
b (Sij ? Tij )2 = (Sij ? Tij )2 ,
E Sij ? Tij
?? E
(3)
where xit is the tth observation of variable i. While the estimator eq. (3) is unbiased even under a
violation of the i.i.d. assumption, the estimator eq. (2) is based on
!
n
1X
i.i.d. 1
Var
xit xjt = Var (xit xjt ) .
n t=1
n
If the data are autocorrelated, cross terms cannot be ignored and we obtain
!
n
n
1X
1 X
Var
xit xjt = 2
Cov(xit xjt , xis xjs )
n t=1
n s,t=1
=
=:
n?1
1
2 X n?s
Cov(xit xjt , xit xjt ) +
Cov(xit xjt , xi,t+s xj,t+s )
n
n s=1 n
n?1
1
2X
?ij (0) +
?ij (s)
n
n s=1
(4)
Figure 2 illustrates the effect of ignoring the cross terms for increasing autocorrelation (larger ARcoefficients, see section 3 for details on the simulation). It compares standard shrinkage to an oracle
shrinkage based on the population variance of the sample covariance1 . The population variance of S
1
calculated by resampling.
2
0.04
0.02
0
0
0.2
0.4
0.6
0.8
1
1
0.8
0.6
0.4
0.2
0
0
0.2
0.4
0.6
0.8
1
1
population
sample
pop. var(S) shrinkage
standard shrinkage
15
0.8
variance
0.06
impr. over sample cov.
shrinkage intensity ?
norm. ?ij var(Sij)
0.1
0.08
0.6
0.4
0.2
0
0.2
0.4
0.6
0.8
AR?coefficients
1
10
AR?coeff. = 0.7
5
0
0
50
100
150
200
#sample eigendirection
Figure 2: Dependency of shrinkage on autocorrelation. p = 200, n = 250.
increases because the effective sample size is reduced [TZ84], yet the standard shrinkage variance
estimator eq. (2) does not increase (outer left). As a consequence, for oracle shrinkage the shrinkage
intensity increases, for the standard shrinkage estimator it even decreases because the denominator
in eq. (1) grows (middle left). With increasing autocorrelation, the sample covariance becomes
a less precise estimator: for optimal (stronger) shrinkage more improvement becomes possible,
yet standard shrinkage does not improve (middle right). Looking at the variance estimates in the
sample eigendirections for AR-coefficients of 0.7, we see that the bias of standard shrinkage is only
marginally smaller than the bias of the sample covariance, while oracle shrinkage yields a substantial
bias reduction (outer right).
Sancetta-estimator An estimator for eq. (4) was proposed by [San08]:
n?s
X
? San (s) := 1
?
(xit xjt ? Sij ) (xi,t+s xj,t+s ? Sij ) ,
ij
n t=1
!
n?1
X
San,b
1 ? San
San
d
?
:=
Var Sij
?ij (0) + 2
?(s/b)?ij (s) , b > 0,
n
s=1
(5)
where ? is a kernel which has to fulfill Assumption B in [And91]. We will restrict our analysis to
the truncated kernel ?TR (x) = {1 for |x| ? 1, 0 otherwise} to obtain less cluttered formulas2 . The
kernel parameter b describes how many time lags are taken into account.
The Sancetta estimator behaves well in the high dimensional limit: the main theoretical result states
that for (i) a fixed decay of the autocorrelation, (ii) b, n ? ? and (iii) b2 increasing at a lower rate
than n, the estimator is consistent independently of the rate of p (for details, see [San08]). This is
in line with the results in [LW04, CWEH10, BM13]: as long as n increases, all of these shrinkage
estimators are consistent.
Bias of the Sancetta-estimator In the following we will show that the Sancetta-estimator is suboptimal in finite samples: it has a non-negligible bias. To understand this, consider a lag s large
enough to have ?ij (s) ? 0. If we approximate the expectation of the Sancetta-estimator, we see that
it is biased downwards:
" n?s
#
h
i
X
1
2
? San
E ?
xit xjt xi,t+s xj,t+s ? Sij
.
ij (s) ? E
n t=1
2
n?s 2
n?s
?
E [Sij ] ? E Sij
=?
Var (Sij ) < 0.
n
n
Bias-corrected (BC) estimator We propose a bias-corrected estimator for the variance of the
entries in the sample covariance matrix:
n?s
1X
2
? BC
:=
?
(s)
xit xjt xi,t+s xj,t+s ? Sij
,
ij
n t=1
d Sij
Var
2
BC,b
1
:=
n ? 1 ? 2b + b(b + 1)/n
? BC
?
ij (0) + 2
(6)
n?1
X
!
BC
?
?TR (s/b)?ij (s) ,
b > 0.
s=1
in his simulations, Sancetta uses the Bartlett kernel. For fixed b, this increases the truncation bias.
3
? BC (s) is very similar to ?
? San (s), but slightly easier to compute. The main difference
The estimator ?
ij
ij
BC,b
d Sij
is the denominator in Var
: it is smaller than n and thus corrects the downwards bias.
2.1
Theoretical results
It is straightforward to extend the theoretical results on the Sancetta estimator ([San08], see summary above) to our proposed estimator. In the following, to better understand the limitations of the
Sancetta estimator, we will provide a complementary theoretical analysis on the behaviour of the
estimator for finite n.
Our theoretical results are based on the analysis of a sequence of statistical models indexed by p. Xp
denotes a p ? n matrix of n observations of p variables with mean zero and covariance matrix Cp .
Y p = R>
p Xp denotes the same observations rotated in their eigenbasis, having diagonal covariance
p
p
3
?p = R>
p Cp Rp . Lower case letters xit and yit denote the entries of Xp and Yp , respectively . The
analysis is based on the following assumptions:
Assumption 1 (A1, bound on average eighth moment). There exists a constant K1 independent of
p such that
p
1X
E[(xpi1 )8 ] ? K1 .
p i=1
Assumption 2 (A2, uncorrelatedness of higher moments). Let Q denote the set of quadruples
{i,j,k,l} of distinct integers.
P
p
p
2 p p
i,j,kl,l?Qp Cov [yi1 yj1 , yk,1+s yl,1+s ]
= O p?1 ,
|Qp |
and
h
i
p p 2
p
p
2
Cov
(y
y
)
,
(y
y
)
i1 j1
i,j,kl,l?Qp
k,1+s l,1+s
P
?s :
|Qp |
= O p?1 ,
hold.
Assumption 3 (A3, non-degeneracy). There exists a constant K2 such that
p
1X
E[(xpi1 )2 ] ? K2 .
p i=1
Assumption 4 (A4, moment relation). There exist constants ?4 , ?8 , ?4 and ?8 such that
E[yi8 ]
?
(1 + ?8 )E2 [yi4 ],
E[yi4 ]
?
(1 + ?4 )E2 [yi2 ],
E[yi8 ]
?
(1 + ?8 )E2 [yi4 ],
E[yi4 ]
?
(1 + ?4 )E2 [yi2 ].
Remarks on the assumptions A restriction on the eighth moment (assumption A1) is necessary
because the estimators eq. (2), (3), (5) and (6) contain fourth moments, their variances therefore
contain eighths moments. Note that, contrary to the similar assumption in the eigenbasis in [LW04],
A1 poses no restriction on the covariance structure [BM13]. To quantify the effect of averaging
over dimensions, assumption A2 restricts the correlations of higher moments in the eigenbasis. This
assumption is trivially fulfilled for Gaussian data, but much weaker (see [LW04]). Assumption A3
rules out the degenerate case of adding observation channels without any variance and assumption
A4 excludes distributions with arbitrarily heavy tails.
Based on these assumptions, we can analyse the difference between the Sancetta-estimator and our
proposed estimator for large p:
Theorem 1 (consistency under ?fixed n?-asympotics). Let A1, A2, A3, A4 hold. We then have
1 X
Var (Sij ) = ?(1)
p2 ij
3
We shall often drop the sequence index p and the observation index t to improve readability of formulas.
4
no AC (b = 10)
?3
?ij var(Sij)/p2
9.5
x 10
low AC (b = 20)
high AC (b = 90)
0.09
30
25
9
0.08
8.5
0.07
8
0.06
7.5
0.05
7
0.04
6.5
0.03
population
shrinkage
Sancetta
bias?corr
20
15
6
0
100
200
300
400
500
10
5
0.02
0
100
200
300
400
500
0
0
dimensionality
dimensionality
100
200
300
400
500
dimensionality
Figure 3: Dependence of the variance estimates on the dimensionality. Averaged over R = 50
models. n = 250.
2
P 2 !
1 X
2
San,b
j ?j
San,b
San,b
d
+O P
E
+ BiasTR
Var
(Sij ) ? Var (Sij )
p2
= Bias
( j ?j )2
ij
2
P 2 !
1 X
BC,b
j ?j
BC,b 2
d
E
2
Var
(Sij ) ? Var (Sij )
= BiasTR
+O P
( j ?j ) 2
p ij
where the ?i denote the eigenvalues of C and
)
(
n
n
b
4 X X X
1 X 1 + 2b ? b(b + 1)/n
San,b
:= ? 2
Cov [xit xjt , xiu xju ]
Var (Sij ) ? 3
Bias
p ij
n
n s=1 t=n?s u=1
:= ?
BiasSan,b
TR
n
1 2 X X n?s
Cov [xit xjt , xi,t+s xj,t+s ]
p2 n ij
n
s=b+1
BiasBC,b
TR := ?
1
2
2
p n ? 1 ? 2b +
X n?1
X
b(b+1)
n
Cov [xit xjt , xi,t+s xj,t+s ]
ij s=b+1
Proof. see the supplemental material.
Remarks on Theorem 1 (i) The mean squared error of both estimators consists of a bias and a
variance term. Both estimators have a truncation bias which is a consequence of including only a
limited number of time lags into the variance estimation. When b is chosen sufficiently high, this
term gets close to zero. (ii) The Sancetta-estimator has an additional bias term which is smaller than
zero in each dimension and therefore does not average out. Simulations will show that, as a consequence, the Sancetta-estimator has a strong bias which gets larger with increasing lag parameter b.
P
P
2
(iii) The variance of both estimators behaves as O( i ?i2 / ( i ?i ) ): the more the variance of the
data is spread over the eigendirections, the smaller the variance of the estimators. This bound is
minimal if the eigenvalues are identical. (iv) Theorem 1 does not make a statement on the relative
sizes of the variances of the estimators. Note that the BC estimator mainly differs by a multiplicative
factor > 1, hence the variance is larger, but not relative to the expectation of the estimator.
3
Simulations
Our simulations are based on those in [San08]: We average over R = 50 multivariate Gaussian
AR(1) models
~xt = A~xt?1 + ~t ,
4
with parameter matrix A = ?AC ? I , with ?no AC = 0, ?low AC = 0.7, and ?high AC = 0.95 (see Figure 1). The innovations it are Gaussian with variances ?i2 drawn from a log-normal distribution
4
more complex parameter matrices or a different generative model do not pose a problem for the biascorrected estimator. The simple model was chosen for clarity of presentation.
5
no AC
?3
?ij var(Sij)/p2
8
x 10
shrinkage intensity ?
high AC
20
0.08
15
0.06
10
0.04
5
6
4
2
0
PRIAL
low AC
0.1
25
50
75
100
0.5
0.4
0.02
0
25
50
75
100
0
0
0.6
1
0.5
0.8
0.4
0.6
0.3
0.4
0.2
0.2
25
50
75
100
25
50
75
100
0.3
0.2
0.1
0
25
50
75
100
0.1
0
25
50
75
100
0
0
0.4
0.7
1
0.35
0.6
0.8
0.3
0.5
0.6
0.25
0.4
0.4
0.2
0.3
0.2
0
25
50
75
100
0.2
0
25
50
75
100
0
0
pop. var(S)
shrinkage
Sancetta
bias?corr
25
50
75
100
lag parameter b
Figure 4: Robustness to the choice of lag parameter b. Variance estimates (upper row), shrinkage
intensities (middle row) and improvement over sample covariance (lower row). Averaged over R =
50 models. p = 200, n = 250.
with mean ? = 1 and scale parameter ? = 0.5. For each model, we generate K = P
50 data sets to
calculate the std. deviations of the estimators and to obtain an approximation of p?2 ij Var (Sij ).
Simulation 1 analyses the dependence of the estimators on the dimensionality of the data. The number of observations is fixed at n = 250 and the lag parameter b chosen by hand such that the whole
autocorrelation is covered5 : bno AC = 10, blow AC = 20 and bhigh AC = 90. Figure 3 shows that the
standard shrinkage estimator is unbiased and has low variance in the no AC-setting, but under the
presence of autocorrelation it strongly underestimates the variance. As predicted by Theorem 1,
the Sancetta estimator is also biased; its remains stays constant for increasing dimensionality. Our
proposed estimator has no visible bias. For increasing dimensionality the variances of all estimators decrease. Relative to the average estimate, there is no visible difference between the standard
deviations of the Sancetta and the BC estimator.
Simulation 2 analyses the dependency on the lag parameter b for fixed dimensionality p = 200
and number of observations n = 250. In addition to variance estimates, Figure 4 reports shrinkage
intensities and the percentage improvement in absolute loss (PRIAL) over the sample covariance
matrix:
EkS ? Ck ? EkC{pop., shr, San., BC} ? Ck
.
PRIAL C{pop., shr, San., BC} =
EkS ? Ck
The three quantities show very similar behaviour. Standard shrinkage performs well in the no ACcase, but is strongly biased in the autocorrelated settings. The Sancetta estimator is very sensitive to
the choice of the lag parameter b. For low AC, the bias at the optimal b is small: only a small number
of biased terms are included. For high AC the optimal b is larger, the higher number of biased terms
causes a larger bias. The BC-estimator is very robust: it performs well for all b large enough to
capture the autocorrelation. For very large b its variance increases slightly, but this has practically
5
for b < 1, optimal in the no AC-setting, Sancetta and BC estimator are equivalent to standard shrinkage.
6
0.8
0.2
0
0.7
0.65
sample cov
standard shrinkage
sancetta
bias?corr
cross?val.
0.6
25
50
0.55
0
75
time lag
5
10
15
0.02
0.01
0
0
20
accuracy
AC
0
200
time lag
300
0.05
10
15
20
0
5
10
15
20
number of trials per class
0.2
0.7
0.65
sample cov
standard shrinkage
sancetta
bias?corr
cross?val.
0.6
100
0.1
0
5
0.8
0.5
0.15
number of trials per class
0.75
b = 300
0.03
0.55
0
5
10
15
20
number of trials per class
0.05
shrinkage intensity
1
?0.5
0
0.04
number of trials per class
accuracy ? acc(sample cov)
?0.5
0
0.05
shrinkage intensity
accuracy
AC
b = 75
0.75
0.5
accuracy ? acc(sample cov)
1
0.04
0.03
0.02
0.01
0
0
0.15
0.1
0.05
0
5
10
15
number of trials per class
20
0
5
10
15
20
number of trials per class
Figure 5: BCI motor imagery data for lag parameter b = 75 (upper row) and b = 300 (lower row).
Averaged over subjects and K = 100 runs.
no effect on the PRIAL. An interesting aspect is that our proposed estimator even outperforms
shrinkage based on the the population Var (Sij ) (calculated by resampling). This results from the
d Sij BC,b with the sample estimate eq. (3) of the denominator in
correlation of the estimator Var
eq. (1).
4
Real World Data: Brain Computer Interface based on Motor Imagery
As an example of autocorrelated data we reanalyzed a data set from a motor imagery experiment. In
the experiment, brain activity for two different imagined movements was measured via EEG (p = 55
channels, 80 subjects, 150 trials per subject, each trial with ntrial = 390 measurements [BSH+ 10]).
The frequency band was optimized for each subject and from the class-wise covariance matrices, 1-3
filters per class were extracted by Common Spatial Patterns (CSP), adaptively chosen by a heuristic
(see [BTL+ 08]). We trained Linear Discriminant Analysis on log-variance features.
To improve the estimate of the class covariances on these highly autocorrelated data, standard shrinkage, Sancetta shrinkage, cross-validation and and our proposed BC shrinkage estimator were applied. The covariance structure is far from diagonal, therefore, for each subject, we used the average
of the class covariances of the other subjects as shrinkage target [BLT+ 11]. Shrinkage is dominated by the influence of high-variance directions [BM13], which are pronounced in this data set.
To reduce this effect, we rescaled, only for the calculation of the shrinkage intensities, the first five
principal components to have the same variance as the sixth principal component.
We analyse the dependency of the four algorithms on the number of supplied training trials. Figure 5
(upper row) shows results for an optimized time lag (b = 75) which captures well the autocorrelation of the data (outer left). Taking the autocorrelation into account makes a clear difference
(middle left/right): while standard shrinkage outperforms the sample covariance, it is clearly outperformed by the autocorrelation-adjusted approaches. The Sancetta-estimator is slightly worse
than our proposed estimator. The shrinkage intensities (outer right) are extremely low for standard
shrinkage and the negative bias of the Sancetta-estimator shows clearly for small numbers of training trials. Figure 5 (lower row) shows results for a too large time lag (b = 300). The performance
of the Sancetta-estimator strongly degrades as its shrinkage intensities get smaller, while our proposed estimator is robust to the choice of b, only for the smallest number of trials we observe a
small degradation in performance. Figure 6 (left) compares our bias-corrected estimator to the four
other approaches for 10 training trials: it significantly outperforms standard shrinkage and Sancetta
shrinkage for both the larger (b = 300, p ? 0.01) and the smaller time lag (b = 75, p ? 0.05).
7
time demand
bias?corr
0.9
1
**90%
**90%
0.9
1
**77.50%
**78.75%
0.9
1
51.25%
53.75%
0.9
*60%
**60%
0.8
0.8
0.8
0.8
0.7
0.7
0.7
0.7
b = 75
b = 300
0.6
*38.75%
**38.75%
0.6
**8.75%
**8.75%
0.5
0.5
0.6
0.7
0.8
0.9
sample covariance
0.6
**21.25%
**20%
0.5
0.5
0.6
0.7
0.8
0.9
0.6
0.5
0.5
47.50%
45.00%
0.6
0.7
standard shrinkage
0.8
CV
0.9
0.5
0.5
0.6
0.7
0.8
0.9
normalized runtime
subject?wise classification accuracies
1
120
100
80
60
40
20
0
SC Shr San BC
CV
Sancetta estimator
Figure 6: Subject-wise BCI classification accuracies for 10 training trials (left) and time demands
(right). ?? /? := significant at p ? 0.01 or p ? 0.05, respectively.
Analytic shrinkage procedures optimize only the mean squared error of the covariance matrix, while
cross-validation directly optimizes the classification performance. Yet, Figure 5 (middle) shows that
for small numbers of training trials our proposed estimator outperforms CV, although the difference
is not significant (see Fig. 6). For larger numbers of training trials CV performs better. This shows
that the MSE is not a very good proxy for classification accuracies in the context of CSP: for optimal MSE, shrinkage intensities decrease with increasing number of observations. CV shrinkage
intensities instead stay on a constant level between 0.1 and 0.15. Figure 6 (right) shows that the
three shrinkage approaches (b = 300) have a huge performance advantage over cross-validation (10
folds/10 parameter candidates) with respect to runtime.
5
Discussion
Analytic Shrinkage estimators are highly useful tools for covariance matrix estimation in timecritical or computationally expensive applications: no time-consuming cross-validation procedure
is required. In addition, it has been observed that in some applications, cross-validation is not a
good predictor for out-of-sample performance [LG11, BKT+ 07]. Its speed and good performance
has made analytic shrinkage widely used: it is, for example, state-of-the-art in ERP experiments
[BLT+ 11]. While standard shrinkage assumes i.i.d. data, many real world data sets, for example
from video, audio, finance, biomedical engineering or energy systems clearly violate this assumption as strong autocorrelation is present. Intuitively this means that the information content per data
point becomes lower, and thus the covariance estimation problem becomes harder: the dimensionality remains unchanged but the effective number of samples available decreases. Thus stronger
regularization is required and standard analytic shrinkage [LW04] needs to be corrected.
Sancetta already moved the first step into this important direction by providing a consistent estimator
under i.i.d. violations [San08]. In this work we analysed finite sample sizes and showed that (i) even
apart from truncation bias ?which results from including a limited number of time lags? Sancetta?s
estimator is biased, (ii) this bias is only negligible if the autocorrelation decays fast compared to
the length of the time series and (iii) the Sancetta estimator is very sensitive to the choice of lag
parameter.
We proposed an alternative estimator which is (i) both consistent and ?apart from truncation bias?
unbiased and (ii) highly robust to the choice of lag parameter: In simulations on toy and real world
data we showed that the proposed estimator yields large improvements for small samples and/or suboptimal lag parameter. Even for optimal lag parameter there is a slight but significant improvement.
Analysing data from BCI motor imagery experiments we see that (i) the BCI data set possesses
significant autocorrelation, that (ii) this adversely affects CSP based on the sample covariance and
standard shrinkage (iii) this effect can be alleviated using our novel estimator, which is shown to
(iv) compare favorably to Sancetta?s estimator.
Acknowledgments
This research was also supported by the National Research Foundation grant (No. 2012-005741)
funded by the Korean government. We thank Johannes H?ohne, Sebastian Bach and Duncan Blythe
for valuable discussions and comments.
8
References
[And91] Donald WK Andrews. Heteroskedasticity and autocorrelation consistent covariance matrix estimation. Econometrica: Journal of the Econometric Society, pages 817?858, 1991.
[BKT+ 07] Benjamin Blankertz, Motoaki Kawanabe, Ryota Tomioka, Friederike Hohlefeld, Klaus-Robert
M?uller, and Vadim V Nikulin. Invariant common spatial patterns: Alleviating nonstationarities in
brain-computer interfacing. In Advances in Neural Information Processing Systems, pages 113?
120, 2007.
[BLT+ 11] Benjamin Blankertz, Steven Lemm, Matthias Treder, Stefan Haufe, and Klaus-Robert M?uller.
Single-trial analysis and classification of ERP components ? a tutorial. NeuroImage, 56(2):814?
825, 2011.
[BM13] Daniel Bartz and Klaus-Robert M?uller. Generalizing analytic shrinkage for arbitrary covariance
structures. In C.J.C. Burges, L. Bottou, M. Welling, Z. Ghahramani, and K.Q. Weinberger, editors,
Advances in Neural Information Processing Systems 26, pages 1869?1877. Curran Associates,
Inc., 2013.
[BS10] Zhidong Bai and Jack William Silverstein. Spectral Analysis of Large Dimensional Random Matrices. Springer Series in Statistics. Springer New York, 2010.
[BSH+ 10] Benjamin Blankertz, Claudia Sannelli, Sebastian Halder, Eva M Hammer, Andrea K?ubler, KlausRobert M?uller, Gabriel Curio, and Thorsten Dickhaus. Neurophysiological predictor of SMRbased BCI performance. Neuroimage, 51(4):1303?1309, 2010.
[BTL+ 08] Benjamin Blankertz, Ryota Tomioka, Steven Lemm, Motoaki Kawanabe, and Klaus-Robert
M?uller. Optimizing spatial filters for robust EEG single-trial analysis. Signal Processing Magazine, IEEE, 25(1):41?56, 2008.
[CWEH10] Yilun Chen, Ami Wiesel, Yonina C Eldar, and Alfred O Hero. Shrinkage algorithms for MMSE
covariance estimation. Signal Processing, IEEE Transactions on, 58(10):5016?5029, 2010.
[ER05] Alan Edelman and N. Raj Rao. Random matrix theory. Acta Numerica, 14:233?297, 2005.
[GLL+ 14] Alexandre Gramfort, Martin Luessi, Eric Larson, Denis A. Engemann, Daniel Strohmeier, Christian Brodbeck, Lauri Parkkonen, and Matti S. H?am?al?ainen. MNE software for processing MEG
and EEG data. NeuroImage, 86(0):446 ? 460, 2014.
[HTF08] Trevor Hastie, Robert Tibshirani, and Jerome Friedman. The Elements of Statistical Learning.
Springer, 2008.
[LG11] Fabien Lotte and Cuntai Guan. Regularizing common spatial patterns to improve BCI designs:
unified theory and new algorithms. Biomedical Engineering, IEEE Transactions on, 58(2):355?
362, 2011.
[LW04] Olivier Ledoit and Michael Wolf. A well-conditioned estimator for large-dimensional covariance
matrices. Journal of Multivariate Analysis, 88(2):365?411, 2004.
[LW12] Olivier Ledoit and Michael Wolf. Nonlinear shrinkage estimation of large-dimensional covariance
matrices. The Annals of Statistics, 40(2):1024?1060, 2012.
[MP67] Vladimir A. Mar?cenko and Leonid A. Pastur. Distribution of eigenvalues for some sets of random
matrices. Mathematics of the USSR-Sbornik, 1(4):457, 1967.
[San08] Alessio Sancetta. Sample covariance shrinkage for high dimensional dependent data. Journal of
Multivariate Analysis, 99(5):949?967, May 2008.
[SBMK13] Wojciech Samek, Duncan Blythe, Klaus-Robert M?uller, and Motoaki Kawanabe. Robust spatial
filtering with beta divergence. In L. Bottou, C.J.C. Burges, M. Welling, Z. Ghahramani, and K.Q.
Weinberger, editors, Advances in Neural Information Processing Systems 26, pages 1007?1015.
2013.
[SS05] Juliane Sch?afer and Korbinian Strimmer. A shrinkage approach to large-scale covariance matrix
estimation and implications for functional genomics. Statistical Applications in Genetics and
Molecular Biology, 4(1):1175?1189, 2005.
[Ste56] Charles Stein. Inadmissibility of the usual estimator for the mean of a multivariate normal distribution. In Proc. 3rd Berkeley Sympos. Math. Statist. Probability, volume 1, pages 197?206,
1956.
[TZ84] H. Jean Thi?ebaux and Francis W. Zwiers. The interpretation and estimation of effective sample
size. Journal of Climate and Applied Meteorology, 23(5):800?811, 1984.
9
| 5399 |@word trial:17 middle:6 inversion:1 wiesel:1 stronger:3 norm:1 simulation:10 covariance:32 tr:4 harder:2 prial:4 moment:7 reduction:2 bai:1 series:4 daniel:4 bc:18 mmse:1 outperforms:5 comparing:1 analysed:1 yet:3 visible:2 j1:1 analytic:9 motor:4 christian:1 drop:1 ainen:1 resampling:2 generative:2 yi1:1 math:1 contribute:2 readability:1 denis:1 five:1 become:2 beta:1 edelman:1 consists:1 autocorrelation:23 expected:2 andrea:1 brain:4 bsh:2 increasing:8 becomes:4 minimizes:1 substantially:1 supplemental:1 unified:1 berkeley:1 finance:1 runtime:2 k2:2 grant:1 eigendirection:2 negligible:3 engineering:2 limit:3 consequence:3 quadruple:1 juliane:1 acta:1 mne:1 gll:2 limited:2 averaged:3 acknowledgment:1 practice:2 differs:1 procedure:3 thi:1 significantly:1 alleviated:1 donald:1 get:3 unfeasible:1 cannot:1 selection:2 close:1 context:1 influence:1 restriction:2 equivalent:1 optimize:1 straightforward:1 cluttered:1 convex:1 independently:1 estimator:74 rule:1 his:1 population:9 annals:1 target:2 alleviating:1 magazine:1 olivier:2 us:1 curran:1 associate:1 element:1 expensive:2 std:1 observed:1 steven:2 capture:2 calculate:1 eva:1 ebaux:1 decrease:4 movement:1 rescaled:1 valuable:1 yk:1 substantial:1 benjamin:4 econometrica:1 trained:1 eric:1 autocorrelated:6 bno:1 distinct:1 fast:2 effective:4 alessio:1 sc:1 klaus:7 sympos:1 jean:1 lag:23 larger:8 heuristic:1 widely:1 otherwise:1 samek:1 bci:6 statistic:3 cov:13 ledoit:5 analyse:2 online:1 sequence:2 eigenvalue:11 advantage:1 matthias:1 propose:3 nikulin:1 tu:4 poorly:1 degenerate:1 moved:1 pronounced:3 eigenbasis:5 inadmissibility:1 rotated:1 andrew:1 ac:24 pose:2 measured:1 ij:26 eq:8 p2:5 strong:3 predicted:1 quantify:1 differ:1 direction:2 motoaki:3 drawback:1 hammer:1 filter:2 material:1 government:1 behaviour:2 adjusted:1 hold:2 practically:1 sufficiently:1 normal:2 a2:3 smallest:1 estimation:12 proc:1 outperformed:1 sensitive:4 tool:1 uller:6 stefan:1 clearly:3 interfacing:2 gaussian:3 csp:3 fulfill:1 ck:3 shrinkage:69 eks:2 derived:1 xit:15 improvement:5 ubler:1 mainly:1 am:1 mueller:1 dependent:1 typically:1 relation:1 i1:1 germany:2 issue:1 classification:5 reanalyzed:1 eldar:1 ussr:1 spatial:5 art:1 gramfort:1 having:1 yonina:1 identical:1 biology:1 report:1 national:1 divergence:1 william:1 friedman:1 huge:1 highly:4 introduces:1 violation:3 strimmer:1 implication:1 accurate:1 necessary:1 korea:2 indexed:1 iv:3 theoretical:5 minimal:1 rao:1 ar:7 cost:1 deviation:2 entry:4 predictor:2 eigendecompositions:1 too:1 dependency:4 xpi1:2 adaptively:1 overestimated:1 stay:2 systematic:1 yl:1 corrects:1 michael:2 squared:4 imagery:4 csh:1 worse:1 adversely:1 wojciech:1 yp:1 toy:2 account:2 de:2 blow:1 b2:1 wk:1 coefficient:2 inc:1 multiplicative:1 optimistic:1 francis:1 accuracy:7 variance:30 yield:3 silverstein:1 marginally:1 acc:2 suffers:1 sebastian:2 trevor:1 sixth:1 underestimate:1 shr:3 energy:1 frequency:1 e2:4 proof:1 degeneracy:1 popular:1 friederike:1 dimensionality:10 alexandre:1 higher:3 response:1 done:1 shrink:1 strongly:3 mar:1 biomedical:2 correlation:2 jerome:1 hand:2 replacing:1 nonlinear:1 autocorrelations:1 quality:1 grows:1 effect:5 contain:2 unbiased:5 normalized:1 hence:2 regularization:3 i2:2 climate:1 claudia:1 larson:1 performs:3 cp:2 interface:1 wise:3 jack:1 novel:1 charles:1 superior:1 common:3 behaves:2 blt:3 functional:1 qp:4 volume:1 imagined:1 extend:1 tail:1 slight:1 interpretation:1 measurement:1 significant:4 cv:8 rd:1 trivially:1 consistency:1 mathematics:1 funded:1 afer:1 heteroskedasticity:1 j:1 uncorrelatedness:1 multivariate:4 recent:2 showed:2 optimizing:1 optimizes:1 apart:2 raj:1 pastur:1 arbitrarily:1 muller:1 additional:1 signal:4 ii:7 xju:1 violate:1 alan:1 faster:1 calculation:1 cross:10 long:1 bach:1 nonstationarities:1 molecular:1 a1:4 variant:1 xjt:13 denominator:3 expectation:3 kernel:4 invert:1 addition:2 underestimated:1 sch:1 biased:9 vadim:1 posse:2 comment:1 subject:8 contrary:1 integer:1 presence:2 iii:6 enough:2 blythe:2 haufe:1 xj:6 affect:1 lauri:1 lotte:1 hastie:1 restrict:1 suboptimal:2 audio:1 reduce:1 bartlett:1 york:1 cause:1 remark:2 ignored:2 tij:3 useful:1 clear:1 gabriel:1 johannes:1 stein:1 band:1 statist:1 meteorology:1 tth:1 reduced:1 generate:1 supplied:1 exist:1 restricts:1 impr:1 percentage:1 tutorial:1 fulfilled:1 per:9 yilun:1 tibshirani:1 alfred:1 hyperparameter:1 shall:1 numerica:1 key:1 yi8:2 four:2 yit:1 drawn:1 clarity:1 erp:2 btl:2 econometric:1 excludes:1 counteract:1 run:1 letter:1 fourth:1 eigendirections:2 duncan:2 coeff:4 bound:2 guaranteed:1 fold:1 treder:1 oracle:3 activity:1 software:1 lemm:2 dominated:1 aspect:1 speed:1 extremely:1 martin:1 hohlefeld:1 department:1 combination:1 smaller:7 describes:1 slightly:3 appealing:1 intuitively:1 invariant:1 sij:28 thorsten:1 taken:1 sannelli:1 computationally:1 remains:2 needed:1 hero:1 serf:1 yj1:1 available:2 multiplied:1 observe:1 kawanabe:3 spectral:1 alternative:3 robustness:1 weinberger:2 rp:1 denotes:2 assumes:1 a4:3 bkt:2 neglect:1 k1:2 especially:1 ghahramani:2 society:1 unchanged:1 already:1 quantity:1 strategy:1 degrades:1 dependence:3 usual:1 diagonal:2 exhibit:1 thank:1 berlin:6 outer:4 discriminant:1 parkkonen:1 meg:1 length:1 index:2 providing:1 vladimir:1 innovation:1 difficult:1 korean:1 robert:8 statement:1 favorably:1 ryota:2 trace:1 negative:1 luessi:1 design:1 perform:1 upper:3 observation:8 finite:5 truncated:1 extended:1 incorporated:1 precise:1 looking:1 arbitrary:2 intensity:15 required:3 kl:2 optimized:2 korbinian:1 pop:4 beyond:1 pattern:3 eighth:3 including:2 video:1 sbornik:1 rely:1 blankertz:4 improve:4 genomics:1 val:2 relative:3 loss:1 interesting:1 limitation:1 filtering:1 var:24 ingredient:1 validation:6 eigendecomposition:1 foundation:1 consistent:8 xp:3 proxy:1 editor:2 heavy:1 row:7 genetics:1 summary:1 supported:1 truncation:4 bias:33 weaker:1 understand:2 burges:2 taking:1 absolute:1 default:1 dimension:4 world:6 calculated:2 autoregressive:1 cenko:1 made:1 san:13 far:1 welling:2 yi4:4 transaction:2 approximate:1 assumed:1 consuming:1 xi:7 spectrum:1 channel:2 matti:1 robust:6 ignoring:2 xjs:1 eeg:4 klausrobert:1 mse:2 bottou:2 bartz:3 complex:1 constructing:1 main:2 yi2:2 spread:1 motivation:1 whole:1 hyperparameters:1 n2:1 complementary:1 fig:1 downwards:2 tomioka:2 neuroimage:3 candidate:1 guan:1 formula:2 theorem:4 xt:2 jt:1 showing:2 decay:2 a3:3 essential:1 exists:2 curio:1 adding:1 corr:5 illustrates:1 conditioned:1 demand:2 chen:1 easier:1 generalizing:1 neurophysiological:1 xiu:1 springer:3 wolf:5 extracted:1 identity:1 presentation:1 towards:1 leonid:1 content:1 analysing:1 included:1 corrected:5 ami:1 averaging:1 principal:2 degradation:1 seoul:1 dickhaus:1 regularizing:1 |
4,859 | 54 | 860
A METHOD FOR THE DESIGN OF STABLE LATERAL INHIBITION
NETWORKS THAT IS ROBUST IN THE PRESENCE
OF CIRCUIT PARASITICS
J.L. WYATT, Jr and D.L. STANDLEY
Department of Electrical Engineering and Computer Science
Massachusetts Institute of Technology
Cambridge, Massachusetts 02139
ABSTRACT
In the analog VLSI implementation of neural systems, it is
sometimes convenient to build lateral inhibition networks by using
a locally connected on-chip resistive grid. A serious problem
of unwanted spontaneous oscillation often arises with these
circuits and renders them unusable in practice. This paper reports
a design approach that guarantees such a system will be stable,
even though the values of designed elements and parasitic elements
in the resistive grid may be unknown. The method is based on a
rigorous, somewhat novel mathematical analysis using Tellegen's
theorem and the idea of Popov multipliers from control theory. It
is thoroughly practical because the criteria are local in the sense
that no overall analysis of the interconnected system is required,
empirical in the sense that they involve only measurable frequency
response data on the individual cells, and robust in the sense that
unmodelled parasitic resistances and capacitances in the interconnection network cannot affect the analysis.
I.
INTRODUCTION
The term "lateral inhibition" first arose in neurophysiology to
describe a common form of neural circuitry in which the output of
each neuron in some population is used to inhibit the response of
each of its neighbors. Perhaps the best understood example is the
horizontal cell layer in the vertebrate retina, in which lateral
inhibition simultaneously enhances intensity edges and acts as an
automatic lain control to extend the dynamic range of the retina
as a whole. The principle has been used in the design of artificial
neural system algorithms by Kohonen 2 and others and in the electronic
design of neural chips by Carver Mead et. al. 3 ,4.
In the VLSI implementation of neural systems, it is convenient
to build lateral inhibition networks by using a locally connected
on-chip resistive grid. Linear resistors fabricated in, e.g.,
polysilicon, yield a very compact realization, and nonlinear
resistive grids, made from MOS transistors, have been found useful
for image segmentation. 4 ,5 Networks of this type can be divided into
two classes: feedback systems and feedforward-only systems. In the
feedforward case one set of amplifiers imposes signal voltages or
? American Institute of Physics 1988
861
currents on the grid and another set reads out the resulting response
for subsequent processing, while the same amplifiers both "write" to
the grid and "read" from it in a feedback arrangement. Feedforward
networks of this type are inherently stable, but feedback networks
need not be.
A practical example is one of Carver Meadls retina chips3 that
achieves edge enhancement by means of lateral inhibition through a
resistive grid. Figure 1 shows a single cell in a continuous-time
version of this chip. Note that the capacitor voltage is affected
both by the local light intensity incident on that cell and by the
capacitor voltages on neighboring cells of identical design. Any
cell drives its neighbors, which drive both their distant neighbors
and the original cell in turn. Thus the necessary ingredients for
instability--active elements and signal feedback--are both present
in this system, and in fact the continuous-time version oscillates
so badly that the original design is scarcely usable in practice
with the lateral inhibition paths enabled. 6 Such oscillations can
I
incident
light
v
out
Figure 1. This photoreceptor and signal processor Circuit, using two
MOS transconductance amplifiers, realizes lateral inhibition by
communicating with similar units through a resistive grid.
readily occur in any resistive grid circuit with active elements and
feedback,even when each individual cell is quite stable. Analysis
of the conditions of instability by straightforward methods appears
hopeless, since any repeated array contains many cells, each of
which influences many others directly or indirectly and is influenced
by them in turn, so that the number of simultaneously active feedback loops is enormous.
This paper reports a practical design approach that rigorously
guarantees such a system will be stable. The very simplest version
of the idea is intuitively obvious: design each individual cell so
that, although internally active, it acts like a passive system as
seen from the resistive grid. In circuit theory language, the
design goal here is that each cellis output impedance should be a
positive-real? function. This is sometimes not too difficult in
practice; we will show that the original network in Fig. 1 satisfies
this condition in the absence of certain parasitic elements. More
important, perhaps, it is a condition one can verify experimentally
862
by frequency-response measurements.
It is physically apparent that a collection of cells that
appear passive at their terminals will form a stable system when
interconnected through a passive medium such as a resistive grid.
The research contributions, reported here in summary form, are
i) a demonstration that this passivity or positive-real condition
is much stronger than we actually need and that weaker conditions,
more easily achieved in practice, suffice to guarantee stability of
the linear network model, and ii) an extension of i) to the nonlinear
domain that furthermore rules out large-signal oscillations under
certain conditions.
II.
FIRST-ORDER LINEAR ANALYSIS OF A SINGLE CELL
We begin with a linear analysis of an elementary model for the
circuit in Fig. 1. For an initial approximation to the output
admittance of the cell we simplify the topology (without loss of
relevant information) and use a naive'model for the transconductance
amplifiers, as shown in Fig. 2.
e
+
Figure 2. Simplified network topology and transconductance amplifier
model for the circuit in Fig. 1. The capacitor in Fig. 1 has been
absorbed into CO2 ?
Straightforward calculations show that the output admittance is
given by
yes)
(1)
This is a positive-real, i.e., passive, admittance since it can always
be realized by a network of the form shown in Fig. 3, where
= (gm2+
-1 -1
-1
Ro2 ) , R2= (gmlgm2Rol)
, and L = COI/gmlgm2?
Although the original circuit contains no inductors, the
realization has both capacitors and inductors and thus is capable
of damped oscillations. Nonetheless, i f the transamp model in
Fig. 2 were perfectly accurate, no network created by interconnecting
such cells through a resistive grid (with parasitic capacitances)
could exhibit sustained oscillations. For element values that may
be typical in practice, the model in Fig. 3 has a lightly damped
resonance around I KHz with a Q ~ 10. This disturbingly high Q
suggests that the cell will be highly sensitive to parasitic elements
not captured by the simple models in Fig. 2. Our preliminary
Rl
863
yes)
Figure 3. Passive network realization of the output admittance (eq.
(1) of the circuit in Fig. 2.
analysis of a much more complex model extracted from a physical
circuit layout created in Carver Mead's laboratory indicates that
the output impedance will not be passive for all values of the transamp bias currents. But a definite explanation of the instability
awaits a more careful circuit modelling effort and perhaps the design
of an on-chip impedance measuring instrument.
III.
POSITIVE-REAL FUNCTIONS, e-POSITlVE FUNCTIONS, AND
STABILITY OF LINEAR NETWORK MODELS
In the following discussion s = cr+jw is a complex variable,
H(s) is a rational function (ratio of polynomials) in s with real
coefficients, and we assume for simplicity that H(s) has no pure
imaginary poles. The term closed right halE plane refers to the set
of complex numbers s with Re{s} > o.
Def. I
The function H(s) is said to be positive-real if a) it has no
poles in the right half plane and b) Re{H(jw)} ~ 0 for all w.
If we know at the outset that H(s) has no right half plane poles,
then Def. I reduces to a simple graphical criterion: H1s} is positivereal if and only if the Nyquist diagram of H(s) (i.e. the plot of
H(jW) for w ~ 0, as in Fig. 4) lies entirely in the closed right half
plane.
Note that positive-real functions are necessarily stable since
they have no right half plane poles, but stable functions are not
necessarily positive-real, as Example 1 will show.
A deep link between positive real functions, physical networks
and passivity is established by the classical result 7 in linear
circuit theory which states that H(s) is positive-real if and only if
it is possible to synthesize a 2-terminal network of positive linear
resistors, capacitors, inductors and ideal transformers that has H(s)
as its driving-point impedance or admittance.
864
Oef. 2
The function H(s) is said to be a-positive for a particular value
of e(e ~ 0, e ~ ~), if a) H{s) has no poles in the right half plane,
and b) the Nyquist plot of H(s) lies strictly to the right of the
straight line passing through the origin at an angle a to the real
positive axis.
Note that every a-positive function is stable and any function
that is e-positive with e = ~/2 is necessarily positive-real.
I {G(jw)}
m
Re{G(jw) }
Figure 4. Nyquist diagram for a fUnction that is a-positive but
not positive-real.
Example 1
The function
G (s)
=
(s+l) (s+40)
(s+5) (s+6) (s+7)
(2)
is a-positive (for any e between about 18? and 68?) and stable, but it
is not positive-real since its Nyquist diagram, shown in Fig. 4,
crosses into the left half plane.
The importance of e-positive functions lies in the following
observations: 1) an interconnection of passive linear resistors and
capacitors and cells with stable linear impedances can result in an
unstable network, b) such an instability cannot result if the
impedances are also positive-real, c) a-positive impedances form a
larger class than positive-real ones and hence a-positivity is a less
demanding synthesis goal, and d) Theorem 1 below shows that such an
instability cannot result if the impedances are a-positive, even if
they are not positive-real.
Theorem 1
Consider a linear network of arbitrary topology, consisting of
any number of passive 2-terminal resistors and capacitors of arbitrary
value driven by any number of active cells. If the output impedances
865
'II" ,
of all the active cells are a-positive for some common a, 0<a 22
then the network is stable.
The proof of Theorem 1 relies on Lemma 1 below.
Lemma 1
If H(s) is a-positive for some fixed a, then for all So in the
closed first quadrant of the complex plane, H(so) lies strictly to
the right of the straight line passing through the origin at an angle
a to the real positive axis, i.e., Re{so} ~ 0 and Im{so} ~ 0 ~
a-'II" < L H (so) < a.
Proof of Lemma 1 (Outline)
Let d be the function that assigns to each s in the closed right
half plane the perpendicular distance des) from H(s) to the line
defined in Def. 2. Note that des) is harmonic in the closed right
half plane, since H is analytic there. It then follows, by application
of the maximum modulus principle8 for harmonic functions, that d takes
its minimum value on the boundary of its domain, which is the
imaginary axis. This establishes Lemma 1.
Proof of Theorem 1 (OUtline)
The network is unstable or marginally stable if and only if it
has a natural frequency in the closed right half plane, and So is a
natural frequency if and only if the network equations have a nonzero
solution at so. Let {Ik} denote the complex branch currents Of such
a solution. By Tellegen I s theorern9 the sum of the complex powers
absorbed by the circuit elements must vanish at such a solution, i.e.,
~
IIk12/s0Ck +
capac~tances
L
cell
terminal pairs
(3)
where the second term is deleted in the special case so=O, since the
complex power into capacitors vanishes at so=O.
If the network has a natural frequency in the closed right half
plane, it must have one in the closed first quadrant since natural
frequencies are either real or else occur in complex conjugate pairs.
But (3) cannot be satisfied for any So in the closed first quadrant,
as we can see by dividing both sides of (3) by
IIkI2, where the
k
sum is taken over all network branches. After this division, (3)
asserts that zero is a convex combination of terms of the form Rk,
terms of the form (CkSo)-I, and terms of the form Zk(So).
Visualize where these terms lie in the complex plane: the first set lies
on the real positive axis, the second set lies in the closed 4-th
~adrant since So lies in the closed 1st quadrant by assumption, and
the third set lies to the right of a line passing through the origin
at an angle a by Lemma 1. Thus all these terms lie strictly to the
right of this line, which implies that no convex combination of them
can equal zero. Hence the network is stable!
866
IV.
STABILITY RESULT FOR NETWORKS WITH NONLINEAR
RESISTORS AND CAPACITORS
The previous result for linear networks can afford some limited
insight into the behavior of nonlinear networks. First the nonlinear
equations are linearized about an equilibrium point and Theorem 1 is
applied to the linear model. If the linearized model is stable, then
the equilibrium point of the original nonlinear network is locally
stable, i.e., the network will return to that equilibrium point if
the initial condition is sufficiently near it. But the result in this
section, in contrast, applies to the full nonlinear circuit model and
allows one to conclude that in certain circumstances the network
cannot oscillate even if the initial state is arbitrarily far from
the equilibrium point.
Def. 3
A function H(s) as described in Section III is said tc satisfy
the Popov criterion lO if there exists a real number r>O such that
Re{(l+jwr) H(jw)} ~ 0 for all w.
Note that positive real functions satisfy the Popov criterion
with r=O. And the reader can easily verify that G(s) in Exam~le I
satisfies the Popov criterion for a range of values of r. The important
effect of the term (l+jwr) in Def. 3 is to rotate the Nyquist plot
counterclockwise by progressively greater amounts up to 90? as w
increases.
Theorem 2
Consider a network consisting of nonlinear 2-terminal resistors
and capacitors, and cells with linear output impedances ~(s). Suppose
i) the resistor curves are characterized by continuously
diffefentiable functions i k = gk(vk ) where gk(O) = 0 and
o < gk(vk ) < G < 00 for all values of k and vk'
ii) the capacitors are characterized by i k = Ck(Vk)~k with
< CI < Ck(v k ) < C2 < 00 for all values of k and vk'
o
iii) the impedances Zk(s) have no poles in the closed right
half plane and all satisfy the Popov criterion for some common
value of r.
If these conditions are satisfied, then the network is stable in the
sense that, for any initial condition,
f oo(
o
I
all branches
i~(t) )
dt
<
00
?
The proof, based on Tellegen's theorem, is rather involved.
will be omitted here and will appear elsewhere.
(4)
It
867
ACKNOWLEDGEMENT
We sincerely thank Professor Carver Mead of Cal Tech for
enthusiastically supporting this work and for making it possible for
us to present an early report on it in this conference proceedings.
This work was supportedJ::? Defense Advanced Research Projects Agency
(DoD), through the Office of Naval Research under ARPA Order No.
3872, Contract No. N00014-80-C-0622 and Defense Advanced Research
Projects Agency (DARPA) Contract No. N00014-87-R-0825.
REFERENCES
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
F.S. Werblin, "The Control of Sensitivity on the Retina,"
Scientific American, Vol. 228, no. 1, Jan. 1983, pp. 70-79.
T. Kohonen, Self-Organization and Associative Memory, (vol. 8 in
the Springer Series in Information Sciences), Springer Verlag,
New York, 1984.
M.A. Sivilotti, M.A. Mahowald, and C.A. Mead, "Real Time Visual
Computations Using Analog CMOS processing Arrays," Advanced
Research in VLSI - Proceedings of the 1987 Stanford Conference,
P. Losleben, ed., MIT Press, 1987, pp. 295-312.
C.A. Mead, Analog VLSI and Neural Systems, Addison-Wesley, to
appear in 1988.
J. Hutchinson, C. Koch, J. Luo and C. Mead, "Computing Motion
Using Analog and Binary Resistive Networks," submitted to IEEE
Transactions on Computers, August 1987.
M. Mahowald, personal communication.
B.D.O. Anderson and S. Vongpanitlerd, Network Analysis and
synthesis - A Modern Systems Theory Approach, Prentice-Hall,
Englewood Cliffs, NJ., 1973.
L.V. Ahlfors, Complex Analysis, McGraw-Hill, New York, 1966,
p. 164.
P. penfield, Jr., R. Spence, and S. Duinker, Tellegen's Theorem
and Electrical Networks, MIT Press, Cambridge, MA,1970.
M. Vidyasagar, Nonlinear Systems Analysis, Prentice-Hall,
Englewood Cliffs, NJ, 1970, pp. 211-217.
| 54 |@word neurophysiology:1 version:3 polynomial:1 stronger:1 linearized:2 initial:4 contains:2 series:1 imaginary:2 current:3 luo:1 must:2 readily:1 distant:1 subsequent:1 analytic:1 designed:1 plot:3 progressively:1 half:11 plane:14 mathematical:1 c2:1 ik:1 resistive:11 sustained:1 behavior:1 terminal:5 vertebrate:1 begin:1 project:2 suffice:1 circuit:14 medium:1 sivilotti:1 fabricated:1 nj:2 guarantee:3 every:1 act:2 unwanted:1 oscillates:1 control:3 unit:1 internally:1 appear:3 werblin:1 positive:30 engineering:1 local:2 understood:1 capac:1 mead:6 cliff:2 path:1 suggests:1 limited:1 range:2 perpendicular:1 practical:3 spence:1 practice:5 definite:1 jan:1 empirical:1 convenient:2 outset:1 refers:1 quadrant:4 cannot:5 cal:1 prentice:2 influence:1 instability:5 transformer:1 measurable:1 straightforward:2 layout:1 convex:2 simplicity:1 assigns:1 pure:1 communicating:1 rule:1 insight:1 array:2 enabled:1 stability:3 population:1 spontaneous:1 suppose:1 origin:3 element:8 synthesize:1 electrical:2 connected:2 inhibit:1 vanishes:1 agency:2 co2:1 rigorously:1 dynamic:1 personal:1 oef:1 division:1 easily:2 darpa:1 chip:5 describe:1 artificial:1 transamp:2 quite:1 apparent:1 larger:1 stanford:1 interconnection:2 associative:1 transistor:1 interconnected:2 kohonen:2 neighboring:1 loop:1 realization:3 relevant:1 ro2:1 asserts:1 enhancement:1 cmos:1 oo:1 exam:1 eq:1 dividing:1 implies:1 coi:1 preliminary:1 elementary:1 im:1 extension:1 strictly:3 around:1 sufficiently:1 koch:1 hall:2 equilibrium:4 mo:2 visualize:1 circuitry:1 driving:1 achieves:1 early:1 omitted:1 realizes:1 sensitive:1 establishes:1 mit:2 always:1 ck:2 arose:1 rather:1 cr:1 voltage:3 passivity:2 office:1 naval:1 vk:5 modelling:1 indicates:1 tech:1 contrast:1 rigorous:1 sense:4 vlsi:4 overall:1 resonance:1 special:1 equal:1 identical:1 report:3 others:2 simplify:1 serious:1 retina:4 modern:1 penfield:1 simultaneously:2 individual:3 consisting:2 amplifier:5 organization:1 englewood:2 highly:1 light:2 damped:2 accurate:1 edge:2 capable:1 popov:5 necessary:1 carver:4 iv:1 re:5 arpa:1 wyatt:1 measuring:1 mahowald:2 pole:6 dod:1 too:1 reported:1 scarcely:1 hutchinson:1 thoroughly:1 st:1 sensitivity:1 contract:2 physic:1 synthesis:2 continuously:1 satisfied:2 positivity:1 american:2 usable:1 return:1 de:2 coefficient:1 satisfy:3 closed:12 contribution:1 yield:1 yes:2 marginally:1 drive:2 straight:2 processor:1 submitted:1 influenced:1 ed:1 nonetheless:1 frequency:6 involved:1 pp:3 obvious:1 proof:4 rational:1 massachusetts:2 segmentation:1 actually:1 appears:1 wesley:1 dt:1 response:4 jw:6 though:1 anderson:1 furthermore:1 horizontal:1 nonlinear:9 perhaps:3 scientific:1 modulus:1 effect:1 verify:2 multiplier:1 hence:2 read:2 laboratory:1 nonzero:1 self:1 standley:1 criterion:6 hill:1 outline:2 motion:1 passive:8 image:1 harmonic:2 novel:1 common:3 rl:1 physical:2 khz:1 analog:4 extend:1 measurement:1 cambridge:2 automatic:1 grid:12 language:1 stable:17 inhibition:8 driven:1 certain:3 n00014:2 verlag:1 binary:1 arbitrarily:1 seen:1 captured:1 minimum:1 somewhat:1 greater:1 signal:4 ii:5 branch:3 full:1 reduces:1 polysilicon:1 calculation:1 cross:1 characterized:2 divided:1 gm2:1 circumstance:1 admittance:5 physically:1 sometimes:2 achieved:1 cell:20 diagram:3 else:1 counterclockwise:1 capacitor:11 near:1 presence:1 ideal:1 feedforward:3 iii:3 affect:1 topology:3 perfectly:1 idea:2 defense:2 effort:1 nyquist:5 render:1 resistance:1 passing:3 afford:1 york:2 oscillate:1 deep:1 useful:1 involve:1 amount:1 locally:3 simplest:1 write:1 vol:2 affected:1 enormous:1 deleted:1 lain:1 sum:2 angle:3 reader:1 electronic:1 oscillation:5 entirely:1 layer:1 def:5 inductor:3 badly:1 occur:2 lightly:1 transconductance:3 department:1 combination:2 conjugate:1 jr:2 making:1 intuitively:1 taken:1 equation:2 sincerely:1 turn:2 know:1 addison:1 instrument:1 awaits:1 indirectly:1 original:5 graphical:1 build:2 classical:1 capacitance:2 arrangement:1 realized:1 said:3 enhances:1 exhibit:1 distance:1 link:1 thank:1 lateral:8 unstable:2 ratio:1 demonstration:1 difficult:1 gk:3 design:10 implementation:2 unknown:1 neuron:1 observation:1 supporting:1 communication:1 arbitrary:2 august:1 intensity:2 pair:2 required:1 established:1 below:2 memory:1 explanation:1 power:2 tances:1 demanding:1 natural:4 vidyasagar:1 advanced:3 technology:1 axis:4 created:2 naive:1 acknowledgement:1 loss:1 ingredient:1 incident:2 imposes:1 principle:1 lo:1 hopeless:1 summary:1 elsewhere:1 bias:1 weaker:1 side:1 institute:2 neighbor:3 feedback:6 boundary:1 curve:1 made:1 collection:1 simplified:1 far:1 transaction:1 compact:1 mcgraw:1 active:6 conclude:1 h1s:1 continuous:2 losleben:1 impedance:11 zk:2 robust:2 inherently:1 complex:10 necessarily:3 domain:2 whole:1 repeated:1 fig:12 interconnecting:1 resistor:7 lie:10 vanish:1 third:1 unmodelled:1 theorem:9 rk:1 unusable:1 hale:1 r2:1 exists:1 importance:1 ci:1 tc:1 absorbed:2 visual:1 applies:1 springer:2 satisfies:2 relies:1 extracted:1 ma:1 goal:2 careful:1 absence:1 professor:1 experimentally:1 typical:1 lemma:5 photoreceptor:1 parasitic:5 arises:1 rotate:1 |
4,860 | 540 | Learning to Segment Images
Using Dynamic Feature Binding
Michael C. Moser
Dept. of Compo Science &
Inst. of Cognitive Science
University of Colorado
Boulder, CO 80309-0430
Richard S. Zemel
Dept. of Compo Science
University of Toronto
Toronto, Ontario
Canada M5S lA4
Marlene Behrmann
Dept. of Psychology &
Faculty of Medicine
University of Toronto
Toronto, Ontario
Canada M5S lAl
Abstract
Despite the fact that complex visual scenes contain multiple, overlapping
objects, people perform object recognition with ease and accuracy. One
operation that facilitates recognition is an early segmentation process in
which features of objects are grouped and labeled according to which object they belong. Current computational systems that perform this operation are based on predefined grouping heuristics. We describe a system
called MAGIC that learn. how to group features based on a set of presegmented examples. In many cases, MAGIC discovers grouping heuristics
similar to those previously proposed, but it also has the capability of finding nonintuitive structural regularities in images. Grouping is performed
by a relaxation network that aUempts to dynamically bind related features. Features transmit a complex-valued signal (amplitude and phase)
to one another; binding can thus be represented by phase locking related
features. MAGIC'S training procedure is a generalization of recurrent back
propagation to complex-valued units.
When a visual image contains multiple, overlapping objects, recognition is difficult
because features in the image are not grouped according to which object they belong.
Without the capability to form such groupings, it would be necessary to undergo a
massive search through all subsets of image features. For this reason, most machine
vision recognition systems include a component that performs feature grouping or
image .egmentation (e.g., Guzman, 1968; Lowe, 1985; Marr, 1982).
436
Learning to Segment Images Using Dynamic Feature Binding
A multitude of heuristics have been proposed for segmenting images. Gestalt psychologists have explored how people group elements of a display and have suggested
a range of grouping principles that govern human perception (Rock &z: Palmer, 1990).
Computer vision researchers have studied the problem from a more computational perspective. They have investigated methods of grouping elements of an image
based on nonaccidental regularitie..-feature combinations that are unlikely to occur
by chance when several objects are juxtaposed, and are thus indicative of a single
object (Kanade, 1981; Lowe &z: Binford, 1982).
In these earlier approaches, the researchers have hypothesized a set of grouping
heuristics and then tested their psychological validity or computational utility. In
our work, we have taken an adaptive approach to the problem of image segmentation in which a system learns how to group features based on a set of examples.
We call the system MAGIC, an acronym for multiple-object !daptive grouping of
image ~omponents. In many cases MAGIC discovers grouping heuristics similar to
those proposed in earlier work, but it also has the capability offinding nonintuitive
structural regularities in images.
is trained on a set of presegmented images containing multiple objects. By
"presegmented," we mean that each image feature is labeled as to which object it
belongs. MAGIC learns to detect configurations of the image features that have a
consistent labeling in relation to one another across the training examples. Identifying these configurations allows MAGIC to then label features in novel, unsegmented
images in a manner consistent with the training examples.
MAGIC
1
REPRESENTING FEATURE LABELINGS
Before describing MAGIC, we must first discuss a representation that allows for
the labeling of features. Von der Malsburg (1981), von der Malsburg &z: Schneider
(1986), Gray et al. (1989), and Eckhorn et al. (1988), among others, have suggested
a biologically plausible mechanism of labeling through temporal correlations among
neural signals, either the relative timing of neuronal spikes or the synchronization of
oscillatory activities in the nervous system. The key idea here is that each processing
unit conveys not just an activation value-average firing frequency in neural termsbut also a second, independent value which represents the relative phcue of firing.
The dynamic grouping or binding of a set of features is accomplished by aligning
the phases of the features. Recent work (Goebel, 1991; Hummel &z: Biederman, in
press) has used this notion of dynamic binding for grouping image features, but has
been based on relatively simple, predetermined grouping heuristics.
2
THE DOMAIN
Our initial work has been conducted in the domain of two-dimensional geometric
contours, including rectangles, diamonds, crosses, triangles, hexagons, and octagons. The contours are constructed from four primitive feature types-oriented
line segments at 0?, 45?, 90?, and 135?-and are laid out on a 15 X 20 grid. At
each location on the grid are units, called feature unib, that detect each of the four
primitive feature types. In our present experiments, images contain two contours.
Contours are not permitted to overlap in their activation of the same feature unit.
437
438
Mozer, Zemel , and Behrmann
hidden
layer
_ _r
Figure 1: The architedure of MAGIC. The lower layer contains the feature units; the
upper layer contains the hidden units. Each layer is arranged in a spatiotopic array
with a number of different feature types at each position in the array. Each plane in
the feature layer corresponds to a different feature type. The grayed hidden units
are reciprocally conneded to all features in the corresponding grayed region of the
feature layer. The lines between layers represent projections in both directions.
3
THE ARCHITECTURE
The input to MAGIC is a paUern of activity over the feature units indicating which
features are present in an image. The initial phases ofthe units are random. MAGIC'S
task is to assign appropriate phase values to the units. Thus, the network performs
a type of paUern completion.
The network architedure consists of two layers of units, as shown in Figure 1. The
lower (input) layer contains the feature units, arranged in spatiotopic arrays with
one array per feature type. The upper layer contains hidden units that help to align
the phases of the feature units; their response properties are determined by training.
Each hidden unit is reciprocally conneded to the units in a local spatial region of
all feature arrays. We refer to this region as a patch; in our current simulations, the
patch has dimensions 4 x 4. For each patch there is a corresponding fixed-size pool
of hidden units. To achieve uniformity of response across the image, the pools are
arranged in a spatiotopic array in which neighboring pools respond to neighboring
patches and the weights of all pools are consbained to be the same.
The feature units activate the hidden units, which in turn feed back to the feature
units. Through a relaxation process, the system settles on an assignment of phases
to the features.
Learning to Segment Images Using Dynamic Feature Binding
4
NETWORK DYNAMICS
Formally, the response of each feature unit i, ~i, is a complex value in polar form,
(<<li, pil, where ?li is the amplitude or activation and Pi is the phase. Similarly, the
response of each hidden unit ;, 11;, has components (b;, q;). The weight connecting
unit i to unit ;, wiiJ is also complex valued, having components (Pii,8ii ). The
activation rule we propose is a generalization of the dot product to the complex
domain:
neti
x?wi
Ei~iWii
([(Ei?lip;i cos(Pi - 8;i?2
t
-1
an
+ (Ei?liPii sin(pi -
8ii ?2] ! ,
[Ei?lip;iSin(Pi - 8ii )])
Ei?liP;i COS(pi - 8;i)
where net; is the net input to hidden unit;. The net input is passed through
a squashing nonlinearity that maps the amplitude of the response from the range
o -+ 00 to 0 -+ 1 but leaves the phase unaffected:
1Ii
neti (1 _ e-Inetjl:l) .
Inet;1
The :Bow of activation from the hidden layer to the feature layer follows the same
dynamics, although in the current implementation the amplitudes of the features
are clamped, hence the top-down How affects only the phases. One could imagine a
more general architecture in which the relaxation process determined not only the
phase values, but cleaned up noise in the feature amplitudes as well.
The intuition underlying the activation rule is as follows. The activity of a hidden
unit, b;, should be monotonically related to how well the feature response pattern
matches the hidden unit weight vector, just as in the standard real-valued activation
rule. Indeed, one can readily see that if the feature and weight phases are equal
(Pi
8;i), the rule for bi reduces to the real-valued case. Even if the feature
and weight phases differ by a constant (Pi
8i i + e), b; is unaffected. This is
a critical property of the activation rule: Because ab.olute phase values have no
inhinsic meaning, the response of a unit should depend only on the relative phases.
The activation rule achieves this by essentially ignoring the average difference in
phase between the feature units and the weights. The hidden phase, q;, reHects this
average difference.
=
5
=
LEARNING ALGORITHM
During training, we would like the hidden units to learn to detect configurations
of features that reliably indicate phase relationships among the features. We have
experimented with a variety of training algorithms. The one with which we have
had greatest success involves running the network for a fixed number of iterations
and, after each iteration, attempting to adjust the weights so that the feature phase
pattern will match a target phase pattern. Each training hial proceeds as follows:
439
440
Mozer, Zemel, and Behrmann
1. A training example is generated at random. This involves selecting two contours and instantiating them in an image. The features of one contour have
target phase 0? and the features of the other contour have target phase 180?.
2. The training example is presented to MAGIC by clamping the amplitude of a
feature unit to 1.0 ifits corresponding image feature is present, or 0.0 otherwise.
The phases ofthe feature units are set to random values in the range 0? to 360?.
3. Activity is allowed to :flow from the feature units to the hidden units and back
to the feature units. Because the feature amplitudes are clamped, they are
unaffected.
4. The new phase pattern over the feature units is compared to the target phase
pattern (see step I), and an error measure is computed:
E = -(Et(l( cos(Pi - Pi))2 - (Eta. sin(Pi - Pi))2,
where p is the target phase pattern. This error ignores the absolute difference
between the target and actual phases. That is, E is minimized when Pi - Pi is
a constant for all i, regardless of the value of Pi - Pi.
5. Using a generalization of back propagation to complex valued units, error gradients are computed for the feature-to-hidden and hidden-to-feature weights.
6. Steps 3-5 are repeated for a maximum of 30 iterations. The trial is terminated
if the error increases on five consecutive iterations.
7. Weights are updated by an amount proportional to the average error gradient
over iterations.
Learning is more robust when the feature-to-hidden weights are constrained to be
symmetric with the hidden-to-feature weights. For complex weights, symmetry
means that the weight from feature unit i to hidden unit j is the complex conjugate of the weight from hidden unit j to feature unit i. Weight symmetry ensures
that MAGIC will converge to a fixed point. (The proof is based on discrete-time
update and a two-layer architecture with sequential layer updates and no intralayer
connections. )
Simulations reported below use a learning rate of .005 for the amplitudes and 0.02
for the phases. About 10,000 learning trials are required for stable performance,
although MAGIC rapidly picks up on the most salient aspects of the domain.
6
SIMULATION RESULTS
We trained a network with 20 hidden units per pool on images containing either
two rectangles, two diamonds, or a rectangle and a diamond. The shapes were of
varying size and appeared in various locations. A subset of the resulting weights are
shown in Figure 2. Each hidden unit attempts to detect and reinstantiate activity
patterns that match its weights. One clear and prevalent pattern in the weights is
the collinear arrangement of segments of a given orientation, all having the same
phase value. When a hidden unit having weights of this form responds to a patch of
the feature array, it tries align the phases of the patch with the phases of its weight
vector. By synchronizing the phases of features, it acts to group the features. Thus,
one can interpret the weight vectors as the rules by which features are grouped.
Learning to Segmem Images Using Dynamic Feature Binding
G
:'" "::OO
,',' 'Q QG)
': : ,',' 'G J::O :O
:' ":'8' ,','
.
..::. . -::.
':-.
..;:.
.:;"
.;..
.;::-
.;:.-
' ~G
:
V
Phase Spectrum
Figure 2: Sample of feature-to-hidden weights learned by MAGIC. The area of a
circle represents the amplitude of a weight, the orientation of the internal tick mark
represents the phase angle. The weights are arranged such that the connections
into each hidden unit are presented on a light gray background. Each hidden unit
has a total of 64 incoming weights--t x 4 locations in its receptive field and four
feature types at each location. The weights are further grouped by feature type,
and for each feature type they are arranged in a 4 X 4 pattern homologous to the
image patch itself.
Whereas traditional grouping principles indicate the conditions under which features
should be bound together as part of the same object, the grouping principles learned
by MAGIC also indicate when features should be segregated into different objects.
For example, the weights of the vertical and horizontal segments are generally 180 0
out of phase with the diagonal segments. This allows MAGIC to segregate the vertical
and horizontal features of a rectangle from the diagonal features of a diamond. We
had anticipated that the weights to each hidden unit would contain two phase
values at most because each image patch contains at most two objects. However,
some units make use of three or more phases, suggesting that the hidden unit is
performing several distinct functions. As is the usual case with hidden unit weights,
these patterns are difficult to interpret.
Figure 3 presents an example of the network segmenting an image. The image
contains two diamonds. The top left panel shows the features of the diamonds and
their initial random phases. The succeeding panels show the network's response
during the relaxation process. The lower right panel shows the network response at
equilibrium. Features of each object have been assigned a uniform phase, and the
two objects are 1800 out of phase. The task here may appear simple, but it is quite
challenging due to the illusory diamond generated by the overlapping diamonds.
441
442
Mozer, Zemel, and Behrmann
"'",
''''?''''tiIi'?''''
<~:~::-::4$s.,'
.,.."
...
..... ,.. "
'"
.
"
:;::.
.~
" ?..... # ",
?" ... ,;..,; ,'
, ,;? "
.;::'
..?
..-?
:.;.
.:::.
~
;::::
~
~~
.:::.
?
.
'::!?
~
Iteration 0
Iteration 2
Iteration 4
Iteration 6
Iteration 10
Iteration 25
Figure 3: An example of MAGIC segmenting an image. The "iteration" refers to
the number of times activity has flowed from the feature units to the hidden units
and back. The phase value of a feature is represented by a gray level. The periodic
phase continuum can only be approximated by the linear gray level continuum, but
the basic information is conveyed nonetheless.
7
CURRENT DIRECTIONS
We are currently extending
MAGIC
in several diredions, which we outline here.
? A natural principle for the hierarchical decomposition of objects emerges from
the relative frequency of feature configurations during training. More frequent
configurations result in a robust hidden representation, and hence the features
forming these configurations will be tightly coupled. A coarse quantization of
phases will lead to parses of the image in which only the highest frequency
configurations are considered as "objeds." Finer quantizations will lead to a
further decomposition of the image. Thus, the continuous phase representation
allows for the construdion of hierarchical descriptions of objeds.
? Spatially local grouping principles are unlikely to be sufficient for the image
segmentation task. Indeed, we have encountered incorred solutions produced
by MAGIC that are locally consistent but globally inconsistent. To solve this
problem, we are investigating an architecture in which the image is processed
at several spatial scales simultaneously.
? Simulations are also underway to examine MAGIC'S performance on real-world
images-overlapping handwriUen leUers and digits-where it is somewhat less
clear to which types of paUerns the hidden units should respond.
? Zemel, Williams, and Mozer (to appear) have proposed a mathematical framework that-with slight modifications to the model-allow it to be interpreted
Learning
to
Segment Images Using Dynamic Feature Binding
as a mean-field approximation to a stochastic phase model .
? Behrmann, Zemel, and Mozer (to appear) are conducting psychological experiments to examine whether limitations of the model match human limitations.
Acknowledgements
This research was supported by NSF Presidential Young Investigator award ffiI-9058450,
grant 90-21 from the James S. McDonnell Foundation, and DEC external research grant
1250 to MM, and by a National Sciences and Engineering Research Council Postgraduate
Scholarship to RZ. Our thanks to Paul Smolensky, Chris Williams, Geoffrey Hinton, and
Jiirgen Schmidhuber for helpful comments regarding this work.
References
Eckhorn, R., Bauer, R., Jordan, W., Brosch, M., Kruse, W., Munk, M., &; Reitboek, H.
J. (1988). Coherent oscillations: A mechanism of feature linking in the visual cortex?
Biological Cybernetic8, 60, 121-130.
Goebel, R. (1991). An oscillatory neural network model of visual attention, pattern recognition, and response generation. Manuscript in preparation.
Gray, C. M., Koenig, P., Engel, A. K., &; Singer, W. (1989). Oscillatory responses in
cat visual cortex exhibit intercolumnar synchronization which reflects global stimulus
properties. Nature (London), 338, 334-337.
Guzman, A. (1968). Decomposition of a visual scene into three-dimensional bodies. AFIPS
Fall Joint Computer Conference, 33, 291-304.
Hummel, J. E., &; Biederman, r. (1992). Dynamic binding in a neural network for shape
recognition. P8ychological Review. In Press.
Kanade, T. (1981). Recovery of the three-dimensional shape of an object from a single
view. Artificial Intelligence, 17, 409-460.
Lowe, D. G. (1985). Perceptual Organization and Vi8ual Recognition. Boston: Kluwer
Academic Publishers.
Lowe, D. G., &; Binford, T. O. (1982). Segmentation and aggregation: An approach to
figure-ground phenomena. In Proceeding8 of the DARPA IU Work8hop (pp. 168-178).
Palo Alto, CA: (null).
Marr, D. (1982). Vi8ion. San Francisco: Freeman.
Rock, I., &; Palmer, S. E. (1990). The legacy of Gestalt psychology. Scientific American,
!63, 84-90.
von der Malsburg, C. (1981). The correlation theory of brain function (Internal Report
81-2). Goettingen: Department of Neurobiology, Max Planck Intitute for Biophysical
Chemistry.
von der Malsburg, C., &; Schneider, W. (1986). A neural cocktail-party processor. Biological Cybernetic8, 54, 29-40.
443
| 540 |@word trial:2 faculty:1 simulation:4 decomposition:3 pick:1 initial:3 configuration:7 contains:7 selecting:1 current:4 activation:9 must:1 readily:1 predetermined:1 shape:3 succeeding:1 update:2 intelligence:1 leaf:1 nervous:1 indicative:1 plane:1 compo:2 coarse:1 toronto:4 location:4 five:1 mathematical:1 constructed:1 consists:1 manner:1 indeed:2 examine:2 brain:1 freeman:1 globally:1 actual:1 grayed:2 underlying:1 panel:3 alto:1 null:1 interpreted:1 finding:1 temporal:1 act:1 unit:53 grant:2 appear:3 planck:1 segmenting:3 before:1 engineering:1 bind:1 timing:1 local:2 pauern:2 despite:1 firing:2 studied:1 dynamically:1 challenging:1 co:4 ease:1 palmer:2 range:3 bi:1 digit:1 procedure:1 area:1 projection:1 refers:1 map:1 primitive:2 regardless:1 williams:2 attention:1 identifying:1 reitboek:1 recovery:1 rule:7 array:7 marr:2 notion:1 transmit:1 updated:1 imagine:1 target:6 colorado:1 massive:1 element:2 recognition:7 approximated:1 labeled:2 region:3 ensures:1 highest:1 mozer:5 intuition:1 govern:1 locking:1 dynamic:10 trained:2 uniformity:1 depend:1 segment:8 proceeding8:1 triangle:1 joint:1 darpa:1 represented:2 various:1 cat:1 distinct:1 describe:1 activate:1 london:1 artificial:1 zemel:6 labeling:3 quite:1 heuristic:6 valued:6 plausible:1 solve:1 otherwise:1 presidential:1 itself:1 la4:1 afips:1 biophysical:1 net:3 rock:2 propose:1 product:1 frequent:1 neighboring:2 rapidly:1 bow:1 ontario:2 achieve:1 description:1 regularity:2 extending:1 object:18 help:1 oo:1 recurrent:1 completion:1 involves:2 indicate:3 pii:1 differ:1 direction:2 stochastic:1 human:2 settle:1 munk:1 assign:1 generalization:3 biological:2 goettingen:1 mm:1 considered:1 ground:1 equilibrium:1 achieves:1 early:1 consecutive:1 continuum:2 polar:1 label:1 currently:1 palo:1 council:1 grouped:4 engel:1 reflects:1 varying:1 prevalent:1 detect:4 inst:1 helpful:1 unlikely:2 hidden:33 relation:1 labelings:1 iu:1 among:3 orientation:2 spatial:2 constrained:1 equal:1 field:2 having:3 represents:3 synchronizing:1 presegmented:3 anticipated:1 minimized:1 guzman:2 others:1 stimulus:1 richard:1 report:1 oriented:1 simultaneously:1 tightly:1 national:1 phase:45 hummel:2 ab:1 attempt:1 organization:1 adjust:1 light:1 predefined:1 necessary:1 circle:1 psychological:2 earlier:2 eta:1 assignment:1 subset:2 uniform:1 conducted:1 reported:1 spatiotopic:3 periodic:1 daptive:1 thanks:1 moser:1 pool:5 michael:1 connecting:1 together:1 von:4 containing:2 cognitive:1 external:1 american:1 li:2 suggesting:1 chemistry:1 performed:1 try:1 lowe:4 view:1 aggregation:1 capability:3 pil:1 accuracy:1 conducting:1 ofthe:2 produced:1 researcher:2 finer:1 m5s:2 unaffected:3 processor:1 oscillatory:3 binford:2 nonetheless:1 frequency:3 pp:1 james:1 jiirgen:1 conveys:1 proof:1 illusory:1 emerges:1 segmentation:4 amplitude:9 back:5 manuscript:1 feed:1 permitted:1 response:11 arranged:5 just:2 correlation:2 koenig:1 horizontal:2 ei:5 unsegmented:1 overlapping:4 propagation:2 gray:5 scientific:1 hypothesized:1 contain:3 validity:1 hence:2 assigned:1 spatially:1 symmetric:1 sin:2 during:3 intercolumnar:1 outline:1 ffii:1 performs:2 tiii:1 image:36 meaning:1 discovers:2 novel:1 belong:2 slight:1 linking:1 kluwer:1 interpret:2 refer:1 goebel:2 cybernetic8:2 eckhorn:2 grid:2 similarly:1 nonlinearity:1 had:2 dot:1 stable:1 cortex:2 align:2 aligning:1 recent:1 perspective:1 belongs:1 schmidhuber:1 success:1 der:4 accomplished:1 somewhat:1 schneider:2 converge:1 kruse:1 monotonically:1 signal:2 ii:4 multiple:4 reduces:1 match:4 academic:1 cross:1 award:1 qg:1 instantiating:1 basic:1 vision:2 essentially:1 iteration:12 represent:1 dec:1 background:1 whereas:1 publisher:1 comment:1 undergo:1 facilitates:1 flow:1 inconsistent:1 jordan:1 call:1 structural:2 variety:1 affect:1 psychology:2 architecture:4 idea:1 regarding:1 whether:1 utility:1 passed:1 collinear:1 cocktail:1 generally:1 clear:2 amount:1 locally:1 processed:1 nsf:1 per:2 discrete:1 group:4 key:1 four:3 salient:1 isin:1 rectangle:4 relaxation:4 angle:1 respond:2 laid:1 patch:8 oscillation:1 hexagon:1 layer:14 bound:1 display:1 encountered:1 activity:6 occur:1 scene:2 aspect:1 attempting:1 juxtaposed:1 performing:1 relatively:1 department:1 according:2 combination:1 mcdonnell:1 conjugate:1 across:2 wi:1 biologically:1 modification:1 psychologist:1 boulder:1 taken:1 brosch:1 previously:1 describing:1 discus:1 mechanism:2 turn:1 neti:2 singer:1 acronym:1 operation:2 hierarchical:2 appropriate:1 rz:1 top:2 running:1 include:1 malsburg:4 medicine:1 scholarship:1 arrangement:1 spike:1 receptive:1 usual:1 responds:1 traditional:1 diagonal:2 exhibit:1 gradient:2 chris:1 reason:1 relationship:1 difficult:2 nonintuitive:2 magic:22 implementation:1 reliably:1 perform:2 diamond:8 upper:2 vertical:2 segregate:1 hinton:1 neurobiology:1 canada:2 biederman:2 cleaned:1 required:1 connection:2 lal:1 coherent:1 learned:2 suggested:2 proceeds:1 below:1 perception:1 pattern:11 appeared:1 smolensky:1 including:1 max:1 reciprocally:2 greatest:1 overlap:1 critical:1 natural:1 homologous:1 representing:1 coupled:1 review:1 geometric:1 acknowledgement:1 segregated:1 underway:1 relative:4 synchronization:2 par:1 generation:1 limitation:2 proportional:1 geoffrey:1 foundation:1 conveyed:1 sufficient:1 consistent:3 principle:5 pi:15 squashing:1 supported:1 legacy:1 tick:1 allow:1 fall:1 absolute:1 bauer:1 dimension:1 world:1 contour:7 ignores:1 adaptive:1 san:1 party:1 gestalt:2 global:1 incoming:1 investigating:1 francisco:1 spectrum:1 search:1 continuous:1 lip:3 kanade:2 nature:1 learn:2 robust:2 ca:1 ignoring:1 symmetry:2 investigated:1 complex:9 intralayer:1 domain:4 terminated:1 noise:1 paul:1 allowed:1 repeated:1 body:1 neuronal:1 position:1 clamped:2 perceptual:1 behrmann:5 learns:2 young:1 down:1 explored:1 experimented:1 multitude:1 grouping:16 quantization:2 postgraduate:1 sequential:1 clamping:1 boston:1 forming:1 visual:6 binding:9 corresponds:1 chance:1 determined:2 called:2 total:1 indicating:1 formally:1 internal:2 people:2 mark:1 preparation:1 investigator:1 dept:3 tested:1 phenomenon:1 |
4,861 | 5,400 | A Dual Algorithm for Olfactory Computation in the
Locust Brain
Sina Tootoonian
[email protected]
M?at?e Lengyel
[email protected]
Computational & Biological Learning Laboratory
Department of Engineering, University of Cambridge
Trumpington Street, Cambridge CB2 1PZ, United Kingdom
Abstract
We study the early locust olfactory system in an attempt to explain its wellcharacterized structure and dynamics. We first propose its computational function
as recovery of high-dimensional sparse olfactory signals from a small number
of measurements. Detailed experimental knowledge about this system rules out
standard algorithmic solutions to this problem. Instead, we show that solving a
dual formulation of the corresponding optimisation problem yields structure and
dynamics in good agreement with biological data. Further biological constraints
lead us to a reduced form of this dual formulation in which the system uses independent component analysis to continuously adapt to its olfactory environment
to allow accurate sparse recovery. Our work demonstrates the challenges and rewards of attempting detailed understanding of experimentally well-characterized
systems.
1
Introduction
Olfaction is perhaps the most widespread sensory modality in the animal kingdom, often crucial for
basic survival behaviours such as foraging, navigation, kin recognition, and mating. Remarkably,
the neural architecture of olfactory systems across phyla is largely conserved [1]. Such convergent
evolution suggests that what we learn studying the problem in small model systems will generalize
to larger ones. Here we study the olfactory system of the locust Schistocerca americana. While
we focus on this system because it is experimentally well-characterized (Section 2), we expect our
results to extend to other olfactory systems with similar architectures. We begin by observing that
although most odors are mixtures of hundreds of molecular species, with typically only a few of
these dominating in concentration ? i.e. odors are sparse in the space of molecular concentrations
(Fig. 1A). We introduce a simple generative model of odors and their effects on odorant receptors
that reflects this sparsity (Section 3). Inspired by recent experimental findings [2], we then propose
that the function of the early olfactory system is maximum a posteriori (MAP) inference of these
concentration vectors from receptor inputs (Section 4). This is basically a sparse signal recovery
problem, but the wealth of biological evidence available about the system rules out standard solutions. We are then led by these constraints to propose a novel solution to this problem in term of
its dual formulation (Section 5), and further to a reduced form of this solution (Section 6) in which
the circuitry uses ICA to continuously adapt itself to the local olfactory environment (Section 7).
We close by discussing predictions of our theory that are amenable to testing in future experiments,
and future extensions of the model to deal with readout and learning simultaneously, and to provide
robustness against noise corrupting sensory signals (Section 8).
1
A
B
C
Odors
~1000
glomeruli
~1000
PNs
~100
bLNs
Behaviour
~300 LNs
Relative concentration
D
50,000
KCs
antennal lobe (AL)
90,000
ORNs
Molecules
E
0
1
Time (s)
2
antenna
mushroom body
(MB) 1 GGN
0
1
Time (s)
2
Figure 1: Odors and the olfactory circuit. (A) Relative concentrations of ? 70 molecules in the
odor of the Festival strawberry cultivar, demonstrating sparseness of odor vectors. (B,C) Diagram
and schematic of the locust olfactory circuit. Inputs from 90,000 ORNs converge onto ? 1000
glomeruli, are processed by the ? 1000 cells (projection neurons, PN, and local internuerons, LNs)
of the antennal lobe, and read out in a feedforward manner by the 50,000 Kenyon cells (KC) of the
mushroom body, whose activity ultimately is read out to produce behavior. (D,E) Odor response
of a PN (D) and a KC (E) to 7 trials of 44 mixtures of 8 monomolecular components (colors)
demonstrating cell- and odor-specific responses. The odor presentation window is in gray. PN
responses are dense and temporally patterned. KC responses are sparse and are often sensitive to
single molecules in a mixture. Panel A is reproduced from [8], B from [6], and D-E from the dataset
in [2].
2
Biological background
A schematic of the locust olfactory system is shown in Figure 1B-C. Axons from ? 90, 000 olfactory
receptor neurons (ORNs) each thought to express one type of olfactory receptor (OR) converge onto
approximately 1000 spherical neuropilar structures called ?glomeruli?, presumably by the ?1-OR-to1-glomerulus? rule observed in flies and mice. The functional role of this convergence is thought to
be noise reduction through averaging.
The glomeruli are sampled by the approximately 800 excitatory projection neurons (PNs) and 300
inhibitory local interneurons (LNs) of the antennal lobe (AL). LNs are densely connected to other
LNs and to the PNs; PNs are connected to each-other only indirectly via their dense connections
to LNs [3]. In response to odors, the AL exhibits 20 Hz local field potential oscillations and odorand cell-specific activity patterns in its PNs and LNs (Fig. 1D). The PNs form the only output of
the AL and project densely [4] to the 50,000 Kenyon cells (KCs) of the mushroom body (MB).
The KCs decode the PNs in a memoryless fashion every oscillation cycle, converting the dense
and promiscuous PN odor code into a very sparse and selective KC code [5], often sensitive to
a single component in a complex odor mixture [2] (Fig. 1E). KCs make axo-axonal connections
with neighbouring KCs [6] but otherwise only communicate with one-another indirectly via global
inhibition mediated by the giant GABA-ergic neuron [7]. Thus, while the AL has rich recurrency,
there is no feedback from the KCs back to the AL: the PN to KC circuit is strictly feedforward. As
we shall see below, this presents a fundamental challenge to theories of AL-MB computation.
2
3
Generative model
Natural odors are mixtures of hundreds of different types of molecules at various concentrations (e.g.
[8]), and can be represented as points in RN
+ , where each dimension represents the concentration
of one of the N molecular species in ?odor space?. Often a few of these will be at a much higher
concentration than the others, i.e. natural odors are sparse. Because the AL responds similarly across
concentrations [9] , we will ignore concentration in our odor model and consider odors as binary
vectors x ? {0, 1}N . We will also assume that molecules appear in odor vectors independently of
one-another with probability k/N , where k is the average complexity of odors (# of molecules/odor,
equivalently the Hamming weight of x) in odor space.
We assume a linear noise-free observation model y = Ax for the M -dimensional glomerular activity vector (we discuss observation noise in Section 7). A is an M ? N affinity matrix representing
the response of each of the M glomeruli to each of the N molecular odor components and has elements drawn iid. from a zero-mean Gaussian with variance 1/M . Our generative model for odors
and observations is summarized as
x = {x1 , . . . , xN }, xi ? Bernoulli(k/N ),
4
y = Ax,
Aij ? N (0, M ?1 )
(1)
Basic MAP inference
Inspired by the sensitivity of KCs to monomolecular odors [2], we propose that the locust olfactory
system acts as a spectrum analyzer which uses MAP inference to recover the sparse N -dimensional
odor vector x responsible for the dense M -dimensional glomerular observations y, with M
N e.g. O(1000) vs. O(10000) in the locust. Thus, the computational problem is akin to one in
compressed sensing [10], which we will exploit in Section 5. We posit that each KC encodes the
presence of a single molecular species in the odor, so that the overall KC activity vector represents
the system?s estimate of the odor that produced the observations y.
To perform MAP inference on binary x from y given A, a standard approach is to relax x to the
positive orthant RN
+ [11], smoothen the observation model with isotropic Gaussian noise of variance
? 2 and perform gradient descent on the log posterior
log p(x|y, A, k) = C ? ?kxk1 ?
where ? = log((1 ? q)/q), q = k/N , kxk1 =
of the posterior determines the x dynamics:
PM
i=1
1
ky ? Axk22
2? 2
(2)
xi for x 0, and C is a constant. The gradient
1
AT (y ? Ax)
(3)
2? 2
Given our assumed 1-to-1 mapping of KCs to (decoded) elements of x, these dynamics fundamentally violate the known biology for two reasons. First, they stipulate KC dynamics where there are
none. Second, they require all-to-all connectivity of KCs via AT A where none exist. In reality, the
dynamics in the circuit occur in the lower (? M ) dimensional measurement space of the antennal
lobe, and hence we need a way of solving the inference problem there rather than directly in the high
(N ) dimensional space of KC activites.
x? ? ?x log p = ?? sgn(x) +
5
Low dimensional dynamics from duality
To compute the MAP solution using lower-dimensional dynamics, we consider the following compressed sensing (CS) problem:
minimize kxk1 ,
subject to ky ? Axk22 = 0
(4)
whose Lagrangian has the form
L(x, ?) = kxk1 + ?ky ? Axk22
(5)
where ? is a scalar Lagrange multiplier. This is exactly the equation for our (negative) log posterior
(Eq. 2) with the constants absorbed by ?. We will assume that because x is binary, the two systems
will have the same solution, and will henceforth work with the CS problem.
3
To derive low dimensional dynamics, we first reformulate the constraint and solve
minimize kxk1 ,
subject to y = Ax
(6)
with Lagrangian
L(x, ?) = kxk1 + ?T (y ? Ax)
(7)
where now ? is a vector of Lagrange multipliers. Note that we are still solving an N -dimensional
minimization problem with M N constraints, while we need M -dimensional dynamics. Therefore, we consider the dual optimization problem of maximizing g(?) where g(?) = inf x L(x, ?)
is the dual Lagrangian of the problem. If strong duality holds, the primal and dual objectives have
the same value at the solution, and the primal solution can be found by minimizing the Lagrangian
at the optimal value of ? [11]. Were x ? RN , strong duality would hold for our problem by Slater?s
sufficiency condition [11]. The binary nature of x robs our problem of the convexity required for
this sufficiency condition to be applicable. Nevertheless we proceed assuming strong duality holds.
The dual Lagrangian has a closed-form expression for our problem. To see this, let b = AT ?.
Then, exploiting the form of the 1-norm and x being binary, we obtain the following:
g(?)??T y = inf kxk1 ?bT x = inf
x
x
T
M
X
(|xi |?bi xi ) =
i=1
M
X
i=1
inf (|xi |?bi xi ) = ?
xi
M
X
[bi ?1]+ (8)
i=1
T
or, in vector form, g(?) = ? y ? 1 [b ? 1]+ , where [?]+ is the positive rectifying function.
Maximizing g(?) by gradient descent yields M dimensional dynamics in ?:
?? ? ?? g = y ? A ?(AT ? ? 1)
(9)
where ?(?) is the Heaviside function. The solution to the CS problem ? the odor vector that produced
the measurements y ? is then read out at the convergence of these dynamics to ?? as
x? = argminx L(x, ?? ) = ?(AT ?? ? 1)
(10)
A natural mapping of equations 9 and 10 to antennal lobe dynamics is for the output of the M
glomeruli to represent y, the PNs to represent ?, and the KCs to represent (the output of) ?, and
hence eventually x? . Note that this would still require the connectivity between PNs and KCs to
be negative reciprocal (and determined by the affinity matrix A). We term the circuit under this
mapping the full dual circuit (Fig. 2B). These dynamics allow neuronal firing rates to be both
positive and negative, hence they can be implemented in real neurons as e.g. deviations relative to a
baseline rate [12], which is subtracted out at readout.
We measured the performance of a full dual network of M = 100 PNs in recovering binary odor
vectors containing an average of k = 1 to 10 components out of a possible N = 1000. The
results in Figure 2E (blue) show that the dynamics exhibit perfect recovery.1 For comparison, we
have included the performance of the purely feedforward circuit (Fig. 2A), in which the glomerular
vector y is merely scaled by the k-specific amount that yields minimum error before being read
out by the KCs (Fig. 2E, black). In principle, no recurrent circuit should perform worse than
this feedfoward network, otherwise we have added substantial (energetic and time) costs without
computational benefits.
6
The reduced dual circuit
The full dual antennal lobe circuit described by Equations 9 and 10 is in better agreement with the
known biology of the locust olfactory system than 2 for a number of reasons:
1. Dynamics are in the lower dimensional space of the antennal lobe PNs (?) rather than the
mushroom body KCs (x).
2. Each PN ?i receives private glomerular input yi
3. There are no direct connections between PNs; their only interaction with other PNs is
indirect via inhibition provided by ?.
1
See the the Supplementary Material for considerations when simulating the piecewise linear dynamics of
9.
4
Full Dual
PNs
KCs
D
PN activation
glom.
Odor
B
Feedforward Circuit
0.04
0.02
0
LN activation
Odor
A
1
0.8
0.6
0.4
0.2
0
E
Time (a.u.)
8
Feedforward
Full Dual
Reduced Dual
7
6
Distance
Reduced Dual
Odor
C
5
4
3
2
1
LNs
0
1
2
3
4
5
k
6
7
8
9
10
Figure 2: Performance of the feedforward and the dual circuits. (A-C) Circuit schematics. Arrows
(circles) indicate excitatory (inhibitory) connections. (D) Example PN and LN odor-evoked dynamics for the reduced dual circuit. Top: PNs receive cell-specific excitation or inhibition whose strength
is changed as different LNs are activated, yielding cell-specific temporal patterning. Bottom: The
LNs whose corresponding KCs encode the odor (red) are strongly excited and eventually breach
the threshold (dashed line), causing changes to the dynamics (time points marked with dots). The
excitation of the other LNs (pink) remains subthreshold. (E) Hamming distance between recovered
and true odor vector as a function of odor density k. The dual circuits generally outperform the
feedforward system over the entire range tested. Points are means, bars are s.e.m., computed for 200
trials (feedforward) and all trials from 200 attempts in which the steady-state solution was found
(dual circuits, greater than 90%).
4. The KCs serve merely as a readout stage and are not interconnected.2
However, there is also a crucial disagreement of the full dual dynamics with biology: the requirement
for feedback from the KCs to the PNs. The mapping of ? to PNs and ? to the KCs in Equation 9
implies negative reciprocal connectivity of PNs and KCs, i.e. a feedforward connection of Aij from
PN i to KC j, and a feedback connection of ?Aij from KC j to PN i. This latter connection from
KCs to PNs violates biological fact ? no such direct and specific connectivity from KCs to PNs exists
in the locust system, and even if it did, it would most likely be excitatory rather than inhibitory, as
KCs are excitatory.
Although KCs are not inhibitory, antennal lobe LNs are and connect densely to the PNs. Hence they
could provide the feedback required to guide PN dynamics. Unfortunately, the number of LNs is on
the order of that of the PNs, i.e. much fewer than the number of the KCs, making it a priori unlikely
that they could replace the KCs in providing the detailed pattern of feedback that the PNs require
under the full dual dynamics.
To circumvent this problem, we make two assumptions about the odor environment. The first is
that any given environment contains a small fraction of the set of all possible molecules in odor
space. This implies the potential activation of only a small number of KCs, whose feedback patterns
(columns of A) could then be provided by the LNs. The second assumption is that the environment
changes sufficiently slowly that the animal has time to learn it, i.e. that the LNs can update their
feedback patterns to match the change in required KC activations.
This yields the reduced dual circuit, in which the reciprocal interaction of the PNs with the KCs via
the matrix A is replaced with interaction with the M LNs via the square matrix B. The activity of
the LNs represents the activity of the KCs encoding the molecules in the current odor environment,
2
Although axo-axonal connections between neighbouring KC axons in the mushroom body peduncle are
known to exist [6], see also Section 2.
5
and the columns of B are the corresponding columns of the full A matrix:
?? ? y ? B ?(BT ? ? 1),
x = ?(AT ? ? 1)
(11)
Note that instantaneous readout of the PNs is still performed by the KCs as in the full dual. The
performance of the reduced dual is shown in red in Figure 2E, demonstrating better performance
than the feedforward circuit, though not the perfect recovery of the full dual. This is because the
solution sets of the two equations are not the same: Suppose that B = A:,1:M , and that y =
Pk
T
i=1 A:,i . The corresponding solution set for reduced dual is ?1 (y) = {? : (B:,1:k ) ? > 1 ?
T
T
T
(B:,k+1:M ) ? < 1}, equivalently ?1 (y) = {? : (A:,1:k ) ? > 1 ? (A:,k+1:M ) ? < 1}. On the
other hand, the solution set for the full dual is ?0 (y) = {? : (A:,1:k )T ? > 1 ? (A:,k+1:M )T ? <
1 ? (A:,M +1:N )T ? < 1}. Note the additional requirement that the projection of ? onto columns
M + 1 to N of A must also be less than 1. Hence any solution to the full dual is a solution to the
reduced dual , but not necessarily vise-versa: ?0 (y) ? ?1 (y). Since only the former are solutions to
the full problem, not all solutions to the reduced dual will solve it, leading to the reduced peformance
observed. This analysis also implies that increasing (or decreasing) the number of columns in B, so
that it is no longer square, will improve (worsen) the performance of the reduced dual, by making
its solution-set a smaller (larger) superset of ?0 (y).
7
Learning via ICA
Figure 2 demonstrates that the reduced dual has reasonable performance when the B matrix is
correct, i.e. it contains the columns of A for the KCs that would be active in the current odor
environment. How would this matrix be learned before birth, when presumably little is known about
the local environment, or as the animal moves from one odor environment to another?
Recall that, according to our generative model (Section 2) and the additional assumptions made for
deriving the reduced dual circuit (Section 6), molecules appear independently at random in odors
of a given odor environment and the mapping from odors x to glomerular responses y is linear
in x via the square mixing matrix B. Hence, our problem of learning B is precisely that of ICA
(or more precisely, sparse coding, as the observation noise variance is assumed to be ? 2 > 0 for
inference), with binary latent variables x. We solve this problem using MAP inference via EM
with a mean-field variational approximation q(x) to the posterior p(x|y, B) [13], where q(x) ,
QM
QM xi
1?xi
. The E-step, after observing that for binary x,
i=1 Bernoulli(xi ; qi ) =
i=1 qi (1 ? qi )
q
1
2
T
x = x, is ?q ? ?? ? log 1?q + ?2 B y ? ?12 Cq, with ? = ?1 + 2?1 2 c, ? = log((1 ? q0 )/q0 ),
q0 = k/M , the vector c = diag(BT B), and C = BT B ? diag(c), i.e. C is BT B with the
diagonal elements set to zero. To yield more plausible neural dynamics, we change variables to
? As vi is monotonically increasing
v = log(q/(1 ? q)). By the chain rule v? = diag(?vi /?qi )q.
in qi , and so the corresponding partial derivatives are all positive, and the resulting diagonal matrix
is positive definite, we can ignore it in performing gradient descent and still minimize the same
objective. Hence we have
?v ? ?? ? v +
1 T
1
B y ? 2 Cq(v),
?2
?
q(v) =
1
,
1 + exp(?v)
(12)
with the obvious mapping of v to LN membrane potentials, and q as the sigmoidal output function
representing graded voltage-dependent transmitter release observed in locust LNs.
The M-step update is made by changing B to increase log p(B) + Eq log p(x, y|B), yielding
1
1
B + 2 (rqT + B diag(q(1 ? q))),
M
?
Note that this update rule takes the form of a local learning rule.
?B ? ?
r , y ? Bq.
(13)
Empirically, we observed convergence within around 10,000 iterations using a fixed step size of
dt ? 10?2 , and ? ? 0.2 for M in the range of 20?100 and k in the range of 1?5. In cases when
the algorithm did not converge, lowering ? slightly typically solved the problem. The performance
of the algorithm is shown in figure 3. Although the B matrix is learned to high accuracy, it is
not learned exactly. The resulting algorithmic noise renders the performance of the dual shown in
Fig. 2E an upper bound, since there the exact B matrix was used.
6
?2
MSE
10
?4
10
Column of Btrue
?6
10
C
0
2000
4000
6000
Iteration
8000
10000
1
Coefficient of Blearned
B
0
10
Coefficient of Binitial
A
0
Column of Btrue
-1
Figure 3: ICA performance for M = 40, k = 1, dt = 10?2 . (A) Time course of mean squared
error between the elements of the estimate B and their true values for 10 different random seeds.
? = 0.162 for six of the seeds, 0.15 for three, and 0.14 for one. (B,C) Projection of the columns of
Btrue into the basis of the columns of B before (B) and after learning (C), for one of the random
seeds. Plotted values before learning are clipped to the -1?1 range.
8
8.1
Discussion
Biological evidence and predictions
Our work is consistent with much of the known anatomy of the locust olfactory system, e.g. the
lack of connectivity between PNs and dense connectivity between LNs, and between LNs and PNs
[3]; direct ORN inputs to LNs (observed in flies [14]; unknown in locust); dense connectivity from
PNs to KCs [4]; odor-evoked dynamics in the antennal lobe [2], vs. memoryless readout in the KCs
[5]. In addition, we require gradient descent PN dynamics (untested directly, but consistent with PN
dynamics reaching fixed-points upon prolonged odor presentation [15]), and short-term plasticity in
the antennal lobe for ICA (a direct search for ICA has not been performed, but short-term plasticity
is present in trial-to-trial dynamics [16]).
Our model also makes detailed predictions about circuit connectivity. First, it predicts a specific
structure for the PN-to-KC connectivity matrix, namely AT , the transpose of the affinity matrix.
This is superficially at odds with recent work in flies suggesting random connectivity between PNs
and KCs (detailed connectivity information is not present in the locust). Murthy and colleagues
[17] examined a small population of genetically identifiable KCs and found no evidence of response
stereotypy across flies, unlike that present at earlier stages in the system. Our model is agnostic
to permutations of the output vector as these reassign the mapping between KCs and molecules
and affect neither information content nor its format, so our results would be consistent with [17]
under animal-specific permutations. Caron and co-workers [18] analysed the structural connectivity of single KCs to glomeruli and found it consistent with random connectivity conditioned on a
glomerulus-specific connection probability. This is also consistent with our model, with the observed randomness reflecting that of the affinity matrix itself. Our model would predict (a) the
observation of repeated connectivity motifs if enough KCs (across animals) were observed, and that
(b) each connectivity motif corresponds to the (binarized) glomerular response vector evoked by a
particular molecule. In addition we predict symmetric inhibitory connectivity between LNs (BT B),
and negative reciprocal connectivity between PNs and LNs (Bij from PN i to LN j and ?Bij from
LN to PN).
8.2
Combining learning and readout
We have presented two mechanisms above ? the reduced dual for readout and and ICA for learning
? both of which need to be at play to guarantee high performance. In fact, these two mechanisms
must be active simultaneously in the animal. Here we sketch a possible mechanism for combining
them. The key is equation 12, which we repeat below, augmented with an additional term from the
PNs:
1 T
1
?v ? ?v + ?? + 2 B y ? 2 C q(v) + BT ? ? 1 = ?v + Ilearning + Ireadout .
?
?
7
A8
B
Feedforward
Full Dual
Reduced Dual
7
?0.5
6
Distance
5
4
0
3
2
1
0
1
2
3
4
5
k
6
7
8
9
0.5
?0.5
10
0
0.5
Figure 4: Effects of noise. (A) As in Figure 2E but with a small amount of additive noise in the
observations. The full dual still outperforms the feedforward circuit which in turn outperforms the
reduced dual over nearly half the tested range. (B) The feedback surface hinting at noise sensitivity.
PN phase space is colored according to activation of each of the KCs and a 2D projection around
the origin is shown. The average size of a zone with a uniform color is quite small, suggesting that
small perturbations would change the configuration of KCs activated by a PN, and hence the readout
performance.
Suppose (a) the two input channels were segregated e.g. on separate dendritic compartments, and
such that (b) the readout component was fast but weak, while (c) the learning component was slow
but strong, and (d) the v time constant was faster than both. Early after odor presentation, the main
input to the LN would be from the readout circuit, driving the PNs to their fixed point. The input
from the learning circuit would eventually catch up and dominate that of the readout circuit, driving
the LN dynamics for learning. Importantly, if B has already been learned, then the output of the
LNs, q(v), would remain essentially unchanged throughout, as both the learning and readout circuits
would produce the same (steady-state) activation vector in the LNs. If the matrix is incorrect, then
the readout is likely to be incorrect already, and so the important aspect is the learning update which
would eventually dominate. This is just one possibility for combining learning and readout. Indeed,
even the ICA updates themselves are non-trivial to implement. We leave the details of both to future
work.
8.3
Noise sensitivity
Although our derivations for serving inference and learning rules assumed observation noise, the
data that we provided to the models contained none. Adding a small amount of noise reduces
the performance of the dual circuits, particularly that of the reduced dual, as shown in Figure 4A.
Though this may partially be attributed to numerical integration issues (Supplementary Material),
there is likely a fundamental theoretical cause underlying it. This is hinted at by the plot in figure
4B of a 2D projection in PN space of the overlayed halfspaces defined by the activation of each of
? As ? crosses into
the N KCs. In the central void no KC is active and ? can change freely along ?.
?
a halfspace, the corresponding KC is activated, changing ? and the trajectory of ?. The different
colored zones indicate different patterns of KC activation and correspondingly different changes to
? The small size of these zones suggests that small changes in the trajectory of ? caused e.g. by
?.
noise could result in very different patterns of KC activation. For the reduced dual, most of these
halfspaces are absent for the dynamics since B has only a small subset of the columns of A, but
are present during readout, exacerbating the problem. How the biological system overcomes this
apparently fundamental sensitivity is an important question for future work.
Acknowledgements This work was supported by the Wellcome Trust (ST, ML).
8
References
[1] Eisthen HL. Why are olfactory systems of different animals so similar?, Brain, behavior and
evolution 59:273, 2002.
[2] Shen K, et al. Encoding of mixtures in a simple olfactory system, Neuron 80:1246, 2013.
[3] Jortner RA. Personal communication.
[4] Jortner RA, et al. A simple connectivity scheme for sparse coding in an olfactory system, The
Journal of neuroscience 27:1659, 2007.
[5] Perez-Orive J, et al. Oscillations and sparsening of odor representations in the mushroom body,
Science 297:359, 2002.
[6] Leitch B, Laurent G. Gabaergic synapses in the antennal lobe and mushroom body of the locust
olfactory system, The Journal of comparative neurology 372:487, 1996.
[7] Papadopoulou M, et al. Normalization for sparse encoding of odors by a wide-field interneuron, Science 332:721, 2011.
[8] Jouquand C, et al. A sensory and chemical analysis of fresh strawberries over harvest dates
and seasons reveals factors that affect eating quality, Journal of the American Society for Horticultural Science 133:859, 2008.
[9] Stopfer M, et al. Intensity versus identity coding in an olfactory system, Neuron 39:991, 2003.
[10] Foucart S, Rauhut H. A mathematical introduction to compressive sensing. Springer, 2013.
[11] Boyd SP, Vandenberghe L. Convex optimization. Cambridge University Press, 2004.
[12] Dayan P, Abbott L. Theoretical Neuroscience. Massachusetts Institute of Technology Press,
2005.
[13] Neal RM, Hinton GE. In Learning in graphical models, 355, 1998.
[14] Ng M, et al. Transmission of olfactory information between three populations of neurons in
the antennal lobe of the fly, Neuron 36:463, 2002.
[15] Mazor O, Laurent G. Transient dynamics versus fixed points in odor representations by locust
antennal lobe projection neurons, Neuron 48:661, 2005.
[16] Stopfer M, Laurent G. Short-term memory in olfactory network dynamics, Nature 402:664,
1999.
[17] Murthy M, et al. Testing odor response stereotypy in the Drosophila mushroom body, Neuron
59:1009, 2008.
[18] Caron SJ, et al. Random convergence of olfactory inputs in the drosophila mushroom body,
Nature 497:113, 2013.
9
| 5400 |@word trial:5 private:1 norm:1 lobe:13 eng:2 excited:1 reduction:1 configuration:1 contains:2 united:1 outperforms:2 recovered:1 current:2 analysed:1 activation:9 mushroom:9 must:2 axk22:3 numerical:1 additive:1 plasticity:2 plot:1 update:5 v:2 generative:4 fewer:1 patterning:1 half:1 isotropic:1 reciprocal:4 feedfoward:1 short:3 colored:2 sigmoidal:1 mathematical:1 along:1 direct:4 incorrect:2 olfactory:25 manner:1 introduce:1 indeed:1 ra:2 ica:8 themselves:1 nor:1 behavior:2 brain:2 inspired:2 spherical:1 decreasing:1 prolonged:1 little:1 window:1 increasing:2 begin:1 project:1 provided:3 underlying:1 circuit:26 panel:1 agnostic:1 what:1 compressive:1 finding:1 giant:1 guarantee:1 temporal:1 every:1 binarized:1 act:1 exactly:2 demonstrates:2 scaled:1 uk:2 qm:2 rm:1 appear:2 positive:5 before:4 engineering:1 local:6 receptor:4 encoding:3 glomerular:6 laurent:3 firing:1 approximately:2 black:1 examined:1 evoked:3 suggests:2 co:1 patterned:1 bi:3 range:5 locust:15 responsible:1 testing:2 definite:1 implement:1 cb2:1 binitial:1 thought:2 projection:7 boyd:1 onto:3 close:1 map:6 lagrangian:5 maximizing:2 independently:2 convex:1 shen:1 recovery:5 rule:7 importantly:1 deriving:1 dominate:2 vandenberghe:1 population:2 suppose:2 play:1 decode:1 exact:1 neighbouring:2 us:3 origin:1 agreement:2 element:4 recognition:1 particularly:1 slater:1 predicts:1 observed:7 role:1 kxk1:7 fly:5 bottom:1 solved:1 readout:15 connected:2 cycle:1 halfspaces:2 substantial:1 environment:10 convexity:1 complexity:1 reward:1 cam:2 dynamic:31 personal:1 ultimately:1 solving:3 purely:1 serve:1 upon:1 basis:1 indirect:1 various:1 represented:1 derivation:1 fast:1 birth:1 whose:5 quite:1 larger:2 dominating:1 solve:3 supplementary:2 relax:1 otherwise:2 compressed:2 plausible:1 itself:2 antenna:1 reproduced:1 propose:4 interaction:3 mb:3 interconnected:1 stipulate:1 causing:1 combining:3 date:1 to1:1 promiscuous:1 mixing:1 ky:3 exploiting:1 convergence:4 smoothen:1 requirement:2 transmission:1 produce:2 comparative:1 perfect:2 leave:1 axo:2 derive:1 recurrent:1 ac:2 measured:1 eq:2 strong:4 implemented:1 c:3 recovering:1 indicate:2 implies:3 posit:1 anatomy:1 untested:1 correct:1 sgn:1 transient:1 material:2 violates:1 orn:1 require:4 behaviour:2 drosophila:2 biological:8 dendritic:1 hinted:1 extension:1 strictly:1 hold:3 sufficiently:1 around:2 exp:1 presumably:2 seed:3 algorithmic:2 mapping:7 predict:2 circuitry:1 driving:2 early:3 applicable:1 festival:1 sensitive:2 reflects:1 minimization:1 gaussian:2 rather:3 reaching:1 pn:19 season:1 voltage:1 eating:1 encode:1 ax:5 focus:1 release:1 bernoulli:2 transmitter:1 baseline:1 posteriori:1 inference:8 dependent:1 motif:2 dayan:1 typically:2 bt:7 entire:1 unlikely:1 kc:58 selective:1 overall:1 dual:43 issue:1 priori:1 animal:7 integration:1 field:3 ng:1 biology:3 represents:3 nearly:1 future:4 others:1 fundamentally:1 piecewise:1 few:2 simultaneously:2 densely:3 replaced:1 phase:1 argminx:1 overlayed:1 olfaction:1 attempt:2 interneurons:1 possibility:1 navigation:1 mixture:6 yielding:2 perez:1 primal:2 activated:3 ergic:1 chain:1 amenable:1 accurate:1 partial:1 worker:1 bq:1 circle:1 plotted:1 mazor:1 theoretical:2 column:11 earlier:1 cost:1 deviation:1 subset:1 hundred:2 uniform:1 connect:1 foraging:1 st:1 density:1 fundamental:3 sensitivity:4 continuously:2 mouse:1 connectivity:18 squared:1 central:1 containing:1 slowly:1 henceforth:1 worse:1 american:1 derivative:1 leading:1 suggesting:2 potential:3 summarized:1 coding:3 coefficient:2 caused:1 vi:2 performed:2 closed:1 observing:2 apparently:1 red:2 recover:1 worsen:1 rectifying:1 halfspace:1 minimize:3 square:3 compartment:1 accuracy:1 variance:3 largely:1 yield:5 subthreshold:1 generalize:1 weak:1 produced:2 basically:1 iid:1 none:3 trajectory:2 rauhut:1 lengyel:2 randomness:1 murthy:2 explain:1 synapsis:1 mating:1 against:1 colleague:1 obvious:1 attributed:1 hamming:2 sampled:1 dataset:1 massachusetts:1 recall:1 knowledge:1 color:2 back:1 reflecting:1 higher:1 dt:2 response:10 formulation:3 sufficiency:2 though:2 strongly:1 just:1 stage:2 hand:1 receives:1 sketch:1 trust:1 lack:1 widespread:1 quality:1 gray:1 perhaps:1 sparsening:1 effect:2 kenyon:2 multiplier:2 true:2 evolution:2 hence:8 former:1 chemical:1 read:4 memoryless:2 laboratory:1 q0:3 symmetric:1 odorant:1 ilearning:1 deal:1 neal:1 during:1 excitation:2 steady:2 variational:1 consideration:1 novel:1 instantaneous:1 functional:1 empirically:1 extend:1 measurement:3 cambridge:3 versa:1 caron:2 glomerulus:9 pm:1 similarly:1 analyzer:1 dot:1 longer:1 surface:1 inhibition:3 posterior:4 recent:2 inf:4 pns:32 binary:8 discussing:1 yi:1 conserved:1 minimum:1 greater:1 additional:3 converting:1 freely:1 converge:3 monotonically:1 signal:3 dashed:1 violate:1 full:15 reduces:1 ggn:1 match:1 adapt:2 characterized:2 faster:1 cross:1 molecular:5 schematic:3 prediction:3 qi:5 basic:2 optimisation:1 essentially:1 iteration:2 represent:3 normalization:1 cell:7 receive:1 background:1 remarkably:1 addition:2 wealth:1 diagram:1 void:1 modality:1 crucial:2 unlike:1 hz:1 subject:2 odds:1 axonal:2 structural:1 presence:1 feedforward:12 superset:1 enough:1 peformance:1 affect:2 architecture:2 absent:1 expression:1 six:1 robs:1 akin:1 energetic:1 harvest:1 render:1 proceed:1 cause:1 reassign:1 generally:1 detailed:5 amount:3 processed:1 reduced:20 outperform:1 exist:2 inhibitory:5 neuroscience:2 blue:1 serving:1 shall:1 express:1 key:1 demonstrating:3 nevertheless:1 threshold:1 drawn:1 changing:2 neither:1 abbott:1 lowering:1 merely:2 fraction:1 communicate:1 clipped:1 throughout:1 reasonable:1 oscillation:3 bound:1 convergent:1 identifiable:1 activity:6 strength:1 occur:1 constraint:4 precisely:2 encodes:1 aspect:1 attempting:1 performing:1 format:1 department:1 trumpington:1 according:2 gaba:1 pink:1 membrane:1 across:4 smaller:1 em:1 slightly:1 remain:1 making:2 hl:1 wellcome:1 ln:7 equation:6 remains:1 discus:1 eventually:4 mechanism:3 turn:1 ge:1 studying:1 available:1 indirectly:2 disagreement:1 simulating:1 recurrency:1 lns:25 subtracted:1 odor:52 robustness:1 top:1 graphical:1 exploit:1 rqt:1 graded:1 society:1 unchanged:1 objective:2 move:1 added:1 already:2 question:1 concentration:10 responds:1 diagonal:2 exhibit:2 affinity:4 gradient:5 distance:3 separate:1 street:1 strawberry:2 trivial:1 reason:2 fresh:1 assuming:1 code:2 reformulate:1 providing:1 minimizing:1 cq:2 equivalently:2 kingdom:2 unfortunately:1 negative:5 unknown:1 perform:3 upper:1 neuron:12 observation:10 descent:4 orthant:1 hinton:1 communication:1 rn:3 perturbation:1 intensity:1 namely:1 required:3 connection:9 learned:4 bar:1 below:2 pattern:6 sparsity:1 challenge:2 genetically:1 memory:1 natural:3 circumvent:1 representing:2 scheme:1 improve:1 technology:1 temporally:1 gabaergic:1 mediated:1 catch:1 breach:1 understanding:1 acknowledgement:1 segregated:1 relative:3 expect:1 permutation:2 antennal:13 versus:2 consistent:5 principle:1 corrupting:1 excitatory:4 changed:1 course:1 repeat:1 supported:1 free:1 transpose:1 aij:3 guide:1 allow:2 institute:1 wide:1 correspondingly:1 sparse:11 benefit:1 feedback:8 dimension:1 xn:1 superficially:1 rich:1 sensory:3 made:2 sj:1 ignore:2 stopfer:2 overcomes:1 ml:1 global:1 active:3 reveals:1 assumed:3 xi:10 neurology:1 spectrum:1 search:1 latent:1 why:1 reality:1 learn:2 nature:3 molecule:11 channel:1 mse:1 complex:1 necessarily:1 diag:4 did:2 pk:1 dense:6 main:1 sp:1 arrow:1 noise:14 repeated:1 body:9 x1:1 fig:7 neuronal:1 augmented:1 fashion:1 slow:1 axon:2 decoded:1 orns:3 bij:2 kin:1 specific:9 sensing:3 americana:1 pz:1 hinting:1 foucart:1 survival:1 evidence:3 exists:1 adding:1 conditioned:1 sparseness:1 interneuron:1 led:1 vise:1 likely:3 absorbed:1 lagrange:2 contained:1 partially:1 scalar:1 springer:1 corresponds:1 a8:1 determines:1 marked:1 presentation:3 identity:1 replace:1 content:1 experimentally:2 change:8 included:1 determined:1 averaging:1 called:1 specie:3 duality:4 experimental:2 zone:3 latter:1 sina:1 heaviside:1 tested:2 |
4,862 | 5,401 | Online Optimization for Max-Norm Regularization
Jie Shen
Dept. of Computer Science
Rutgers University
Piscataway, NJ 08854
Huan Xu
Dept. of Mech. Engineering
National Univ. of Singapore
Singapore 117575
Ping Li
Dept. of Statistics
Dept. of Computer Science
Rutgers University
[email protected]
[email protected]
[email protected]
Abstract
Max-norm regularizer has been extensively studied in the last decade as it promotes an effective low rank estimation of the underlying data. However, maxnorm regularized problems are typically formulated and solved in a batch manner,
which prevents it from processing big data due to possible memory bottleneck.
In this paper, we propose an online algorithm for solving max-norm regularized
problems that is scalable to large problems. Particularly, we consider the matrix
decomposition problem as an example, although our analysis can also be applied
in other problems such as matrix completion. The key technique in our algorithm
is to reformulate the max-norm into a matrix factorization form, consisting of a
basis component and a coefficients one. In this way, we can solve the optimal
basis and coefficients alternatively. We prove that the basis produced by our algorithm converges to a stationary point asymptotically. Experiments demonstrate
encouraging results for the effectiveness and robustness of our algorithm.
See the full paper at arXiv:1406.3190.
1 Introduction
In the last decade, estimating low rank matrices has attracted increasing attention in the machine
learning community owing to its successful applications in a wide range of domains including subspace clustering [13], collaborative filtering [9] and visual texture analysis [25], to name a few.
Suppose that we are given an observed data matrix Z of size p ? n, i.e., n observations in p ambient
dimensions, with each observation being i.i.d. sampled from some unknown distribution, we aim
to learn a prediction matrix X with a low rank structure to approximate Z. This problem, together
with its many variants, typically involves minimizing a weighted combination of the residual error
and matrix rank regularization term.
Generally speaking, it is intractable to optimize a matrix rank [15]. To tackle this challenge, researchers suggest alternative convex relaxations to the matrix rank. The two most widely used convex surrogates are the nuclear norm 1 [15] and the max-norm 2 [19]. In the work of [6], Cand`es et al.
proved that under mild conditions, solving a convex optimization problem consisting of a nuclear
norm regularization and a weighted ?1 norm penalty can exactly recover the low-rank component of
the underlying data even if a constant fraction of the entries are arbitrarily corrupted. In [20], Srebro
and Shraibman studied collaborative filtering and proved that the max-norm regularization formulation enjoyed a lower generalization error than the nuclear norm. Moreover, the max-norm was
shown to empirically outperform the nuclear norm in certain practical applications as well [11, 12].
To optimize a max-norm regularized problem, however, algorithms proposed in prior work [12, 16,
19] require to access all the data. In a large scale setting, the applicability of such batch optimiza1
2
Also known as the trace norm, the Ky-Fan n-norm and the Schatten 1-norm.
Also known as the ?2 -norm.
1
tion methods will be hindered by the memory bottleneck. In this paper, by utilizing the matrix
factorization form of the max-norm, we propose an online algorithm to solve max-norm regularized
problems. The main advantage of online algorithms is that the memory cost is independent from the
sample size, which makes online algorithms a good fit for the big data era [14, 18].
Specifically, we are interested in the max-norm regularized matrix decomposition (MRMD) problem. Assume that the observed data matrix Z can be decomposed into a low rank component X and
a sparse one E, we aim to simultaneously and accurately estimate the two components, by solving
the following convex program:
min
X,E
?1
1
2
?Z ? X ? E?F + ?X?2max + ?2 ?E?1,1 .
2
2
(1.1)
Here ???F denotes the Frobenius norm, ???max is the max-norm (which promotes low rank), ???1,1
is the ?1 norm of a matrix seen as a vector, and ?1 and ?2 are two non-negative parameters.
Our main contributions are two-folds: 1) We develop an online method to solve this MRMD
problems, making it scalable to big data. 2) We prove that the solutions produced by our algorithm
converge to a stationary point asymptotically.
1.1 Connection to Matrix Completion
While we mainly focus on the matrix decomposition problem, our method can be extended to the
matrix completion (MC) problem [4, 7] with max-norm regularization [5], which is another popular
topic in machine learning and signal processing. The MC problem can be described as follows:
min
X
1
?
2
2
?P? (Z ? X)?F + ?X?max ,
2
2
where ? is the set of indices of observed entries in Z and P? (M ) is the orthogonal projector onto
the span of matrices vanishing outside of ? so that the (i, j)-th entry of P? (M ) is equal to Mij if
(i, j) ? ? and zero otherwise. Interestingly, the max-norm regularized MC problem can be cast into
our framework. To see this, let us introduce an auxiliary matrix M , with Mij = C > 0 if (i, j) ? ?
and Mij = C1 otherwise. The following reformulated MC problem,
min
X,E
1
?
2
2
?Z ? X ? E?F + ?X?max + ?M ? E?1,1 ,
2
2
where ??? denotes the entry-wise product, is equivalent to our MRMD formulation (1.1). Furthermore, when C tends to infinity, the reformulated problem converges to the original MC problem.
1.2 Related Work
Here we discuss some relevant work in the literature. Most previous works on max-norm focused
on showing that the max-norm was empirically superior to the nuclear norm in a wide range of applications, such as collaborative filtering [19] and clustering [11]. Moreover, in [17], Salakhutdinov
and Srebro studied the influence of data distribution for the max-norm regularization and observed
good performance even when the data were sampled non-uniformly.
There are also studies which investigated the connection between the max-norm and the nuclear
norm. A comprehensive study on this problem, in the context of collaborative filtering, can be found
in [20], which established and compared the generalization bounds for the nuclear norm regularization and max-norm regularization, and showed that the generalization bound of the max-norm
regularization scheme is superior. More recently, Foygel et al. [9] attempted to unify the nuclear
norm and max-norm for gaining further insights on these two important regularization schemes.
There are few works to develop efficient algorithms for solving max-norm regularized problems, particularly large scale ones. Rennie and Srebro [16] devised a gradient-based optimization method and
empirically showed promising results on large collaborative filtering datasets. In [12], the authors
presented large scale optimization methods for max-norm constrained and max-norm regularized
problems with a theoretical guarantee to a stationary point. Nevertheless, all those methods were
formulated in a batch manner, which can be hindered by the memory bottleneck.
2
From a high level, the goal of this paper is similar to that of [8]. Motivated by the celebrated Robust
Principal Component Analysis (RPCA) problem [6, 23, 24], the authors of [8] developed an online
implementation for the nuclear-norm regularized matrix decomposition. Yet, since the max-norm
is a much more complicated mathematical entity (e.g., even the subgradient of the max-norm is not
completely characterized to the best of our knowledge), new techniques and insights are needed in
order to develop online methods for the max-norm regularization. For example, after taking the
max-norm with its matrix factorization form, the data are still coupled and we propose to convert
the problem to a constrained one for stochastic optimization.
The main technical contribution of this paper is to convert max-norm regularization to an appropriate
matrix factorization problem amenable to online implementation. Part of our proof ideas are inspired
by [14], which also studied online matrix factorization. In contrast to [14], our formulation contains
an additive sparse noise matrix, which enjoys the benefit of robustness to sparse contamination. Our
proof techniques are also different. For example, to prove the convergence of the dictionary and
to well define their problem, [14] needs to assume that the magnitude of the learned dictionary is
constrained. In contrast, in our setup we prove that the optimal basis is uniformly bounded, and
hence our problem is naturally well defined.
2
Problem Setup
We first introduce our notations. We use bold letters to denote vectors. The i-th row and j-th column
of a matrix M are denoted by m(i) and mj , respectively. The ?1 norm and ?2 norm of a vector v
are denoted by ?v?1 and ?v?2 , respectively. The ?2,? norm of a matrix is defined as the maximum
?2 row norm. Finally, the trace of a square matrix M is denoted as Tr(M ).
We are interested in developing an online algorithm for the MRMD Problem (1.1). By taking the
matrix factorization form of the max-norm [19]:
?X?max , min{?L?2,? ? ?R?2,? : X = LR? , L ? Rp?d , R ? Rn?d },
L,R
(2.1)
where d is the intrinsic dimension of the underlying data, we can rewrite Problem (1.1) into the
following equivalent form:
min
L,R,E
1
?1
?Z ? LRT ? E?2F + ?L?22,? ?R?22,? + ?2 ?E?1,1 .
2
2
(2.2)
Intuitively, the variable L corresponds to a basis and the variable R is a coefficients matrix with
each row corresponding to the coefficients. At a first sight, the problem can only be optimized
in a batch manner as the term ?R?22,? couples all the samples. In other words, to compute the
optimal coefficients of the i-th sample, we are required to compute the subgradient of ?R?2,? ,
which needs to access all the data. Fortunately, we have the following proposition that alleviates the
inter-dependency among samples.
Proposition 2.1. Problem (2.2) is equivalent to the following constrained program:
minimize
L,R,E
subject to
1
?1
?Z ? LRT ? E?2F + ?L?22,? + ?2 ?E?1,1 ,
2
2
?R?22,? = 1.
(2.3)
Proposition 2.1 states that our primal MRMD problem can be transformed to an equivalent constrained one. In the new formulation (2.3), the coefficients of each individual sample (i.e., a row of
the matrix R) is uniformly constrained. Thus, the samples are decoupled. Consequently, we can,
equipped with Proposition 2.1, rewrite the original problem in an online fashion, with each sample
being separately processed:
minimize
L,R,E
n
n
?
1?
?1
?ei ?1 ,
?zi ? Lri ? ei ?22 + ?L?22,? + ?2
2 i=1
2
i=1
subject to ?i ? 1, 2, . . . , n, ?ri ?22 ? 1,
3
where zi is the i-th observed sample, ri is the coefficients and ei is the sparse error. Combining the
first and third terms in the above equation, we have
n
?
? i , L, ri , ei ) + ?1 ?L?2 ,
minimize
?(z
2,?
L,R,E
2
(2.4)
i=1
subject to
?i ? 1, 2, . . . , n, ?ri ?22 ? 1,
where
? L, r, e) , 1 ?z ? Lr ? e?2 + ?2 ?e?1 .
?(z,
2
2
This is indeed equivalent to optimizing (i.e., minimizing) the empirical loss function:
n
1?
?1
fn (L) ,
?L?22,? ,
?(zi , L) +
n i=1
2n
where
?(z, L) =
min
r,e,?r?22 ?1
? L, r, e).
?(z,
(2.5)
(2.6)
(2.7)
When n goes to infinity, the empirical loss converges to the expected loss, defined as follows
f (L) = lim fn (L) = Ez [?(z, L)].
n?+?
3
(2.8)
Algorithm
We now present our online implementation to solve the MRMD problem. The detailed algorithm
is listed in Algorithm 1. Here we first briefly explain the underlying intuition: We optimize the
coefficients r, the sparse noise e and the basis L in an alternating manner, which is known to be a
successful strategy [8, 10, 14]. At the t-th iteration, given the basis Lt?1 , we can optimize over r
and e by examining the Karush Kuhn Tucker (KKT) conditions. To update the basis Lt , we then
optimize the following objective function:
1??
?1
?(zi , L, ri , ei ) + ?L?22,? ,
t i=1
2t
t
gt (L) ,
(3.1)
where {ri }ti=1 and {ei }ti=1 have been computed in previous iterations. It is easy to verify that
Eq. (3.1) is a surrogate function of the empirical cost function ft (L) defined in Eq. (2.6). The basis
Lt can be optimized by block coordinate decent, with Lt?1 being a warm start for efficiency.
4
Main Theoretical Results and Proof Outline
In this section we present our main theoretic result regarding the validity of the proposed algorithm.
We first discuss some necessary assumptions.
4.1
Assumptions
1. The observed data are i.i.d. generated from a distribution with compact support Z.
2. The surrogate functions gt (L) in Eq. (3.1) are strongly convex. Particularly, we assume that
the smallest eigenvalue of the positive semi-definite matrix 1t At defined in Algorithm 1 is
not smaller than some positive constant ?1 . Note that we can easily enforce this assumption
by adding a term ?21 ?L?2F to gt (L).
? L, r, e) is strongly convex
3. The minimizer for Problem (2.7) is unique. Notice that ?(z,
w.r.t. e and convex w.r.t. r. Hence, we can easily enforce this assumption by adding a term
??r?22 , where ? is a small positive constant.
4.2
Main Theorem
The following theorem is the main theoretical result of this work. It states that when t tends to
infinity, the basis Lt produced by Algorithm 1 converges to a stationary point.
Theorem 4.1 (Convergence to a stationary point of Lt ). Assume 1, 2 and 3. Given that the intrinsic
dimension of the underlying data is d, the optimal basis Lt produced by Algorithm 1 asymptotically
converges to a stationary point of Problem (2.8) when t tends to infinity.
4
Algorithm 1 Online Max-Norm Regularized Matrix Decomposition
Input: Z ? Rp?n (observed samples), parameters ?1 and ?2 , L0 ? Rp?d (initial basis), zero
matrices A0 ? Rd?d and B0 ? Rp?d
Output: optimal basis Lt
1: for t = 1 to n do
2:
Access the t-th sample zt .
3:
Compute the coefficient and noise:
? t , Lt?1 , r, e).
{rt , et } = arg min ?(z
(3.2)
r,e,?r?22 ?1
4:
5:
Compute the accumulation matrices At and Bt :
At
?
At?1 + rt r?
t ,
Bt
?
Bt?1 + (zt ? et ) r?
t .
Compute the basis Lt by optimizing the surrogate function (3.1):
1??
?1
?(zi , L, ri , ei ) + ?L?22,?
t i=1
2t
L
(
)
)
( ? )
?1
1 1 ( ?
= arg min
Tr L LAt ? Tr L Bt + ?L?22,? .
t
2
2t
L
t
Lt = arg min
(3.3)
6: end for
4.3
Proof Outline for Theorem 4.1
The essential tools for our analysis are from stochastic approximation [3] and asymptotic statistics [21]. There are three main steps in our proof:
(I) We show that the positive stochastic process gt (Lt ) defined in Eq. (3.1) converges almost surely.
(II) Then we prove that the empirical loss function, ft (Lt ) defined in Eq. (2.6) converges almost
surely to the same limit of its surrogate gt (Lt ). According to the central limit theorem, we can
expect that ft (Lt ) also converges almost surely to the expected loss f (Lt ) defined in Eq. (2.8),
implying that gt (Lt ) and f (Lt ) converge to the same limit.
(III) Finally, by taking a simple Taylor expansion, it justifies that the gradient of f (L) taking at Lt
vanishes as t tends to infinity, which concludes Theorem 4.1.
Theorem 4.2 (Convergence of the surrogate function gt (Lt )). The surrogate function gt (Lt ) we
defined in Eq. (3.1) converges almost surely, where Lt is the solution produced by Algorithm 1.
To establish the convergence of gt (Lt ), we verify that gt (Lt ) is a quasi-martingale [3] that converges
almost surely. To this end, we show that the expectation of the difference of gt+1 (Lt+1 ) and gt (Lt )
can be upper bounded by a family of functions ?(?, L) indexed by L ? L, where L is a compact
set. Then we show that the family of functions satisfy the hypotheses in the corollary of Donsker
Theorem [21] and thus can be uniformly upper bounded. Therefore, we conclude that gt (Lt ) is a
quasi-martingale and converges almost surely.
Now let us verify the hypotheses in the corollary of Donsker Theorem. First we prove that the index
set L is uniformly bounded.
Proposition 4.3. Let rt , et and Lt be the optimal solutions produced by Algorithm 1. Then,
1. The optimal solutions rt and et are uniformly bounded.
2. The matrices 1t At and 1t Bt are uniformly bounded.
5
3. There exists a compact set L, such that for all Lt produced by Algorithm 1, Lt ? L. That
is, there exists a positive constant Lmax that is uniform over t, such that for all t > 0,
?Lt ? ? Lmax .
To prove the third claim (which is required for our proof of convergence of gt (Lt )), we should prove
that for all t > 0, rt , et , 1t At and 1t Bt can be uniformly bounded, which can easily be verified.
Then, by utilizing the first order optimal condition of Problem (3.3), we can build an equation that
connects Lt with the four items we mentioned in the first and second claim. From Assumption 2, we
know that the nuclear norm of 1t At can be uniformly lower bounded. This property provides us the
way to show that Lt can be uniformly upper bounded. Note that in [8, 14], both papers assumed that
the dictionary (or basis) is uniformly bounded. In contrast, here in the third claim of Proposition 4.3,
we prove that such condition naturally holds in our problem.
Next, we show that the family of functions ?(z, L) is uniformly Lipschitz w.r.t. L.
Proposition 4.4. Let L ? L and denote the minimizer of ?(z, L, r, e) defined in (2.7) as:
1
{r? , e? } = arg min ?z ? Lr ? e?22 + ?2 ?e?1 .
r,e,?r?2 ?1 2
Then, the function ?(z, L) defined in Problem (2.7) is continuously differentiable and
?L ?(z, L) = (Lr? + e? ? z)r?T .
Furthermore, ?(z, ?) is uniformly Lipschitz and bounded.
By utilizing the corollary of Theorem 4.1 from [2], we can verify the differentiability of ?(z, L) and
the form of its gradient. As all of the items in the gradient are uniformly bounded (Assumption 1
and Proposition 4.3), we show that ?(z, L) is uniformly Lipschitz and bounded.
Based on Proposition 4.3 and 4.4, we verify that all the hypotheses in the corollary of Donsker
Theorem [21] are satisfied. This implies the convergence of gt (Lt ). We now move to step (II).
Theorem 4.5 (Convergence of f (Lt )). Let f (Lt ) be the expected loss function defined in Eq. (2.8)
and Lt is the solution produced by the Algorithm 1. Then,
1. gt (Lt ) ? ft (Lt ) converges almost surely to 0.
2. ft (Lt ) defined in Eq. (2.6) converges almost surely.
3. f (Lt ) converges almost surely to the same limit of ft (Lt ).
We apply Lemma 8 from [14] to prove the first claim. We denote the difference of gt (Lt ) and ft (Lt )
by bt . First we show that bt is uniformly Lipschitz. Then we show that the difference between Lt+1
and Lt is O( 1t ), making bt+1 ? bt be uniformly upper bounded by O( 1t ). Finally, we verify the
convergence of the summation of the serial { 1t bt }?
t=1 . Thus, Lemma 8 from [14] applies.
Proposition 4.6. Let {Lt } be the basis sequence produced by the Algorithm 1. Then,
1
?Lt+1 ? Lt ?F = O( ).
t
(4.1)
Proposition 4.6 can be proved by combining the strong convexity of gt (L) (Assumption 2 in Section 4.1) and the Lipschitz of gt (L); see the full paper for details.
Equipped with Proposition 4.6, we can verify that the difference of the sequence bt = gt (Lt ) ?
ft (Lt ) can be upper bounded by O( 1t ). The convergence of the summation of the serial { 1t bt }?
t=1
can be examined by the expectation convergence property of quasi-martingale gt (Lt ), stated in [3].
Applying the Lemma 8 from [14], we conclude that gt (Lt ) ? ft (Lt ) converges to zero a.s..
After the first claim of Theorem 4.5 being proved, the second claim follows immediately, as gt (Lt )
converges a.s. (Theorem 4.2). By the central limit theorem, the third claim can be verified.
According to Theorem 4.5, we can see that gt (Lt ) and f (Lt ) converge to the same limit a.s. Let
t tends to infinity, as Lt is uniformly bounded (Proposition 4.3), the term ?2t1 ?Lt ?22,? in gt (Lt )
vanishes. Thus gt (Lt ) becomes differentiable. On the other hand, we have the following proposition
about the gradient of f (L).
6
Proposition 4.7 (Gradient of f (L)). Let f (L) be the expected loss function defined in Eq. (2.8).
Then, f (L) is continuously differentiable and ?f (L) = Ez [?L ?(z, L)]. Moreover, ?f (L) is uniformly Lipschitz on L.
Thus, taking a first order Taylor expansion for f (Lt ) and gt (Lt ), we can show that the gradient of
f (Lt ) equals to that of gt (Lt ) when t tends to infinity. Since Lt is the minimizer for gt (L), we know
that the gradient of f (Lt ) vanishes. Therefore, we have proved Theorem 4.1.
5
Experiments
In this section, we report some simulation results on synthetic data to demonstrate the effectiveness
and robustness of our online max-norm regularized matrix decomposition (OMRMD) algorithm.
Data Generation. The simulation data are generated by following a similar procedure in [6]. The
clean data matrix X is produced by X = U V T , where U ? Rp?d and V ? Rn?d . The entries
of U and V are i.i.d. sampled from the Gaussian distribution N (0, 1). We introduce a parameter ?
to control the sparsity of the corruption matrix E, i.e., a ?-fraction of the entries are non-zero and
following an i.i.d. uniform distribution over [?1000, 1000]. Finally, the observation matrix Z is
produced by Z = X + E.
Evaluation Metric. Our goal is to estimate the correct subspace for the underlying data. Here, we
evaluate the fitness of our estimated subspace basis L and the ground truth basis U by the Expressed
Variance (EV) [22]:
Tr(LT U U T L)
EV(U, L) ,
.
Tr(U U T )
The values of EV range in [0, 1] and a higher EV value indicates a more accurate subspace recovery.
Other Settings. Through the experiments, we set the ambient dimension p = 400 and the total
number
? of samples n = 5000 unless otherwise specified. We fix the tunable parameter ?1 = ?2 =
1/ p, and use default parameters for all baseline algorithms we compare with. Each experiment is
repeated 10 times and we report the averaged EV as the result.
0.5
fraction of corruption
fraction of corruption
0.5
0.38
0.26
0.14
0.02
0.02
0.38
0.26
0.14
0.02
0.02
0.14
0.26
0.38
0.5
rank / ambient dimension
(a) OMRMD
0.14
0.26
0.38
0.5
rank / ambient dimension
(b) OR-PCA
Figure 1: Performance of subspace recovery under different rank and corruption fraction. Brighter
color means better performance.
We first study the effectiveness of the algorithm, measured by the EV value of its output after the last
sample, and compare it to the nuclear norm based online RPCA (OR-PCA) algorithm [8]. Specifically, we vary the intrinsic dimension d from 0.02p to 0.5p, with a step size 0.04p, and the corruption
fraction ? from 0.02 to 0.5, with a step size 0.04. The results are reported in Figure 1 where brighter
color means higher EV (hence better performance). We observe that for easier tasks (i.e., when
corruption and rank are low), both algorithms perform comparably. On the other hand, for more
difficult cases, OMRMD outperforms OR-PCA. This is possibly because the max-norm is a tighter
approximation to the matrix rank.
We next study the convergence of OMRMD by plotting the EV curve against the number of samples. Besides OR-PCA, we also add Principal Component Pursuit (PCP) [6] and an online PCA
7
1
0.8
0.8
0.6
OMRMD
OR?PCA
PCP
Online PCA
0.4
EV
EV
1
0.2
0
1
1000 2000 3000 4000 5000
Number of Samples
(a) ? = 0.01
1
0.8
0.8
0.6
0.6
EV
EV
1000 2000 3000 4000 5000
Number of Samples
(b) ? = 0.3
1
0.4
0.4
OMRMD
OR?PCA
PCP
Online PCA
0.2
0
1
OMRMD
OR?PCA
PCP
Online PCA
0.4
0.2
0
1
0.6
0.2
0
1000 2000 3000 4000 5000
Number of Samples
(c) ? = 0.5
OMRMD
OR?PCA
PCP
2
4
6
8
10
Number of Samples x 104
(d) p = 3000, d = 300, ? = 0.3
Figure 2: EV value against number of samples. p = 400 and d = 80 in (a) to (c).
algorithm [1] as baseline algorithms to compare with. The results are reported in Figure 2. As expected, PCP achieves the best performance since it is a batch method and needs to access all the data
throughout the algorithm. Online PCA degrades significantly even with low corruption (Figure 2a).
OMRMD is comparable to OR-PCA when the corruption is low (Figure 2a), and converges significantly faster when the corruption is high (Figure 2b and 2c). Indeed, this is true even with high
dimension and as many as 100, 000 samples (Figure 2d). This observation agrees with Figure 1,
and again suggests that for large corruption, max-norm may be a better fit than the nuclear norm.
Additional experimental results are available in the full paper.
6
Conclusion
In this paper, we developed an online algorithm for max-norm regularized matrix decomposition
problem. Using the matrix factorization form of the max-norm, we convert the original problem
to a constrained one which facilitates an online implementation for solving the original problem.
We established theoretical guarantees that the solutions will converge to a stationary point asymptotically. Moreover, we empirically compared our proposed algorithm with OR-PCA, which is a
recently proposed online algorithm for nuclear-norm based matrix decomposition. The simulation
results suggest that the proposed algorithm outperforms OR-PCA, in particular for harder task (i.e.,
when a large fraction of entries are corrupted). Our experiments, to an extent, empirically suggest
that the max-norm might be a tighter relaxation of the rank function compared to the nuclear norm.
Acknowledgments
The research of Jie Shen and Ping Li is partially supported by NSF-DMS-1444124, NSF-III1360971, NSF-Bigdata-1419210, ONR-N00014-13-1-0764, and AFOSR-FA9550-13-1-0137. Part
of the work of Jie Shen was conducted at Shanghai Jiao Tong University. The work of Huan Xu is
partially supported by the Ministry of Education of Singapore through AcRF Tier Two grant R-265000-443-112.
8
References
[1] Matej Artac, Matjaz Jogan, and Ales Leonardis. Incremental pca for on-line visual learning and recognition. In Pattern Recognition, 2002. Proceedings. 16th International Conference on, volume 3, pages
781?784. IEEE, 2002.
[2] J Fr?ed?eric Bonnans and Alexander Shapiro. Optimization problems with perturbations: A guided tour.
SIAM review, 40(2):228?264, 1998.
[3] L?eon Bottou. Online learning and stochastic approximations. On-line learning in neural networks, 17(9),
1998.
[4] Jian-Feng Cai, Emmanuel J Cand`es, and Zuowei Shen. A singular value thresholding algorithm for matrix
completion. SIAM Journal on Optimization, 20(4):1956?1982, 2010.
[5] Tony Cai and Wen-Xin Zhou. A max-norm constrained minimization approach to 1-bit matrix completion.
Journal of Machine Learning Research, 14:3619?3647, 2014.
[6] Emmanuel J Cand`es, Xiaodong Li, Yi Ma, and John Wright. Robust principal component analysis?
Journal of the ACM (JACM), 58(3):11, 2011.
[7] Emmanuel J Cand`es and Benjamin Recht. Exact matrix completion via convex optimization. Foundations
of Computational mathematics, 9(6):717?772, 2009.
[8] Jiashi Feng, Huan Xu, and Shuicheng Yan. Online robust pca via stochastic optimization. In Advances in
Neural Information Processing Systems, pages 404?412, 2013.
[9] Rina Foygel, Nathan Srebro, and Ruslan Salakhutdinov. Matrix reconstruction with the local max norm.
In NIPS, pages 944?952, 2012.
[10] Prateek Jain, Praneeth Netrapalli, and Sujay Sanghavi. Low-rank matrix completion using alternating
minimization. In Proceedings of the 45th annual ACM symposium on Symposium on theory of computing,
pages 665?674. ACM, 2013.
[11] Ali Jalali and Nathan Srebro. Clustering using max-norm constrained optimization. In ICML, 2012.
[12] Jason D Lee, Ben Recht, Ruslan Salakhutdinov, Nathan Srebro, and Joel A Tropp. Practical large-scale
optimization for max-norm regularization. In NIPS, pages 1297?1305, 2010.
[13] Guangcan Liu, Zhouchen Lin, Shuicheng Yan, Ju Sun, Yong Yu, and Yi Ma. Robust recovery of subspace
structures by low-rank representation. Pattern Analysis and Machine Intelligence, IEEE Transactions on,
35(1):171?184, 2013.
[14] Julien Mairal, Francis Bach, Jean Ponce, and Guillermo Sapiro. Online learning for matrix factorization
and sparse coding. The Journal of Machine Learning Research, 11:19?60, 2010.
[15] Benjamin Recht, Maryam Fazel, and Pablo A Parrilo. Guaranteed minimum-rank solutions of linear
matrix equations via nuclear norm minimization. SIAM review, 52(3):471?501, 2010.
[16] Jasson DM Rennie and Nathan Srebro. Fast maximum margin matrix factorization for collaborative
prediction. In Proceedings of the 22nd international conference on Machine learning, pages 713?719.
ACM, 2005.
[17] Ruslan Salakhutdinov and Nathan Srebro. Collaborative filtering in a non-uniform world: Learning with
the weighted trace norm. tc (X), 10:2, 2010.
[18] Shai Shalev-Shwartz, Yoram Singer, Nathan Srebro, and Andrew Cotter. Pegasos: Primal estimated subgradient solver for svm. Mathematical programming, 127(1):3?30, 2011.
[19] Nathan Srebro, Jason DM Rennie, and Tommi Jaakkola. Maximum-margin matrix factorization. In NIPS,
volume 17, pages 1329?1336, 2004.
[20] Nathan Srebro and Adi Shraibman. Rank, trace-norm and max-norm. In Learning Theory, pages 545?
560. Springer, 2005.
[21] Aad W Van der Vaart. Asymptotic statistics, volume 3. Cambridge university press, 2000.
[22] Huan Xu, Constantine Caramanis, and Shie Mannor. Principal component analysis with contaminated
data: The high dimensional case. In COLT, pages 490?502, 2010.
[23] Huan Xu, Constantine Caramanis, and Shie Mannor. Outlier-robust pca: the high-dimensional case.
Information Theory, IEEE Transactions on, 59(1):546?572, 2013.
[24] Huan Xu, Constantine Caramanis, and Sujay Sanghavi. Robust pca via outlier pursuit. IEEE Transactions
on Information Theory, 58(5):3047?3064, 2012.
[25] Zhengdong Zhang, Arvind Ganesh, Xiao Liang, and Yi Ma. Tilt: transform invariant low-rank textures.
International Journal of Computer Vision, 99(1):1?24, 2012.
9
| 5401 |@word mild:1 briefly:1 norm:70 nd:1 shuicheng:2 simulation:3 decomposition:8 tr:5 iii1360971:1 harder:1 initial:1 celebrated:1 contains:1 liu:1 mpexuh:1 interestingly:1 outperforms:2 yet:1 attracted:1 pcp:6 john:1 fn:2 additive:1 update:1 stationary:7 implying:1 intelligence:1 item:2 vanishing:1 fa9550:1 lr:4 provides:1 mannor:2 zhang:1 mathematical:2 symposium:2 prove:10 introduce:3 manner:4 inter:1 expected:5 indeed:2 cand:4 salakhutdinov:4 inspired:1 decomposed:1 encouraging:1 equipped:2 solver:1 increasing:1 becomes:1 estimating:1 underlying:6 moreover:4 bounded:16 notation:1 prateek:1 developed:2 shraibman:2 nj:1 guarantee:2 sapiro:1 ti:2 tackle:1 exactly:1 control:1 grant:1 positive:5 t1:1 engineering:1 local:1 tends:6 limit:6 era:1 might:1 studied:4 examined:1 suggests:1 factorization:10 range:3 averaged:1 fazel:1 practical:2 unique:1 acknowledgment:1 block:1 definite:1 procedure:1 mech:1 empirical:4 yan:2 significantly:2 jasson:1 word:1 suggest:3 onto:1 pegasos:1 context:1 influence:1 applying:1 optimize:5 equivalent:5 projector:1 accumulation:1 go:1 attention:1 convex:8 focused:1 shen:4 unify:1 recovery:3 immediately:1 insight:2 utilizing:3 nuclear:15 coordinate:1 suppose:1 exact:1 programming:1 hypothesis:3 jogan:1 recognition:2 particularly:3 observed:7 ft:9 solved:1 rina:1 sun:1 contamination:1 mentioned:1 intuition:1 vanishes:3 convexity:1 benjamin:2 solving:5 rewrite:2 ali:1 efficiency:1 eric:1 basis:18 completely:1 easily:3 caramanis:3 regularizer:1 univ:1 jiao:1 jain:1 effective:1 fast:1 outside:1 shalev:1 jean:1 widely:1 solve:4 rennie:3 otherwise:3 statistic:3 vaart:1 transform:1 online:27 advantage:1 eigenvalue:1 differentiable:3 sequence:2 cai:2 propose:3 reconstruction:1 maryam:1 product:1 fr:1 relevant:1 combining:2 alleviates:1 frobenius:1 ky:1 convergence:11 incremental:1 converges:17 ben:1 develop:3 completion:7 stat:1 andrew:1 measured:1 b0:1 eq:10 strong:1 netrapalli:1 auxiliary:1 involves:1 implies:1 tommi:1 kuhn:1 guided:1 correct:1 owing:1 stochastic:5 education:1 require:1 bonnans:1 fix:1 generalization:3 karush:1 proposition:15 tighter:2 summation:2 hold:1 ground:1 wright:1 claim:7 dictionary:3 vary:1 smallest:1 achieves:1 estimation:1 ruslan:3 rpca:2 agrees:1 tool:1 weighted:3 cotter:1 minimization:3 gaussian:1 sight:1 aim:2 zhou:1 jaakkola:1 corollary:4 l0:1 focus:1 ponce:1 rank:20 indicates:1 mainly:1 contrast:3 lri:1 baseline:2 typically:2 bt:13 a0:1 transformed:1 quasi:3 interested:2 lrt:2 arg:4 among:1 colt:1 denoted:3 constrained:9 equal:2 yu:1 icml:1 report:2 sanghavi:2 contaminated:1 few:2 wen:1 simultaneously:1 national:1 comprehensive:1 individual:1 fitness:1 consisting:2 connects:1 evaluation:1 joel:1 primal:2 amenable:1 accurate:1 ambient:4 necessary:1 huan:6 orthogonal:1 decoupled:1 indexed:1 unless:1 taylor:2 theoretical:4 column:1 applicability:1 cost:2 entry:7 tour:1 uniform:3 successful:2 examining:1 conducted:1 jiashi:1 reported:2 dependency:1 corrupted:2 synthetic:1 recht:3 ju:1 international:3 siam:3 lee:1 together:1 continuously:2 again:1 central:2 satisfied:1 possibly:1 li:3 parrilo:1 bold:1 coding:1 coefficient:9 satisfy:1 tion:1 jason:2 francis:1 start:1 recover:1 complicated:1 shai:1 guangcan:1 collaborative:7 contribution:2 square:1 minimize:3 variance:1 zhengdong:1 accurately:1 produced:11 comparably:1 mc:5 researcher:1 corruption:10 ping:2 explain:1 ed:1 against:2 tucker:1 dm:3 naturally:2 proof:6 couple:1 sampled:3 proved:5 tunable:1 popular:1 knowledge:1 lim:1 color:2 matej:1 higher:2 formulation:4 strongly:2 furthermore:2 hand:2 tropp:1 ei:7 ganesh:1 acrf:1 name:1 xiaodong:1 validity:1 verify:7 true:1 regularization:13 hence:3 alternating:2 outline:2 theoretic:1 demonstrate:2 wise:1 recently:2 superior:2 empirically:5 shanghai:1 tilt:1 volume:3 cambridge:1 enjoyed:1 rd:1 sujay:2 mathematics:1 zhouchen:1 access:4 gt:29 add:1 showed:2 optimizing:2 constantine:3 certain:1 n00014:1 onr:1 arbitrarily:1 yi:3 der:1 seen:1 ministry:1 fortunately:1 additional:1 zuowei:1 minimum:1 surely:9 converge:4 signal:1 semi:1 ii:2 full:3 ale:1 technical:1 faster:1 characterized:1 bach:1 arvind:1 lin:1 dept:4 devised:1 serial:2 promotes:2 prediction:2 scalable:2 variant:1 vision:1 expectation:2 rutgers:4 metric:1 arxiv:1 iteration:2 c1:1 separately:1 singular:1 jian:1 subject:3 facilitates:1 shie:2 effectiveness:3 iii:1 easy:1 decent:1 fit:2 zi:5 brighter:2 hindered:2 idea:1 regarding:1 praneeth:1 bottleneck:3 motivated:1 pca:20 penalty:1 reformulated:2 speaking:1 jie:3 generally:1 detailed:1 listed:1 extensively:1 processed:1 differentiability:1 shapiro:1 outperform:1 nsf:3 singapore:3 notice:1 estimated:2 key:1 four:1 nevertheless:1 clean:1 verified:2 asymptotically:4 relaxation:2 subgradient:3 fraction:7 convert:3 letter:1 almost:9 family:3 throughout:1 comparable:1 bit:1 bound:2 guaranteed:1 fan:1 fold:1 annual:1 infinity:7 ri:7 yong:1 nathan:8 span:1 min:10 developing:1 according:2 piscataway:1 combination:1 smaller:1 making:2 intuitively:1 outlier:2 invariant:1 tier:1 equation:3 foygel:2 discus:2 needed:1 know:2 singer:1 end:2 pursuit:2 available:1 apply:1 observe:1 appropriate:1 enforce:2 batch:5 robustness:3 alternative:1 rp:5 original:4 denotes:2 clustering:3 tony:1 lat:1 yoram:1 eon:1 emmanuel:3 build:1 establish:1 feng:2 objective:1 move:1 pingli:1 strategy:1 degrades:1 rt:5 jalali:1 surrogate:7 gradient:8 subspace:6 schatten:1 entity:1 topic:1 extent:1 besides:1 index:2 reformulate:1 minimizing:2 liang:1 setup:2 difficult:1 trace:4 negative:1 stated:1 implementation:4 zt:2 unknown:1 perform:1 upper:5 observation:4 datasets:1 extended:1 rn:2 perturbation:1 community:1 pablo:1 cast:1 required:2 specified:1 connection:2 optimized:2 learned:1 established:2 nu:1 nip:3 leonardis:1 pattern:2 ev:13 sparsity:1 challenge:1 program:2 max:47 memory:4 including:1 gaining:1 warm:1 regularized:12 residual:1 scheme:2 julien:1 mrmd:6 concludes:1 coupled:1 prior:1 sg:1 literature:1 review:2 asymptotic:2 afosr:1 loss:7 expect:1 generation:1 filtering:6 srebro:11 foundation:1 xiao:1 plotting:1 thresholding:1 row:4 lmax:2 guillermo:1 supported:2 last:3 enjoys:1 aad:1 wide:2 taking:5 sparse:6 benefit:1 van:1 curve:1 dimension:8 default:1 world:1 author:2 transaction:3 approximate:1 compact:3 kkt:1 maxnorm:1 mairal:1 conclude:2 assumed:1 shwartz:1 alternatively:1 decade:2 promising:1 learn:1 optimiza1:1 robust:6 mj:1 adi:1 expansion:2 investigated:1 bottou:1 domain:1 main:8 big:3 noise:3 repeated:1 xu:6 fashion:1 martingale:3 tong:1 donsker:3 third:4 theorem:17 showing:1 svm:1 intractable:1 intrinsic:3 essential:1 exists:2 adding:2 texture:2 magnitude:1 justifies:1 margin:2 easier:1 tc:1 lt:69 jacm:1 ez:2 visual:2 prevents:1 expressed:1 partially:2 applies:1 springer:1 mij:3 corresponds:1 minimizer:3 truth:1 acm:4 ma:3 goal:2 formulated:2 consequently:1 lipschitz:6 specifically:2 uniformly:19 principal:4 lemma:3 total:1 e:4 experimental:1 attempted:1 xin:1 support:1 alexander:1 bigdata:1 evaluate:1 |
4,863 | 5,402 | Finding a sparse vector in a subspace:
Linear sparsity using alternating directions
Qing Qu, Ju Sun, and John Wright
{qq2105, js4038, jw2966}@columbia.edu
Dept. of Electrical Engineering, Columbia University, New York City, NY, USA, 10027
Abstract
We consider the problem of recovering the sparsest vector in a subspace S ? Rp
with dim (S) = n. This problem can be considered a homogeneous variant of
the sparse recovery problem, and finds applications in sparse dictionary learning,
sparse PCA, and other problems in signal processing and machine learning. Simple
convex heuristics for this problem provably break down when ?
the fraction of
nonzero entries in the target sparse vector substantially exceeds 1/ n. In contrast,
we exhibit a relatively simple nonconvex approach based on alternating directions,
which provably succeeds even when the fraction of nonzero entries is ?(1). To
our knowledge, this is the first practical algorithm to achieve this linear scaling.
This result assumes a planted sparse model, in which the target sparse vector is
embedded in an otherwise random subspace. Empirically, our proposed algorithm
also succeeds in more challenging data models arising, e.g., from sparse dictionary
learning.
1
Introduction
Suppose we are given a linear subspace S of a high-dimensional space Rp , which contains a sparse
vector x0 6= 0. Given arbitrary basis of S, can we efficiently recover x0 ? Equivalently, provided a
matrix A ? R(p?n)?p , can we efficiently find a nonzero sparse vector x such that Ax = 0? In the
language of sparse approximation, can we solve
min kxk0 s.t. Ax = 0, x 6= 0
?
(1)
x
Variants of this problem have been studied in the context of applications to numerical linear algebra
[15], graphical model learning [27], nonrigid structure from motion [16], spectral estimation and
Prony?s problem [11], sparse PCA [29], blind source separation [28], dictionary learning [24],
graphical model learning [3], and sparse coding on manifolds [21].
However, in contrast to the standard sparse regression problem (Ax = b, b 6= 0), for which convex
relaxations perform nearly optimally for broad classes of designs A [14, 18], the computational
properties of problem (1) are not nearly as well understood. It has been known for several decades
that the basic formulation
min kxk0 , s.t. x ? S \ {0},
(2)
x
is NP-hard [15]. However, it is only recently that efficient computational surrogates with nontrivial
recovery guarantees have been discovered. In the context of sparse dictionary learning, Spielman et
al. [24] introduced a relaxation which replaces the nonconvex problem (2) with a sequence of linear
programs:
min kxk1 , s.t. xi = 1, x ? S, 1 ? i ? p,
(3)
x
and proved that when S is generated as a span of n random sparse vectors, with high probability
the relaxation
? recovers these vectors, provided the probability of an entry being nonzero is at most
? ? O (1/ n).
1
In a planted sparse model, in which S consists of a single sparse vector x0 embedded in a ?generic?
subspace, Hand et al. proved
that (3) also correctly recovers x0 , provided the fraction of nonzeros in
?
x0 scales as ? ? O (1/ n) [19].
?
Unfortunately, the results of [24, 19] are essentially sharp: when ? substantially exceeds 1/ n,
in both models the relaxation (3) provably breaks down. Moreover, the most natural semidefinite
programming relaxation of (1),
>
min kXk1 , s.t.
A A, X = 0, trace[X] = 1, X 0.
(4)
X
?
also breaks down at exactly the same threshold of ? ? 1/ n.1
?
One might naturally conjecture that this 1/ n threshold is simply an intrinsic price we must pay for
having an efficient algorithm, even in these random models. Some evidence towards this conjecture
might be borrowed from the surface similarity of (2)-(4) and sparse PCA [29]. In sparse PCA, there is
a substantial gap between what can be achieved with efficient algorithms and the information theoretic
?
optimum [10]. Is this also the case for recovering a sparse vector in a subspace? Is ? ? O (1/ n)
simply the best we can do with efficient, guaranteed algorithms?
Remarkably, this is not the case. Recently, Barak et al. introduced a new rounding technique for
sum-of-squares relaxations, and showed that the sparse vector x0 in the planted sparse model can be
recovered when p ? ? n2 and ? ? ?(1) [8]. It is perhaps surprising that this is possible at all with
a polynomial time algorithm. Unfortunately, the runtime of this approach is a high-degree polynomial
in p, and so for machine learning problems in which p is either a feature dimension or sample size,
this algorithm is of theoretical interest only. However, it raises an interesting algorithmic
question: Is
?
there a practical algorithm that provably recovers a sparse vector with ? 1/ n nonzeros from a
generic subspace S?
In this paper, we address this problem, under the following hypotheses: we assume the planted
sparse model, in which a target sparse vector x0 is embedded in an otherwise random n-dimensional
subspace of Rp . We allow x0 to have up to ?0 p nonzero entries, where ?0 is a constant. We provide
a relatively simple
algorithm which, with very high probability, exactly recovers x0 , provided that
p ? ? n4 log2 n .
Our algorithm is based on alternating directions, with two special twists. First, we introduce a
special data driven initialization, which seems to be important for achieving ? = ?(1). Second, our
theoretical results require a second, linear programming based rounding phase, which is similar to
[24]. Our core algorithm has very simple iterations, of linear complexity in the size of the data, and
hence should be scalable to moderate-to-large scale problems.
In addition to enjoying theoretical guarantees in a regime (? = ?(1)) that is out of the reach
of previous practical algorithms, it performs well in simulations ? succeeding empirically with
p ? ? (n log n). It also performs well empirically on more challenging data models, such as the
dictionary learning model, in?which the subspace of interest contains not one, but n target sparse
vectors. Breaking the O(1/ n) sparsity barrier with a practical algorithm is an important open
problem in the nascent literature on algorithmic guarantees for dictionary learning [5, 4, 2, 1]. We are
optimistic that the techniques introduced here will be applicable in this direction.
2
Problem Formulation and Global Optimality
We study the problem of recovering a sparse vector x0 6= 0 (up to scale), which is an element of a
known subspace S ? Rp of dimension n, provided an arbitrary orthonormal basis Y ? Rp?n for S.
Our starting point is the nonconvex formulation (2). Both the objective and constraint are nonconvex,
and hence not easy to optimize over. We relax (2) by replacing the `0 norm with the `1 norm. For the
constraint x 6= 0, which is necessary to avoid a trivial solution, we force x to live on the unit sphere
kxk2 = 1, giving
min kxk1 , s.t. x ? S, kxk2 = 1.
(5)
x
1
This breakdown behavior is again in sharp contrast to the standard sparse approximation problem (with
b 6= 0), in which it is possible to handle very large fractions of nonzeros (say, ? = ?(1/ log n), or even
? = ?(1)) using a very simple `1 relaxation [14, 18]
2
This formulation is still nonconvex, and so we should not expect to obtain an efficient algorithm
that can solve it globally for general inputs S. Nevertheless, the geometry of the sphere is benign
enough that for well-structured inputs it actually will be possible to give algorithms that find the
global optimum of this problem.
The formulation (5) can be contrasted with (3), in which we optimize the `1 norm subject to the
constraint kxk? = 1. Because k?k? is polyhedral, that formulation immediately yields a sequence
of linear programs. This is very convenient for computation
and analysis, but suffers from the
?
aforementioned breakdown behavior around kx0 k0 ? p/ n.
In contrast, the sphere kxk2 = 1 is a more complicated geometric constraint, but will allow much
larger numbers of nonzeros in x0 . For example, if we consider the global optimizer of a variant of
(5):
min kYqk1 ,
q?Rn
s.t.
kqk2 = 1,
(6)
under the planted sparse model (detailed below), e1 is the unique to (6) with very high probability:
?
Theorem 2.1 (`1 /`2 recovery, planted sparse model). There exists a constant ?0 ? (1/ n, 1/2)
such that if the subspace S follows the planted sparse model
S = span (x0 , g1 , . . . , gn?1 ) ? Rp ,
(7)
x0 ?i.i.d. ?1?p Ber(?),
with gi ?i.i.d. N (0, 1/p), and
with x0 , g1 , . . . , gn?1 mutually independent and
?
1/ n < ? < ?0 , then ?e0 are the only global minimizers to (6) if Y = [x0 , g1 , . . . , gn?1 ], provided
p ? ? (n log n).
Hence, if we could find the global optimizer of (6), we would be able to recover x0 whose number of
nonzero entries is quite large ? even linear in the dimension p (? = ?(1)). On the other hand, it is
not obvious that this should be possible: (6) is nonconvex. In the next section, we will describe a
simple heuristic algorithm for (a near approximation of) the `1 /`2 problem (6), which guarantees
to find a stationary point. More surprisingly, we will then prove that for a class of random problem
instances, this algorithm, plus an auxiliary rounding technique, actually recovers the global optimum
? the target sparse vector x0 . The proof requires a detailed probabilistic analysis, which is sketched in
Section 4.2.
Before continuing, it is worth noting that the formulation (5) is in no way novel ? see, e.g., the work
of [28] in blind source separation for precedent. However, the novelty originates from our algorithms
and subsequent analysis.
3
Algorithm based on Alternating Direction Method (ADM)
To develop an algorithm for solving (6), we work with the orthonormal basis Y ? Rp?n for S. For
numerical purposes, and also for coping with noise in practical application, it is useful to consider a
slight relaxation of (6), in which we introduce an auxiliary variable x ? Yq:
1
2
(8)
min kYq ? xk2 + ? kxk1 , s.t. kqk2 = 1,
q,x 2
Here, ? > 0 is a penalty parameter. It is not difficult to see that this problem is equivalent to
minimizing the Huber m-estimator over Yq. This relaxation makes it possible to apply alternating
direction method to this problem, which, starting from some initial point q(0) , alternates between
optimizing with respect to x and optimizing with respect to q:
2
1
x(k+1) = arg min
Yq(k) ? x
+ ? kxk1 ,
(9)
2
2
x
2
1
(10)
q(k+1) = arg min
Yq ? x(k+1)
s.t. kqk2 = 1.
2
2
q
Both (9) and (10) have simple closed form solutions:
x(k+1) = S? [Yq(k) ],
Y> x(k+1)
q(k+1) =
Y> x(k+1)
,
2
3
(11)
Algorithm 1 Nonconvex ADM
Input: A matrix Y ? Rp?n with Y> Y = I, initialization q(0) , threshold ? > 0.
? 0 = Yq(k)
Output: The recovered sparse vector x
1: Set k = 0,
2: while not converged do
3:
x(k+1) = S? [Yq(k) ],
> (k+1)
4:
q(k+1) = YY> xx(k+1) ,
k
k2
5:
Set k = k + 1.
6: end while
where S? [x] = sign(x) max {|x| ? ?, 0} is the soft-thresholding operator. The proposed ADM
algorithm is summarized in Algorithm 1.
For general input Y and initialization q(0) , Algorithm 1 is guaranteed to produce a stationary point
of problem (8). This is a consequence of recent general analyses of alternating direction methods
for nonsmooth and nonconvex problems ? see [6, 7]. However, if our goal is to recover the sparsest
vector x0 , some additional tricks are needed.
Initialization. Because the problem (6) is nonconvex, an arbitrary or random initialization is
unlikely to produce a global minimizer.2 Therefore, good initializations are critical for the proposed
ADM algorithm to succeed. For this purpose, we suggest to use every normalized row of Y as
initializations for q, and solve a sequence of p nonconvex programs (6) by the ADM algorithm.
To get an intuition of why our initialization works, recall the planted sparse model: S =
p?n
?? | g
. Suppose we take a row zi of Z,
span(x0 , g1 , . . . , gn?1 ). Write Z = [x0 | g1 | ??
n?1 ] ? R
in which x0 (i) is nonzero, then x0 (i) = ? 1/ ?p . Meanwhile, the entries of g1 (i), . . . gn?1 (i)
?
are all N (0, 1/p), and so have size about 1/ p. Hence, when ? is not too large, x0 (i) will be
somewhat bigger than most of the other entries in zi . Put another way, zi is biased towards the first
standard basis vector e1 .
Now, under our probabilistic assumptions, Z is very well conditioned: Z> Z ? I.3 Using, e.g.,
? for S of the form
Gram-Schmidt, we can find a basis Y
? = ZR,
Y
(12)
where R is upper triangular, and R is itself well-conditioned: R ? I. Since the i-th row of Z is
? i is also biased in the direction
biased in the direction of e1 and R is well-conditioned, the i-th row y
of e1 .
? ? = x0 . Since Ze1 = x0 , we have
We know that the global optimizer q? should satisfy Yq
q? = R?1 e1 ? e1 . Here, the approximation comes from R ? I. Hence, for this particular choice
of Y, described in (12), the i-th row is biased in the direction of the global optimizer. This is what
makes the rows of Y a particularly effective choice for initialization.
?
What if we are handed some other basis Y = YU,
where U is an orthogonal matrix? Suppose
? then it is easy to check that, with input matrix
q? is a global optimizer to (6) with input matrix Y,
Y, U> q? is also a global optimizer to (6), which implies that our initialization is invariant to any
rotation of the basis. Hence, even if we are handed an arbitrary basis for S, the i-th row is still biased
in the direction of the global optimizer.
? denote the output of Algorithm 1. We will prove that with our particular initializaRounding. Let q
tion and an appropriate choice of ?, the solution of our ADM algorithm falls within a certain radius
of the globally optimal solution q? to (6). To recover q? , or equivalently to recover the sparse vector
x0 = Yq? , we solve the linear program
min kYqk1
q
s.t.
hr, qi = 1,
(13)
2
More precisely, in our models, random initialization does work, but only when the subspace dimension n is
extremely low compared to the ambient dimension p.
3
This is the common heuristic that ?tall random matrices are well conditioned? [25].
4
? . We will prove that if r is close enough to q? , then this relaxation exactly recovers q? ,
with r = q
and hence x0 .
4
4.1
Analysis
Main Results
In this section, we describe our main theoretical result, which shows that with high probability, the
algorithm described in the previous section succeeds.
Theorem 4.1. Suppose that S satisfies the planted sparse model, and let Y be an arbitrary basis?
for
S. Let y1 . . . yp ? Rn denote the (transposes of) the rows of Y. Apply Algorithm 1 with ? = 1/ p,
?1, . . . , q
? p . Solve the linear program
using initializations q(0) = y1 , . . . , yp , to produce outputs q
?1, . . . , q
? p , to produce q
?1, . . . , q
? p . Set i? ? arg mini kY?
(13) with r = q
qi k0 . Then
Y?
qi? = ?x0 ,
(14)
for some ? 6= 0, with overwhelming probability, provided
p > Cn4 log2 n,
and
1
? ? ? ? ?0 .
4 n
(15)
Here, C and ?0 > 0 are universal constants.
We can see that the result in Theorem 4.1 is suboptimal compared to the global optimality condition
and Barak et al.?s result in the sense of the sampling complexity that we require p ? Cn4 log2 n.
While for the global optimality condition, we only need p > Cn to guarantee a global optimal
solution exists with high probability. For Barak et al.?s result, we need p > Cn2 . Nonetheless,
compared to Barak et al., we believe this is the first practical and efficient method that is guaranteed
to achieve ? ? O(1) rate. The lower bound on ? in Theorem 4.1 is mostly for convenience in the
proof; in fact,
? the LP rounding stage of our algorithm already succeeds with high probability when
? ? O (1/ n).
4.2
A Sketch of Analysis
The proof of our main result requires rather detailed technical analysis of the iteration-by-iteration
properties of Algorithm 1. In this subsection, we briefly sketch the main ideas. For detailed proofs,
please see the technical supplement to this paper.
As noted in Section 3, the ADM algorithm is invariant to change of basis. So, we can assume without
? = ZR defined in that section. In
loss of generality that we are working with the particular basis Y
order to further streamline the presentation, we are going to sketch the proof under the assumption
that
Y = [x0 | g1 | ? ? ? | gn?1 ],
(16)
? This may seem plausible, but when p is large Y is already
rather than the orthogonalized version Y.
? In fact, in our proof, we simply carry through the
nearly orthogonal, and hence Y is very close to Y.
?
argument for Y, and then note that Y and Y are close enough that all steps of the proof still hold
? With that noted, let y1 , . . . , yp ? Rn denote the transposes of the rows of Y,
with Y replaced by Y.
and note that these are independent random vectors. From (11), we can see one step of the ADM
algorithm takes the form:
Pp
1
i
i > (k)
q ]
i=1 y S? [ y
p
(k+1)
.
q
=
(17)
P
1 p
>
p i=1 yi S? [(yi ) q(k) ]
2
This is a very favorable form for analysis: if q is viewed as fixed, the term in the numerator is a sum
of p independent random vectors. To this end, we define a vector valued random process Q(q) on
q ? Sn?1 , via
p
>
1X i
Q(q) =
y S? [ yi q].
(18)
p i=1
5
We study the behavior of the iteration (17) through the random process Q(q). We wish to show
that w.h.p. in our choice of Y, q(k) converges to (?e1 ), so that the algorithm successfully retrieves
the sparse vector x0 = Ye1 . Thus, we hope that in general,
Q(q)
is more concentrated on the first
q1
coordinate than q. Let us partition the vector q as q =
, with q1 ? R and q2 ? Rn?1 , and
q2
Q1 (q)
correspondingly partition Q(q) =
, where
Q2 (q)
p
Q1 (q) =
h > i
1X
x0i S? yi q
p i=1
p
Q2 (q) =
and
1 X i h i > i
g S? y
q .
p i=1
(19)
The inner product of Q(q)/ kQ(q)k2 and e1 is strictly larger than the inner product of q and e1 if
and only if
kQ2 (q)k2
|Q1 (q)|
>
.
(20)
|q1 |
kq2 k2
In the appendix, we show that with high probability, this inequality holds uniformly over a significant
portion of the sphere, so the algorithm moves in the correct direction. To complete the proof of
Theorem 4.1, we combine the following observations:
1. Algorithm 1 converges.
?
2. Rounding succeeds when |r1 | > 2 ?. With high probability, the linear programming based
rounding (13) will produce ?x0 , up
? to scale, whenever it is provided with an input r whose first
coordinate has magnitude at least 2 ?.
?
3. No jumps away from the caps. With high probability, for all q such that |q|1 > C? ?,
?
|Q1 (q)|
q
? 2 ?.
(21)
2
|Q1 (q)2 | + kQ2 (q)k2
4. Uniform progress away from the equator. With high probability, for every q such that
?
|q1 | ? C? ?, the bound
c
|Q1 (q)| kQ2 (q)k2
?
>
|q1 |
kqk2
np
(k)
holds. This implies that if at any iteration k of the algorithm, |q1 | >
?
0
(k0 )
eventually obtain a point q(k ) , k 0 > k, for which |q1 | > C? ?.4
?1 ,
2 ?n
?1
2 ?n
?
(22)
the algorithm will
5. Location of stationary points. Steps 1, 3 and 4 above imply that if Algorithm 1 ever obtains a
?
?
(k)
point q(k) with |q1 | > 2?1?n , it will converge to a point q? with q?1 > C? ?, provided 2?1?n < 2 ?
(i.e., ? > 4?1 n ).
(0)
6. Good initializers. With high probability, at least one of the initializers q(0) satisfies |q1 | >
?1 .
2 ?n
Taken together, these claims imply that from at least one of the initializers q(0) , the ADM algorithm
? which is accurate enough for LP rounding to exactly return x0 , up to scale.
will produce an output q
As x0 is the sparsest nonzero vector in the subspace S with overwhelming probability, it will be
selected as Yqi? , and hence produced by the algorithm.
5
Experimental Results
In this section, we show the performance of the proposed ADM algorithm on both synthetic and real
datasets. On the synthetic dataset, we show the phase transition of our algorithm on both the planted
sparse vector and dictionary learning models; for the real dataset, we demonstrate how seeking sparse
vectors can help discover interesting patterns.
4
In fact, the rate of progress guaranteed in (22) can be used to bound the complexity of the algorithm; we do
not dwell on this here.
6
5.1
Phase Transition on Synthetic Data
For the planted sparse model, for each pair of (k, p), we generate the n dimensional subspace
S ? Rp by a k sparse vector x0 with nonzero entries equal to 1 and a random Gaussian matrix
i.i.d.
G ? Rp?(n?1) with Gij ? N (0, 1/p), so that the basis Yof the subspace S can be constructed
by Y = GS ([x0 , G]) U, where GS (?) denotes the Gram-Schmidt orthonormalization operator and
U ? Rn?n is an arbitrary orthogonal matrix. We fix the?relationship between n and p as p = 5n log n,
and set the regularization parameter in (8) as ? = 1/ p. We use all the normalized rows of Y as
initializations of q for the proposed ADM algorithm, and
for 5000 iterations. We
run every program
x0
assume the proposed method to be success whenever
kx0 k ? Yq
? for at least one of the p
2
2
programs, for some error tolerance = 10?3 . For each pair of (k, p), we repeat the simulation for 5
times.
Figure 1: Phase transition for the planted sparse model (left) and dictionary learning (right) using the ADM
algorithm, with fixed relationship between p and n: p = 5n log n. White indicates success and black indicates
failure.
Second, we consider the same dictionary learning model as in [24]. Specifically, the observation is
assumed to be Y = A0 X0 where A0 is a square, invertible matrix, and X0 a n ? p sparse matrix.
Since A0 is invertible, the row space of Y is the same as that of X0 . For each pair of (k, n), we
>
generate X0 = [x1 , ? ? ? , xn ] , where each vector xi ? Rp is k-sparse with every nonzero
>entry
following i.i.d. Gaussian distribution, and construct the observation by Y> = GS X>
0 U . We
repeat the same experiment as for the planted sparse model presented above. The only difference is
that we assume the proposed method to be success as long as one sparse row of X0 is recovered by
those p programs.
Fig. 1 shows the phase transition between the sparsity level k = ?p and p for both models. It seems
clear for both problems our algorithm can work well into (beyond) the linear regime in sparsity level.
Hence for the planted sparse model, to close the gap between our algorithm and practice is one future
direction. Also, how to extend our analysis for dictionary learning is another interesting direction.
5.2
Exploratory Experiments on Faces
It is well known in computer vision convex objects only subject to illumination changes produce
image collection that can be well approximated by low-dimensional space in raw-pixel space [9].
We will play with face subspaces here. First, we extract face images of one person (65 images)
under different illumination conditions. Then we apply robust principal component analysis [12]
to the data and get a low dimensional subspace of dimension 10, i.e., the basis Y ? R32256?10 . We
apply the ADM algorithm to find the sparsest element in such a subspace, by randomly selecting
10% rows as initializations for q. We judge the sparsity in a `1 /`2 sense, that is, the sparsest vector
? 0 = Yq? should produce the smallest kYqk1 / kYqk2 among all results. Once some sparse vectors
x
are found, we project the subspace onto orthogonal complement of the sparse vectors already found,
and continue the seeking process in the projected subspace. Fig. 2 shows the first four sparse vectors
we get from the data. We can see they correspond well to different extreme illumination conditions.
Second, we manually select ten different persons? faces under the normal lighting condition. Again,
the dimension of the subspace is 10 and Y ? R32256?10 . We repeat the same experiment as stated
above. Fig. 3 shows four sparse vectors we get from the data. Interestingly, the sparse vectors roughly
7
Figure 2: Four sparse vectors extracted by the ADM algorithm for one person in the Yale B database under
different illuminations.
correspond to differences of face images concentrated around facial parts that different people tend to
differ from each other.
Figure 3: Four sparse vectors extracted by the ADM algorithm for 10 persons in the Yale B database under
normal illuminations.
In sum, our algorithm seems to find useful sparse vectors for potential applications, like peculiar
discovery in first setting, and locating differences in second setting. Netherless, the main goal of this
experiment is to invite readers to think about similar pattern discovery problems that might be cast as
searching for a sparse vector in a subspace. The experiment also demonstrates in a concrete way the
practicality of our algorithm, both in handling data sets of realistic size and in producing attractive
results even outside of the (idealized) planted sparse model that we adopt for analysis.
6
Discussion
The random models we assume for the subspace can be easily extended to other random models,
particularly for dictionary learning. Moreover we believe the algorithm paradigm works far beyond
the idealized models, as our preliminary experiments on face data have clearly shown. For the
particular planted sparse model, the performance gap in terms of (p, n, ?) between the empirical
simulation and our result is likely due to analysis itself. Advanced techniques to bound the empirical
process, such as decoupling [17] techniques, can be deployed in place of our crude union bound
to cover all iterates. Our algorithmic paradigm as a whole sits well in the recent surge of research
endeavors in provable and practical nonconvex approaches towards many problems of interest, often
in large-scale setting [13, 22, 20, 23, 26]. We believe this line of research will become increasingly
important in theory and practice. On the application side, the potential of seeking sparse/structured
element in a subspace seems largely unexplored, despite the cases we mentioned at the start. We
hope this work can invite more application ideas.
References
[1] AGARWAL , A., A NANDKUMAR , A., JAIN , P., N ETRAPALLI , P., AND TANDON , R. Learning sparsely
used overcomplete dictionaries via alternating minimization. arXiv preprint arXiv:1310.7991 (2013).
[2] AGARWAL , A., A NANDKUMAR , A., AND N ETRAPALLI , P. Exact recovery of sparsely used overcomplete
dictionaries. arXiv preprint arXiv:1309.1952 (2013).
[3] A NANDKUMAR , A., H SU , D., JANZAMIN , M., AND K AKADE , S. M. When are overcomplete topic
models identifiable? uniqueness of tensor tucker decompositions with structured sparsity. In Advances in
Neural Information Processing Systems (2013), pp. 1986?1994.
8
[4] A RORA , S., B HASKARA , A., G E , R., AND M A , T. More algorithms for provable dictionary learning.
arXiv preprint arXiv:1401.0579 (2014).
[5] A RORA , S., G E , R., AND M OITRA , A. New algorithms for learning incoherent and overcomplete
dictionaries. arXiv preprint arXiv:1308.6273 (2013).
[6] ATTOUCH , H., B OLTE , J., R EDONT, P., AND S OUBEYRAN , A. Proximal alternating minimization and
projection methods for nonconvex problems: An approach based on the kurdyka-lojasiewicz inequality.
Mathematics of Operations Research 35, 2 (2010), 438?457.
[7] ATTOUCH , H., B OLTE , J., AND S VAITER , B. F. Convergence of descent methods for semi-algebraic and
tame problems: proximal algorithms, forward?backward splitting, and regularized gauss?seidel methods.
Mathematical Programming 137, 1-2 (2013), 91?129.
[8] BARAK , B., K ELNER , J.,
arXiv:1312.6652 (2013).
AND
S TEURER , D. Rounding sum-of-squares relaxations. arXiv preprint
[9] BASRI , R., AND JACOBS , D. W. Lambertian reflectance and linear subspaces. Pattern Analysis and
Machine Intelligence, IEEE Transactions on 25, 2 (2003), 218?233.
[10] B ERTHET, Q., AND R IGOLLET, P. Complexity theoretic lower bounds for sparse principal component
detection. In Conference on Learning Theory (2013), pp. 1046?1066.
? , L. On approximation of functions by exponential sums. Applied and
[11] B EYLKIN , G., AND M ONZ ON
Computational Harmonic Analysis 19, 1 (2005), 17?48.
[12] C AND E` S , E., L I , X., M A , Y., AND W RIGHT, J. Robust principal component analysis? Journal of the
ACM 58, 3 (May 2011).
[13] C AND E` S , E. J., L I , X., AND S OLTANOLKOTABI , M. Phase retrieval via wirtinger flow: Theory and
algorithms. arXiv preprint arXiv:1407.1065 (2014).
[14] C ANDES , E. J., AND TAO , T. Decoding by linear programming. Information Theory, IEEE Transactions
on 51, 12 (2005), 4203?4215.
[15] C OLEMAN , T. F., AND P OTHEN , A. The null space problem i. complexity. SIAM Journal on Algebraic
Discrete Methods 7, 4 (1986), 527?537.
[16] DAI , Y., L I , H., AND H E , M. A simple prior-free method for non-rigid structure-from-motion factorization.
In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on (2012), IEEE, pp. 2018?
2025.
[17] D E LA P ENA , V., AND G IN E? , E. Decoupling: from dependence to independence. Springer, 1999.
[18] D ONOHO , D. L. For most large underdetermined systems of linear equations the minimal `1 -norm solution
is also the sparsest solution. Communications on pure and applied mathematics 59, 6 (2006), 797?829.
[19] H AND , P., AND D EMANET, L.
arXiv:1310.1654 (2013).
Recovering the sparsest element in a subspace. arXiv preprint
[20] H ARDT, M. On the provable convergence of alternating minimization for matrix completion. arXiv
preprint arXiv:1312.0925 (2013).
[21] H O , J., X IE , Y., AND V EMURI , B. On a nonlinear generalization of sparse coding and dictionary learning.
In Proceedings of The 30th International Conference on Machine Learning (2013), pp. 1480?1488.
[22] JAIN , P., N ETRAPALLI , P., AND S ANGHAVI , S. Low-rank matrix completion using alternating minimization. In Proceedings of the 45th annual ACM symposium on Symposium on theory of computing (2013),
ACM, pp. 665?674.
[23] N ETRAPALLI , P., JAIN , P., AND S ANGHAVI , S. Phase retrieval using alternating minimization. In
Advances in Neural Information Processing Systems (2013), pp. 2796?2804.
[24] S PIELMAN , D. A., WANG , H., AND W RIGHT, J. Exact recovery of sparsely-used dictionaries. In
Proceedings of the 25th Annual Conference on Learning Theory (2012).
[25] V ERSHYNIN , R. Introduction to the non-asymptotic analysis of random matrices. arXiv preprint
arXiv:1011.3027 (2010).
[26] Y I , X., C ARAMANIS , C., AND S ANGHAVI , S. Alternating minimization for mixed linear regression.
arXiv preprint arXiv:1310.3745 (2013).
[27] Z HAO , Y.-B., AND F UKUSHIMA , M. Rank-one solutions for homogeneous linear matrix equations over
the positive semidefinite cone. Applied Mathematics and Computation 219, 10 (2013), 5569?5583.
[28] Z IBULEVSKY, M., AND P EARLMUTTER , B. A. Blind source separation by sparse decomposition in a
signal dictionary. Neural computation 13, 4 (2001), 863?882.
[29] Z OU , H., H ASTIE , T., AND T IBSHIRANI , R. Sparse principal component analysis. Journal of computational and graphical statistics 15, 2 (2006), 265?286.
9
| 5402 |@word version:1 briefly:1 polynomial:2 seems:4 norm:4 open:1 simulation:3 decomposition:2 jacob:1 q1:15 carry:1 initial:1 contains:2 selecting:1 interestingly:1 kx0:2 recovered:3 surprising:1 must:1 john:1 numerical:2 subsequent:1 partition:2 benign:1 realistic:1 succeeding:1 stationary:3 intelligence:1 selected:1 core:1 iterates:1 location:1 sits:1 mathematical:1 constructed:1 become:1 symposium:2 consists:1 prove:3 combine:1 polyhedral:1 introduce:2 x0:41 huber:1 roughly:1 behavior:3 surge:1 globally:2 overwhelming:2 provided:9 xx:1 moreover:2 discover:1 project:1 null:1 what:3 substantially:2 q2:4 finding:1 guarantee:5 every:4 unexplored:1 runtime:1 exactly:4 k2:6 demonstrates:1 unit:1 originates:1 producing:1 before:1 positive:1 engineering:1 understood:1 consequence:1 despite:1 haskara:1 might:3 plus:1 black:1 initialization:14 studied:1 challenging:2 ye1:1 factorization:1 practical:7 unique:1 practice:2 union:1 orthonormalization:1 coping:1 universal:1 empirical:2 convenient:1 projection:1 suggest:1 get:4 convenience:1 close:4 onto:1 operator:2 put:1 context:2 live:1 optimize:2 equivalent:1 starting:2 convex:3 recovery:5 immediately:1 splitting:1 pure:1 yqi:1 estimator:1 orthonormal:2 handle:1 exploratory:1 coordinate:2 searching:1 target:5 suppose:4 play:1 tandon:1 exact:2 programming:5 homogeneous:2 hypothesis:1 trick:1 element:4 approximated:1 particularly:2 recognition:1 breakdown:2 sparsely:3 database:2 kxk1:5 preprint:10 electrical:1 wang:1 sun:1 andes:1 substantial:1 intuition:1 mentioned:1 tame:1 complexity:5 raise:1 solving:1 algebra:1 basis:13 easily:1 k0:3 retrieves:1 jain:3 describe:2 effective:1 outside:1 whose:2 heuristic:3 larger:2 solve:5 cvpr:1 quite:1 say:1 otherwise:2 relax:1 triangular:1 plausible:1 valued:1 gi:1 g1:7 statistic:1 think:1 itself:2 sequence:3 product:2 achieve:2 ky:1 convergence:2 optimum:3 r1:1 produce:8 converges:2 object:1 tall:1 help:1 develop:1 completion:2 x0i:1 progress:2 borrowed:1 recovering:4 auxiliary:2 streamline:1 come:1 implies:2 judge:1 differ:1 direction:14 radius:1 correct:1 require:2 fix:1 generalization:1 preliminary:1 underdetermined:1 strictly:1 hold:3 around:2 considered:1 wright:1 normal:2 algorithmic:3 claim:1 dictionary:18 optimizer:7 smallest:1 xk2:1 adopt:1 purpose:2 uniqueness:1 estimation:1 favorable:1 akade:1 applicable:1 city:1 successfully:1 hope:2 minimization:6 clearly:1 gaussian:2 rather:2 avoid:1 ax:3 rank:2 check:1 indicates:2 contrast:4 sense:2 dim:1 minimizers:1 rigid:1 unlikely:1 a0:3 going:1 tao:1 provably:4 pixel:1 sketched:1 adm:15 aforementioned:1 arg:3 among:1 special:2 equal:1 construct:1 once:1 having:1 sampling:1 manually:1 broad:1 yu:1 nearly:3 future:1 np:2 nonsmooth:1 randomly:1 qing:1 replaced:1 phase:7 geometry:1 onz:1 initializers:3 detection:1 interest:3 extreme:1 semidefinite:2 accurate:1 ambient:1 peculiar:1 necessary:1 janzamin:1 facial:1 orthogonal:4 enjoying:1 continuing:1 e0:1 orthogonalized:1 theoretical:4 minimal:1 overcomplete:4 instance:1 handed:2 soft:1 gn:6 cover:1 entry:9 kq:1 uniform:1 rounding:8 too:1 optimally:1 proximal:2 synthetic:3 ju:1 person:4 international:1 siam:1 ie:1 probabilistic:2 othen:1 decoding:1 invertible:2 together:1 concrete:1 again:2 return:1 yp:3 potential:2 coding:2 summarized:1 vaiter:1 satisfy:1 blind:3 idealized:2 tion:1 break:3 closed:1 optimistic:1 portion:1 start:1 recover:5 complicated:1 square:3 largely:1 efficiently:2 yield:1 correspond:2 raw:1 produced:1 worth:1 lighting:1 converged:1 reach:1 suffers:1 whenever:2 failure:1 nonetheless:1 pp:7 tucker:1 obvious:1 naturally:1 proof:8 recovers:6 proved:2 dataset:2 recall:1 knowledge:1 subsection:1 cap:1 ou:1 actually:2 formulation:7 generality:1 stage:1 hand:2 sketch:3 working:1 replacing:1 invite:2 su:1 nonlinear:1 perhaps:1 believe:3 usa:1 attouch:2 normalized:2 hence:10 regularization:1 alternating:12 nonzero:10 white:1 attractive:1 numerator:1 please:1 noted:2 nonrigid:1 theoretic:2 complete:1 demonstrate:1 performs:2 motion:2 image:4 harmonic:1 novel:1 recently:2 common:1 rotation:1 empirically:3 twist:1 extend:1 slight:1 significant:1 ena:1 mathematics:3 language:1 similarity:1 surface:1 showed:1 recent:2 optimizing:2 moderate:1 driven:1 certain:1 nonconvex:12 inequality:2 success:3 continue:1 yi:4 additional:1 somewhat:1 dai:1 kxk0:2 novelty:1 converge:1 paradigm:2 signal:2 semi:1 ibshirani:1 nonzeros:4 kq2:4 seidel:1 exceeds:2 technical:2 sphere:4 long:1 retrieval:2 e1:9 bigger:1 qi:3 variant:3 regression:2 basic:1 scalable:1 essentially:1 vision:2 arxiv:20 iteration:6 agarwal:2 achieved:1 equator:1 addition:1 remarkably:1 source:3 biased:5 subject:2 tend:1 flow:1 seem:1 near:1 noting:1 wirtinger:1 easy:2 enough:4 independence:1 zi:3 suboptimal:1 inner:2 idea:2 cn:1 pca:4 penalty:1 locating:1 algebraic:2 york:1 useful:2 detailed:4 clear:1 ten:1 concentrated:2 generate:2 sign:1 arising:1 correctly:1 yy:1 write:1 discrete:1 four:4 threshold:3 nevertheless:1 achieving:1 backward:1 relaxation:11 fraction:4 sum:5 cone:1 run:1 nascent:1 anghavi:3 place:1 reader:1 separation:3 appendix:1 scaling:1 bound:6 pay:1 guaranteed:4 dwell:1 yale:2 replaces:1 g:3 identifiable:1 nontrivial:1 annual:2 constraint:4 precisely:1 argument:1 min:10 span:3 optimality:3 extremely:1 relatively:2 conjecture:2 structured:3 alternate:1 increasingly:1 lp:2 qu:1 n4:1 invariant:2 taken:1 equation:2 mutually:1 eventually:1 needed:1 know:1 end:2 lojasiewicz:1 operation:1 apply:4 lambertian:1 away:2 spectral:1 generic:2 appropriate:1 rora:2 schmidt:2 rp:11 assumes:1 denotes:1 graphical:3 log2:3 giving:1 practicality:1 reflectance:1 seeking:3 objective:1 move:1 question:1 already:3 tensor:1 planted:16 dependence:1 surrogate:1 exhibit:1 subspace:26 topic:1 manifold:1 trivial:1 provable:3 relationship:2 mini:1 minimizing:1 equivalently:2 difficult:1 unfortunately:2 mostly:1 hao:1 trace:1 stated:1 design:1 perform:1 upper:1 observation:3 datasets:1 descent:1 extended:1 ever:1 communication:1 y1:3 discovered:1 rn:5 arbitrary:6 sharp:2 introduced:3 complement:1 pair:3 cast:1 address:1 able:1 beyond:2 below:1 pattern:4 regime:2 sparsity:6 program:8 max:1 prony:1 critical:1 natural:1 force:1 regularized:1 zr:2 hr:1 advanced:1 yq:11 imply:2 incoherent:1 extract:1 columbia:2 sn:1 prior:1 literature:1 geometric:1 discovery:2 precedent:1 asymptotic:1 embedded:3 loss:1 expect:1 mixed:1 interesting:3 degree:1 cn2:1 thresholding:1 etrapalli:4 row:13 surprisingly:1 repeat:3 transpose:2 free:1 side:1 allow:2 barak:5 ber:1 fall:1 face:6 barrier:1 correspondingly:1 sparse:64 tolerance:1 dimension:7 xn:1 gram:2 transition:4 forward:1 collection:1 jump:1 projected:1 far:1 transaction:2 obtains:1 basri:1 global:15 astie:1 assumed:1 xi:2 kqk2:4 decade:1 why:1 robust:2 decoupling:2 meanwhile:1 main:5 whole:1 noise:1 n2:1 x1:1 fig:3 deployed:1 ny:1 sparsest:7 wish:1 exponential:1 kxk2:3 crude:1 breaking:1 down:3 theorem:5 evidence:1 intrinsic:1 exists:2 supplement:1 magnitude:1 illumination:5 conditioned:4 gap:3 simply:3 likely:1 kurdyka:1 kxk:1 springer:1 minimizer:1 satisfies:2 extracted:2 acm:3 succeed:1 goal:2 presentation:1 viewed:1 endeavor:1 towards:3 price:1 hard:1 change:2 specifically:1 nandkumar:3 contrasted:1 uniformly:1 principal:4 gij:1 experimental:1 gauss:1 succeeds:5 la:1 select:1 people:1 spielman:1 dept:1 handling:1 |
4,864 | 5,403 | Compressive Sensing of Signals from a GMM with
Sparse Precision Matrices
1
1
2
1
Jianbo Yang
Xuejun Liao
Minhua Chen
Lawrence Carin
1
Department of Electrical and Computer Engineering, Duke University
2
Department of Statistics & Department of Computer Science, University of Chicago
{jianbo.yang;xjliao;lcarin@[email protected]},{[email protected]}
Abstract
This paper is concerned with compressive sensing of signals drawn from a Gaussian mixture model (GMM) with sparse precision matrices. Previous work has
shown: (i) a signal drawn from a given GMM can be perfectly reconstructed from
r noise-free measurements if the (dominant) rank of each covariance matrix is
less than r; (ii) a sparse Gaussian graphical model can be efficiently estimated
from fully-observed training signals using graphical lasso. This paper addresses a
problem more challenging than both (i) and (ii), by assuming that the GMM is unknown and each signal is only observed through incomplete linear measurements.
Under these challenging assumptions, we develop a hierarchical Bayesian method
to simultaneously estimate the GMM and recover the signals using solely the incomplete measurements and a Bayesian shrinkage prior that promotes sparsity of
the Gaussian precision matrices. In addition, we provide theoretical performance
bounds to relate the reconstruction error to the number of signals for which measurements are available, the sparsity level of precision matrices, and the ?incompleteness? of measurements. The proposed method is demonstrated extensively
on compressive sensing of imagery and video, and the results with simulated and
hardware-acquired real measurements show significant performance improvement
over state-of-the-art methods.
1
Introduction
Gaussian mixture models (GMMs) [1, 2, 3] have become a popular signal model for compressive
sensing [4, 5] of imagery and video, partly because the information domain in these problems can
be decomposed into subdomains known as pixel/voxel patches [3, 6]. A GMM employs a Gaussian
precision matrix to capture the statistical relations between local pixels/voxels within a patch, and
meanwhile captures the global statistics between patches using its clustering mechanism.
Compressive sensing (CS) of signals drawn from a GMM admits closed-form minimum mean
squared error (MMSE) reconstruction from linear measurements. Recent theoretical analysis in
[7] shows that, given a sensing matrix with entries i.i.d. drawn from a zero-mean, fixed-variance,
Gaussian distribution or Bernoulli distribution with parameter 0.5, if the GMM is known and the
(dominant) rank of each covariance matrix is less than r, each signal can be perfectly reconstructed
from r noise-free measurements. Though this is a much less stringent reconstruction condition than
that prescribed by standard restricted-isometry-property (RIP) bounds, it relies on the assumption
of knowing the exact GMM. If a sufficient number of fully observed signals are available beforehand, one can use maximum likelihood (ML) estimators to train a GMM [8, 9, 7, 1, 10] for use in
reconstructing the signals in question. Unfortunately, finding an accurate GMM a priori is usually a
challenge in practice, because it is difficult to obtain training signals that match the statistics of the
interrogated signals.
1
Recent work [2] on GMM-based methods proposes to solve this problem by estimating the Gaussian components, based on measurements of the signals under interrogation, without resorting to
any fully-observed signals to train a model in advance. The method of [2] has two drawbacks: (i)
it estimates full dense Gaussian covariance matrices, with the number of free parameters to be estimated growing quadratically fast with the signal dimensionality n; (ii) it does not have performance
guarantees, because all previous theoretical results, including those in [7], assume the GMM is given
and thus are no longer applicable to the method of [2]. This paper addresses these two issues.
First, we effectively reduce the number of GMM parameters by restricting the GMM to have sparse
precision matrices with group sparsity patterns, making the GMM a mixture of group-sparse Gaussian graphical models. The group sparsity is motivated by the Markov random field (MRF) property
of natural images and video [11, 12, 13]. Instead of having n2 parameters for each Gaussian component as in [2], we have only n + s parameters, where s is the number of nonzero off-diagonals
of the precision matrix. We develop a variational maximum-marginal-likelihood estimator (variational MMLE) to simultaneously estimate the GMM and reconstruct the signals, with a Bayesian
shrinkage prior used to promote sparsity of the Gaussian precision matrices. Our variational MMLE maximizes the marginal likelihood of the GMM given only the linear measurements, with the
unknown signals treated as random variables and integrated out of the likelihood. A key step of
the variational MMLE is using Bayesian graphical lasso to reestimate the sparse Gaussian precision
matrices based on a posteriori signal samples conditional on the linear measurements.
Second, we provide theoretical performance bounds under the assumption that the GMM is not
exactly known. Assuming the GMM has sparse precision matrices, our theoretical results relate
the signal reconstruction error to the number of signals for which measurements are available, the
sparsity level of the precision matrices, and the ?incompleteness? of measurements, where the last
is defined as the uncertainty (variance) of a signal given its linear measurements.
In the experiments, we present reconstruction results of the proposed method on both simulated
measurements and real measurements acquired by actual hardware [6]. The proposed method outperforms the state-of-art CS reconstruction algorithms by significant margins.
Notations. Let N (x|?, ??1 ) denote a Gaussian density of x with mean ? and precision matrix ?,
kM kF denote the Frobenius matrix norm of matrix M , kM kmax denote the largest entry of M in
terms of magnitude, tr(M ) denote the trace of M , ?0 = ??1
0 denote the true precision matrix (i.e.,
the inverse of true covariance matrix ?0 ), ?? denote the estimate of ?0 by the proposed model.
Herein, the eigenvalues of ?0 are assumed to be bounded in a constant interval [?1 , ?2 ] ? (0, ?), to
guarantee the existence of ?0 . For functions f (x) and g(x), we write f (x) g(x) when f (x) =
O(g(x)) and g(x) = O(f (x)) hold simultaneously.
2
2.1
Learning a GMM of Unknown Signals from Linear Measurements
Signal Reconstruction with a Given GMM
The linear measurement of an unknown signal x ? Rn can be written as y = ?x + , where
? ? Rm?n is a sensing matrix, and ? Rm denote measurement noises (we are interested in
m < n). Assuming ? N (|0, R), one has p(y|x) = N (y|?x, R). We further assume R to be a
scaled identity matrix, R = ??1 I, and thus the noise is white Gaussian.
PK
?1
If x is governed by a GMM, i.e., p(x) = z=1 ? (z) N (x|?(z) , ?(z) ), one may obtain
p(y, x, z) = ? (z) N (y|?x, R)N (x|?(z) , ?(z)
p(y) =
K
X
? (z) N (y|??(z) , R + ??(z)
?1
?0 ),
?1
),
p(x, z|y) = ?(z) N (x|? (z) , (C(z) )?1 ),
(1)
z=1
where
C(z) =
?0 R?1 ? + ?(z)
?1
? (z) = ?z + C(z) ?0 R?1 (y ? ??z ),
,
?1
? (z) N (y|??(z) , R + ??(z) ?0 )
?(z) = PK
?1 0 .
(l) N (y|??(l) , R + ??(l)
?)
l=1 ?
(2)
When the GMM is exactly known, the signal is reconstructed analytically as the conditional mean,
b , E(x|y) =
x
PK
2
z=1 ?
(z)
? (z) .
(3)
It has been shown in [7] that, if the (dominant) rank of each Gaussian covariance matrix is less than
r, the signal can be perfectly reconstructed from only r measurements in the low-noise regime.
2.2 Restriction of the GMM to a mixture of Gaussian Markov Random Fields
A Markov random field (MRF), also known as an undirected graphical model, provides a graphical
representation of the joint probability distribution over multiple random variables, by considering
the conditional dependences among the variables [11, 12, 13]. In image analysis, each node of
an MRF corresponds to a pixel of the image in question, and an edge between two nodes is often
modeled by a potential function to characterize the conditional dependence between the associated
pixels. Because of the local smoothness structure of images, the edges of an MRF are usually
chosen based on a pairwise neighborhood structure: each pixel only has edge connections with
its neighbors. The widely used scheme is that each pixel only has edge connections with its four
immediate neighboring pixels to the left, right, top and bottom [11]. Therefore, an MRF for image
representation is an undirected graph with only a limited number of edges between its nodes.
Generally, learning and inference of an MRF are nontrivial, due to the nonlinearity and nonconvexity of the potential functions [14]. A popular special case of MRF is the Gaussian Markov
random field (GMRF) which is an MRF with a multivariate Gaussian distribution over node variables. The best-known advantage of a GMRF is its simplicity of learning and inference, because
of the nice properties of a multivariate Gaussian distribution. According to Hammersley-Clifford?s
theorem [15], the conditional dependence of the node variables in a GMRF is encoded in the precision matrix. As mentioned before, an MRF is sparse for image analysis problems, on account of
the neighborhood structure in the pixel domain. Therefore, the multivariate Gaussian distribution
associated with a GMRF has a sparse precision matrix. This property of a GMRF in image analysis
is demonstrated in Section 1 of the Supplementary Material.
Inspired by the GMRF interpretation, we place a shrinkage prior on each precision matrix to promote
sparsity when estimating the GMM. The Laplacian shrinkage prior used in [16] is chosen, but other
shrinkage priors [17] could also be used. Specifically, we impose a Laplacian shrinkage prior on the
off-diagonal elements of each of K precision matrices,
p(?
(k)
)=
n Y
Y
q
(k)
? (k) ?ij
2
i=1 j<i
(k)
q
(k)
(k)
exp(? ? (k) ?ij |?ij |), ?k = 1, . . . , K,
(4)
(k)
with the symmetry constraints ?ij = ?ji . In (4), ? (k) > 0 is a ?global? scaling parameter for all
(k)
(k)
the elements of {?ij |i = 1, ..., n, j < i} and generally fixed to be one [18], and ?ij is a ?local?
(k)
weight for the element ?ij . With the Laplacian prior (4), many off-diagonal elements of ?(k) are
encouraged to be close to zero. However, in the inference procedure, the above Laplacian shrinkage
prior (4) is inconvenient due to the lack of analytic updating expressions. This issue is overcome by
using an equivalent scale mixture of normals representation [16] of (4) as shown below:
q
(k)
? (k) ?ij
2
Z
(k)
q
?ij
?1 (k) ?1
(k)
(k)
(k)
(k)
(k)
(k)
exp(? ? ?ij |?ij |) = N (?ij |0, ? (k) ?ij
)InvGa(?ij |1,
)d?ij
2
(5)
(k)
where ?ij is an augmented variable drawn from an inverse gamma distribution. Further, one may
(k)
place a gamma prior on ?ij . Then, a draw of the precision matrix may be represented by
?(k) ?
n Y
Y
(k)
N (?ij |0, ? (k)
?1
(k) ?1
?ij
(k)
(k)
(k)
), ?ij ? InvGa(?ij |1,
i=1 j<i
?ij
(k)
(k)
), ?ij ? Ga(?ij |a0 , b0 )
2
(6)
where a0 , b0 are the hyperparameters.
?1
Suppose {xi }N
are samples drawn from N (x|0, ?(k) ) and S denotes the empirical covariance
PNi=1
(k)
1
matrix N i=1 (xi ? x)(xi ? x)0 where x is the empirical mean of {xi }N
i=1 . If the elements ?
are drawn as in (6), the logarithm of the joint likelihood can be expressed as
(k)
log p({xi }N
)
i=1 , ?
N
?
2
log det(?
(k)
) ? tr(S?
(k)
!
q
n X
X
2
(k)
(k)
(k)
)?
? ?ij |?ij | .
N
i=1 j<i
(7)
From the optimization perspective, the maximum a posterior (MAP) estimations of ?(k) in (7) is
known as the adaptive graphical lasso problem [18].
3
2.3 Group sparsity based on banding patterns
The Bayesian adaptive graphical lasso described above assumes the precision matrix is sparse, and
the same Laplacian prior is imposed on all off-diagonal elements of the precision matrix without any
discrimination. However, the aforementioned neighborhood structure of image pixels implies that
the entries of the precision matrix corresponding to the pairs between neighboring pixels tend to have
significant values. This is consistent with the observations as seen from the demonstration in Section
1 of the Supplementary Material: (i) the bands scattered along a few lines above or below the main
diagonal are constituted by the entries with significant values in the precision matrix; (ii) the entries
in the bands correspond to the pairwise neighborhood structure of the graph, since vectorization of
an image patch is constituted by stacking all columns of pixels in a patch on the top of each other;
(iii) the existence of multiple bands in some Gaussian components reveals that, besides the four
immediate neighboring pixels, other indirected neighboring pixels may also lead to nonnegligible
conditional dependence, though the entries in the associated bands have relatively smaller values.
Inspired by the banding patterns mentioned above, we categorize the elements in the set
(k)
(k)
(k)
{?ij }ni=1,j<i into two groups {?ij |(i, j) ? L1 } and {?ij |(i, j) ? L2 }, where L1 denotes the
set of indices corresponding to the elements in the bands and L2 represents the set of indices for the
(k)
elements not in the bands. For the elements in the group {?ij |(i, j) ? L2 }, the Laplacian prior is
(k)
used to encourage a sparse precision matrix. For the elements in the group {?ij |(i, j) ? L1 } , the
sparsity is not desired so a normal prior with Gamma hyperparameters is used instead. Accordingly,
the expressions in (6) can be replaced by
?(k) ?
n Y
Y
(k)
N (?ij |0, ? (k)
?1
(k) ?1
?ij
)
i=1 i<j
(
(k)
?ij
?
(k)
Ga(?ij |c0 , d0 ),
if (i, j) ? L1
(k)
(k)
InvGa(?ij |1,
?ij
2
(k)
(k)
), ?ij ? Ga(?ij |a0 , b0 ),
(8)
.
if (i, j) ? L2
With the prior distribution of ?(k) in (6) replaced with that in (8), the joint log-likelihood in (7)
changes to
(k)
log p({xi }N
)
i=1 , ?
?
N
? ?log det(?(k) ) ? tr(S?(k) ) ?
2
X
(i,j)?L1
2 (k) (k) (k) 2
? ?ij k?ij k ?
N
X
(i,j)?L2
2
N
?
q
(k)
(k)
? (k) ?ij |?ij |? .
(9)
To the best of our knowledge, the maximum a posterior (MAP) estimations of ?(k) in (9) has not
been studied in the family of graphical lasso or its variants, from the optimization perspective.
2.4 Hierarchical Bayesian model and inference
We consider the collective compressive sensing of the signals X = {xi ? Rn }N
i=1 that are drawn
from an unknown GMM. The noisy linear measurements of X are given by Y = {y i ? Rm : y i =
?i xi + i }N
i=1 . We assume the sensing matrices to be signal-dependent to account for generality
(i.e., ?i depends on the signal index i).
The unification of signal reconstruction with a given GMM (presented in Section 2.1) and GMRF learning with fully-observed training signals (presented in Section 2.2) leads to the following
Bayesian model,
y i |xi ? N (y i |?i xi , ??1 I), xi ?
K
X
? (z) N (xi |?(z) , ?(z)
?1
), ? ? Ga(?|e0 , f0 )
(10)
z=1
?(k) ?
n Y
Y
(k)
N (?ij |0, ? (k)
?1
(k) ?1
?ij
(k)
(k)
(k)
), ?ij ? InvGa(?ij |1,
i=1 i<j
?ij
(k)
(k)
), ?ij ? Ga(?ij |a0 , b0 ), (11)
2
The expression in (11) could be replaced by (8) if the group sparsity is considered in the precision
matrix. In addition to the precision matrices, we further add the following standard priors on the
other parameters of the GMM to make the proposed model a full hierarchical Bayesian model,
?(k) ? N (?(k) |m0 , (?0 ?(k) )?1 ), ? ? Dirichlet(? (1) , . . . , ? (K) |a0 ),
4
(12)
where m0 , a0 and ?0 are hyperparameters.
We develop the inference procedure for the proposed Bayesian hierarchical model. Let the symbols
Z, ?, ?, ?, ?, ? denote the sets {zi }, {?(k) }, {?(k) }, {? (k) }, {?(k) }, {? (k) } respectively. The
marginalized likelihood function is written as
Z
L(?) = ln
p(Y, ?, ?)d?
where ? , {X, Z, ?, ?} and ? , {?, ?, ?, ?} denote the set of the latent variables and parameters of the model, respectively. An expectation-maximization (EM) algorithm [19] could be used to
find the optimal ? by alternating the following two steps
? E-step: Find p(?|Y, ?? ) with ?? computed at the M-step, and obtain the expected complete log-likelihood E? (ln p(Y, ?, ?? )).
? M-step: Find an improved estimate of ?? by maximizing the expected complete loglikelihood given at the E-step.
However, it is intractable to compute the exact posterior p (?|Y, ?) at the E step. We develop a
variational inference approach to overcome the intractability. Based on the mean field theory [20],
we approximate the posterior distribution p (?|Y, ?) by a proposal distribution q(?) that factorizes
over the variables as follows
q(?) = q(X, Z, ?, ?) = q(X, Z)q(?)q(?).
(13)
Then, we find an optimal distribution q(?) that minimizes the Kullback-Leibler (KL) divergence
R
q(?)
KL(q(?)||p(?|Y, ?)) = q(?) ln p(?|Y,?)
d?, or equivalently, maximizes the evidence lower
bound (ELBO) of the log-marginal data likelihood [21], denoted by F(q(?), ?),
Z
ln p(Y, ?) = ln
q(?)
p (Y, ?, ?)
d? ?
q(?)
Z
q(?) ln
p (Y, ?, ?)
d? , F(q(?), ?)
q(?)
(14)
where the inequality is held based on the Jensen?s inequality.
With the above approximation, the entire algorithm becomes a variational EM algorithm and it
iterates between the following VE-step and VM-step until convergence:
? VE-step: Find the optimal posterior distribution q ? (?) that maximizes F(q(?), ?? ) with
?? computed at the VM-step.
? VM-step: Find the optimal ?? that maximizes F(q ? (?), ?) with q ? (?) computed at the
VE-step.
The full update equations of the variational EM algorithm are given in Section 2 of the Supplementary Material.
3
Theoretical Analysis
The proposed hierarchical Bayesian model unifies the task of signal recovery and the task of estimating the mixture of GMRF, with a common goal of maximizing the ELBO of the log-marginal
likelihood of the measurements. This section provides a theoretical analysis to further reveal the
mutual influence between these two tasks (Theorem 1 and Theorem 2), and establish a theoretical
performance bound (Theorem 3) to relate the reconstruction error to the number of signals being
measured, the sparsity level of precision matrices, and the ?incompleteness? of measurements. The
proofs of these theorems are presented in Sections 3-5 of the Supplementary Material. For convenience, we consider the single Gaussian case, so the superscript (k) is omitted in the sequel. We
begin with the definitions and assumptions used in the theorems.
e i and x
b i be the signals estimated from measurement y i , using the true precision
Definition 3.1 Let x
matrix ?0 and the estimated precision matrix ?? respectively, according to (3),
b i =? + ?0 + ?0i R?1 ?i
x
?1
?0i R?1 (yi ? ?i ?) = ? + Ci ?0i R?1 (yi ? ?i ?)
?1 0 ?1
?1 0 ?1
ei =? + ?0 + ? + ?0i R?1 ?i
x
?i R (yi ? ?i ?) = ? + C?1
+?
?i R (yi ? ?i ?) .
i
b i as the
Assuming yi ? Rr is noise-free and the (dominant) rank of ?0 is less than r, one obtains x
b i = xi . Then the reconstruction error of x
e i is k? i k2 , where ? i = x
ei ? x
bi.
true signal xi [7], i.e., x
5
Definition 3.2 The estimation error of ?? is defined as k?kF where ? = ?? ? ?0 .
At each VM-step of the variational EM algorithm developed in Section 2.4, ?? is updated based on
the empirical covariance matrix ?em computed from {e
xi }, i.e.,
?em =
N
N
N
N
1 X
1 X
1 X
1 X
eix
e 0i +
bix
b 0i +
x
Ci =
x
(2b
xi ?i0 + ?i ?i0 + Ci ),
N i=1
N i=1
N i=1
N i=1
{z
} |
{z
}
|
?0
em
(15)
?de
where {b
xi } and {e
xi } are considered to both have zero mean, as one can always center the signals
with respect to their means [2].
Definition 3.3 The deviation of empirical matrix ?0em is defined as ?de = ?em ??0em according to
(15), and we use ?
?de , k?de kmax to measure this deviation. Considering
q the developed variational
EM algorithm can converge to a local minimum, we assume ?
?de ? c
log n
N
for a constant c > 01 .
3.1 Theoretical results
Theorem 1 Assuming kCi kF k?kF < 1, the reconstruction error of the i-th signal is upper boundkCi kF k?kF
kb
xi k2 .
ed as k? i k2 ? 1?kC
i k k?k
F
F
Theorem 1 establishes the error bound of signal recovery in terms of ?. In this theorem, ?? can be
obtained by any GMRF estimation methods, including [1, 2] and the proposed method.
?
??
?
??
Let ? = min(i,j)?S c N ij , ? = max(i,j)?S N ij , S = {(i, j) : ?ij 6= 0, i 6= j}, S c = {(i, j) :
?ij = 0, i 6= j} and the cardinality of S be s. The following theorem establishes an upper bound of
k?kF on account of ?de .
q
log n
Theorem 2 Given the empirical covariance matrix ?em , if ?, ?
?de , then we have
N + ?
p
?
?de }.
k?kF = Op { (n + s) log n/N + n + s?
Note that the standard graphical lasso and its variants [18, 23] assume the true signal samples {xi }
are fully observed when estimating ?? , so they correspond to the simple case that ?
?de = 0. Loh
and Wainwright [22, Corollary 5] also provides an upper bound of k?kF taking ?de into account.
However, they assume ?0em is attainable and the proof of their corollary relies on their proposed
GMRF estimation algorithm, so the theoretical result in [22] cannot be used here.
PN
PN
b max = supi kb
Let 0 = N1 i=1 kb
xi ? ?k2 , ? = N1 i=1 tr(Ci ), ?max = supi k?i k2 , x
xi k2 and
? = maxi kCi kF . A combination of Theorem 1 and 2 leads to the following theorem which relates the error bound of signal reconstruction to the number of partially-observed signals (observed
through incomplete linear measurements), the sparsity level of precision matrices, and the uncertainty of signal reconstruction (i.e., ? and ?) which represent the ?incompleteness? of the measurements.
q
log n
?de , ?k?kF < ?
Theorem 3 Given the empirical covariance matrix ?em , if ?, ?
N + ?
?
where ? is a constant and (1 ?p
?)/ n + s > M 0 (?max + 2b
xmax )? with M being an appropriate
?
constant to make k?kF ? M (n + s) log n/N
+
M
n
+
s?
?de hold with high probability, then
?
PN
(log n)/N +?
1
bi k2 ? (1??)/?n+s?M (? +2bx )? M 0 ?.
we obtain that N i=1 ke
xi ? x
0
max
max
From Theorem 3, we find that when the number of partially-observed signals N tends to infinity and
the uncertainty of signal reconstruction tr(Ci ) tends to zero ? i, the average reconstruction error
PN
1
bi k2 is close to zero with high probability.
xi ? x
i=1 ke
N
4
Experiments
The performance of the proposed methods is evaluated on the problems of compressive
sensing (CS) of imagery and high-speed video2 . For convenience, the proposed method
is termed as Sparse-GMM when using the non-group sparsity described in Section 2.2,
1
2
A similar assumption is made in expression (3.13) of [22].
The complete results can be found at the website: https://sites.google.com/site/nipssgmm/.
6
and is termed Sparse-GMM(G) when using the group sparsity described in Section 2.3.
For Sparse-GMM(G), we construct the two groups L1 and L2 as follows : L1 =
{(i, j) : pixel i is one of four immediate neighbors, in the spatial domain, of pixel j, i 6= j} and
L2 = {(i, j) : i, j = 1, 2, ? ? ? , n, i 6= j} \ L1 . The proposed methods are compared with state-ofthe-art methods, including: a GMM pre-trained from training patches (GMM-TP) [7, 8], a piecewise
linear estimator (PLE) [2], generalized alternating projection (GAP) [24], Two-step Iterative Shrinkage/Thresholding (TwIST) [25], KSVD-OMP [26].
For
of the scaled mixture of Gaussians are set as
p the proposed methods, the hyperparameters
a0 /b0 /N ? 300, c0 = d0 = 10?6 , the hyperparameter of Dirichlet prior ?0 is set as a vector with all elements being one, the hyperparameters of the mean of each Gaussian component are
?6
set as ?0 = 1, and m0 is set to the mean of the initialization of {b
xi }N
for
i=1 . We fixed ? = 10
the proposed methods, GMM-TP and PLE. The number of dictionary elements in KSVD is set to
the best in {64, 128, 256, 512}. The TwIST adopts the total-variation (TV) norm, and the results of
TwIST reported here represented the best among the different settings of regularization parameter in
the range of [10?4 , 1]. In GAP, the spatial transform is chosen between DCT and waveletes and the
one with the best result is reported, and the temporal transform for video is fixed to be DCT.
4.1 Simulated measurements
Compressive sensing of still images. Following the single pixel camera [27], an image xi is projected onto the rows of a random sensing matrix ?i ? Rm?n to obtain the compressive measurements y i for i = 1, . . . , N . Each sensing matrix ?i is constituted by the elements drawn
from a uniform distribution in [0, 1]. The USPS handwritten digits dataset 3 and the
face dataset [28] are used in this experiment. In each dataset, we randomly select 300 images
and each image is resized to the scale of 12 ? 12. Eight settings of CS ratios are adopted with
m
n ? {0.05, 0.10, 0.15, 0.20, 0.25, 0.30, 0.35, 0.40}. Since signal xi in the single pixel camera represents an entire image which generally has unique statistics, it is infeasible to find suitable training
data in practice. Therefore, GMM-TP and KSVD-OMP are not compared to in this experiment4 . For
PLE, Sparse-GMM and Sparse-GMM(G), the minimum-norm estimates from the measurements,
b i = arg minx {kxk22 : ?i x = y i } = ?0i (?i ?0i )?1 y i , i = 1, . . . , N , are used to initialize the
x
GMM. The number of GMM components K in PLE, Sparse-GMM, and Sparse-GMM(G) is tuned
among 2 ? 10 based on Bayesian information criterion (BIC).
GAP
TwIST
PLE
Sparse-GMM
Sparse-GMM(G)
14
12
32
GAP
TwIST
PLE
Sparse-GMM
Sparse-GMM(G)
25
20
PSRN (dB)
PSRN (dB)
16
15
GAP (23.72)
TwIST (24.81)
GMM-TP (24.47)
KSVD-OMP (22.37)
PLE (25.35)
Sparse-GMM (27.3)
Sparse-GMM(G) (28.05)
30
28
PSRN (dB)
18
26
24
10
10
22
8
0.05
0.1
0.15
0.2
0.25
0.3
CS measurements fraction in a patch
0.35
0.4
5
0.05
0.1
0.15
0.2
0.25
0.3
CS measurements fraction in a patch
0.35
0.4
20
5
10
15
20
Frames
25
30
Figure 1: A comparison of reconstruction performances, in terms of PSNR, among different methods
for CS of imagery on USPS handwritten digits (left) and face datasets (middle), and CS
of video on NBA game dataset (right), with the average PSNR over frames shown in the brackets.
Compressive sensing of high-speed video. Following the Coded Aperture Compressive Temporal
Imaging (CACTI) system [6], each frame of video to be reconstructed is encoded with a shifted
binary mask which is designed by randomly drawing values from {0, 1} at every pixel location,
with a 0.5 probability of drawing 1. Each signal xi represents the vectorization of T consecutive
spatial frames, obtained by first vectorizing each frame into a column and then stacking the resulting
T columns on top of each other. The measurement y i is constituted by y i = ?i xi where ?i =
[?i,1 , . . . , ?i,T ] and ?i,t is a diagonal matrix with its diagonal being the mask that is applied to
the t-th frame. A video containing NBA game scenes is used in the experiment. It has 32 frames,
each of size 256 ? 256, and T is set to be 8. For GMM-TP, KSVD-OMP, PLE, Sparse-GMM and
Sparse-GMM(G), we partition each 256 ? 256 measurement frame into a set of 64 ? 64 blocks,
and each block is considered as if it were a small frame and is processed independently of other
blocks.5 The patch is of size 4 ? 4 ? T . Since each block is only 64 ? 64, a small number of GMM
components are sufficient to capture its statistics, and we find the results are robust to K as long as
2 ? K ? 5 for PLE, Sparse-GMM and Sparse-GMM(G). Following [8, 26], we use the patches
3
It is downloaded from http://cs.nyu.edu/?roweis/data.html.
The results of other settings can be found at https://sites.google.com/site/nipssgmm/.
5
This subimage processing strategy has also been used in [2].
4
7
Max-Max Max-Max
20
40
40
60
60
80
80
100
100
120
140
1
0.8
0.8
0.6
0.6
0.4
0.4
120
20
20
40
40
60
60
80
80
100
100
120
120
1
1
0.8
0.8
0.6
0.6
0.4
0.4
20
MMLE-MFA
MMLE-MFA Max-Max
1
20
20
1
Sparse-GMM
Sparse-GMM
MMLE-GMM
1
1
20
20
0.8
0.8
0.6
0.6
0.4
0.4
20
1
Sparse-GMM(G)
Sparse-GMM(G)
MMLE-MFA
1
1
20
20
0.8
0.8
0.8
0.6
0.6
0.4
0.4
20
0.2
0.2
0
40
40
60
60
80
80
100
100
120
140
1
120
0.2
140
0
20 40 6020 8040100
60120
80140
100 120 140
1
0.2
0
40
40
40
60
60
60
80
80
80
100
100
100
120
120
120
140
1
40
40
40
60 0.6
60
60
80
80
80
100
100
120
120
0.4
100
0.2
0.2
120 0.2
140
140
0
0
20 40 6020 8040100
6020
120
8040
140
100
60120
80140
100 120 140
1
1
140
1
40
40
60 0.6
60
60
80
80
80
100
100
120
120
100
0.2
0.2
120 0.2
140
140
0
0
20 40 6020 8040100
6020
120
8040
140
100
60120
80140
100 120 140
1
1
0
40
0.4
140
0
1
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
140
0
20 40 6020 8040100
60120
80140
100 120 140
0
20
40
40
60
60
80
80
100
100
120
120
140
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
140
0
20 40 6020 8040100
60120
80140
100 120 140
0
20
20
20
40
40
40
60
60
60
80
80
80
100
100
100
120
120
120
140
20
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
20
20
40
40
40
60 0.6
60
60
80
80
80
100
100
100
120 0.2
120
120
140
140
0
0
20 40 6020 8040100
6020
120
8040
140
100
60120
80140
100 120 140
140
0.4
0
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
20
20
40
40
40
60 0.6
60
60
80
80
80
100
100
100
120 0.2
120
120
0.8
140
140
0
0
20 40 6020 8040100
6020
120
8040
140
100
60120
80140
100 120 140
140
0.4
0
GMM-TP
#2
#3
#5
#1
#6
#11
#7
#12
#8
#9
#13
#10
#14
#2
#3
#4
#5
#7
#8
#9
#10
#11
#12
#13
#14
#6
#11
#11
#6
#11
5
#2
#3
#4
#5
#7
#8
#9
#10
#12
#13
#14
Sparse-GMM
TwIST
#2
#3
#4
#5
#7
#8
#9
#10
#12
#13
#14
#1
#6
#11
#2
#7
#12
#3
#8
#13
#4
#9
#5
#1
#6
#10
#11
#14
Sparse-GMM(G)
#1
#1
#6
#6
GAP
#1
PLE
KSVD-OMP
#4
#2
#3
#4
#5
#7
#8
#9
#10
#12
#13
#14
Raw measurement
(Coded image)
#2
#3
#4
#5
#7
#8
#9
#10
#12
#13
#14
GMM-TP
GAP
TwIST
Sparse-GMM
Sparse-GMM(G)
Conclusions
The success of compressive sensing of signals from a GMM highly depends on the quality of the
estimator of the unknown GMM. In this paper, we have developed a hierarchical Bayesian method
to simultaneously estimate the GMM and recover the signals, all based on using only incomplete
linear measurements and a Bayesian shrinkage prior for promoting sparsity of the Gaussian precision matrices. In addition, we have obtained theoretical results under the challenging assumption
that the underlying GMM is unknown and has to be estimated from measurements that contain only
incomplete information about the signals. Our results extend substantially from previous theoretical
results in [7] which assume the GMM is exactly known. The experimental results with both simulated and hardware-acquired measurements show the proposed method significantly outperforms
state-of-the-art methods.
Acknowledgement
The research reported here was funded in part by ARO, DARPA, DOE, NGA and ONR.
6
The results of the training videos containing general scenes can be found at the aforementioned website.
8
0.6
0.6
0.6
0.4
0.4
0.4
40
60
80
100
0.2
0.2
0.2
0
120
140
20 40 60 80 100 120
1
20
0.8
0.8
0.8
0.6
0.6
0.6
0.4
0.4
0.4
0.2
0.2
0.2
40
60
80
100
140
140
0
0
20 40 6020 8040100
6020
120
8040
140
100
60120
80140
100 120 140
4.2 Real measurements
We demonstrate the efficacy of
the proposed methods on the CS
of video, with the measurements
acquired by the actual hardware
of CACTI camera [6]. A letter is
placed on the blades of a chopper wheel that rotates at an angular velocity of 15 blades per second. The training data are obtained from the videos of a chopper wheel rotating at several orientations, positions and velocities. These training videos are
captured by a regular camcorder
at frame-rates that are different from the high-speed frame rate Figure 3: Reconstructed images 256 ? 256 ? T by differenachieved by CACTI reconstruc- t methods from the ?raw measurement? acquired from CACTI
tion. Other settings of the meth- with T = 14. The region in the red boxes are enlarged and
ods are the same as in the experi- shown at the right bottom part for better comparison.
ments on simulated data. The reconstruction results are shown in Figure 3, which shows that SparseGMM(G) generally yields sharper reconstructed frames with less ghost effects than other methods.
#1
0.8
140
140
0
0
20 40 6020 8040100
6020
120
8040
140
100
60120
80140
100 120 140
1
1
20
0.8
1
0.8
20
Results. From the results shown in
Figure 1, we observe that the proposed
methods, especially Sparse-GMM(G),
outperforms other methods with significant margins in all considered settings. The better performance of SparseGMM(G) over Sparse-GMM validates Figure 2: Plots of an example precision matrix (in magthe advantage of considering group s- nitude) learned by different GMM methods on the Face
parsity in the model. Figure 2 shows the dataset with m/n = 0.4. It is preferred to view the figure
an example precision matrix of one of K electronically. The magnitudes in each precision matrix
Gaussian components that are learned are scaled to the range of [0, 1].
by the methods of PLE, Sparse-GMM, and Sparse-GMM(G) on the face dataset. From this figure,
we can see that Sparse-GMM and Sparse-GMM(G) show much clearer groups sparsity than PLE,
demonstrating the benifits of using group sparsity constructed from the banding patterns.
20
Sparse-GMM
1
0.8
20
0.8
of a randomly-selected video containing traffic scenes6 , which are irrelevant to the NBA game, as
training data to learn a GMM for GMM-TP with 20 components, and we use it to initialize PLE,
Sparse-GMM, and Sparse-GMM(G). The same training data are used to learn the dictionaries for
KSVD-OMP.
PLE
Sparse-GMM
Sparse-GMM(G)
140
0
20 40 6020 8040100
60120
80140
100 120 140
1
20
140
MMLE-GMM
MMLE-GMM
1
20
0
120
140
20 40 60 80 100 120
References
[1] M. Chen, J. Silva, J. Paisley, C. Wang, D. Dunson, and L. Carin, ?Compressive sensing on manifolds
using a nonparametric mixture of factor analyzers: Algorithm and performance bounds,? IEEE Trans. on
Signal Processing, 2010.
[2] G. Yu, G. Sapiro, and S. Mallat, ?Solving inverse problems with piecewise linear estimators: From Gaussian mixture models to structured sparsity,? IEEE Trans. on Image Processing, 2012.
[3] G. Yu and G. Sapiro, ?Statistical compressed sensing of Gaussian mixture models,? IEEE Trans. on Signal
Processing, 2011.
[4] E. J. Cand`es, J. Romberg, and T. Tao, ?Robust uncertainty principles: Exact signal reconstruction from
highly incomplete frequency information,? IEEE Trans. on Inform. Theory, 2006.
[5] D. L. Donoho, ?Compressed sensing,? IEEE Trans. on Inform. Theory, 2006.
[6] P. Llull, X. Liao, X. Yuan, J. Yang, D. Kittle, L. Carin, G. Sapiro, and D. J. Brady, ?Coded aperture
compressive temporal imaging,? Optics Express, 2013.
[7] F. Renna, R. Calderbank, L. Carin, and M. Rodrigues, ?Reconstruction of signals drawn from a Gaussian
mixture via noisy compressive measurements,? IEEE Trans. Signal Processing, 2014.
[8] J. Yang, X. Yuan, X. Liao, P. Llull, G. Sapiro, D. J. Brady, and L. Carin, ?Video compressive sensing using
Gaussian mixture models,? IEEE Trans. on Image Processing, vol. 23, no. 11, pp. 4863?4878, 2014.
[9] ??, ?Gaussian mixture model for video compressive sensing,? ICIP, pp. 19?23, 2013.
[10] D. Zoran and Y. Weiss, ?From learning models of natural image patches to whole image restoration,? in
ICCV, 2011.
[11] S. Roth and M. J. Black, ?Fields of experts,? Int. J. Comput. Vision, 2009.
[12] F. Heitz and P. Bouthemy, ?Multimodal estimation of discontinuous optical flow using Markov random
fields.? IEEE Trans. Pattern Anal. Mach. Intell., 1993.
[13] V. Cevher, P. Indyk, L. Carin, and R. Baraniuk, ?Sparse signal recovery and acquisition with graphical
models,? IEEE Signal Processing Magazine, 2010.
[14] M. Tappen, C. Liu, E. Adelson, and W. Freeman, ?Learning Gaussian conditional random fields for lowlevel vision,? in CVPR, 2007.
[15] H. Rue and L. Held, Gaussian Markov Random Fields: Theory and Applications, 2005.
[16] T. Park and G. Casella, ?The Bayesian lasso,? Journal of the American Statistical Association, 2008.
[17] N. G. Polson and J. G. Scott, ?Shrink globally, act locally: Sparse Bayesian regularization and prediction,?
Bayesian Statistics, 2010.
[18] J. Fan, Y. Feng, and Y. Wu, ?Network exploration via the adaptive lasso and scad penalties,? Ann. Appl.
Stat., 2009.
[19] A. P. Dempster, N. M. Laird, and D. B. Rubin, ?Maximum likelihood from incomplete data via the EM
algorithm,? Journal of the Royal Statistical Society: Series B, 1977.
[20] G. Parisi, Statistical Field Theory.
Addison-Wesley, 1998.
[21] M. I. Jordan, Z. Ghahramani, T. S. Jaakkola, and L. K. Saul, ?An introduction to variational methods for
graphical models,? Machine Learning, 1999.
[22] P.-L. Loh and M. J. Wainwright, ?High-dimensional regression with noisy and missing data: Provable
guarantees with nonconvexity,? Ann. Statist., 2012.
[23] J. Friedman, T. Hastie, and R. Tibshirani, ?Sparse inverse covariance estimation with the graphical lasso,?
Biostatistics, 2008.
[24] X. Liao, H. Li, and L. Carin, ?Generalized alternating projection for weighted-`2,1 minimization with
applications to model-based compressive sensing,? SIAM Journal on Imaging Sciences, 2014.
[25] J. Bioucas-Dias and M. Figueiredo, ?A new TwIST: Two-step iterative shrinkage/thresholding algorithms
for image restoration,? IEEE Trans. on Image Processing, 2007.
[26] Y. Hitomi, J. Gu, M. Gupta, T. Mitsunaga, and S. K. Nayar, ?Video from a single coded exposure photograph using a learned over-complete dictionary,? ICCV, 2011.
[27] M. F. Duarte, M. A.Davenport, D. Takhar, J. N. Laska, S. Ting, K. F. Kelly, and R. G. Baraniuk, ?Singlepixel imaging via compressive sampling,? IEEE Signal Processing Magazine, 2008.
[28] J. B. Tenenbaum, V. Silva, and J. C. Langford, ?A global geometric framework for nonlinear dimensionality reduction,? Science, 2000.
9
| 5403 |@word middle:1 norm:3 c0:2 km:2 covariance:10 attainable:1 tr:5 blade:2 reduction:1 liu:1 series:1 efficacy:1 tuned:1 mmse:1 outperforms:3 com:3 od:1 gmail:1 written:2 dct:2 chicago:1 partition:1 analytic:1 designed:1 plot:1 update:1 discrimination:1 selected:1 website:2 accordingly:1 provides:3 iterates:1 node:5 location:1 along:1 constructed:1 become:1 ksvd:7 yuan:2 pairwise:2 acquired:5 video2:1 mask:2 expected:2 cand:1 growing:1 inspired:2 freeman:1 decomposed:1 globally:1 actual:2 considering:3 cardinality:1 becomes:1 begin:1 estimating:4 notation:1 bounded:1 maximizes:4 underlying:1 biostatistics:1 banding:3 minimizes:1 substantially:1 developed:3 compressive:19 finding:1 brady:2 guarantee:3 temporal:3 sapiro:4 every:1 act:1 exactly:3 jianbo:2 rm:4 scaled:3 nonnegligible:1 k2:8 before:1 bioucas:1 engineering:1 local:4 tends:2 mach:1 solely:1 black:1 initialization:1 studied:1 challenging:3 bix:1 appl:1 limited:1 bi:3 range:2 unique:1 camera:3 practice:2 block:4 digit:2 lcarin:1 procedure:2 empirical:6 significantly:1 projection:2 pre:1 regular:1 convenience:2 close:2 ga:5 cannot:1 onto:1 wheel:2 kmax:2 influence:1 romberg:1 restriction:1 equivalent:1 map:2 demonstrated:2 imposed:1 maximizing:2 center:1 roth:1 missing:1 lowlevel:1 independently:1 exposure:1 ke:2 xuejun:1 simplicity:1 gmrf:10 recovery:3 estimator:5 eix:1 variation:1 updated:1 suppose:1 mallat:1 rip:1 exact:3 duke:3 magazine:2 rodrigues:1 element:14 velocity:2 tappen:1 updating:1 bouthemy:1 observed:9 bottom:2 electrical:1 capture:3 wang:1 region:1 mentioned:2 dempster:1 trained:1 zoran:1 solving:1 usps:2 gu:1 multimodal:1 joint:3 darpa:1 represented:2 train:2 fast:1 neighborhood:4 encoded:2 widely:1 solve:1 supplementary:4 loglikelihood:1 drawing:2 reconstruct:1 elbo:2 compressed:2 cvpr:1 statistic:6 transform:2 noisy:3 validates:1 superscript:1 indyk:1 laird:1 advantage:2 eigenvalue:1 rr:1 parisi:1 reconstruction:19 aro:1 cactus:4 neighboring:4 roweis:1 frobenius:1 convergence:1 develop:4 clearer:1 stat:1 measured:1 ij:54 op:1 b0:5 c:10 implies:1 drawback:1 discontinuous:1 kb:3 exploration:1 stringent:1 material:4 hold:2 considered:4 normal:2 exp:2 lawrence:1 m0:3 dictionary:3 consecutive:1 omitted:1 estimation:7 applicable:1 largest:1 establishes:2 weighted:1 minimization:1 gaussian:32 always:1 pn:4 shrinkage:10 resized:1 factorizes:1 jaakkola:1 corollary:2 llull:2 improvement:1 rank:4 bernoulli:1 likelihood:11 duarte:1 posteriori:1 inference:6 dependent:1 i0:2 integrated:1 entire:2 a0:7 relation:1 kc:1 interested:1 tao:1 pixel:18 arg:1 issue:2 among:4 aforementioned:2 denoted:1 priori:1 html:1 proposes:1 orientation:1 art:4 special:1 initialize:2 mutual:1 marginal:4 laska:1 construct:1 spatial:3 having:1 field:10 sampling:1 encouraged:1 represents:3 park:1 yu:2 adelson:1 carin:7 promote:2 piecewise:2 employ:1 few:1 randomly:3 simultaneously:4 gamma:3 divergence:1 ve:3 intell:1 replaced:3 n1:2 friedman:1 highly:2 mixture:13 bracket:1 held:2 accurate:1 beforehand:1 edge:5 encourage:1 unification:1 incomplete:7 logarithm:1 rotating:1 desired:1 inconvenient:1 e0:1 theoretical:12 cevher:1 column:3 tp:8 restoration:2 maximization:1 stacking:2 deviation:2 entry:6 uniform:1 characterize:1 reported:3 density:1 siam:1 sequel:1 off:4 vm:4 imagery:4 squared:1 clifford:1 containing:3 davenport:1 expert:1 american:1 bx:1 li:1 account:4 potential:2 de:12 int:1 depends:2 tion:1 view:1 closed:1 traffic:1 red:1 recover:2 ni:1 variance:2 efficiently:1 correspond:2 ofthe:1 yield:1 bayesian:16 unifies:1 handwritten:2 raw:2 reestimate:1 inform:2 casella:1 ed:1 definition:4 acquisition:1 frequency:1 pp:2 associated:3 proof:2 dataset:6 popular:2 knowledge:1 dimensionality:2 psnr:2 wesley:1 improved:1 wei:1 evaluated:1 though:2 box:1 generality:1 shrink:1 angular:1 until:1 langford:1 ei:2 nonlinear:1 lack:1 google:2 quality:1 reveal:1 effect:1 contain:1 true:5 analytically:1 regularization:2 alternating:3 nonzero:1 leibler:1 white:1 game:3 criterion:1 generalized:2 complete:4 demonstrate:1 l1:8 silva:2 image:22 variational:10 common:1 ji:1 twist:9 experiment4:1 extend:1 interpretation:1 association:1 measurement:41 significant:5 paisley:1 smoothness:1 resorting:1 nonlinearity:1 analyzer:1 funded:1 f0:1 longer:1 add:1 dominant:4 multivariate:3 isometry:1 recent:2 posterior:5 perspective:2 camcorder:1 irrelevant:1 termed:2 inequality:2 binary:1 success:1 onr:1 yi:5 seen:1 minimum:3 captured:1 impose:1 omp:6 converge:1 signal:59 ii:4 relates:1 full:3 multiple:2 d0:2 match:1 long:1 promotes:1 coded:4 laplacian:6 prediction:1 mrf:9 variant:2 regression:1 liao:4 supi:2 expectation:1 vision:2 represent:1 xmax:1 proposal:1 addition:3 interval:1 tend:1 undirected:2 db:3 flow:1 gmms:1 jordan:1 yang:4 iii:1 concerned:1 bic:1 zi:1 hastie:1 lasso:9 perfectly:3 reduce:1 knowing:1 det:2 motivated:1 expression:4 penalty:1 loh:2 generally:4 nonparametric:1 extensively:1 band:6 hardware:4 processed:1 locally:1 statist:1 tenenbaum:1 http:3 shifted:1 estimated:5 per:1 tibshirani:1 write:1 hyperparameter:1 vol:1 express:1 group:14 key:1 four:3 kci:2 demonstrating:1 drawn:10 gmm:87 nonconvexity:2 imaging:4 graph:2 fraction:2 nga:1 inverse:4 letter:1 uncertainty:4 baraniuk:2 place:2 family:1 mmle:9 wu:1 patch:11 draw:1 scaling:1 incompleteness:4 nba:3 bound:10 fan:1 nontrivial:1 optic:1 constraint:1 infinity:1 scene:2 speed:3 prescribed:1 min:1 optical:1 relatively:1 department:3 tv:1 according:3 structured:1 combination:1 scad:1 smaller:1 reconstructing:1 em:15 making:1 restricted:1 iccv:2 xjliao:1 ln:6 equation:1 mechanism:1 addison:1 dia:1 adopted:1 available:3 gaussians:1 eight:1 promoting:1 hierarchical:6 observe:1 appropriate:1 existence:2 subdomains:1 top:3 clustering:1 denotes:2 assumes:1 dirichlet:2 graphical:13 marginalized:1 ting:1 ghahramani:1 especially:1 establish:1 society:1 feng:1 question:2 strategy:1 dependence:4 diagonal:7 minx:1 rotates:1 simulated:5 nitude:1 manifold:1 provable:1 assuming:5 besides:1 modeled:1 index:3 ratio:1 demonstration:1 equivalently:1 difficult:1 unfortunately:1 dunson:1 sharper:1 relate:3 takhar:1 trace:1 polson:1 anal:1 collective:1 unknown:7 upper:3 observation:1 markov:6 datasets:1 immediate:3 frame:12 rn:2 pair:1 kl:2 connection:2 icip:1 quadratically:1 herein:1 learned:3 trans:9 address:2 usually:2 pattern:5 below:2 scott:1 regime:1 sparsity:18 challenge:1 hammersley:1 ghost:1 royal:1 including:3 video:16 max:12 wainwright:2 mfa:3 suitable:1 natural:2 treated:1 meth:1 scheme:1 kxk22:1 prior:16 voxels:1 nice:1 l2:7 kf:12 vectorizing:1 acknowledgement:1 kelly:1 geometric:1 fully:5 calderbank:1 interrogation:1 downloaded:1 sufficient:2 consistent:1 experi:1 rubin:1 thresholding:2 principle:1 intractability:1 row:1 placed:1 last:1 free:4 electronically:1 infeasible:1 figueiredo:1 neighbor:2 pni:1 taking:1 face:4 subimage:1 saul:1 sparse:50 overcome:2 heitz:1 adopts:1 made:1 adaptive:3 projected:1 ple:14 voxel:1 reconstructed:7 approximate:1 obtains:1 preferred:1 kullback:1 aperture:2 ml:1 global:3 reveals:1 assumed:1 xi:29 vectorization:2 latent:1 iterative:2 learn:2 robust:2 symmetry:1 meanwhile:1 domain:3 rue:1 pk:3 dense:1 main:1 constituted:4 whole:1 noise:6 hyperparameters:5 n2:1 augmented:1 site:4 enlarged:1 scattered:1 precision:33 position:1 comput:1 governed:1 theorem:15 jensen:1 sensing:21 symbol:1 maxi:1 admits:1 nyu:1 ments:1 evidence:1 gupta:1 intractable:1 restricting:1 effectively:1 ci:5 magnitude:2 margin:2 chen:2 gap:7 chopper:2 photograph:1 expressed:1 partially:2 corresponds:1 relies:2 conditional:7 identity:1 goal:1 donoho:1 ann:2 change:1 specifically:1 total:1 partly:1 experimental:1 e:1 select:1 categorize:1 nayar:1 |
4,865 | 5,404 | On the Relationship Between LFP & Spiking Data
David E. Carlson1 , Jana Schaich Borg2 , Kafui Dzirasa2 , and Lawrence Carin1
1
Department of Electrical and Computer Engineering
2
Department of Psychiatry and Behavioral Sciences
Duke University
Duham, NC 27701
{david.carlson, jana.borg, kafui.dzirasa, lcarin}@duke.edu
Abstract
One of the goals of neuroscience is to identify neural networks that correlate with
important behaviors, environments, or genotypes. This work proposes a strategy
for identifying neural networks characterized by time- and frequency-dependent
connectivity patterns, using convolutional dictionary learning that links spike-train
data to local field potentials (LFPs) across multiple areas of the brain. Analytical
contributions are: (i) modeling dynamic relationships between LFPs and spikes;
(ii) describing the relationships between spikes and LFPs, by analyzing the ability
to predict LFP data from one region based on spiking information from across the
brain; and (iii) development of a clustering methodology that allows inference
of similarities in neurons from multiple regions. Results are based on data sets in
which spike and LFP data are recorded simultaneously from up to 16 brain regions
in a mouse.
1
Introduction
One of the most fundamental challenges in neuroscience is the ?large-scale integration problem?:
how does distributed neural activity lead to precise, unified cognitive moments [1]. This paper seeks
to examine this challenge from the perspective of extracellular electrodes inserted into the brain. An
extracellular electrode inserted into the brain picks up two types of signals: (1) the local field potential (LFP), which represents local oscillations in frequencies below 200 Hz; and (2) single neuron
action potentials (also known as ?spikes?), which typically occur in frequencies of 0.5 kHz. LFPs
represent network activity summed over long distances, whereas action potentials represent the precise activity of cells near the tip of an electrode. Although action potentials are often treated as the
?currency? of information transfer in the brain, relationships between behaviors and LFP activity
can be equally precise, and sometimes even more precise, than those with the activity of individual
neurons [2, 3]. Further, LFP network disruptions are highly implicated in many forms of psychiatric
disease [4]. This has led to much interest in understanding the mechanisms of how LFPs and action
potentials interact to create specific types of behaviors. New multisite recording techniques that
allow simultaneous recordings from a large number of brain regions provide unprecedented opportunities to study these interactions. However, this type of multi-dimensional data poses significant
challenges that require new analysis techniques.
Three of the most challenging characteristics of multisite recordings are that: 1) the networks they
represent are dynamic in space and time, 2) subpopulations of neurons within a local area can have
different functions and may therefore relate to LFP oscillations in specific ways, and 3) different frequencies of LFP oscillations often relate to single neurons in specific ways [5]. Here new models are
proposed to examine the relationship between neurons and neural networks that accommodate these
characteristics. First, each LFP in a brain region is modeled as convolutions between a bounded-time
dictionary element and the observed spike trains. Critically, the convolutional factors are allowed
to be dynamic, by binning the LFP and spike time series, and modeling the dictionary element for
1
each bin of the time series. Next, a clustering model is proposed making each neuron?s dictionary
element a scaled version of an autoregressive template shared among all neurons in a cluster. This
allows one to identify sub-populations of neurons that have similar dynamics over their functional
connectivity to a brain region. Finally, we provide a strategy for exploring which frequency bands
characterize spike-to-LFP functional connectivity. We show, using two novel multi-region electrophysiology datasets from mice, how these models can be used to identify coordinated interactions
within and between different neuronal subsystems, defined jointly by the activity of single cells and
LFPs. These methods may lead to better understanding of the relationship between brain activity
and behavior, as well as the pathology underlying brain diseases.
2
Model
2.1 Data and notation
The data used here consists of multiple LFP and spike-train time series, measured simultaneously
from M regions of a mouse brain. Spike sorting is performed on the spiking data by a VB implementation of [6], from which J single units are assumed detected from across the multiple regions
(henceforth we refer to single units as ?neurons?); the number of observed neurons J depends on
the data considered, and is inferred as discussed in [6]. Since multiple microwires are inserted into
single brain regions in our experiments (described in [7]), we typically detect between 4-50 neurons
for each of the M regions in which the microwires are inserted (discussed further when presenting
results). The analysis objective is to examine the degree to which one may relate (predict) the LFP
data from one brain region using the J-neuron spiking data from all brain regions. This analysis allows the identification of multi-site neural networks through the examination of the degree to which
neurons in one region are predictive of LFPs in another.
Let x ? RT represent a time series of LFP data measured from a particular brain region. The T
samples are recorded on a regular grid, with temporal interval ?. The spike trains from J different neurons (after sorting) are represented by the set of vectors {y1 , . . . , yJ }, binned in the same
manner temporally as the LFP data. Each yj ? ZT+ is reflective of the number of times neuron
j ? {1, . . . , J} fired within each of the T time bins, where Z+ represents nonnegative integers.
In the proposed model LFP data x are represented as a superposition of signals associated with each
neuron yj , plus a residual that captures LFP signal unrelated to the spiking data. The contribution
to x from information in yj is assumed generated by the convolution of yj with a bounded-time
dictionary element dj (residing within the interval -L to L, with L T ). This model is related to
convolutional dictionary learning [8], where the observed (after spike sorting) signal yj represents
the signal we convolve the learned dictionary dj against.
We model dj as time evolving, motivated by the expectation that neuron j may contribute differently
to specified LFP data, based upon the latent state of the brain (which will be related to observed
animal activity). The time series x is binned into a set of B equal-size contiguous windows, where
x = vec([x1 , . . . , xB ]), and likewise y = vec([yj1 , . . . , yjB ]). The dictionary element for neuron
j is similarly binned as {dj1 , . . . , djB }, and the contribution of neuron j to xb is represented as a
convolution of djb and yjb . This bin size is a trade-off between how finely time is discretized and
the computational costs.
In the experiments, in one example the bins are chosen to be 30 seconds wide (novel-environment
data) and in the other 1 minute (sleep-cycle data), and these are principally chosen for computational
convenience (the second data set is nine times longer). Similar results were found with windows as
narrow as 10 second, or as wide as 2 minutes.
2.2 Modeling the LFP contribution of multiple neurons jointly
Given {y1 , . . . , yJ }, the LFP voltage at time window b is represented as
xb =
J
X
yjb ? djb + b
(1)
j=1
where ? represents the convolution operator. Let Dj = [dj1 , . . . , djB ] ? R(2L+1)?B represent
the sequence of dictionary elements used to represent the LFP data over the B windows, from the
perspective of neuron j. We impose the clustering prior
Dj = ?j Aj , Aj ? G, G ? DP(?, G0 )
2
(2)
where G is a draw from a Dirichlet process (DP) [9, 10], with scale parameter ? > 0 and base
probability measure G0 . Note that we cluster the shape of the dictionary elements, and each neuron
has its own scaling ? ? R. Concerning the base measure, we impose an autoregressive prior on the
temporal dynamics, and therefore G0 is defined by an AR(?, ?) process
ab = ?ab?1 + ?t , ?t ? N (0, ? ?1 I)
(3)
where I is the identity matrix. ThisP
AR prior is used to constitute the B columns of the DP ?atoms?
?
A?h = (a?h1 , . . . , a?hB ), with G = k=1 ?k ?A?k . The elements
of the vector ? = (?1 , ?2 , . . . ) are
Q
drawn from the ?stick-breaking? [9] process ?h = Vh i<h (1 ? Vi ) with Vh ? Beta(1, ?). We
place the prior Gamma(a? , b? ) on ?, and priors Uniform(0,1) and Gamma(a? , b? ) respectively on
? and ?. To complete the model, we place the prior N (0, ? ?1 I) on b , and ?j ? N (0, 1).
In the implementation, a truncated stick-breaking representation is employed for G, using K ?sticks?
(VK = 1), which simplifies the implementation and has been shown to be effective in practice [9] if
K is made large enough, and the size of K is inferred during the inference algorithm.
Special cases of this model are clear. For example, if the Aj are simply drawn i.i.d. from G0 , rather
than from the DP, each neuron is allowed to contribute its own unique dictionary shape to represent
xb , called a non-clustering model in the results. In [11] the authors considered a similar model, but
the time evolution of dj was not considered (each neuron was assumed to contribute in the same way
to represent the LFP, independent of time). Further, in [11] only a single neuron was considered, and
therefore no clustering was considered. A multi-neuron version of this model is inferred by setting
B = 1.
3
3.1
Inference
Mean-field Variational Inference
Letting ? = {z, ?, A1,...,K , V1,...,K , ?, ?, ?}, the full likelihood of the clustering model
p(x, ?)
=
B
Y
[p(xb |?)]
J
Y
[p(zj |?)p(?j )]
j=1
b=1
K
Y
[p(A?k |?, ?)p(Vk |?)] p(?, ?, ?)
(4)
k=1
The non-clustering model can be recovered by setting zj = ?j and the truncation level in the stickbreaking process K to J. The time-invariant model is recovered by setting the number of bins B
to 1, with or without clustering. The model of [11] is recovered by using a single bin and a single
neuron.
Many recent methods [12, 13] have been proposed to provide quick approximations to the Dirichlet
process mixture model. Critically, in these models the latent assignment variables are conditionally
independent when the DP parameters are given. However, in the proposed model this assumption
does not hold because the observation x is the superposition of the convolved draws from the Dirichlet process.
A factorized variational distribution q is proposed to approximate the posterior distribution, and
the non-clustering model arises as a special case of the clustering model. The inference to fit the
distribution q is based on Bayesian Hierarchical Clustering [13] and the VB Dirichlet Process SplitMerge method [12]. The proposed model does not fit in either of these frameworks, so a method
to learn K by merging clusters by adapting [12, 13] is presented in Section 3.1.1. The factorized
distribution q takes the form: "
#
Y
Y
Y
q(?) =
q(zj )
q(?jk ) q(?)q(?)q(?)
[q(A?k )q(Vk )]
(5)
j
k
k
Standard forms on these distributions are assumed, with q(zj ) = Categorical(rj ), q(?) =
? kB ), ??1
Gamma(a0? , b0? ), q(?) = N(0,1) (?
? , ???1 ), q(Ak ) = N (vec(Ak ); vec(?
ak1 , . . . , a
k ), ?k =
0
0
?k , and q(?) = Gamma(a? , b? ). To facilitate inference, the distribution on ?j is split into
?1
q(?jk ) = N (?jk , ?jk
), the variational distribution for ? on the j th spike train given that it is in
cluster k. The non-clustering model can be represented as a special case of the clustering model
where q(?jk ) = ?1 , and q(zj ) = ?j . As noted in [12], this factorized posterior has the property that
a q with K 0 clusters is nested in a representation of q for K clusters for K ? K 0 , so any number of
clusters up to K 0 is represented.
3
Variational algorithms find a q that minimizes the KL divergence from the true, intractable posterior
[14], finding a q that locally maximizes the evidence lower bound (ELBO) objective:
log p(x|?) ? L(q) = Eq [log p(x, z, ?, A?1,...,K , ?, ?, ?|?) ? log q(z, ?, A?1,...,K , ?, ?, ?)] (6)
To facilitate inference, approximations to p(y|?) are developed. Let Tb be the number of time
PTb
points in bin b, and define Rjib ? R(2L+1)?(2L+1) with entries Rjib,ik = T1b t=1
yjb,t yib,t+k?i ;
P
P
?j
? kb ), or
yjb,t is yj at time point t in window/bin b. Let xb = xb ? j 0 6=j yb ? ( k rjk ?jk a
th
the residual after all but the contribution from the j neuron have been removed, and define let
PTb
?j
j
?jb
? R2L+1 with entries ?ji
= T1b t=1
yjb,t xb,t+i for i ? {?L, . . . , L}. Both Rjb and ?jb
can be efficiently estimated with the FFT. For each time bin b, we can write: log p(x?j
b |yjb , djb ) =
PL
?j
?j T
? Tb
?
2
T
const ? 2 (xb,t ? `=?L yj,b,t+` dj,b,?` ) ' const ? 2 (djb Rjjb d ? 2(?jb ) djb )
P
P
0
0 ?
To define the key updates, let ykb
= j rjk ?jk yjb , and x?k
kb . ?kbb0 denotes
j 0 6=j yk ? a
b = xb ?
0
the block in ?k indexing the b and the b bins, which is efficiently calculated because ??1
k is a block
tri-diagonal matrix from the first-order autoregressive process, and explicit equations exist. Letting
?k = P rjk , then q(Vk ) is updated by are ak = 1 + N
? , bk = ?? + PK0
?0
N
j
k =k+1 Nk . For q(?jk ), the
P
?j
?1 P
T
? kb + ?kbb )), and ?jk = ?jk
? Tkb Rjb ?jb
parameters are updated ?jk = 1 + b trace(Rjb (?
akb a
.
ba
The clustering latent variables are updated sequentially by:
?X
?1
T
? kb a
? Tkb ))?2?jk (x?j
? kb ))]+Eq [q(?)]
[(?jk +?jk
)tr(Rjb (Tb ?kbb + a
log(rjk ) ? ?
b ) (yb Rjbb a
2
b
x?k
b
yb?k
and
can be used to calculate q(A?k ). The mean of the distribution q(Ak ) is evaluated
using the forward filtering-backward smoothing algorithm, and ??1
k is a block tridiagonal matrix,
enabling efficient computations. Further details on updating q(A?k ) are found in Section A of the
Supplemental Material. Approximating distributions q(?), q(?) and q(?) are standard [14, 15].
3.1.1 Merge steps
The model is initialized to K = J clusters and the algorithm first finds q for the non-clustering
model. This initialization is important because of the superposition measurement model. The algorithm proceeds to merge down to K 0 , where K 0 is a local mode of the VB algorithm. The procedure
is as follows: (i) Randomly choose two clusters k and k 0 to merge. (ii) Propose a new variational distribution q? with K ? 1 clusters. (iii) Calculate the change in the variational lower bound,
L(?
q ) ? L(q), and accept the merge if the variational lower bound increases. As in [12], intelligent
sampling of k and k 0 significantly improves performance. Here, we sample k and k 0 with weight
proportional to exp(?K(Ak , Ak0 ; c0 )), where K(?, ?; c0 ) is the radial basis function. In [13] all
pairwise clusterings were considered, but that is computationally infeasible in this problem. This
approach for merging clusters is similar to that developed in [12].
This algorithm requires efficient estimation of the difference in the lower bound. For a proposed k
and k 0 , a new variational distribution q? is proposed, with q?(zj = k) = q(zj = k) + q(zj = k 0 )
?
6=k0 ?
?k + N
?k0 , b0 + PK,k
?
0
and q?(zj = k 0 ) = 0, q?(?k ) = Beta(a0P+ P
N
k? =k+1 Nk ), q(?k ) = ?0 , and
q(Ak ) is calculated. Letting H(q) = ? j k rjk log rjk , the difference in the lower bound can
be calculated:
p(Ak |?, ?) p(?k ))
? H(?
q ) + H(?
p)
(7)
L(?
q ) ? L(q) = Eq? log p(y|A1,...,K , ?, ? )
q?(Ak ) q(?k )
p(Ak |?, ?)p(A0k |?, ?) p(?k )p(?k0 )
? Eq log p(y|A1,...,K , ?, ? )
+ H(q) ? H(p)
q(Ak )q(A0k )
q(?k )q(?k0 )
Explicit details on the calculations of these variables are found in Section A of the Supplementary
Material, and the block tridiagonal nature of ?k allows the complete calculation of this value in
?k + N
?k0 ) + L3 )). This is linear in the amount of data used in the model. The algorithm
O(BTb ((N
is stopped after 10 merges in a row are rejected.
3.2 Integrated Nested Laplacian Approximation for the Non-Clustering Model
The VB inference method assumes a separable posterior. In the non-clustering model, Integrated
Nested Laplacian Approximation (INLA) [16] was used to estimate of the joint posterior, without
4
Animal
1
2
3
4
5
6
Invariant
0.1394
0.1465
0.2251
0.0867
0.1238
0.0675
Non-Cluster
0.1968
0.2382
0.3050
0.1433
0.1867
0.1407
Clustering
0.2094
0.2340
0.3414
0.1434
0.1882
0.1351
Animal
7
8
9
10
11
Invariant
0.1385
0.0902
0.1597
0.0311
0.675
Non-Cluster
0.2567
0.3440
0.1881
0.0803
0.1064
Clustering
0.2442
0.3182
0.2362
0.0865
0.1161
Table 1: Mean held-out RFE of the multi-cell models predicting the Hippocampus LFP. ?Invariant? denotes the
time-invariant model, ?Non-cluster? and ?clustering? denote the dynamic model without and with clustering.
0.1
5 Min
15 Min
38 Min
Dynamic
0.02
0.01
0.05
0.5
Invar iant
Non-Clus ter
Clus ter ing
0.4
Hold-out RFE
Amplitude, a.u.
0.04
0.03
Joint Model Prediction in HP
Dictionary Element of a VTA Cell
Single Neuron Hold-out RFE
0.3
0
0.2
?0.05
0.1
0
0
0.01
0.02
0.03
Time-Invariant
0.04
?0.1
?0.5
0
Time, seconds
0.5
0
0
10
20
30
40
Experiment Time, Minutes
Figure 1: (Left) Mean single-cell holdout RFE predicting mouse 3?s Nucleus Accumbens LFP comparing the
dynamic and time-invariant model. Each point is a single neuron. (Middle) Convolutional dictionary for a
VTA cell predicting mouse 3?s Nucleus Accumbens LFP at 5 minutes, 15 minutes, and 38 minutes after the
experiment start. (Right) Hold-out RFE over experiment time with the time-invariant, non-clustering, and the
clustering model to predict mouse 3?s Hippocampus LFP.
assuming separability. Comparisons to INLA constitute an independent validation of VB, for inference in the non-clustering version of the model. The INLA inference procedure is detailed in
Supplemental Section B. INLA inference was found to be significantly slower than the VB approximation, so experimental results below are shown for VB. The INLA and VB predictive performance
were quantitatively similar for the non-clustering model, providing confidence in the VB results.
4
Experiments
4.1 Results on Mice Introduced to a Novel Environment
This data set is from a group of 12 mice consisting of male Clock-?19 (mouse numbers 7-12)
and male wild-type littermate controls (mouse numbers 1-6) (further described in [7]). For each
animal, 32-48 total microwires were implanted, with 6-16 wires in each of the Nucleus Accumbens,
Hippocampus (HP), Prelimbic Cortex (PrL), Thalamus, and the Ventral Tegmental Area (VTA).
LFPs were averaged over all electrodes in an area and filtered from 3-50Hz and sampled at 125
Hz. Neuronal activity was recorded using a Multi-Neuron Acquisition Processor (Plexon). 99-192
individual spike trains (single units) were detected per animal. In this dataset animals begin in their
home cage, and after 10 minutes are placed in a novel environment for 30 minutes. For analysis, this
40 minute data sequence was binned into 30 second chunks, giving 80 bins. For all experiments we
choose L such that the dictionary element covered 0.5 seconds before and after each spike event.
Cross-validation was performed using leave-one-out analysis over time bins, using the error metric
of reduction in fractional error (RFE), 1 ? ||xb ? x?b ||22 /||xb ||22 . Figure 1(left) shows the average
hold-out RFE for the time-invariant model and the dynamic model for single spike train predicting
mouse 3?s Nucleus Accumbens, showing that the dynamic model can give strong improvements
on the scale of a single cell (these results are typical). The dynamic model has a higher hold-out
RFE on 98.4% of detected cells across all animals and all regions, indicating that the dynamic
model generally outperforms the time-invariant model. A dynamic dictionary element from a VTA
cell predicting mouse 3?s Nucleus Accumbens is shown in Figure 1(middle). At the beginning
of the experiment, this cell is linked with a slow, high-amplitude oscillation. After the animal is
initially placed into a new environment (illustrated by the 15-minute data point), the amplitude of the
dictionary element drops close to zero. Once the animal becomes accustomed to its new environment
(illustrated by the 38-minute data point), the cell?s original periodic dictionary element begins to
appear again. This example shows how cells and LFPs clearly have time-evolving relationships.
The leave-one-out performance of the time-invariant, non-clustering, and clustering models predicting animal 3?s Hippocampus LFP with 182 neurons is shown in Figure 1(right). These results show
5
10
5
0
Accumbens HP
PrL Thalamus VTA
0
?0.02
10
20
30
Experiment Time, Minutes
Cluster?s Cell Locations
Number of Cells
40
0.02
40
6
4
2
0
Accumbens HP
Dictionary, Seconds
?0.02
Cluster Factor Evolution
?0.4
?0.2
0
0.2
0.4
Cluster Factor Evolution
PrL Thalamus VTA
0.05
?0.4
?0.2
0
0.2
0.4
Number of Cells
0
Dictionary, Seconds
0.02
10
20
30
Experiment Time, Minutes
Cluster?s Cell Locations
Number of Cells
Dictionary, Seconds
Cluster Factor Evolution
?0.4
?0.2
0
0.2
0.4
0
10
20
30
Experiment Time, Minutes
Cluster?s Cell Locations
40
?0.05
20
10
0
Accumbens HP
PrL Thalamus VTA
Figure 2: Example clusters predicting mouse 3?s Hippocampus LFP. The top part shows the convolutional factor throughout the duration of the experiment, and the bottom part shows the location of the cells in the cluster.
Some of the clusters are dynamic whereas others were consistent through the duration of the experiment.
Hippocampus Cells Predicting Thalamus LFP 25-35Hz
0.6
8
13
0.5
18
0.4
Cluster Contribution
Raw Energy
Residual
500
8
28
0.2
33
300
200
18
0.15
23
0.1
28
33
0.1
38
43
Frequency, Hz
0.3
RFE
23
0.2
13
400
Energy, a.u.
Frequency, Hz
600
RFE
Hipp ocampus Cells Predicting Thalamus LFP
10
20
30
Exp erimental Time, min
40
100
0
0
0.05
38
43
10
20
30
Experimental Time, min
40
10
20
30
Exp erimental Time, min
40
Figure 3: (Left) RFE as a function of time bin and frequency bin for all Hippocampus cells predicting the
Thalamus LFP. There is a change in the predictive properties around 10 minutes. (Middle) Total energy versus
the unexplained residual for the Hippocampus cells predicting the Thalamus LFP for the frequency band 25-35
Hz. (Right) RFE using only the cluster of cells shown in Figure 2(right).
that predictability changes over time, and indicate that there is a strong increase in LFP predictability
when the mouse is placed in the novel environment. Using dynamics improves the results dramatically, and the clustering hold-out results showed further improvements in hold-out performance.
The mean hold-out RFE results for the Hippocampus for 11 animals are shown in Table 1 (1 animal
was missing this region recording). Results for other regions are shown in Supplemental Tables 1,
2, 3, and 4, and show similar results.
In this dataset, there is little quantitative difference between the clustering and non-clustering models; however, the clustering result is much better for interpretation. One reason for this is that
spike-sorting procedures are notoriously imprecise, and often under- or over-cluster. A clustering
model with equivalent performance is evidence that many neurons have the same shapes and dynamics, and repeated dynamic patterns reduces concerns that dynamics are the result of failure to
distinguish distinct neurons. Similarly, clustering of neuron shapes in a single electrode could be
the result of over-clustering from the spike-sorting algorithm, but clustering across electrodes gives
strong evidence that truly different neurons are clustering together. Additionally, neural action potential shapes drift over time [6, 17], but since cells in a cluster come from different electrodes and
regions, this is strong evidence that the dynamics are not due to over-sorting drifting neurons.
Each cluster has both a dynamic shape result as well as well as a neural distribution over regions.
Example clustering shapes and histogram cell locations for clusters predicting mouse 3?s Thalamus
LFP are shown in Figure 2. The top part of this figure shows the base dictionary element evolution
over the duration of the experiment. Note that both the (left) and (middle) plots show a dynamic
effect around 10 minutes, and the cells primarily come from the Ventral Tegmental Area. The (right)
plot shows a fairly stable factor, and its cells are mostly in the Hippocampus region.
The ability to predict the LFP constitutes functional connectivity between a neuron and the neuronal
circuit around the electrode for the LFP [18]. Neural circuits have been shown to transfer information
through specific frequencies of oscillations, so it is of scientific interest to know the functional
connectivity of a group of neurons as a function of frequency [5]. Frequency relationships were
explored by filtering the LFP signal after the predicted signal has been removed, using a notch filter
at 1 Hz intervals with a 1 Hz bandwidth, and the RFE was calculated for each held-out time bin and
frequency bin.
All cells in the Thalamus were used to predict each frequency band in mouse 3?s Hippocampus LFP,
and this result is shown in Figure 3(left). This figure shows an increase in RFE of the 25-35 Hz band
after the animal has been moved to a new location. The RFE on the band from 25-35 Hz is shown
6
Region
Time-Invariant
Non-Clustering
Clustering
PrLCx
0.1055
0.1686
0.1749
MOFCCx
0.1304
0.1994
0.2029
NAcShell
0.0904
0.1599
0.1609
NAcCore
0.1076
0.1796
0.1814
Amyg
0.0883
0.1422
0.1390
Hipp
0.2091
0.2662
0.2798
V1
0.1366
0.1972
0.2020
VTA
0.1317
0.1907
0.1923
Region
Time-Invariant
Non-Clustering
Clustering
Subnigra
0.1309
0.1939
0.1950
Thal
0.1550
0.2188
0.2204
LHb
0.1240
0.1801
0.1813
DLS
0.1237
0.1973
0.2012
DMS
0.1518
0.2363
0.2378
M1
0.1350
0.2034
0.2080
OFC
0.1878
0.2695
0.2723
FrA
0.1164
0.1894
0.1912
Table 2: Mean held-out RFE of the animal going through sleep cycles in each region.
Mean Factors for Cell in V1
Mean Factors for Cell in NAcShell
0.04
0.03
0.1
0.02
0.02
0.05
0
?0.05
?0.1
V1
HP
MDThal
VTA
?0.15
?0.2
?0.5
0
Time, Seconds
0.5
0
?0.02
?0.04
V1
HP
MDThal
VTA
?0.06
?0.08
?0.5
0
Time, Seconds
0.5
Amplitude, a.u.
0.15
Amplitude, a.u.
Amplitude, a.u.
Mean Factors for Cell in HP
0.01
0
?0.01
?0.02
V1
HP
MDThal
VTA
?0.03
?0.04
?0.5
0
Time, Seconds
0.5
Figure 4: The predictive patterns of individual neurons predicting multiple regions. (Left) A Hippocampus
cell is the best single cell predictor of the V1 LFP (Middle) A V1 cell with a relationship only to the V1 LFP.
(Right) A Nucleus Accumbens Shell cell that is equivalent in predictive ability to the best V1 cell.
in Figure 3(middle), and shows that while the raw energy in this frequency band is much higher
after the move to the novel environment, the cells from the Hippocampus can explain much of the
additional energy in this band. In Figure 3(right), we show the same result using only the cluster in
Figure 2. Note that there is a change around 10 minutes that is due to both a slight change in the
convolutional dictionary and a change in the neural firing patterns.
4.2 Results on Sleep Data Set
The second data set was recorded from one mouse going through different sleep cycles over 6 hours.
64 microwires were implanted in 16 different regions of the brain, using the Prelimbix Cortex (PrL),
Medial Orbital Frontal Cortex (MOFCCx), the core and shell of the Nucleus Accumbens (NAc),
Basal Amygdala (Amy), Hippocampus (Hipp), V1, Ventral Tegmental Area (VTA), Substantia nigra (Subnigra), Medial Dorsal Thalamus (MDThal), Lateral Habenula (LHb), Dorsolateral Striatum (DLS), Dorsomedial Striatum (DMS), Motor Cortex (M1), Orbital Frontal Cortex (OFC), and
Frontal Association Cortex (FrA). LFPs were averaged over all electrodes in an area and filtered
from 3-50Hz and sampled at 125Hz, and L was set to 0.5 seconds. 163 total neurons (single units)
were detected using spike sorting, and the data were split into 360 1-minute time bins. The leaveone-out predictive performance was higher for the dynamic single cell model on 159 out of 163
neurons predicting the Hippocampus LFP. The mean hold-out RFEs for all recorded regions of the
brain are shown in Table 2 for all models, and the clustering model is the best performing model in
15 of the 16 regions.
Previously published work looked at the predictability of the V1 LFP signal from individual V1
neurons [11,18,19]. Our experiments find that the dictionary elements for all V1 cells (4 electrodes,
4 cells in this dataset) are time-invariant and match the single-cell time-invariant dictionary shape
of [11]. The dictionary elements for a single V1 cell predicting multiple regions are shown in Figure
4(middle; for simplicity, only a subset of brain regions recorded from are shown). This suggests that
the V1 cell has a connection to the V1 region, but no other brain region that was recorded from in
this model. However, cells in other brain regions showed functional connectivity to V1. The best
individual predictor is a cell in the Hippocampus shown in Figure 4(left). An additional example
cell is a cell in the Nucleus Accumbens shell that has the same RFE as the best V1 cell, and its shape
is shown in Figure 4(right).
Sleep states are typically defined by dynamic changes in functional connectivity across brain regions
as measured by EEG (LFPs recorded from the scalp) [20], but little is known about how single neurons contribute to, or interact with, these network changes. To get sleep covariates, each second of
data was scored into ?awake? or ?sleep? states using the methods in [21], and the sleep state was
averaged over the time bin. We defined a time bin to be a sleep state if ? 95% of the individual sec7
0
?0.02
?0.2
0
0.2
Dictionary Element, s
0.4
6
4
Su
rL
A
P
M
O
D
Fr
0
C
2
F
Numb er of Cells
P
H
rL
0.02
?0.04
?0.4
0.4
5
P
TA
yg
1
m
V
0
0.2
Dictionary Element, s
10
0
V
M
A
T
ha
FC
0
l
2
?0.2
Awake
Sleep
LS
?0.05
?0.4
0.4
0.04
bn
0.2
Numb er of Cells
0
Time, s
4
O
Number of Cells
?0.2
0
ra
?0.1
Awake
Sleep
ig
?0.05
0.05
Amplitude, a.u.
Amplitude, a.u.
Amplitude, a.u.
0
?0.4
Anti-Sleep Cluster
Pro-Sleep Cluster
Cluster predicting V1 Region
0.05
Figure 5: (Left) The cluster predicting the V1 region of the brain, matching known pattern for individual V1
cells [11, 18]. (Middle,Right) Clusters predicting the motor cortex that show positive (pro) and negative (anti)
relationships between amplitude and sleep.
Sleep-Increased Cluster RFE by Frequency
Sleep-Neutral Cluster RFE by Frequency
0.025
Awake
Sleep
0.01
0.03
0.1
0.02
0.05
0.005
10
20
30
Frequency, Hz
40
50
Awake
Sleep
Mean RFE
0.015
Sleep-Decreased Cluster RFE by Frequency
0.05
0.04
0.15
Mean RFE
Mean RFE
0.02
0
0
0.2
Awake
Sleep
0
0
0.01
10
20
30
Frequency, Hz
40
50
0
0
10
20
30
Frequency, Hz
40
50
Figure 6: Mean RFE when the animal is awake and when it is asleep. (Left) Cluster?s convolution factor is
stable, and shows only minor differences between sleep and awake prediction. (Middle and Right) Clusters
shown in Figure 5 (left and right), depicting varying patterns with the mouse?s sleep state
onds are scored as a sleep state, and the animal is awake if ? 5% of the individual seconds are scored
as a sleep state. In Figure 5(middle) we show a cluster that is most strongly positively correlated
with sleep (pro-sleep), and in Figure 5(right) we show a cluster that is most negatively correlated
with sleep (pro-awake). Both figures show the neuron locations as well as the mean waveform shape
during sleep and wake. In this case, the pro-sleep cluster is dominantly Hippocampus cells and the
anti-sleep cluster comes from many different regions. There may be concern that because these are
the maximally correlated clusters, that these results may be atypical. To address this concern, the
p-value for finding a cluster this strongly correlated has a p-value 4 ? 10?6 for Pearson correlation
with the Bonferroni correction for multiple tests. Furthermore, 4 of the 25 clusters detected showed
correlation above .4 between amplitude and sleep state, so this is not an isolated phenomena.
The RFE changes as both a function of frequency and sleep state for some clusters of neurons. Using
1Hz bandwidth frequency bins, in Figure 6 (middle and right) we show the mean RFE using only
the clusters in Figure 5 (middle and right). The cluster associated positively with sleeping shifts
its frequency peak and increases its ability to predict when the animal is sleeping. Likewise, the
sleep-decreased cluster performs worst at predicting when the animal is asleep. For comparison, in
Figure 6 (left) we include the frequency results for cluster with a stable dictionary element. The
total RFE is comparable and there is a not a dramatic shift in the peak frequency between the sleep
and awake states.
5
Conclusions
Novel models and methods are developed here to account for time-varying relationships between
neurons and LFPs. Within the context of our experiments, significantly improved predictive performance is realized when one accounts for temporal dynamics in the neuron-LFP interrelationship.
Further, the clustering model reveals which neurons have similar relationships to a specific brain region, and the frequencies that are predictable in the LFP change with known dynamics of the animal
state. In future work, these ideas can be incorporated with attempts to learn network structure, and
LFPs can be considered a common input when exploring networks of neurons [19, 22, 23]. Moreover, future experiments are being designed to place additional electrodes in a single brain region,
with the goal of detecting 100 neurons in a single brain region while recording LFPs in up to 20
regions. The methods proposed here will facilitate exploration of both the diversity of neurons and
the differences in functional connectivity on an individual neuron scale.
Acknowlegements The research reported here was funded in part by ARO, DARPA, DOE, NGA
and ONR. We thank the reviewers for their helpful comments.
8
References
[1] F. Varela, J.P. Lachaux, E. Rodriguez, and J. Martinerie. The brainweb: phase synchronization and largescale integration. Nature Rev. Neuro., 2001.
[2] B. Pesaran, J.S. Pezaris, M. Sahani, P.P. Mitra, and R.A. Andersen. Temporal structure in neuronal activity
during working memory in macaque parietal cortex. Nature neuroscience, 5:805?811, 2002.
[3] C. Mehring, J. Rickert, E. Vaadia, S. Cardosa de Oliveira, A. Aertsen, and S. Rotter. Inference of hand
movements from local field potentials in monkey motor cortex. Nature neuroscience, 6(12):1253?1254,
2003.
[4] P.J. Uhlhaas and W. Singer. Abnormal neural oscillations and synchrony in schizophrenia. Nature Rev.
Neuro., 2010.
[5] M. Le Van Quyen and A. Bragin. Analysis of dynamic brain oscillations: methodological advances.
Trends in Neurosciences, 2007.
[6] D.E. Carlson, J.T. Vogelstein, Q. Wu, W. Lian, M. Zhou, C.R. Stoetzner, D. Kipke, D. Weber, D.B.
Dunson, and L. Carin. Multichannel Electrophysiological Spike Sorting via Joint Dictionary Learning &
Mixture Modeling. IEEE TBME, 2013.
[7] K Dzirasa, L Coque, MM Sidor, S Kumar, EA Dancy, JS Takahashi, C.A. McClung, and M.A.L. Nicolelis.
Lithium ameliorates nucleus accumbens phase-signaling dysfunction in a genetic mouse model of mania.
J. Neurosci., December 2010.
[8] B. Chen, G. Polatkan, G. Sapiro, D. Dunson, and L. Carin. The hierarchical beta process for convolutional
factor analysis and deep learning. ICML, 2011.
[9] H. Ishwaran and L.F. James. Gibbs Sampling Methods for Stick-Breaking Priors. JASA, March 2001.
[10] T.S. Ferguson. A Bayesian Analysis of Some Nonparametric Problems. Annals Stat., March 1973.
[11] M. Rasch, N.K. Logothetis, and G. Kreiman. From neurons to circuits: linear estimation of local field
potentials. J. Neurosci., November 2009.
[12] M.C. Hughes and E.B. Sudderth. Memoized Online Variational Inference for Dirichlet Process Mixture
Models. NIPS, 2013.
[13] K.A. Heller and Z. Ghahramani. Bayesian Hierarchical Clustering. ICML, 2005.
[14] D.M. Blei and M.I. Jordan. Variational inference for Dirichlet process mixtures. Bayesian Analysis, 2006.
[15] S.J. Roberts and W.D. Penny. Variational Bayes for generalized autoregressive models. IEEE Trans.
Signal Process., September 2002.
[16] H. Rue, S. Martino, and N. Chopin. Approximate Bayesian inference for latent Gaussian models by using
integrated nested Laplace approximations. J. Royal Stat. Soc., 2009.
[17] A. Calabrese and L. Paniski. Kalman filter mixture model for spike sorting of non-stationary data. J.
Neurosci. Methods, 2010.
[18] I. Nauhaus, L. Busse, M. Carandini, and D.L. Ringach. Stimulus contrast modulates functional connectivity in visual cortex. Nature Neuro., January 2009.
[19] R.C. Kelly, M.A. Smith, R.E. Kass, and T.S. Lee. Local field potentials indicate network state and account
for neuronal response variability. J. Comp. Neurosci., December 2010.
[20] L.J. Larson-Prior, J.M. Zempel, T.S. Nolan, F.W. Prior, A.Z. Snyder, and M.E. Raichle. Cortical network
functional connectivity in the descent to sleep. PNAS, March 2009.
[21] K. Dzirasa, S. Ribeiro, R. Costa, L.M. Santos, S.-C. Lin, A. Grosmark, T.D. Sotnikova, R.R. Gainetdinov,
M.G. Caron, and M.A.L. Nicolelis. Dopaminergic control of sleep-wake states. J. Neurosci., October
2006.
[22] J.W. Pillow and P. Latham. Neural characterization in partially observed populations of spiking neurons.
NIPS, 2007.
[23] M.J. Rasch, A. Gretton, Y. Murayama, W. Maass, and N.K. Logothetis. Inferring spike trains from local
field potentials. J. Neurophysiology, March 2008.
[24] D.I. Kim, P. Gopalan, D.M. Blei, and E.B. Sudderth. Efficient Online Inference for Bayesian Nonparametric Relational Models. NIPS, 2012.
9
| 5404 |@word neurophysiology:1 version:3 middle:12 hippocampus:17 c0:2 seek:1 bn:1 splitmerge:1 pick:1 dramatic:1 acknowlegements:1 tr:1 accommodate:1 reduction:1 moment:1 series:5 genetic:1 outperforms:1 recovered:3 comparing:1 ka:1 shape:10 motor:3 drop:1 plot:2 update:1 medial:2 designed:1 stationary:1 beginning:1 smith:1 core:1 filtered:2 blei:2 detecting:1 characterization:1 contribute:4 location:7 borg:1 beta:3 ik:1 consists:1 wild:1 behavioral:1 manner:1 orbital:2 pairwise:1 ra:1 behavior:4 examine:3 busse:1 multi:6 brain:28 discretized:1 ptb:2 little:2 window:5 becomes:1 begin:2 unrelated:1 bounded:2 underlying:1 notation:1 factorized:3 moreover:1 circuit:3 santos:1 maximizes:1 minimizes:1 monkey:1 accumbens:12 developed:3 supplemental:3 unified:1 finding:2 temporal:4 quantitative:1 yjb:8 sapiro:1 onds:1 scaled:1 stick:4 control:2 unit:4 appear:1 before:1 positive:1 engineering:1 local:9 mitra:1 striatum:2 ak:10 analyzing:1 firing:1 merge:4 plus:1 initialization:1 suggests:1 challenging:1 averaged:3 unique:1 yj:9 lfp:44 practice:1 block:4 hughes:1 signaling:1 lcarin:1 procedure:3 substantia:1 area:7 polatkan:1 evolving:2 adapting:1 significantly:3 matching:1 imprecise:1 confidence:1 radial:1 subpopulation:1 regular:1 psychiatric:1 get:1 convenience:1 close:1 subsystem:1 operator:1 context:1 equivalent:2 mcclung:1 quick:1 missing:1 reviewer:1 pk0:1 kbb:2 duration:3 l:1 simplicity:1 identifying:1 amy:1 population:2 laplace:1 updated:3 annals:1 logothetis:2 duke:2 element:19 trend:1 jk:14 updating:1 binning:1 observed:5 inserted:4 bottom:1 electrical:1 capture:1 worst:1 calculate:2 region:41 cycle:3 trade:1 removed:2 movement:1 yk:1 disease:2 environment:8 predictable:1 covariates:1 dynamic:26 predictive:7 upon:1 negatively:1 basis:1 joint:3 darpa:1 differently:1 k0:5 represented:6 train:8 distinct:1 effective:1 detected:5 pearson:1 supplementary:1 elbo:1 nolan:1 ability:4 dzirasa:3 lfps:14 jointly:2 online:2 sequence:2 vaadia:1 unprecedented:1 analytical:1 propose:1 aro:1 interaction:2 tbme:1 fr:1 murayama:1 fired:1 moved:1 rfe:28 electrode:11 cluster:52 leave:2 stat:2 pose:1 measured:3 minor:1 b0:2 thal:1 eq:4 strong:4 soc:1 predicted:1 indicate:2 come:3 rasch:2 waveform:1 filter:2 kb:6 exploration:1 a0k:2 material:2 bin:20 require:1 exploring:2 pl:1 correction:1 hold:10 mm:1 around:4 considered:7 residing:1 exp:3 lawrence:1 predict:6 inla:5 dictionary:29 ventral:3 numb:2 estimation:2 superposition:3 rjb:4 unexplained:1 stickbreaking:1 ofc:2 create:1 clearly:1 tegmental:3 gaussian:1 rather:1 martinerie:1 zhou:1 varying:2 voltage:1 vk:4 improvement:2 methodological:1 likelihood:1 martino:1 contrast:1 psychiatry:1 kim:1 detect:1 helpful:1 inference:16 dependent:1 ferguson:1 typically:3 integrated:3 a0:1 accept:1 initially:1 raichle:1 chopin:1 going:2 among:1 proposes:1 development:1 smoothing:1 animal:19 integration:2 summed:1 special:3 field:7 equal:1 once:1 fairly:1 vta:12 atom:1 sampling:2 represents:4 icml:2 constitutes:1 carin:2 future:2 jb:4 stimulus:1 others:1 intelligent:1 quantitatively:1 primarily:1 randomly:1 simultaneously:2 gamma:4 divergence:1 individual:9 phase:2 consisting:1 ab:2 attempt:1 interest:2 highly:1 male:2 mixture:5 truly:1 genotype:1 held:3 xb:12 initialized:1 isolated:1 stopped:1 increased:1 column:1 modeling:4 contiguous:1 ar:2 assignment:1 cost:1 sidor:1 subset:1 entry:2 neutral:1 uniform:1 predictor:2 tridiagonal:2 characterize:1 reported:1 periodic:1 chunk:1 fundamental:1 peak:2 lee:1 off:1 tip:1 together:1 mouse:19 yg:1 connectivity:10 again:1 andersen:1 recorded:8 choose:2 henceforth:1 cognitive:1 account:3 potential:11 takahashi:1 diversity:1 de:1 coordinated:1 depends:1 vi:1 performed:2 h1:1 linked:1 start:1 bayes:1 synchrony:1 contribution:6 convolutional:7 characteristic:2 likewise:2 efficiently:2 identify:3 identification:1 bayesian:6 raw:2 critically:2 calabrese:1 notoriously:1 comp:1 published:1 processor:1 simultaneous:1 explain:1 against:1 failure:1 energy:5 acquisition:1 frequency:27 james:1 dm:2 nauhaus:1 associated:2 sampled:2 costa:1 holdout:1 dataset:3 carandini:1 fractional:1 improves:2 electrophysiological:1 amplitude:11 ea:1 higher:3 ta:1 methodology:1 response:1 maximally:1 improved:1 yb:3 evaluated:1 strongly:2 furthermore:1 rejected:1 clock:1 correlation:2 working:1 hand:1 su:1 cage:1 rodriguez:1 mode:1 aj:3 sotnikova:1 scientific:1 nac:1 facilitate:3 effect:1 true:1 evolution:5 maass:1 illustrated:2 pezaris:1 conditionally:1 ringach:1 during:3 bonferroni:1 dysfunction:1 jana:2 noted:1 larson:1 generalized:1 btb:1 presenting:1 complete:2 latham:1 performs:1 interrelationship:1 pro:5 disruption:1 weber:1 variational:11 novel:7 common:1 functional:9 spiking:6 ji:1 rl:2 khz:1 prelimbic:1 discussed:2 interpretation:1 m1:2 slight:1 association:1 significant:1 refer:1 measurement:1 caron:1 vec:4 gibbs:1 grid:1 similarly:2 hp:9 pathology:1 dj:7 funded:1 l3:1 stable:3 invar:1 similarity:1 longer:1 cortex:10 base:3 mania:1 dorsomedial:1 j:1 posterior:5 own:2 recent:1 showed:3 perspective:2 gainetdinov:1 onr:1 rotter:1 additional:3 impose:2 nigra:1 employed:1 brainweb:1 signal:9 ii:2 vogelstein:1 multiple:9 currency:1 pnas:1 gretton:1 reduces:1 full:1 rj:1 ing:1 thalamus:11 characterized:1 calculation:2 cross:1 long:1 match:1 lin:1 concerning:1 equally:1 schizophrenia:1 a1:3 laplacian:2 prediction:2 neuro:3 ameliorates:1 implanted:2 expectation:1 metric:1 histogram:1 represent:8 sometimes:1 cell:54 sleeping:2 whereas:2 interval:3 decreased:2 wake:2 sudderth:2 finely:1 tkb:2 tri:1 comment:1 hz:17 recording:5 thisp:1 december:2 jordan:1 integer:1 reflective:1 near:1 ter:2 prl:5 iii:2 enough:1 split:2 hb:1 fft:1 erimental:2 fit:2 bandwidth:2 simplifies:1 idea:1 shift:2 motivated:1 notch:1 nine:1 constitute:2 action:5 deep:1 dramatically:1 generally:1 clear:1 detailed:1 covered:1 gopalan:1 amount:1 nonparametric:2 oliveira:1 band:7 locally:1 multichannel:1 exist:1 zj:9 neuroscience:5 estimated:1 per:1 write:1 snyder:1 group:2 key:1 basal:1 varela:1 drawn:2 backward:1 v1:22 nga:1 place:3 throughout:1 wu:1 oscillation:7 dj1:2 draw:2 home:1 scaling:1 vb:9 dorsolateral:1 comparable:1 bound:5 abnormal:1 distinguish:1 sleep:36 nonnegative:1 activity:10 scalp:1 binned:4 occur:1 kreiman:1 awake:11 min:6 kumar:1 performing:1 separable:1 extracellular:2 dopaminergic:1 department:2 march:4 across:6 separability:1 rev:2 making:1 invariant:15 indexing:1 principally:1 computationally:1 equation:1 previously:1 describing:1 mechanism:1 microwires:4 know:1 letting:3 singer:1 yj1:1 ishwaran:1 hierarchical:3 slower:1 convolved:1 drifting:1 original:1 convolve:1 clustering:45 dirichlet:6 denotes:2 assumes:1 top:2 opportunity:1 include:1 const:2 carlson:2 giving:1 ghahramani:1 approximating:1 objective:2 g0:4 move:1 realized:1 spike:22 looked:1 strategy:2 rt:1 diagonal:1 aertsen:1 september:1 ak1:1 dp:5 r2l:1 distance:1 link:1 thank:1 lateral:1 reason:1 rjk:6 assuming:1 kalman:1 modeled:1 relationship:12 providing:1 nc:1 mostly:1 dunson:2 robert:1 october:1 relate:3 trace:1 negative:1 ba:1 implementation:3 lachaux:1 zt:1 neuron:56 quyen:1 convolution:5 datasets:1 observation:1 enabling:1 wire:1 descent:1 anti:3 parietal:1 truncated:1 november:1 january:1 relational:1 incorporated:1 precise:4 variability:1 y1:2 carin1:1 drift:1 inferred:3 david:2 bk:1 introduced:1 subnigra:2 specified:1 kl:1 connection:1 rickert:1 learned:1 narrow:1 merges:1 hour:1 macaque:1 nip:3 address:1 trans:1 proceeds:1 below:2 pattern:6 memoized:1 challenge:3 tb:3 royal:1 memory:1 event:1 treated:1 examination:1 nicolelis:2 predicting:19 largescale:1 residual:4 temporally:1 categorical:1 vh:2 sahani:1 prior:9 understanding:2 heller:1 kelly:1 synchronization:1 filtering:2 proportional:1 kipke:1 versus:1 validation:2 nucleus:9 degree:2 jasa:1 consistent:1 row:1 placed:3 truncation:1 infeasible:1 implicated:1 dominantly:1 allow:1 wide:2 template:1 leaveone:1 penny:1 distributed:1 van:1 calculated:4 ak0:1 amygdala:1 cortical:1 pillow:1 autoregressive:4 author:1 made:1 forward:1 ig:1 ribeiro:1 correlate:1 approximate:2 sequentially:1 hipp:3 reveals:1 assumed:4 mehring:1 latent:4 table:5 additionally:1 learn:2 transfer:2 nature:6 eeg:1 depicting:1 interact:2 rue:1 pk:1 accustomed:1 neurosci:5 scored:3 allowed:2 repeated:1 x1:1 neuronal:5 site:1 clus:2 positively:2 slow:1 predictability:3 sub:1 inferring:1 yib:1 explicit:2 atypical:1 breaking:3 minute:18 down:1 specific:5 fra:2 showing:1 er:2 explored:1 evidence:4 concern:3 intractable:1 dl:2 merging:2 modulates:1 stoetzner:1 nk:2 sorting:9 multisite:2 chen:1 led:1 electrophysiology:1 simply:1 fc:1 visual:1 partially:1 nested:4 grosmark:1 asleep:2 shell:3 goal:2 identity:1 shared:1 change:10 typical:1 kafui:2 called:1 total:4 experimental:2 indicating:1 arises:1 dorsal:1 frontal:3 lian:1 phenomenon:1 correlated:4 |
4,866 | 5,405 | A Synaptical Story of Persistent Activity with
Graded Lifetime in a Neural System
Yuanyuan Mi,
Luozheng Li
State Key Laboratory of Cognitive Neuroscience & Learning,
Beijing Normal University, Beijing 100875, China
[email protected], [email protected]
Dahui Wang
State Key Laboratory of Cognitive Neuroscience & Learning,
School of System Science, Beijing Normal University,Beijing 100875, China
[email protected]
Si Wu
State Key Laboratory of Cognitive Neuroscience & Learning,
IDG/McGovern Institute for Brain Research,
Beijing Normal University ,Beijing 100875, China
[email protected]
Abstract
Persistent activity refers to the phenomenon that cortical neurons keep firing even
after the stimulus triggering the initial neuronal responses is moved. Persistent
activity is widely believed to be the substrate for a neural system retaining a memory trace of the stimulus information. In a conventional view, persistent activity is
regarded as an attractor of the network dynamics, but it faces a challenge of how
to be closed properly. Here, in contrast to the view of attractor, we consider that
the stimulus information is encoded in a marginally unstable state of the network
which decays very slowly and exhibits persistent firing for a prolonged duration.
We propose a simple yet effective mechanism to achieve this goal, which utilizes
the property of short-term plasticity (STP) of neuronal synapses. STP has two
forms, short-term depression (STD) and short-term facilitation (STF), which have
opposite effects on retaining neuronal responses. We find that by properly combining STF and STD, a neural system can hold persistent activity of graded lifetime,
and that persistent activity fades away naturally without relying on an external
drive. The implications of these results on neural information representation are
discussed.
1
Introduction
Stimulus information is encoded in neuronal responses. Persistent activity refers to the phenomenon
that cortical neurons keep firing even after the stimulus triggering the initial neural responses is
removed [1, 2, 3]. It has been widely suggested that persistent activity is the substrate for a neural system to retain a memory trace of the stimulus information [4]. For instance, in the classical
delayed-response task where an animal needs to memorize the stimulus location for a given period of time before taking an action, it was found that neurons in the prefrontal cortex retained
high-frequency firing during this waiting period, indicating that persistent activity may serve as the
1
neural substrate of working memory [2]. Understanding the mechanism of how persistent activity is
generated in neural systems has been at the core of theoretical neuroscience for decades [5, 6, 7].
In a conventional view, persistent activity is regarded as an emergent property of network dynamics: neurons in a network are reciprocally connected with each other via excitatory synapses, which
form a positive feedback loop to maintain neural responses in the absence of an external drive; and
meanwhile a matched inhibition process suppresses otherwise explosive neural activities. Mathematically, this view is expressed as the dynamics of an attractor network, in which persistent activity
corresponds to a stationary state (i.e., an attractor) of the network. The notion of attractor dynamics
is appealing, which qualitatively describes a number of brain functions, but its detailed implementation in neural systems remains to be carefully evaluated.
A long-standing debate on the feasibility of attractor dynamics is on how to properly close the attractor states in a network: once a neural system is evolved into a self-sustained active state, it will
stay there forever until an external force pulls it out. Solutions including applying a strong global inhibitory input to shut-down all neurons simultaneously, or applying a strong global excitatory input
to excite all neurons and force them to fall into the refractory period simultaneously, were suggested [9], but none of them appears to be natural or feasible in all conditions. From the computational
point of view, it is also unnecessary for a neural system to hold a mathematically perfect attractor
state lasting forever. In reality, the brain only needs to hold the stimulus information for a finite
amount of time necessary for the task. For instance, in the delayed-response task, the animal only
needed to memorize the stimulus location for the waiting period [1].
To address the above issues, here we propose a novel mechanism to retain persistent activity in neural systems, which gives up the concept of prefect attractor, but rather consider that a neural system
is in a marginally unstable state which decays very slowly and exhibits persistent firing for a prolonged period. The proposed mechanism utilizes a general feature of neuronal interaction, i.e., the
short-term plasticity (STP) of synapses [10, 11]. STP has two forms: short-term depression (STD)
and short-term facilitation (STF). The former is due to depletion of neurotransmitters after neural
firing, and the latter is due to elevation of calcium level after neural firing which increases the release probability of neurotransmitters. STD and STP have opposite effects on retaining prolonged
neuronal responses: the former weakens neuronal interaction and hence tends to suppress neuronal
activities; whereas, the latter strengthens neuronal interaction and tends to enhance neuronal activities. Interestingly, we find that the interplay between the two processes endows a neural system
with the capacity of holding persistent activity with desirable properties, including: 1) the lifetime
of persistent activity can be arbitrarily long depending on the parameters; and 2) persistent activity
fades away naturally in a network without relying on an external force. The implications of these
results on neural information representation are discussed.
2
The Model
Without loss of generality, we consider a homogeneous network in which neurons are randomly and
sparsely connected with each other with a small probability p. The dynamics of a single neuron is
described by an integrate-and-fire process, which is given by
?
dvi
= ?(vi ? VL ) + Rm hi ,
dt
for i = 1 . . . N,
(1)
where vi is the membrane potential of the ith neuron and ? the membrane time constant. VL is the
resting potential. hi is the synaptic current and Rm the membrane resistance. A neuron fires when
its potential exceeds the threshold, i.e., vi > Vth , and after that vi is reset to be VL . N the number
of neurons.
The dynamics of the synaptic current is given by
1 ?
dhi
sp
?
ext
= ?hi +
?s
Jij u+
?(t ? text
i ),
j xj ?(t ? tj ) + I
dt
Np j
(2)
where ?s is the synaptic time constant, which is about 2 ? 5ms. Jij is the absolute synaptic efficacy
from neurons j to i. Jij = J0 if there is a connection from the neurons j to i, and Jij = 0 otherwise.
tsp
j denotes the spiking moment of the neuron j. All neurons in the network receive an external input
2
in the form of Poisson spike train. I ext represents the external input strength and text
the moment
i
of the Poisson spike train the neuron i receives.
The variables uj and xj measure, respectively, the STF and STD effects on the synapses of the jth
neuron, whose dynamics are given by [12, 13]
duj
dt
dxj
?d
dt
?f
sp
= ?uj + ?f U (1 ? u?
j )?(t ? tj ),
(3)
sp
?
= 1 ? xj ? ?d u+
j xj ?(t ? tj ),
(4)
?
where uj is the release probability of neurotransmitters, with u+
j and uj denoting, respectively,
the values of uj just after and just before the arrival of a spike. ?f is the time constant of STF.
?
U controls the increment of uj produced by a spike. Upon the arrival of a spike, u+
j = uj +
+
?
U (1 ? u?
j ). xj represents the fraction of available neurotransmitters, with xj and xj denoting,
respectively, the values of xj just after and just before the arrival of a spike. ?d is the recover time
?
+ ?
of neurotransmitters. Upon the arrival of a spike, x+
j = xj ? uj xj . The time constants ?f and ?d
are typically in the time order of hundreds to thousands of milliseconds, much larger than ? and ?s ,
that is, STP is a slow process compared to neural firing.
2.1
Mean-field approximation
As to be confirmed by simulation, neuronal firings in the state of persistent activity are irregular and
largely independent to each other. Therefore, we can assume that the responses of individual neurons
are statistically equivalent in the state of persistent activity. Under this mean-field approximation,
the dynamics of a single neuron, and so does the mean activity of the network, can be written as [7]
dh
(5)
= ?h + J0 uxR + I,
dt
du
?f
= ?u + ?f U (1 ? u)R,
(6)
dt
dx
?d
= 1 ? x ? ?d uxR,
(7)
dt
where the state variables are the same for all neurons. R is the firing rate of a neuron, which is also
the mean activity of the neuron ensemble. I = I ext ? denotes the external input with ? the rate of
the Poisson spike train. The exact relationship between the firing rate R and the synaptic input h is
difficult to obtain. Here, we assume it to be of the form,
?s
R = max(?h, 0),
(8)
with ? a positive constant.
3
The Mechanism
By using the mean-field model, we first elucidate the working mechanism underlying the generation
of persistent activity of finite lifetime. Later we carry out simulation to confirm the theoretical
analysis.
3.1
How to generate persistent activity of finite lifetime
For the illustration purpose, we only study the dynamics of the firing rate R and assume that the
variables u and x reach to their steady values instantly. This approximation is in general inaccurate,
since u and x are slow variables compared to R. Nevertheless, it gives us insight into understanding
the network dynamics.
By setting du/dt = 0 and dx/dt = 0 in Eqs.(6,7) and substituting them into Eqs.(5,8), we get that,
for I = 0 and R ? 0,
?s
dR
J0 ??f U R2
= ?R +
? F (R).
dt
1 + ?f U R + ?d ?f U R2
3
(9)
J 0<Jc
J >Jc
0
J =Jc
F(R)
0
R*
0
R
Figure 1: The steady states of the network, i.e., the solutions of Eq.(9), have three forms depending
on the parameter values. The three lines correspond to the different neuronal connection strenghths,
which are J0 = 4, 4.38, 5, respectively. The other parameters are: ?s = 5ms, ?d = 100ms, ?f =
700ms, ? = 1, U = 0.05 and Jc = 4.38.
(
)
?
Define a critical connection strength Jc ? 1 + 2 ?d /(?f U ) /?, which is the point the network
dynamics experiences saddle-node bifurcation (see Figure 1). Depending on the parameters, the
steady states of the network have three forms
? When J0 < Jc , F (R) = 0 has only one solution at R = 0, i.e., the network is only stable
at the silent state;
? When J0 > Jc , F (R) = 0 has three solutions, and the network can be stable at the silent
state and an active state;
? When J0 = Jc , F (R) = 0 has two solutions, one is the stable silent state, and the other is
a neutral stable state, referred to as R? .
The interesting behavior occurs at J0 = Jc? , i.e., J0 is slightly smaller than the critical connection
strength Jc . In this case, the network is only stable at the silent state. However, since near to the
state R? , F (R) is very close to zero (and so does |dR/dt|), the decay of the network activity is very
slow in this region (Figure 2A). Suppose that the network is initially at a state R > R? , under the
network dynamics, the system will take a considerable amount of time to pass through the state R?
before reaching to silence. This is manifested by that the decay of the network activity exhibits a
long plateau around R? before dropping to silence rapidly (Figure 2B). Thus, persistent activity of
finite lifetime is achieved.
The lifetime of persistent activity, which is dominated by the time of the network state passing
through the point R? , is calculated to be (see Appendix A),
T ??
2?s
F (R? )F ?? (R? )
,
(10)
where F ?? (R? ) = d2 F (R)/d2 R|R? . By varying the STP effects, such as ?d and ?f , the value of
F (R? )F ?? (R? ) is changed, and the lifetime of persistent activity can be adjusted.
3.2
Persistent activity of graded lifetime
We formally analyze the condition for the network holding persistent activity of finite lifetime.
Inspired by the result in the proceeding section, we focus on the parameter regime of J0 = Jc , i.e.,
the situation when the network has the stable silent state and a neutral stable active state.
Denote (R? , u? , x? ) to be the neutral stable state of the network at J0 = Jc . Linearizing the network
dynamics at this point, we obtain
(
)
(
)
R ? R?
R ? R?
d
u ? u?
? A u ? u?
,
(11)
dt
x ? x?
x ? x?
4
A
B
0
R(Hz)
F(R)
R*
R*
0
R
0
2
4
6
8
t(s)
Figure 2: Persistent activity of finite lifetime. Obtained by solving Eqs.(5-8). (A) When J0 = Jc? ,
the function F (R), and so does dR/dt, is very close to zero at the state R? . Around this point, the
network activity decays very slowly. The inset shows the fine structure in the vicinity of R? . (B) An
external input (indicated by the red bar) triggers the network response. After removing the external
input, the network activity first decays quickly, and then experiences a long plateau before dropping
to silence rapidly. The parameters are: ?s = 5ms, ?d = 10ms, ?f = 800ms, ? = 1, U = 0.5,
I = 10, Jc = 1.316 and J0 = 1.315.
where A is the Jacobian matrix (see Appendix B).
It turns out that the matrix A always has one eigenvector with vanishing eigenvalue, a property due to
that (R? , u? , x? ) is the neutral stable state of the network dynamics. As demonstrated in Sec.3.1, by
choosing J0 = Jc? , we expect that the network state will decay very slowly along the eigenvector of
vanishing eigenvalue, which we call the decay-direction. To ensure this always happens, it requires
that the real parts of the other two eigenvalues of A are negative, so that any perturbation of the
network state away from the decay-direction will be pulled back; otherwise, the network state may
approach to silence rapidly via other routes avoiding the state (R? , u? , x? ). This idea is illustrated
in Fig.3.
The condition for the real parts of the other two eigenvalues of A being smaller than zero is calculated to be (see Appendix B):
?
2
1
U
1
1
1
?
+
+
?
> 0.
(12)
?f ?d
?d ?f ?d
?d ?s 1 + ?f U
?f ?s
?d
This inequality together with J0 =
of finite lifetime.
Jc?
form the condition for the network holding persistent activity
R* R
3-D view
Decay-direction
t
Figure 3: Illustration of the slow-decaying process of the network activity. The network dynamics
experiences a long plateau before dropping to silence quickly. The inset presents a 3-D view of the
local dynamics in the plateau region, where the network state is attracted to the decay-direction to
ensure slow-decaying.
By solving the network dynamics Eqs.(5-8), we calculate how the lifetime of persistent activity
changes with the STP effect. Fig.4A presents the results of fixing U and J0 and varying ?d and
5
?f , We see that below the critical line J0 = Jc , which is the region for J0 > Jc , the network has
prefect attractor states never decaying; and above the critical line, the network has only the stable
silent state. Close to the critical line, the network activity decays slowly and displays persistent
activity of finite lifetime. Fig.4B shows a case that when the STF strength (?f ) is fixed, the lifetime
of persistent activity decreases with the STD strength (?d ). This is understandable, since STD tends
to suppress neuronal responses. Fig.4C shows a case that when ?d is fixed, the lifetime of persistent
activity increases with ?f , due to that STF enhances neuronal responses. These results demonstrate
that by regulating the effects of STF and STD, the lifetime of persistent activity can be adjusted.
A
Decay time (s)
B
Decay time (s)
C
attractor
10
5
0
0
0.5
1
1.5
1
1.5
?d (s)
10
5
0
0
0.5
?f (s)
Figure 4: (A). The lifetimes of the network states with respect to ?f and ?d when U and J0 are fixed.
We use an external input to trigger a strong response of the network and then remove the input. The
lifetime of a network state is measured from the offset of the external input to the moment when the
network returns to silence. The white line corresponds to the condition of J0 = Jc , below which
the network has attractors lasting forever; and above which, the lifetime of a network state gradually
decreases (coded by colour). (B) When ?f = 1250ms is fixed, the lifetime of persistent activity
decreases with ?d (the vertical dashed line in A). (C) When ?d = 260ms is fixed, the lifetime of
persistent activity increases with ?f (the horizontal dashed line in A). The other parameters are:
?s = 5ms, ? = 1, U = 0.05 and J0 = 5.
4
Simulation Results
We carry out simulation with the spiking neuron network model given by Eqs.(1-4) to further confirm
the above theoretical analysis. A homogenous network with N = 1000 neurons is used, and in the
network neurons are randomly and sparsely connected with each other with a probability p = 0.1.
At the state of persistent activity, neurons fire irregularly (the mean value of Coefficient of Variation
is 1.29)and largely independent to each other(the mean correlation of all spike train pairs is 0.30)
with each other (Fig.5A). Fig.5 present the examples of the network holding persistent activity with
varied lifetimes, through different combinations of STF and STD satisfying the condition Eq.(12).
5
Conclusions
In the present study, we have proposed a simple yet effective mechanism to generate persistent
activity of graded lifetime in a neural system. The proposed mechanism utilizes the property of STP,
a general feature of neuronal synapses, and that STF and STD have opposite effects on retaining
neuronal responses. We find that with properly combined STF and STD, a neural system can be in a
marginally unstable state which decays very slowly and exhibits persistent firing for a finite lifetime.
This persistent activity fades away naturally without relying on an external force, and hence avoids
the difficulty of closing an active state faced by the conventional attractor networks.
STP has been widely observed in the cortex and displays large diversity in different regions [14, 15, 16]. Compared to static ones, dynamical synapses with STP greatly enriches the response
patterns and dynamical behaviors of neural networks, which endows neural systems with information processing capacities which are otherwise difficult to implement using purely static synapses.
The research on the computational roles of STP is receiving increasing attention in the field [12]. In
6
A
C
D
B
E
Figure 5: The simulation results of the spiking neural network. (A) A raster plot of the responses of
50 example neurons randomly chosen from the network. The external input is applied for the first
0.5 second. The persistent activity lasts about 1100ms. The parameters are: ?f = 800ms, ?d =
500ms, U = 0.5, J0 = 28.6. (B) The firing rate of the network for the case (A). (C) An example
of persistent activity of negligible lifetime. The parameters are:?f = 800ms, ?d = 1800ms, U =
0.5, J0 = 28.6. (D) An example of persistent activity of around 400ms lifetime. The parameters
are:?f = 600ms, ?d = 500ms, U = 0.5, J0 = 28.6. (E) An example of the network holding an
attractor lasting forever. The parameters are: ?f = 800ms, ?d = 490ms, U = 0.5, J0 = 28.6.
terms of information presentation, a number of appealing functions contributed by STP were proposed. For instances, Mongillo et al. proposed an economical way of using the facilitated synapses
due to STF to realize working memory in the prefrontal cortex without recruiting neural firing [8];
Pfister et al. suggested that STP enables a neuron to estimate the membrane potential information
of the pre-synaptic neuron based on the spike train it receives [17]. Torres et al. found that STD
induces instability of attractor states in a network, which could be useful for memory searching [18];
Fung et al. found that STD enables a continuous attractor network to have a slow-decaying state in
the time order of STD, which could serve for passive sensory memory [19]. Here, our study reveals
that through combining STF and STD properly, a neural system can hold stimulus information for
an arbitrary time, serving for different computational purposes. In particular, STF tends to increase
the lifetime of persistent activity; whereas, STD tends to decrease the lifetime of persistent activity.
This property may justify the diverse distribution of STF and STD in different cortical regions. For
instances, in the prefrontal cortex where the stimulus information often needs to be held for a long
time in order to realize higher cognitive functions, such as working memory, STF is found to be
dominating; whereas, in the sensory cortex where the stimulus information will be forwarded to
higher cortical regions shortly, STD is found to be dominating. Furthermore, our findings suggest
that a neural system may actively regulate the combination of STF and STD, e.g., by applying appropriate neural modulators [10], so that it can hold the stimulus information for a flexible amount
of time depending on the actual computational requirement. Further experimental and theoretical
studies are needed to clarify these interesting issues.
6
Acknowledgments
This work is supported by grants from National Key Basic Research Program of China
(NO.2014CB846101), and National Foundation of Natural Science of China (No.11305112, Y.Y.M.;
No.31261160495, S.W.; No.31271169,D.H.W.), and the Fundamental Research Funds for the central Universities (No.31221003, S.W.), and SRFDP (No.20130003110022, S.W), and Natural Science Foundation of Jiangsu Province BK20130282.
7
Appendix A: The lifetime of persistent activity
Consider the network dynamics Eq.(9). When J0 = Jc , the network has a stable silent state (R = 0)
and an unstable active state, referred to as R? (Fig.1). We consider that J0 = Jc? . In this case,
F (R? ) is slightly smaller than zero (Fig.2A). Starting from a state R > R? , the network will take
a considerable amount of time to cross the point R? , since dR/dt is very small in this region, and
the network exhibits persistent activity for a considerable amount of time. We estimate the time
consuming for the network crossing the point R? .
According to Eq.(9), we have
?
?
?T
?R+
dt
=
?
R?
0
=
?
=
?
?s
dR
F (R)
?R+
?
?
R?
F (R? )
[
2?s
F (R? )F ?? (R? )
2?s
F (R? )F ?? (R? )
?s dR
,
+ (R ? R? )2 F ?? (R? )/2
?
R+
? R?
?
R?
? R?
]
arctg ?
? arctg ?
,
F (R? )/F ?? (R? )
F (R? )/F ?? (R? )
G(R? ),
(13)
?
?
where R+
and R?
denote, respectively, the points slightly larger or smaller than R? , F ? (R? ) =
dF (R)/dR|R? , and F ?? (R? ) = dF ? (R)/dR|R? . To get the above result, we used the second-order
Taylor expansion of F (R) at R? , and the condition F ? (R? ) = 0.
In the limit of F (R? ) ? 0, the value of G(R? ) is bounded. Thus, the lifetime of persistent activity
is in the order of
2?s
T ??
.
(14)
F (R? )F ?? (R? )
Appendix B: The condition for the network holding persistent activity of finite
lifetime
Denote (R? , u? , x? ) to be the neutral stable state of the network when J0 = Jc , which is calculated
to be (by solving Eqs.(5-8)),
?
?f U R?
1 + ?f U R?
?
R? = 1/?f ?d U , u? =
(15)
,
x
=
.
1 + ?f U R ?
1 + ?f U R? + ?f ?d U R? 2
Linearizing the network dynamics at this point, we obtain Eq.(12), in which the Jacobian matrix A
is given by
)
(
(J0 u? x? ? 1)/?s ,
J0 x? R? /?s ,
J0 u? R? /?s
?
?
U (1 ? u ),
?1/?f ? U R ,
0
.
(16)
A=
?u? x? ,
?x? R? ,
?1/?d ? u? R?
The eigenvalues of the Jacobian matrix satisfy the equality |A ? ?I| = 0. Utilizing Eqs.(15), this
equality becomes
?(?2 + b? + c?) = 0,
(17)
where the coefficients b and c are given by
b
=
c
=
1
1
+
+ u? R ? + U R ? ,
?d
?f
?
2
1
U
1
1
1
?
+
+
?
.
?f ?d
?d ?f ?d
?d ?s 1 + ?f U
?f ?s
(18)
(19)
?d
From Eq.(17), we see that the matrix A has three eigenvalues. One eigenvalue, referred to as ?1 , is
always zero. The other two eigenvalues satisfy that ?2 + ?3 = ?b and ?2 ?3 = c. Since b > 0, the
condition for the real parts of ?2 and ?3 being negative is c > 0.
8
References
[1] J. Fuster and G. Alexander. Neuron activity related to short-term memory. Science 173, 652654 (1971).
[2] S. Funahashi, C. J. Bruce and P.S. Goldman-Rakic. Mnemonic coding of visual space in the
monkeys dorsolateral prefrontal cortex. J. Neurophysiol. 61, 331-349 (1989).
[3] R. Romo, C. D. Brody, A. Hernandez. Lemus L. Neuronal correlates of parametric working
memory in the prefrontal cortex. Nature 399, 470-473 (1999).
[4] D.J. Amit. Modelling brain function. New York: Cambridge University Press. (1989)
[5] S. Amari. Dynamics of pattern formation in lateral-inhibition type neural fields. Biol. Cybern.
27, 77-87 (1977).
[6] X.J. Wang. Synaptic basis of cortical persistent activity: the importance of NMDA receptors
to working memory. J. Neurosci. 19, 9587-9603 (1999).
[7] O. Barak and M. Tsodyks. Persistent Activity in Neural Networks with Dynamic Synapses.
PLoS Computational Biology.3(2): e35(2007).
[8] G. Mongillo, O. Barak and M. Tsodyks. Synaptic theory of working memory. Science
319,1543-1546(2008).
[9] B. Gutkin, C. Laing, C. Colby, C. Chow, and B. Ermentrout. Turning on and off with excitation:
the role of spike-timing asynchrony and synchrony in sustained neural activity. J. Comput.
Neurosci. 11, 121C134 (2001).
[10] H. Markram and M. Tsodyks. Redistribution of synaptic efficacy between neocortical pyramidal neurons. Nature. 382(6594): 807-810(1996).
[11] L. F. Abbott and W. G. Regehr. Synaptic computation. Nature. 431(7010): 796-803(2004).
[12] M. Tsodyks and S. Wu. Short-Term Synaptic Plasticity. Scholarpedia, 8(10): 3153(2013).
[13] M. Tsodyks, K. Pawelzik and H. Markram. Neural Networks with Dynamic Synapses. Neural
Computation. 10(4): 821-835(1998).
[14] H. Markram, Y. Wang and M. Tsodyks. Differential signaling via the same axon of neocortical pyramidal neurons. Proceedings of the National Academy of Sciences. 95(9): 53235328(1998).
[15] J. S. Dittman, A. C. Kreitzer and W. G. Regehr. Interplay between facilitation, depression, and
residual calcium at three presynaptic terminals. J. Neurosci. 20: 1374-1385(2000).
[16] Y. Wang, H. Markram, P. H. Goodman, T. K. Berger, J. Y. Ma and P. S. Goldman-Rakic.
Heterogeneity in the pyramidal network of the medial prefrontal cortex. Nature Neuroscience.
9(4): 534-542(2006).
[17] J. P. Pfister, P. Dayan and M. Lengyel. Synapses with short-term plasticity are optimal estimators of presynaptic membrane potentials. Nature Neuroscience 13,1271-1275(2010).
[18] J.J. Torres, J. M. Cortes, J. Marro and H. J. Kappen. Competition Between Synaptic Depression
and Facilitation in Attractor Neural Networks. Neural Computation. 19(10): 2739-2755(2007).
[19] C. C. Fung, K. Y. Michael Wong, H. Wang and S. Wu. Dynamical Synapses Enhance Neural Information Processing: Gracefulness, Accuracy and Mobility. Neural Computation 24(5):11471185(2012).
9
| 5405 |@word d2:2 simulation:5 colby:1 carry:2 kappen:1 moment:3 initial:2 efficacy:2 idg:1 denoting:2 interestingly:1 current:2 com:1 si:1 yet:2 dx:2 written:1 attracted:1 realize:2 plasticity:4 enables:2 remove:1 plot:1 fund:1 medial:1 stationary:1 shut:1 ith:1 vanishing:2 short:9 core:1 funahashi:1 node:1 location:2 along:1 differential:1 persistent:52 sustained:2 behavior:2 brain:4 terminal:1 inspired:1 relying:3 goldman:2 prolonged:3 actual:1 pawelzik:1 increasing:1 becomes:1 matched:1 underlying:1 bounded:1 evolved:1 eigenvector:2 suppresses:1 monkey:1 finding:1 rm:2 control:1 grant:1 before:7 positive:2 negligible:1 local:1 timing:1 tends:5 limit:1 ext:3 receptor:1 firing:15 hernandez:1 china:5 statistically:1 acknowledgment:1 implement:1 signaling:1 j0:31 pre:1 refers:2 suggest:1 get:2 close:4 applying:3 instability:1 cybern:1 wong:1 conventional:3 equivalent:1 demonstrated:1 romo:1 attention:1 starting:1 duration:1 fade:3 insight:1 estimator:1 utilizing:1 regarded:2 pull:1 facilitation:4 searching:1 notion:1 variation:1 increment:1 elucidate:1 suppose:1 trigger:2 exact:1 substrate:3 homogeneous:1 crossing:1 strengthens:1 satisfying:1 std:19 sparsely:2 observed:1 role:2 wang:5 thousand:1 calculate:1 region:7 tsodyks:6 connected:3 plo:1 decrease:4 removed:1 ermentrout:1 dynamic:23 solving:3 serve:2 upon:2 purely:1 basis:1 neurophysiol:1 emergent:1 neurotransmitter:5 train:5 effective:2 modulators:1 mcgovern:1 formation:1 choosing:1 whose:1 encoded:2 widely:3 larger:2 dominating:2 otherwise:4 forwarded:1 amari:1 tsp:1 interplay:2 eigenvalue:8 propose:2 interaction:3 jij:4 reset:1 combining:2 loop:1 rapidly:3 achieve:1 academy:1 yuanyuan:1 moved:1 competition:1 requirement:1 perfect:1 weakens:1 depending:4 fixing:1 measured:1 school:1 eq:13 strong:3 memorize:2 direction:4 redistribution:1 elevation:1 mathematically:2 adjusted:2 clarify:1 hold:5 around:3 normal:3 substituting:1 recruiting:1 purpose:2 laing:1 always:3 rather:1 reaching:1 varying:2 release:2 focus:1 properly:5 modelling:1 lemus:1 greatly:1 contrast:1 dayan:1 vl:3 typically:1 inaccurate:1 chow:1 initially:1 synaptical:1 issue:2 stp:14 flexible:1 retaining:4 animal:2 bifurcation:1 homogenous:1 field:5 once:1 never:1 biology:1 represents:2 np:1 stimulus:13 randomly:3 simultaneously:2 national:3 individual:1 delayed:2 fire:3 attractor:17 maintain:1 explosive:1 regulating:1 tj:3 held:1 implication:2 necessary:1 experience:3 mobility:1 taylor:1 theoretical:4 instance:4 neutral:5 hundred:1 jiangsu:1 combined:1 fundamental:1 retain:2 standing:1 stay:1 off:1 receiving:1 michael:1 enhance:2 together:1 quickly:2 dhi:1 central:1 slowly:6 prefrontal:6 dr:8 cognitive:4 external:13 cb846101:1 return:1 li:1 actively:1 potential:5 diversity:1 sec:1 coding:1 coefficient:2 satisfy:2 jc:22 vi:4 scholarpedia:1 later:1 view:7 closed:1 analyze:1 red:1 recover:1 decaying:4 mongillo:2 bruce:1 synchrony:1 accuracy:1 largely:2 ensemble:1 correspond:1 produced:1 marginally:3 bnu:3 none:1 confirmed:1 drive:2 economical:1 lengyel:1 plateau:4 synapsis:12 reach:1 synaptic:12 raster:1 frequency:1 naturally:3 mi:1 static:2 nmda:1 carefully:1 back:1 appears:1 higher:2 dt:15 response:16 prefect:2 evaluated:1 generality:1 lifetime:32 just:4 furthermore:1 until:1 correlation:1 working:7 receives:2 horizontal:1 indicated:1 asynchrony:1 effect:7 concept:1 regehr:2 former:2 hence:2 vicinity:1 equality:2 laboratory:3 illustrated:1 white:1 rakic:2 during:1 self:1 steady:3 excitation:1 m:20 linearizing:2 neocortical:2 demonstrate:1 passive:1 novel:1 enriches:1 spiking:3 refractory:1 discussed:2 resting:1 cambridge:1 closing:1 gutkin:1 stable:12 cortex:8 inhibition:2 route:1 manifested:1 inequality:1 arbitrarily:1 period:5 dashed:2 desirable:1 exceeds:1 believed:1 long:6 cross:1 mnemonic:1 coded:1 feasibility:1 basic:1 poisson:3 df:2 achieved:1 irregular:1 receive:1 whereas:3 fine:1 pyramidal:3 goodman:1 hz:1 dxj:1 call:1 near:1 xj:10 triggering:2 opposite:3 silent:7 idea:1 cn:3 colour:1 resistance:1 passing:1 york:1 action:1 depression:4 useful:1 detailed:1 amount:5 induces:1 generate:2 inhibitory:1 millisecond:1 neuroscience:6 instantly:1 serving:1 diverse:1 dropping:3 waiting:2 key:4 threshold:1 nevertheless:1 abbott:1 fraction:1 beijing:6 facilitated:1 wu:3 utilizes:3 appendix:5 dorsolateral:1 brody:1 hi:3 display:2 activity:62 strength:5 dominated:1 fung:2 according:1 combination:2 membrane:5 describes:1 slightly:3 smaller:4 appealing:2 happens:1 lasting:3 gradually:1 depletion:1 remains:1 turn:1 mechanism:8 needed:2 irregularly:1 available:1 away:4 regulate:1 appropriate:1 shortly:1 denotes:2 ensure:2 uj:8 graded:4 amit:1 classical:1 spike:11 occurs:1 parametric:1 exhibit:5 enhances:1 lateral:1 capacity:2 mail:1 presynaptic:2 unstable:4 retained:1 relationship:1 illustration:2 berger:1 difficult:2 holding:6 debate:1 trace:2 negative:2 suppress:2 implementation:1 calcium:2 understandable:1 contributed:1 vertical:1 neuron:32 finite:10 situation:1 heterogeneity:1 perturbation:1 varied:1 arbitrary:1 pair:1 connection:4 address:1 suggested:3 bar:1 below:2 dynamical:3 pattern:2 regime:1 challenge:1 program:1 including:2 memory:11 reciprocally:1 max:1 critical:5 natural:3 force:4 endows:2 difficulty:1 turning:1 residual:1 text:2 faced:1 understanding:2 stf:17 loss:1 expect:1 generation:1 interesting:2 foundation:2 integrate:1 dvi:1 story:1 excitatory:2 changed:1 supported:1 last:1 jth:1 silence:6 pulled:1 barak:2 institute:1 fall:1 face:1 taking:1 arctg:2 absolute:1 markram:4 feedback:1 calculated:3 cortical:5 avoids:1 sensory:2 qualitatively:1 correlate:1 forever:4 gracefulness:1 keep:2 confirm:2 global:2 active:5 reveals:1 unnecessary:1 excite:1 consuming:1 continuous:1 fuster:1 decade:1 reality:1 nature:5 du:2 expansion:1 meanwhile:1 sp:3 neurosci:3 arrival:4 neuronal:17 fig:8 referred:3 torres:2 slow:6 axon:1 comput:1 jacobian:3 down:1 removing:1 inset:2 r2:2 decay:15 offset:1 cortes:1 wusi:1 importance:1 province:1 saddle:1 visual:1 vth:1 expressed:1 corresponds:2 dh:1 ma:1 goal:1 presentation:1 absence:1 feasible:1 considerable:3 change:1 justify:1 pfister:2 pas:1 experimental:1 indicating:1 formally:1 latter:2 alexander:1 avoiding:1 phenomenon:2 biol:1 |
4,867 | 5,406 | Sparse PCA via Covariance Thresholding
Andrea Montanari
Electrical Engineering and Statistics
Stanford University
[email protected]
Yash Deshpande
Electrical Engineering
Stanford University
[email protected]
Abstract
In sparse principal component analysis we are given noisy observations of a lowrank matrix of dimension n ? p and seek to reconstruct it under additional sparsity assumptions. In particular, we assume here that the principal components
v1 , . . . , vr have at most k1 , ? ? ? , kq non-zero entries respectively, and study the
high-dimensional regime in which p is of the same order as n.
In an influential paper, Johnstone and Lu [JL04] introduced a simple algorithm
that estimates the support of the principal vectors v1 , . . . , vr by the largest entries
in the diagonal of the empirical covariance.
This method can be shown to succeed
p
with highpprobability if kq ? C1 n/ log p, and to fail with high probability if
kq ? C2 n/ log p for two constants 0 < C1 , C2 < ?. Despite a considerable
amount of work over the last ten years, no practical algorithm exists with provably
better support recovery guarantees.
Here we analyze a covariance thresholding algorithm that was recently proposed
by Krauthgamer, Nadler and Vilenchik [KNV13]. We confirm empirical evidence
presented by these authors and rigorously
prove that the algorithm succeeds with
?
high probability for k of order n. Recent conditional lower bounds [BR13]
suggest that it might be impossible to do significantly better.
The key technical component of our analysis develops new bounds on the norm of
kernel random matrices, in regimes that were not considered before.
1
Introduction
In the spiked covariance model proposed by [JL04], we are given data x1 , x2 , . . . , xn with xi ? Rp
of the form1 :
xi =
r
X
p
?q uq,i vq + zi ,
(1)
q=1
Here v1 , . . . , vr ? Rp is a set of orthonormal vectors, that we want to estimate, while uq,i ?
N(0, 1) and zi ? N(0, Ip ) are independent and identically distributed. The quantity ?q ? R>0
quantifies the signal-to-noise ratio. We are interested in the high-dimensional limit n, p ? ? with
limn?? p/n = ? ? (0, ?). In the rest of this introduction we will refer to the rank one case, in
order to simplify the exposition, and drop the subscript q = {1, 2, . . . , r}. Our results and proofs
hold for general bounded rank.
The standard method
Pn of principal component analysis involves computing the sample covariance
matrix G = n?1 i=1 xi xT
i and estimates v = v1 by its principal eigenvector vPC (G). It is a
well-known fact that, in the high dimensional asymptotic p/n ? ? > 0, this yields an inconsistent
1
Throughout the paper, we follow the convention of denoting scalars by lowercase, vectors by lowercase
boldface, and matrices by uppercase boldface letters.
1
estimate [JL09]. Namely kvPC ? vk2 6? 0 in the high-dimensional asymptotic limit, unless ? ? 0
(i.e. p/n ? 0). Even worse, Baik, Ben-Arous
and P?ech?e [BBAP05] and Paul [Pau07] demonstrate
?
a phase transition phenomenon: if ? < ?
? the estimate is asymptotically orthogonal to the signal
hvPC , vi ? 0. On the other hand, for ? > ?, hvPC , vi remains strictly positive as n, p ? ?. This
phase transition phenomenon has attracted considerable attention recently within random matrix
theory [FP07, CDMF09, BGN11, KY13].
These inconsistency results motivated several efforts to exploit additional structural information on
the signal v. In two influential papers, Johnstone and Lu [JL04, JL09] considered the case of a
signal v that is sparse in a suitable basis, e.g. in the wavelet domain. Without loss of generality, we
will assume here that v is sparse in the canonical basis e1 , . . . ep . In a nutshell, [JL09] proposes the
following:
1. Order the diagonal entries of the Gram matrix Gi(1),i(1) ? Gi(2),i(2) ? ? ? ? ? Gi(p),i(p) ,
and let J ? {i(1), i(2), . . . , i(k)} be the set of indices corresponding to the k largest
entries.
2. Set to zero all the entries Gi,j of G unless i, j ? J, and estimate v with the principal
eigenvector of the resulting matrix.
Johnstone and Lu formalized the sparsity assumption by requiring that v belongs to a weak `q -ball
with q ? (0, 1). Instead, here we consider a strict
? sparsity constraint where v has exactly k non-zero
entries, with magnitudes bounded below by ?/ k for some constant ? > 0. The case of ? = 1 was
studied by Amini and Wainwright in [AW09].
Within this model, it was proved that diagonalp
thresholding successfully recovers the support of
v provided v is sparse enough, namely k ? C n/ log p with C = C(?, ?) a constant [AW09].
(Throughout the paper we denote by C constants that can change from point to point.) This result is
a striking improvement over vanilla PCA. While the latter requires a number of samples scaling as
the number of parameters2 n & p, sparse PCA via diagonal thresholding achieves the same objective
with a number of samples scaling as the number of non-zero parameters, n & k 2 log p.
At the same time, this result is not as optimistic as might have been expected. By searching exhaustively over all possible supports of size k (a method that has complexity of order pk ) the correct
support can be identified with high probability as soon as n & k log p. On the other hand, no method
can succeed for much smaller n, because of information theoretic obstructions [AW09].
Over the last ten years, a significant effort has been devoted to developing practical algorithms
that outperform diagonal thresholding, see e.g. [MWA05, ZHT06, dEGJL07, dBG08, WTH09]. In
particular, d?Aspremont et al. [dEGJL07] developed a promising M-estimator based on a semidefinite programming (SDP) relaxation. Amini and Wainwright [AW09] carried out an analysis of this
method and proved that, if (i) k ? C(?) n/ log p, and (ii) if the SDP solution has rank one, then the
SDP relaxation provides a consistent estimator of the support of v.
At first sight, this appears as a satisfactory solution of the original problem. No procedure can
estimate the support of v from less than k log p samples, and the SDP relaxation succeeds in doing it
from ?at most? a constant factor more samples. This picture was upset by a recent, remarkable result
by Krauthgamer, Nadler and Vilenchik [KNV13]
? who showed that the rank-one condition assumed
by Amini and Wainwright does not hold for n . k . (n/ log p). This result is consistent with
recent work of Berthet and Rigollet [BR13]
demonstrating that sparse PCA cannot be performed in
?
polynomial time in the regime k & n, under a certain computational complexity conjecture for
the so-called planted clique problem.
In summary, the problem of support recovery in sparse PCA demonstrates a fascinating interplay
between computational and statistical barriers.
From a statistical perspective, and disregarding computational considerations, the support of v
can be estimated consistently if and only if k . n/ log p. This can be done, for instance,
by exhaustive search over all the kp possible supports of v. (See [VL12, CMW+ 13] for a
minimax analysis.)
2
Throughout the introduction, we write f (n) & g(n) as a shorthand of ?f (n) ? C g(n) for a some constant
C = C(?, ?)?. Further C denotes a constant that may change from point to point.
2
From a computational perspective, the problem appears to be much more difficult. There is rigorous evidence
? [BR13, MW13] that no polynomial algorithm can reconstruct the support
unless k . n. On the positive p
side, a very simple algorithm (Johnstone and Lu?s diagonal
thresholding) succeeds for k . n/ log p.
Of course, several elements are still missing in this emerging picture. In the present paper we address
one of them, providing an answer to the following question:
Is there a polynomial time algorithm
to solve the sparse PCA
p that is guaranteed
?
problem with high probability for n/ log p . k . n?
We answer this question positively by analyzing a covariance thresholding algorithm that proceeds,
briefly, as follows. (A precise, general definition, with some technical changes is given in the next
section.)
1. Form
? the Gram matrix G and set to zero all its entries that are in modulus smaller than
? / n, for ? a suitably chosen constant.
b1 of this thresholded matrix.
2. Compute the principal eigenvector v
b1 .
3. Denote by B ? {1, . . . , p} be the set of indices corresponding to the k largest entries of v
4. Estimate the support of v by ?cleaning? the set B. (Briefly, v is estimated by thresholding
bB obtained by zeroing the entries outside B.)
Gb
vB with v
Such a covariance thresholding approach was proposed in [KNV13], and is in turn related to earlier
work by Bickel and Levina [BL08]. The formulation discussed in the next section presents some
technical differences that have been introduced to simplify the analysis. Notice that, to simplify
proofs, we assume k to be known: This issue is discussed in the next two sections.
The rest of the paper is organized as follows. In the next section we provide a detailed description of
the algorithm and state our main results. Our theoretical results assume full knowledge of problem
parameters for ease of proof. In light of this, in Section 3 we discuss a practical implementation
of the same idea that does not require prior knowledge of problem parameters, and is entirely datadriven. We also illustrate the method through simulations. The complete proofs are available in the
accompanying supplement, in Sections 1, 2 and 3 respectively.
2
Algorithm and main result
For notational convenience, we shall assume hereafter that 2n sample vectors are given (instead of
n): {xi }1?i?2n . These are distributed according to the model (1). The number of spikes r will be
treated as a known parameter in the problem.
We will make the following assumptions:
A1 The number of spikes r and the signal strengths ?1 , . . . , ?r are fixed as n, p ? ?.
Further ?1 > ?2 > . . . ?r are all distinct.
A2 Let Qq and kq denote the support of vq and its size respectively. We let Q = ?q Qqpand
P
k = q kq throughout. Then the non-zero entries of the spikes satisfy |vq,i | ? ?/ kq
for all i ? Qq for some ? > 0. Further, for any q, q 0 we assume |vq,i /vq0 ,i | ? ? for every
i ? Qq ? Qq0 , for some constant ? > 1.
As before, we are interested in the high-dimensional limit of n, p ? ? with p/n ? ?. A more
detailed description of the covariance thresholding algorithm for the general model (1) is given in
Algorithm 1. We describe the basic intuition for the simpler rank-one case (omitting the subscript
q ? {1, 2, . . . , r}), while stating results in generality.
We start by splitting the data into two halves: (xi )1?i?n and (xi )n<i?2n and compute the respective
sample covariance matrices G and G0 respectively. As we will see, the matrix G is used to obtain
a good estimate for the spike v. This estimate, along with the (independent) second part G0 , is then
used to construct a consistent estimator for the supports of v.
3
Let us focus on the first phase of the algorithm, which aims to obtain a good estimate of v. We
b = G ? I. For ? > ??, the principal eigenvector of G, and hence of ?
b is
first compute ?
?
b
positively correlated with v, i.e. limn?? hb
v1 (?), vi > 0. However, for ? < ?, the noise
b dominates and the two vectors become asymptotically orthogonal, i.e. for instance
component in ?
b vi = 0. In order to reduce the noise level, we exploit the sparsity of the spike v.
limn?? hb
v1 (?),
Denoting by X ? Rn?p the matrix with rows x1 , . . . xn , by Z ? Rn?p the matrix with rows z1 ,
. . . zn , and letting u = (u1 , u2 , . . . , un ), the model (1) can be rewritten as
p
X = ? u vT + Z .
(2)
?
Hence, letting ? 0 ? ?kuk2 /n ? ?, and w ? ?ZT u/n
b = ? 0 vvT + v wT + w vT + 1 ZT Z ? Ip , .
?
n
(3)
For a moment, let us neglect the cross terms (vwT + wvT ). The ?signal? component
? 0 vvT is
?
2
sparse with
? k entries of magnitude ?/k, which (in the regime of interest k = n/C)
? is equivalent
to ?
C?/ n. The ?noise? component ZT Z/n ? Ip is dense with entries of order 1/ n. Assuming
k/ n a small enough constant,?it should be possible to remove most of the noise by thresholding
the entries at level of order 1/ n. For technical reasons, we use the soft thresholding function
? : R ? R?0 ? R, ?(z; ? ) = sgn(z)(|z| ? ? )+ . We will omit the second argument wherever it is
clear from context. Classical denoising theory [DJ94, Joh02] provides upper bounds the estimation
error of such a procedure. Note however that these results fall short of our goal. Classical theory
measures estimation error by (element-wise) `p norm, while here we are interested in the resulting
principal eigenvector. This would require bounding, for instance, the error in operator norm.
?
?
Since the soft thresholding function ?(z; ? / n) is affine when z ? / n, we would expect that
the following decomposition holds approximately (for instance, in operator norm):
1 T
0
T
b
?(?) ? ? ? vv + ?
Z Z ? Ip .
(4)
n
The main technical challenge now is to control the operator norm of the perturbation ?(ZT Z/n?Ip ).
It is easy to see that ?(ZT Z/n ? Ip ) has entries of variance ?(? )/n, for ?(? ) ? 0 as ? ? ?. If
entries were independent with mild decay, this would imply ?with high probability?
? 1 ZT Z
. C?(? ),
(5)
n
2
for some constant C. Further, the first component in the decomposition (4) is still approximately
rank one with norm of the order of ? 0 ? ?. Consequently, with standard linear algebra results on
the perturbation of eigenspaces [DK70], we obtain an error bound kb
v ? vk . ?(? )/C 0 ?
Our first theorem formalizes this intuition and provides a bound on the estimation error in the prinb
cipal components of ?(?).
bq denote
Theorem 1. Under the spiked covariance model Eq. (1) satisfying Assumption A1, let v
b using threshold ? . For every ?, (?q )r ? (0, ?), integer r and every
the q th eigenvector of ?(?)
q=1
? > 0 there exist constants, ? = ? (?, ?, (?q )rq=1 , r, ?) and 0 < c? = c? (?, ?, (?q )rq=1 , r, ?) < ?
P
P
?
such that, if q kq = q |supp(vq )| ? c? n), then
n
o
?
P min(kb
vq ? vq k , kb
vq + vq k) ? ? ?q ? {1, . . . , r} ? 1 ? 4 .
(6)
n
Random matrices of the type ?(ZT Z/n ? Ip ) are called inner-product kernel random matrices and
have attracted recent interest within probability theory [EK10a, EK10b, CS12]. In this literature, the
asymptotic eigenvalue distribution of a matrix f (ZT Z/n) is the object of study. Here f : R ? R
is a kernel function and is applied entry-wise to the matrix ZT Z/n, with Z a matrix as above.
Unfortunately, these results cannot be applied to our problem for the following reasons:
? The results [EK10a, EK10b] are perturbative and provide conditions under which the spectrum of f (ZT Z/n) is close to a rescaling of the spectrum of (ZT Z/n) (with rescaling
4
Algorithm 1 Covariance Thresholding
1: Input: Data (xi )1?i?2n , parameters r, (kq )q?r ? N, ?, ?, ? ? R?0 ;
Pn
P2n
0
T
2: Compute the Gram matrices G ? i=1 xi xT
i /n , G ?
i=n+1 xi xi /n;
b = G ? Ip (resp. ?
b 0 = G0 ? Ip );
3: Compute ?
b by soft-thresholding the entries of ?:
b
4: Compute the matrix ?(?)
?
?
?
b
b
? / n,
?
??ij ? ?n if ?ij ?
?
b ij = 0
b ij < ? /?n,
?(?)
if ?? / n < ?
?
??
b ij ? ?? /?n,
b ij + ??
if ?
n
b
5: Let (b
vq )q?r be the first r eigenvectors
of ?(p
?);
6: Define sq ? Rp by sq,i = v
bq,i I(vbq,i ? ?/2 kq );
b = {i ? [p] : ? q s.t. |(?
b 0 sq )i | ? ?}.
7: Output: Q
factors depending on the Taylor expansion of f close to 0). We are interested instead in a
non-perturbative regime in which the spectrum of f (ZT Z/n) is very different from the one
of (ZT Z/n) (and the Taylor expansion is trivial).
? [CS12] consider n-dependent kernels, but their results are asymptotic and concern the weak
limit of the empirical spectral distribution of f (ZT Z/n). This does not yield an upper
bound on the spectral norm3 of f (ZT Z/n).
Our approach to prove Theorem 1 follows instead the so-called ?-net method: we develop high
probability bounds on the maximum Rayleigh quotient:
X h?
zi , ?
zj i ?
T
?
max hy, ?(Z Z/n)yi = max
;?
yi yj ,
n
n
y?S p?1
y?S p?1
i,j
Here S p?1 = {y ? Rp : kyk = 1} is the unit sphere and ?
zi denote the columns of Z. Since
?(ZT Z/n) is not Lipschitz continuous in the underlying Gaussian variables Z, concentration does
not follow immediately from Gaussian isoperimetry. We have to develop more careful (non-uniform)
bounds on the gradient of ?(ZT Z/n) and show that they imply concentration as required.
b is a reasonable estimate of the spike v in `2 distance (up to
While Theorem 1 guarantees that v
a sign flip), it does not provide a consistent estimator of its support. This brings us to the second
b is not even expected to be sparse, it is easy to see that the largest
phase of our algorithm. Although v
b should have significant overlap with supp(v). Steps 6, 7 and 8 of the algorithm exploit
entries of v
b of the support of the spike v. Our second theorem
this property to construct a consistent estimator Q
guarantees that this estimator is indeed consistent.
Theorem 2. Consider the spiked covariance model of Eq. (1) satisfying Assumptions A1, A2. For
any ?, (?q )q?r ? (0, ?), ?, ? P
> 0 and integer r, there ?
exist constants c? , ?, ? dependent on
?, (?q )q?r , ?, ?, r, such that, if q kq = |supp(vq )| ? c? n, the Covariance Thresholding algorithm of Table 1 recovers the joint supports of vq with high probability.
Explicitly, there exists a constant C > 0 such that
n
o
b = ?q supp(vq ) ? 1 ? C .
P Q
n4
(7)
Given the results above, it is useful to pause for a few remarks.
Remark 2.1. We focus on a consistent estimation of the joint supports ?q supp(vq ) of the spikes.
In the rank-one case, this obviously corresponds to the standard support recovery. Once this is
accomplished, estimating the individual supports poses no additional difficulty: indeed, since | ?q
?
b yields estimates for vq
supp(vq ))| = O( n) an extra step with n fresh samples xi restricted to Q
3
Note that [CS12] also provide a finite n bound for the spectral norm of f (ZT Z/n) via the moment method,
but this bound diverges with n and does not give a result of the type of Eq. (5).
5
p
with `2 error of order k/n. This implies consistent estimation of supp(vq ) when vq have entries
bounded below as in Assumption A2.
Remark 2.2. Recovering the signed supports Qq,+ = {i ? [p] : vq,i > 0} and Qq,? = {i ?
[p] : vq,i < 0} is possible using the same technique as recovering the supports supp(vq ) above, and
poses no additional difficulty.
p
Remark 2.3. Assumption A2 requires |vq,i | ? ?/ kq for all i ? Qq . This is a standard requirement
in the support recovery literature [Wai09, MB06]. The second part of assumption A2 guarantees
that when the supports of two spikes overlap, their entries are roughly of the same order. This is
necessary for our proof technique to go through. Avoiding such an assumption altogether remains
an open question.
Our covariance thresholding algorithm assumes knowledge of the correct support sizes kq . Notice
that the same assumption is made in earlier theoretical, e.g. in the analysis of SDP-based reconstruction by Amini and Wainwright [AW09]. While this assumption is not realistic in applications,
it helps to focus our exposition
on the most challenging aspects of the problem. Further, a ballpark
P
estimate of kq (indeed q kq ) is actually sufficient, with which we use the following steps in lieu of
Steps 7, 8 of Algorithm 1.
7: Define s0q ? Rp by
s0q,i =
?
if |b
vq,i | > ?/(2 k0 )
otherwise.
vbq,i
0
(8)
8: Return
b = ?q {i ? [p] : |(?
b 0 s0 )i | ? ?} .
Q
q
(9)
P
The next theorem shows that this procedure is effective even if k0 overestimates q kq by an order
of magnitude. Its proof is deferred to Section 2.
Theorem 3. Consider the spiked covariance model of Eq.
? (0, ?),Plet constants
P (1). For any ?, ? ?
c? , ?, ? be given as in Theorem 2. Further assume k = q |supp(vq )| ? c? n, and q k ? k0 ?
P
20 q kq . Then, the Covariance Thresholding algorithm of Table 1, with the definitions in Eqs. (8)
and (9), recovers the joint supports of vq successfully, i.e.
b = ?q supp(vq ) ? 1 ? C .
(10)
P Q
n4
3
Practical aspects and empirical results
Specializing to the rank one case, Theorems 1 and 2 show that Covariance Thresholding succeeds
with high probability for a number of samples n & k 2 , while Diagonal Thresholding requires n &
k 2 log p. The reader might wonder whether eliminating the log p factor has any practical relevance
or is a purely conceptual improvement. Figure 1 presents simulations on synthetic data under the
strictly sparse model, and the Covariance Thresholding algorithm of Table 1, used in the proof of
Theorem 2. The objective is to check whether the log p factor has an impact at moderate p. We
compare this with Diagonal Thresholding.
?
We plot the empirical success probability as a function of k/ n for several values of p, with p = n.
The empirical success probability was computed by using 100 independent instances of the problem.
A few observations are of interest: (i) Covariance Thresholding appears to have a significantly
larger success probability in the ?difficult? regime where Diagonal Thresholding starts to fail; (ii)
The curves for Diagonal
Thresholding appear to decrease monotonically with p indicating that k
?
proportional to n is not the right scaling for this algorithm (as is known from theory); (iii) In
contrast, the curves for Covariance Thresholding become
steeper for larger p, and, in particular,
?
the success
? probability increases with p for k ? 1.1 n. This indicates a sharp threshold for k =
const ? n, as suggested by our theory.
In terms of practical applicability, our algorithm in Table 1 has the shortcomings of requiring knowledge of problem parameters ?q , r, kq . Furthermore, the thresholds ?, ? suggested by theory need not
6
0.6
0.4
0.2
p = 625
p = 1250
p = 2500
p = 5000
1
0.8
0.6
0.4
0.2
0
0.5
1
?
k/ n
1.5
2
0.5
1
?
k/ n
1.5
2
Fraction of support recovered
0.8
Fraction of support recovered
Fraction of support recovered
p = 625
p = 1250
p = 2500
p = 5000
1
p = 625
p = 1250
p = 2500
p = 5000
1
0.8
0.6
0.4
0.2
0
0.5
1
?
k/ n
1.5
2
Figure 1: The support recovery phase transitions for Diagonal Thresholding (left) and Covariance
Thresholding (center) and the data-driven version of Section 3 (right). For Covariance Threshold?
ing, the fraction of support recovered correctly increases monotonically with p, as long as k ? c n
with c ? 1.1. Further, it appears to converge to one throughout this region. For Diagonal Thresholding,
the fraction of support recovered correctly decreases monotonically with p for all k of order
?
n. This confirms that Covariance Thresholding
(with or without knowledge of the support size k)
?
succeeds with high probability for k ? c n, while Diagonal Thresholding requires a significantly
sparser principal component.
be optimal. We next describe a principled approach to estimating (where possible) the parameters of
interest and running the algorithm in a purely data-dependent manner. Assume the following model,
for i ? [n]
Xp
?q uq,i vq + ?zi ,
xi = ? +
q
p
where ? ? R is a fixed mean vector, ui have mean 0 and variance 1, and zi have mean 0 and covariance Ip . Note that our focus in this section is not on rigorous analysis, but instead to demonstrate
a principled approach to applying covariance thresholding in practice. We proceed as follows:
P
b = ni=1 xi /n be the empirical mean estimate for ?. Further letting
Estimating ?, ?: We let ?
P
X = X ? 1b
?T we see that pn ? ( q kq )n ? pn entries of X are mean 0 and variance
? 2 . We let ?
b = MAD(X)/? where MAD( ? ) denotes the median absolute deviation of
the entries of the matrix in the argument, and ? is a constant scale factor. Guided by the
Gaussian case, we take ? = ??1 (3/4) ? 0.6745.
Choosing ? : Although in the statement of the theorem, our choice of ? depends on the SNR
?/? 2 , we believe this is an artifact of our analysis. Indeed it is reasonable to threshold
?at the noise level?, as follows. The noise component of entry i, j of the sample covariance (ignoring lower order terms) is given by ? 2 hzi , zj i/n. By the central limit theo? d
4
rem, hzi , zj i/ n ? N(0, 1). Consequently, ? 2 hz
? i , zj i/n ? N(0, ? /n), and we need to
choose the (rescaled) threshold proportional to ? 4 = ? 2 . Using previous estimates, we
let ? = ? 0 ? ?
b2 for a constant ? 0 . In simulations, a choice 3 . ? 0 . 4 appears to work well.
b = XT X/n ? ? 2 Ip and soft threshold it to get ?(?)
b using ? as above.
Estimating r: We define ?
b has r eigenvalues that are separated
Our proof of Theorem 1 relies on the fact that ?(?)
from the bulk of the spectrum4 . Hence, we estimate r using rb: the number of eigenvalues
b
separated from the bulk in ?(?).
b Our theoretical analysis indicates that
bq denote the q th eigenvector of ?(?).
Estimating vq : Let v
bq is expected to be close to vq . In order to denoise v
bq , we assume v
bq ? (1 ? ?)vq + ?q ,
v
where ?q is additive random noise. We then threshold vq ?at the noise level? to recover a better estimate of vq . To do this, we estimate the standard deviation of the
?noise? ? by ?
c? = MAD(vq )/?. Here we set ?again guided by the Gaussian heuristic?
? ? 0.6745. Since vq is sparse, this procedure returns a good estimate for the size of the
bq : set
noise deviation. We let ?H (b
vq ) denote the vector obtained by hard thresholding v
4
The support of the bulk spectrum can be computed numerically from the results of [CS12].
7
bq,i if |b
bq? = ?(b
(?H (b
vq ))i = v
vq,i | ? ? 0 ?c
vq )/ k?(b
vq )k and
?q and 0 otherwise. We then let v
bq? as our estimate for vq .
return v
Note that ?while different in several respects? this empirical approach shares the same philosophy
of the algorithm in Table 1. On the other hand, the data-driven algorithm presented in this section is
less straightforward to analyze, a task that we defer to future work.
Figure 1 also shows results of a support recovery experiment using the ?data-driven? version of
this section.
Covariance thresholding in this form also appears to work for supports of size k ?
?
const n. Figure 2 shows the performance of vanilla PCA, Diagonal Thresholding and Covariance
Thresholding on the ?Three Peak? example of Johnstone and Lu [JL04]. This signal is sparse in
the wavelet domain and the simulations employ the data-driven version of covariance thresholding.
A similar experiment with the ?box? example of Johnstone and Lu is provided in the supplement.
These experiments demonstrate that, while for large values of n both Diagonal Thresholding and
Covariance Thresholding perform well, the latter appears superior for smaller values of n.
PCA
DT
CT
0.3
0.1
0.1
0.2
5 ? 10?2
n = 1024
5 ? 10?2
0.1
0
0
0
?5 ? 10?2
0
1,000
2,000
3,000
4,000
0
0.1
0.1
5 ? 10?2
n = 1625
5 ? 10?2
0
0
1,000
2,000
3,000
4,000
0
1,000
2,000
3,000
4,000
0
1,000
2,000
3,000
4,000
0
1,000
2,000
3,000
4,000
0
1,000
2,000
3,000
4,000
0.1
5 ? 10?2
0
?5 ? 10?2
0
1,000
2,000
3,000
4,000
0
1,000
2,000
3,000
4,000
0.1
0.1
0.1
5 ? 10?2
n = 2580
5 ? 10?2
5 ? 10?2
0
0
?5 ? 10?2
0
1,000
2,000
3,000
0
4,000
0.1
5 ? 10?2
n = 4096
0
1,000
2,000
3,000
4,000
0.1
0.1
5 ? 10?2
5 ? 10?2
0
0
0
0
1,000
2,000
3,000
4,000
0
1,000
2,000
3,000
4,000
Figure 2: The results of Simple PCA, Diagonal Thresholding and Covariance Thresholding (respectively) for the ?Three Peak? example of Johnstone and Lu [JL09] (see Figure 1 of the paper). The
signal is sparse in the ?Symmlet 8? basis. We use ? = 1.4, p = 4096, and the rows correspond to
sample sizes n = 1024, 1625, 2580, 4096 respectively. Parameters for Covariance Thresholding are
chosen as in Section 3, with ? 0 = 4.5. Parameters for Diagonal Thresholding are from [JL09]. On
each curve, we superpose the clean signal (dotted).
References
[AW09]
Arash A Amini and Martin J Wainwright, High-dimensional analysis of semidefinite relaxations
for sparse principal components, The Annals of Statistics 37 (2009), no. 5B, 2877?2921.
[BBAP05]
Jinho Baik, G?erard Ben Arous, and Sandrine P?ech?e, Phase transition of the largest eigenvalue for
nonnull complex sample covariance matrices, Annals of Probability (2005), 1643?1697.
8
[BGN11]
Florent Benaych-Georges and Raj Rao Nadakuditi, The eigenvalues and eigenvectors of finite,
low rank perturbations of large random matrices, Advances in Mathematics 227 (2011), no. 1,
494?521.
[BL08]
Peter J Bickel and Elizaveta Levina, Regularized estimation of large covariance matrices, The
Annals of Statistics (2008), 199?227.
[BR13]
Quentin Berthet and Philippe Rigollet, Computational lower bounds for sparse pca, arXiv
preprint arXiv:1304.0828 (2013).
[CDMF09] Mireille Capitaine, Catherine Donati-Martin, and Delphine F?eral, The largest eigenvalues of finite
rank deformation of large wigner matrices: convergence and nonuniversality of the fluctuations,
The Annals of Probability 37 (2009), no. 1, 1?47.
[CMW+ 13] T Tony Cai, Zongming Ma, Yihong Wu, et al., Sparse pca: Optimal rates and adaptive estimation,
The Annals of Statistics 41 (2013), no. 6, 3074?3110.
[CS12]
Xiuyuan Cheng and Amit Singer, The spectrum of random inner-product kernel matrices, arXiv
preprint arXiv:1202.3155 (2012).
[dBG08]
Alexandre d?Aspremont, Francis Bach, and Laurent El Ghaoui, Optimal solutions for sparse principal component analysis, The Journal of Machine Learning Research 9 (2008), 1269?1294.
[dEGJL07] Alexandre d?Aspremont, Laurent El Ghaoui, Michael I Jordan, and Gert RG Lanckriet, A direct
formulation for sparse pca using semidefinite programming, SIAM review 49 (2007), no. 3, 434?
448.
[DJ94]
D. L. Donoho and I. M. Johnstone, Minimax risk over lp balls, Prob. Th. and Rel. Fields 99
(1994), 277?303.
[DK70]
Chandler Davis and W. M. Kahan, The rotation of eigenvectors by a perturbation. iii, SIAM
Journal on Numerical Analysis 7 (1970), no. 1, pp. 1?46.
[EK10a]
Noureddine El Karoui, On information plus noise kernel random matrices, The Annals of Statistics 38 (2010), no. 5, 3191?3216.
[EK10b]
, The spectrum of kernel random matrices, The Annals of Statistics 38 (2010), no. 1, 1?50.
[FP07]
Delphine F?eral and Sandrine P?ech?e, The largest eigenvalue of rank one deformation of large
wigner matrices, Communications in mathematical physics 272 (2007), no. 1, 185?228.
[JL04]
Iain M Johnstone and Arthur Yu Lu, Sparse principal components analysis, Unpublished
manuscript (2004).
, On consistency and sparsity for principal components analysis in high dimensions, Jour[JL09]
nal of the American Statistical Association 104 (2009), no. 486.
[Joh02]
IM Johnstone, Function estimation and gaussian sequence models, Unpublished manuscript 2
(2002), no. 5.3, 2.
[KNV13]
R. Krauthgamer, B. Nadler, and D. Vilenchik, Do semidefinite relaxations really solve sparse
pca?, CoRR abs/1306:3690 (2013).
[KY13]
Antti Knowles and Jun Yin, The isotropic semicircle law and deformation of wigner matrices,
Communications on Pure and Applied Mathematics (2013).
[MB06]
Nicolai Meinshausen and Peter B?uhlmann, High-dimensional graphs and variable selection with
the lasso, The Annals of Statistics (2006), 1436?1462.
[MW13]
Zongming Ma and Yihong Wu, Computational barriers in minimax submatrix detection, arXiv
preprint arXiv:1309.5914 (2013).
[MWA05] Baback Moghaddam, Yair Weiss, and Shai Avidan, Spectral bounds for sparse pca: Exact and
greedy algorithms, Advances in neural information processing systems, 2005, pp. 915?922.
[Pau07]
Debashis Paul, Asymptotics of sample eigenstructure for a large dimensional spiked covariance
model, Statistica Sinica 17 (2007), no. 4, 1617.
[VL12]
Vincent Q Vu and Jing Lei, Minimax rates of estimation for sparse pca in high dimensions, Proceedings of the 15th International Conference on Artificial Intelligence and Statistics (AISTATS)
2012, 2012.
[Wai09]
Martin J Wainwright, Sharp thresholds for high-dimensional and noisy sparsity recovery usingconstrained quadratic programming (lasso), Information Theory, IEEE Transactions on 55
(2009), no. 5, 2183?2202.
[WTH09] Daniela M Witten, Robert Tibshirani, and Trevor Hastie, A penalized matrix decomposition, with
applications to sparse principal components and canonical correlation analysis, Biostatistics 10
(2009), no. 3, 515?534.
[ZHT06]
Hui Zou, Trevor Hastie, and Robert Tibshirani, Sparse principal component analysis, Journal of
computational and graphical statistics 15 (2006), no. 2, 265?286.
9
| 5406 |@word mild:1 briefly:2 eliminating:1 polynomial:3 norm:7 version:3 suitably:1 open:1 confirms:1 seek:1 simulation:4 covariance:35 decomposition:3 arous:2 moment:2 hereafter:1 denoting:2 recovered:5 nicolai:1 karoui:1 perturbative:2 attracted:2 realistic:1 additive:1 numerical:1 remove:1 drop:1 plot:1 half:1 greedy:1 intelligence:1 kyk:1 isotropic:1 short:1 provides:3 simpler:1 mathematical:1 along:1 c2:2 direct:1 become:2 prove:2 shorthand:1 manner:1 datadriven:1 indeed:4 expected:3 roughly:1 andrea:1 sdp:5 rem:1 provided:2 estimating:5 bounded:3 underlying:1 biostatistics:1 ballpark:1 eigenvector:7 emerging:1 developed:1 guarantee:4 formalizes:1 every:3 nutshell:1 debashis:1 exactly:1 demonstrates:1 control:1 unit:1 omit:1 appear:1 wai09:2 eigenstructure:1 overestimate:1 before:2 positive:2 engineering:2 limit:5 despite:1 analyzing:1 subscript:2 fluctuation:1 laurent:2 approximately:2 might:3 signed:1 plus:1 studied:1 meinshausen:1 challenging:1 ease:1 practical:6 yj:1 vu:1 practice:1 sq:3 procedure:4 asymptotics:1 empirical:8 semicircle:1 significantly:3 suggest:1 get:1 cannot:2 convenience:1 close:3 operator:3 selection:1 context:1 impossible:1 applying:1 risk:1 equivalent:1 missing:1 center:1 go:1 attention:1 straightforward:1 formalized:1 recovery:7 splitting:1 immediately:1 pure:1 estimator:6 iain:1 orthonormal:1 quentin:1 searching:1 gert:1 qq:6 resp:1 annals:8 cleaning:1 programming:3 exact:1 superpose:1 lanckriet:1 element:2 satisfying:2 ep:1 preprint:3 electrical:2 region:1 qq0:1 decrease:2 rescaled:1 rq:2 intuition:2 principled:2 complexity:2 ui:1 rigorously:1 exhaustively:1 algebra:1 purely:2 basis:3 joint:3 k0:3 yash:1 separated:2 distinct:1 ech:3 describe:2 effective:1 shortcoming:1 kp:1 cmw:2 artificial:1 outside:1 choosing:1 exhaustive:1 heuristic:1 stanford:4 solve:2 larger:2 reconstruct:2 otherwise:2 statistic:9 gi:4 kahan:1 noisy:2 ip:11 obviously:1 interplay:1 sequence:1 eigenvalue:7 net:1 cai:1 reconstruction:1 product:2 description:2 convergence:1 requirement:1 diverges:1 jing:1 ben:2 object:1 help:1 illustrate:1 depending:1 stating:1 pose:2 develop:2 ij:6 lowrank:1 eq:5 recovering:2 quotient:1 involves:1 implies:1 convention:1 guided:2 correct:2 chandler:1 kb:3 arash:1 sgn:1 require:2 really:1 parameters2:1 im:1 strictly:2 hold:3 accompanying:1 considered:2 capitaine:1 nadler:3 achieves:1 bickel:2 a2:5 estimation:9 uhlmann:1 largest:7 successfully:2 gaussian:5 sight:1 aim:1 pn:4 focus:4 improvement:2 consistently:1 rank:11 notational:1 vk:1 check:1 indicates:2 contrast:1 rigorous:2 vk2:1 dependent:3 el:3 lowercase:2 wvt:1 interested:4 provably:1 issue:1 proposes:1 field:1 construct:2 once:1 vvt:2 yu:1 future:1 develops:1 simplify:3 employ:1 few:2 individual:1 phase:6 ab:1 detection:1 interest:4 bl08:2 deferred:1 semidefinite:4 light:1 uppercase:1 devoted:1 moghaddam:1 necessary:1 arthur:1 respective:1 eigenspaces:1 orthogonal:2 unless:3 bq:10 nadakuditi:1 taylor:2 deformation:3 theoretical:3 instance:5 column:1 earlier:2 soft:4 rao:1 zn:1 applicability:1 deviation:3 entry:23 snr:1 kq:18 uniform:1 wonder:1 answer:2 synthetic:1 jour:1 peak:2 siam:2 international:1 physic:1 michael:1 again:1 central:1 choose:1 hzi:2 worse:1 american:1 rescaling:2 return:3 supp:10 b2:1 satisfy:1 explicitly:1 vi:4 depends:1 performed:1 optimistic:1 analyze:2 francis:1 doing:1 start:2 steeper:1 recover:1 shai:1 defer:1 ni:1 variance:3 who:1 yield:3 correspond:1 weak:2 norm3:1 vincent:1 lu:8 trevor:2 definition:2 delphine:2 pp:2 deshpande:1 proof:8 sandrine:2 recovers:3 proved:2 knowledge:5 organized:1 actually:1 appears:7 alexandre:2 manuscript:2 vl12:2 dt:1 follow:2 wei:1 formulation:2 done:1 box:1 generality:2 furthermore:1 correlation:1 hand:3 brings:1 artifact:1 lei:1 believe:1 modulus:1 omitting:1 requiring:2 hence:3 satisfactory:1 davis:1 theoretic:1 demonstrate:3 complete:1 wigner:3 wise:2 consideration:1 recently:2 superior:1 rotation:1 witten:1 rigollet:2 discussed:2 association:1 numerically:1 refer:1 significant:2 vanilla:2 consistency:1 mathematics:2 zeroing:1 recent:4 showed:1 perspective:2 raj:1 belongs:1 moderate:1 driven:4 catherine:1 certain:1 success:4 vt:2 inconsistency:1 yi:2 accomplished:1 additional:4 george:1 baback:1 converge:1 monotonically:3 signal:9 ii:2 full:1 ing:1 technical:5 levina:2 cross:1 sphere:1 long:1 bach:1 e1:1 a1:3 specializing:1 impact:1 basic:1 avidan:1 arxiv:6 kernel:7 c1:2 want:1 median:1 mb06:2 limn:3 extra:1 rest:2 benaych:1 strict:1 hz:1 inconsistent:1 jordan:1 integer:2 structural:1 iii:2 identically:1 baik:2 enough:2 hb:2 easy:2 zi:6 hastie:2 identified:1 florent:1 lasso:2 reduce:1 idea:1 inner:2 yihong:2 whether:2 motivated:1 pca:15 gb:1 effort:2 peter:2 proceed:1 remark:4 useful:1 detailed:2 clear:1 eigenvectors:3 amount:1 obstruction:1 ten:2 outperform:1 exist:2 canonical:2 zj:4 notice:2 dotted:1 sign:1 estimated:2 correctly:2 rb:1 bulk:3 tibshirani:2 write:1 shall:1 key:1 upset:1 demonstrating:1 threshold:9 clean:1 thresholded:1 nal:1 v1:6 asymptotically:2 relaxation:5 graph:1 fraction:5 year:2 prob:1 letter:1 striking:1 throughout:5 reasonable:2 reader:1 wu:2 knowles:1 scaling:3 vb:1 eral:2 submatrix:1 entirely:1 bound:12 ct:1 guaranteed:1 cheng:1 quadratic:1 fascinating:1 strength:1 constraint:1 x2:1 dk70:2 hy:1 u1:1 aspect:2 argument:2 min:1 martin:3 conjecture:1 influential:2 developing:1 according:1 ball:2 smaller:3 lp:1 n4:2 wherever:1 spiked:5 restricted:1 ghaoui:2 vq:40 remains:2 turn:1 discus:1 fail:2 daniela:1 singer:1 letting:3 flip:1 lieu:1 available:1 rewritten:1 spectral:4 amini:5 uq:3 yair:1 altogether:1 rp:5 original:1 denotes:2 assumes:1 running:1 tony:1 krauthgamer:3 graphical:1 const:2 neglect:1 exploit:3 k1:1 amit:1 classical:2 objective:2 g0:3 question:3 quantity:1 spike:9 planted:1 concentration:2 diagonal:16 gradient:1 elizaveta:1 distance:1 trivial:1 reason:2 boldface:2 fresh:1 mad:3 assuming:1 index:2 ratio:1 providing:1 difficult:2 unfortunately:1 sinica:1 robert:2 statement:1 implementation:1 zt:18 perform:1 upper:2 observation:2 zongming:2 finite:3 philippe:1 communication:2 precise:1 rn:2 perturbation:4 sharp:2 introduced:2 namely:2 required:1 unpublished:2 vbq:2 z1:1 address:1 suggested:2 proceeds:1 below:2 regime:6 sparsity:6 challenge:1 max:2 wainwright:6 suitable:1 overlap:2 treated:1 difficulty:2 regularized:1 isoperimetry:1 pause:1 minimax:4 imply:2 picture:2 carried:1 aspremont:3 jun:1 form1:1 vpc:1 prior:1 literature:2 review:1 asymptotic:4 law:1 loss:1 expect:1 plet:1 proportional:2 remarkable:1 affine:1 sufficient:1 consistent:8 s0:1 xp:1 thresholding:43 share:1 row:3 course:1 summary:1 penalized:1 last:2 soon:1 nonnull:1 theo:1 antti:1 side:1 vv:1 johnstone:10 fall:1 barrier:2 absolute:1 sparse:26 distributed:2 curve:3 dimension:3 xn:2 transition:4 gram:3 author:1 berthet:2 made:1 adaptive:1 transaction:1 bb:1 confirm:1 clique:1 b1:2 conceptual:1 assumed:1 xi:13 spectrum:6 search:1 un:1 p2n:1 quantifies:1 continuous:1 table:5 promising:1 vilenchik:3 ignoring:1 expansion:2 complex:1 zou:1 domain:2 vwt:1 aistats:1 pk:1 main:3 montanari:2 dense:1 statistica:1 bounding:1 noise:12 paul:2 denoise:1 x1:2 positively:2 vr:3 xiuyuan:1 wavelet:2 theorem:13 kuk2:1 xt:3 disregarding:1 decay:1 evidence:2 dominates:1 exists:2 concern:1 noureddine:1 rel:1 corr:1 hui:1 supplement:2 magnitude:3 sparser:1 rg:1 rayleigh:1 yin:1 erard:1 scalar:1 u2:1 corresponds:1 relies:1 ma:2 succeed:2 conditional:1 goal:1 consequently:2 exposition:2 careful:1 donoho:1 lipschitz:1 considerable:2 change:3 hard:1 wt:1 denoising:1 principal:16 called:3 succeeds:5 indicating:1 support:36 latter:2 relevance:1 phenomenon:2 philosophy:1 avoiding:1 correlated:1 |
4,868 | 5,407 | Low Rank Approximation Lower Bounds in
Row-Update Streams
David P. Woodruff
IBM Research Almaden
[email protected]
Abstract
We study low-rank approximation in the streaming model in which the rows of
an n ? d matrix A are presented one at a time in an arbitrary order. At the end
of the stream, the streaming algorithm should output a k ? d matrix R so that
kA ? AR? Rk2F ? (1 + )kA ? Ak k2F , where Ak is the best rank-k approximation
to A. A deterministic streaming algorithm of Liberty (KDD, 2013), with an improved analysis of Ghashami and Phillips (SODA, 2014), provides such a streaming algorithm using O(dk/) words of space. A natural question is if smaller
space is possible. We give an almost matching lower bound of ?(dk/) bits of
space, even for randomized algorithms which succeed only with constant probability. Our lower bound matches the upper bound of Ghashami and Phillips up to
the word size, improving on a simple ?(dk) space lower bound.
1
Introduction
In the last decade many algorithms for numerical linear algebra problems have been proposed, often
providing substantial gains over more traditional algorithms based on the singular value decomposition (SVD). Much of this work was influenced by the seminal work of Frieze, Kannan, and Vempala
[8]. These include algorithms for matrix product, low rank approximation, regression, and many
other problems. These algorithms are typically approximate and succeed with high probability.
Moreover, they also generally only require one or a small number of passes over the data.
When the algorithm only makes a single pass over the data and uses a small amount of memory,
it is typically referred to as a streaming algorithm. The memory restriction is especially important
for large-scale data sets, e.g., matrices whose elements arrive online and/or are too large to fit in
main memory. These elements may be in the form of an entry or entire row seen at a time; we
refer to the former as the entry-update model and the latter as the row-update model. The rowupdate model often makes sense when the rows correspond to individual entities. Typically one is
interested in designing robust streaming algorithms which do not need to assume a particular order
of the arriving elements for their correctness. Indeed, if data is collected online, such an assumption
may be unrealistic.
Muthukrishnan asked the question of determining the memory required of data stream algorithms
for numerical linear algebra problems, including best rank-k approximation, matrix product, eigenvalues, determinants, and inverses [18]. This question was posed again by Sarl?os [21]. A number
of exciting streaming algorithms now exist for matrix problems. Sarl?os [21] gave 2-pass algorithms
for matrix product, low rank approximation, and regression, which were sharpened by Clarkson and
Woodruff [5], who also proved lower bounds in the entry-update model for a number of these problems. See also work by Andoni and Nguyen for estimating eigenvalues in a stream [2], and work in
[1, 4, 6] which implicitly provides algorithms for approximate matrix product.
In this work we focus on the low rank approximation problem. In this problem we are given an
n ? d matrix A and would like to compute a matrix B of rank at most k for which kA ? BkF ?
1
(1 + )kA ? Ak kF . Here, for a matrix A, kAkF denotes its Frobenius norm
Ak is the best rank-k approximation to A in this norm given by the SVD.
qP
n
i=1
Pd
j=1
A2i,j and
Clarkson and Woodruff [5] show in the entry-update model, one can compute a factorization B =
L ? U ? R with L ? Rn?k , U ? Rk?k , and R ? Rk?d , with a streaming algorithm using O(k?2 (n +
d/2 ) log(nd)) bits of space. They also show a lower bound of ?(k?1 (n + d) log(nd)) bits of
space. One limitation of these bounds is that they hold only when the algorithm is required to output
a factorization L ? U ? R. In many cases n d, and using memory that grows linearly with n (as
the above lower bounds show is unavoidable) is prohibitive. As observed in previous work [9, 16],
in downstream applications we are often only interested in an approximation to the top k principal
components, i.e., the matrix R above, and so the lower bounds of Clarkson and Woodruff can be
too restrictive. For example, in PCA the goal is to compute the most important directions in the row
space of A.
By reanalyzing an algorithm of Liberty [16], Ghashami and Phillips [9] were able to overcome this
restriction in the row-update model, showing that Liberty?s algorithm is a streaming algorithm which
finds a k ? d matrix R for which kA ? AR? RkF ? (1 + )kA ? Ak kF using only O(dk/) words of
space. Here R? is the Moore-Penrose pseudoinverse of R and R? R denotes the projection onto the
row space of R. Importantly, this space bound no longer depends on n. Moreover, their algorithm
is deterministic and achieves relative error. We note that Liberty?s algorithm itself is similar in spirit
to earlier work on incremental PCA [3, 10, 11, 15, 19], but that work missed the idea of using a
Misra-Gries heavy hitters subroutine [17] which is used to bound the additive error (which was then
improved to relative error by Ghashami and Phillips). It also seems possible to obtain a streaming
algorithm using O(dk(log n)/) words of space, using the coreset approach in an earlier paper by
Feldman et al. [7].
This work is motivated by the following questions: Is the O(dk/) space bound tight or can one
achieve an even smaller amount of space? What if one also allows randomization?
In this work we answer the above questions. Our main theorem is the following.
Theorem 1. Any, possibly randomized, streaming algorithm in the row-update model which outputs
a k ? d matrix R and guarantees that kA ? AR? Rk2F ? (1 + )kA ? Ak k2F with probability at least
2/3, must use ?(kd/) bits of space.
Up to a factor of the word size (which is typically O(log(nd)) bits), our main theorem shows that the
algorithm of Liberty is optimal. It also shows that allowing for randomization and a small probability
of error does not significantly help in reducing the memory required. We note that a simple argument
gives an ?(kd) bit lower bound, see Lemma 2 below, which intuitively follows from the fact that
if A itself is rank-k, then R needs to have the same rowspace of A, and specifying a random kdimensional subspace of Rd requires ?(kd) bits. Hence, the main interest here is improving upon
this lower bound to ?(kd/) bits of space. This extra 1/ factor is significant for small values of ,
e.g., if one wants approximations as close to machine precision as possible with a given amount of
memory.
The only other lower bounds for streaming algorithms for low rank approximation that we know of
are due to Clarkson and Woodruff [5]. As in their work, we use the Index problem in communication
complexity to establish our bounds, which is a communication game between two players Alice and
Bob, holding a string x ? {0, 1}r and an index i ? [r] =: {1, 2, . . . , r}, respectively. In this
game Alice sends a single message to Bob who should output xi with constant probability. It is
known (see, e.g., [13]) that this problem requires Alice?s message to be ?(r) bits long. If Alg is a
streaming algorithm for low rank approximation, and Alice can create a matrix Ax while Bob can
create a matrix Bi (depending on their respective inputs x and i), then if from the output of Alg
on the concatenated matrix [Ax ; Bi ] Bob can output xi with constant probability, then the memory
required of Alg is ?(r) bits, since Alice?s message is the state of Alg after running it on Ax .
The main technical challenges are thus in showing how to choose Ax and Bi , as well as showing
how the output of Alg on [Ax ; Bi ] can be used to solve Index. This is where our work departs
significantly from that of Clarkson and Woodruff [5]. Indeed, a major challenge is that in Theorem
1, we only require the output to be the matrix R, whereas in Clarkson and Woodruff?s work from
the output one can reconstruct AR? R. This causes technical complications, since there is much less
information in the output of the algorithm to use to solve the communication game.
2
The intuition behind the proof of Theorem 1 is that given a 2 ? d matrix A = [1, x; 1, 0d ], where
x is a random unit vector, then if P = R? R is a sufficiently good projection matrix for the low
rank approximation problem on A, then the second row of AP actually reveals a lot of information
about x. This may be counterintuitive at first, since one may think that [1, 0d ; 1, 0d ] is a perfectly
good low rank approximation. However, it turns out that [1, x/2; 1, x/2] is a much better low rank
approximation in Frobenius norm, and even this is not optimal. Therefore, Bob, who has [1, 0d ]
together with the output P , can compute the second row of AP , which necessarily reveals a lot of
information about x (e.g., if AP ? [1, x/2; 1, x/2], its second row would reveal a lot of information
about x), and therefore one could hope to embed an instance of the Index problem into x. Most of
the technical work is about reducing the general problem to this 2 ? d primitive problem.
2
Main Theorem
This section is devoted to proving Theorem 1. We start with a simple lemma showing an ?(kd)
lower bound, which we will refer to. The proof of this lemma is in the full version.
Lemma 2. Any streaming algorithm which, for every input A, with constant probability (over its
internal randomness) succeeds in outputting a matrix R for which kA ? AR? RkF ? (1 + )kA ?
Ak kF must use ?(kd) bits of space.
Returning to the proof of Theorem 1, let c > 0 be a small constant to be determined. We consider
the following two player problem between Alice and Bob: Alice has a ck/ ? d matrix A which
can be written as a block matrix [I, R], where I is the ck/ ? ck/ identity matrix, and R is a
ck/ ? (d ? ck/) matrix in which the entries are in {?1/(d ? ck/)1/2 , +1/(d ? ck/)1/2 }. Here
[I, R] means we append the columns of I to the left of the columns of R. Bob is given a set of k
standard unit vectors ei1 , . . . , eik , for distinct i1 , . . . , ik ? [ck/] = {1, 2, . . . , ck/}. Here we need
c/ > 1, but we can assume is less than a sufficiently small constant, as otherwise we would just
need to prove an ?(kd) lower bound, which is established by Lemma 2.
Let B be the matrix [A; ei1 , . . . , eik ] obtained by stacking A on top of the vectors ei1 , . . . , eik .
The goal is for Bob to output a rank-k projection matrix P ? Rd?d for which kB ? BP kF ?
(1 + )kB ? Bk kF .
Denote this problem by f . We will show the randomized 1-way communication complexity of this
1?way
problem R1/4
(f ), in which Alice sends a single message to Bob and Bob fails with probability
at most 1/4, is ?(kd/) bits. More precisely, let ? be the following product distribution on Alice
and Bob?s inputs: the entries of R are chosen independently and uniformly at random in {?1/(d ?
ck/)1/2 , +1/(d ? ck/)1/2 }, while {i1 , . . . , ik } is a uniformly random set among all sets of k
1?way
1?way
distinct indices in [ck/]. We will show that D?,1/4
(f ) = ?(kd/), where D?,1/4
(f ) denotes
the minimum communication cost over all deterministic 1-way (from Alice to Bob) protocols which
fail with probability at most 1/4 when the inputs are distributed according to ?. By Yao?s minimax
1?way
1?way
principle (see, e.g., [14]), R1/4
(f ) ? D?,1/4
(f ).
1?way
We use the following two-player problem Index in order to lower bound D?,1/4
(f ). In this probr
lem Alice is given a string x ? {0, 1} , while Bob is given an index i ? [r]. Alice sends a single
message to Bob, who needs to output xi with probability at least 2/3. Again by Yao?s minimax prin1?way
1?way
ciple, we have that R1/3
(Index) ? D?,1/3
(Index), where ? is the distribution for which x and
i are chosen independently and uniformly at random from their respective domains. The following
is well-known.
1?way
Fact 3. [13] D?,1/3
(Index) = ?(r).
1?way
Theorem 4. For c a small enough positive constant, and d ? k/, we have D?,1/4
(f ) = ?(dk/).
Proof. We will reduce from the Index problem with r = (ck/)(d ? ck/). Alice, given her string x
to Index, creates the ck/ ? d matrix A = [I, R] as follows. The matrix I is the ck/ ? ck/ identity
matrix, while the matrix R is a ck/?(d?ck/) matrix with entries in {?1/(d?ck/)1/2 , +1/(d?
ck/)1/2 }. For an arbitrary bijection between the coordinates of x and the entries of R, Alice sets a
3
ck/"
B2
ck/"
B1
k
1
0
0
0
0
0
0
0
1
0
0
0
0
1
0
0
1
0
0
0
0
0
0
0
1
0
1
0
d
0
0
0
0
1
0
0
ck/"
RS
R
Alice
RT
0
Bob
S
R
T
given entry in R to ?1/(d ? ck/)1/2 if the corresponding coordinate of x is 0, otherwise Alice sets
the given entry in R to +1/(d ? ck/)1/2 . In the Index problem, Bob is given an index, which under
the bijection between coordinates of x and entries of R, corresponds to being given a row index i
and an entry j in the i-th row of R that he needs to recover. He sets i` = i for a random ` ? [k],
and chooses k ? 1 distinct and random indices ij ? [ck/] \ {i` }, for j ? [k] \ {`}. Observe that
if (x, i) ? ?, then (R, i1 , . . . , ik ) ? ?. Suppose there is a protocol in which Alice sends a single
message to Bob who solves f with probability at least 3/4 under ?. We show that this can be used
to solve Index with probability at least 2/3 under ?. The theorem will follow by Fact 3. Consider
the matrix B which is the matrix A stacked on top of the rows ei1 , . . . , eik , in that order, so that B
has ck/ + k rows.
We proceed to lower bound kB ? BP k2F in a certain way, which will allow our reduction to Index
to be carried out. We need the following fact:
?
Fact 5. ((2.4) of [20]) Let
? A be an m ? n matrix with i.i.d. entries which are each +1/ n with
probability 1/2 and ?1/ n with probability 1/2, and suppose m/n < 1. Then for all t > 0,
p
0
3/2
Pr[kAk2 > 1 + t + m/n] ? ?e?? nt .
where ?, ?0 > 0 are absolute constants. Here kAk2 is the operator norm supx kAxk/kxk of A.
We apply Fact 5 to the matrix R, which implies,
p
?
0
3/4
Pr[kRk2 > 1 + c + (ck/)/(d ? (ck/))] ? ?e?? (d?(ck/))c ,
and using that d ? k/ and c > 0 is a sufficiently small constant, this implies
?
Pr[kRk2 > 1 + 3 c] ? e??d ,
(1)
where ?
? > 0 is an ?
absolute constant (depending on c). Note
? that for c > 0 sufficiently small,
(1 + 3 c)2 ? 1 + 7 c. Let E be the event that kRk22 ? 1 + 7 c, which we condition on.
We partition the rows of B into B1 and B2 , where B1 contains those rows whose projection onto
the first ck/ coordinates equals ei for some i ?
/ {i1 , . . . , ik }. Note that B1 is (ck/ ? k) ? d and
B2 is 2k ? d. Here, B2 is 2k ? d since it includes the rows in A indexed by i1 , . . . , ik , together with
the rows ei1 , . . . , eik . Let us also partition the rows of R into RT and RS , so that the union of the
rows in RT and in RS is equal to R, where the rows of RT are the rows of R in B1 , and the rows
of RS are the non-zero rows of R in B2 (note that k of the rows are non-zero and k are zero in B2
restricted to the columns in R).
Lemma 6. For any unit vector u, write u = uR + uS + uT , where S = {i1 , . . . , ik }, T = [ck/] \ S,
and R = [d] \ [ck/], and ?
where uA for a set A is 0 on indices j ?
/ A. Then, conditioned on E
occurring, kBuk2 ? (1 + 7 c)(2 ? kuT k2 ? kuR k2 + 2kuS + uT kkuR k).
4
Proof. Let C be the matrix consisting of the top ck/ rows of B, so that C has the form [I, R],
where I is a ck/ ? ck/ identity matrix. By construction of B, kBuk2 = kuS k2 + kCuk2 . Now,
Cu = uS + uT + RuR , and so
kCuk22
and so
= kuS + uT k2 + kRuR k2 + 2(us + uT )T RuR
?
? kuS + uT k2 + (1 + 7 c)kuR k2 + 2kuS + uT kkRuR k
?
?
? (1 + 7 c)(kuS k2 + kuT k2 + kuR k2 ) + (1 + 3 c)2kuS + uT kkuR k
?
? (1 + 7 c)(1 + 2kuS + uT kkuR k),
kBuk2
?
? (1 + 7 c)(1 + kuS k2 + 2kuS + uT kkuR k)
?
= (1 + 7 c)(2 ? kuR k2 ? kuT k2 + 2kuS + UT kkuR k).
We will also make use of the following simple but tedious fact, shown in the full version.
?
Fact
For x ? [0, 1], the function f (x) = 2x 1 ? x2 ? x2 is maximized when x =
q 7. ?
?
?
1/2 ? 5/10. We define ? to be the value of f (x) at its maximum, where ? = 2/ 5 + 5/10 ?
1/2 ? .618.
?
Corollary 8. Conditioned on E occurring, kBk22 ? (1 + 7 c)(2 + ?).
Proof. By Lemma 6, for any unit vector u,
?
kBuk2 ? (1 + 7 c)(2 ? kuT k2 ? kuR k2 + 2kuS + uT kkuR k).
Suppose we replace the vector uS + uT with an arbitrary vector supported on coordinates in S with
the same norm as uS +uT . Then the right hand side of this expression cannot increase,
which means
p
?
2
2
it is maximized when kuT k = 0, for which it equals (1 + 7 c)(2 ? kuR k + 2 1 ?
?kuR k kuR k),
and setting kuR k to equal the x in Fact 7, we see that this expression is at most (1+7 c)(2+?).
Write the projection matrix P output by the streaming algorithm as U U T , where U is d ? k with
orthonormal columns ui (so R? R = P in the notation of Section 1). Applying Lemma 6 and Fact 7
to each of the columns ui , we show in the full version:
kBP k2F
?
k
X
?
(1 + 7 c)((2 + ?)k ?
kuiT k2 ).
(2)
i=1
Using the matrix Pythagorean theorem, we thus have,
kB ? BP k2F
= kBk2F ? kBP k2F
k
X
?
kuiT k2 ) using kBk2F = 2ck/ + k
? 2ck/ + k ? (1 + 7 c)((2 + ?)k ?
i=1
?
k
?
? X
2ck/ + k ? (1 + 7 c)(2 + ?)k + (1 + 7 c)
kuiT k2 .
(3)
i=1
We now argue that kB ? BP k2F cannot be too large if Alice and Bob succeed in solving f . First, we
?k of rank-k and bound kB ? B
?k k2 .
need to upper bound kB ? Bk k2F . To do so, we create a matrix B
F
?k will be 0 on the rows in B1 . We can group the rows of B2 into k pairs so that each pair
Matrix B
has the form ei + v i , ei , where i ? [ck/] and v i is a unit vector supported on [d] \ [ck/]. We let
Yi be the optimal (in Frobenius norm) rank-1 approximation to the matrix [ei + v i ; ei ]. By direct
?k then
computation 1 the maximum squared singular value of this matrix is 2 + ?. Our matrix B
?
consists of a single Yi for each pair in B2 . Observe that Bk has rank at most k and
?k k2F ? 2ck/ + k ? (2 + ?)k,
kB ? Bk k2F ? kB ? B
1
For an online SVD calculator, see http://www.bluebit.gr/matrix-calculator/
5
Therefore, if Bob succeeds in solving f on input B, then,
kB ? BP k2F
? (1 + )(2ck/ + k ? (2 + ?)k) ? 2ck/ + k ? (2 + ?)k + 2ck.
(4)
Comparing (3) and (4), we arrive at, conditioned on E:
k
X
i=1
kuiT k2 ?
?
1
? ? (7 c(2 + ?)k + 2ck) ? c1 k,
1+7 c
(5)
where c1 > 0 is a constant that can be made arbitrarily small by making c > 0 an arbitrarily small.
? +U
? , where the vectors in U
? are supported
Since P is a projector, kBP kF = kBU kF . Write U = U
?
on T , and the vectors in U are supported on [d] \ T . We have,
?
? k2F ? kBk22 c1 k ? (1 + 7 c)(2 + ?)c1 k ? c2 k,
kB U
? kF ? kBk2 kU
? kF and (5), the second inequality uses that event
where the first inequality uses kB U
E occurs, and the third inequality holds for a constant c2 > 0 that can be made arbitrarily small by
making the constant c > 0 arbitrarily small.
Combining with (4) and using the triangle inequality,
? kF ? kBP kF ? kB U
? kF using the triangle inequality
kB U
p
? k2F
? kBP kF ? c2 k using our bound on kB U
q
p
=
kBk2F ? kB ? BP k2F ? c2 k by the matrix Pythagorean theorem
p
p
?
(2 + ?)k ? 2ck ? c2 k by (4)
p
?
(2 + ?)k ? c3 k,
(6)
where c3 > 0 is a constant that can be made arbitrarily small for c > 0 an arbitrarily small constant
(note that c2 > 0 also becomes arbitrarily small as c > 0 becomes arbitrarily small). Hence,
? k2 ? k ? c4 k for a
? k2 ? (2 + ?)k ? c3 k, and together with Corollary 8, that implies kU
kB U
F
F
constant c4 that can be made arbitrarily small by making c > 0 arbitrarily small.
?,
? k2 . Consider any column u
? k2 is almost as large as kB U
? of U
Our next goal is to show that kB2 U
F
F
and write it as u
?S + u
?R . Hence,
kB u
?k2
= kRT u
?R k2 + kB2 u
?k2 using B1 u
? = RT u
?R
?
kRT u
?R k2 + k?
uS + RS u
?R k2 + k?
uS k2 by definition of the components
= kR?
uR k2 + 2k?
uS k2 + 2?
uTS RS u
?R using the Pythagorean theorem
?
2
? 1 + 7 c + k?
uS k + 2k?
uS kkRS u
?R k
?
using kR?
uR k2 ? (1 + 7 c)k?
uR k2 and k?
uR k2 + k?
uS k2 ? 1
(also using Cauchy-Schwarz to bound the other term).
?
Suppose kRS u
?R k = ? k?
uR k for a value 0 ? ? ? 1 + 7 c. Then
p
?
kB u
?k2 ? 1 + 7 c + k?
uS k2 + 2? k?
uS k 1 ? k?
uS k2 .
We thus have,
kB u
?k2
?
?
?
p
?
1 + 7 c + (1 ? ? )k?
uS k2 + ? (k?
uS k2 + 2k?
uS k 1 ? k?
uS k2 )
?
1 + 7 c + (1 ? ? ) + ? (1 + ?) by Fact 7
?
2 + ? ? + 7 c,
(7)
?
and hence, letting ?1 , . . . , ?k denote the corresponding values of ? for the k columns of U , we have
k
X
?
? k2F ? (2 + 7 c)k + ?
?i .
kB U
(8)
i=1
Comparing the square of (6) with (8), we have
k
X
i=1
?i
? k ? c5 k,
6
(9)
where c5 > 0 is a constant that can be made arbitrarily small by making c > 0 an arbitrarily small
? k2 ? k ? c4 k as shown above, while since kRs u
constant. Now, kU
?R k = ?i k?
uR k if u
?R is the i-th
F
?
column of U , by (9) we have
?R k2F ? (1 ? c6 )k
kRS U
(10)
for a constant c6 that can be made arbitrarily small by making c > 0 an arbitarily small constant.
?R k2 ? (1 + 7?c)k since event E occurs, and kRU
?R k2 = kRT U
?R k2 + kRS U
?R k2 since
Now kRU
F
F
F
F
the rows of R are the concatenation of rows of RS and RT , so combining with (10), we arrive at
?R k2
kRT U
F
? c7 k,
(11)
for a constant c7 > 0 that can be made arbitrarily small by making c > 0 arbitrarily small.
Combining the square of (6) with (11), we thus have
? k2F
kB2 U
? k2F ? kB1 U
? k2F = kB U
? k2F ? kRT U
?R k2F ? (2 + ?)k ? c3 k ? c7 k
= kB U
? (2 + ?)k ? c8 k,
(12)
where the constant c8 > 0 can be made arbitrarily small by making c > 0 arbitrarily small.
By the triangle inequality,
? kF ? kB2 U
? kF ? ((2 + ?)k ? c8 k)1/2 ? (c2 k)1/2 .
kB2 U kF ? kB2 U
(13)
Hence,
kB2 ? B2 P kF
q
kB2 k2F ? kB2 U k2F Matrix Pythagorean, kB2 U kF = kB2 P kF
q
? kF ? kB2 U
? kF )2 Triangle Inequality
?
kB2 k2F ? (kB2 U
q
?
3k ? (((2 + ?)k ? c8 k)1/2 ? (c2 k)1/2 )2 Using (13),kB2 k2F = 3k,(14)
=
(15)
or equivalently, kB2 ? B2 P k2F ? 3k ? ((2 + ?)k ? c8 k) ? (c2 k) + 2k(((2 + ?) ? c8 )c2 )1/2 ?
(1 ? ?)k + c8 k + 2k(((2 + ?) ? c8 )c2 )1/2 ? (1 ? ?)k + c9 k for a constant c9 > 0 that can be made
arbitrarily small by making the constant c > 0 small enough. This intuitively says that P provides a
good low rank approximation for the matrix B2 . Notice that by (14),
kB2 P k2F
= kB2 k2F ? kB2 ? B2 P k2F ? 3k ? (1 ? ?)k ? c9 k ? (2 + ?)k ? c9 k.
(16)
Now B2 is a 2k ? d matrix and we can partition its rows into k pairs of rows of the form Z` =
(ei` +Ri` , ei` ), for ` = 1, . . . , k. Here we abuse notation and think of Ri` as a d-dimensional vector,
its first ck/ coordinates set to 0. Each such pair of rows is a rank-2 matrix, which we abuse notation
and call Z`T . By direct computation2 Z`T has squared maximum singular value 2 + ?. We would
like to argue that the projection of P onto the row span of most Z` has length very close to 1. To
this end, for each Z` consider the orthonormal basis V`T of right singular vectors for its row space
T
T
(which is span(ei` , Ri` )). We let v`,1
, v`,2
be these two right singular vectors with corresponding
singular values ?1 and ?2 (which will be the same for all `, see below). We are interested in the
Pk
quantity ? = `=1 kV`T P k2F which intuitively measures how much of P gets projected onto the
row spaces of the Z`T . The following lemma and corollary are shown in the full version.
Lemma 9. Conditioned on event E, ? ? [k ? c10 k, k + c10 k], where c10 > 0 is a constant that can
be made arbitrarily small by making c > 0 arbitrarily small.
The following corollary is shown in the full version.
?
Corollary 10. Conditioned on event E, for a 1? c9 + 2c10 fraction of ` ? [k], kV`T P k2F ? 1+c11 ,
and for a 99/100 fraction of ` ? [k], we have kV`T P k2F ? 1 ? c11 , where c11 > 0 is a constant that
can be made arbitrarily small by making the constant c > 0 arbitrarily small.
2
We again used the calculator at http://www.bluebit.gr/matrix-calculator/
7
Recall that Bob holds i = i` for a random ` ? [k]. It follows (conditioned on E) by a union bound
that with probability at least 49/50, kV`T P k2F ? [1 ? c11 , 1 + c11 ], which we call the event F and
condition on. We also condition on event G that kZ`T P k2F ? (2+?)?c12 , for a constant c12 > 0 that
can be made arbitrarily small by making c > 0 an arbitrarily small constant. Combining the first part
of Corollary 10 together with (16), event G holds with probability at least 99.5/100, provided c > 0
is a sufficiently small constant. By a union bound it follows that E, F, and G occur simultaneously
with probability at least 49/51.
T
T
As kZ`T P k2F = ?12 kv`,1
P k2 + ?22 kv`,2
P k2 , with ?12 = 2 + ? and ?12 = 1 ? ?, events E, F, and G
T
2
imply that kv`,1 P k ? 1 ? c13 , where c13 > 0 is a constant that can be made arbitrarily small by
T
making the constant c > 0 arbitrarily small. Observe that kv`,1
P k2 = hv`,1 , zi2 , where z is a unit
vector in the direction of the projection of v`,1 onto P .
By the Pythagorean theorem, kv`,1 ? hv`,1 , zizk2 = 1 ? hv`,1 , zi2 , and so
kv`,1 ? hv`,1 , zizk2 ? c14 ,
(17)
for a constant c14 > 0 that can be made arbitrarily small by making c > 0 arbitrarily small.
We thus have Z`T P = ?1 hv`,1 , ziu`,1 z T + ?2 hv`,2 , wiu`,2 wT , where w is a unit vector in the
direction of the projection of of v`,2 onto P , and u`,1 , u`,2 are the left singular vectors of Z`T . Since
F occurs, we have that |hv`,2 , wi| ? c11 , where c11 > 0 is a constant that can be made arbitrarily
small by making the constant c > 0 arbitrarily small. It follows now by (17) that
t
kZ`T P ? ?1 u`,1 v`,1
k2F ? c15 ,
(18)
where c15 > 0 is a constant that can be made arbitrarily small by making the constant c > 0
arbitrarily small.
By direct calculation3 , u`,1 = ?.851ei` ? .526Ri` and v`,1 = ?.851ei` ? .526Ri` . It follows that
kZ`T P ? (2 + ?)[.724ei` + .448Ri` ; .448ei` + .277Ri` ]k2F ? c15 . Since ei` is the second row of
Z`T , it follows that keTi` P ? (2 + ?)(.448ei` + .277Ri` )k2 ? c15 .
Observe that Bob has ei` and P , and can therefore compute eTi` P . Moreover, as c15 > 0 can be made
arbitrarily small by making the constant c > 0 arbitrarily small, it follows that a 1 ? c16 fraction of
the signs of coordinates of eTi` P , restricted to coordinates in [d] \ [ck/], must agree with those of
(2 + ?).277Ri` , which in turn agree with those of Ri` . Here c16 > 0 is a constant that can be made
arbitrarily small by making the constant c > 0 arbitrarily small. Hence, in particular, the sign of the
j-th coordinate of Ri` , which Bob needs to output, agrees with that of the j-th coordinate of eTi` P
with probability at least 1 ? c16 . Call this event H.
By a union bound over the occurrence of events E, F, G, and H, and the streaming algorithm succeeding (which occurs with probability 3/4), it follows that Bob succeeds in solving Index with
probability at least 49/51 ? 1/4 ? c16 > 2/3, as required. This completes the proof.
3
Conclusion
We have shown an ?(dk/) bit lower bound for streaming algorithms in the row-update model for
outputting a k ? d matrix R with kA ? AR? RkF ? (1 + )kA ? Ak kF , thus showing that the
algorithm of [9] is optimal up to the word size. The next natural goal would be to obtain multi-pass
lower bounds, which seem quite challenging. Such lower bound techniques may also be useful for
showing the optimality of a constant-round O(sdk/) + (sk/)O(1) communication protocol in [12]
for low-rank approximation in the distributed communication model.
Acknowledgments. I would like to thank Edo Liberty and Jeff Phillips for many useful discusions
and detailed comments on this work (thanks to Jeff for the figure!). I would also like to thank
the XDATA program of the Defense Advanced Research Projects Agency (DARPA), administered
through Air Force Research Laboratory contract FA8750-12-C0323 for supporting this work.
3
Using the online calculator in earlier footnotes.
8
References
[1] N. Alon, P. B. Gibbons, Y. Matias, and M. Szegedy. Tracking join and self-join sizes in limited
storage. J. Comput. Syst. Sci., 64(3):719?747, 2002.
[2] A. Andoni and H. L. Nguyen. Eigenvalues of a matrix in the streaming model. In SODA, pages
1729?1737, 2013.
[3] M. Brand. Incremental singular value decomposition of uncertain data with missing values. In
ECCV (1), pages 707?720, 2002.
[4] M. Charikar, K. Chen, and M. Farach-Colton. Finding frequent items in data streams. Theor.
Comput. Sci., 312(1):3?15, 2004.
[5] K. L. Clarkson and D. P. Woodruff. Numerical linear algebra in the streaming model. In STOC,
pages 205?214, 2009.
[6] G. Cormode and S. Muthukrishnan. An improved data stream summary: the count-min sketch
and its applications. J. Algorithms, 55(1):58?75, 2005.
[7] D. Feldman, M. Schmidt, and C. Sohler. Turning big data into tiny data: Constant-size coresets
for k-means, pca and projective clustering. In SODA, pages 1434?1453, 2013.
[8] A. M. Frieze, R. Kannan, and S. Vempala. Fast monte-carlo algorithms for finding low-rank
approximations. J. ACM, 51(6):1025?1041, 2004.
[9] M. Ghashami and J. M. Phillips. Relative errors for deterministic low-rank matrix approximations. In SODA, pages 707?717, 2014.
[10] G. H. Golub and C. F. van Loan. Matrix computations (3. ed.). Johns Hopkins University
Press, 1996.
[11] P. M. Hall, A. D. Marshall, and R. R. Martin. Incremental eigenanalysis for classification. In
BMVC, pages 1?10, 1998.
[12] R. Kannan, S. Vempala, and D. P. Woodruff. Nimble algorithms for cloud computing. CoRR,
2013.
[13] I. Kremer, N. Nisan, and D. Ron. On randomized one-round communication complexity.
Computational Complexity, 8(1):21?49, 1999.
[14] E. Kushilevitz and N. Nisan. Communication complexity. Cambridge University Press, 1997.
[15] A. Levy and M. Lindenbaum. Efficient sequential karhunen-loeve basis extraction. In ICCV,
page 739, 2001.
[16] E. Liberty. Simple and deterministic matrix sketching. In KDD, pages 581?588, 2013.
[17] J. Misra and D. Gries. Finding repeated elements. Sci. Comput. Program., 2(2):143?152, 1982.
[18] S. Muthukrishnan. Data streams: Algorithms and applications. Foundations and Trends in
Theoretical Computer Science, 1(2), 2005.
[19] D. A. Ross, J. Lim, R.-S. Lin, and M.-H. Yang. Incremental learning for robust visual tracking.
International Journal of Computer Vision, 77(1-3):125?141, 2008.
[20] M. Rudelson and R. Vershynin. Non-asymptotic theory of random matrices: extreme singular
values. CoRR, 2010.
[21] T. Sarl?os. Improved approximation algorithms for large matrices via random projections. In
FOCS, pages 143?152, 2006.
9
| 5407 |@word determinant:1 version:5 cu:1 seems:1 norm:6 nd:3 tedious:1 r:7 decomposition:2 reduction:1 contains:1 woodruff:9 fa8750:1 ka:12 com:1 nt:1 comparing:2 must:3 written:1 john:1 numerical:3 additive:1 partition:3 kdd:2 succeeding:1 update:8 prohibitive:1 item:1 cormode:1 provides:3 complication:1 bijection:2 kaxk:1 ron:1 c6:2 c2:11 direct:3 ik:6 focs:1 prove:1 consists:1 indeed:2 multi:1 ua:1 becomes:2 provided:1 estimating:1 moreover:3 notation:3 project:1 what:1 string:3 c13:2 finding:3 guarantee:1 every:1 returning:1 k2:53 unit:7 positive:1 ak:8 kuit:4 ap:3 abuse:2 specifying:1 challenging:1 alice:18 factorization:2 limited:1 projective:1 bi:4 acknowledgment:1 union:4 block:1 krt:5 significantly:2 matching:1 projection:9 word:6 get:1 onto:6 lindenbaum:1 close:2 operator:1 cannot:2 storage:1 applying:1 seminal:1 kb1:1 restriction:2 www:2 deterministic:5 projector:1 missing:1 primitive:1 independently:2 coreset:1 kushilevitz:1 importantly:1 counterintuitive:1 orthonormal:2 proving:1 coordinate:10 construction:1 suppose:4 us:3 designing:1 element:4 trend:1 observed:1 cloud:1 hv:7 substantial:1 intuition:1 pd:1 agency:1 complexity:5 ui:2 gibbon:1 asked:1 krk22:1 tight:1 solving:3 algebra:3 upon:1 creates:1 basis:2 triangle:4 darpa:1 muthukrishnan:3 stacked:1 distinct:3 fast:1 monte:1 sarl:3 whose:2 quite:1 posed:1 solve:3 say:1 reconstruct:1 otherwise:2 think:2 itself:2 online:4 eigenvalue:3 outputting:2 product:5 frequent:1 combining:4 achieve:1 frobenius:3 kv:10 r1:3 incremental:4 help:1 depending:2 alon:1 ij:1 c10:4 solves:1 implies:3 liberty:7 direction:3 kb:24 require:2 randomization:2 theor:1 hold:4 sufficiently:5 hall:1 major:1 achieves:1 ross:1 schwarz:1 agrees:1 correctness:1 create:3 hope:1 eti:3 rkf:3 ck:51 corollary:6 ax:5 focus:1 rank:24 sense:1 streaming:19 typically:4 entire:1 her:1 subroutine:1 interested:3 i1:6 among:1 c16:4 classification:1 almaden:1 equal:4 extraction:1 sohler:1 k2f:37 eik:5 frieze:2 simultaneously:1 individual:1 consisting:1 interest:1 message:6 golub:1 extreme:1 behind:1 devoted:1 respective:2 indexed:1 theoretical:1 uncertain:1 instance:1 column:8 earlier:3 marshall:1 ar:6 stacking:1 cost:1 entry:13 wiu:1 gr:2 too:3 c14:2 answer:1 supx:1 chooses:1 vershynin:1 thanks:1 international:1 randomized:4 kut:5 contract:1 kdimensional:1 together:4 hopkins:1 sketching:1 yao:2 again:3 sharpened:1 unavoidable:1 squared:2 choose:1 possibly:1 szegedy:1 syst:1 c12:2 b2:13 includes:1 coresets:1 depends:1 stream:7 nisan:2 lot:3 start:1 recover:1 square:2 air:1 calculator:5 who:5 maximized:2 correspond:1 farach:1 carlo:1 bob:23 randomness:1 footnote:1 influenced:1 edo:1 ed:1 definition:1 c7:3 c15:5 matias:1 proof:7 gain:1 dpwoodru:1 proved:1 recall:1 lim:1 ut:15 actually:1 follow:1 improved:4 bmvc:1 just:1 hand:1 sketch:1 ei:15 o:3 reveal:1 grows:1 former:1 hence:6 moore:1 laboratory:1 round:2 game:3 self:1 probr:1 qp:1 he:2 refer:2 significant:1 cambridge:1 feldman:2 phillips:6 rd:2 xdata:1 longer:1 misra:2 certain:1 inequality:7 arbitrarily:36 yi:2 seen:1 minimum:1 c11:7 ghashami:5 full:5 technical:3 match:1 long:1 lin:1 regression:2 vision:1 c1:4 whereas:1 want:1 completes:1 singular:9 sends:4 extra:1 pass:1 comment:1 spirit:1 seem:1 call:3 yang:1 enough:2 fit:1 gave:1 perfectly:1 reduce:1 idea:1 administered:1 motivated:1 pca:3 expression:2 defense:1 clarkson:7 proceed:1 cause:1 kbk2f:3 generally:1 useful:2 detailed:1 amount:3 http:2 exist:1 notice:1 sign:2 write:4 group:1 downstream:1 fraction:3 inverse:1 soda:4 arrive:3 almost:2 kbp:5 missed:1 bit:13 bound:31 occur:1 precisely:1 bp:6 x2:2 ri:11 loeve:1 argument:1 rur:2 c8:8 span:2 optimality:1 min:1 vempala:3 martin:1 charikar:1 according:1 kd:9 smaller:2 ur:7 wi:1 making:17 lem:1 intuitively:3 restricted:2 pr:3 iccv:1 agree:2 turn:2 count:1 fail:1 know:1 letting:1 end:2 hitter:1 kur:9 apply:1 observe:4 occurrence:1 zi2:2 schmidt:1 a2i:1 denotes:3 top:4 include:1 running:1 clustering:1 rudelson:1 restrictive:1 concatenated:1 especially:1 establish:1 question:5 quantity:1 occurs:4 rt:6 kak2:2 traditional:1 subspace:1 thank:2 sci:3 entity:1 concatenation:1 ei1:5 argue:2 collected:1 cauchy:1 kannan:3 length:1 index:20 providing:1 equivalently:1 stoc:1 holding:1 append:1 allowing:1 upper:2 supporting:1 communication:9 rn:1 arbitrary:3 david:1 bk:4 pair:5 required:5 c3:4 c4:3 established:1 able:1 below:2 challenge:2 program:2 kus:12 including:1 memory:8 unrealistic:1 event:11 natural:2 force:1 turning:1 advanced:1 minimax:2 imply:1 carried:1 kb2:19 kf:22 determining:1 relative:3 asymptotic:1 kakf:1 limitation:1 foundation:1 exciting:1 principle:1 tiny:1 heavy:1 ibm:2 row:40 eccv:1 summary:1 supported:4 last:1 kremer:1 arriving:1 side:1 allow:1 absolute:2 distributed:2 van:1 overcome:1 kz:4 made:18 c5:2 projected:1 nguyen:2 approximate:2 implicitly:1 pseudoinverse:1 colton:1 reveals:2 b1:7 xi:3 decade:1 sk:1 ku:3 robust:2 improving:2 alg:5 necessarily:1 protocol:3 domain:1 pk:1 main:6 linearly:1 big:1 repeated:1 referred:1 join:2 precision:1 fails:1 kbk22:2 comput:3 krk2:2 levy:1 third:1 rk:2 theorem:14 departs:1 embed:1 showing:6 dk:8 eigenanalysis:1 andoni:2 sequential:1 corr:2 kr:6 conditioned:6 occurring:2 karhunen:1 chen:1 arbitarily:1 kbk2:1 penrose:1 visual:1 kxk:1 tracking:2 sdk:1 corresponds:1 acm:1 succeed:3 goal:4 identity:3 jeff:2 replace:1 loan:1 determined:1 reducing:2 uniformly:3 wt:1 principal:1 lemma:10 pas:3 svd:3 player:3 succeeds:3 brand:1 internal:1 latter:1 pythagorean:5 c9:5 bkf:1 |
4,869 | 5,408 | Tight convex relaxations for sparse matrix
factorization
Emile Richard
Electrical Engineering
Stanford University
Guillaume Obozinski
Universit?e Paris-Est
Ecole des Ponts - ParisTech
Jean-Philippe Vert
MINES ParisTech
Institut Curie
Abstract
Based on a new atomic norm, we propose a new convex formulation for sparse
matrix factorization problems in which the number of non-zero elements of the
factors is assumed fixed and known. The formulation counts sparse PCA with
multiple factors, subspace clustering and low-rank sparse bilinear regression as
potential applications. We compute slow rates and an upper bound on the statistical dimension [1] of the suggested norm for rank 1 matrices, showing that
its statistical dimension is an order of magnitude smaller than the usual `1 -norm,
trace norm and their combinations. Even though our convex formulation is in theory hard and does not lead to provably polynomial time algorithmic schemes, we
propose an active set algorithm leveraging the structure of the convex problem to
solve it and show promising numerical results.
1
Introduction
A range of machine learning problems such as link prediction in graphs containing community structure [16], phase retrieval [5], subspace clustering [18] or dictionary learning [12] amount to solve
sparse matrix factorization problems, i.e., to infer a low-rank matrix that can be factorized as the
product of two sparse matrices with few columns (left factor) and few rows (right factor). Such
a factorization allows more efficient storage, faster computation, more interpretable solutions and
especially leads to more accurate estimates in many situations. In the case of interaction networks,
for example, this is related to the assumption that the network is organized as a collection of highly
connected communities which can overlap. More generally, considering sparse low-rank matrices
combines two natural forms of sparsity, in the spectrum and in the support, which can be motivated
by the need to explain systems behaviors by a superposition of latent processes which only involve
a few parameters. Landmark applications of sparse matrix factorization are sparse principal components analysis (SPCA) [8, 21] or sparse canonical correlation analysis (SCCA)[19], which are
widely used to analyze high-dimensional data such as genomic data.
In this paper, we propose new convex formulations for the estimation of sparse low-rank matrices. In
particular, we assume that the matrix of interest should be factorized as the sum of rank one factors
that are the product of column and row vectors with respectively k and q non zero-entries, where k
and q are known. We first introduce below the (k, q)-rank of a matrix as the minimum number of
left and right factors, having respectively k and q non-zeros, required to reconstruct a matrix. This
index is a more involved complexity measure for matrices than the rank in that it conditions on the
number of non-zero elements of the left and right factors of the matrix. Based on this index, we
propose a new atomic norm for matrices [7] by considering its convex hull restricted to the unit ball
of the operator norm, resulting in convex surrogates to low (k, q)-rank matrix estimation problem.
We analyze the statistical dimension of the new norm and compare it to that of linear combinations
of the `1 and trace norms. In the vector case, our atomic norm actually reduces to k-support norm
introduced by [2] and our analysis shows that its statistical power is not better than that of the `1 1
norm. By contrast, in the matrix case, the statistical dimension of our norm is at least one order of
magnitude better than combinations of the `1 -norm and the trace norm.
However, while in the vector case the computation remains feasible in polynomial time, the norm we
introduce for matrices can not be evaluated in polynomial time. We propose algorithmic schemes
to approximately learn with the new norm. The same norm and meta-algorithms can be used as
a regularizer in supervised problems such as multitask learning or quadratic regression and phase
retrieval, highlighting the fact that our algorithmic contribution does not consist in providing more
efficient solutions to the rank-1 SPCA problem, but to combine atoms found by the rank-1 solvers
in a principled way.
2
Tight convex relaxations of sparse factorization constraints
In this section we propose a new matrix norm allowing to formulate various sparse matrix factorization problems as convex optimization problems. We start by defining the (k, q)-rank of a matrix
in section 2.1, a useful generalization of the rank which also quantifies the sparseness of a matrix
factorization. We then introduce in section 2.2 the (k, q)-trace norm, an atomic norm defined as the
convex relaxations of the (k, q)-rank over the operator norm ball. We discuss further properties and
potential applications of this norm used as a regularizer in section 2.3.
2.1
The (k, q)-rank of a matrix
The rank of a matrix Z ? Rm1 ?m2 is theP
minimum number of rank-1 matrices needed to express Z
r
as a linear combination of the form Z = i=1 ai b>
i . The following definition generalizes this rank
to incorporate conditions on the sparseness of the rank-1 elements:
m1 ?m2
Definition 1 ((k, q)-sparse decomposition and (k, q)-rank) For a matrix
, we call
Pr Z ? R
(k, q)-sparse decomposition of Z any decomposition of the form Z = i=1 ci ai b>
i where ai (resp.
bi ) are unit vectors with at most k (resp. q) non-zero elements, and with minimal r, which we call
the (k, q)-rank of Z.
The (k, q)-rank and (k, q)-sparse decomposition of Z can equivalently be defined as the optimal
value and a solution of the optimization problem:
min kck0
s.t.
Z=
?
X
ci ai b>
i ,
m2
1
(ai , bi , ci ) ? Am
k ? A q ? R+ ,
(1)
i=1
where for any 1 ? j ? n, Anj = {a ? Rn | kak0 ? j, kak2 = 1}. Since Ani ? Anj when i ? j, we
have for any k and q rank(Z) ? (k, q)-rank(Z) ? kZk0 . The (k, q)-rank is useful to formalize
problems such as sparse matrix factorization, which can be defined as approximating the solution of
a matrix valued problem by a matrix having low (k, q)-rank. For instance the standard rank-1 SPCA
problem consists in finding the symmetric matrix with (k, k)-rank equal to 1 and providing the best
approximation of the sample covariance matrix [21].
2.2
A convex relaxation for the (k, q)-rank
The (k, q)-rank is a discrete, nonconvex index, like the rank or the cardinality, leading to computational difficulties if one wants to learn matrices with small (k, q)-rank. We propose a convex
relaxation of the (k, q)-rank aimed at mitigating these difficulties. For that purpose, we consider an
atomic norm [7] that provides a convex relaxation of the (k, q)-trace norm, just like the `1 norm and
the trace norm are convex relaxations of the `0 semi-norm and the rank, respectively. An atomic
norm is a convex function defined based on a small set of elements called atoms which constitute
a basis on which an object of interest can be sparsely decomposed. The function (a norm if the set
is centrally symmetric) is defined as the gauge of the convex hull of atoms. In other terms, its unit
ball or level-set of value 1 is formed by the convex envelope of atoms. In case of atoms of interest,
namely rank-1 factors of given sparsities k and q, we define
m2
1
Definition 2 ((k, q)-trace norm) Let Ak,q be a set of atoms Ak,q = ab> : a ? Am
.
k , b ? Aq
For a matrix Z ? Rm1 ?m2 , the (k, q)-trace norm ?k,q (Z) is the atomic norm induced by Ak,q , i.e.,
2
n X
?k,q (Z) = inf
A?Ak,q
o
cA A, cA ? 0, ?A ? Ak,q .
X
cA : Z =
(2)
A?Ak,q
In words, Ak,q is the set of matrices A ? Rm1 ?m2 such that (k, q)-rank(A) = 1 and kAkop = 1.
The next lemma provides an explicit formulation for the (k, q)-trace norm and its dual:
Lemma 1 For any Z, K ? Rm1 ?m2 , and denoting Gkm = {I ? [[1, m]] : |I| = k}, we have
n
o
X
X
?k,q (Z) = inf
kZ (I,J) k? : Z =
Z (I,J) , supp(Z (I,J) ) ? I ? J ,
m1
m
?Gq 2
and
2.3
(3)
(I,J)
(I,J)?Gk
??k,q (K) = max kKI,J kop : I ? Gkm1 , J ? Gqm2 .
Learning matrices with sparse factors
In this section, we briefly discuss how the (k, q)-trace norm norm can be used to formulate various
problems involving the estimation of sparse low-rank matrices. A way to learn a matrix Z with
low empirical risk L(Z) and with low (k, q)-rank is to use ?k,q as a regularizer and minimize an
objective of the form
min
L(Z) + ??k,q (Z).
(4)
m ?m
Z?R
1
2
A number of problems can be formulated as variants of (4).
Bilinear regression. In bilinear regression, given two inputs x ? Rm1 and x0 ? Rm2 one observes as output a noisy version of y = x> Zx0 . Assuming that Z has low (k, q)-rank means that
the noiseless response is a sum of a small number of terms, each involving only a small number
of features from either of the input vectors. To estimate within such a model from observations
(xi , x0i , yi )i=1,...,n one can consider the following formulation, in which ` is a convex loss :
X
0
min
` x>
(5)
i Zxi , yi + ??k,q (Z) .
Z?Rm1 ?m2
i
Subspace clustering. In subspace clustering, one assumes that the data can be clustered in such a
way that the points in each cluster belong to a low dimensional space. If we have a design matrix
X ? Rn?p with each row corresponding to an observation, then the previous assumption means
that if X (j) ? Rnj ?p is a matrix formed by the rows of cluster j, there exist a low rank matrix
Z (j) ? Rnj ?nj such that Z (j) X (j) = X (j) . This means that there exists a block-diagonal matrix Z
such that ZX = X and with low-rank diagonal blocks. This idea, exploited recently by [18] implies
that Z is a sum of low rank sparse matrices; and this property still holds if the clustering is unknown.
We therefore suggest that if all subspaces are of dimension k, Z may be estimated via
min ?k,k (Z) s.t. ZX = X .
Z?Rn?n
Sparse PCA. One possible formulation of sparse PCA with multiple factors is the problem of ap? n by a low-rank matrix with sparse factors. This
proximation of an empirical covariance matrix ?
suggests to formulate sparse PCA as follows:
? n ? ZkF : (k, k)-rank(Z) ? r and Z 0 ,
min k?
(6)
Z
where q is the maximum number of non-zero coefficients allowed in each principal direction. By
contrast to sequential approaches that estimate the principal components one-by-one [11], this formulation requires to find simultaneously a set of complementary factors. If we require the decomposition of Z to be a sum of positive semi-definite (k, k)-sparse rank one factors (which is a stronger
assumption than assuming that Z is p.s.d.), the positivity constraint on Z is no longer necessary and
a natural convex relaxation for (6) using another atomic norm (in fact only a gauge here) is
min
Z?Rm?m
? n ? Zk2F + ??k, (Z) ,
k?
where ?k, is the gauge of the set of atoms Ak, := {aa> , a ? Am
k }.
3
(7)
3
Performance of the (k, q)-trace norm for denoising
In this section, we consider the problem of denoising a low-rank matrix Z ? ? Rm1 ?m2 with sparse
factors corrupted by additive Gaussian noise, that is noisy observations Y ? Rm1 ?m2 of the form
Y = Z ? + ?G , where ? > 0 and G is a random matrix with i.i.d. N (0, 1) entries. For a convex
penalty ? : Rm1 ?m2 ? R, we consider, for any ? > 0, the estimator
1
Z??? = arg min kZ ? Y k2F + ??(Z) .
(8)
Z 2
The following result is a straightforward generalization to any norm ? of the so-called slow rates
that are well know for the `1 norms and other norms such as the trace-norm (see e.g. [10]).
Lemma 2 If ? ? ??? (G)
then
?
Z?? ? Z ?
2 ? 4??(Z ? ) .
F
To derive an upper bound in estimation error from these inequalities, and to keep the argument as
simple as possible we consider the oracle1 estimate Z??Oracle equal to Z??? where ? = ??? (G). From
Lemma 2 we immediately get
(9)
E
Z??Oracle ? Z ? k2F ? 4? ?(Z ? ) E ??(G) .
This upper bound can be computed
for Z ? = ab> ? Ak,q for different norms. In particular, for
?
?
>
?(Z ), we have kab k1 ? kq and ?k,q (ab> ) = kab> k? = 1 which lead to the corollary:
Corollary 1 When Z ? = ab> ? Ak,q is an atom, the expected errors of the oracle estimators
Z??Oracle
, Z?1Oracle and Z??Oracle using respectively the (k, q)-trace norm, the `1 norm and the trace norm
k,q
are upper bounded as follows:
r
r
m1
m2
? 2
E kZ??Oracle
?
Z
k
?
8
?
k
log
q
log
+
2k
+
+
2q
,
F
k,q
k
q
p
p
(10)
E kZ?1Oracle ? Z ? k2F ? 2?kZ ? k1 2 log(m1 m2 ) ? 2? 2kq log(m1 m2 ) ,
?
?
E kZ??Oracle ? Z ? k2F ? 2?( m1 + m2 ) .
When the smallest p
entry in absolute value of a or b is close to 0, then the expected error is smaller for
Oracle
?
, reaching ? 2 log(m1 m2 ) on e1 e>
Z1
1 while not changing for the two other norms. But under
the?assumption that the smallest nonzero entries in absolute value of a and b are lower bounded by
c/ kq with c a constant, the upper bound on the rates obtained for the (k, q)-trace norm is at least
an order of magnitude larger than for the other norms. We report the order of magnitude of these
upper bounds in Table 1 for m1 =?
m2 = m and k = q and assuming that nonzeros coefficients are
lower bounded in magnitude by c/ kq.
Obviously the comparison of upper bounds is not enough to conclude to the superiority of
(k, q)-trace norm and, admittedly, the problem of denoising considered here is a special instance
of linear regression in which the design matrix is the identity, and, since this is a case in which the
design is trivially incoherent, it is possible to obtain fast rates for decomposable norms such as the
`1 or trace norm [13]; however, the slow rates obtained are the same if instead of Y a linear transformation of Z with incoherent design is observed, or when the signal to recover is only weakly
sparse, which is not the case for the fast rates. Moreover, Lemma 2 applies to matrices of any rank
and Corollary 1 generalizes to rank greater than 1. We present in the next section more sophisticated
results, based on bounds on the so-called statistical dimension of different norms [1].
4
A bound on the statistical dimension of the (k, q)-trace norm
The squared Gaussian width [7, and ref. therein] and the statistical dimension introduced recently
by Amelunxen et al. [1], provide quantified estimation guarantees. The two quantities are equal
1
We call it oracle estimate because the choice of ? depends on the unknown noise level. Virtually identical bounds (up to constants) holding with large probability could be derived for the non-oracle estimator by
controlling the deviations of ?? (G) from its expectation.
4
up to an additive term smaller than 1 and we thus present results only in terms of the statistical
dimension. The sample complexity of exact recovery and robust recovery are characterized by this
quantity [7]. It is also equal to the signal to noise ratio necessary for denoising [6] and demixing
[1] (see supplementary section 3). The statistical dimension is defined as follows: if T? (A) is the
tangent cone of a matrix norm ? : Rm1 ?m2 ? R+ at A, then, the statistical dimension of T? (A) is
h
2 i
S(Z, ?) := E
?T? (Z) (G)
F ,
where G ? Rm1 ?m2 is a random matrix with i.i.d. standard normal entries and ?T? (Z) (G) is the
orthogonal projection of G onto the cone T? (Z). In this section, we compute an upper bound on
the statistical dimension of ?k,q at an atoms A of Ak,q , which we will denote by S(A, ?k,q ), and
compare it to results known for linear combinations of the `1 and the trace norm of the form ?? with
?
?? (Z) := ? kZk1 + (1 ? ?)kZk? ,
(11)
kq
which are norms that have been used in the literature to infer sparse low-rank matrices [17]. The
ability to recover the support of a sparse vector typically depends on the size of its smallest non-zero
coefficient. For the recovery of a sparse rank 1 matrix, this motivates the following definition
?? ? [0, 1], ?Z ? Rm1 ?m2 ,
Definition 3 Let A = ab> ? Ak,q with I0 = supp(a) and J0 = supp(b). Denote a2min = min a2i and
i?I0
b2min = min b2j . We define the strength ?(a, b) ? (0, 1] as ?(a, b) := (k a2min ) ? (q b2min ).
j?J0
?
?
The strength of an atom takes the maximal value of 1 when |ai | = 1/ k, i ? I and |bj | = 1/ q, j ?
J where I and J are the supports of a and b. On the contrary, its strength is close to 0 as soon as one
of its nonzero entries is close to zero. We can now present our main result: a bound on the statistical
dimension of ?k,q on Ak,q .
Proposition 1 For A = ab> ? Ak,q with strength ? = ?(a, b), there exist universal constants
c1 , c2 , independent of m1 , m2 , k, q such that
c2
c1
S(A, ?k,q ) ? 2 (k + q) + (k + q) log(m1 ? m2 ) .
?
?
Our proof, presented in the appendix, follows the scheme proposed in [7] and used for the trace
norm and `1 norm. However, ?k,q is not decomposable and requires some work to obtain precise
upper bounds on various quantities.
Note first that S must be larger than the number of degrees of freedoms of elements of Ak,q which
is k + q ? 1. So the bound could not possibly be improved beyond logarithmic factors, besides
the logarithmic dependence on the dimension of the overall space is expected. To appreciate the
result, it should be compared with the statistical dimension for the `1 -norm which scales as the
product of the size of the support with the logarithm of the dimension of the ambient space, that is
as kq log(m1 m2 ). Using Landau notation, we report in Table 1 the upper and lower bounds known
for the statistical dimension of other norms in the case where k = q and m1 = m2 = m. The
rates are known exactly up to constants for the `1 and the trace norm (see e.g. [1]). Of particular
relevance is the comparison with norms of the form ?? which have been introduced with the aim
of improving over both the `1 -norm and the trace norm and have been the object of a significant
literature [17, 15, 9]. Using theorem 3.2 in [15], we prove in appendix 4 a lower bound on the
statistical dimension of ?? of order kq ? (m1 + m2 ) which holds for all values of ?, and which show
that, up to logarithmic factors, ?k,q is an order of magnitude smaller in term of k ? q.
In the right column of Table 1 we also report results in the vector case, that is, when m2 = q = 1. In
fact, in that case, ?k,1 is exactly the k-support norm proposed by [2]. But the statistical dimension
of that norm and the `1 norm is the same as it is known that the rate k log kp cannot be improved
([4]). So, perhaps surprisingly, there improvement in the matrix case but not in the vector case.
5
Algorithm
In this section, we present a working set algorithm that attempts to solve problem (4). Injecting the
variational form (3) of ?k,q in (4) and eliminating the variable Z from the optimization problem
5
Matrix norm
(k, q)-trace
`1
trace-norm
`1 + trace-n.
S
O(k log m)
?(k 2 log km2 )
?(m)
? k2 ? m
?(Z ? )E ?? (G)
1/2
(k log m
k)
2
(k log m)1/2
m1/2
1/2
O m ? (k 2 log m)1/2
Vector norm
k-support
`1
`2
elastic net
S
?(k log kp )
?(k log kp )
p
?(k log kp )
Table 1: Scaling of the statistical dimension S and of the upper bound ?(Z ? ) E?? (G) in estimation
error (slow-rates) of different matrix norms for elements of Ak,q with strength (see Definition
? 3)
lower bounded by a constant (or equivalently with nonzero coefficient lower bounded by c/ kq
for c a constant). Leftmost columns: scalings for matrices with k = q, m = m1 = m2 ; rightmost
columns: scalings for vectors with m1 = p and m2 = q = 1. We use the notations ? and ? with
f = ?(g) meaning g = O(f ) and f = ?(g) to mean that both g = O(f ) and f = O(g).
using Z =
min
Z (IJ) ?Rm1 m2
Z (IJ) , one obtains that, when S = Gkm1 ? Gqm2 , problem (4) is equivalent to
X
X
L
Z (IJ) + ?
kZ (IJ) k? , s.t. Supp(Z (IJ) ) ? I ? J, (I, J) ? S. (PS )
P
(I,J)?S
(I,J)?S
(I,J)?S
At the optimum of (4) however, most of the variables Z (IJ) are equal to zero, and the solution is
the same as the solution obtained from (PS ) in which S is reduced to the set of non-zero matrices
Z (IJ) obtained at optimality, that are often called the active components. We thus propose to solve
problem (4) using a so-called working set algorithm which solves a sequence of problems of the form
(PS ) for a growing sequence of working sets S, so as to keep a small number of non-zero matrices
Z (IJ) throughout. Problem (PS ) is solved easily using approximate block coordinate descent on
the (Z (IJ) )(I,J)?S [3, Chap. 4] , which consists in iterating proximal operators of the trace norm
on blocks I ? J. The principle of the working set algorithm is to solve problem (PS ) for the
current working set S and to check whether a new component should be added. It can be shown
that a component with support I ? J should be added if and only if k[?L(Z)]IJ kop > ? for the
current value of Z. If such a component is found, the corresponding (I, J) pair is added in S and
problem (PS ) is solved again. Given that for any component in S, we have k[?L(Z)]IJ kop ? ? at
the optimum of (PS ), the algorithm terminates if ??k,q (?L(Z)) ? ?.
m2
1
The main difficulty is that ??k,q (K) = max{a> Kb | a ? Am
k , b ? Aq }, which is NP-hard to
compute, since it reduces in particular to rank 1 sparse PCA when k = q and K is p.s.d.. This
implies that determining when the algorithm should stop and which new component to add is hard.
However, a significant amount of research has been carried out on sparse PCA recently, and we
thus propose to leverage some of the recently proposed relaxations and heuristics to solve this rank
1 sparse PCA problem (see [8, 20] and references therein). In particular, the Truncated Power
iteration (TPI) algorithm of [20] can easily be modified to compute ??k,q which corresponds to a
generalization of the rank 1 sparse PCA in which in general a 6= b and k 6= q.
In our numerical experiments, we used a variant of Truncated Power Iteration with multiple restarts,
keeping track of the highest found variance. It should be noted that under RIP conditions on the
matrix, [20] shows that the solution returned by TPI is guaranteed to solve the rank 1 sparse PCA
problem. Also, even if TPI finds a pair (I, J) which is suboptimal, adding it in S does not hurt as
the algorithm might determine subsequently that it is not necessary. However TPI might fail to find
some of the components violating the optimality conditions and terminate the algorithm early.
The proposed algorithm cannot be guaranteed to solve (4) if ??k,q is not computed exactly, but it
exploits as much as possible the structure of the convex optimization problem to find a candidate
solution. A similar active set algorithm can be designed to solve problems regularized by ?k, .
Formulations regularized by the trace norm require to compute its proximal operator, and thus to
compute an SVD. However, even when m1 , m2 are large, solving PS involves the computation of
trace norms of matrices of size only k ? q and so the SVDs that need to be computed are fairly
small. This means that the computational bottleneck of the algorithm is clearly in finding candidate
supports. It has been proved [20] that, under some conditions, the problem can be solved in linear
time. Multiple restarts allow to find good candidate supports in practice.
6
6
5
10
700
600
10
3
10
NMSE
500
Trace
?k,q
3
10
400
90 % overlap
Trace
?k,q
5
10
4
10
10
3
10
300
2
2
10
2
10
1
2
10
3
10
k
k
1
100
1
10
10
(k,q)?rank
200
k
10
0
10
l1
No overlap
4
NMSE
4
10
l1
800
Trace
?k,q
5
10
6
10
(k,k)?rank = 1
NMSE
900
l1
NMSE
6
10
1
1.5
2
2.5
3
3.5
4
4.5
5
5.5
6
10
0
10
1
1
10
2
10
3
10
10
0
10
1
10
2
10
Figure 1: Estimates of the statistical dimensions of the `1 , trace and ?k,q norms at a matrix Z ?
R1000?1000 in different setting. From left to right: (a): Z is an atom in Aek,k for different values of
k. (b) Z is a sum of r atoms in Aek,k with non-overlapping support, with k = 10 and varying r,
(c) Z is a sum of 3 atoms in Aek,k with non-overlapping support, for varying k. (d) Z is a sum of 3
atoms in Aek,k with overlapping support, for varying k.
6
Numerical experiments
In this section we report experimental results to assess the performance of sparse low-rank matrix
estimation using different techniques. We start in Section 6.1 with simulations that confirm and
illustrate the theoretical results on statistical dimension of ?k,q and assess how they generalize to
matrices with (k, q)-rank larger than 1. In Section 6.2 we compare several techniques for sparse
PCA on simulated data.
6.1
Empirical estimates of the statistical dimension.
In order to numerically estimate the statistical dimension S(Z, ?) of a regularizer ? at a matrix
Z, we add to Z a random Gaussian noise matrix and observe Y = Z + ?G where G has normal
i.i.d. entries following N (0, 1). We then denoise Y to form an estimate Z? of Z. For small ?, the
normalized mean-squared error (NMSE) defined as NMSE(?) := EkZ? ?Zk2F /? 2 is a good estimate
of the statistical dimension, since [14] show that S(Z, ?) = lim??0 NMSE(?) . Numerically, we
therefore estimate S(Z, ?) with the empirical NMSE(?) for ? = 10?4 , averaged over 20 replicates.
We consider square matrices with m1 = m2 = 1000, and estimate the statistical dimension of ?k,q ,
the `1 and the trace norms at different matrices Z. The constrained denoiser has a simple closed-form
for the `1 and the trace norm. For ?k,q , it can be obtained by a sequence of proximal projections
? has the correct value ?k,q (Z). Since the noise is small,
with different parameters ? until ?k,q (Z)
we found that it was sufficient and faster to perform a (k, q)-SVD of Y by computing a proximal of
?k,q with a small ?, and then apply the `1 constrained denoiser to the set of (k, q)-sparse singular
values.
We first estimate the statistical dimensions of the three norms at an atom Z in Aek,q for different
?
values of k = q, where Aek,q = {ab> ? Ak,q | kab> k? = 1/ kq} is the set of elements of
Ak,q with nonzero entries of constant magnitude . Figure 1.a shows the results, which confirm
the theoretical bounds summarized in Table 1. The statistical dimension of the trace norm does
not depend on k, while that of the `1 norm increases almost quadratically with k and that of ?k,q
increases linearly with k. The linear versus quadratic dependence of the statistical dimension on
k are reflected by the slopes of the curves in the log-log plot in Figure 1.a. As expected, ?k,q
interpolates between the `1 norm (for k = 1) and the trace norm (for k = m1 ), and outperforms both
norms for intermediate values of k. This experiments therefore confirms that our upper bound (1) on
S(Z, ?k,q ) captures the correct order in k, although the constants can certainly be much improved,
and that our algorithm manages, in this simple setting, to correctly approximate the solution of the
convex minimization problem.
Second, we estimate the statistical dimension of ?k,q on matrices with (k, q)-rank larger than 1,
a setting for which we proved no theoretical result. Figure 1.b shows the numerical estimate of
S(Z, ?k,q ) for matrices Z which are sums of r atoms in Aek,k with non-overlapping support, for
k = 10 and varying r. We observe that the increase in statistical dimension is roughly linear in
the (k, q)-rank. For a fixed (k, q)-rank of 3, Figures 1.c and 1.d compare the estimated statistical
dimensions of the three regularizers on matrices Z which are sums of 3 atoms in Aek,k with re7
Sample covariance
4.20 ? 0.02
Trace
0.98 ? 0.01
`1
2.07 ? 0.01
Trace + `1
0.96 ? 0.01
Sequential
0.93 ? 0.08
?k,
0.59 ? 0.03
Table 2: Relative error of covariance estimation with different methods.
spectively non-overlapping or overlapping supports. The shapes of the different curves are overall
similar to the rank 1 case, although the performance of ?k,q degrades when the supports of atoms
overlap. In both cases, ?k,q consistently outperforms the two other norms. Overall these experiments suggest that the statistical dimension of ?k,q at a linear combination of r atoms increases as
Cr (k log m1 + q log m2 ) where the coefficient C increases with the overlap among the supports of
the atoms.
6.2
Comparison of algorithms for sparse PCA
In this section we compare the performance of different algorithms in estimating a sparsely factored
covariance matrix that we denote ?? . The observed sample consists of n i.i.d. random vectors generated according to N (0, ?? + ? 2 Idp ), where (k, k)-rank(?? ) = 3. The matrix ?? is formed by
>
adding 3 blocks of rank 1, ?? = a1 a>
a3 a>
1 + a2 a2 +?
3 , having all the same sparsity kai k0 = k = 10,
3 ? 3 overlaps and nonzero entries equal to 1/ k. The noise level ? = 0.8 is set in order to make
the signal to noise ratio below the level ? = 1 where a spectral gap appears and makes the spectral
baseline (penalizing the trace of the PSD matrix) work. In our experiments the number of variables
is p = 200 and n = 80 points are observed. To estimate the true covariance matrix from the noisy
? n = 1 Pn xi x> , and given as input
observation, first the sample covariance matrix is formed as ?
i
i=1
n
? The methods we compared are the following:
to various algorithms which provide a new estimate ?.
? n as the estimate of the covariance.
? Sample covariance. Output ?
? n elementwise.
? `1 penalty. Soft-threshold ?
? n k2 + ? Tr Z .
? Trace penalty on the PSD cone. minZ0 12 kZ ? ?
F
? n k2 + ??? (Z).
? Trace + `1 penalty. minZ0 12 kZ ? ?
F
? n k2 + ??k, (Z) , with ?k, the gauge associated with Ak,
? ?k, penalty. minZ?Rp?p 12 kZ ? ?
F
introduced in Section 2.3.
? Sequential sparse PCA. This is the standard way of estimating multiple sparse principal components which consists of solving the problem for a single component at each step t = 1 . . . r, and
deflate to switch to the next (t + 1)st component. The deflation step used in this algorithm is the
>
orthogonal projection Zt+1 = (Idp ? ut u>
t ) Zt (Idp ? ut ut ) . The tuning parameters for this approach are the sparsity level k and the number of principal components r. The hyperparameters
were chosen by leaving one portion of the train data off (validation) and selecting the parameter
which allows to build an estimator approximating the best the validation set?s empirical covariance.
We assumed the true value of k known in advance for all algorithms.
? ? ?? kF /k?? kF over 10 runs of our experiments in Table 2. The
We report the relative errors k?
results indicate that sparse PCA methods, whether based on ?k, or the sequential method with
deflation steps, outperform spectral and `1 baselines, and that penalizing ?k, is superior to the
sequential approach. This was to be expected since our algorithm minimizes a loss function close to
the error measure used, whereas the sequential scheme does not optimize a well-defined objective.
7
Conclusion
We formulated the problem of matrix factorization with sparse factors of known sparsity as the
minimization of an index, the (k, q)-rank which tight convex relaxation is the (k, q)-trace norm
regularizer. This penalty is proved to have near optimal statistical performance. Despite theoretical
computational hardness in the worst-case scenario, exploiting the convex geometry of the problem
allowed us to build an efficient algorithm to minimize it. Future work will consist of relaxing the
constraint on the blocks size, and exploring applications such as finding small comminuties in large
random graph background.
Acknowlegments This project was partially funded by Agence Nationale de la Recherche grant
ANR-13-MONU-005-10 (CHORUS project) and by ERC grant SMAC-ERC-280032.
8
References
[1] D. Amelunxen, M. Lotz, M. McCoy, and J. Tropp. Living on the edge: a geometric theory of
phase transition in convex optimization. Information and Inference, 3(3):224?294, 2014.
[2] A. Argyriou, R. Foygel, and N. Srebro. Sparse prediction with the k-support norm. In Advances
in Neural Information Processing Systems (NIPS), pages 1466?1474, 2012.
R in Machine
[3] F. Bach. Optimization with sparsity-inducing penalties. Foundations and Trends
Learning, 4(1):1?106, 2011.
[4] E. J. Candes and M. A. Davenport. How well can we estimate a sparse vector? Applied and
Computational Harmonic Analysis 34, 34(2):317?323, 2013.
[5] E.J. Cand`es, T Strohmer, and V. Voroninski. Phaselift: Exact and stable signal recovery from
magnitude measurements via convex programming. Comm. Pure Appl. Math., 2011.
[6] V. Chandrasekaran and M. I. Jordan. Computational and statistical tradeoffs via convex relaxation. Proc. Natl. Acad. Sci. USA, 2013.
[7] V. Chandrasekaran, B. Recht, P.A. Parrilo, and A.S. Willsky. The convex geometry of linear
inverse problems. Foundations of Computational Mathematics, 12, 2012.
[8] A. d?Aspremont, L. El Ghaoui, M.I. Jordan, and G.R.G. Lanckriet. A direct formulation for
sparse PCA using semidefinite programming. SIAM review, 2007.
[9] X.V. Doan and S.A. Vavasis. Finding approximately rank-one submatrices with the nuclear
norm and l1 norm. SIAM Journal on Optimization, 23:2502 ? 2540, 2013.
[10] V. Koltchinskii, K. Lounici, and A. Tsybakov. Nuclear-norm penalization and optimal rates
for noisy low-rank matrix completion. Ann. Stat., 39(5):2302?2329, 2011.
[11] L. Mackey. Deflation methods for sparse PCA. Advances in Neural Information Processing
Systems (NIPS), 21:1017?1024, 2008.
[12] J. Mairal, F. Bach, J. Ponce, and G. Sapiro. Online learning for matrix factorization and sparse
coding. J. Mach. Learn. Res., 11:19?60, 2010.
[13] S. N Negahban, P. Ravikumar, M. J Wainwright, and B. Yu. A unified framework for highdimensional analysis of M-estimators. Statistical Science, 27(4):538?557, 2012.
[14] S. Oymak and B. Hassibi. Sharp MSE bounds for proximal denoising. Technical Report
1305.2714, arXiv, 2013.
[15] S. Oymak, A. Jalali, M. Fazel, Y. Eldar, and B. Hassibi. Simultaneously structured models
with application to sparse and low-rank matrices. arXiv preprint 1212.3753, 2012.
[16] E. Richard, S. Gaiffas, and N. Vayatis. Link prediction in graphs with autoregressive features.
In Neural Information Processing Systems, volume 15, pages 565?593. MIT Press, 2012.
[17] E. Richard, P.-A. Savalle, and N. Vayatis. Estimation of simultaneously sparse and low-rank
matrices. In International Conference on Machine Learning (ICML), 2012.
[18] Y.X. Wang, H. Xu, and C. Leng. Provable subspace clustering: When LRR meets SSC. In
Advances in Neural Information Processing Systems (NIPS), pages 64?72. 2013.
[19] D Witten, R Tibshirani, and T Hastie. A penalized matrix decomposition, with applications to
sparse PCA and CCA. Biostatistics, 10(3):515?534, 2009.
[20] X.-T. Yuan and T. Zhang. Truncated power method for sparse eigenvalue problems. J. Mach.
Learn. Res., 14(1):899?925, 2013.
[21] H. Zou, T. Hastie, and R. Tibshirani. Sparse principal component analysis. Journal of computational and graphical statistics, 15(2):265?286, 2006.
9
| 5408 |@word multitask:1 version:1 briefly:1 polynomial:3 norm:89 stronger:1 eliminating:1 zkf:1 confirms:1 simulation:1 decomposition:6 covariance:10 tr:1 selecting:1 ecole:1 denoting:1 rightmost:1 outperforms:2 current:2 must:1 numerical:4 additive:2 shape:1 designed:1 interpretable:1 plot:1 mackey:1 recherche:1 provides:2 math:1 zhang:1 c2:2 direct:1 yuan:1 consists:4 prove:1 combine:2 introduce:3 x0:1 hardness:1 expected:5 roughly:1 cand:1 behavior:1 growing:1 decomposed:1 chap:1 landau:1 considering:2 solver:1 cardinality:1 project:2 estimating:2 bounded:5 moreover:1 notation:2 factorized:2 spectively:1 biostatistics:1 anj:2 minimizes:1 deflate:1 unified:1 finding:4 transformation:1 savalle:1 nj:1 guarantee:1 sapiro:1 exactly:3 universit:1 rm:1 k2:4 unit:3 grant:2 superiority:1 positive:1 engineering:1 acad:1 bilinear:3 despite:1 ak:19 mach:2 ponts:1 r1000:1 meet:1 approximately:2 ap:1 might:2 therein:2 koltchinskii:1 quantified:1 suggests:1 relaxing:1 appl:1 factorization:11 range:1 bi:2 averaged:1 lrr:1 fazel:1 atomic:8 practice:1 block:6 definite:1 j0:2 empirical:5 universal:1 submatrices:1 vert:1 projection:3 word:1 suggest:2 get:1 onto:1 close:4 cannot:2 operator:4 storage:1 risk:1 optimize:1 equivalent:1 straightforward:1 rnj:2 convex:28 formulate:3 decomposable:2 recovery:4 immediately:1 pure:1 kzk0:1 m2:33 estimator:5 factored:1 nuclear:2 coordinate:1 hurt:1 resp:2 controlling:1 rip:1 exact:2 programming:2 lanckriet:1 element:8 trend:1 sparsely:2 observed:3 preprint:1 electrical:1 solved:3 capture:1 svds:1 worst:1 wang:1 connected:1 highest:1 observes:1 principled:1 comm:1 complexity:2 mine:1 weakly:1 tight:3 solving:2 depend:1 basis:1 easily:2 k0:1 various:4 regularizer:5 train:1 fast:2 kp:4 jean:1 heuristic:1 stanford:1 solve:9 widely:1 valued:1 larger:4 reconstruct:1 supplementary:1 kai:1 ability:1 anr:1 statistic:1 noisy:4 online:1 obviously:1 sequence:3 eigenvalue:1 net:1 tpi:4 propose:9 interaction:1 product:3 gq:1 maximal:1 km2:1 re7:1 inducing:1 exploiting:1 cluster:2 p:8 optimum:2 object:2 derive:1 illustrate:1 completion:1 stat:1 ij:11 x0i:1 solves:1 involves:1 implies:2 indicate:1 idp:3 direction:1 correct:2 hull:2 kb:1 subsequently:1 b2j:1 require:2 generalization:3 clustered:1 proposition:1 exploring:1 hold:2 considered:1 normal:2 algorithmic:3 kki:1 bj:1 dictionary:1 early:1 smallest:3 a2:2 purpose:1 estimation:9 proc:1 injecting:1 superposition:1 gauge:4 minimization:2 mit:1 clearly:1 genomic:1 gaussian:3 aim:1 modified:1 reaching:1 pn:1 cr:1 varying:4 mccoy:1 corollary:3 derived:1 ponce:1 improvement:1 consistently:1 rank:69 check:1 contrast:2 baseline:2 am:4 amelunxen:2 inference:1 el:1 i0:2 typically:1 lotz:1 voroninski:1 provably:1 mitigating:1 overall:3 dual:1 among:1 eldar:1 arg:1 constrained:2 special:1 fairly:1 equal:6 having:3 atom:20 identical:1 yu:1 k2f:4 icml:1 future:1 report:6 np:1 richard:3 few:3 simultaneously:3 phase:3 geometry:2 ab:7 freedom:1 attempt:1 psd:2 interest:3 highly:1 certainly:1 replicates:1 semidefinite:1 natl:1 regularizers:1 strohmer:1 accurate:1 ambient:1 edge:1 necessary:3 institut:1 orthogonal:2 phaselift:1 logarithm:1 re:2 theoretical:4 minimal:1 instance:2 column:5 soft:1 deviation:1 entry:9 kq:9 corrupted:1 proximal:5 kck0:1 st:1 recht:1 international:1 siam:2 negahban:1 oymak:2 off:1 squared:2 again:1 containing:1 possibly:1 positivity:1 davenport:1 ssc:1 leading:1 supp:4 potential:2 parrilo:1 de:2 summarized:1 coding:1 coefficient:5 depends:2 closed:1 analyze:2 portion:1 start:2 recover:2 candes:1 slope:1 curie:1 contribution:1 minimize:2 formed:4 ass:2 square:1 variance:1 generalize:1 manages:1 zx:2 explain:1 definition:6 involved:1 proof:1 associated:1 stop:1 proved:3 lim:1 ut:3 organized:1 formalize:1 sophisticated:1 actually:1 appears:1 supervised:1 restarts:2 violating:1 response:1 improved:3 reflected:1 formulation:10 evaluated:1 though:1 lounici:1 just:1 correlation:1 until:1 working:5 tropp:1 overlapping:6 perhaps:1 kab:3 usa:1 normalized:1 true:2 symmetric:2 nonzero:5 width:1 noted:1 leftmost:1 l1:4 meaning:1 variational:1 harmonic:1 recently:4 superior:1 witten:1 volume:1 belong:1 m1:20 elementwise:1 numerically:2 significant:2 measurement:1 ai:6 tuning:1 trivially:1 mathematics:1 erc:2 gaiffas:1 aq:2 funded:1 stable:1 longer:1 add:2 agence:1 inf:2 scenario:1 nonconvex:1 meta:1 inequality:1 yi:2 exploited:1 minimum:2 greater:1 determine:1 signal:4 semi:2 living:1 multiple:5 aek:8 infer:2 reduces:2 nonzeros:1 technical:1 faster:2 characterized:1 bach:2 retrieval:2 e1:1 ravikumar:1 a1:1 prediction:3 involving:2 regression:5 variant:2 noiseless:1 expectation:1 arxiv:2 iteration:2 c1:2 vayatis:2 whereas:1 want:1 background:1 chorus:1 singular:1 leaving:1 envelope:1 induced:1 virtually:1 contrary:1 leveraging:1 gkm:1 jordan:2 call:3 near:1 leverage:1 spca:3 intermediate:1 enough:1 switch:1 hastie:2 suboptimal:1 idea:1 tradeoff:1 bottleneck:1 whether:2 motivated:1 pca:16 penalty:7 returned:1 interpolates:1 constitute:1 generally:1 useful:2 iterating:1 involve:1 aimed:1 amount:2 tsybakov:1 reduced:1 vavasis:1 outperform:1 exist:2 canonical:1 estimated:2 track:1 correctly:1 tibshirani:2 discrete:1 express:1 threshold:1 changing:1 penalizing:2 ani:1 graph:3 relaxation:11 sum:9 cone:3 run:1 inverse:1 throughout:1 almost:1 chandrasekaran:2 appendix:2 scaling:3 cca:1 bound:19 guaranteed:2 centrally:1 quadratic:2 oracle:12 strength:5 constraint:3 argument:1 min:10 optimality:2 structured:1 according:1 combination:6 ball:3 smaller:4 terminates:1 smac:1 restricted:1 pr:1 ghaoui:1 remains:1 foygel:1 discus:2 count:1 fail:1 deflation:3 needed:1 know:1 generalizes:2 apply:1 observe:2 spectral:3 a2i:1 rp:1 assumes:1 clustering:6 graphical:1 exploit:1 k1:2 especially:1 build:2 approximating:2 appreciate:1 objective:2 added:3 quantity:3 degrades:1 dependence:2 usual:1 kak2:1 surrogate:1 diagonal:2 jalali:1 subspace:6 link:2 simulated:1 sci:1 landmark:1 provable:1 willsky:1 assuming:3 denoiser:2 besides:1 index:4 providing:2 ratio:2 equivalently:2 holding:1 kzk1:1 gk:1 trace:42 design:4 motivates:1 zt:2 unknown:2 zk2f:2 allowing:1 upper:12 perform:1 observation:4 zx0:1 descent:1 philippe:1 truncated:3 situation:1 defining:1 precise:1 rn:3 zxi:1 sharp:1 community:2 introduced:4 namely:1 paris:1 required:1 pair:2 z1:1 quadratically:1 nip:3 beyond:1 suggested:1 below:2 sparsity:6 max:2 wainwright:1 power:4 overlap:6 difficulty:3 natural:2 regularized:2 scheme:4 carried:1 incoherent:2 aspremont:1 review:1 literature:2 geometric:1 tangent:1 kf:2 determining:1 relative:2 loss:2 srebro:1 versus:1 emile:1 validation:2 foundation:2 penalization:1 degree:1 sufficient:1 doan:1 principle:1 row:4 penalized:1 surprisingly:1 soon:1 keeping:1 allow:1 rm2:1 absolute:2 sparse:53 kzk:1 dimension:33 curve:2 transition:1 kz:10 acknowlegments:1 autoregressive:1 collection:1 leng:1 approximate:2 obtains:1 keep:2 confirm:2 active:3 proximation:1 mairal:1 assumed:2 conclude:1 xi:2 thep:1 spectrum:1 latent:1 quantifies:1 table:7 promising:1 terminate:1 learn:5 robust:1 ca:3 elastic:1 improving:1 mse:1 zou:1 main:2 linearly:1 noise:7 hyperparameters:1 denoise:1 allowed:2 complementary:1 ref:1 nmse:8 xu:1 slow:4 hassibi:2 explicit:1 candidate:3 minz:3 kop:3 theorem:1 showing:1 a3:1 demixing:1 consist:2 exists:1 sequential:6 adding:2 ci:3 magnitude:8 sparseness:2 nationale:1 gap:1 logarithmic:3 highlighting:1 partially:1 applies:1 aa:1 corresponds:1 obozinski:1 kak0:1 identity:1 formulated:2 ann:1 rm1:13 feasible:1 paristech:2 hard:3 denoising:5 principal:6 lemma:5 called:5 admittedly:1 svd:2 experimental:1 la:1 est:1 e:1 guillaume:1 highdimensional:1 support:18 relevance:1 incorporate:1 argyriou:1 |
4,870 | 5,409 | Robust Tensor Decomposition with Gross Corruption
Huan Gui? Jiawei Han
Department of Computer Science
University of Illinois
at Urbana-Champaign
Urbana, IL 61801
{huangui2,hanj}@illinois.edu
Quanquan Gu?
Department of Operations Research
and Financial Engineering
Princeton University
Princeton, NJ 08544
[email protected]
Abstract
In this paper, we study the statistical performance of robust tensor decomposition
with gross corruption. The observations are noisy realization of the superposition
of a low-rank tensor W ? and an entrywise sparse corruption tensor V ? . Unlike
conventional noise with bounded variance in previous convex tensor decomposition
analysis, the magnitude of the gross corruption can be arbitrary large. We show
that under certain conditions, the true low-rank tensor as well as the sparse corruption tensor can be recovered simultaneously. Our theory yields nonasymptotic
Frobenius-norm estimation error bounds for each tensor separately. We show
through numerical experiments that our theory can precisely predict the scaling
behavior in practice.
1
Introduction
Tensor data analysis have witnessed increasing applications in machine learning, data mining and
computer vision. For example, an ensemble of face images can be modeled as a tensor, whose mode
corresponds to pixels, subjects, illumination and viewpoint [23]. Traditional tensor decomposition
methods such as Tucker decomposition and CANDECOMP/PARAFAC(CP) decomposition [14, 13]
aim to factorize an input tensor into a number of low-rank factors. However, they are prone to local
optima because they are solving essentially non-convex optimization problems. In order to address
this problem, [15] [20] extended the trace norm of matrices [19] to tensors, and generalized convex
matrix completion [8] [7] and matrix decomposition [6] to convex tensor completion/decomposition.
For example, the goal of tensor decomposition aims to accurately estimate a low-rank tensor W ?
Rn1 ?...?nK from the noisy observation tensor Y ? Rn1 ?...?nK that is contaminated by dense
noises, i.e., Y = W ? + E, where W ? ? Rn1 ?...?nK is a low-rank tensor, E ? Rn1 ?...?nK is a
noise tensor whose entries are i.i.d. Gaussian noise with zero mean and bounded variance ? 2 , i.e.,
Ei1 ,...,iK ? N (0, ? 2 ). [22] [21] analyzed the statistical performance of convex tensor decomposition
under different extensions of trace norm. They showed that, under certain conditions, the estimation
error scales with the rank of the true tensor W ? . Furthermore, they demonstrated that given a noisy
tensor, the true low-rank tensor can be recovered under restricted strong convexity assumption [18].
However, all these algorithms [15] [20] and theoretical results [22] [21] reply on the assumption that
the observation noise has a bounded variance ? 2 . Without this assumption, we are not able to identify
c could be very far from the true tensor
the rank of W ? , and therefore the estimated low-rank tensor W
W ?.
On the other hand, in many practical applications such as face recognition and image/video denoising,
a portion of the observation tensor Y might be contaminated by gross error due to illumination,
occlusion or pepper/salt noise. This scenario is not covered by finite variance noise assumption,
therefore new mathematical models are demanded to address this problem. This motivates us to study
?
Equal Contribution
1
convex tensor decomposition with gross corruption. It is clear that if all the entries of a tensor are
corrupted by large error, there is no hope to recover the underlying low-rank tensor. To overcome
this problem, one common assumption is that the gross corruption is sparse. Under this assumption,
together with previous low-rank assumption, we formalize the noisy linear observation model as
follows:
Y = W ? + V ? + E,
(1)
where W ? ? Rn1 ?...?nK is a low-rank tensor, V ? ? Rn1 ?...?nK is a sparse corruption tensor, where
the locations of nonzero entries are unknown and the magnitudes of the nonzero entries can be
arbitrarily large, and E ? Rn1 ?...?nK is a noise tensor whose entries are i.i.d. Gaussian noise with
zero mean and bounded variance ? 2 , and thus dense. Our goal is to recover the low-rank tensor W ? ,
as well as the sparse corruption tensor V ? . Note that in some applications, the corruption tensor is of
independent interest and needs to be recovered.
Given the observation model in (1), and the low-rank as well as sparse assumptions on W ? and E ?
respectively, we propose the following convex minimization to estimate the unknown low-rank tensor
W ? and the sparse corruption tensor E ? simultaneously:
2
arg min |||Y ? W ? V|||F + ?M |||W|||S1 + ?M |||V|||1 ,
W,V
(2)
where |||?|||S1 is tensor Schatten-1 norm [22], |||?|||1 is entry-wise `1 norm of tensors, and ?M and ?M
are positive regularization parameters. We call this optimization Robust Tensor Decomposition, which
can been seen as a generalization of convex tensor decomposition in [15] [20] [22]. The regularization
associated with the E encourages sparsity on the corruption tensor, where parameter ?M controls the
sparsity level. In this paper, we focus on the following questions: under what conditions for the size
of the tensor, the rank of the tensor, and the fraction (sparsity level) of the corruption so that: (i) (2) is
able to recover W ? and V ? with small estimator error? (ii) (2) is able to recover the exact rank of
W ? and the support of V ? ? We will present nonasymptotic error bounds to answer these questions.
Experiments on synthetic datasets validate our theoretical results.
The rest of this paper is arranged as follows. Related work is discussed in Section 2. Section 3
introduces the background and notations. Section 4 presents the main results. Section 5 provides
an ADMM algorithm to solve the problem, followed by the numerical experiments in Section 6.
Section 7 concludes this work with remarks.
2
Related Work
The problem of recovering the data under gross error has gained many attentions recently in matrix
decomposition. A large body of work have been proposed and analyzed statistically. For example,
[9] considered the problem of recovering an unknown low-rank and an unknown sparse matrix, given
the sum of the two matrices. [5] proposed a similar problem, namely robust principal component
analysis (RPCA), which studies the problem of recovering the low-rank and sparse matrices by
solving a convex program. [10] studied multi-task regression which decomposes the coefficient
matrix into two matrices, and imposes different group sparse regularization on two matrices. [25]
considered more general case, where the parameter matrix could be the superposition of more than
two matrices with different structurally constraints. Our paper extends [5] from two perspective: we
extend the problem from matrices to high-order tensors, and consider the additional noise setting.
We notice that [16] extended RPCA to tensors, which aims to recover the low-rank and sparse
tensors by solving a constrained convex program. However, our formulation departs from [16] in
that we consider not only the sparse corruption, but also the dense noise. We also note that low-rank
noisy matrix completion [17] and robust matrix decomposition [1] [12] have been studied in in
the high dimensional setting as well. Our model can be seen as the high-order extension of robust
matrix decomposition. This extension is nontrivial, because the treatment of the tensor trace norm
(Schatten-1 norm) is more complicated. More importantly, for the robust matrix decomposition
problem considered [1], only the sum of error bound of two matrices (low-rank matrix and the sparse
corruption matrix) can be obtained under the assumption of restricted strongly convexity. In contrast,
under a different condition, our analysis provides error bound for each tensor component (low-rank
tensor and the sparse corruption tensor) separately, making our results more appealing in practice
and of independent theoretical interest. Since the problem in [1] is a special case of our problem, our
2
technical tool can be directly applied to their problem and yields new error bounds on the low-rank
matrix as well as the sparse corruption matrix separately.
3
Notation and Background
Before proceeding, we define our notation and state assumptions that will appear in various parts of
the analysis. For more details about tensor algebra, please refer to [14].
Scalars are denoted by lower case letters (a, b, . . .), vectors by bold lower case letters (a, b, . . .),
matrices by bold upper case letters (A, B, . . .), and high-order tensors by calligraphic upper case
letters (A, B, . . .). A tensor is a higher order generalization of a vector (first order tensor) and a matrix
(second order tensor). From a multi-linear algebra view, tensor is a multi-linear mapping over a set of
vector spaces. The order of tensor A ? Rn1 ?...?n2 ?...?nK is K, where nk is the dimensionality of
the k-th order. Elements of A are denoted as Ai1 ...ik ...in , 1 ? ik ? nk . We denote the number of
QK
elements in A by N = k=1 nk .
The mode-k vectors of a K order tensor A are the nk dimensional vectors obtained from A by
varying index ik while keeping the other indices fixed. The mode-k vectors are the column vectors
of mode-k flattening matrix A(k) ? Rnk ?(n1 ...nk?1 nk+1 ...nK ) that results by mode-k flattening the
tensor A. For example, matrix column vectors are referred to as mode-1 vectors and matrix row
vectors are referred to as mode-2 vectors.
n1 ...n2 ...nK
The
, is defined as hA, Bi =
P scalar
P product of two tensors A, B ? R
.
.
.
A
B
=
vec(A)vec(B),
where
vec(?)
is
a
vectorization. The Frobenius
i
...i
i
...i
1
1
K
K
i1
iK
p
norm of a tensor A is |||A|||F = hA, Ai.
There are multiple ways to define tensor rank. In this paper, following [22], we define the rank of
a tensor based on the mode-k rank of a tensor. More specifically, the mode-k rank of a tensor X ,
denoted by rankk (X), is the rank of the mode-k unfolding X(k) (note that X(k) is a matrix, so its
rank is well-defined). Based on mode-k rank, we define the rank of tensor X as r(X ) = (r1 , . . . , rk )
if the mode-k rank is rk for k = 1, . . . , K. Note that the mode-k rank can be computed in polynomial
time, because it boils down to computing a matrix rank, whereas computing tensor rank [14] is NP
complete.
Similarly, we extend the trace norm (a.k.a. nuclear norm) of matrices [19] to tensors. The overlapped
PK
1
Schatten-1 norm is defined as |||X |||S1 = K
X is the mode-k unfolding
k=1 kX(k) kS1 , where
Pr (k)
of X , and k ? kS1 is the Schatten-1 norm for a matrix, kXkS1 = j=1 ?j (X), where ?j (X) is the
j-th largest singular value of X. The dual norm of the Schatten-1 norm is Schatten-? norm (a.k.a.,
spectral norm) as kXkS? = maxj=1,...,r ?j (X).
By H?older?s inequality, we have |hW, Xi| ? kWkS1 kXkS? . It is easy to prove a similar result for
the overlapped Schatten-1 norm and its dual norm. We have the following H?older-like inequality [22]:
|hW, X i| ? |||W|||S1 |||X |||S ? ? |||W|||S1 |||X |||mean ,
1
where |||X |||mean :=
1
K
PK
k=1
(3)
kX(k) kS? .
Pn
Pn
Moreover, we define `1 -norm and `? -norm for tensors that |||X |||1 = i11=1 . . . iKK=1 |Xi1 ,...,iK |,
|||X |||? = max1?i1 ?n1 . . . max1?iK ?nK |Xi1 ,...,iK |. By H?older?s inequality, we have |hW, X i| ?
|||W|||1 |||X |||? , and the following inequality relates the overlapped Schatten-1 norm with the Frobenius norm,
|||X |||S1 ?
K
X
?
rk |||X |||F .
(4)
k=1
Let W ? ? Rn1 ?...?nK be the low-rank tensor that we wish to recover. We assume that W ? is
?
of rank (r1 , . . . , rK ). Thus, for each k, we have W(k)
= Uk Sk Vk> , where Uk ? Rnk ?rk and
?
Vk ? Rrk ?nk are orthogonal matrices, which consist of left and right singular vectors of W(k)
,
rk ?rk
n1 ?...?nK
Sk ? R
is a diagonal matrix whose diagonal elements are singular values. Let ? ? R
3
be an arbitrary tensor, we define the mode-k orthogonal complement ?00k of its mode-k unfolding
?
?(k) ? Rnk ?N\k with respect to the true low-rank tensor W ? as follows
>
?00k = (Ink ? Uk U>
?\k ? Vk Vk ).
k )?(k) (IN
?0k
(5)
?00k
In addition
= ?(k) ?
is the component which has overlapped row/column space with the
?
unfolding of the true tensor W(k)
. Note that the decomposition ?(k) = ?0k + ?00k is defined for
each mode.
In [18], the concept of decomposibility and a large class of decomposable norms are discussed
at length. Of particular relevance to us is the decomposability of the Schatten-1 norm and `1 norm. We have the following equality, i.e., mode-k decomposibility of the Schatten-1 norm that
?
?
+ ?00k kS1 = kW(k)
kS1 + k?00k kS1 , k = 1, . . . , K. To note that the decomposibility is defined
kW(k)
on each mode. It is also easy to check the decomposibility of the `1 -norm.
Let V ? ? Rn1 ?...?nK be the gross corruption tensor that we wish to recover. We assume the the
?
gross
corruption is sparse, in that the cardinality s = |supp(V
)| of its support, S = supp(V ? ) =
?
(i1 , i2 , . . . , iK ) ? [n1 ] ? . . . ? [nK ]|Vi1 ,...,iK 6= 0 . This assumption leads to the inequality
?
between the `1 norm and the Forbenius norm that |||V ? |||1 ? s |||V ? |||F . Moreover, we have
?
?
n1 ?...?nK
|||V |||1 = |||VS |||1 . For any D ? R
, we have |||D|||1 = |||DS |||1 + |||DS c |||1 .
4
Main Results
To get a deep theoretical insight into the recovery property of robust tensor decomposition, we will
now present a set of estimation error bounds. Unlike the analysis in [1], where only the summation
of the estimation errors on the low-rank matrix and gross corruption matrix are analyzed, we aim
at obtaining the estimation error bounds on each tensor (the low-rank tensor and corrupted tensor)
separately. All the proofs can be found in the longer version of this paper.
Instead of considering the observation model in 1, we consider the following more general observation
model
yi = hW ? , Xi i + hV ? , Xi i + i , i = 1, . . . , M,
(6)
where Xi can be seen as an observation operator, and i ?s are i.i.d. zero mean Gaussian noise with
variance ? 2 . Our goal is to estimate an unknown rank (r1 , . . . , rk ) of tensor W ? ? Rn1 ?...?nK ,
as well as the unknown support of tensor V ? , from observations yi , i = 1, . . . , M . We propose
the following convex minimization to estimate the unknown low-rank tensor W ? and the sparse
corruption tensor V ? simultaneously, with composite regularizers on W and V as follows:
c V)
b = arg min 1 ky ? X(W + V)k22 + ?M |||W||| + ?M |||V||| ,
(W,
1
S1
W,V 2M
(7)
where y = (y1 , . . . , yM )> is the collection of observations, X(W) is the linear observation model
that X(W) = [hW, X1 i, . . . , hW, XM i]> . Note that (2) is a special case of (7), where the linear
operator the identity tensor, we have yi as observation of each element in the summation of tensors
W ? + V ?.
? >
We also define y? = (y1? , . . . , yM
) , where yi? = hW ? + V ? , Xi i, is the true evaluation. Due to the
noise of observation model, we have y = y? + . In addition, we define the adjoint operator of X as
PM
X? : RM ? Rn1 ?...?nK that X? () = i=1 i Xi .
4.1
Deterministic Bounds
c?W ?
This section is devoted to obtain the deterministic bound of the residual low-rank tensor ? = W
?
b
and residual corruption tensor D = V ? V separately, which makes our analysis unique.
c ? W ? and D = V
b ? V ? , obtained
We begin with a key technical lemma on residual tensors ? = W
from the convex problem in (7).
c and V
b be the solution of minimization problem (7) with ?M ? 2 |||X? ()|||
Lemma 1. Let W
mean /M ,
?
?M ? 2 |||X ()|||? /M , we have
4
1. rank(?0k ) ? 2rk .
2. There exist ?1 ? 3 and ?2 ? 3, such that
|||DS c |||1 ? ?2 |||DS |||1 .
PK
k=1
k?00k kS1 ? ?1
PK
k=1
k?0k kS1 and
c and V,
b as well as the decomposibility of
The lemma can be obtained by utilizing the optimality of W
Schatten-1 norm and `1 -norm of tensors.
Also, we obtain the key property of the optimal solution of (7), presented in the following theorem.
c and V
b be the solution of minimization problem (7) with ?M ?
Theorem 1. Let W
?
2 |||X ()|||mean /M , ?M ? 2 |||X? ()|||? /M , we have
K
1
3?M X
3?M
kX(? + D)k22 ?
k?0k kS1 +
|||DS |||1 .
2M
2K
2
(8)
k=1
Theorem 1 provides a deterministic prediction error bound for model (7). This is a very general
result, and can be applied to any linear operator X, including the robust tensor decomposition case
that we are particularly interested in this paper. It also covers, for example, tensor regression, tensor
compressive sensing, to mention a few.
Furthermore, we impose an assumption on the linear operator and the residual low-rank tensor and
residue sparse corruption tensor, which generalized the restricted eigenvalue assumption [2] [10].
PK
PK
Assumption 1. Defining ? = {(?, D)| k=1 k?00k kS1 ? ?1 k=1 k?0k kS1 , |||DS c |||1 ?
?2 |||DS |||1 }, we assume there exist positive scalars ?1 , ?2 that
kX(? + D)k2
kX(? + D)k2
> 0, ?2 = min ?
> 0.
?1 = min ?
?,D??
?,D??
M |||?|||F
M |||D|||F
Note that Assumption 1 is also related to restricted strong convexity assumption, which is proposed
in [18] to analyze the statistical properties of general M-estimators in the high dimensional setting.
Combing the results in Theorem 1 and Assumption 1, we have the following theorem, which
summarizes our main result.
cV
b be an optimal solution of (7), and take the regularization parameters ?M ?
Theorem 2. Let W,
2 |||X? ()|||mean /M , ?M ? 2 |||X? ()|||? /M . Then the following results hold:
?
? !
K
3
1 X ?M 2rk
?M s
c
?
+
,
(9)
W ? W ?
?1 K
?1
?2
F
k=1
?
? !
K
1 X ?M 2rk
?M s
3
b
?
.
(10)
+
V ? V ?
?2 K
?1
?2
F
k=1
Theorem 2 provides us with the error bounds of each tensor separately. Specifically, these bounds not
only measure how well our decomposition model can approximate the observation model defined
in (6), but also measure how well the decomposition of the true low-rank tensor and gross corruption
tensor is. When s = 0, our theoretical results reduce to that proposed in [22], which is a special case
of our problem, i.e., noisy low-rank tensor decomposition without corruption.
On the other hand, the results obtained in Theorem 2 are very appealing both practically and
theoretically. From the perspective of applications, this result is quite useful as it helps us to better
understand the behavior of each tensor separately. From the theoretical point of view, this result is
novel, and is incomparable with previous results [1][17] or simple generalization of previous results.
c and V,
b it is unclear whether the rank
Though Theorem 2 has provided estimation error bounds of W
?
?
of W and the support of V can be exactly recovered. We show that under some assumptions about
the true tensors, both of them can be exactly recovered.
Corollary 1. Under the same conditions of Theorem 2, if the following condition holds:
PK ?
?
? !
K
X
6(1
+
?
)
2r
1
?
2r
?
s
1
k
k
M
M
?
k=1
?rk (W(k)
)>
+
,
(11)
?1 M K
K
?1
?2
k=1
5
?
?
where ?rk (W(k)
) is the rk -th largest singular value of W(k)
, then
PK ?
?
? !
K
X
3(1
+
?
)
2r
1
?
2r
?
s
1
k
M
k
M
k=1
c (k) ) >
rbk = arg max ?r (W
+
r
?1 M K
K
?1
?2
k=1
?
recovers the rank of W(k)
for all k.
Furthermore, if the following condition holds:
min
i1 ,...,iK
then
Sb =
|Vi?1 ,...,iK |
?
6(1 + ?2 ) s
>
?2 M
?
? !
K
1 X ?M 2rk
?M s
,
+
K
?1
?2
(12)
k=1
?
3(1 + ?2 ) s
b
(i1 , i2 , . . . , iK ) : Vi1 ,...,iK >
?2 M
?
? !
K
1 X ?M 2rk
?M s
+
K
?1
?2
k=1
recovers the true support of V ? .
Corollary 1, basically states that, under the assumption that the singular values of the low-rank tensor
W ? , and the entry values of corruption tensor V ? are above the noise level (e.g., (11) and (12)), we
can recover the rank and the support successfully.
4.2
Noisy Tensor Decomposition
Now we are going back to study robust tensor decomposition with corruption in (2), which is a special
case of (7), where the linear operator is identity tensor. As the linear operator X is a vectorization such
that M = N , and kX(? + D)k
? 2 = |||? + D|||F . In addition, it is easy to show that Assumption 1
holds with ?1 = ?2 = O(1/ N ). It remains to bound |||X? ()|||mean and |||X? ()|||? , as shown in
the following lemma [1] [24].
Lemma 2. Suppose that X : Rn1 ?????nK ? RN is a vectorization of a tensor. Then we have with
?\k )) ? 1/N that
probability at least 1 ? 2 exp(?C(nk + N
K
q
? X ?
?
?
|||X ()|||mean ?
nk + N\k ,
K
k=1
p
|||X? ()|||? ? 4? log N ,
where C is a universal constant.
With Theorem 2 and Lemma 2, we immediately have the following estimation error bounds for robust
tensor decomposition.
is a vectorization
of a tensor. Then for the
Theorem 3. Suppose that X : Rn1 ?????nK ? RN q
?
PK
?
?
regularization constants ?N ? 2?
nk + N\k /(N K), ?N > 8? log N /N , with
k=1
?\k )) ? 1/N , any solution of (2) have the following error
probability at least 1 ? 2 exp(?C(nk + N
bound:
q
PK ?
!
?
?\k ?2rk
K
?
n
+
N
X
k
k=1
6
1
4? s log N
c
?
+
,
W ? W ?
?1 K
?1 N K
?2 N
F
k=1
q
?
PK ?
!
?
?
K ?
n
+
N
2rk
X
k
\k
k=1
6
1
4? s log N
b
?
+
.
V ? V ?
?2 K
?1 N K
?2 N
F
k=1
c
In the special case that n1 = . . . = nK = n and r1 = . . . = rK = r, we have W
? W ? =
F
?
?
?
?
b
O ? rnK?1 + ? Ks log n and V
? V ? = O ? rnK?1 + ? Ks log n , which matches
F
the error bound of robust matrix decomposition [1] when K = 2.
Note that the high probability support and rank recovery guarantee for the special case of tensor
decomposition follows immediately from Corollary 1. Due to the space limit, we omit the result here.
6
5
Algorithm
In this section, we present an algorithm to solve (2). Since (2) is a special case of (7), we consider
the more general problem (7). It is easy to show that (7) is equivalent to the following problem with
auxiliary variables ?, ?:
min
W,V,Y,Z
K
K
1
?M X
?M X
ky ? x> (w + v)k22 +
|||?k |||S1 +
|||?k |||1 ,
2M
K
K
k=1
k=1
subject to Pk w = ?k , Pk v = ?k ,
PM
where x, w, v, ?k , ?k are the vectorizations of i=1 Xi , W, V, ?k , ?k respectively, and Pk is the
transformation matrix that change the order of rows and columns so that Pk w = ?k .
The augmented Lagrangian (AL) function of the above minimization problem with respect to the
primal variables (W t , V t ) is given as follows:
K
K
K
L? (W, V, {?k }K
k=1 , {?k }k=1 , {?k }k=1 , {?k }k=1 )
K
K
?M M X
1
?M M X
= ky ? x> (w + v)k22 +
|||?k |||S1 +
|||?k |||1
2
K
K
k=1
k=1
!
X
X
1
1
2
>
2
>
(?k (Pk v ? ?k ) + kPk v ? ?k k2 ) ,
+?
(?k (Pk w ? ?k ) + kPk w ? ?k k2 ) +
2
2
k
k
t
t
where ? , ? are Lagrangian multiplier vectors, and ? > 0 is a penalty parameter.
We then apply the algorithm of Alternating Direction Method of Multipliers (ADMM)
[3, 20] to solve the above optimization problem.
Starting from initial points
0 K
0 K
0 K
(w0 , v0 , {?0k }K
,
{?
}
,
{?
}
,
{?
}
),
ADMM
performs
the following updates
k=1
k k=1
k k=1
k k=1
iteratively:
!
K
X
t
t
wt+1 = (x> y ? x> xvt ) + ?
P>
k (?k ? ?k ) / (1 + ?K) ,
k=1
v
t+1
=
>
>
(x y ? x xw
t+1
)+?
K
X
!
t
P>
k (?k
?
?kt )
/ (1 + ?K) ,
k=1
t+1
?t+1
= proxtr
+ ?tk ),
?M (Pk w
k
?K
?t+1
= ?t+1
+ (Pk wt+1 ? ?kt+1 )
k
k
1
?t+1
= prox`?M
(Pk vt+1 + ?kt )
k
k = 1, . . . , K,
?K
?kt+1 = ?kt+1 + (Pk vt+1 ? ?t+1
k )
k = 1, . . . , K,
`1
where proxtr
? (?) is the soft-thresholding operator for trace norm, and prox? (?) is the soft-thresholding
operator for `1 norm [4, 11]. The stopping criterion is that all the partial (sub)gradients are (near)
zero, under which condition we obtain the saddle point of the augmented Lagrangian function. Since
(7) is strictly convex, the saddle point is the global optima for the primal problem.
6
Experiments
In this section, we conduct numerical experiments to confirm our analysis in previous sections. The
experiments are conducted under the setting of robust noisy tensor decomposition.
We follow the procedure described in [22] for the experimental part. We randomly generate low-rank
tensors of dimensions n(1) = (50, 50, 20) ( results are shown in Figure 1(a, b, c)) and n(2) =
(100, 100, 50)( results are shown in Figure 1(d, e, f)) for various rank (r1 , r2 , ..., rk ). Given a specific
rank, we first generated the ?core tensor? with elements r1 ? . . . ? rK from the standard normal
distribution, and then multiplied each mode of the core tensor with an orthonormal factor randomly
drawn from the Haar measure. For the gross corruption, we randomly generated the sparsity of
the corruption matrix s, and then randomly selected s elements in which we put values randomly
generated from uniform distribution. The additive independent Gaussian noise with variance ? 2
7
Nr = 4.0
3
2
1
0
mean error of low?rank tensor
15
20
25
30
35
40
Ns
|||?|||F
M
?5
x 10
N = 2.9
r
11
10
Nr = 4.0
9
Nr = 4.9
8
7
6
5
4
10
15
20
25
Ns
|||?|||F
M
5
30
35
against Ns of size n(2) . (e)
7
Ns = 25
4
6
Ns = 35
5
3
4
3
2
2
1
1
0
3
3.5
4
4.5
5
0
5.5
0
1
2
3
Nr
|||?|||F
M
against Ns of size n(1) . (b)
12
?6
x 10
8
Ns = 17
4
5
6
7
?6
x 10
?1
against Nr of size n(1) .
(c) ?1 against ?2 of size n(1) .
?5
x 10
9
?6
x 10
Ns = 15.8
4
8
Ns = 22.4
3.5
7
6
Ns = 31.6
3
2.5
?2
10
(a)
(d)
Nr = 5.4
4
?4
x 10
6
?2
5
mean error of low?rank tensor
Nr = 2.9
mean error of low?rank tensor
mean error of low?rank tensor
?4
x 10
6
5
2
4
1.5
3
1
2
1
1.5
2
2.5
3
3.5
0.5
0.5
4
Nr
|||?|||F
M
against Nr of size n(2) .
1
1.5
2
2.5
?1
3
3.5
4
?6
x 10
(f) ?1 against ?2 of size n(2) .
Figure 1: Results of robust noisy tensor decomposition with corruption, under different sizes.
was added to the observations of elements. We use the alternating direction method of multipliers
(ADMM) to solve the minimization problem (2). The whole experiments were repeated 50 times and
the averaged results are reported.
?
?
The results are shown in Figure 1, where Nr = K
Ns = s. In Figure 1(a, d), we first fix
k=1
rk /K ,and
c
Nr at different values, and then draw the value of W ? W ? /N against Ns . Similarly, in Figure 1(b,
F
c ? W ? /N against Nr . In Figure 1(c, f), we
e), we first fix Ns at different values, and then draw W
F
c ? W ? /N scales linearly
study the values of ?1 and ?2 at various settings. We can see that W
F
with both Ns and Nr . Similar scalings of Vb ? V ? /N can be observed, hence we omit them due
F
to space limitation. We can also observe from
Figure1(c, f) that,
under
various settings, ?1 ? ?2 ,
c
this finding is consistent with the fact that W ? W ? /N ? Vb ? V ? /N . All these results are
F
F
consistent with each other, validating our theoretical analysis.
P
7
Conclusions
In this paper, we analyzed the statistical performance of robust noisy tensor decomposition with
corruption. Our goal is to recover a pair of tensors, based on observing a noisy contaminated version
of their sum. It is based on solving a convex optimization with composite regularizations of Schatten-1
norm and `1 norm defined on tensors. We provided a general nonasymptotic estimator error bounds on
the underly low-rank tensor and sparse corruption tensor. Furthermore, the error bound we obtained
in this paper is new, and non-comparable with previous theoretical analysis.
Acknowledgement
We would like to thank the anonymous reviewers for their helpful comments. Research was sponsored
in part by the Army Research Lab, under Cooperative Agreement No. W911NF-09-2-0053 (NSCTA),
the Army Research Office under Cooperative Agreement No. W911NF-13-1-0193, National Science
Foundation IIS-1017362, IIS-1320617, and IIS-1354329, HDTRA1-10-1-0120, and MIAS, a DHSIDS Center for Multimodal Information Access and Synthesis at UIUC.
8
References
[1] A. Agarwal, S. Negahban, and M. J. Wainwright. Noisy matrix decomposition via convex relaxation:
Optimal rates in high dimensions. The Annals of Statistics, 40(2):1171?1197, 04 2012.
[2] P. J. Bickel, Y. Ritov, and A. B. Tsybakov. Simultaneous analysis of lasso and dantzig selector. The Annals
of Statistics, pages 1705?1732, 2009.
[3] S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein. Distributed optimization and statistical learning
R in Machine Learning,
via the alternating direction method of multipliers. Foundations and Trends
3(1):1?122, 2011.
[4] J.-F. Cai, E. J. Cand`es, and Z. Shen. A singular value thresholding algorithm for matrix completion. SIAM
Journal on Optimization, 20(4):1956?1982, 2010.
[5] E. J. Cand`es, X. Li, Y. Ma, and J. Wright. Robust principal component analysis? J. ACM, 58(3):11, 2011.
[6] E. J. Cand`es and Y. Plan. Matrix completion with noise. Proceedings of the IEEE, 98(6):925?936, 2010.
[7] E. J. Cand`es and B. Recht. Exact matrix completion via convex optimization. Commun. ACM, 55(6):111?
119, 2012.
[8] E. J. Cand`es and T. Tao. The power of convex relaxation: near-optimal matrix completion. IEEE
Transactions on Information Theory, 56(5):2053?2080, 2010.
[9] V. Chandrasekaran, S. Sanghavi, P. A. Parrilo, and A. S. Willsky. Rank-sparsity incoherence for matrix
decomposition. SIAM Journal on Optimization, 21(2):572?596, 2011.
[10] P. Gong, J. Ye, and C. Zhang. Robust multi-task feature learning. In Proceedings of the 18th ACM SIGKDD
international conference on Knowledge discovery and data mining, pages 895?903. ACM, 2012.
[11] E. T. Hale, W. Yin, and Y. Zhang. Fixed-point continuation for \ell 1-minimization: Methodology and
convergence. SIAM Journal on Optimization, 19(3):1107?1130, 2008.
[12] D. Hsu, S. M. Kakade, and T. Zhang. Robust matrix decomposition with sparse corruptions. IEEE
Transactions on Information Theory, 57(11):7221?7234, 2011.
[13] T. G. Kolda and B. W. Bader. Tensor decompositions and applications. SIAM Review, 51(3):455?500,
2009.
[14] L. D. Lathauwer, B. D. Moor, and J. Vandewalle. On the best rank-1 and rank-(r1,r2,. . .,rn) approximation
of higher-order tensors. SIAM J. Matrix Anal. Appl., 21(4):1324?1342, 2000.
[15] J. Liu, P. Musialski, P. Wonka, and J. Ye. Tensor completion for estimating missing values in visual data.
IEEE Trans. Pattern Anal. Mach. Intell., 35(1):208?220, 2013.
[16] C. Mu, B. Huang, J. Wright, and D. Goldfarb. Square deal: Lower bounds and improved relaxations for
tensor recovery. CoRR, abs/1307.5870, 2013.
[17] S. Negahban and M. J. Wainwright. Estimation of (near) low-rank matrices with noise and high-dimensional
scaling. The Annals of Statistics, 39(2):1069?1097, 04 2011.
[18] S. N. Negahban, P. Ravikumar, M. J. Wainwright, and B. Yu. A unified framework for high-dimensional
analysis of m-estimators with decomposable regularizers. Statistical Science, 27(4):538?557, 11 2012.
[19] N. Srebro and A. Shraibman. Rank, trace-norm and max-norm. In COLT, pages 545?560, 2005.
[20] R. Tomioka, K. Hayashi, and H. Kashima. Estimation of low-rank tensors via convex optimization. 2010.
[21] R. Tomioka and T. Suzuki. Convex tensor decomposition via structured schatten norm regularization. In
NIPS, pages 1331?1339, 2013.
[22] R. Tomioka, T. Suzuki, K. Hayashi, and H. Kashima. Statistical performance of convex tensor decomposition. In NIPS, pages 972?980, 2011.
[23] M. A. O. Vasilescu and D. Terzopoulos. Multilinear analysis of image ensembles: Tensorfaces. In ECCV
(1), pages 447?460, 2002.
[24] R. Vershynin. Introduction to the non-asymptotic analysis of random matrices.
arXiv:1011.3027, 2010.
[25] E. Yang and P. D. Ravikumar. Dirty statistical models. In NIPS, pages 611?619, 2013.
9
arXiv preprint
| 5409 |@word version:2 polynomial:1 norm:38 vi1:2 nscta:1 decomposition:37 mention:1 initial:1 liu:1 recovered:5 chu:1 additive:1 underly:1 numerical:3 sponsored:1 update:1 v:1 selected:1 core:2 provides:4 location:1 zhang:3 mathematical:1 lathauwer:1 ik:14 prove:1 theoretically:1 behavior:2 cand:5 uiuc:1 multi:4 cardinality:1 increasing:1 considering:1 begin:1 provided:2 bounded:4 underlying:1 notation:3 moreover:2 estimating:1 what:1 compressive:1 unified:1 finding:1 transformation:1 shraibman:1 nj:1 guarantee:1 exactly:2 rm:1 k2:4 uk:3 control:1 omit:2 appear:1 positive:2 before:1 engineering:1 local:1 limit:1 mach:1 incoherence:1 might:1 studied:2 k:3 dantzig:1 appl:1 bi:1 statistically:1 averaged:1 practical:1 unique:1 practice:2 procedure:1 universal:1 composite:2 boyd:1 get:1 operator:9 put:1 equivalent:1 conventional:1 deterministic:3 demonstrated:1 lagrangian:3 reviewer:1 center:1 attention:1 starting:1 missing:1 convex:20 shen:1 decomposable:2 recovery:3 immediately:2 estimator:4 insight:1 utilizing:1 importantly:1 nuclear:1 orthonormal:1 financial:1 annals:3 kolda:1 suppose:2 exact:2 agreement:2 overlapped:4 element:7 trend:1 recognition:1 particularly:1 decomposibility:5 cooperative:2 observed:1 preprint:1 hv:1 gross:12 hanj:1 convexity:3 mu:1 solving:4 algebra:2 max1:2 gu:1 multimodal:1 various:4 whose:4 quite:1 solve:4 statistic:3 noisy:12 eigenvalue:1 cai:1 propose:2 product:1 realization:1 adjoint:1 frobenius:3 validate:1 ky:3 convergence:1 optimum:2 r1:7 tk:1 help:1 completion:8 gong:1 strong:2 recovering:3 auxiliary:1 direction:3 bader:1 fix:2 generalization:3 anonymous:1 multilinear:1 summation:2 extension:3 strictly:1 hold:4 practically:1 considered:3 wright:2 normal:1 exp:2 mapping:1 predict:1 bickel:1 estimation:9 rpca:2 superposition:2 quanquan:1 largest:2 successfully:1 tool:1 moor:1 hope:1 minimization:7 unfolding:4 gaussian:4 aim:4 pn:2 varying:1 ikk:1 office:1 corollary:3 parafac:1 focus:1 vk:4 rank:67 check:1 contrast:1 sigkdd:1 helpful:1 stopping:1 sb:1 jiawei:1 going:1 i1:5 tao:1 interested:1 pixel:1 arg:3 dual:2 colt:1 denoted:3 plan:1 constrained:1 special:7 ell:1 equal:1 xvt:1 kw:2 yu:1 mias:1 hdtra1:1 contaminated:3 np:1 sanghavi:1 few:1 randomly:5 simultaneously:3 national:1 intell:1 maxj:1 occlusion:1 n1:7 gui:1 ab:1 interest:2 mining:2 ai1:1 evaluation:1 introduces:1 analyzed:4 primal:2 regularizers:2 devoted:1 kt:5 partial:1 huan:1 orthogonal:2 conduct:1 theoretical:8 witnessed:1 column:4 soft:2 cover:1 w911nf:2 entry:7 decomposability:1 uniform:1 kxks:2 vandewalle:1 conducted:1 reported:1 answer:1 corrupted:2 synthetic:1 vershynin:1 recht:1 international:1 negahban:3 siam:5 xi1:2 together:1 ym:2 synthesis:1 rn1:14 huang:1 combing:1 supp:2 li:1 nonasymptotic:3 prox:2 parrilo:1 bold:2 coefficient:1 vi:1 view:2 lab:1 analyze:1 observing:1 portion:1 recover:9 complicated:1 contribution:1 il:1 square:1 variance:7 qk:1 ensemble:2 yield:2 identify:1 accurately:1 basically:1 corruption:33 simultaneous:1 kpk:2 vasilescu:1 against:8 tucker:1 associated:1 proof:1 recovers:2 boil:1 hsu:1 treatment:1 knowledge:1 dimensionality:1 formalize:1 musialski:1 back:1 higher:2 follow:1 methodology:1 improved:1 entrywise:1 arranged:1 formulation:1 though:1 strongly:1 ritov:1 furthermore:4 reply:1 d:7 hand:2 mode:20 ye:2 concept:1 true:10 multiplier:4 k22:4 regularization:7 equality:1 rbk:1 alternating:3 hence:1 nonzero:2 iteratively:1 goldfarb:1 i2:2 deal:1 encourages:1 please:1 criterion:1 generalized:2 complete:1 performs:1 cp:1 image:3 wise:1 novel:1 recently:1 parikh:1 common:1 salt:1 discussed:2 extend:2 refer:1 vec:3 ai:1 cv:1 pm:2 similarly:2 illinois:2 access:1 han:1 longer:1 v0:1 showed:1 perspective:2 commun:1 scenario:1 certain:2 inequality:5 calligraphic:1 arbitrarily:1 vt:2 yi:4 seen:3 additional:1 impose:1 ii:4 relates:1 multiple:1 champaign:1 technical:2 match:1 ravikumar:2 prediction:1 regression:2 vision:1 essentially:1 arxiv:2 agarwal:1 background:2 whereas:1 separately:7 addition:3 residue:1 singular:6 rest:1 unlike:2 comment:1 subject:2 validating:1 call:1 near:3 yang:1 easy:4 pepper:1 lasso:1 reduce:1 incomparable:1 whether:1 penalty:1 remark:1 deep:1 useful:1 covered:1 clear:1 rankk:1 tsybakov:1 generate:1 continuation:1 exist:2 notice:1 estimated:1 group:1 key:2 tensorfaces:1 drawn:1 rrk:1 relaxation:3 fraction:1 sum:3 letter:4 extends:1 chandrasekaran:1 draw:2 summarizes:1 scaling:3 vb:2 comparable:1 rnk:5 bound:20 followed:1 nontrivial:1 precisely:1 constraint:1 min:6 optimality:1 department:2 structured:1 appealing:2 kakade:1 making:1 s1:9 ks1:10 restricted:4 pr:1 remains:1 vectorizations:1 operation:1 multiplied:1 apply:1 observe:1 spectral:1 kashima:2 dirty:1 xw:1 ink:1 tensor:130 question:2 added:1 traditional:1 diagonal:2 unclear:1 nr:13 gradient:1 thank:1 schatten:13 ei1:1 w0:1 willsky:1 length:1 modeled:1 index:2 trace:6 wonka:1 anal:2 motivates:1 unknown:7 upper:2 observation:16 datasets:1 urbana:2 finite:1 defining:1 extended:2 y1:2 rn:3 arbitrary:2 peleato:1 complement:1 namely:1 pair:1 eckstein:1 nip:3 trans:1 address:2 able:3 pattern:1 xm:1 candecomp:1 sparsity:5 program:2 including:1 max:2 video:1 wainwright:3 power:1 haar:1 residual:4 older:3 concludes:1 review:1 acknowledgement:1 discovery:1 asymptotic:1 limitation:1 srebro:1 foundation:2 consistent:2 imposes:1 thresholding:3 viewpoint:1 row:3 prone:1 eccv:1 keeping:1 understand:1 terzopoulos:1 face:2 sparse:20 distributed:1 overcome:1 dimension:2 collection:1 suzuki:2 far:1 transaction:2 approximate:1 selector:1 confirm:1 global:1 factorize:1 xi:7 demanded:1 vectorization:4 decomposes:1 sk:2 robust:18 obtaining:1 flattening:2 pk:21 dense:3 main:3 linearly:1 whole:1 noise:17 n2:2 repeated:1 body:1 x1:1 augmented:2 referred:2 n:14 tomioka:3 structurally:1 sub:1 wish:2 hw:7 rk:22 departs:1 down:1 theorem:12 specific:1 hale:1 sensing:1 r2:2 consist:1 i11:1 corr:1 gained:1 magnitude:2 illumination:2 kx:6 nk:32 yin:1 saddle:2 army:2 forbenius:1 visual:1 scalar:3 hayashi:2 corresponds:1 acm:4 ma:1 goal:4 identity:2 admm:4 change:1 specifically:2 wt:2 denoising:1 principal:2 lemma:6 experimental:1 e:5 support:7 relevance:1 princeton:3 |
4,871 | 541 | Network activity determines
spatio-temporal integration in single cells
Ojvind Bernander, Christof Koch *
Computation and Neural Systems Program,
California Institut.e of Technology,
Pasadena, Ca 91125, USA.
Rodney J. Douglas
Anatomical Neuropharmacology Unit,
Dept. Pharmacology,
Oxford, UK.
Abstract
Single nerve cells with static properties have traditionally been viewed
as the building blocks for networks that show emergent phenomena. In
contrast to this approach, we study here how the overall network activity
can control single cell parameters such as input resistance, as well as time
and space constants, parameters that are crucial for excitability and spatiotemporal integration. Using detailed computer simulations of neocortical
pyramidal cells, we show that the spontaneous background firing of the
network provides a means for setting these parameters. The mechanism
for this control is through the large conductance change of the membrane
that is induced by both non-NMDA and NMDA excitatory and inhibitory
synapses activated by the spontaneous background activity.
1
INTRODUCTION
Biological neurons display a complexity rarely heeded in abstract network models.
Dendritic trees allow for local interactions, attenuation, and delays. Voltage- and
*To whom all correspondence should be a.ddressed.
43
44
Bernander, Koch, and Douglas
time-dependent conductances can give rise to adaptation, burst-firing, and other
non-linear effects. The extent of temporal integration is determined by the time
constant, and spatial integration by the "leakiness" of the membrane. It is unclear
which cell properties are computationally significant and which are not relevant
for information processing, even though they may be important for the proper
functioning of the cell. However, it is crucial to understand the function of the
component cells in order to make relevant abstractions when modeling biological
systems. In this paper we study how the spontaneous background firing of the
network as a whole can strongly influence some of the basic integration properties
of single cells.
1.1
Controlling parameters via background synaptic activity
The input resistance, RJn, is defined as ~, where dV is the steady state voltage
change in response to a small current step of amplitude dI. RJn will vary throughout
the cell, and is typically much larger in a long, narrow dendrite than in the soma.
However, the somatic input resistance is more relevant to the spiking behavior of
the neuron, since spikes are initiated at or close to the soma, and hence Rin,.oma
(henceforth simply referred to as Rin) will tell us something of the sensitivity of the
cell to charge reaching the soma.
The time constant, Tm , for a passive membrane patch is Rm . em, the membrane
resistance times the membrane capacitance. For membranes containing voltagedependent non-linearities, exponentials are fitted to the step response and the
largest. time constant is taken to be the membrane time constant. A large time
constant implies that any injected charge leaks away very slowly, and hence the cell
has a longer "memory" of previous events.
The parameters discussed above (Rin, Tm) clearly have computational significance
and it would be convenient to be able to chanfe them dynamically. Both depend
directly on the membrane conductance G m = Jr.;' so any change in G m will change
the ?parameters. Traditionally, however, G m has been viewed as static, so these
parameters have also been considered static. How can we change G m dynamically?
In traditional models, G m has two components: active (time- and voltagedependent) conductances and a passive "leak" conductance. Synapses are modeled as conductance changes, but if only a few are activated, the cable structure
of the cell will hardly change at all. However, it is well known that neocortical
neurons spike spontaneously, in the absence of sensory stimuli, at rates from 0 to
10 Hz. Since neocortical neurons receive on the order of 5,000 to 15,000 excitatory
synapses (Larkman, 1991), this spontaneous firing is likely to add up to a large total
conductance (Holmes & Woody, 1989) . This synaptic conductance becomes crucial
if the non-synaptic conductance components are small. Recent evidence show indeed that the non-synaptic conductances are relatively small (when the cell is not
spiking) (Anderson et aI., 1990). Our model uses a leak Rm
100,000 kOcm 2 ,
instead of more conventional values in the range of 2,500-10,000 kOcm 2 ? These
two facts, high Rm and synaptic background activity, allow R in and Tm to change
by more than ten-fold, as described below in this paper.
=
Nerwork activity determines spatio-temporal integration in single cells
2
MODEL
A typical layer V pyramidal cell (fig. 2) in striate cortex was filled with HRP during in vivo experiments in the anesthetized, adult cat (Douglas et aI., 1991). The
3-D coordinates and diameters of the dendritic tree were measured by a computerassisted method and each branch was replaced by a single equivalent cylinder. This
morphological data was fed into a modified version of NEURON, an efficient single cell simulator developed by Hines (1989). The dendrites were passive, while the
soma contained seven active conductances, underlying spike generation, adaptation,
and slow onset for weak stimuli. The model included two sodium conductances (a
fast spiking current and a. fJlower non-inactivating current), one calcium conductance, and four potassium conductances (delayed rectifier, slow 'M' and 'A' type
currents, and a calcium-dependent current). The active conductances were modeled
using a Hodgkin-Huxley-like formalism.
The model used a total of 5,000 synapses. The synaptic conductance change
in time was modeled with an alpha function, get) = ~.,.,... e te- tlt .,.... 4,000
synapses were fast excitatory non-NMDA or AMPA-type (tped 1.5 msec, gpeaJ: =
0.5 nS, E retJ
0 mV), 500 were medium-slow inhibitory GABAA (tpe4k =
10 msec, gpeole 1.0 nS, E retJ -70 mV), and 500 were slow inhibitory GABAB
(tpeok 40 msec, gpeok 0.1 nS, E,.etJ -95 mV). The excitatory synapses were
less concentrated towards the soma, while the inhibitory ones were more so. For a
more detailed description of the model, see Bernander et al. (1991).
...=.
=
=
=
=
=
=
120r------------------------------,
~~-----------------------,
RIn, no~NMDA
Rin, no~MDA and NMDA
100
60
i
140
i
40
20
./
""
-
.....
.....
--- --- --
20
'"--
2
3
4
5
6
Background frequency (Hz)
7
2
,3
4
5
6
7
Background frequency (Hz)
Figure 1: Input resistance and time constant as a function of background
frequency. In (a), the solid line corresponds to the "standard" model with passive
dendrites, while the dashed line includes active NMDA synapses as described in the
text.
45
46
Bernander, Koch, and Douglas
3
3.1
RESULTS
R,n and
Tm
change with background frequency
Fig. 1 illustrates what happens to ~n and Tm when the synaptic background activities of all synaptic types are varied simultaneously. In the absence of any synaptic
input, ~n
110 Mn and Tm
80 msec. At 1 Hz background activity, on average 5 synaptic events are impinging on the cell every msec, contributing a total
of 24 nS to the somatic input conductance Gin. Because of the reversal potential
of the excitatory synapses (0 mV), the membrane potential throughout the cell is
pulled towards more depolarizing potentials, activating additional active currents.
Although these trends continue as f is increased, the largest change can be observed
between 0 and 2 Hz.
=
=
Figure 2: Spatial integration as a function of background frequency.
Each dendrite has been "stretched" so that its apparent length corresponds to its
electrotonic length. The synaptic background frequency was 0 Hz (left) and 2 Hz
(right). The scale bar corresponds to 1 A (length constant).
Activating synaptic input has two distinct effects: the conductance of the postsynaptic membrane increases and the membrane is depolarized. The system can,
at least in principle, independently control these two effects by differentially varying the spontaneous firing frequencies of excitatory versus inhibitory inputs. Thus,
increasing f selectively for the GABAB inhibition will further increase the membrane conductance but move the resting potential towards more hyperpolarizing
Network activity determines spatio-temporal integration in single cells
potentials.
Note that the 0 Hz ca?c corresponds to experiments made with in vitro slice preparations or culture. In this case incoming fibers have been cut off and the spontaneous
firing rate is very small. Careful studies have shown very large values for Rin and
Tm under these circumstances (e.g. Spruston &. Johnston, 1991). In vivo preparations, on the other hand, leave the cortical circuitry intact and much smaller values
of R,n and Tm are usually recorded.
3.2
Spatial integration
Varying synaptic background activity can have a significant impact on the electrotonic structure of the cell (fig. 2). We plot the electrotonic distance of any particular
point from the cell body, that is the sum of the electrotonic length's L, = Ej(lj/Aj)
=J
associated with each dendritic segment i, where Aj
~m.R~j is the electrotonic
length constant of compartment i, Ij its anatomical length and the sum is taken
over all compartments between the soma and compartment i.
=
=
Increasing the synaptic background activity from I
0 to f
2 Hz has the effect
of stretching the "distance" of any particular synapse t.o the soma by a factor of
about 3, on average. Thus, while a distal synapse has an associated L value of
about 2.6 at 2 Hz it shrinks to 1.2 if all network activity is shut off, while for a
synapse at the tip of a basal dendrite, L shrinks from 0.7 t.o 0.2. In fact, the EPSP
induced by a single excitatory synapse at that location goes from 39 to 151 J,lV, a
decrease of about 4. Thus, when the overall network activity is low, synapses in the
superficial layer of cortex could have a significant effect on somatic discharge, while
having only a weak modulatory effect on the soma if the overall network activity is
high. Note that basal dendrites, which receive a larger number of synapses, stretch
more than apical dendrites.
3.3
Temporal integration
That the synaptic background activity can also modify the temporal integration
behavior of the cell is demonstrated in fig. 3. At any part.icular background frequency I, we compute the minimal number of additional excitatory synapses (at
gpeal: = 0.5 nS) necessary to barely generate one action potential. These synapses
were chosen randomly from among all excitatory synapses throughout the cell. We
compare the case in which all synapses are activated simultaneously (solid line)
with the case in which the inputs arrive asynchronously, smeared out over 25 msec
(dashed line). If I = 0, it requires 115 synapses firing simultaneously to generate
a single action potential, while 145 are needed if the input is desynchronized. This
small difference between inputs arriving synchronized and at random is due to the
long integration period of the cell.
If the background activity increases to f = 1 Hz, 113 synchronized synaptic
inputs-spread out all over the cell-are sufficient to fire the cell. If, however,
the synaptic input is spread out over 25 msec, 202 synapses are now needed in
order to trigger a response from the cell. This is mainly due to the much smaller
value of Tm relative to the period over which the synaptic input is spread out. Note
47
48
Bernander, Koch, and Douglas
that the difference in number of simultaneous synaptic inputs needed to fire the
cell for f
0 compared to f = 1 is small (i.e. 113 vs. 115), in spite of the more
than five-fold decrease in somatic input resistance. The effect of the smaller size of
the individual EPSP at higher values of f is compensated for by the fact that the
resting potential of the cell has been shifted towards the firing threshold of the cell
(about -49 mY).
=
,~----------------------------------------------,
-:
800
Unsynchronized Input
Synchronized input
o
I
!
600
...=-
-
J
E
:;,
z
?OL-------'------~2------~3~----~4------~5~----~I~----~7
Background frequency (Hz)
Figure 3: Phase detection.
A variable number of excitatory synapses were fired superimposed onto a constant
background frequency of 1 Hz. They fired either simultaneously (solid line) or
spread out in time uniformly during a 25 msec interval (dashed line). The y axis
shows the minimum number of synapses necessary to cause the cell to fire.
3.4
NMDA synapses
Fast excitatory synaptic input in cortex is mediated by both AMPA or non-NMDA
as well as NMDA receptors (Miller et aI., 1989). As opposed to the AMPA synapse,
the NMDA conductance change depends not only on time but also on the postsynaptic voltage:
(1)
where '1'1 = 40 msec, '1'2 = 0.335 msec, '1 = 0.33 mM-t, [M g2+] - 1 mM,
r = 0.06 mV-1. During spontaneous background activity many inputs impinge
on the cell and we can time-average the equation above. We will then be left with
a purely voltage-dependent conductance.
We measured the somatic input resistance, Rin, by injecting a small current pulse in
the soma (fig. 4) in the standard model. All synapses fired at a 0.5 Hz background
frequency. Next we added 4,000 NMDA synapses in addition to the 4,000 non-
Network activity determines spatio-temporal integration in single cells
NMDA synapses, also at 0.5 Hz, and again injected a current pulse. The voltage
response is now larger by about 65%, corresponding to a smaller input conducta.nce,
even though we are adding the positive NMDA conductance. This seeming paradox
depends on two effects. First, the input conductance is, by definition, ~
G(V)+
(V - Ern), where G(V) is the conductance specified in eq. (1). For the
N DA synapse this derivative is negative below about -35 mV. Second, due to the
excitation the membrane voltage has drifted towards more depolarized values. This
will cause a change in the activation of the other voltage-dependent currents. Even
though the summed conductance of these active currents will be larger at the new
voltage, the derivative
will be smaller at that point. In other words, activation
of NMDA synapses gives a negative contribution to the input conductance, even
though more conductances have opened up.
=
di) .
'*"
Next we replaced 2,000 of the 4,000 non-NMDA synapses in the old model with
2,000 NMDA synapses and recomputed the input resistance as a function of synaptic background activity. The result is overlaid in figure 1a (dashed line). The curve
shifts toward larger values of Rin for most values of f. This shift varies between
50 % - 200 %. The cell is more excitable than before.
-60
>
E
E
>
-61
-62
-63
-64
-65
-66
0
400
200
t
600
800
1000
lmsecl
Figure 4: Negative input conductance from NMDA activation.
At times t
250 msec and t
750 msec a 0.05 nA current pulse was injected
at the soma and the somatic voltage response was recorded. At t = 500 msec,
one NMDA synapse was activated for each non-NMDA synapse, for a total of 8,000
excitatory synaptic inputs. The background frequency was 0.5 Hz for all synapses.
=
4
=
DISCUSSION
We have seen that parameters such as Rtn, 7'm, and L are not static, but can
vary over about one order of magnitude under network control. The potential
computational possibilities could be significant.
49
50
Bernander, Koch, and Douglas
For example, if a low-contrast stimulus is presented within the receptive field of
the cell, the synaptic input rate will be small and the signal-t~noise ratio (SNR)
low. In this case, to make the cell more sensitive to the inputs we might want to
increase R;n. This would automatically be achieved as the total network activation
is low. We can improve the SNR by integrating over a longer time period, i.e. by
increasing Tm. This would also be a consequence of the reduced network activity.
The converse argument can be made for high-contrast stimuli, associated with high
overall network activity and low R;n and Tm values.
Many cortical cells are tuned for various properties of the stimulus, such as orientation, direction, and binocular disparity. As the effective membrane conductance,
G m , changes, the tuning curves are expected to change. Depending on the exact
circuitry and implementation of the tuning properties, this change in background
frequency could take many forms. One example of phase-tuning was given above.
In this case the temporal tuning increases with background frequency.
Acknowledgements
This work was supported by the Office of Naval Research, the National Science
Foundation, the James McDonnell Foundation and the International Human Frontier Science Program Organization. Thanks to Tom Tromey for writing the graphic
software and to Mike Hines for providing us with NEURON.
References
P. Anderson, M. Raastad &, J. F. Storm. (1990) Excitatory synaptic integration in
hippocampal pyramids and dentate granule cells. Symp. Quant. Bioi. 55, Cold
Spring Harbor Press, pp. 81-86.
O.
Bernander, R. J. Douglas, K. A. C. Martin &, C. Koch. (1991) Synaptic
background activity influences spatiotemporal integration in single pyramidal cells.
P.N.A.S, USA 88: 11569-11573.
R. J. Douglas, K. A. C. Martin &, D. Whitteridge. (1991) An intracellular analysis
of the visual responses of neurones in cat visual cortex. J. Physiol. 440: 659-696.
M. Hines. (1989) A program for simulation of nerve equations with branching
geometries. Int. J. Biomed. Comput. 24: 55-68.
W. R. Holmes &, C. D. Woody. (1989) Effects of uniform and non-uniform synaptic activation-distributions on the cable properties of modeled cortical pyramidal
neurons. Brain Research 505: 12-22.
A. U. Larkman. (1991) Dendritic morphology of pyramidal neurones of the visual
cortex of the rat: III. Spine distributions. J. Compo Neurol. 306: 332-343.
K. D. Miller, B. Chapman &, M. P. Stryker. (1989) Responses of cells in cat visual
cortex depend on NMDA receptors. P.N.A.S. 86: 5183-5187.
N. Spruston &, D. Johnston. (1992) Perforated patch-clamp analysis of the passive
membrane properties of three classes of hippocampal neurons. J. Netlrophysiol., in
press.
| 541 |@word version:1 pulse:3 simulation:2 solid:3 disparity:1 tuned:1 current:11 activation:5 physiol:1 hyperpolarizing:1 plot:1 v:1 shut:1 compo:1 leakiness:1 provides:1 location:1 five:1 burst:1 symp:1 expected:1 indeed:1 spine:1 behavior:2 simulator:1 ol:1 brain:1 morphology:1 automatically:1 increasing:3 becomes:1 linearity:1 underlying:1 medium:1 what:1 developed:1 temporal:8 kocm:2 every:1 attenuation:1 charge:2 rm:3 uk:1 control:4 unit:1 converse:1 christof:1 inactivating:1 positive:1 before:1 local:1 modify:1 consequence:1 receptor:2 oxford:1 initiated:1 firing:8 might:1 dynamically:2 range:1 spontaneously:1 block:1 cold:1 convenient:1 word:1 integrating:1 spite:1 get:1 onto:1 close:1 influence:2 writing:1 conventional:1 equivalent:1 demonstrated:1 compensated:1 go:1 independently:1 holmes:2 ojvind:1 traditionally:2 coordinate:1 discharge:1 spontaneous:7 controlling:1 trigger:1 exact:1 us:1 larkman:2 trend:1 cut:1 observed:1 mike:1 morphological:1 decrease:2 leak:3 complexity:1 depend:2 segment:1 purely:1 gabaa:1 rin:8 emergent:1 cat:3 fiber:1 various:1 distinct:1 fast:3 effective:1 tell:1 apparent:1 larger:5 asynchronously:1 clamp:1 interaction:1 epsp:2 adaptation:2 relevant:3 fired:3 description:1 differentially:1 potassium:1 etj:1 leave:1 depending:1 measured:2 ij:1 eq:1 implies:1 synchronized:3 direction:1 opened:1 human:1 activating:2 biological:2 dendritic:4 frontier:1 stretch:1 mm:2 koch:6 considered:1 overlaid:1 dentate:1 circuitry:2 vary:2 injecting:1 sensitive:1 largest:2 smeared:1 clearly:1 modified:1 reaching:1 ej:1 varying:2 voltage:9 office:1 naval:1 superimposed:1 mainly:1 contrast:3 dependent:4 abstraction:1 typically:1 lj:1 pasadena:1 biomed:1 overall:4 among:1 orientation:1 spatial:3 integration:15 summed:1 field:1 having:1 chapman:1 stimulus:5 few:1 randomly:1 simultaneously:4 national:1 individual:1 delayed:1 hrp:1 geometry:1 replaced:2 phase:2 fire:3 cylinder:1 conductance:29 detection:1 organization:1 possibility:1 activated:4 necessary:2 culture:1 institut:1 tree:2 filled:1 spruston:2 old:1 minimal:1 fitted:1 increased:1 formalism:1 modeling:1 apical:1 snr:2 uniform:2 delay:1 graphic:1 varies:1 spatiotemporal:2 my:1 thanks:1 international:1 sensitivity:1 off:2 tip:1 na:1 again:1 recorded:2 containing:1 opposed:1 slowly:1 henceforth:1 derivative:2 potential:9 seeming:1 includes:1 int:1 mv:6 onset:1 depends:2 rodney:1 depolarizing:1 vivo:2 contribution:1 compartment:3 stretching:1 miller:2 oma:1 weak:2 simultaneous:1 synapsis:26 synaptic:26 definition:1 frequency:14 pp:1 james:1 storm:1 associated:3 di:2 static:4 gabab:2 nmda:20 amplitude:1 nerve:2 higher:1 tom:1 response:7 synapse:8 though:4 strongly:1 anderson:2 shrink:2 binocular:1 hand:1 unsynchronized:1 aj:2 building:1 effect:9 usa:2 functioning:1 hence:2 excitability:1 distal:1 during:3 branching:1 steady:1 excitation:1 rat:1 hippocampal:2 neocortical:3 passive:5 spiking:3 vitro:1 woody:2 discussed:1 resting:2 neuropharmacology:1 significant:4 ai:3 stretched:1 whitteridge:1 tuning:4 longer:2 cortex:6 inhibition:1 add:1 something:1 recent:1 continue:1 seen:1 minimum:1 additional:2 period:3 dashed:4 signal:1 branch:1 long:2 impact:1 basic:1 circumstance:1 pyramid:1 achieved:1 cell:40 receive:2 background:27 want:1 addition:1 interval:1 johnston:2 pyramidal:5 desynchronized:1 crucial:3 depolarized:2 induced:2 hz:16 iii:1 harbor:1 quant:1 tm:11 shift:2 resistance:8 neurones:2 cause:2 hardly:1 action:2 electrotonic:5 modulatory:1 detailed:2 ten:1 concentrated:1 diameter:1 reduced:1 generate:2 inhibitory:5 shifted:1 anatomical:2 icular:1 basal:2 recomputed:1 four:1 soma:10 threshold:1 douglas:8 rjn:2 sum:2 nce:1 injected:3 hodgkin:1 arrive:1 throughout:3 patch:2 layer:2 display:1 correspondence:1 fold:2 activity:22 mda:1 huxley:1 software:1 argument:1 spring:1 relatively:1 martin:2 ern:1 mcdonnell:1 membrane:15 jr:1 smaller:5 em:1 postsynaptic:2 voltagedependent:2 cable:2 happens:1 dv:1 taken:2 computationally:1 equation:2 mechanism:1 needed:3 fed:1 reversal:1 away:1 granule:1 move:1 capacitance:1 added:1 spike:3 receptive:1 stryker:1 striate:1 traditional:1 unclear:1 gin:1 distance:2 seven:1 whom:1 impinge:1 extent:1 barely:1 toward:1 length:6 modeled:4 ratio:1 providing:1 negative:3 rise:1 implementation:1 proper:1 calcium:2 neuron:8 paradox:1 varied:1 somatic:6 specified:1 california:1 narrow:1 adult:1 able:1 bar:1 below:2 usually:1 program:3 memory:1 event:2 sodium:1 mn:1 improve:1 technology:1 axis:1 rtn:1 bernander:7 mediated:1 excitable:1 perforated:1 text:1 acknowledgement:1 contributing:1 relative:1 generation:1 versus:1 lv:1 foundation:2 sufficient:1 principle:1 excitatory:13 supported:1 arriving:1 allow:2 understand:1 pulled:1 anesthetized:1 slice:1 curve:2 cortical:3 sensory:1 made:2 tlt:1 alpha:1 active:6 incoming:1 spatio:4 superficial:1 ca:2 dendrite:7 ampa:3 impinging:1 da:1 significance:1 spread:4 intracellular:1 whole:1 noise:1 pharmacology:1 body:1 fig:5 referred:1 slow:4 n:5 msec:13 exponential:1 comput:1 rectifier:1 neurol:1 evidence:1 adding:1 magnitude:1 te:1 illustrates:1 simply:1 likely:1 visual:4 drifted:1 contained:1 g2:1 corresponds:4 determines:4 hines:3 bioi:1 viewed:2 careful:1 towards:5 absence:2 change:16 included:1 determined:1 typical:1 uniformly:1 total:5 intact:1 rarely:1 selectively:1 preparation:2 dept:1 phenomenon:1 |
4,872 | 5,410 | PEWA: Patch-based Exponentially Weighted
Aggregation for image denoising
Charles Kervrann
Inria Rennes - Bretagne Atlantique
Serpico Project-Team
Campus Universitaire de Beaulieu, 35 042 Rennes Cedex, France
[email protected]
Abstract
Patch-based methods have been widely used for noise reduction in recent years.
In this paper, we propose a general statistical aggregation method which combines
image patches denoised with several commonly-used algorithms. We show that
weakly denoised versions of the input image obtained with standard methods, can
serve to compute an efficient patch-based aggregated estimator. In our approach,
we evaluate the Stein?s Unbiased Risk Estimator (SURE) of each denoised candidate image patch and use this information to compute the exponential weighted
aggregation (EWA) estimator. The aggregation method is flexible enough to combine any standard denoising algorithm and has an interpretation with Gibbs distribution. The denoising algorithm (PEWA) is based on a MCMC sampling and is
able to produce results that are comparable to the current state-of-the-art.
1
Introduction
Several methods have been proposed to solve the image denoising problem including anisotropic
diffusion [15], frequency-based methods [26], Bayesian and Markov Random Fields methods [20],
locally adaptive kernel-based methods [17] and sparse representation [10]. The objective is to estimate a clean image generally assumed to be corrupted with additive white Gaussian (AWG) noise.
In recent years, state-of-the-art results have been considerably improved and the theoretical limits of denoising algorithms are currently discussed in the literature [4, 14]. The most competitive
methods are mostly patch-based methods, such as BM3D [6], LSSC [16], EPLL [28], NL-Bayes
[12], inspired from the N(on)L(ocal)-means [2]. In the NL-means method, each patch is replaced
by a weighted mean of the most similar patches found in the noisy input image. BM3D combines
clustering of noisy patches, DCT-based transform and shrinkage operation to achieve the current
state-of-the-art results [6]. PLOW [5], S-PLE [24] and NL-Bayes [12], falling in the same category of the so-called internal methods, are able to produce very comparable results. Unlike BM3D,
covariances matrices of clustered noisy patches are empirically estimated to compute a Maximum
A Posteriori (MAP) or a Minimum-Mean-Squared-Error (MMSE) estimate. The aforementioned
algorithms need two iterations [6, 12, 18] and the performances are surprisingly very close to the
state-of-the-art in average while the motivation and the modeling frameworks are quite different. In
this paper, the proposed Patch-based Exponential Weighted Aggregation (PEWA) algorithm, requiring no patch clustering, achieves also the state-of-the-art results.
A second category of patch-based external methods (e.g. FoE [20], EPLL [28], MLP [3]) has been
also investigated. The principle is to approximate the noisy patches using a set of patches of an
external learned dictionary. The statistics of a noise-free training set of image patches, serve as
priors for denoising. EPLL computes a prior from a mixture of Gaussians trained with a database
of clean image patches [28]; denoising is then performed by maximizing the so-called Expected
Patch Log Likelihood (EPLL) criteria using an optimization algorithm. In this line of work, a multi1
layer perceptron (MLP) procedure exploiting a training set of noisy and noise-free patches was able
to achieve the state-of-the-art performance [3]. Nevertheless, the training procedure is dedicated
to handle a fixed noise level and the denoising method is not flexible enough, especially for real
applications when the signal-to-noise ratio is not known.
Recently, the similarity of patch pairs extracted from the input noisy image and from clean patch
dataset has been studied in [27]. The authors observed that more repetitions are found in the same
noisy image than in a clean image patch database of natural images; also, it is not necessary to
examine patches far from the current patch to find good matching. While the external methods
are attractive, computation is not always feasible since a very large collection of clean patches are
required to denoise all patches in the input image. Other authors have previously proposed to learn
a dictionary on the noisy image [10] or to combine internal and external information (LSSC) [16].
In this paper, we focus on internal methods since they are more flexible for real applications than
external methods. They are less computationally demanding and remain the most competitive.
Our approach consists in estimating an image patch from ?weakly? denoised image patches in the
input image. We consider the general problem of combining multiple basic estimators to achieve
an estimation accuracy not much worse than that of the ?best? single estimator in some sense. This
problem is important for practical applications because single estimators often do not perform as
well as their combinations. The most important and widely studied aggregation method that achieves
the optimal average risk is the Exponential Weighted Aggregation (EWA) algorithm [13, 7, 19].
Salmon & Le Pennec have already interpreted the NL-means as a special case of the EWA procedure
but the results of the extended version described in [21] were similar to [2].
Our estimator combination is then achieved through a two-step procedure, where multiple estimators
are first computed and are then combined in a second separate computing step. We shall see that
the proposed method can be thought as a boosting procedure [22] since the performance of the precomputed estimators involved in the first step are rather poor, both visually and in terms of peak
signal-to-noise ratio (PSNR). Our contributions are the following ones:
1. We show that ?weak? denoised versions of the input noisy images can be combined to get
a boosted estimator.
2. A spatial Bayesian prior and a Gibbs energy enable to select good candidate patches.
3. We propose a dedicated Monte Carlo Markov Chain (MCMC) sampling procedure to compute efficiently the PEWA estimator.
The experimental results are comparable to BM3D [6] and the method is implemented efficiently
since all patches can be processed independently.
2
Patch-based image representation and SURE estimation
Formally, we represent a n-dimensional image patch at location x ? X ? R2 as a vector f (x) ? Rn .
We define the observation patch v(x) ? Rn as: v(x) = f (x) + ?(x) where ?(x) ? N (0, ? 2 In?n )
represents the errors. We are interested in an estimator fb(x) of f (x) assumed to be independent of
f (x) that achieves a small L2 risk. We consider the Stein?s Unbiased Risk Estimator
R(fb(x)) = kv(x) ? fb(x)k2n ? n? 2
in the Mean Square Error sense such that E[R(fb(x))] = E[kf (x) ? fb(x)k2n ] (E denotes the mathematical expectation). SURE has been already investigated for image denoising using NL-means
[23, 9, 22, 24] and for image deconvolution in [25].
3
Aggregation by exponential weights
Assume a family {f? (x), ? ? ?} of functions such that the mapping ? ? f? (x) is measurable
and ? = {1, ? ? ? , M }. Functions f? (x) can be viewed as some pre-computed estimators of f (x) or
?weak? denoisers independent of observations v(x), and considered as frozen in the following. The
set of M estimators is assumed to be very large, that is composed of several hundreds of thousands
2
of candidates. In this paper, we consider aggregates that are weighted averages of the functions in
the set {f? (x), ? ? ?} with some data-dependent weights:
fb(x) =
M
X
w? (x) f? (x) such that w? (x) ? 0 and
?=1
M
X
w? (x) = 1.
(1)
?=1
As suggested in [19], we can associate two probability measures w(x) = {w1 (x), ? ? ? , wM (x)} and
?(x) = {?1 (x), ? ? ? , ?M (x)} on {1, ? ? ? , M } and we define the Kullback-Leibler divergence as:
DKL (w(x), ?(x)) =
M
X
w? (x) log
?=1
w? (x)
?? (x)
.
(2)
The exponential weights are obtained as the solution of the following optimization problem:
(M
)
X
b
w(x)
= arg min
w? (x)?(R(f? (x))) + ? DKL (w(x), ?(x)) subject to (1)
w(x)?RM
(3)
?=1
where ? > 0 and ?(z) is a function of the following form ?(z) = |z|. From the Karush-KuhnTucker conditions, the unique closed-form solution is
exp(??(R(f? (x)))/?) ?? (x)
,
w? (x) = PM
0
0
?0 =1 exp(??(R(f? (x)))/?) ?? (x)
(4)
where ? can be interpreted as a ?temperature? parameter. This estimator satisfies oracle inequalities
of the following form [7]:
(M
)
X
b
E[R(f (x))] ?
min
w? (x)?(R(f? (x))) + ? DKL (w(x), ?(x)) .
(5)
w(x)?RM
?=1
The role of the distribution ? is to put a prior weight on the functions in the set. When there is no
preference, the uniform prior is a common choice but other choices are possible (see [7]).
In the proposed approach, we define the set of estimators as the set of patches taken in denoised
versions of the input image v. The next question is to develop a method to efficiently compute the
sum in (1) since the collection can be very large. For a typical image of N = 512 ? 512 pixels,
we could potentially consider M = L ? N pre-computed estimators if we apply L denoisers to the
input image v.
4
PEWA: Patch-based EWA estimator
Suppose that we are given a large collection of M competing estimators. These basis estimators
can be chosen arbitrarily among the researchers favorite denoising algorithm: Gaussian, Bilateral,
Wiener, Discrete Cosine Transform or other transform-based filterings. Let us emphasize here that
the number of basic estimators M is not expected to grow and is typically very large (M is chosen
on the order of several hundreds of thousands). In addition, the essential idea is that these basic
estimators only slightly improve the PSNR values of a few dBs.
Let us consider u` , ` = 1, ? ? ? , L denoised versions of v. A given pre-computed patch estimator
f? (x) is then a n-dimensional patch taken in the denoised image u` at any location y ? X , in the
spirit of the NL-means algorithm which considers only the noisy input patches for denoising. The
proposed estimator is then more general since a set of denoised patches at a given location are used.
Our estimator is then of the following form if we choose ?(z) = |z|:
fb(x) =
L
1 X X ?|R(u` (y))|/?
e
?` (y) u` (y),
Z(x)
`=1 y?X
Z(x) =
L X
X
0
e?|R(u`0 (y ))|/? ?`0 (y) (6)
`0 =1 y 0 ?X
where Z(x) is a normalization constant. Instead of considering a uniform prior over the set of
denoised patches taken in the whole image, it is appropriate to encourage patches located in the
3
neighborhood of x [27]. This can be achieved by introducing a spatial Gaussian prior G? (z) ?
2
2
e?z /(2? ) in the definition as
fbPEWA (x)
=
L
1 X X ?|R(u` (y))|/?
e
G? (x ? y) u` (y).
Z(x)
(7)
`=1 y?X
The Gaussian prior has a significant impact on the performance of the EWA estimator. Moreover, the
practical performance of the estimator strongly relies on an appropriate choice of ?. This important
question has been thoroughly discussed in [13] and ? = 4? 2 is motivated by the authors. Finally,
our patch-based EWA (PEWA) estimator can be written in terms of energies and Gibbs distributions
as:
fbPEWA (x)
=
L
1 X X ?E(u` (y))
e
u` (y),
Z(x)
Z(x) =
L X
X
0
e?E(u`0 (y )) ,
(8)
`0 =1 y 0 ?X
`=1 y?X
kx ? yk22
|kv(x) ? u` (y)k2n ? n? 2 |
+
.
2
4?
2? 2
The sums in (8) cannot be computed, especially when we consider a large collection of estimators.
In that sense, it differs from the NL-means methods [2, 11, 23, 9] which exploits patches generally
taken in a neighborhood of fixed size. Instead, we propose a Monte-Carlo sampling method to
approximately compute such an EWA when the number of aggregated estimators is large [1, 19].
E(u` (y))
4.1
=
Monte-Carlo simulations for computation
Because of the high dimensionality of the problem, we need efficient computational algorithms,
and therefore we suggest a stochastic approach to compute the PEWA estimator. Let us consider a random process (Fn (x))n?0 consisting in an initial noisy patch F0 (x) = v(x). The proposed Monte-Carlo procedure recommended to compute the estimator is based on the following
Metropolis-Hastings algorithm:
Draw a patch by considering a two-stage drawing procedure:
? draw uniformly a value ` in the set {1, 2, ? ? ? , L}.
? draw a pixel y = yc + ?, y ? X , with ? ? N (0, I2?2 ? 2 ) and yc is the position of the
current patch. At the initialization yc = x.
u` (y) if ? ? e??E(u` (y)),Fn (x))
Define Fn+1 (x) as: Fn+1 (x) =
(9)
Fn (x) otherwise
4
where ? is a random variable: ? ? U [0, 1] and ?E(u` (y), Fn (x)) = E(u` (y)) ? E(Fn (x)).
If we assume the Markov chain is ergodic, homogeneous, reductible, reversible and stationary, for
any F0 (x), we have almost surely
lim
T ?+?
T
X
1
Fn (x) ? fbPEWA (x)
T ? Tb
(10)
n=Tb
where T is the maximum number of samples of the Monte-Carlo procedure. It is also recommended
to introduce a burn-in phase to get a more satisfying estimator. Hence, the first Tb samples are
discarded in the average The Metropolis-Hastings rule allows reversibility and then stationarity of
the Markov chain. The chain is irreducible since it is possible to reach any patch in the set of possible
considered patches. The convergence is ensured when T tends to infinity. In practice, T is assumed
to be high to get a reasonable approximation of fbPEWA (x). In our implementation, we set T ? 1000
and Tb = 250 to produce fast and satisfying results. To improve convergence speed, we can use
several chains instead of only one [21].
In the Metropolis-Hastings dynamics, some patches are more frequently selected than others at a
given location. The number of occurrences of a particular candidate patch can be then evaluated. In
constant image areas, there is probably no preference for any one patch over any other and a low
number of candidate patches is expected along image contours and discontinuities.
4
4.2
Patch overlapping and iterations
The next step is to extend the PEWA procedure at every position of the entire image. To avoid
block effects at the patch boundaries, we overlap the patches. As a result, for the pixels lying
in the overlapping regions, we obtain multiple EWA estimates. These competing estimates must
be fused or aggregated into the single final estimate. The final aggregation can be performed by a
weighted average of the multiple EWA estimates as suggested in [21, 5, 22]. The simplest method of
aggregating such multiple estimates is to average them using equal weights. Such uniform averaging
provided the best results in our experiments and amounts to fusing n independent Markov chains.
The proposed implementation proceeds in two identical iterations. At the first iteration, the estimation is performed using several denoised versions of the noisy image. At the second iteration,
the first estimator is used as an additional denoised image in the procedure to improve locally the
estimation as in [6, 12]. The second iteration improves the PSNR values in the range of 0.2 to 0.5
dB as demonstrated by the experiments presented in the next section. Note that the first iteration is
able to produce very satisfying results for low and medium levels of noise. In practical imaging, we
use the method described in [11] to estimate the noise variance ? 2 for real-world noisy images.
5
Experimental results
We evaluated the PEWA algorithm on 25 natural images showing natural, man-made, indoor and
outdoor scenes (see Fig. 1). Each original image was corrupted with white Gaussian noise with zero
mean and variance ? 2 . In our experiments, the best results are obtained with n = 7 ? 7 patches
and L = 4 images ul denoised with DCT-based transform [26] ; we consider three different DCT
shrinkage thresholds: 1.25?, 1.5? and 1.75? to improve the PSNR of 1 to 6 db at most, depending
on ? and images (see Figs. 2-3). The fourth image is the noisy input image itself. We evaluated
the algorithm with a larger number L of denoised images and the quality drops by 0.1 db to 0.3 db,
which is visually imperceptible. Increasing L suggest also to considering more than 1000 samples
since the space of candidate patches is larger. The prior neighborhood size corresponds to a disk of
radius ? = 7 pixels but it can be smaller.
Performances of PEWA and other methods are quantified in terms of PSNR values for several noise
levels (see Tables 1-3). Table 1 reports the results obtained with PEWA on each individual image for
different values of standard deviation of noise. Table 2 compares the average PSNR values on these
25 images obtained by PEWA (after 1 and 2 iterations) and two state-of-the-art denoising methods
[6, 12]. We used the implementations provided by the authors: BM3D (http://www.cs.tut.fi/?foi/GCFBM3D/) and NL-Bayes (www.ipol.im). The best PSNR values are in bold and the results are quantitatively quite comparable except for very high levels of noise. We compared PEWA to the baseline
NL-means [2] and DCT [26] (using the implementation of www.ipol.im) since they form the core of
PEWA. The PSNR values increases of 1.5 db and 1.35 db on average over NL-means and DCT respectively. Finally, we compared the results to the recent S-PLE method which uses SURE to guide
the probabilistic patch-based filtering described in [24]. Figure 2 shows the denoising results on the
noisy Valdemossa (? = 15), Man (? = 20) and Castle (? = 25) images denoised with BM3D,
NL-Bayes and PEWA. Visual quality of methods is comparable.
Table 3 presents the denoising results with PEWA if the pre-computed estimators are obtained with
a Wiener filtering (spatial domain1 ) and DCT-based transform [26]. The results of PEWA with 5 ? 5
or 7 ? 7 patches are also given in Table 3, for one and two iterations. Note that NL-means can be
considered as a special case of the proposed method in which the original noisy patches constitute
the set of ?weak? estimators. The MCMC-based procedure can be then considered as an alternative
procedure to the usual implementation of NL-means to accelerate summation. Accordingly, in Table
3 we added a fair comparison (7?7 patches) with the implementation of NL-means algorithm (IPOL
(ipol.im)) which restricts the search of similar patches in a neighborhood of 21 ? 21 pixels. In these
experiments, ?PEWA basic? (1 iteration) produced better results especially for ? ? 10. Finally we
compared these results with the most popular and competitive methods on the same images. The
PSNR values are selected from publications cited in the literature. LSSC and BM3D are the most
var(v(x)) ? a` ? 2
u` (x) = mean(v(x)) + max 0,
? (v(x) ? mean(v(x))), where ` = {1, 2, 3} and
var(v(x))
a1 = 0.15, a2 = 0.20, a3 = 0.25.
1
5
cameraman
(256 ? 256)
peppers
(256 ? 256)
house
(256 ? 256)
Lena
(512 ? 512)
maya
(313 ? 473)
asia
(313 ? 473)
aircraft
(473 ? 313)
panther
(473 ? 313)
castle
(313 ? 473)
young man
(313 ? 473)
tiger
(473 ? 313)
man on wall picture
(473 ? 313)
barbara
(512 ? 512)
boat
(512 ? 512)
man
(512 ? 512)
couple
(512 ? 512)
hill
(512 ? 512)
alley
(192 ? 128)
computer
(704 ? 469)
dice
(704 ? 469)
flowers
(704 ? 469)
girl
(704 ? 469)
traffic
(704 ? 469)
trees
(192 ? 128)
valldemossa
(769 ? 338)
?
Figure 1: Set of 25 tested images. Top left: images from the BM3D website (cs.tut.fi/foi/GCFBM3D/) ; Bottom left: images from IPOL (ipol.im); Right: images from the Berkeley segmentation
database (eecs.berkeley.edu/Research/Projects/CS/ vision/bsds/).
performant but PEWA is able to produce better results on several piecewise smooth images while
BM3D is more appropriate for textured images.
In terms of computational complexity, denoising a 512 ? 512 grayscale image with an unoptimized
implementation of our method in C++ take about 2 mins (Intel Core i7 64-bit CPU 2.4 Ghz). Recently, PEWA has been implemented in parallel since every patch can be processed independently
and the computational times become a few seconds.
6
Conclusion
We presented a new general two-step denoising algorithm based on non-local image statistics and
patch repetition, that combines ideas from the popular NL-means [6] and BM3D algorithms [6] and
theoretical results from the statistical literature on Exponentially Weighted Aggregation [7, 21]. The
first step of PEWA involves the computation of denoised images obtained with a separate collection of multiple denoisers (Wiener, DCT... ) applied to the input image. In the second step, the
set of denoised image patches are selectively exploited to compute an aggregated estimator. We
showed that the estimator can be computed in reasonable time using a Monte-Carlo Markov Chain
(MCMC) sampling procedure. If we consider DCT-based transform [6] in the first step, the results
are comparable in average to the state-of-the-art results. The PEWA method generalizes the NLmeans algorithm in some sense but share also common features with BM3D (e.g. DCT transform,
two-stage collaborative filtering). tches, contrary to NL-Bayes and BM3D. For future work, waveletbased transform, multiple image patch sizes, robust statistics and sparse priors will be investigated
to improve the results of the flexible PEWA method.
6
noisy (PSNR = 24.61)
PEWA (PSNR = 29.25)
BM3D [6] (PSNR = 29.19)
NL-Bayes [12] (PSNR = 29.22)
Figure 2: Comparison of algorithms. Valldemossa image corrupted with white Gaussian noise (? =
15). The PSNR values of the three images denoised with DCT-based transform [26] are combined
with PEWA are 27.78, 27.04 and 26.26.)
noisy
(PSNR = 20.18)
PEWA
(PSNR = 29.49)
BM3D [6]
(PSNR = 29.36)
NL-Bayes [12]
(PSNR = 29.48)
noisy
(PSNR = 22.11)
PEWA
(PSNR = 30.50)
BM3D [6]
(PSNR = 30.59)
NL-Bayes [12]
(PSNR = 30.60)
Figure 3: Comparison of algorithms. First row: Castle image corrupted with white Gaussian noise
(? = 25). The PSNR values of the three images denoised with DCT-based transform [26] and
combined with PEWA are 25.77, 24.26 and 22.85. Second row: Man image corrupted with white
Gaussian noise (? = 20). The PSNR values of the three images denoised with DCT-based transform
[26] and combined with PEWA are 27.42, 26.00 and 24.67.
7
Cameraman
Peppers
House
Lena
Barbara
Boat
Man
Couple
Hill
Alley
Computer
Dice
Flowers
Girl
Traffic
Trees
Valldemossa
Aircraft
Asia
Castle
Man Picture
Maya
Panther
Tiger
Young man
Average
?=5
38.20
38.00
39.56
38.57
38.09
37.12
37.68
37.35
37.01
36.29
39.04
46.82
43.48
43.95
37.85
34.88
36.65
37.59
38.67
38.06
37.78
34.72
38.53
36.92
40.79
38.54
? = 10
34.23
34.68
36.40
35.78
34.73
33.75
33.93
33.91
33.52
32.20
35.13
43.87
39.67
41.22
33.54
29.93
31.79
34.62
34.46
34.13
33.58
29.64
33.91
32.85
37.36
34.75
? = 15
31.98
32.75
34.86
34.12
32.86
31.94
31.93
31.98
31.69
29.98
32.81
42.05
37.47
39.52
31.13
27.49
29.25
33.00
32.25
32.02
31.27
27.17
31.56
30.63
35.58
32.67
? = 20
30.60
31.40
33.72
32.90
31.43
30.64
30.50
30.57
30.50
28.54
31.23
40.58
35.90
38.27
29.58
25.86
27.59
31.75
30.73
30.56
29.73
25.42
30.02
29.13
34.30
31.26
? = 25
29.48
30.30
32.77
31.89
30.28
29.65
29.50
29.48
29.56
27.46
30.01
39.36
34.55
37.33
28.48
24.69
26.37
30.72
29.60
29.49
28.44
24.28
28.83
27.99
33.25
30.15
? = 50
26.25
26.69
29.29
28.83
26.58
26.64
26.67
26.02
26.92
24.13
26.38
35.33
30.81
34.14
25.50
21.78
23.18
27.68
26.63
26.15
24.65
22.85
25.59
24.63
29.59
26.95
? = 100
22.81
22.84
25.35
25.65
22.95
23.63
24.15
23.27
24.49
21.37
23.27
30.82
27.53
30.50
22.90
20.03
20.71
24.99
24.32
23.09
21.50
18.17
22.75
21.90
25.20
23.76
Table 1: Denoising results on the 25 tested images for several values of ?. The PSNR values are
averaged over 3 experiments corresponding to 3 different noise realizations.
PEWA 1
PEWA 2
BM3D [6]
NL-Bayes [12]
S-PLE [24]
NL-means [2]
DCT [26]
?=5
38.27
38.54
38.64
38.60
38.17
37.44
37.81
? = 10
34.39
34.75
34.78
34.75
34.38
33.35
33.57
? = 15
32.26
32.67
32.68
32.48
32.35
31.00
31.87
? = 20
30.76
31.26
31.25
31.22
30.67
30.16
29.95
? = 25
29.62
30.15
30.19
30.12
29.77
28.96
28.97
? = 50
26.00
26.95
26.97
26.90
26.46
25.53
25.91
? = 100
22.35
23.76
24.08
23.65
23.21
22.29
23.08
Table 2: Average of denoising results over the 25 tested images for several values of ?. The experiments with NL-Bayes [12], S-PLE[24], NL-means [2] and DCT [26] have been performed using the
using the implementation of IPOL (ipol.im). The best PSNR values are in bold.
Image
Peppers
House
Lena
Barbara
(256 ? 256)
(256 ? 256)
(512 ? 512)
(512 ? 512)
?
5.00 15.00 25.00 50.00
5.00 15.00 25.00 50.00
5.00 15.00 25.00 50.00
5.00 15.00 25.00 50.00
PEWA 1 (W) (5?5)
PEWA 2 (W) (5?5)
PEWA 1 (W) (7 ?7)
PEWA 2 (W) (7 ?7)
PEWA 1 (D) (5 ?5)
PEWA 2 (D) (5 ?5)
PEWA 1 (D) (7 ?7)
PEWA 2 (D) (7 ?7)
PEWA Basic (7?7)
NL-means [2] (7?7)
BM3D [6]
NL-Bayes [12]
ND-SAFIR [11]
K-SVD [10]
LSSC [16]
PLOW [5]
SOP [18]
36.69 30.58 27.50 22.85
37.45 32.20 29.72 26.09
36.72 30.60 27.60 22.82
37.34 32.34 30.11 26.53
37.70 32.45 29.83 26.01
37.95 32.80 30.20 26.66
37.71 32.43 29.87 26.00
38.00 32.75 30.30 26.69
36.88 31.34 29.47 26.02
36.77 30.93 28.76 24.24
38.12 32.70 30.16 26.68
38.09 32.26 29.79 26.10
37.34 32.13 29.73 25.29
37.80 32.23 29.81 26.24
38.18 32.82 30.21 26.62
37.69 31.82 29.53 26.32
37.63 32.40 30.01 26.75
37.89 31.88 28.55 23.49
38.98 34.27 32.13 28.35
37.90 31.90 28.59 23.52
39.00 34.57 32.51 29.04
39.28 34.23 31.79 27.72
39.46 34.74 31.67 29.15
39.27 34.26 31.79 27.71
39.56 34.83 32.77 29.29
37.88 34.13 32.14 28.25
37.75 32.36 31.11 27.54
39.83 34.94 32.86 29.69
39.39 33.77 31.36 27.62
37.62 34.08 32.22 28.67
39.33 34.19 31.97 28.01
39.93 35.35 33.15 30.04
39.52 34.72 32.70 29.08
38.76 34.35 32.54 29.64
37.27 31.43 28.30 23.45
38.05 33.40 31.11 27.80
37.26 31.45 28.33 23.45
38.00 33.65 31.56 28.40
38.46 33.72 31.33 27.59
38.57 33.96 31.81 28.43
38.45 33.72 31.25 27.62
38.58 34.12 31.89 28.83
37.39 33.26 31.20 27.92
36.65 32.00 30.45 27.32
38.72 34.27 32.08 29.05
38.75 33.51 31.16 27.62
37.91 33.70 31.73 28.38
38.63 33.76 31.35 27.85
38.69 34.15 31.87 28.87
38.66 33.90 31.92 28.32
38.31 33.84 31.80 28.96
36.39 30.18 29.31 22.71
37.13 31.94 29.47 25.58
36.40 30.18 27.32 22.71
37.00 32.10 30.00 26.20
37.71 32.20 29.55 25.58
38.03 32.70 30.03 26.01
37.70 32.30 29.84 26.20
38.09 32.86 30.28 26.58
36.80 31.89 29.76 25.83
36.79 30.65 28.99 25.63
38.31 33.11 30.72 27.23
38.38 32.47 30.02 26.45
37.12 31.80 29.24 24.09
38.08 32.33 29.54 25.43
38.48 33.00 30.47 27.06
37.98 21.17 30.20 26.29
37.74 32.65 30.37 27.35
Table 3: Comparison of several versions of PEWA (W (Wiener), D (DCT), Basic) and competitive
methods on a few standard images corrupted with white Gaussian noise. The best PSNR values are
in bold (PSNR values from publications cited in the literature).
8
References
[1] Alquier, P., Lounici, K. (2011) PAC-Bayesian bounds for sparse regression estimation with exponential
weights. Electronic Journal of Statistics 5:127-145.
[2] Buades A., Coll, B. & Morel, J.-M. (2005) A review of image denoising algorithms, with a new one. SIAM
J. Multiscale Modeling & Simulation, 4(2):490-530.
[3] Burger, H., Schuler, C. & Harmeling, S. (2012) Image denoising: can plain neural networks compete with
BM3D ? In IEEE Conf. Comp. Vis. Patt. Recogn. (CVPR?12), pp. 2392-2399, Providence, Rhodes Island.
[4] Chatterjee, P. & Milanfar, P. (2010) Is denoising dead?, IEEE Transactions on Image Processing,
19(4):895-911.
[5] Chatterjee, P. & Milanfar, P. (2012) Patch-based near-optimal image denoising, IEEE Transactions on
Image Processing, 21(4):1635-1649.
[6] Dabov, K., Foi, A., Katkovnik, V. & Egiazarian, K. (2007) Image denoising by sparse 3D transform-domain
collaborative filtering, IEEE Transactions on Image Processing, 16(8):2080-2095.
[7] Dalayan, A.S. & Tsybakov, A. B. (2008) Aggregation by exponential weighting, sharp PAC-Bayesian
bounds and sparsity. Machine Learning 72:39-61.
[8] Dalayan, A.S. & Tsybakov, A. B. (2009) Sparse regression learning by aggregation and Langevin Monte
Carlo. Available at ArXiv:0903.1223
[9] Deledalle, C.-A., Duval, V., Salmon, J. (2012) Non-local methods with shape-adaptive patches (NLMSAP). J. Mathematical Imaging and Vision, 43:103-120.
[10] Elad, M. & Aharon, M. (2006) Image denoising via sparse and redundant representations over learned
dictionaries. IEEE Transactions on Image Processing, 15(12):3736-3745.
[11] Kervrann, C. & Boulanger, J. (2006) Optimal spatial adaptation for patch-based image denoising. IEEE
Transcations on Image Processing, 15(10):2866-2878.
[12] Lebrun, M., Buades, A. & Morel, J.-M. (2013) Implementation of the ?Non-Local Bayes? (NL-Bayes)
image denoising algorithm, Image Processing On Line, 3:1-42. http://dx.doi.org/10.5201/ipol.2013.16
[13] Leung, G. & Barron, A. R. (2006) Information theory and mixing least-squares regressions. IEEE Transactions on Information Theory 52:3396-3410.
[14] Levin, A., Nadler, B., Durand, F. & Freeman, W.T. (2012) Patch complexity, finite pixel correlations and
optimal denoising. Europ. Conf. Comp. Vis. (ECCV?12), pp. 73-86, Firenze, Italy.
[15] Louchet, C. & Moisan, L. (2011) Total Variation as a local filter. SIAM J. Imaging Sciences 4(2):651-694.
[16] Mairal, J., Bach, F., Ponce, J., Sapiro, G. &. Zisserman, A. (2009) Non-local sparse models for image
restoration. In IEEE Int. Conf. Comp. Vis. (ICCV?09), pp. 2272-2279, Tokyo, Japan.
[17] Milanfar, P. (2013) A Tour of modern image filtering. IEEE Signal Processing Magazine, 30(1):106-128.
[18] Ram, I., Elad, M. & Cohen, I. (2013) Image processing using smooth ordering of its patches. IEEE
Transactions on Image Processing, 22(7):2764?2774
[19] Rigollet, P. & Tsybakov, A. B. (2012) Sparse estimation by exponential weighting. Statistical Science
27(4):558-575.
[20] Roth, S. & Black, M.J. (2005) Fields of experts: A framework for learning image priors. In IEEE Conf.
Comp. Vis. Patt. Recogn. (CVPR?05), vol. 2, pp. 860-867, San Diego, CA.
[21] Salmon, J. & Le Pennec, E. (2009) NL-Means and aggregation procedures. In IEEE Int. Conf. Image
Process. (ICIP?09), pp. 2977-2980, Cairo, Egypt.
[22] Talebi, H., Xhu, X. & Milanfar, P. (2013) How to SAIF-ly boost denoising performance. In IEEE Transactions on Image Processing 22(4):1470-1485.
[23] Van De Ville, D. & Kocher, M. (2009) SURE based non-local means. IEEE Signal Processing Letters,
16(11):973-976, 2009.
[24] Wang, Y.-Q. & Morel, J.-M. (2013). SURE guided Gaussian mixture image denoising. SIAM J. Imaging
Sciences, 6(2):999-1034.
[25] Xue, F., Luisier, F. & Blu T. (2013) Multi-wiener SURE-LET deconvolution. IEEE Transactions on Image
Processing, 22(5):1954-1968.
[26] Yu G. & Sapiro G. (2011) . DCT image denoising: a simple and effective image denoising algorithm.
Image Processing On Line (http://dx.doi.org/10.5201/ipol.2011.ys-dct).
[27] Zontak, M. & Irani, M. (2011) Internal statistics of a single natural image. In IEEE Comp. Vis. Patt.
Recogn. (CVPR?11), pp. 977-984, Colorado Springs, CO.
[28] Zoran, D. & Weiss, Y. (2011) From learning models of natural image patches to whole image restoration.
In IEEE Int. Conf. Comp. Vis. (ICCV?11), pp. 479-486, Barcelona, Spain.
9
| 5410 |@word aircraft:2 version:7 blu:1 nd:1 disk:1 simulation:2 covariance:1 reduction:1 initial:1 mmse:1 current:4 dx:2 written:1 must:1 fn:8 dct:17 additive:1 shape:1 drop:1 stationary:1 selected:2 website:1 accordingly:1 core:2 awg:1 boosting:1 location:4 preference:2 org:2 mathematical:2 along:1 become:1 consists:1 combine:5 introduce:1 expected:3 examine:1 frequently:1 multi:1 bm3d:18 lena:3 inspired:1 freeman:1 cpu:1 considering:3 increasing:1 project:2 estimating:1 campus:1 moreover:1 provided:2 medium:1 buades:2 burger:1 spain:1 plow:2 interpreted:2 sapiro:2 berkeley:2 every:2 ensured:1 rm:2 ly:1 aggregating:1 local:6 tends:1 limit:1 approximately:1 inria:2 burn:1 black:1 initialization:1 studied:2 quantified:1 co:1 range:1 averaged:1 practical:3 unique:1 harmeling:1 practice:1 block:1 differs:1 firenze:1 procedure:15 area:1 dice:2 thought:1 matching:1 pre:4 suggest:2 get:3 cannot:1 close:1 put:1 risk:4 www:3 measurable:1 map:1 demonstrated:1 roth:1 maximizing:1 independently:2 ergodic:1 filterings:1 estimator:39 rule:1 handle:1 variation:1 diego:1 suppose:1 colorado:1 magazine:1 homogeneous:1 epll:4 us:1 associate:1 satisfying:3 located:1 database:3 observed:1 role:1 domain1:1 bottom:1 wang:1 thousand:2 region:1 ordering:1 complexity:2 dynamic:1 trained:1 weakly:2 zoran:1 serve:2 basis:1 textured:1 girl:2 accelerate:1 panther:2 recogn:3 fast:1 effective:1 monte:7 universitaire:1 doi:2 aggregate:1 neighborhood:4 quite:2 widely:2 solve:1 larger:2 cvpr:3 drawing:1 otherwise:1 elad:2 statistic:5 transform:12 noisy:19 itself:1 final:2 frozen:1 propose:3 fr:1 adaptation:1 combining:1 realization:1 mixing:1 achieve:3 kv:2 exploiting:1 convergence:2 produce:5 depending:1 develop:1 lssc:4 implemented:2 c:3 involves:1 europ:1 guided:1 radius:1 tokyo:1 filter:1 stochastic:1 enable:1 clustered:1 karush:1 wall:1 summation:1 im:5 lying:1 considered:4 visually:2 exp:2 bsds:1 mapping:1 nadler:1 achieves:3 dictionary:3 a2:1 estimation:6 rhodes:1 currently:1 repetition:2 weighted:8 morel:3 gaussian:10 always:1 rather:1 avoid:1 shrinkage:2 boosted:1 publication:2 focus:1 ponce:1 denoisers:3 likelihood:1 baseline:1 sense:4 posteriori:1 dependent:1 leung:1 typically:1 entire:1 waveletbased:1 unoptimized:1 france:1 interested:1 pixel:6 arg:1 aforementioned:1 flexible:4 among:1 art:8 special:2 spatial:4 field:2 equal:1 reversibility:1 sampling:4 identical:1 represents:1 yu:1 future:1 others:1 report:1 quantitatively:1 piecewise:1 few:3 irreducible:1 alley:2 modern:1 composed:1 divergence:1 individual:1 replaced:1 phase:1 consisting:1 stationarity:1 mlp:2 mixture:2 nl:27 tut:2 chain:7 encourage:1 necessary:1 tree:2 bretagne:1 theoretical:2 modeling:2 restoration:2 introducing:1 fusing:1 deviation:1 tour:1 hundred:2 uniform:3 levin:1 providence:1 corrupted:6 eec:1 xue:1 considerably:1 combined:5 thoroughly:1 cited:2 peak:1 siam:3 probabilistic:1 fused:1 talebi:1 w1:1 squared:1 choose:1 worse:1 ocal:1 external:5 castle:4 conf:6 dead:1 expert:1 japan:1 de:2 bold:3 int:3 vi:6 performed:4 bilateral:1 closed:1 traffic:2 competitive:4 aggregation:13 denoised:20 bayes:13 wm:1 parallel:1 contribution:1 collaborative:2 square:2 accuracy:1 wiener:5 variance:2 egiazarian:1 efficiently:3 weak:3 bayesian:4 foi:3 produced:1 carlo:7 dabov:1 comp:6 researcher:1 foe:1 reach:1 definition:1 energy:2 frequency:1 involved:1 pp:7 couple:2 dataset:1 popular:2 lim:1 dimensionality:1 psnr:28 improves:1 segmentation:1 asia:2 zisserman:1 improved:1 wei:1 evaluated:3 lounici:1 strongly:1 stage:2 correlation:1 hastings:3 multiscale:1 reversible:1 overlapping:2 quality:2 effect:1 alquier:1 requiring:1 unbiased:2 hence:1 irani:1 leibler:1 i2:1 white:6 attractive:1 cosine:1 criterion:1 hill:2 dedicated:2 temperature:1 egypt:1 image:95 salmon:3 recently:2 charles:2 fi:2 common:2 rigollet:1 empirically:1 cohen:1 exponentially:2 anisotropic:1 discussed:2 interpretation:1 extend:1 significant:1 gibbs:3 pm:1 f0:2 similarity:1 recent:3 showed:1 italy:1 barbara:3 pennec:2 inequality:1 durand:1 arbitrarily:1 exploited:1 minimum:1 additional:1 surely:1 aggregated:4 redundant:1 recommended:2 signal:4 multiple:7 smooth:2 imperceptible:1 bach:1 y:1 dkl:3 a1:1 impact:1 basic:6 regression:3 vision:2 expectation:1 arxiv:1 iteration:10 kernel:1 represent:1 normalization:1 achieved:2 ewa:9 addition:1 grow:1 rennes:2 unlike:1 sure:7 cedex:1 subject:1 probably:1 db:7 contrary:1 spirit:1 near:1 yk22:1 enough:2 pepper:3 competing:2 idea:2 i7:1 motivated:1 ul:1 milanfar:4 constitute:1 generally:2 k2n:3 amount:1 stein:2 tsybakov:3 locally:2 processed:2 category:2 simplest:1 http:3 restricts:1 estimated:1 patt:3 discrete:1 shall:1 vol:1 nevertheless:1 threshold:1 falling:1 clean:5 diffusion:1 imaging:4 ram:1 luisier:1 ville:1 year:2 sum:2 compete:1 letter:1 fourth:1 family:1 almost:1 reasonable:2 electronic:1 patch:72 draw:3 comparable:6 bit:1 layer:1 bound:2 maya:2 duval:1 oracle:1 infinity:1 scene:1 speed:1 min:3 spring:1 combination:2 poor:1 remain:1 slightly:1 smaller:1 island:1 metropolis:3 iccv:2 taken:4 computationally:1 previously:1 cameraman:2 precomputed:1 generalizes:1 operation:1 gaussians:1 available:1 aharon:1 apply:1 barron:1 appropriate:3 occurrence:1 alternative:1 original:2 denotes:1 clustering:2 top:1 exploit:1 especially:3 boulanger:1 objective:1 already:2 question:2 added:1 usual:1 separate:2 considers:1 performant:1 ratio:2 mostly:1 potentially:1 sop:1 implementation:9 perform:1 observation:2 markov:6 discarded:1 finite:1 langevin:1 extended:1 team:1 rn:2 sharp:1 pair:1 required:1 icip:1 learned:2 boost:1 barcelona:1 discontinuity:1 able:5 suggested:2 proceeds:1 flower:2 yc:3 indoor:1 sparsity:1 tb:4 including:1 max:1 overlap:1 demanding:1 natural:5 transcations:1 boat:2 improve:5 picture:2 prior:11 literature:4 l2:1 review:1 kf:1 filtering:5 var:2 principle:1 share:1 row:2 eccv:1 surprisingly:1 free:2 guide:1 katkovnik:1 perceptron:1 kuhntucker:1 sparse:8 ghz:1 van:1 boundary:1 plain:1 world:1 contour:1 computes:1 fb:7 author:4 commonly:1 adaptive:2 collection:5 made:1 coll:1 ple:4 far:1 san:1 transaction:8 approximate:1 emphasize:1 kullback:1 mairal:1 assumed:4 grayscale:1 search:1 table:9 favorite:1 learn:1 schuler:1 robust:1 ca:1 investigated:3 domain:1 motivation:1 noise:18 whole:2 denoise:1 fair:1 fig:2 intel:1 position:2 exponential:8 candidate:6 outdoor:1 house:3 weighting:2 young:2 beaulieu:1 showing:1 pac:2 r2:1 a3:1 deconvolution:2 essential:1 chatterjee:2 kx:1 atlantique:1 visual:1 corresponds:1 satisfies:1 relies:1 extracted:1 viewed:1 man:9 feasible:1 tiger:2 typical:1 except:1 uniformly:1 averaging:1 denoising:31 called:2 total:1 experimental:2 svd:1 select:1 formally:1 selectively:1 internal:4 evaluate:1 mcmc:4 tested:3 |
4,873 | 5,411 | A Multi-World Approach to Question Answering
about Real-World Scenes based on Uncertain Input
Mateusz Malinowski
Mario Fritz
Max Planck Institute for Informatics
Saarbr?ucken, Germany
{mmalinow,mfritz}@mpi-inf.mpg.de
Abstract
We propose a method for automatically answering questions about images by
bringing together recent advances from natural language processing and computer
vision. We combine discrete reasoning with uncertain predictions by a multiworld approach that represents uncertainty about the perceived world in a bayesian
framework. Our approach can handle human questions of high complexity about
realistic scenes and replies with range of answer like counts, object classes, instances and lists of them. The system is directly trained from question-answer
pairs. We establish a first benchmark for this task that can be seen as a modern
attempt at a visual turing test.
1
Introduction
As vision techniques like segmentation and object recognition begin to mature, there has been an
increasing interest in broadening the scope of research to full scene understanding. But what is
meant by ?understanding? of a scene and how do we measure the degree of ?understanding?? Most
often ?understanding? refers to a correct labeling of pixels, regions or bounding boxes in terms of
semantic annotations. All predictions made by such methods inevitably come with uncertainties
attached due to limitations in features or data or even inherent ambiguity of the visual input.
Equally strong progress has been made on the language side, where methods have been proposed
that can learn to answer questions solely from question-answer pairs [1]. These methods operate on
a set of facts given to the system, which is refered to as a world. Based on that knowledge the answer
is inferred by marginalizing over multiple interpretations of the question. However, the correctness
of the facts is a core assumption.
We like to unite those two research directions by addressing a question answering task based on realworld images. To combine the probabilistic output of state-of-the-art scene segmentation algorithms,
we propose a Bayesian formulation that marginalizes over multiple possible worlds that correspond
to different interpretations of the scene.
To date, we are lacking a substantial dataset that serves as a benchmark for question answering on
real-world images. Such a test has high demands on ?understanding? the visual input and tests a
whole chain of perception, language understanding and deduction. This very much relates to the
?AI-dream? of building a turing test for vision. While we are still not ready to test our vision system
on completely unconstrained settings that were envisioned in early days of AI, we argue that a
question-answering task on complex indoor scenes is a timely step in this direction.
Contributions: In this paper we combine automatic, semantic segmentations of real-world scenes
with symbolic reasoning about questions in a Bayesian framework by proposing a multi-world
approach for automatic question answering. We introduce a novel dataset of more than 12,000
1
question-answer pairs on RGBD images produced by humans, as a modern approach to a visual
turing test. We benchmark our approach on this new challenge and show the advantages of our
multi-world approach. Furthermore, we provide additional insights regarding the challenges that lie
ahead of us by factoring out sources of error from different components.
2
Related work
Semantic parsers: Our work is mainly inspired by [1] that learns the semantic representation for
the question answering task solely based on questions and answers in natural language. Although
the architecture learns the mapping from weak supervision, it achieves comparable results to the
semantic parsers that rely on manual annotations of logical forms ([2], [3]). In contrast to our work,
[1] has never used the semantic parser to connect the natural language to the perceived world.
Language and perception: Previous work [4, 5] has proposed models for the language grounding
problem with the goal of connecting the meaning of the natural language sentences to a perceived
world. Both methods use images as the representation of the physical world, but concentrate rather
on constrained domain with images consisting of very few objects. For instance [5] considers only
two mugs, monitor and table in their dataset, whereas [4] examines objects such as blocks, plastic
food, and building bricks. In contrast, our work focuses on a diverse collection of real-world indoor
RGBD images [6] - with many more objects in the scene and more complex spatial relationship
between them. Moreover, our paper considers complex questions - beyond the scope of [4] and [5]
- and reasoning across different images using only textual question-answer pairs for training. This
imposes additional challenges for the question-answering engines such as scalability of the semantic
parser, good scene representation, dealing with uncertainty in the language and perception, efficient
inference and spatial reasoning. Although others [7, 8] propose interesting alternatives for learning
the language binding, it is unclear if such approaches can be used to provide answers on questions.
Integrated systems that execute commands: Others [9, 10, 11, 12, 13] focus on the task of learning the representation of natural language in the restricted setting of executing commands. In such
scenario, the integrated systems execute commands given natural language input with the goal of using them in navigation. In our work, we aim for less restrictive scenario with the question-answering
system in the mind. For instance, the user may ask our architecture about counting and colors (?How
many green tables are in the image??), negations (?Which images do not have tables??) and superlatives (?What is the largest object in the image??).
Probabilistic databases: Similarly to [14] that reduces Named Entity Recognition problem into the
inference problem from probabilistic database, we sample multiple-worlds based on the uncertainty
introduced by the semantic segmentation algorithm that we apply to the visual input.
3
Method
Our method answers on questions based on images by combining natural language input with output
from visual scene analysis in a probabilistic framework as illustrated in Figure 1. In the single world
approach, we generate a single perceived world W based on segmentations - a unique interpretation
of a visual scene. In contrast, our multi-world approach integrates over many latent worlds W, and
hence taking different interpretations of the scene and question into account.
Single-world approach for question answering problem We build on recent progress on end-toend question answering systems that are solely trained on question-answer pairs (Q, A) [1]. Top part
of Figure 1 outlines how we build on [1] by modeling the logical forms associated with a question as
latent variable T given a single world W. More formally the task of predicting an answer A given
a question Q and a world W is performed by computing the following posterior which marginalizes
over the latent logical forms (semantic trees in [1]) T :
X
P (A|Q, W) :=
P (A|T , W)P (T |Q).
(1)
T
P (A|T , W) corresponds to denotation of a logical form T on the world W. In this setting,
the answer is unique given the logical form and the world: P (A|T , W) = 1[A ? ?W (T )]
with the evaluation function ?W , which evaluates a logical form on the world W. Following
[1] we use DCS Trees that yield the following recursive evaluation function ?W : ?W (T ) :=
2
single?
world
approach
Semantic
parsing
Q
Semantic
evaluation
T
A
logical
form
question
S
W
world
answer
sofa (1,brown, image 1, X,Y,Z)
table (1,brown, image 1,X,Y,Z)
wall (1,white, image 1, X,Y,Z)
Scene
analysis
bed (1, white, image 2 X,Y,Z)
chair (1,brown, image 4, X,Y,Z)
chair (2,brown, image 4, X,Y,Z)
chair (1,brown, image 5, X,Y,Z)
?
multi-world
approach
Semantic
parsing
Q
Semantic
evaluation
T
A
logical
form
question
W
S
latent
worlds
answer
segmentation
Figure 1: Overview of our approach to question answering with multiple latent worlds in contrast to single
world approach.
Td
j {v : v ? ?W (p), t ? ?W (Tj ), Rj (v, t)} where T := hp, (T1 , R1 ), (T2 , R2 ), ..., (Td , Rd )i is
the semantic tree with a predicate p associated with the current node, its subtrees T1 , T2 , ..., Td , and
relations Rj that define the relationship between the current node and a subtree Tj .
In the predictions, we use a log-linear distribution P (T |Q) ? exp(?T ?(Q, T )) over the logical
forms with a feature vector ? measuring compatibility between Q and T and parameters ? learnt
from training data. Every component ?j is the number of times that a specific feature template
occurs in (Q, T ). We use the same templates as [1]: string triggers a predicate, string is under a
relation, string is under a trace predicate, two predicates are linked via relation and a predicate has
a child. The model learns by alternating between searching over a restricted space of valid trees
and gradient descent updates of the model parameters ?. We use the Datalog inference engine to
produce the answers from the latent logical forms. The linguistic phenomena such as superlatives
and negations are handled by the logical forms and the inference engine. For a detailed exposition,
we refer the reader to [1].
Question answering on real-world images based on a perceived world Similar to [5], we
extend the work of [1] to operate now on what we call perceived world W. This still corresponds to the single world approach in our overview Figure 1. However our world is now populated with ?facts? derived from automatic, semantic image segmentations S. For this purpose, we
build the world by running a state-of-the-art semantic segmentation algorithm [15] over the images and collect the recognized information about objects such as object class, 3D position, and
50
5050
50
50
100
50
100
100
100
150
100
150
150
100
150
150
200
200
150
200
200
200
250
250
250
200
250
250
300
300
300
250
300
350
300
350
350
350
300
400
350
400
400
400
350
400
400
50
50
550
5050 100
100
150
200
250
300
350
400
450
500
550
100150
150200
200250
250300
300350
350400
400450
450500
500
550
100
150
150
150150
150
100
100
200
200
200200
200
150
150
250
250
300
300
5050 50
100
100
100
50 50
100100
100
50 50
100100
100
5050
150
150
150
100
100
200
200
200
150
150
250
250
250
200
200
300
300
300
250
250
350
350
350
300
300
400
400
400
350
350
150150
150
100
100
200200
200
150
150
250250
250
200
200
300300
300
250
250
250250
250
200
200
300300
300
250
250
350350
350
300
350350
350
300
300
400400
400
350
5050 100
100
150
200200250
250250300
300300350
350350400
400400450
450450500
500500
550550
350
50
550
100150
150200
400400
400
350
400
300
400
50
400
400
50
100 150 200 250 300 350 400 450 500 550
100 150 200 250 300 350 400 450 500 550
300 350
350
300
400
400
450
450
500
500
350
50
100 150 200 250 300 350
50
100 150 200 250 300 350
(a) Sampled worlds.
50
50
100 150 200 250 300 350 400 450 500 550
100 150 200 250 300 350 400 450 500 550
400
400
450
450
500
500
550
550
100
150
200
250
300
350
400
450
500
550
50
100
150
200
250
300
350
400
450
500
550
350
400
350
100100
150150
200200
250250
300300
100
150
200
250
300
350350350
400400400
450450450
500500500
550550550
100100
150150
200200
250250
300300
350350
400400
450450
500500
5505505050 50
5050 50100
150
200
250
300
350
400
450
500
550
400
400
50
50
100
550
550
100 150
150 200
200 250
250
5050 100
5050 50
5050
50
400
50
100
150
200
250
300
350
400
450
500
550
(b) Object?s coordinates.
Figure 2: Fig. 2a shows a few sampled worlds where only segments of the class ?person? are shown. In the
clock-wise order: original picture, most confident world, and three possible worlds (gray-scale values denote
the class confidence). Although, at first glance the most confident world seems to be a reasonable approach,
our experiments show opposite - we can benefit from imperfect but multiple worlds. Fig. 2b shows object?s
coordinates (original and Z, Y , X images in the clock-wise order), which better represent the spatial location
of the objects than the image coordinates.
3
auxiliary relations
spatial
Predicate
closeAbove(A, B)
closeLef tOf (A, B)
closeInF rontOf (A, B)
Xaux (A, B)
Zaux (A, B)
haux (A, B)
vaux (A, B)
daux (A, B)
lef tOf (A, B)
above(A, B)
inF rontOf (A, B)
on(A, B)
close(A, B)
Definition
above(A, B) and (Ymin (B) < Ymax (A) + )
lef tOf (A, B) and (Xmin (B) < Xmax (A) + )
inF rontOf (A, B) and (Zmin (B) < Zmax (A) + )
Xmean (A) < Xmax (B) and Xmin (B) < Xmean (A)
Zmean (A) < Zmax (B) and Zmin (B) < Zmean (A)
closeAbove(A, B) or closeBelow(A, B)
closeLef tOf (A, B) or closeRightOf (A, B)
closeInF rontOf (A, B) or closeBehind(A, B)
Xmean (A) < Xmean (B))
Ymean (A) < Ymean (B)
Zmean (A) < Zmean (B))
closeAbove(A, B) and Zaux (A, B) and Xaux (A, B)
haux (A, B) or vaux (A, B) or daux (A, B)
Table 1: Predicates defining spatial relations between A and B. Auxiliary relations define actual spatial relations. The Y axis points downwards, functions Xmax , Xmin , ... take appropriate values from the tuple
predicate, and is a ?small? amount. Symmetrical relations such as rightOf , below, behind, etc. can readily
be defined in terms of other relations (i.e. below(A, B) = above(B, A)).
color [16] (Figure 1 - middle part). Every object hypothesis is therefore represented as an n-tuple:
predicate(instance id, image id, color, spatial loc) where predicate ? {bag, bed, books, ...},
instance id is the object?s id, image id is id of the image containing the object, color is estimated color of the object [16], and spatial loc is the object?s position in the image. Latter is
represented as (Xmin , Xmax , Xmean , Ymin , Ymax , Ymean , Zmin , Zmax , Zmean ) and defines minimal, maximal, and mean location of the object along X, Y, Z axes. To obtain the coordinates we fit
axis parallel cuboids to the cropped 3d objects based on the semantic segmentation. Note that the
X, Y, Z coordinate system is aligned with direction of gravity [15]. As shown in Figure 2b, this is
a more meaningful representation of the object?s coordinates over simple image coordinates. The
complete schema will be documented together with the code release.
We realize that the skilled use of spatial relations is a complex task and grounding spatial relations
is a research thread on its own (e.g. [17], [18] and [19]). For our purposes, we focus on predefined
relations shown in Table 1, while the association of them as well as the object classes are still dealt
within the question answering architecture.
Multi-worlds approach for combining uncertain visual perception and symbolic reasoning
Up to now we have considered the output of the semantic segmentation as ?hard facts?, and hence
ignored uncertainty in the class labeling. Every such labeling of the segments corresponds to different interpretation of the scene - different perceived world. Drawing on ideas from probabilistic
databases [14], we propose a multi-world approach (Figure 1 - lower part) that marginalizes over
multiple possible worlds W - multiple interpretations of a visual scene - derived from the segmentation S. Therefore the posterior over the answer A given question Q and semantic segmentation S
of the image marginalizes over the latent worlds W and logical forms T :
XX
P (A | Q, S) =
P (A | W, T )P (W | S) P (T | Q)
(2)
W
T
The semantic segmentation of the image is a set of segments si with the associated probabilities
pij over the C object categories cj . More precisely S = {(s1 , L1 ), (s2 , L2 ), ..., (sk , Lk )} where
C
Li = {(cj , pij )}j=1 , P (si = cj ) = pij , and k is the number of segments of given image. Let
S?f = (s1 , cf (1) ), (s2 , cf (2) ), ..., (sk , cf (k) )) be an assignment of the categories into segments of
the image according to the binding function f ? F = {1, ..., C}{1,...,k} . With such notation, for
aQfixed binding function f , a world W is a set of tuples consistent with S?f , and define P (W |S) =
k
i p(i,f (i)) . Hence we have as many possible worlds as binding functions, that is C . Eq. 2 becomes
quickly intractable for k and C seen in practice, wherefore we use a sampling strategy that draws a
~ = (W1 , W2 , ..., WN ) from P (?|S) under an assumption that for each segment si
finite sample W
every object?s category cj is drawn independently according to pij . A few sampled perceived worlds
are shown in Figure 2a.
P
Regarding the computational efficiency, computing T P (A | Wi , T )P (T | Q) can be done independently for every Wi , and therefore in parallel without any need for synchronization. Since for
small N the computational costs of summing up computed probabilities is marginal, the overall cost
is about the same as single inference modulo parallelism. The presented multi-world approach to
question answering on real-world scenes is still an end-to-end architecture that is trained solely on
the question-answer pairs.
4
50
50
50
100
100
100
150
150
150
200
200
200
250
250
250
300
300
300
350
350
350
400
400
400
50
100
150
200
250
300
350
400
450
500
550
50
100
150
200
250
300
350
400
450
500
550
50
50
50
100
100
100
150
150
150
200
200
200
250
250
250
300
300
300
350
350
400
400
50
100
150
200
250
300
350
400
450
500
550
50
100
150
200
250
300
350
400
450
500
550
50
100
150
200
250
300
350
400
450
500
550
350
400
50
100
150
200
250
300
350
400
450
500
550
set
Individual
Figure 3: NYU-Depth V2 dataset: image, Z axis, ground truth and predicted semantic segmentations.
Description
counting
counting and colors
room type
superlatives
counting and colors
negations type 1
negations type 2
negations type 3
Template
How many {object} are in {image id}?
How many {color} {object} are in {image id}?
Which type of the room is depicted in {image id}?
What is the largest {object} in {image id}?
How many {color} {object}?
Which images do not have {object}?
Which images are not {room type}?
Which images have {object} but do not have a {object}?
Example
How many cabinets are in image1?
How many gray cabinets are in image1?
Which type of the room is depicted in image1?
What is the largest object in image1?
How many black bags?
Which images do not have sofa?
Which images are not bedroom?
Which images have desk but do not have a lamp?
Table 2: Synthetic question-answer pairs. The questions can be about individual images or the sets of images.
Implementation and Scalability For worlds containing many facts and spatial relations the induction step becomes computationally demanding as it considers all pairs of the facts (we have
about 4 million predicates in the worst case). Therefore we use a batch-based approximation in such
situations. Every image induces a set of facts that we call a batch of facts. For every test image, we
find k nearest neighbors in the space of training batches with a boolean variant of TF.IDF to measure similarity [20]. This is equivalent to building a training world from k images with most similar
content to the perceived world of the test image. We use k = 3 and 25 worlds in our experiments.
Dataset and the source code can be found in our website 1 .
4
Experiments
4.1 DAtaset for QUestion Answering on Real-world images (DAQUAR)
Images and Semantic Segmentation Our new dataset for question answering is built on top of
the NYU-Depth V2 dataset [6]. NYU-Depth V2 contains 1449 RGBD images together with annotated semantic segmentations (Figure 3) where every pixel is labeled into some object class with a
confidence score. Originally 894 classes are considered. According to [15], we preprocess the data
to obtain canonical views of the scenes and use X, Y , Z coordinates from the depth sensor to define
spatial placement of the objects in 3D. To investigate the impact of uncertainty in the visual analysis
of the scenes, we also employ computer vision techniques for automatic semantic segmentation. We
use a state-of-the-art scene analysis method [15] which maps every pixel into 40 classes: 37 informative object classes as well as ?other structure?, ?other furniture? and ?other prop?. We ignore the
latter three. We use the same data split as [15]: 795 training and 654 test images. To use our spatial
representation on the image content, we fit 3d cuboids to the segmentations.
New dataset of questions and answers In the spirit of a visual turing test, we collect question
answer pairs from human annotators for the NYU dataset. In our work, we consider two types of the
annotations: synthetic and human. The synthetic question-answer pairs are automatically generated
question-answer pairs, which are based on the templates shown in Table 2. These templates are then
instantiated with facts from the database. To collect 12468 human question-answer pairs we ask 5
in-house participants to provide questions and answers. They were instructed to give valid answers
that are either basic colors [16], numbers or objects (894 categories) or sets of those. Besides the
answers, we don?t impose any constraints on the questions. We also don?t correct the questions as
we believe that the semantic parsers should be robust under the human errors. Finally, we use 6794
training and 5674 test question-answer pairs ? about 9 pairs per image on average (8.63, 8.75)2 .
1
https://www.d2.mpi-inf.mpg.de/visual-turing-challenge
Our notation (x, y) denotes mean x and trimean y. We use Tukey?s trimean 41 (Q1 + 2Q2 + Q3 ), where Qj
denotes the j-th quartile [21]. This measure combines the benefits of both median (robustness to the extremes)
and empirical mean (attention to the hinge values).
2
5
The database exhibit some biases showing humans tend to focus on a few prominent objects. For
instance we have more than 400 occurrences of table and chair in the answers. In average the
object?s category occurs (14.25, 4) times in training set and (22.48, 5.75) times in total. Figure 4
shows example question-answer pairs together with the corresponding image that illustrate some of
the challenges captured in this dataset.
Performance Measure While the quality of an answer that the system produces can be measured
in terms of accuracy w.r.t. the ground truth (correct/wrong), we propose, inspired from the work
on Fuzzy Sets [22], a soft measure based on the WUP score [23], which we call WUPS (WUP
Set) score. As the number of classes grows, the semantic boundaries between them are becoming
more fuzzy. For example, both concepts ?carton? and ?box? have similar meaning, or ?cup? and
?cup of coffee? are almost indifferent. Therefore we seek a metric that measures the quality of an
answer and penalizes naive solutions where the architecture outputs too many or too few answers.
PN
Standard Accuracy is defined as: N1 i=1 1{Ai = T i } ? 100 where Ai , T i are i-th answer and
ground-truth respectively. Since both the answers may include more than one object, it is beneficial
to represent them as sets of the objects T = {t1 , t2 , ...}. From this point of view we have for every
i ? {1, 2, ..., N }:
1{Ai = T i } = 1{Ai ? T i ? T i ? Ai } = min{1{Ai ? T i }, 1{T i ? Ai }}
Y
Y
Y
Y
= min{
1{a ? T i },
1{t ? Ai }} ? min{
?(a ? T i ),
?(t ? Ai )}
a?Ai
t?T i
a?Ai
(3)
(4)
t?T i
We use a soft equivalent of the intersection operator in Eq. 3, and a set membership measure ?,
with properties ?(x ? X) = 1 if x ? X, ?(x ? X) = maxy?X ?(x = y) and ?(x = y) ? [0, 1],
in Eq. 4 with equality whenever ? = 1. For ? we use a variant of Wu-Palmer similarity [23, 24].
WUP(a, b) calculates similarity based on the depth of two words a and b in the taxonomy[25, 26],
and define the WUPS score:
N
Y
Y
1 X
WUPS(A, T ) =
min{
maxi WUP(a, t),
max WUP(a, t)} ? 100
(5)
N i=1
t?T
a?Ai
i
i
a?A
t?T
Empirically, we have found that in our task a WUP score of around 0.9 is required for precise
answers. Therefore we have implemented down-weighting WUP(a, b) by one order of magnitude
(0.1 ? WUP) whenever WUP(a, b) < t for a threshold t. We plot a curve over thresholds t ranging
from 0 to 1 (Figure 5). Since ?WUPS at 0? refers to the most ?forgivable? measure without any downweighting and ?WUPS at 1.0? corresponds to plain accuracy. Figure 5 benchmarks architectures by
requiring answers with precision ranging from low to high. Here we show some examples of the pure
WUP score to give intuitions about the range: WUP(curtain, blinds) = 0.94, WUP(carton, box) =
0.94, WUP(stove, fire extinguisher) = 0.82.
4.2 Quantitative results
We perform a series of experiments to highlight particular challenges like uncertain segmentations, unknown true logical forms, some linguistic phenomena as well as show the advantages of
our proposed multi-world approach. In particular, we distinguish between experiments on synthetic question-answer pairs (SynthQA) based on templates and those collected by annotators (HumanQA), automatic scene segmentation (AutoSeg) with a computer vision algorithm [15] and human segmentations (HumanSeg) based on the ground-truth annotations in the NYU dataset as well
as single world (single) and multi-world (multi) approaches.
4.2.1 Synthetic question-answer pairs (SynthQA)
Based on human segmentations (HumanSeg, 37 classes) (1st and 2nd rows in Table 3) uses automatically generated questions (we use templates shown in Table 2) and human segmentations.
We have generated 20 training and 40 test question-answer pairs per template category, in total 140
training and 280 test pairs (as an exception negations type 1 and 2 have 10 training and 20 test examples each). This experiment shows how the architecture generalizes across similar type of questions
provided that we have human annotation of the image segments. We have further removed negations
of type 3 in the experiments as they have turned out to be particularly computationally demanding.
Performance increases hereby from 56% to 59.9% with about 80% training Accuracy. Since some
incorrect derivations give correct answers, the semantic parser learns wrong associations. Other difficulties stem from the limited training data and unseen object categories during training.
Based on automatic segmentations (AutoSeg, 37 classes, single) (3rd row in Table 3) tests the architecture based on uncertain facts obtained from automatic semantic segmentation [15] where the
6
most likely object labels are used to create a single world. Here, we are experiencing a severe drop
in performance from 59.9% to 11.25% by switching from human to automatic segmentation. Note
that there are only 37 classes available to us. This result suggests that the vision part is a serious
bottleneck of the whole architecture.
Based on automatic segmentations using multi-world approach (AutoSeg, 37 classes, multi)
(4th row in Table 3) shows the benefits of using our multiple worlds approach to predict the answer. Here we recover part of the lost performance by an explicit treatment of the uncertainty in the
segmentations. Performance increases from 11.25% to 13.75%.
4.3 Human question-answer pairs (HumanQA)
Based on human segmentations 894 classes (HumanSeg, 894 classes) (1st row in Table 4) switching to human generated question-answer pairs. The increase in complexity is twofold. First, the
human annotations exhibit more variations than the synthetic approach based on templates. Second,
the questions are typically longer and include more spatially related objects. Figure 4 shows a few
samples from our dataset that highlights challenges including complex and nested spatial reference
and use of reference frames. We yield an accuracy of 7.86% in this scenario. As argued above,
we also evaluate the experiments on the human data under the softer WUPS scores given different
thresholds (Table 4 and Figure 5). In order to put these numbers in perspective, we also show performance numbers for two simple methods: predicting the most popular answer yields 4.4% Accuracy,
and our untrained architecture gives 0.18% and 1.3% Accuracy and WUPS (at 0.9).
Based on human segmentations 37 classes (HumanSeg, 37 classes) (2nd row in Table 4) uses human segmentation and question-answer pairs. Since only 37 classes are supported by our automatic
segmentation algorithm, we run on a subset of the whole dataset. We choose the 25 test images
yielding a total of 286 question answer pairs for the following experiments. This yields 12.47% and
15.89% Accuracy and WUPS at 0.9 respectively.
Based on automatic segmentations (AutoSeg, 37 classes) (3rd row in Table 4) Switching from the
human segmentations to the automatic yields again a drop from 12.47% to 9.69% in Accuracy and
we observe a similar trend for the whole spectrum of the WUPS scores.
Based on automatic segmentations using multi-world approach (AutoSeg, 37 classes, multi)
(4th row in Table 4) Similar to the synthetic experiments our proposed multi-world approach yields
an improvement across all the measure that we investigate.
Human baseline (5th and 6th rows in Table 4 for 894 and 37 classes) shows human predictions on
our dataset. We ask independent annotators to provide answers on the questions we have collected.
They are instructed to answer with a number, basic colors [16], or objects (from 37 or 894 categories) or set of those. This performance gives a practical upper bound for the question-answering
algorithms with an accuracy of 60.27% for the 37 class case and 50.20% for the 894 class case.
We also ask to compare the answers of the AutoSeg single world approach with HumanSeg single
world and AutoSeg multi-worlds methods. We use a two-sided binomial test to check if difference
in preferences is statistically significant. As a result AutoSeg single world is the least preferred
method with the p-value below 0.01 in both cases. Hence the human preferences are aligned with
our accuracy measures in Table 4.
4.4 Qualitative results
We choose examples in Fig. 6 to illustrate different failure cases - including last example where all
methods fail. Since our multi-world approach generates different sets of facts about the perceived
worlds, we observe a trend towards a better representation of high level concepts like ?counting?
(leftmost the figure) as well as language associations. A substantial part of incorrect answers is
attributed to missing segments, e.g. no pillow detection in third example in Fig. 6.
5
Summary
We propose a system and a dataset for question answering about real-world scenes that is reminiscent
of a visual turing test. Despite the complexity in uncertain visual perception, language understanding
and program induction, our results indicate promising progress in this direction. We bring ideas
together from automatic scene analysis, semantic parsing with symbolic reasoning, and combine
them under a multi-world approach. As we have mature techniques in machine learning, computer
vision, natural language processing and deduction at our disposal, it seems timely to bring these
disciplines together on this open challenge.
7
QA: (What is behind the table?, window)!
Spatial relation like ?behind? are dependent
on the reference frame. Here the annotator
uses observer-centric view.!
The annotators are using different names to
call the same things. The names of the
brown object near the bed include ?night
stand?, ?stool?, and ?cabinet?.
QA: (what is beneath the candle holder,
decorative plate)!
Some annotators use variations on spatial
relations that are similar, e.g. ?beneath? is
closely related to ?below?.!
Some objects, like the table on the left of
image, are severely occluded or truncated.
Yet, the annotators refer to them in the
questions.
QA: (What is in front of toilet?, door)!
Here the ?open door? to the restroom is not
clearly visible, yet captured by the annotator.!
!
QA: (what is in front of the wall divider?,
cabinet)?
Annotators use additional properties to
clarify object references (i.e. wall divider).
Moreover, the perspective plays an
important role in these spatial relations
interpretations.
!
QA1:(How many doors are in the image?, 1)!QA: (How many drawers are there?, 8)!
QA2:(How many doors are in the image?, 5)!The annotators use their common-sense
knowledge for amodal completion. Here the
Different interpretation of ?door? results in
different counts: 1 door at the end of the hall ? annotator infers the 8th drawer from the
context
vs. 5 doors including lockers
QA: (what is behind the table?, sofa)!
Spatial relations exhibit different reference
frames. Some annotations use observercentric, others object-centric view!
QA: (how many lights are on?, 6)!
Moreover, some questions require detection
of states ?light on or off??
QA: (What is the shape of the green
chair?, horse shaped)!
In this example, an annotator refers to a
?horse shaped chair? which requires a quite
abstract reasoning about the shapes.!
QA1: (what is in front of the curtain behind
the armchair?, guitar)!
!
QA2: (what is in front of the curtain?,
guitar)!
!
Q: what is at the back side of the sofas?!
Annotators use wide range spatial relations,
such as ?backside? which is object-centric.
Spatial relations matter more in complex
environments where reference resolution
becomes more relevant. In cluttered scenes,
pragmatism starts playing a more important
role
QA: (What is the object on the counter in
the corner?, microwave)!
References like ?corner? are difficult to
resolve given current computer vision
models. Yet such scene features are
frequently used by humans.!
QA: (How many doors are open?, 1)!
Notion of states of object (like open) is not
well captured by current vision techniques.
Annotators use such attributes frequently
for disambiguation.!
QA: (Where is oven?, on the right side of
refrigerator)!
On some occasions, the annotators prefer to
use more complex responses. With spatial
relations, we can increase the answer?s
precision.!
Figure 4: Examples of human generated question-answer pairs illustrating the associated challenges. In the
descriptions we use following notation: ?A? - answer, ?Q? - question, ?QA? - question-answer pair. Last two
examples (bottom-right column) are from the extended dataset not used in our experiments.
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
0.4
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
0.0
0
0.1
0.2
0.3
?
?
?
?
0.4
0.5
?
?
?
?
?
?
?
?
?
?
?
?
0.8
0.9
1
synthetic question-answer pairs (SynthQA)
Segmentation
World(s)
# classes Accuracy
HumanSeg Single with Neg. 3
37
56.0%
HumanSeg
Single
37
59.5%
AutoSeg
Single
37
11.25%
AutoSeg
Multi
37
13.75%
?
?
?
?
HumanSeg, Single, 894
HumanSeg, Single, 37
AutoSeg, Single, 37
AutoSeg, Multi, 37
Human Baseline, 894
Human Baseline, 37
0.2
WUPS
0.6
0.8
HumanQA
?
0.6
0.7
?
?
?
Threshold
Figure 5: WUPS scores for different thresholds.
Table 3: Accuracy results for the experiments with synthetic question-answer pairs.
Human question-answer pairs (HumanQA)
Segmentation World(s) #classes Accuracy WUPS at 0.9
HumanSeg
Single
894
7.86%
11.86%
HumanSeg
Single
37
12.47%
16.49%
AutoSeg
Single
37
9.69%
14.73%
AutoSeg
Multi
37
12.73%
18.10%
Human Baseline
894
50.20%
50.82%
Human Baseline
37
60.27%
61.04%
WUPS at 0
38.79%
50.28%
48.57%
51.47%
67.27%
78.96%
Table 4: Accuracy and WUPS scores for the experiments with human question-answer pairs. We show WUPS
scores at two opposite sides of the WUPS spectrum.
Q: How many red chairs are there?!
H: ()!
M: 6!
C: blinds!
!
Q: How many chairs are at the table?!
H: wall?
M: 4!
C: chair
Q: What is on the right side of cabinet?!
H: picture?
M: bed!
C: bed
Q: What is on the wall?!
H: mirror!
M: bed!
C: picture
Q: What is the object on the chair?!
H: pillow!
M: floor, wall!
C: wall
Q: What is on the right side of the table?!
H: chair?
M: window, floor, wall!
C: floor
Q: What is behind the television?!
H: lamp?
M: brown, pink, purple!
C: picture
Q: What is in front of television?!
H: pillow!
M: chair!
C: picture
Figure 6: Questions and predicted answers. Notation: ?Q? - question, ?H? - architecture based on human
segmentation, ?M? - architecture with multiple worlds, ?C? - most confident architecture, ?()? - no answer. Red
color denotes correct answer.
8
References
[1] Liang, P., Jordan, M.I., Klein, D.: Learning dependency-based compositional semantics. Computational Linguistics (2013)
[2] Kwiatkowski, T., Zettlemoyer, L., Goldwater, S., Steedman, M.: Inducing probabilistic ccg
grammars from logical form with higher-order unification. In: EMNLP. (2010)
[3] Zettlemoyer, L.S., Collins, M.: Online learning of relaxed ccg grammars for parsing to logical
form. In: EMNLP-CoNLL-2007. (2007)
[4] Matuszek, C., Fitzgerald, N., Zettlemoyer, L., Bo, L., Fox, D.: A joint model of language and
perception for grounded attribute learning. In: ICML. (2012)
[5] Krishnamurthy, J., Kollar, T.: Jointly learning to parse and perceive: Connecting natural
language to the physical world. TACL (2013)
[6] Silberman, N., Hoiem, D., Kohli, P., Fergus, R.: Indoor segmentation and support inference
from rgbd images. In: ECCV. (2012)
[7] Kong, C., Lin, D., Bansal, M., Urtasun, R., Fidler, S.: What are you talking about? text-toimage coreference. In: CVPR. (2014)
[8] Karpathy, A., Joulin, A., Fei-Fei, L.: Deep fragment embeddings for bidirectional image
sentence mapping. In: NIPS. (2014)
[9] Matuszek, C., Herbst, E., Zettlemoyer, L., Fox, D.: Learning to parse natural language commands to a robot control system. In: Experimental Robotics. (2013)
[10] Levit, M., Roy, D.: Interpretation of spatial language in a map navigation task. Systems, Man,
and Cybernetics, Part B: Cybernetics, IEEE Transactions on (2007)
[11] Vogel, A., Jurafsky, D.: Learning to follow navigational directions. In: ACL. (2010)
[12] Tellex, S., Kollar, T., Dickerson, S., Walter, M.R., Banerjee, A.G., Teller, S.J., Roy, N.: Understanding natural language commands for robotic navigation and mobile manipulation. In:
AAAI. (2011)
[13] Kruijff, G.J.M., Zender, H., Jensfelt, P., Christensen, H.I.: Situated dialogue and spatial organization: What, where... and why. IJARS (2007)
[14] Wick, M., McCallum, A., Miklau, G.: Scalable probabilistic databases with factor graphs and
mcmc. In: VLDB. (2010)
[15] Gupta, S., Arbelaez, P., Malik, J.: Perceptual organization and recognition of indoor scenes
from rgb-d images. In: CVPR. (2013)
[16] Van De Weijer, J., Schmid, C., Verbeek, J.: Learning color names from real-world images. In:
CVPR. (2007)
[17] Regier, T., Carlson, L.A.: Grounding spatial language in perception: an empirical and computational investigation. Journal of Experimental Psychology: General (2001)
[18] Lan, T., Yang, W., Wang, Y., Mori, G.: Image retrieval with structured object queries using
latent ranking svm. In: ECCV. (2012)
[19] Guadarrama, S., Riano, L., Golland, D., Gouhring, D., Jia, Y., Klein, D., Abbeel, P., Darrell,
T.: Grounding spatial relations for human-robot interaction. In: IROS. (2013)
[20] Manning, C.D., Raghavan, P., Sch?utze, H.: Introduction to information retrieval. Cambridge
university press Cambridge (2008)
[21] Tukey, J.W.: Exploratory data analysis. (1977)
[22] Zadeh, L.A.: Fuzzy sets. Information and control (1965)
[23] Wu, Z., Palmer, M.: Verbs semantics and lexical selection. In: ACL. (1994)
[24] Guadarrama, S., Krishnamoorthy, N., Malkarnenkar, G., Mooney, R., Darrell, T., Saenko, K.:
Youtube2text: Recognizing and describing arbitrary activities using semantic hierarchies and
zero-shot recognition. In: ICCV. (2013)
[25] Miller, G.A.: Wordnet: a lexical database for english. CACM (1995)
[26] Fellbaum, C.: WordNet. Wiley Online Library (1999)
9
| 5411 |@word kohli:1 kong:1 illustrating:1 middle:1 seems:2 nd:2 open:4 d2:1 vldb:1 seek:1 rgb:1 q1:1 shot:1 loc:2 contains:1 score:11 series:1 hoiem:1 fragment:1 miklau:1 current:4 guadarrama:2 si:3 yet:3 wherefore:1 reminiscent:1 parsing:4 readily:1 realize:1 wup:13 realistic:1 visible:1 informative:1 shape:2 plot:1 drop:2 update:1 v:1 website:1 mccallum:1 lamp:2 zmax:3 core:1 node:2 location:2 preference:2 along:1 skilled:1 incorrect:2 qualitative:1 combine:5 introduce:1 zmin:3 mpg:2 frequently:2 multi:22 inspired:2 automatically:3 food:1 td:3 ucken:1 actual:1 window:2 increasing:1 becomes:3 begin:1 xx:1 moreover:3 notation:4 provided:1 what:23 string:3 q2:1 fuzzy:3 proposing:1 quantitative:1 every:10 gravity:1 wrong:2 control:2 planck:1 t1:3 severely:1 switching:3 despite:1 id:10 solely:4 becoming:1 black:1 acl:2 collect:3 suggests:1 jurafsky:1 limited:1 palmer:2 range:3 statistically:1 unique:2 practical:1 recursive:1 block:1 practice:1 lost:1 empirical:2 confidence:2 word:1 refers:3 symbolic:3 close:1 selection:1 operator:1 put:1 context:1 www:1 equivalent:2 map:2 missing:1 lexical:2 attention:1 independently:2 cluttered:1 resolution:1 pure:1 perceive:1 insight:1 examines:1 handle:1 searching:1 coordinate:8 variation:2 notion:1 krishnamurthy:1 exploratory:1 hierarchy:1 trigger:1 parser:6 user:1 modulo:1 experiencing:1 play:1 us:3 hypothesis:1 wups:16 trend:2 roy:2 recognition:4 particularly:1 database:7 labeled:1 bottom:1 role:2 wang:1 refrigerator:1 worst:1 region:1 counter:1 removed:1 xmin:4 substantial:2 envisioned:1 intuition:1 environment:1 complexity:3 fitzgerald:1 occluded:1 trained:3 segment:8 coreference:1 efficiency:1 toilet:1 completely:1 joint:1 represented:2 derivation:1 walter:1 instantiated:1 query:1 labeling:3 horse:2 cacm:1 quite:1 cvpr:3 drawing:1 grammar:2 unseen:1 jointly:1 online:2 advantage:2 propose:6 interaction:1 maximal:1 drawer:2 aligned:2 combining:2 turned:1 date:1 beneath:2 relevant:1 ymax:2 bed:6 description:2 inducing:1 scalability:2 darrell:2 r1:1 produce:2 executing:1 object:51 illustrate:2 completion:1 measured:1 nearest:1 progress:3 eq:3 strong:1 auxiliary:2 predicted:2 implemented:1 come:1 indicate:1 direction:5 concentrate:1 closely:1 correct:5 annotated:1 attribute:2 quartile:1 human:33 raghavan:1 softer:1 argued:1 require:1 abbeel:1 wall:8 investigation:1 clarify:1 around:1 considered:2 ground:4 hall:1 exp:1 scope:2 mapping:2 predict:1 achieves:1 early:1 utze:1 purpose:2 perceived:10 integrates:1 sofa:4 bag:2 label:1 largest:3 correctness:1 tf:1 create:1 clearly:1 sensor:1 aim:1 rather:1 curtain:3 pn:1 mobile:1 command:5 linguistic:2 derived:2 focus:4 ax:1 release:1 q3:1 improvement:1 check:1 mainly:1 contrast:4 baseline:5 sense:1 inference:6 dependent:1 factoring:1 membership:1 integrated:2 typically:1 relation:21 deduction:2 germany:1 semantics:2 pixel:3 compatibility:1 overall:1 art:3 constrained:1 spatial:25 daquar:1 marginal:1 weijer:1 never:1 shaped:2 sampling:1 represents:1 icml:1 others:3 t2:3 inherent:1 few:6 employ:1 modern:2 serious:1 individual:2 consisting:1 fire:1 n1:1 negation:7 attempt:1 detection:2 organization:2 interest:1 investigate:2 steedman:1 evaluation:4 severe:1 indifferent:1 navigation:3 extreme:1 yielding:1 light:2 behind:6 tj:2 chain:1 subtrees:1 predefined:1 microwave:1 tuple:2 unification:1 fox:2 tree:4 unite:1 penalizes:1 minimal:1 uncertain:6 instance:6 brick:1 modeling:1 boolean:1 soft:2 amodal:1 column:1 measuring:1 assignment:1 cost:2 addressing:1 subset:1 recognizing:1 predicate:11 too:2 front:5 connect:1 answer:62 dependency:1 learnt:1 synthetic:9 confident:3 person:1 fritz:1 st:2 armchair:1 probabilistic:7 off:1 informatics:1 discipline:1 together:6 connecting:2 quickly:1 w1:1 again:1 ambiguity:1 aaai:1 containing:2 choose:2 marginalizes:4 emnlp:2 corner:2 book:1 dialogue:1 li:1 account:1 de:3 matter:1 ranking:1 blind:2 performed:1 view:4 observer:1 mario:1 linked:1 schema:1 tof:4 tukey:2 participant:1 parallel:2 recover:1 annotation:7 timely:2 start:1 red:2 levit:1 jia:1 contribution:1 purple:1 accuracy:15 holder:1 miller:1 correspond:1 yield:6 preprocess:1 dealt:1 weak:1 bayesian:3 goldwater:1 plastic:1 krishnamoorthy:1 produced:1 cybernetics:2 mooney:1 manual:1 whenever:2 definition:1 evaluates:1 failure:1 hereby:1 associated:4 cabinet:5 attributed:1 sampled:3 dataset:17 treatment:1 popular:1 ask:4 logical:15 knowledge:2 color:13 infers:1 segmentation:39 cj:4 back:1 fellbaum:1 centric:3 bidirectional:1 disposal:1 originally:1 higher:1 day:1 follow:1 response:1 formulation:1 execute:2 box:3 done:1 furthermore:1 reply:1 clock:2 tacl:1 night:1 parse:2 banerjee:1 glance:1 defines:1 quality:2 gray:2 believe:1 grows:1 grounding:4 building:3 name:3 brown:7 concept:2 requiring:1 true:1 divider:2 hence:4 equality:1 fidler:1 alternating:1 spatially:1 semantic:30 illustrated:1 white:2 mug:1 regier:1 during:1 mpi:2 leftmost:1 prominent:1 occasion:1 plate:1 bansal:1 outline:1 complete:1 l1:1 bring:2 reasoning:7 image:67 meaning:2 wise:2 novel:1 ranging:2 common:1 physical:2 overview:2 empirically:1 attached:1 million:1 extend:1 interpretation:9 association:3 refer:2 significant:1 cup:2 cambridge:2 ai:14 stool:1 automatic:14 unconstrained:1 rd:3 populated:1 similarly:1 hp:1 language:22 robot:2 supervision:1 similarity:3 longer:1 etc:1 posterior:2 own:1 recent:2 oven:1 perspective:2 inf:4 scenario:3 manipulation:1 herbst:1 neg:1 seen:2 captured:3 additional:3 relaxed:1 impose:1 floor:3 recognized:1 relates:1 full:1 multiple:9 rj:2 reduces:1 stem:1 lin:1 retrieval:2 equally:1 impact:1 prediction:4 variant:2 basic:2 calculates:1 scalable:1 vision:10 metric:1 verbeek:1 represent:2 grounded:1 xmax:4 robotics:1 golland:1 whereas:1 cropped:1 resolve:1 zettlemoyer:4 median:1 source:2 sch:1 w2:1 operate:2 vogel:1 bringing:1 tend:1 mature:2 thing:1 kollar:2 spirit:1 jordan:1 call:4 near:1 counting:5 door:8 yang:1 split:1 embeddings:1 wn:1 fit:2 psychology:1 bedroom:1 architecture:13 opposite:2 imperfect:1 regarding:2 idea:2 image1:4 qj:1 bottleneck:1 thread:1 handled:1 compositional:1 deep:1 ignored:1 malinowski:1 detailed:1 karpathy:1 amount:1 desk:1 situated:1 induces:1 category:8 documented:1 generate:1 http:1 canonical:1 toend:1 estimated:1 per:2 klein:2 diverse:1 discrete:1 wick:1 threshold:5 lan:1 monitor:1 drawn:1 downweighting:1 iros:1 graph:1 realworld:1 turing:6 run:1 uncertainty:7 you:1 named:1 almost:1 reader:1 reasonable:1 wu:2 draw:1 datalog:1 disambiguation:1 prefer:1 zadeh:1 conll:1 comparable:1 matuszek:2 bound:1 carton:2 furniture:1 stove:1 distinguish:1 activity:1 ahead:1 denotation:1 precisely:1 idf:1 placement:1 constraint:1 scene:26 fei:2 generates:1 chair:12 min:4 toimage:1 xmean:5 structured:1 according:3 manning:1 pink:1 across:3 beneficial:1 wi:2 s1:2 candle:1 maxy:1 christensen:1 refered:1 restricted:2 iccv:1 sided:1 computationally:2 mori:1 describing:1 count:2 fail:1 mind:1 serf:1 end:4 generalizes:1 available:1 apply:1 observe:2 v2:3 appropriate:1 occurrence:1 alternative:1 batch:3 robustness:1 original:2 top:2 running:1 cf:3 denotes:3 superlative:3 include:3 binomial:1 hinge:1 linguistics:1 carlson:1 restrictive:1 build:3 establish:1 coffee:1 silberman:1 malik:1 question:73 occurs:2 strategy:1 unclear:1 exhibit:3 gradient:1 arbelaez:1 entity:1 argue:1 considers:3 collected:2 urtasun:1 induction:2 dream:1 code:2 besides:1 relationship:2 liang:1 difficult:1 taxonomy:1 trace:1 locker:1 implementation:1 unknown:1 perform:1 upper:1 benchmark:4 finite:1 descent:1 inevitably:1 truncated:1 defining:1 situation:1 extended:1 precise:1 dc:1 frame:3 verb:1 arbitrary:1 inferred:1 introduced:1 pair:29 required:1 sentence:2 engine:3 textual:1 saarbr:1 nip:1 qa:12 beyond:1 below:4 perception:7 parallelism:1 indoor:4 mateusz:1 challenge:9 program:1 navigational:1 built:1 max:2 green:2 including:3 demanding:2 natural:11 rely:1 difficulty:1 predicting:2 library:1 picture:5 axis:3 ymin:2 ready:1 lk:1 naive:1 schmid:1 text:1 understanding:8 l2:1 teller:1 marginalizing:1 lacking:1 synchronization:1 highlight:2 interesting:1 limitation:1 annotator:15 degree:1 pij:4 consistent:1 imposes:1 playing:1 row:8 eccv:2 summary:1 supported:1 last:2 english:1 lef:2 side:6 bias:1 institute:1 neighbor:1 template:9 taking:1 wide:1 benefit:3 van:1 boundary:1 depth:5 curve:1 world:78 valid:2 plain:1 pillow:3 stand:1 instructed:2 made:2 collection:1 transaction:1 ignore:1 preferred:1 dealing:1 cuboid:2 robotic:1 summing:1 symmetrical:1 tuples:1 fergus:1 don:2 spectrum:2 latent:8 sk:2 why:1 table:27 promising:1 learn:1 robust:1 broadening:1 untrained:1 complex:7 domain:1 joulin:1 bounding:1 whole:4 s2:2 child:1 rgbd:4 fig:4 downwards:1 zmean:5 wiley:1 precision:2 position:2 explicit:1 lie:1 house:1 answering:19 perceptual:1 weighting:1 third:1 learns:4 down:1 specific:1 showing:1 maxi:1 list:1 r2:1 nyu:5 guitar:2 gupta:1 svm:1 intractable:1 mirror:1 ccg:2 magnitude:1 subtree:1 zaux:2 television:2 demand:1 depicted:2 intersection:1 likely:1 visual:14 bo:1 talking:1 binding:4 corresponds:4 truth:4 nested:1 prop:1 goal:2 exposition:1 towards:1 room:4 twofold:1 man:1 content:2 hard:1 wordnet:2 total:3 experimental:2 meaningful:1 saenko:1 exception:1 formally:1 support:1 latter:2 meant:1 collins:1 evaluate:1 mcmc:1 phenomenon:2 |
4,874 | 5,412 | Quantized Kernel Learning for Feature Matching
Danfeng Qin
ETH Z?urich
Xuanli Chen
TU Munich
Matthieu Guillaumin
ETH Z?urich
Luc Van Gool
ETH Z?urich
{qind, guillaumin, vangool}@vision.ee.ethz.ch, [email protected]
Abstract
Matching local visual features is a crucial problem in computer vision and its
accuracy greatly depends on the choice of similarity measure. As it is generally
very difficult to design by hand a similarity or a kernel perfectly adapted to the
data of interest, learning it automatically with as few assumptions as possible is
preferable. However, available techniques for kernel learning suffer from several
limitations, such as restrictive parametrization or scalability.
In this paper, we introduce a simple and flexible family of non-linear kernels
which we refer to as Quantized Kernels (QK). QKs are arbitrary kernels in the
index space of a data quantizer, i.e., piecewise constant similarities in the original feature space. Quantization allows to compress features and keep the learning
tractable. As a result, we obtain state-of-the-art matching performance on a standard benchmark dataset with just a few bits to represent each feature dimension.
QKs also have explicit non-linear, low-dimensional feature mappings that grant
access to Euclidean geometry for uncompressed features.
1
Introduction
Matching local visual features is a core problem in computer vision with a vast range of applications
such as image registration [28], image alignment and stitching [6] and structure-from-motion [1].
To cope with the geometric transformations and photometric distorsions that images exhibit, many
robust feature descriptors have been proposed. In particular, histograms of oriented gradients such
as SIFT [15] have proved successful in many of the above tasks. Despite these results, they are
inherently limited by their design choices. Hence, we have witnessed an increasing amount of work
focusing on automatically learning visual descriptors from data via discriminative embeddings [11,
4] or hyper-parameter optimization [5, 21, 23, 22].
A dual aspect of visual description is the measure of visual (dis-)similarity, which is responsible
for deciding whether a pair of features matches or not. In image registration, retrieval and 3D
reconstruction, for instance, nearest neighbor search builds on such measures to establish point
correspondences. Thus, the choice of similarity or kernel impacts the performance of a system as
much as the choice of visual features [2, 16, 18]. Designing a good similarity measure for matching
is difficult and commonly used kernels such as the linear, intersection, ?2 and RBF kernels are not
ideal as their inherent properties (e.g., stationarity, homogeneity) may not fit the data well.
Existing techniques for automatically learning similarity measures suffer from different limitations.
Metric learning approaches [25] learn to project the data to a lower-dimensional and more discriminative space where the Euclidean geometry can be used. However, these methods are inherently
linear. Multiple Kernel Learning (MKL) [3] is able to combine multiple base kernels in an optimal
way, but its complexity limits the amount of data that can be used and forces the user to pre-select
or design a small number of kernels that are likely to perform well. Additionally, the resulting kernel may not be easily represented in a reasonably small Euclidean space. This is problematic, as
many efficient algorithms (e.g. approximate nearest neighbor techniques) heavily rely on Euclidean
geometry and have non-intuitive behavior in higher dimensions.
1
In this paper, we introduce a simple yet powerful family of kernels, Quantized Kernels (QK), which
(a) model non-linearities and heterogeneities in the data, (b) lead to compact representations that
can be easily decompressed into a reasonably-sized Euclidean space and (c) are efficient to learn so
that large-scale data can be exploited. In essence, we build on the fact that vector quantizers project
data into a finite set of N elements, the index space, and on the simple observation that kernels on
finite sets are fully specified by the N?N Gram matrix of these elements (the kernel matrix), which
we propose to learn directly. Thus, QKs are piecewise constant but otherwise arbitrary, making
them very flexible. Since the learnt kernel matrices are positive semi-definite, we directly obtain the
corresponding explicit feature mappings and exploit their potential low-rankness.
In the remainder of the paper, we first further discuss related work (Sec. 2), then present QKs in detail
(Sec. 3). As important contributions, we show how to efficiently learn the quantizer and the kernel
matrix so as to maximize the matching performance (Sec. 3.2), using an exact linear-time inference
subroutine (Sec. 3.3), and devise practical techniques for users to incorporate knowledge about the
structure of the data (Sec. 3.4) and reduce the number of parameters of the system. Our experiments
in Sec. 4 show that our kernels yield state-of-the-art performance on a standard feature matching
benchmark and improve over kernels used in the literature for several descriptors, including one
based on metric learning. Our compressed features are very compact, using only 1 to 4 bits per
dimension of the original features. For instance, on SIFT descriptors, our QK yields about 10%
improvement on matching compared to the dot product, while compressing features by a factor 8.
2
Related work
Our work relates to a vast literature on kernel selection and tuning, descriptor, similarity, distance
and kernel learning. We present a selection of such works below.
Basic kernels and kernel tuning. A common approach for choosing a kernel is to pick one from
the literature: dot product, Gaussian RBF, intersection [16], ?2 , Hellinger, etc. These generic kernels
have been extensively studied [24] and have properties such as homogeneity or stationarity. These
properties may be inadequate for the data of interest and thus the kernels will not yield optimal
performance. Efficient yet approximate versions of such kernels [9, 20, 24] are similarly inadequate.
Descriptor learning. Early work on descriptor learning improved SIFT by exploring its parameter space [26]. Later, automatic parameter selection was proposed with a non-convex objective [5].
Recently, significant improvements in local description for matching have been obtained by optimizing feature encoding [4] and descriptor pooling [21, 23]. These works maximize the matching
performance directly via convex optimization [21] or boosting [23]. As we show in our experiments,
our approach improves matching even for such optimized descriptors.
Distance, similarity and kernel learning. Mahalanobis metrics (e.g., [25]) are probably the most
widely used family of (dis-)similarities in supervised settings. They extend the Euclidean metric
by accounting for correlations between input dimensions and are equivalent to projecting data to
a new, potentially smaller, Euclidean space. Learning the projection improves discrimination and
compresses feature vectors, but the projection is inherently linear.1 There are several attempts to
learn more powerful non-linear kernels from data. Multiple Kernel Learning (MKL) [3] operates
on a parametric family of kernels: it learns a convex combination of a few base kernels so as to
maximize classification accuracy. Recent advances now allow to combine thousands of kernels in
MKL [17] or exploit specialized families of kernels to derive faster algorithms [19]. In that work, the
authors combine binary base kernels based on randomized indicator functions but restricted them
to XNOR-like kernels. Our QK framework can also be seen as an efficient and robust MKL on
a specific family of binary base kernels. However, our binary base kernels originate from more
general quantizations: they correspond to their regions of constantness. As a consequence, the
resulting optimization problem is also more involves and thus calls for approximate solutions.
In parallel to MKL approaches, Non-Parametric Kernel Learning (NPKL) [10] has emerged as a
flexible kernel learning alternative. Without any assumption on the form of the kernel, these methods
aim at learning the Gram matrix of the data directly. The optimization problem is a semi-definite
program whose size is quadratic in the number of samples. Scalability is therefore an issue, and
approximation techniques must be used to compute the kernel on unobserved data. Like NPKL, we
learn the values of the kernel matrix directly. However, we do it in the index space instead of the
1
Metric learning can be kernelized, but then one has to choose the kernel.
2
original space. Hence, we restrict our family of kernels to piecewise constant ones2 , but, contrary to
NPKL, the complexity of the problems we solve does not grow with the number of data points but
with the refinement of the quantization and our kernels trivially generalize to unobserved inputs.
3
Quantized kernels
In this section, we present the framework of quantized kernels (QK). We start in Sec. 3.1 by defining
QKs and looking at some of their properties. We then present in Sec. 3.2 a general alternating
learning algorithm. A key step is to optimize the quantizer itself. We present in Sec. 3.3 our scheme
for quantization optimization for a single dimensional feature and how to generalize it to higher
dimensions in Sec. 3.4.
3.1
Definition and properties
D
D
Formally, quantized kernels QKD
N are the set of kernels kq on R ?R such that:
?q : RD 7? {1, . . . , N },
?K ? RN ?N 0,
?x, y ? RD ,
kq (x, y) = K(q(x), q(y)), (1)
where q is a quantization function which projects x ? RD to the finite index space {1, . . . , N },
and K 0 denotes that K is a positive semi-definite (PSD) matrix. As discussed above, quantized
kernels are an efficient parametrization of piecewise constant functions, where q defines the regions
of constantness. Moreover, the N ? N matrix K is unique for a given choice of kq , as it simply
accounts for the N (N+1)/2 possible values of the kernel and is the Gram matrix of the N elements
of the index space. We can also see q as a 1-of-N coding feature map ?q , such that:
kq (x, y) = K(q(x), q(y)) = ?q (x)> K?q (y).
(2)
The components of the matrix K fully parametrize the family of quantized kernels based on q, and
it is a PSD matrix if and only if kq is a PSD kernel. An explicit feature mapping of kq is easily
computed from the Cholesky decomposition of the PSD matrix K = P> P:
kq (x, y) = ?q (x)> K?q (y) = ?qP (x), ?qP (y) ,
(3)
where ?qP (x) = P?q (x). It is of particular interest to limit the rank N 0 ? N of K, and hence the
number of rows in P. In their compressed form, vectors require only log2 (N ) bits of memory for
0
storing q(x) and they can be decompressed in RN using P?q (x). Not only is this decompressed
vector smaller than one based on ?q , but it is also associated with the Euclidean geometry rather than
the kernel one. This allows the exploitation of the large literature of efficient methods specialized to
Euclidean spaces.
3.2 Learning quantized kernels
In this section, we describe a general alternating algorithm to learn a quantized kernel kq for feature
matching. This problem can be formulated as quadruple-wise constraints of the following form:
kq (x, y) > kq (u, v),
?(x, y) ? P,
?(u, v) ? N ,
(4)
where P denotes the set of positive feature pairs, and N is the negative one. The positive set contains
feature pairs that should be visually matched, while the negative pairs are mismatches.
We adopt a large-margin formulation of the above constraints using the trace-norm regularization
k ? k? on K, which is the tightest convex surrogate to low-rank regularization [8]. Using M training
pairs {(xj , yj )}j=1...M , we obtain the following optimization problem:
argmin
K0, q?QD
N
E(K, q) =
M
X
?
kKk? +
max 0, 1 ? lj ?q (xj )> K?q (yj ) ,
2
j=1
(5)
D
where QD
N denotes the set of quantizers q : R 7? {1, . . . , N }, the pair label lj ? {?1, 1} denotes
whether the feature pair (xj , yj ) is in N or P respectively. The parameter ? controls the trade-off
between the regularization and the empirical loss. Solving Eq. (5) directly is intractable. We thus
propose to alternate between the optimization of K and q. We describe the former below, and the
latter in the next section.
2
As any continuous function on an interval is the uniform limit of a series of piecewise constant functions,
this assumption does not inherently limit the flexibility of the family.
3
Optimizing K with fixed q. When fixing q in Eq. (5), the objective function becomes convex in
K but is not differentiable, so we resort to stochastic sub-gradient descent for optimization. Similar
to [21], we used Regularised Dual Averaging (RDA) [27] to optimize K iteratively. At iteration
t + 1, the kernel matrix Kt+1 is updated with the following rule:
?
t
Kt+1 = ? ?
Gt + ?I
?
(6)
Pt
where ? > 0 and Gt = 1t t0 =1 Gt0 is the rolling average of subgradients Gt0 of the loss computed
at step t0 from one sample pair. I is the identity matrix and ? is the projection onto the PSD cone.
3.3 Interval quantization optimization for a single dimension
To optimize an objective like Eq. (5) when K is fixed, we must consider how to design and
parametrize the elements of QD
N . In this work, we adopt interval quantizers, and in this section
we assume D = 1, i.e., restrict the study of quantization to R.
Interval quantizers. An interval quantizer q over R is defined by a set of N + 1 boundaries
bi ? R with b0 = ??, bN = ? and q(x) = i if and only if bi?1 < x ? bi . Importantly, interval
quantizers are monotonous, x ? y ? q(x) ? q(y), and boundaries bi can be set to any value between
maxq(x)=i x (included) and minq(x)=i+1 x (excluded). Therefore, Eq. (5) can be viewed as a data
labelling problem, where each value xj or yj takes a label in [1, N ], with a monotonicity constraint.
Thus, let us now consider the graph (V, E) where nodes V = {vt }t=1...2M represent the list of all
xj and yj in a sorted order and the edges E = {(vs , vt )} connect all pairs (xj , yj ). Then Eq. (5) with
fixed K is equivalent to the following discrete pairwise energy minimization problem:
argmin
q?[1,N ]2M
E 0 (q) =
X
Est (q(vs ), q(vt )) +
2M
X
Ct (q(vt?1 ), q(vt )),
(7)
t=2
(s,t)?E
where Est (q(vs ), q(vt )) = Ej (q(xj ), q(yj )) = max (0, 1 ? lj K(q(xj ), q(yj ))) and Ct is ? for
q(vt ) < q(vt?1 ) and 0 otherwise (i.e., it encodes the monotonicity of q in the sorted list of vt ).
The optimization of Eq. (7) is an NP-hard problem as the energies Est are arbitrary and the graph
does not have a bounded treewidth, in general. Hence, we iterate the individual optimization of each
of the boundaries using an exact linear-time algorithm, which we present below.
Exact linear-time optimization of a binary interval quantizer. We now consider solving equations of the form of Eq. (7) for the binary label case (N = 2). The main observation is that the
monotonicity constraint means that labels are 1 until a certain node t and then 2 from node t + 1,
and this switch can occur only once on the entire sequence, where vt ? b1 < vt+1 . This means
that there are only 2M +1 possible labellings and we can order them from (1, . . . , 1), (1, . . . , 1, 2)
to (2, . . . , 2). A na??ve algorithm consists in computing the 2M +1 energies explicitly. Since each
energy computation is linear in the number of edges, this results in a quadratic complexity overall.
A linear-time algorithm exist. It stems from the observation that the energies of two consecutive
labellings (e.g., switching the label of vt from 1 to 2) differ only by a constant number of terms:
E(q(vt?1 ) = 1, q(vt ) = 2, q(vt+1 ) = 2) = E(q(vt?1 ) = 1, q(vt ) = 1, q(vt+1 ) = 2)
+ Ct (1, 2) ? Ct (1, 1) + Ct+1 (2, 2) ? Ct+1 (1, 2) + Est (q(vs ), 2) ? Est (q(vs ), 1)
(8)
where, w.l.o.g., we have assumed (s, t) ? E. After finding the optimal labelling, i.e. finding the
label change (vt , vt+1 ), we set b1 = (vt +vt+1 )/2 to obtain the best possible generalization.
Finite spaces. When the input feature space has a finite number of different values (e.g., x ?
[1, T ]), then we can use linear-time sorting and merge all nodes with equal value in Eq. (7): this
results in considering at most T + 1 labellings, which is potentially much smaller than 2M + 1.
Extension to the multilabel case. Optimizing a single boundary bi of a multilabel interval quantization is essentially the same binary problem as above, where we limit the optimization to the
values currently assigned to i and i + 1 and keep the other assignments q fixed. We use unaries
Ej (q(xj ), q(yj )) or Ej (q(xj ), q(yj )) to model half-fixed pairs for xj or yj , respectively.
3.4 Learning higher dimensional quantized kernels
We now want to generalize interval quantizers to higher dimensions. This is readily feasible via
product quantization [13], using interval quantizers for each individual dimension.
4
Interval product quantization. An interval product quantizer q(x) : RD 7? {1, . . . , N } is of
the form q(x) = (q1 (x1 ), . . . , qD (xD )), where q1 , . . . , qD are interval quantizers with N1 , . . . , ND
QD
bins respectively, i.e., N = d=1 Nd . The learning algorithm devised above trivially generalizes
to interval product quantization by fixing all but one boundary of a single component quantizer qd .
However, learning K ? RN ? RN when N is very large becomes problematic: not only does RDA
scale unfavourably, but the lack of training data will eventually lead to severe overfitting. To address
these issues, we devise below variants of QKs that have practical advantages for robust learning.
Additive quantized kernels (AQK). We can drastically reduce the number of parameters by restricting product quantized kernels to additive ones, which consists in decomposing over dimensions:
kq (x, y) =
D
X
d=1
kqd (xd , yd ) =
D
X
?qd (xd )> Kd ?qd (yd ) = ?q (x)> K?q (y),
(9)
d=1
where qd ? Q1Nd , ?qd is the 1-of-Nd coding of dimension d, Kd is theP
Nd ? Nd Gram
P matrix of
dimension d, ?q is the concatenation of the D mappings ?qd , and K is a ( d Nd )?( d Nd ) blockdiagonal matrix of K
Q1 , . . . , KD . ThePbenefits of AQK are twofold. First, the explicit feature space is
reduced
N = d Nd to N 0 = d Nd . Second, the number of parameters to learn
P from
Pin K is now
only d Nd2 instead of N 2 . The compression ratio is unchanged since log2 (N ) = d log2 (Nd ).
To learn K in Eq. (9), we simply set the off-block-diagonal elements of Gt0 to zero in each iteration,
and iteratively update K as describe in Sec. 3.2. To optimize a product quantizer, we iterate the
optimization of each 1d quantizer qd following Sec. 3.3, while fixing qc for c 6= d. This leads to
using the following energy Ej for a pair (xj , yj ):
where ?j,d
Ej,d (qd (xj,d ), qd (yj,d )) = max (0, ?j,d ? lj Kd (qd (xj,d ), qd (yj,d ))) ,
P
= 1 ? lj c6=d Kc (qc (xc ), qc (yc )) acts as an adaptive margin.
(10)
Block quantized kernels (BQK). Although the additive assumption in AQK greatly reduces the
number of parameters, it is also very restrictive, as it assumes independent data dimensions. A
simple way to extend additive quantized kernels to model the inter-dependencies of dimensions is
to allow the off-diagonal elements of K in Eq. (9) to be nonzero. As a trade-off between a blockdiagonal (AQK) and a full matrix, in this work we also consider the grouping of the feature dimensions into B blocks, and only learn off-block-diagonal elements within each block, leading to Block
Quantized Kernels (BQK). In this way, assuming ?d Nd = n, the number of parameters in K is
B times smaller than for the full matrix. As a matter of fact, many features such as SIFT descriptors
exhibit block structure. SIFT is composed of a 4?4 grid of 8 orientation bins. Components within
the same spatial cell correlate more strongly than others and, thus, only modeling those jointly may
prove sufficient. The optimization of K and q are straightforwardly adapted from the AQK case.
Additional parameter sharing. Commonly, the different dimensions of a descriptor are generated by the same procedure and hence share similar properties. This results in block matrices
K1 , . . . , KD in AQK that are quite similar as well. We propose to exploit this observation and share
the kernel matrix for groups of dimensions, further reducing the number of parameters. Specifically,
we cluster dimensions based on their variances into G equally sized groups and use a single block
matrix for each group. During optimization,
dimensions sharingPthe same block matrix can conP
veniently be merged, i.e. ?q (x) = [ d s.t. Kd =K0 ?qd (xd ), . . . , d s.t. Kd =K0 ?qd (xd )], and then
1
G
K = diag(K01 , . . . , K0G ) is learnt following the procedure already described for AQK. Notably, the
quantizers themselves are not shared, so the kernel still adapts uniquely to every dimension of the
data, and the optimization of quantizers is not changed either. This parameter sharing strategy can
be readily applied to BQK as well.
4
Results
We now present our experimental results, starting with a description of our protocol. We then explore
parameters and properties of our kernels (optimization of quantizers, explicit feature maps). Finally,
we compare to the state-of-the-art in performance and compactness.
Dataset and evaluation protocol. We evaluate our method using the dataset of Brown et al. [5].
It contains three sets of patches extracted from Liberty, Notre Dame and Yosemite using the Difference of Gaussians (DoG) interest point detector. The patches are rectified with respect to the scale
5
Table 1: Impact of quantization optimization for different quantization
strategies
FPR @ 95% recall [%]
Optimized
21.68
25.70
14.29
FPR @ 95% recall [%]
Uniform
Adaptive
Adaptive+
Initial
24.84
25.99
14.62
20
18
16
14
2
6
10
14
#intervals
18
1
2
3
4
5
6
7
8
50
50
100
150
200
250
100 150 200 250
0.4
0.3
0.2
0.1
0
?0.1
?0.2
?0.3
?0.4
50
SIFT[15]
14
12
PR-proj[18]
10
8
1
SQ-4-DAISY[4]
2
3
#groups
4
Figure 1: Impact of N , the num- Figure 2: Impact of G, the number of quantization intervals
1 2 3 4 5 6 7 8
16
100
150
200
250
0.4
0.3
0.2
0.1
0
?0.1
?0.2
?0.3
?0.4
50
100
150
200
ber of dimension groups
250
1
2
3
4
5
6
7
8
1 2 3 4 5 6 7 8
1
2
3
4
5
6
7
8
1 2 3 4 5 6 7 8
0.1
0.08
0.06
0.04
0.02
0
?0.02
?0.04
?0.06
?0.08
(a)
(b)
(c)
(d)
(e)
(f)
Figure 3: Our learned feature maps and additive quantized kernel of a single dimension. (a) shows the quantized kernel in index space, while (b) is in the original feature space for the first quantizer. (c,d) show the two
corresponding feature maps, and (e,f) the related rank-1 kernels.
and dominant orientation, and pairwise correspondences are computed using a multi-view stereo
algorithm. In our experiments, we use the standard evaluation protocol [5] and state-of-the-art descriptors: SIFT [15], PR-proj [21] and SQ-4-DAISY [4]. M =500k feature pairs are used for training
on each dataset, with as many positives as negatives. We report the false positive rate (FPR) at 95%
recall on the test set of 100k pairs. A challenge for this dataset is the bias in local patch appearance
for each set, so a key factor for performance is the ability to generalize and adapt across sets.
Below, in absence of other mention, AQKs are trained for SIFT on Yosemite and tested on Liberty.
Interval quantization and optimization. We first study the influence of initialization and optimization on the generalization ability of the interval quantizers. For initialization, we have used two
different schemes: a) Uniform quantization, i.e. the quantization with equal intervals; b) Adaptive
quantization, i.e. the quantization with intervals with equal number of samples. In both cases, it
allows to learn a first kernel matrix, and we can then iterate with boundary optimization (Sec. 3.3).
Typically, convergence is very fast (2-3 iterations) and takes less than 5 minutes in total (i.e., about
2s per feature dimension) with 1M nodes. We see in Table 1 that uniform binning outperforms the
adaptive one and that further optimization benefits the uniform case more. This may seem paradoxical at first, but this is due to the train/test bias problem: intervals with equal number of samples
are very different across sets, so refinements will not transfer well. Hence, following [7], we first
normalize the features with respect to their rank, separately for the training and test sets. We refer to
this process as Adaptive+. As Table 1 shows, not only does it bring a significant improvement, but
further optimization of the quantization boundaries is more beneficial than for the Adaptive case. In
the following, we thus adopt this strategy.
Number of quantization intervals. In Fig. 1, we show the impact of the number of intervals N
of the quantizer on the matching accuracy, using a single shared kernel submatrix (G = 1). This
number balances the flexibility of the model and its compression ratio. As we can see, using too few
intervals limits the performance of QK, and using too many eventually leads to overfitting. The best
performance for SIFT is obtained with between 8 and 16 intervals.
Explicit feature maps. Fig. 3a shows the additive quantized kernel learnt for SIFT with N = 8
and G = 1. Interestingly, the kernel has negative values far from the diagonal and positive values
near the diagonal. This is typical of stationary kernels: when both features have similar values,
they contribute more to the similarity. However, contrary to stationary kernels, diagonal elements
are far from being constant. There is a mode on small values and another one on large ones. The
second one is stronger: i.e., the co-occurrence of large values yields greater similarity. This is consistent with the voting nature of SIFT descriptors, where strong feature presences are both rarer and
more informative than their absences. The negative values far from the diagonal actually penalize
inconsistent observations, thus confirming existing results [12]. Looking at the values in the original space in Fig. 3b, we see that the quantizer has learnt that fine intervals are needed in the lower
6
Descriptor
Kernel
Dimensionality
Train on Yosemite
Train on Notredame
Notredame
Liberty
Yosemite
Liberty
Mean
SIFT[15]
SIFT[15]
SIFT[15]
SIFT[15]
SIFT[15]
Euclidean
?2
AQK(8)
AQK(8)
BQK(8)
128
128
128
256
256
24.02
17.65
10.72
9.26
8.05
31.34
22.84
16.90
14.48
13.31
27.96
23.50
10.72
10.16
9.88
31.34
22.84
16.85
14.43
13.16
28.66
21.71
13.80
12.08
11.10
SQ-4-DAISY [4]
SQ-4-DAISY [4]
SQ-4-DAISY [4]
SQ-4-DAISY [4]
Euclidean
?2
SQ [4]
AQK(8)
1360
1360
1360
?1813
10.08
10.61
8.42
4.96
16.90
16.25
15.58
9.41
10.47
12.19
9.25
5.60
16.90
16.25
15.58
9.77
13.58
13.82
12.21
7.43
Euclidean[21]
AQK(16)
<64
?102
7.11
5.41
14.82
10.90
10.54
7.65
12.88
10.54
11.34
8.63
PR-proj [21]
PR-proj [21]
Table 2: Performance of kernels on different datasets with different descriptors. AQK(N) denotes the additive
quantized kernel with N quantization intervals. Following [6], we report the False positive rate (%) at 95%
recall. The best results for each descriptor are in bold.
values, while larger ones are enough for larger values. This is consistent with previous observations
that the contribution of large values in SIFT should not grow proportionally [2, 18, 14].
In this experiment, the learnt kernel has rank 2. We show in Fig. 3c, 3d, 3e and 3f the corresponding
feature mappings and their associated rank 1 kernels. The map for the largest eigenvalue (Fig. 3c)
is monotonous but starts with negative values. This impacts dot product significantly, and accounts
for the above observation that negative similarities occur when inputs disagree. This rank 1 kernel
cannot allot enough contribution to similar mid-range values. This is compensated by the second
rank (Fig. 3f).
Number of groups. Fig. 2 now shows the influence of the number of groups G on performance,
for the three different descriptors (N = 8 for SIFT and SQ-4-DAISY, N = 16 for PR-proj). As for
intervals, using more groups adds flexibility to the model, but as less data is available to learn each
parameter, over-fitting will hurt performance. We choose G = 3 for the rest of the experiments.
Comparison to the state of the art. Table 2 reports the matching performance of different kernels
using different descriptors, for all sets, as well as the dimensionality of the corresponding explicit
feature maps. For all three descriptors and on all sets, our quantized kernels significantly and consistently outperform the best reported result in the literature. Indeed, AQK improves the mean error
rate at 95% recall from 28.66% to 12.08% for SIFT, from 13.58% to 7.43% for SQ-4-DAISY and
from 11.34% to 8.63% for PR-proj compared to the Euclidean distance, and about as much for the
?2 kernel. Note that PR-proj already integrates metric learning in its design ([21] thus recommends
using the Euclidean distance): as a consequence our experiments show that modelling non-linearities
can bring significant improvements. When comparing to sparse quantization (SQ) with hamming
distance as done in [4], the error is significantly reduced from 12.21% to 7.43%. This is a notable
achievement considering that [4] is the previous state of the art.
The SIFT descriptor has a grid block design which makes it particularly suited for the use of BQK.
Hence, we also evaluated our BQK variant for that descriptor. With BQK(8), we observed a relative
improvement of 8%, from 12.08% for AQK(8) to 11.1%.
We provide in Fig. 4 the ROC curves for the three descriptors when training on Yosemite and testing
on Notre Dame and Liberty. These figures show that the improvement in recall is consistent over the
full range of false positive rates. For further comparisons, our data and code are available online.3
Compactness of our kernels. In many applications of feature matching, the compactness of the
descriptor is important. In Table 3, we compare to other methods by grouping them according to
their memory footprint. As a reference, the best method reported in Table 2 (AQK(8) on SQ-4DAISY) uses 4080 bits per descriptor. As expected, error rates increase as fewer bits are used, the
original features being significantly altered. Notably, QKs consistently yield the best performance in
all groups. Even with a crude binary quantization of SQ-4-DAISY, our quantized kernel outperform
the state-of-the-art SQ of [4] by 3 to 4%. When considering the most compact encodings (? 64 bits),
our AQK(2) does not improve over BinBoost [22], a descriptor designed for extreme compactness, or
the product quantization (PQ [13]) encoding as used in [21]. This is because our current framework
does not yet allow for joint compression of multiple dimensions. Hence, it is unable to use less
3
See: http://www.vision.ee.ethz.ch/?qind/QuantizedKernel.html
7
BQK(8)
AQK(8)
AQK(2)
L2
80
75
5
10
15
20
25
False Positive Rate [%]
SIFT
90
85
BQK(8)
AQK(8)
AQK(2)
L2
80
75
5
10
15
90
85
80
AQK(8)
AQK(2)
SQ
75
5
20
25
False Positive Rate [%]
30
10
15
20
25
False Positive Rate [%]
90
85
80
AQK(8)
AQK(2)
SC
75
5
10
15
90
85
80
AQK(16)
AQK(4)
L2
75
5
10
20
25
False Positive Rate [%]
20
25
30
PR?proj
95
90
85
80
AQK(16)
AQK(4)
L2
75
70
0
30
15
False Positive Rate [%]
100
95
70
0
95
70
0
30
SQ-4-DAISY
100
95
70
0
95
70
0
30
PR?proj
100
True Positive Rate [%]
85
100
True Positive Rate [%]
True Positive Rate [%]
90
True Positive Rate [%]
True Positive Rate [%]
95
70
0
SQ-4-DAISY
100
True Positive Rate [%]
SIFT
100
5
10
15
20
25
False Positive Rate [%]
30
Figure 4: ROC curves when evaluating Notre Dame (top) and Liberty (bottom) from Yosemite
Descriptor
Encoding
Memory (bits)
Train on Yosemite
Train on Notredame
Notredame
Liberty
Yosemite
Liberty
Mean
SQ-4-DAISY [4]
SQ-4-DAISY [4]
SQ [4]
AQK(2)
1360
1360
8.42
5.86
15.58
10.81
9.25
6.36
15.58
10.94
12.21
8.49
SIFT[15]
PR-proj [21]
PR-proj [21]
AQK(8)
Bin [21]
AQK(16)
384
1024
<256
9.26
7.09
5.41
14.48
15.15
10.90
10.16
8.5
7.65
14.43
12.16
10.54
12.08
10.73
8.63
SIFT[15]
PR-proj [21]
PR-proj [21]
AQK(2)
Bin [21]
AQK(4)
128
128
<128
14.62
10.00
7.18
19.72
18.64
13.02
15.65
13.41
10.29
19.45
16.39
13.18
17.36
14.61
10.92
BinBoost[22]
PR-proj [21]
PR-proj [21]
PR-proj [21]
BinBoost[22]
AQK(2)
PQ [21]
PCA+AQK(4)
64
<64
64
64
14.54
14.80
12.91
10.74
21.67
20.59
20.15
17.46
18.97
19.38
19.32
14.44
20.49
22.24
17.97
17.60
18.92
19.26
17.59
15.06
Table 3: Performance comparison of different compact feature encoding. The number in the table is
reported as False positive rate (%) at 95% recall. The best results for each group are in bold.
than 1 bit per original dimension, and is not optimal in that case. To better understand the potential
benefits of decorrelating features and joint compression in future work, we pre-processed the data
with PCA, projecting to 32 dimensions and then using AQK(4). This simple procedure obtained
state-of-the-art performance with 15% error rate, now outperforming [22] and [21].
Although QKs yield very compact descriptors and achieve the best performance across many experimental setups, the computation of similarity values is slower than for competitors: in the binary
case, we double the complexity of hamming distance for the 2 ? 2 table look-up.
5
Conclusion
In this paper, we have introduced the simple yet powerful family of quantized kernels (QK), and
presented an efficient algorithm to learn its parameters, i.e. the kernel matrix and the quantization
boundaries. Despite their apparent simplicity, QKs have numerous advantages: they are very flexible, can model non-linearities in the data and provide explicit low-dimensional feature mappings
that grant access to the Euclidean geometry. Above all, they achieve state-of-the-art performance
on the main visual feature matching benchmark. We think that QKs have a lot of potential for further improvements. In future work, we want to explore new learning algorithms to obtain higher
compression ratios ? e.g. by jointly compressing feature dimensions ? and find the weight sharing
patterns that would further improve the matching performance automatically.
Acknowledgements
We gratefully thank the KIC-Climate project Modeling City Systems.
8
References
[1] Sameer Agarwal, Yasutaka Furukawa, Noah Snavely, Ian Simon, Brian Curless, Steven M Seitz, and
Richard Szeliski. Building rome in a day. Communications of the ACM, 54(10):105?112, 2011.
[2] Relja Arandjelovic and Andrew Zisserman. Three things everyone should know to improve object retrieval. In IEEE Conference onComputer Vision and Pattern Recognition (CVPR), 2012.
[3] Francis R Bach, Gert RG Lanckriet, and Michael I Jordan. Multiple kernel learning, conic duality, and
the smo algorithm. In Proceedings of the International Conference on Machine learning. ACM, 2004.
[4] Xavier Boix, Michael Gygli, Gemma Roig, and Luc Van Gool. Sparse quantization for patch description.
In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2013.
[5] Matthew Brown, Gang Hua, and Simon Winder. Discriminative learning of local image descriptors.
Pattern Analysis and Machine Intelligence, IEEE Transactions on, 33(1):43?57, 2011.
[6] Matthew Brown and David G Lowe. Automatic panoramic image stitching using invariant features.
International Journal of Computer Vision, 74(1):59?73, 2007.
[7] Thomas Dean, Mark A Ruzon, Mark Segal, Jonathon Shlens, Sudheendra Vijayanarasimhan, and Jay
Yagnik. Fast, accurate detection of 100,000 object classes on a single machine. In CVPR, 2013.
[8] Maryam Fazel. Matrix rank minimization with applications. PhD thesis, 2002.
[9] Yunchao Gong, Sanjiv Kumar, Vishal Verma, and Svetlana Lazebnik. Angular quantization-based binary
codes for fast similarity search. In NIPS, pages 1196?1204, 2012.
[10] Steven CH Hoi, Rong Jin, and Michael R Lyu. Learning nonparametric kernel matrices from pairwise
constraints. In Proceedings of the International Conference on Machine learning. ACM, 2007.
[11] Gang Hua, Matthew Brown, and Simon Winder. Discriminant embedding for local image descriptors. In
ICCV 2007. IEEE, 2007.
[12] Herv?e J?egou and Ond?rej Chum. Negative evidences and co-occurences in image retrieval: The benefit of
pca and whitening. In Computer Vision?ECCV 2012, pages 774?787. Springer, 2012.
[13] Herve Jegou, Matthijs Douze, and Cordelia Schmid. Product quantization for nearest neighbor search.
Pattern Analysis and Machine Intelligence, IEEE Transactions on, 33(1):117?128, 2011.
[14] Herv?e J?egou, Matthijs Douze, Cordelia Schmid, and Patrick P?erez. Aggregating local descriptors into a
compact image representation. In CVPR, pages 3304?3311. IEEE, 2010.
[15] David G Lowe. Distinctive image features from scale-invariant keypoints. IJCV, 2004.
[16] Subhransu Maji, Alexander C Berg, and Jitendra Malik. Efficient classification for additive kernel svms.
IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(1):66?77, 2013.
[17] Francesco Orabona and Luo Jie. Ultra-fast optimization algorithm for sparse multi kernel learning. In Proceedings of the 28th International Conference on Machine Learning (ICML-11), pages 249?256, 2011.
[18] Florent Perronnin, Jorge S?anchez, and Thomas Mensink. Improving the fisher kernel for large-scale image
classification. In European Conference on Computer Vision (ECCV). 2010.
[19] Gemma Roig, Xavier Boix, and Luc Van Gool. Random binary mappings for kernel learning and efficient
SVM. arXiv preprint arXiv:1307.5161, 2013.
[20] Dimitris Achlioptas Frank McSherry Bernhard Scholkopf. Sampling techniques for kernel methods. In
NIPS 2001, volume 1, page 335. MIT Press, 2002.
[21] Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. Learning local feature descriptors using convex optimisation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2014.
[22] Tomasz Trzcinski, Mario Christoudias, Pascal Fua, and Vincent Lepetit. Boosting binary keypoint descriptors. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2013.
[23] Tomasz Trzcinski, Mario Christoudias, Vincent Lepetit, and Pascal Fua. Learning image descriptors with
the boosting-trick. In NIPS, 2012.
[24] Andrea Vedaldi and Andrew Zisserman. Efficient additive kernels via explicit feature maps. Pattern
Analysis and Machine Intelligence, IEEE Transactions on, 34(3):480?492, 2012.
[25] Kilian Weinberger, John Blitzer, and Lawrence Saul. Distance metric learning for large margin nearest
neighbor classification. Advances in neural information processing systems, 18:1473, 2006.
[26] Simon AJ Winder and Matthew Brown. Learning local image descriptors. In CVPR, 2007.
[27] Lin Xiao et al. Dual averaging methods for regularized stochastic learning and online optimization.
Journal of Machine Learning Research, 11(2543-2596):4, 2010.
[28] Zheng Yi, Cao Zhiguo, and Xiao Yang. Multi-spectral remote image registration based on sift. Electronics
Letters, 44(2):107?108, 2008.
9
| 5412 |@word exploitation:1 version:1 compression:5 norm:1 stronger:1 nd:11 seitz:1 bn:1 accounting:1 decomposition:1 q1:3 pick:1 egou:2 mention:1 lepetit:2 electronics:1 initial:1 contains:2 series:1 interestingly:1 outperforms:1 existing:2 current:1 comparing:1 luo:1 yet:4 must:2 readily:2 john:1 additive:9 sanjiv:1 informative:1 confirming:1 designed:1 update:1 discrimination:1 v:5 half:1 stationary:2 fewer:1 intelligence:5 fpr:3 parametrization:2 core:1 num:1 quantizer:12 quantized:23 boosting:3 node:5 contribute:1 c6:1 scholkopf:1 consists:2 prove:1 ijcv:1 combine:3 fitting:1 hellinger:1 introduce:2 pairwise:3 inter:1 notably:2 indeed:1 expected:1 andrea:2 themselves:1 unaries:1 behavior:1 multi:3 jegou:1 automatically:4 considering:3 increasing:1 becomes:2 project:4 linearity:3 moreover:1 matched:1 bounded:1 argmin:2 unobserved:2 transformation:1 finding:2 every:1 act:1 voting:1 xd:5 preferable:1 control:1 grant:2 positive:22 local:9 aggregating:1 limit:6 consequence:2 switching:1 despite:2 encoding:5 quadruple:1 merge:1 yd:2 initialization:2 studied:1 co:2 limited:1 range:3 bi:5 fazel:1 practical:2 responsible:1 unique:1 yj:14 testing:1 ond:1 block:11 definite:3 sq:19 procedure:3 footprint:1 empirical:1 eth:3 significantly:4 vedaldi:2 matching:17 projection:3 pre:2 sudheendra:1 onto:1 cannot:1 selection:3 vijayanarasimhan:1 influence:2 optimize:4 equivalent:2 map:8 www:1 compensated:1 dean:1 urich:3 starting:1 minq:1 convex:6 qc:3 simplicity:1 matthieu:1 occurences:1 rule:1 importantly:1 shlens:1 embedding:1 gert:1 hurt:1 gygli:1 updated:1 roig:2 pt:1 heavily:1 user:2 exact:3 us:1 designing:1 regularised:1 lanckriet:1 trick:1 element:8 recognition:3 particularly:1 binning:1 observed:1 bottom:1 steven:2 preprint:1 thousand:1 region:2 compressing:2 kilian:1 remote:1 trade:2 complexity:4 multilabel:2 trained:1 solving:2 distinctive:1 gt0:3 easily:3 joint:2 k0:2 represented:1 maji:1 train:5 fast:4 describe:3 sc:1 hyper:1 choosing:1 whose:1 emerged:1 widely:1 solve:1 quite:1 larger:2 apparent:1 otherwise:2 compressed:2 cvpr:6 ability:2 simonyan:1 think:1 jointly:2 itself:1 online:2 sequence:1 differentiable:1 advantage:2 eigenvalue:1 reconstruction:1 propose:3 maryam:1 product:11 douze:2 remainder:1 qin:1 tu:1 cao:1 flexibility:3 achieve:2 adapts:1 k01:1 description:4 intuitive:1 christoudias:2 normalize:1 scalability:2 achievement:1 convergence:1 cluster:1 double:1 gemma:2 object:2 derive:1 andrew:3 blitzer:1 gong:1 fixing:3 nearest:4 b0:1 eq:10 strong:1 involves:1 treewidth:1 qd:19 differ:1 liberty:8 merged:1 stochastic:2 jonathon:1 hoi:1 bin:4 require:1 generalization:2 rda:2 brian:1 ultra:1 exploring:1 extension:1 rong:1 deciding:1 visually:1 lawrence:1 mapping:7 lyu:1 matthew:4 early:1 adopt:3 consecutive:1 integrates:1 label:6 currently:1 largest:1 city:1 minimization:2 bqk:9 mit:1 gaussian:1 aim:1 rather:1 ej:5 improvement:7 nd2:1 rank:9 consistently:2 panoramic:1 modelling:1 greatly:2 inference:1 perronnin:1 lj:5 entire:1 typically:1 compactness:4 kernelized:1 kc:1 proj:16 subroutine:1 subhransu:1 issue:2 dual:3 flexible:4 classification:4 orientation:2 html:1 pascal:2 overall:1 art:9 spatial:1 equal:4 once:1 cordelia:2 sampling:1 look:1 uncompressed:1 icml:1 photometric:1 future:2 np:1 others:1 piecewise:5 inherent:1 few:4 report:3 richard:1 oriented:1 composed:1 ve:1 homogeneity:2 individual:2 qks:10 geometry:5 n1:1 attempt:1 psd:5 detection:1 stationarity:2 interest:4 zheng:1 evaluation:2 severe:1 alignment:1 notre:3 extreme:1 mcsherry:1 kt:2 accurate:1 edge:2 herve:1 euclidean:15 monotonous:2 witnessed:1 instance:2 modeling:2 yosemite:8 assignment:1 rolling:1 kq:11 uniform:5 successful:1 inadequate:2 too:2 reported:3 straightforwardly:1 connect:1 dependency:1 learnt:5 international:4 randomized:1 matthijs:2 off:5 michael:3 na:1 thesis:1 choose:2 allot:1 resort:1 leading:1 winder:3 account:2 potential:3 segal:1 de:1 sec:13 coding:2 bold:2 matter:1 jitendra:1 notable:1 explicitly:1 depends:1 later:1 view:1 lot:1 lowe:2 mario:2 francis:1 start:2 parallel:1 tomasz:2 simon:4 daisy:14 contribution:3 accuracy:3 qk:7 descriptor:34 efficiently:1 variance:1 yield:6 correspond:1 generalize:4 vincent:2 curless:1 rectified:1 detector:1 sharing:3 guillaumin:2 definition:1 competitor:1 energy:6 associated:2 hamming:2 dataset:5 proved:1 kic:1 recall:7 knowledge:1 improves:3 dimensionality:2 actually:1 focusing:1 tum:1 higher:5 supervised:1 day:1 zisserman:3 improved:1 fua:2 formulation:1 done:1 notredame:4 strongly:1 evaluated:1 decorrelating:1 just:1 angular:1 mensink:1 achlioptas:1 correlation:1 until:1 hand:1 lack:1 mkl:5 defines:1 mode:1 aj:1 building:1 brown:5 true:6 former:1 hence:8 regularization:3 assigned:1 alternating:2 excluded:1 iteratively:2 nonzero:1 xavier:2 xnor:1 climate:1 mahalanobis:1 during:1 uniquely:1 essence:1 motion:1 bring:2 image:14 wise:1 lazebnik:1 recently:1 common:1 specialized:2 qp:3 volume:1 extend:2 discussed:1 qkd:1 refer:2 significant:3 tuning:2 automatic:2 trivially:2 rd:4 similarly:1 grid:2 erez:1 arandjelovic:1 gratefully:1 dot:3 pq:2 access:2 similarity:15 whitening:1 etc:1 base:5 gt:2 dominant:1 add:1 patrick:1 recent:1 optimizing:3 certain:1 binary:11 outperforming:1 yagnik:1 vt:22 jorge:1 yi:1 exploited:1 devise:2 furukawa:1 seen:1 additional:1 greater:1 maximize:3 semi:3 relates:1 multiple:5 full:3 keypoints:1 reduces:1 stem:1 sameer:1 match:1 faster:1 adapt:1 bach:1 retrieval:3 lin:1 devised:1 equally:1 impact:6 variant:2 basic:1 optimisation:1 vision:10 metric:7 essentially:1 arxiv:2 histogram:1 kernel:99 represent:2 iteration:3 agarwal:1 cell:1 penalize:1 want:2 separately:1 fine:1 interval:28 grow:2 crucial:1 rest:1 probably:1 pooling:1 thing:1 contrary:2 inconsistent:1 seem:1 jordan:1 call:1 ee:2 near:1 yang:1 presence:1 ideal:1 embeddings:1 enough:2 recommends:1 iterate:3 xj:14 fit:1 switch:1 perfectly:1 restrict:2 florent:1 reduce:2 t0:2 whether:2 herv:2 pca:3 stereo:1 suffer:2 karen:1 jie:1 generally:1 proportionally:1 amount:2 nonparametric:1 mid:1 extensively:1 processed:1 svms:1 decompressed:3 reduced:2 http:1 outperform:2 exist:1 problematic:2 chum:1 per:4 discrete:1 group:10 key:2 registration:3 vangool:1 vast:2 graph:2 cone:1 letter:1 powerful:3 svetlana:1 family:10 patch:4 bit:8 submatrix:1 ct:6 dame:3 correspondence:2 quadratic:2 adapted:2 occur:2 noah:1 constraint:5 gang:2 encodes:1 aspect:1 kumar:1 subgradients:1 munich:1 according:1 alternate:1 combination:1 kd:7 smaller:4 across:3 beneficial:1 labellings:3 making:1 projecting:2 iccv:1 restricted:1 pr:16 invariant:2 equation:1 discus:1 eventually:2 pin:1 needed:1 know:1 tractable:1 stitching:2 available:3 parametrize:2 tightest:1 generalizes:1 decomposing:1 gaussians:1 generic:1 spectral:1 occurrence:1 ruzon:1 alternative:1 weinberger:1 slower:1 rej:1 original:7 compress:2 denotes:5 assumes:1 top:1 thomas:2 log2:3 paradoxical:1 xc:1 exploit:3 restrictive:2 k1:1 build:2 establish:1 unfavourably:1 unchanged:1 objective:3 malik:1 already:2 yunchao:1 parametric:2 strategy:3 snavely:1 diagonal:7 surrogate:1 exhibit:2 gradient:2 distance:7 unable:1 thank:1 concatenation:1 originate:1 discriminant:1 assuming:1 code:2 index:6 ratio:3 balance:1 difficult:2 setup:1 quantizers:12 potentially:2 frank:1 trace:1 negative:8 design:6 perform:1 disagree:1 observation:7 francesco:1 datasets:1 anchez:1 benchmark:3 finite:5 descent:1 jin:1 heterogeneity:1 defining:1 looking:2 communication:1 rome:1 rn:4 arbitrary:3 rarer:1 introduced:1 pair:13 dog:1 specified:1 david:2 optimized:2 smo:1 learned:1 maxq:1 nip:3 address:1 able:1 below:5 pattern:9 mismatch:1 yc:1 dimitris:1 distorsions:1 challenge:1 program:1 including:1 memory:3 max:3 everyone:1 gool:3 force:1 rely:1 regularized:1 indicator:1 scheme:2 improve:4 altered:1 keypoint:1 numerous:1 conic:1 schmid:2 geometric:1 literature:5 blockdiagonal:2 l2:4 acknowledgement:1 relative:1 fully:2 loss:2 limitation:2 conp:1 sufficient:1 consistent:3 trzcinski:2 xiao:2 storing:1 share:2 verma:1 row:1 eccv:2 changed:1 dis:2 drastically:1 bias:2 allow:3 understand:1 ber:1 szeliski:1 neighbor:4 saul:1 sparse:3 van:3 benefit:3 boundary:8 dimension:26 curve:2 gram:4 evaluating:1 author:1 commonly:2 refinement:2 adaptive:7 far:3 cope:1 correlate:1 transaction:5 approximate:3 compact:6 bernhard:1 keep:2 monotonicity:3 overfitting:2 b1:2 assumed:1 discriminative:3 thep:1 search:3 continuous:1 table:10 additionally:1 learn:13 reasonably:2 robust:3 transfer:1 inherently:4 nature:1 improving:1 european:1 protocol:3 diag:1 main:2 x1:1 fig:8 boix:2 roc:2 sub:1 explicit:9 crude:1 jay:1 learns:1 vishal:1 ian:1 minute:1 specific:1 sift:25 kkk:1 list:2 svm:1 evidence:1 grouping:2 intractable:1 quantization:29 restricting:1 false:10 phd:1 labelling:2 margin:3 rankness:1 chen:2 sorting:1 suited:1 rg:1 intersection:2 simply:2 likely:1 explore:2 appearance:1 visual:7 hua:2 springer:1 ch:3 extracted:1 acm:3 sized:2 formulated:1 identity:1 viewed:1 rbf:2 sorted:2 orabona:1 twofold:1 luc:3 fisher:1 shared:2 feasible:1 hard:1 change:1 included:1 kqd:1 specifically:1 operates:1 reducing:1 averaging:2 absence:2 typical:1 total:1 duality:1 experimental:2 est:5 select:1 formally:1 berg:1 cholesky:1 mark:2 latter:1 alexander:1 ethz:2 incorporate:1 evaluate:1 tested:1 |
4,875 | 5,413 | Diverse Sequential Subset Selection for
Supervised Video Summarization
Boqing Gong?
Department of Computer Science
University of Southern California
Los Angeles, CA 90089
[email protected]
Wei-Lun Chao?
Department of Computer Science
University of Southern California
Los Angeles, CA 90089
[email protected]
Kristen Grauman
Department of Computer Science
University of Texas at Austin
Austin, TX 78701
[email protected]
Fei Sha
Department of Computer Science
University of Southern California
Los Angeles, CA 90089
[email protected]
Abstract
Video summarization is a challenging problem with great application potential.
Whereas prior approaches, largely unsupervised in nature, focus on sampling useful frames and assembling them as summaries, we consider video summarization
as a supervised subset selection problem. Our idea is to teach the system to learn
from human-created summaries how to select informative and diverse subsets, so
as to best meet evaluation metrics derived from human-perceived quality. To this
end, we propose the sequential determinantal point process (seqDPP), a probabilistic model for diverse sequential subset selection. Our novel seqDPP heeds the
inherent sequential structures in video data, thus overcoming the deficiency of the
standard DPP, which treats video frames as randomly permutable items. Meanwhile, seqDPP retains the power of modeling diverse subsets, essential for summarization. Our extensive results of summarizing videos from 3 datasets demonstrate
the superior performance of our method, compared to not only existing unsupervised methods but also naive applications of the standard DPP model.
1
Introduction
It is an impressive yet alarming fact that there is far more video being captured?by consumers, scientists, defense analysts, and others?than can ever be watched or browsed efficiently. For example,
144,000 hours of video are uploaded to YouTube daily; lifeloggers with wearable cameras amass
Gigabytes of video daily; 422,000 CCTV cameras perched around London survey happenings in
the city 24/7. With this explosion of video data comes an ever-pressing need to develop automatic
video summarization algorithms. By taking a long video as input and producing a short video (or
keyframe sequence) as output, video summarization has great potential to reign in raw video and
make it substantially more browseable and searchable.
Video summarization methods often pose the problem in terms of subset selection: among all the
frames (subshots) in the video, which key frames (subshots) should be kept in the output summary?
There is a rich literature in computer vision and multimedia developing a variety of ways to answer
this question [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]. Existing techniques explore a plethora of properties that a
good summary should capture, designing criteria that the algorithm should prioritize when deciding
?
Equal contribution
1
which subset of frames (or subshots) to select. These include representativeness (the frames should
depict the main contents of the videos) [1, 2, 10], diversity (they should not be redundant) [4, 11],
interestingness (they should have salient motion/appearance [2, 3, 6] or trackable objects [5, 12, 7]),
or importance (they should contain important objects that drive the visual narrative) [8, 9].
Despite valuable progress in developing the desirable properties of a summary, prior approaches are
impeded by their unsupervised nature. Typically the selection algorithm favors extracting content
that satisfies criteria like the above (diversity, importance, etc.), and performs some sort of frame
clustering to discover events. Unfortunately, this often requires some hand-crafting to combine the
criteria effectively. After all, the success of a summary ultimately depends on human perception.
Furthermore, due to the large number of possible subsets that could be selected, it is difficult to
directly optimize the criteria jointly on the selected frames as a subset; instead, sampling methods
that identify independently useful frames (or subshots) are common.
To address these limitations, we propose to consider video summarization as a supervised subset
selection problem. The main idea is to use examples of human-created summaries?together with
their original source videos?to teach the system how to select informative subsets. In doing so,
we can escape the hand-crafting often necessary for summarization, and instead directly optimize
the (learned) factors that best meet evaluation metrics derived from human-perceived quality. Furthermore, rather than independently select ?high scoring? frames, we aim to capture the interlocked
dependencies between a given frame and all others that could be chosen.
To this end, we propose the sequential determinantal point process (seqDPP), a new probabilistic
model for sequential and diverse subset selection. The determinantal point process (DPP) has recently emerged as a powerful method for selecting a diverse subset from a ?ground set? of items [13],
with applications including document summarization [14] and information retrieval [15]. However,
existing DPP techniques have a fatal modeling flaw if applied to video (or documents) for summarization: they fail to capture their inherent sequential nature. That is, a standard DPP treats the
inputs as bags of randomly permutable items agnostic to any temporal structure. Our novel seqDPP
overcomes this deficiency, making it possible to faithfully represent the temporal dependencies in
video data. At the same time, it lets us pose summarization as a supervised learning problem.
While learning how to summarize from examples sounds appealing, why should it be possible?
particularly if the input videos are expected to vary substantially in their subject matter?1 Unlike
more familiar supervised visual recognition tasks, where test data can be reasonably expected to
look like the training instances, a supervised approach to video summarization must be able to learn
generic properties that transcend the specific content of the training set. For example, the learner
can recover a ?meta-cue? for representativeness, if the input features record profiles of the similarity between a frame and its increasingly distant neighbor frames. Similarly, category-independent
cues about an object?s placement in the frame, the camera person?s active manipulation of viewpoint/zoom, etc., could play a role. In any such case, we can expect the learning algorithm to focus
on those meta-cues that are shared by the human-selected frames in the training set, even though the
subject matter of the videos may differ.
In short, our main contributions are: a novel learning model (seqDPP) for selecting diverse subsets
from a sequence, its application to video summarization (the model is applicable to other sequential
data as well), an extensive empirical study with three benchmark datasets, and a successful firststep/proof-of-concept towards using human-created video summaries for learning to select subsets.
The rest of the paper is organized as follows. In section 2, we review DPP and its application to
document summarization. In section 3, we describe our seqDPP method, followed by a discussion
of related work in section 4. We report results in section 5, then conclude in section 6.
2
Determinantal point process (DPP)
The DPP was first used to characterize the Pauli exclusion principle, which states that two identical particles cannot occupy the same quantum state simultaneously [16]. The notion of exclusion
has made DPP an appealing tool for modeling diversity in application such as document summarization [14, 13], or image search and ranking [17]. In what follows, we give a brief account on DPP and
how to apply it to document summarization where the goal is to generate a summary by selecting
1
After all, not all videos on YouTube are about cats.
2
several sentences from a long document [18, 19]. The selected sentences should be not only diverse
(i.e., different) to reduce the redundancy in the summary, but also representative of the document.
Background Let Y = {1, 2, ? ? ? , N} be a ground set of N items (eg., sentences). In its simplest
form, a DPP defines a discrete probability distribution over all the 2N subsets of Y. Let Y denote
the random variable of selecting a subset. Y is then distributed according to
det(Ly )
P (Y = y) =
(1)
det(L + I)
for y ? Y. The kernel L ? SN?N
is the DPP?s parameter and is constrained to be positive semidefi+
nite. I is the identity matrix. Ly is the principal minor (sub-matrix) with rows and columns selected
according to the indices in y. The determinant function det(?) gives rise to the interesting property
of pairwise repulsion. To see that, consider selecting a subset of two items i and j. We have
P (Y = {i, j}) ? Lii Ljj ? L2ij .
(2)
If the items i and j are the same, then P (Y = {i, j}) = 0 (because Lij = Lii = Ljj ). Namely,
identical items should not appear together in the same set. A more general case also holds: if i and
j are similar to each other, then the probability of observing i and j in a subset together is going to
be less than that of observing either one of them alone (see the excellent tutorial [13] for details).
The most diverse subset of Y is thus the one that attains the highest probability
y ? = arg maxy P (Y = y) = arg maxy det(Ly ),
(3)
?
where y results from MAP inference. This is a NP-hard combinatorial optimization problem.
However, there are several approaches to obtaining approximate solutions [13, 20].
Learning DPPs for document summarization Suppose we model selecting a subset of sentences
as a DPP over all sentences in a document. We are given a set of training samples in the form of
documents (i.e., ground sets) and the ground-truth summaries. How can we discover the underlying
parameter L so as to use it for generating summaries for new documents?
Note that the new documents will likely have sentences that have not been seen before in the training
samples. Thus, the kernel matrix L needs to be reparameterized in order to generalize to unseen
documents. [14] proposed a special reparameterization called quality/diversity decomposition:
1 T
T
Lij = qi ?i ?j qj , qi = exp
? xi ,
(4)
2
where ?i is the normalized TF-IDF vector of the sentence i so that ?Ti ?j computes the cosine angle
between two sentences. The ?quality? feature vector xi encodes the contextual information about i
and its representativeness of other items. In document summarization, xi are the sentence lengths,
positions of the sentences in the texts, and other meta cues. The parameter ? is then optimized with
maximum likelihood estimation (MLE) such that the target subsets have the highest probabilities
X
? ? = arg max?
log P (Y = yn? ; Ln (?)),
(5)
n
where Ln is the L matrix formulated using sentences in the n-th ground set, and yn? is the corresponding ground-truth summary.
Despite its success in document summarization [14], a direct application of DPP to video summarization is problematic. The DPP model is agnostic about the order of the items. For video (and to a
large degree, text data), it does not respect the inherent sequential structures. The second limitation
is that the quality-diversity decomposition, while cleverly leading to a convex optimization, limits
the power of modeling complex dependencies among items. Specifically, only the quality factor qi
is optimized on the training data. We develop new approaches to overcoming those limitations.
3
Approach
In what follows, we describe our approach for video summarization. Our approach contains three
components: (1) a preparatory yet crucial step that generates ground-truth summaries from multiple
human-created ones (section 3.1); (2) a new probabilistic model?the sequential determinantal point
process (seqDPP)?that models the process of sequentially selecting diverse subsets (section 3.2);
(3) a novel way of re-parameterizing seqDPP that enables learning more flexible and powerful representations for subset selection from standard visual and contextual features (section 3.3).
3
Figure 1: The agreement among human-created summaries is high, as is the agreement between the oracle
summary generated by our algorithm (cf. section 3.1) and human annotations.
3.1
Generating ground-truth summaries
The first challenge we need to address is what to provide to our learning algorithm as ground-truth
summaries. In many video datasets, each video is annotated (manually summarized) by multiple
human users. While the users were often well instructed on the annotation task, discrepancies are
expected due to many uncontrollable individual factors such as whether the person was attentive,
idiosyncratic viewing preferences, etc. There are some studies on how to evaluate automatically
generated summaries in the presence of multiple human-created annotations [21, 22, 23]. However,
for learning, our goal is to generate one single ground-truth or ?oracle? summary per video.
Our main idea is to synthesize the oracle summary that maximally agrees with all annotators. Our
hypothesis is that despite the discrepancies, those summaries nonetheless share the common traits of
reflecting the subject matters in the video. These commonalities, to be discovered by our synthesis
algorithm, will provide strong enough signals for our learning algorithm to be successful.
To begin with, we first describe a few metrics in quantifying the agreement in the simplest setting
where there are only two summaries. These metrics will also be used later in our empirical studies
to evaluate various summarization methods. Using those metrics, we then analyze the consistency
of human-created summaries in two video datasets to validate our hypothesis. Finally, we present
our algorithm for synthesizing one single oracle summary per video.
Evaluation metrics Given two video summaries A and B, we measure how much they are in
agreement by first matching their frames, as they might be of different lengths. Following [24], we
compute the pairwise distances between all frames across the two summaries. Two frames are then
?matched? if their visual difference is below some threshold; a frame is constrained to appear in
the matched pairs at most once. After the matching, we compute the following metrics (commonly
known as Precision, Recall and F-score):
PAB =
#matched frames
,
#frames in A
RAB =
#matched frames
,
#frames in B
FAB =
PAB ? RAB
.
0.5(PAB + RAB )
All of them lie between 0 and 1, and higher values indicate better agreement between A and B. Note
that these metrics are not symmetric ? if we swap A and B, the results will be different.
Our idea of examining the consistency among all summaries is to treat each summary in turn as if it
were the gold-standard (and assign it as B) while treating the other summaries as A?s. We report our
analysis of existing video datasets next.
Consistency in existing video databases We analyze video summaries in two video datasets: 50
videos from the Open Video Project (OVP) [25] and another 50 videos from Youtube [24]. Details
about these two video datasets are in section 5. We briefly point out that the two datasets have very
different subject matters and composition styles. Each of the 100 videos has 5 annotated summaries.
For each video, we compute the pairwise evaluation metrics in precision, recall, and F-score by
forming total 20 pairs of summaries from two different annotators. We then average them per video.
We plot how these averaged metrics distribute in Fig. 1. The plots show the number of videos
(out of 100) whose averaged metrics exceed certain thresholds, marked on the horizontal axes. For
example, more than 80% videos have an averaged F-score greater than 0.6, and 60% more than 0.7.
Note that there are many videos (?20) with averaged F-scores greater than 0.8, indicating that on
average, human-created summaries have a high degree of agreement. Note that the mean values of
the averaged metrics per video are also high.
4
Greedy algorithm for synthesizing an oracle summary Encouraged by our findings, we develop
a greedy algorithm for synthesizing one oracle summary per video, from multiple human-created
ones. This algorithm is adapted from a similar one for document summarization [14]. Specifically,
for each video, we initialize the oracle summary with the empty set y ? = ?. Iteratively, we then add
to y ? one frame i at a time from the video sequence
X
y ? ? y ? ? arg maxi
Fy? ?i,yu .
(6)
u
In words, the frame i is selected to maximally increase the F-score between the new oracle summary
and the human-created summaries yu . To avoid adding all frames in the video sequence, we stop
the greedy process as soon as there is no frame that can increase the F-score.
We measure the quality of the synthesized oracle summaries by computing their mean agreement
with the human annotations. The results are shown in Fig. 1 too. The quality is high: more than
90% of the oracle summaries agree well with other summaries, with an F-score greater than 0.6. In
what follows, we will treat the oracle summaries as ground-truth to inform our learning algorithms.
3.2
Sequential determinantal point processes (seqDPP)
The determinantal point process, as described in section 2, is a powerful tool for modeling diverse
subset selection. However, video frames are more than items in a set. In particular, in DPP, the
ground set is a bag ? items are randomly permutable such that the most diverse subset remains
unchanged. Translating this into video summarization, this modeling property essentially suggests
that we could randomly shuffle video frames and expect to get the same summary!
To address this serious deficiency, we propose sequential DPP, a new probabilistic model to introduce strong dependency structures between items. As a motivating example, consider a video
portraying the sequence of someone leaving home for school, coming back to home for lunch, leaving for market and coming back for dinner. If only visual appearance cues are available, a vanilla
DPP model will likely select only one frame from the home scene and repel other frames occurring
at the home. Our model, on the other hand, will recognize that the temporal span implies those
frames are still diverse despite their visual similarity. Thus, our modeling intuition is that diversity
should be a weaker prior for temporally distant frames but ought to act more strongly for closely
neighboring frames. We now explain how our seqDPP method implements this intuition.
Model definition Given a ground set (a long video sequence) Y, we partition it into T disjoint yet
ST
consecutive short segments t=1 Yt = Y. At time t, we introduce a subset selection variable Yt .
We impose a DPP over two neighboring segments where the ground set is Ut = Yt ? yt?1 , ie., the
union between the video segments and the selected subset in the immediate past. Let ?t denote the
L-matrix defined over the ground set Ut . The conditional distribution of Yt is thus given by,
det ?yt?1 ?yt
P (Yt = yt |Yt?1 = yt?1 ) =
.
(7)
det(?t + It )
As before, the subscript yt?1 ? yt selects the corresponding rows and columns from ?t . It is a
diagonal matrix, the same size as Ut . However, the elements corresponding to yt?1 are zeros and
the elements corresponding to Yt are 1 (see [13] for details). Readers who are familiar with DPP
might identify the conditional distribution is also a DPP, restricted to the ground set Yt .
The conditional probability is defined in such a way that at time t, the subset selected should be
diverse among Yt as well as be diverse from previously selected yt?1 . However, beyond those two
priors, the subset is not constrained by subsets selected in the distant past. Fig. 2 illustrates the idea
in graphical model notation. In particular, the joint distribution of all subsets is factorized
Y
P (Y1 = y1 , Y2 = y2 , ? ? ? , YT = yT ) = P (Y1 = y1 )
P (Yt = yt |Yt?1 = yt?1 ).
(8)
t=2
Inference and learning The MAP inference for the seqDPP model eq. (8) is as hard as the standard DPP model. Thus, we propose to use the following online inference, analogous to Bayesian
belief updates (for Kalman filtering):
y1? = arg maxy?Y1 P (Y1 = y)
y2? = arg maxy?Y2 P (Y2 = y|Y1 = y1? ) ? ? ?
?
yt? = arg maxy?Yt P (Yt = y|Yt?1 = yt?1
) ??????
5
Y1
Y2
Y3
Y1
Y2
Y3
Yt
???
Yt
YT
???
YT
Figure 2: Our sequential DPP for modeling sequential video data, drawn as a Bayesian network
Note that, at each step, the ground set could be quite small; thus an exhaustive search for the most
diverse subset is plausible. The parameter learning is similar to the standard DPP model. We
describe the details in the supplementary material.
3.3
Learning representations for diverse subset selection
As described previously, the kernel L of DPP hinges on the reparameterization with features of
the items that can generalize across different ground sets. The quality-diversity decomposition in
eq. (4), while elegantly leading to convex optimization, is severely limited in its power in modeling
complex items and dependencies among them. In particular, learning the subset selection rests solely
on learning the quality factor, as the diversity component remains handcrafted and fixed.
We overcome this deficiency with more flexible and powerful representations. Concretely, let fi
stand for the feature representation for item (frame) i, including both low-level visual cues and
meta-cues such as contextual information. We reparameterize the L matrix with fi in two ways.
Linear embeddings The simplest way is to linearly transform the original features
Lij = fiT W T W fj ,
(9)
where W is the transformation matrix.
Nonlinear hidden representation We use a one-hidden-layer neural network to infer a hidden
representation for fi
Lij = ziT W T W zj
where zi = tanh(U fi ),
(10)
where tanh(?) stands for the hyperbolic transfer function.
To learn the parameters W or U and W , we use maximum likelihood estimation (cf. eq. (5)), with
gradient-descent to optimize. Details are given in the supplementary material.
4
Related work
Space does not permit a thorough survey of video summarization methods. Broadly speaking, existing approaches develop a variety of selection criteria to prioritize frames for the output summary,
often combined with temporal segmentation. Prior work often aims to retain diverse and representative frames [2, 1, 10, 4, 11], and/or defines novel metrics for object and event saliency [3, 2, 6, 8].
When the camera is known to be stationary, background subtraction and object tracking are valuable
cues (e.g., [5]). Recent developments tackle summarization for dynamic cameras that are worn or
handheld [10, 8, 9] or design online algorithms to process streaming data [7].
Whereas existing methods are largely unsupervised, our idea to explicitly learn subset selection
from human-given summaries is novel. Some prior work includes supervised learning components
that are applied during selection (e.g., to generate learned region saliency metrics [8] or train classifiers for canonical viewpoints [10]), but they do not train/learn the subset selection procedure itself.
Our idea is also distinct from ?interactive? methods, which assume a human is in the loop to give
supervision/feedback on each individual test video [26, 27, 12].
Our focus on the determinantal point process as the building block is largely inspired by its appealing
property in modeling diversity in subset selection, as well as its success in search and ranking [17],
document summarization [14], news headline displaying [28], and pose estimation [29]. Applying
DPP to video summarization, however, is novel to the best of our knowledge.
Our seqDPP is closest in spirit to the recently proposed Markov DPP [28]. While both models enjoy
the Markov property by defining conditional probabilities depending only on the immediate past,
6
Table 1: Performance of various video summarization methods on OVP. Ours and its variants perform the best.
Unsupervised methods
F
P
R
DT
STIMO
VSUMM 1
VSUMM 2
[30]
57.6
67.7
53.2
[31]
63.4
60.3
72.2
[24]
70.3
70.6
75.8
[24]
68.2
73.1
69.1
+ Q/D
[14]
70.8?0.3
71.5?0.4
74.5?0.3
DPP
Supervised subset selection
Ours (seqDPP+)
Q/D
LINEAR
N . NETS
68.5?0.3 75.5?0.4 77.7?0.4
66.9?0.4 77.5?0.5 75.0?0.5
75.8?0.5 78.4?0.5 87.2?0.3
Table 2: Performance of our method with different representation learning
VSUMM 2
Youtube
Kodak
F
55.7
68.9
P
59.7
75.7
[24]
R
58.7
80.6
seqDPP+LINEAR
F
P
R
57.8?0.5 54.2?0.7 69.8?0.5
75.3?0.7 77.8?1.0 80.4?0.9
seqDPP+N . NETS
F
P
R
60.3?0.5 59.4?0.6 64.9?0.5
78.9?0.5 81.9?0.8 81.1?0.9
Markov DPP?s ground set is still the whole video sequence, whereas seqDPP can select diverse sets
from the present time. Thus, one potential drawback of applying Markov DPP is to select video
frames out of temporal order, thus failing to model the sequential nature of the data faithfully.
5
Experiments
We validate our approach of sequential determinantal point processes (seqDPP) for video summarization on several datasets, and obtain superior performance to competing methods.
5.1
Setup
Data We benchmark various methods on 3 video datasets: the Open Video Project (OVP), the
Youtube dataset [24], and the Kodak consumer video dataset [32]. They have 50, 392 , and 18 videos,
respectively. The first two have 5 human-created summaries per video and the last has one humancreated summary per video. Thus, for the first two datasets, we follow the algorithm described in
section 3.1 to create an oracle summary per video. We follow the same procedure as in [24] to
preprocess the video frames. We uniformly sample one frame per second and then apply two stages
of pruning to remove uninformative frames. Details are in the supplementary material.
Features Each frame is encoded with an `2-normalized 8192-dimensional Fisher vector ?i [33],
computed from SIFT features [34]. The Fisher vector represents well the visual appearance of the
video frame, and is hence used to compute the pairwise correlations of the frames in the qualitydiversity decomposition (cf. eq. (4)). We derive the quality features xi by measuring the representativeness of the frame. Specifically, we place a contextual window centered around the frame of
interest, and then compute its mean correlation (using the SIFT Fisher vector) to the other frames in
the window. By varying the size of the windows from 5 to 15, we obtain 12-dimensional contextual
features. We also add features computed from the frame saliency map [35]. To apply our method
for learning representations (cf. section 3.3), however, we do not make a distinction between the two
types, and instead compose a feature vector fi by concatenating xi and ?i . The dimension of our
linear transformed features W fi is 10, 40 and 100 for OVP, Youtube, and Kodak, respectively. For
the neural network, we use 50 hidden units and 50 output units.
Other details For each dataset, we randomly choose 80% of the videos for training and use the
remaining 20% for testing. We run 100 rounds of experiments and report the average performance,
which is evaluated by the aforementioned F-score, Precision, and Recall (cf. section 3.1). For
evaluation, we follow the standard procedure: for each video, we treat each human-created summary
as golden-standard and assess the quality of the summary output by our algorithm. We then average
over all human annotators to obtain the evaluation metrics for that video.
5.2
Results
We contrast our approach to several state-of-the-art methods for video summarization?which include several leading unsupervised methods?as well as the vanilla DPP model that has been successfully used for document summarization but does not model sequential structures. We compare
the methods in greater detail on the OVP dataset. Table 1 shows the results.
2
In total there are 50 Youtube videos. We keep 39 of them after excluding the cartoon videos.
7
Oracle
Youtube
(Video 99)
Sequential LINEAR
(F=70, P=60, R=88)
VSUMM1
(F=59, P=65, R=55)
User
Kodak
(Video 4)
Sequential LINEAR
(F=86, P=75, R=100)
VSUMM1
(F=50, P=100, R=33)
Figure 3: Exemplar video summaries by our seqDPP LINEAR vs.
VSUMM
summary [24].
Unsupervised or supervised? The four unsupervised methods are DT [30], STIMO [31],
VSUMM 1 [24], and VSUMM 2 with a postprocessing step to VSUMM 1 to improve the precision of
the results. We implement VSUMM ourselves using features described in the orignal paper and tune
its parameters to have the best test performance. All 4 methods use clustering-like procedures to
identify key frames as video summaries. Results of DT and STIMO are taken from their original
papers. They generally underperform VSUMM.
What is interesting is that the vanilla DPP does not outperform the unsupervised methods, despite
its success in other tasks. On the other end, our supervised method seqDPP, when coupled with the
linear or neural network representation learning, performs significantly better than all other methods.
We believe the improvement can be attributed to two factors working in concert: (1) modeling sequential structures of the video data, and (2) more flexible and powerful representation learning.
This is evidenced by the rather poor performance of seqDPP with the quality/diversity (Q/D) decomposition, where the representation of the items is severely limited such that modeling temporal
structures alone is simply insufficient.
Linear or nonlinear? Table 2 concentrates on comparing the effectiveness of these two types of
representation learning. The performances of VSUMM are provided for reference only. We see that
learning representations with neural networks generally outperforms the linear representations.
Qualitative results We present exemplar video summaries by different methods in Fig. 3. The
challenging Youtube video illustrates the advantage of sequential diverse subset selection. The visual
variance in the beginning of the video is far greater (due to close-shots of people) than that at the end
(zooming out). Thus the clustering-based VSUMM method is prone to select key frames from the first
half of the video, collapsing the latter part. In contrast, our seqDPP copes with time-varying diversity
very well. The Kodak video demonstrates again our method?s ability in attaining high recall when
users only make diverse selections locally but not globally. VSUMM fails to acknowledge temporally
distant frames can be diverse despite their visual similarities.
6
Conclusion
Our novel learning model seqDPP is a successful first step towards using human-created summaries
for learning to select subsets for the challenging video summarization problem. We just scratched
the surface of this fruit-bearing direction. We plan to investigate how to learn more powerful representations from low-level visual cues.
Acknowledgments B. G., W. C. and F. S. are partially supported by DARPA D11-AP00278, NSF IIS-1065243,
and ARO #W911NF-12-1-0241. K. G. is supported by ONR YIP Award N00014-12-1-0754 and gifts from
Intel and Google. B. G. and W. C. also acknowledge supports from USC Viterbi Doctoral Fellowship and USC
Annenberg Graduate Fellowship. We are grateful to Jiebo Luo for providing the Kodak dataset [32].
8
References
[1] R. Hong, J. Tang, H. Tan, S. Yan, C. Ngo, and T. Chua. Event driven summarization for web videos. In
ACM SIGMM Workshop on Social Media, 2009.
[2] C.-W. Ngo, Y.-F. Ma, and H.-J. Zhang. Automatic video summarization by graph modeling. In ICCV,
2003.
[3] Yu-Fei Ma, Lie Lu, Hong-Jiang Zhang, and Mingjing Li. A user attention model for video summarization.
In ACM MM, 2002.
[4] Tiecheng Liu and John R. Kender. Optimization algorithms for the selection of key frame sequences of
variable length. In ECCV, 2002.
[5] Y. Pritch, A. Rav-Acha, A. Gutman, and S. Peleg. Webcam synopsis: Peeking around the world. In ICCV,
2007.
[6] H. Kang, X. Chen, Y. Matsushita, and Tang X. Space-time video montage. In CVPR, 2006.
[7] Shikun Feng, Zhen Lei, Dong Yi, and Stan Z. Li. Online content-aware video condensation. In CVPR,
2012.
[8] Yong Jae Lee, Joydeep Ghosh, and Kristen Grauman. Discovering important people and objects for
egocentric video summarization. In CVPR, 2012.
[9] Zheng Lu and Kristen Grauman. Story-driven summarization for egocentric video. In CVPR, 2013.
[10] A. Khosla, R. Hamid, C-J. Lin, and N. Sundaresan. Large-scale video summarization using web-image
priors. In CVPR, 2013.
[11] Hong-Jiang Zhang, Jianhua Wu, Di Zhong, and Stephen W. Smoliar. An integrated system for contentbased video retrieval and browsing. Pattern Recognition, 30(4):643?658, 1997.
[12] D. Liu, Gang Hua, and Tsuhan Chen. A hierarchical visual model for video object summarization. PAMI,
32(12):2178?2190, 2010.
[13] Alex Kulesza and Ben Taskar. Determinantal point processes for machine learning. Foundations and
R in Machine Learning, 5(2-3):123?286, 2012.
Trends
[14] Alex Kulesza and Ben Taskar. Learning determinantal point processes. In UAI, 2011.
[15] Jennifer Gillenwater, Alex Kulesza, and Ben Taskar. Discovering diverse and salient threads in document
collections. In EMNLP/CNLL, 2012.
[16] Odile Macchi. The coincidence approach to stochastic point processes. Advances in Applied Probability,
7(1):83?122, 1975.
[17] Alex Kulesza and Ben Taskar. k-dpps: Fixed-size determinantal point processes. In ICML, 2011.
[18] Hoa Trang Dang. Overview of duc 2005. In Document Understanding Conf., 2005.
[19] Hui Lin and Jeff Bilmes. Multi-document summarization via budgeted maximization of submodular
functions. In NAACL/HLT, 2010.
[20] Jennifer Gillenwater, Alex Kulesza, and Ben Taskar. Near-optimal map inference for determinantal point
processes. In NIPS, 2012.
[21] V??ctor Vald?es and Jos?e M Mart??nez. Automatic evaluation of video summaries. ACM Trans. on multimedia computing, communications, and applications, 8(3):25, 2012.
[22] Emilie Dumont and Bernard M?erialdo. Automatic evaluation method for rushes summary content. In
ICME, 2009.
[23] Yingbo Li and Bernard Merialdo. Vert: automatic evaluation of video summaries. In ACM MM, 2010.
[24] Sandra Eliza Fontes de Avila, Ana Paula Brand?ao Lopes, et al. Vsumm: A mechanism designed to
produce static video summaries and a novel evaluation method. Pattern Recognition Letters, 32(1):56?
68, 2011.
[25] Open video project. http://www.open-video.org/.
[26] M. Ellouze, N. Boujemaa, and A. Alimi. Im(s)2: Interactive movie summarization system. J VCIR,
21(4):283?294, 2010.
[27] Dan B Goldman, Brian Curless, and Steven M. Seitz. Schematic storyboarding for video visualization
and editing. In SIGGRAPH, 2006.
[28] R. H. Affandi, A. Kulesza, and E. B. Fox. Markov determinantal point processes. In UAI, 2012.
[29] A. Kulesza and B. Taskar. Structured determinantal point processes. In NIPS, 2011.
[30] Padmavathi Mundur, Yong Rao, and Yelena Yesha. Keyframe-based video summarization using delaunay
clustering. Int?l J. on Digital Libraries, 6(2):219?232, 2006.
[31] Marco Furini, Filippo Geraci, Manuela Montangero, and Marco Pellegrini. Stimo: Still and moving video
storyboard for the web scenario. Multimedia Tools and Applications, 46(1):47?69, 2010.
[32] Jiebo Luo, Christophe Papin, and Kathleen Costello. Towards extracting semantically meaningful key
frames from personal video clips: from humans to computers. IEEE Trans. on Circuits and Systems for
Video Technology, 19(2):289?301, 2009.
[33] Florent Perronnin and Christopher Dance. Fisher kernels on visual vocabularies for image categorization.
In CVPR, 2007.
[34] David G Lowe. Distinctive image features from scale-invariant keypoints. IJCV, 60(2):91?110, 2004.
[35] Esa Rahtu, Juho Kannala, Mikko Salo, and Janne Heikkil. Segmenting salient objects from images and
videos. In ECCV, 2010.
9
| 5413 |@word determinant:1 briefly:1 open:4 underperform:1 seitz:1 decomposition:5 shot:1 liu:2 contains:1 score:8 selecting:7 document:21 ours:2 past:3 existing:7 outperforms:1 contextual:5 comparing:1 luo:2 yet:3 must:1 determinantal:15 john:1 distant:4 partition:1 informative:2 enables:1 remove:1 treating:1 plot:2 depict:1 update:1 v:1 alone:2 cue:9 selected:10 greedy:3 item:17 stationary:1 half:1 rav:1 discovering:2 beginning:1 designed:1 short:3 record:1 chua:1 preference:1 org:1 zhang:3 direct:1 qualitative:1 ijcv:1 combine:1 compose:1 dan:1 introduce:2 pairwise:4 market:1 expected:3 preparatory:1 multi:1 inspired:1 globally:1 montage:1 automatically:1 goldman:1 window:3 dang:1 gift:1 provided:1 begin:1 discover:2 underlying:1 matched:4 project:3 agnostic:2 notation:1 factorized:1 what:5 permutable:3 medium:1 circuit:1 substantially:2 finding:1 transformation:1 ghosh:1 ought:1 temporal:6 thorough:1 y3:2 ti:1 act:1 tackle:1 interactive:2 golden:1 grauman:4 classifier:1 demonstrates:1 unit:2 ly:3 enjoy:1 appear:2 producing:1 interestingness:1 yn:2 positive:1 before:2 scientist:1 segmenting:1 treat:5 vald:1 limit:1 severely:2 despite:6 jiang:2 meet:2 subscript:1 solely:1 perched:1 pami:1 might:2 doctoral:1 suggests:1 challenging:3 someone:1 limited:2 graduate:1 averaged:5 acknowledgment:1 camera:5 merialdo:1 testing:1 union:1 block:1 implement:2 procedure:4 nite:1 empirical:2 yan:1 hyperbolic:1 significantly:1 matching:2 vert:1 word:1 get:1 cannot:1 close:1 selection:21 applying:2 optimize:3 www:1 map:4 yt:33 uploaded:1 worn:1 attention:1 independently:2 convex:2 survey:2 impeded:1 parameterizing:1 gigabyte:1 reparameterization:2 notion:1 analogous:1 target:1 play:1 suppose:1 user:5 tan:1 designing:1 hypothesis:2 mikko:1 agreement:7 geraci:1 synthesize:1 element:2 recognition:3 particularly:1 trend:1 gutman:1 database:1 role:1 taskar:6 steven:1 coincidence:1 capture:3 region:1 news:1 shuffle:1 highest:2 valuable:2 intuition:2 yingbo:1 dynamic:1 personal:1 ultimately:1 grateful:1 segment:3 distinctive:1 learner:1 swap:1 joint:1 darpa:1 siggraph:1 cat:1 tx:1 various:3 train:2 distinct:1 describe:4 london:1 exhaustive:1 whose:1 emerged:1 quite:1 plausible:1 supplementary:3 encoded:1 cvpr:6 pab:3 fatal:1 favor:1 ability:1 unseen:1 jointly:1 transform:1 itself:1 peeking:1 online:3 sequence:8 pressing:1 advantage:1 net:2 propose:5 aro:1 coming:2 neighboring:2 loop:1 gold:1 validate:2 los:3 empty:1 plethora:1 produce:1 generating:2 categorization:1 ben:5 object:8 depending:1 develop:4 derive:1 gong:1 pose:3 exemplar:2 school:1 minor:1 progress:1 eq:4 zit:1 strong:2 c:1 come:1 indicate:1 implies:1 differ:1 concentrate:1 direction:1 peleg:1 closely:1 annotated:2 drawback:1 stochastic:1 centered:1 human:24 viewing:1 translating:1 material:3 ana:1 sandra:1 assign:1 ao:1 uncontrollable:1 kristen:3 hamid:1 brian:1 im:1 hold:1 mm:2 around:3 marco:2 ground:19 deciding:1 great:2 pellegrini:1 viterbi:1 exp:1 vary:1 commonality:1 consecutive:1 perceived:2 narrative:1 failing:1 estimation:3 applicable:1 bag:2 combinatorial:1 tanh:2 utexas:1 headline:1 agrees:1 faithfully:2 city:1 tool:3 tf:1 successfully:1 create:1 feisha:1 aim:2 rather:2 avoid:1 dinner:1 zhong:1 varying:2 derived:2 focus:3 ax:1 improvement:1 likelihood:2 contrast:2 attains:1 summarizing:1 inference:5 flaw:1 repulsion:1 perronnin:1 streaming:1 alarming:1 typically:1 integrated:1 hidden:4 going:1 transformed:1 selects:1 arg:7 among:6 flexible:3 aforementioned:1 development:1 plan:1 constrained:3 special:1 initialize:1 art:1 yip:1 equal:1 aware:1 once:1 sampling:2 manually:1 identical:2 encouraged:1 represents:1 look:1 unsupervised:9 yu:3 cartoon:1 icml:1 discrepancy:2 others:2 report:3 np:1 inherent:3 escape:1 few:1 serious:1 randomly:5 simultaneously:1 recognize:1 zoom:1 individual:2 usc:5 familiar:2 ourselves:1 interest:1 investigate:1 zheng:1 evaluation:10 d11:1 explosion:1 daily:2 necessary:1 ovp:5 fox:1 re:1 rush:1 joydeep:1 instance:1 column:2 modeling:13 rao:1 w911nf:1 retains:1 measuring:1 maximization:1 subset:41 successful:3 examining:1 interlocked:1 too:1 characterize:1 motivating:1 dependency:5 answer:1 trackable:1 combined:1 person:2 st:1 ie:1 retain:1 probabilistic:4 dong:1 lee:1 jos:1 together:3 synthesis:1 again:1 choose:1 prioritize:2 emnlp:1 collapsing:1 conf:1 lii:2 leading:3 style:1 li:3 account:1 potential:3 distribute:1 diversity:11 attaining:1 de:1 summarized:1 representativeness:4 includes:1 matter:4 int:1 explicitly:1 ranking:2 depends:1 scratched:1 later:1 lowe:1 doing:1 observing:2 analyze:2 sort:1 recover:1 annotation:4 reign:1 contribution:2 ass:1 variance:1 largely:3 efficiently:1 who:1 identify:3 saliency:3 preprocess:1 acha:1 generalize:2 world:1 raw:1 bayesian:2 curless:1 lu:2 bilmes:1 drive:1 explain:1 inform:1 emilie:1 hlt:1 definition:1 attentive:1 nonetheless:1 proof:1 attributed:1 di:1 static:1 wearable:1 stop:1 dataset:5 recall:4 knowledge:1 ut:3 organized:1 segmentation:1 reflecting:1 back:2 higher:1 dt:3 supervised:10 follow:3 wei:1 maximally:2 synopsis:1 editing:1 evaluated:1 though:1 strongly:1 furthermore:2 just:1 stage:1 correlation:2 hand:3 working:1 horizontal:1 web:3 duc:1 christopher:1 nonlinear:2 google:1 icme:1 defines:2 rab:3 quality:13 lei:1 believe:1 building:1 naacl:1 contain:1 concept:1 normalized:2 y2:7 hence:1 symmetric:1 iteratively:1 contentbased:1 eg:1 round:1 during:1 cosine:1 criterion:5 hong:3 demonstrate:1 performs:2 motion:1 fj:1 postprocessing:1 image:5 novel:9 recently:2 fi:6 superior:2 common:2 heed:1 overview:1 handcrafted:1 assembling:1 trait:1 synthesized:1 composition:1 dpps:2 automatic:5 vanilla:3 consistency:3 similarly:1 particle:1 gillenwater:2 submodular:1 moving:1 impressive:1 similarity:3 supervision:1 etc:3 add:2 surface:1 delaunay:1 closest:1 exclusion:2 recent:1 boqing:1 driven:2 manipulation:1 scenario:1 certain:1 n00014:1 meta:4 onr:1 success:4 christophe:1 yi:1 scoring:1 captured:1 seen:1 greater:5 handheld:1 impose:1 subtraction:1 redundant:1 signal:1 ii:1 stephen:1 multiple:4 desirable:1 sound:1 infer:1 sundaresan:1 keypoints:1 long:3 retrieval:2 lin:2 mle:1 award:1 watched:1 qi:3 schematic:1 variant:1 vision:1 metric:15 essentially:1 seqdpp:23 represent:1 kernel:4 whereas:3 background:2 uninformative:1 fellowship:2 condensation:1 source:1 leaving:2 crucial:1 rest:2 unlike:1 subject:4 ellouze:1 searchable:1 spirit:1 effectiveness:1 ngo:2 extracting:2 near:1 presence:1 exceed:1 enough:1 embeddings:1 variety:2 fit:1 zi:1 competing:1 florent:1 reduce:1 idea:7 texas:1 angeles:3 det:6 qj:1 whether:1 thread:1 defense:1 eliza:1 speaking:1 useful:2 generally:2 tune:1 locally:1 clip:1 category:1 simplest:3 generate:3 occupy:1 outperform:1 http:1 problematic:1 tutorial:1 zj:1 canonical:1 nsf:1 pritch:1 disjoint:1 per:9 diverse:23 broadly:1 discrete:1 key:5 salient:3 redundancy:1 threshold:2 four:1 drawn:1 budgeted:1 kept:1 graph:1 egocentric:2 run:1 angle:1 letter:1 powerful:6 lope:1 place:1 reader:1 wu:1 home:4 jianhua:1 layer:1 followed:1 matsushita:1 oracle:13 adapted:1 gang:1 placement:1 filippo:1 deficiency:4 fei:2 idf:1 scene:1 alex:5 encodes:1 yong:2 avila:1 generates:1 span:1 reparameterize:1 department:4 developing:2 according:2 structured:1 poor:1 cleverly:1 across:2 increasingly:1 appealing:3 lunch:1 making:1 costello:1 maxy:5 restricted:1 iccv:2 invariant:1 macchi:1 taken:1 ln:2 agree:1 remains:2 previously:2 turn:1 jennifer:2 fail:1 mechanism:1 visualization:1 end:4 available:1 permit:1 apply:3 pauli:1 hierarchical:1 generic:1 juho:1 kodak:6 original:3 clustering:4 include:2 cf:5 remaining:1 graphical:1 hinge:1 webcam:1 unchanged:1 feng:1 crafting:2 question:1 sha:1 diagonal:1 orignal:1 southern:3 gradient:1 distance:1 zooming:1 fy:1 consumer:2 analyst:1 length:3 kalman:1 index:1 insufficient:1 providing:1 difficult:1 unfortunately:1 idiosyncratic:1 setup:1 tsuhan:1 teach:2 rise:1 synthesizing:3 design:1 summarization:44 perform:1 datasets:11 markov:5 benchmark:2 acknowledge:2 ctor:1 descent:1 reparameterized:1 immediate:2 defining:1 ever:2 excluding:1 communication:1 frame:53 discovered:1 y1:11 rahtu:1 jiebo:2 esa:1 overcoming:2 david:1 evidenced:1 namely:1 pair:2 extensive:2 sentence:11 optimized:2 repel:1 kathleen:1 california:3 learned:2 fab:1 distinction:1 kang:1 hour:1 nip:2 trans:2 address:3 able:1 beyond:1 below:1 perception:1 pattern:2 kulesza:7 summarize:1 challenge:1 including:2 max:1 video:118 belief:1 power:3 event:3 improve:1 movie:1 technology:1 brief:1 library:1 temporally:2 stan:1 created:13 zhen:1 naive:1 coupled:1 lij:4 sn:1 ljj:2 chao:1 prior:7 literature:1 review:1 text:2 understanding:1 portraying:1 expect:2 interesting:2 limitation:3 filtering:1 annotator:3 digital:1 foundation:1 concert:1 degree:2 fruit:1 principle:1 viewpoint:2 displaying:1 story:1 share:1 austin:2 row:2 prone:1 summary:60 eccv:2 supported:2 last:1 soon:1 weaker:1 neighbor:1 affandi:1 taking:1 distributed:1 dpp:32 overcome:1 vocabulary:1 feedback:1 stand:2 dimension:1 rich:1 quantum:1 computes:1 instructed:1 made:1 commonly:1 concretely:1 collection:1 far:2 cope:1 social:1 approximate:1 pruning:1 keyframe:2 overcomes:1 keep:1 active:1 sequentially:1 uai:2 manuela:1 conclude:1 xi:5 search:3 khosla:1 why:1 table:4 nature:4 learn:6 reasonably:1 ca:3 transfer:1 obtaining:1 bearing:1 excellent:1 complex:2 meanwhile:1 elegantly:1 main:4 linearly:1 whole:1 profile:1 jae:1 fig:4 representative:2 intel:1 lun:1 sub:1 position:1 precision:4 fails:1 concatenating:1 lie:2 tang:2 specific:1 sift:2 maxi:1 essential:1 workshop:1 sequential:21 effectively:1 importance:2 adding:1 hui:1 illustrates:2 occurring:1 browsing:1 chen:2 simply:1 explore:1 appearance:3 likely:2 forming:1 happening:1 visual:13 nez:1 heikkil:1 tracking:1 partially:1 hua:1 truth:7 satisfies:1 acm:4 ma:2 mart:1 conditional:4 goal:2 identity:1 formulated:1 quantifying:1 marked:1 towards:3 jeff:1 shared:1 fisher:4 content:5 hard:2 youtube:9 specifically:3 uniformly:1 semantically:1 boujemaa:1 principal:1 multimedia:3 called:1 total:2 bernard:2 e:1 brand:1 meaningful:1 indicating:1 select:10 janne:1 people:2 support:1 latter:1 evaluate:2 dance:1 |
4,876 | 5,414 | Grouping-Based Low-Rank Trajectory Completion
and 3D Reconstruction
Marta Salas
Universidad de Zaragoza,
Zaragoza, Spain
[email protected]
Katerina Fragkiadaki
EECS, University of California,
Berkeley, CA 94720
[email protected]
Jitendra Malik
EECS, University of California,
Berkeley, CA 94720
[email protected]
Pablo Arbel?aez
Universidad de los Andes,
Bogot?a, Colombia
[email protected]
Abstract
Extracting 3D shape of deforming objects in monocular videos, a task known as
non-rigid structure-from-motion (NRSfM), has so far been studied only on synthetic datasets and controlled environments. Typically, the objects to reconstruct
are pre-segmented, they exhibit limited rotations and occlusions, or full-length
trajectories are assumed. In order to integrate NRSfM into current video analysis pipelines, one needs to consider as input realistic -thus incomplete- tracking,
and perform spatio-temporal grouping to segment the objects from their surroundings. Furthermore, NRSfM needs to be robust to noise in both segmentation and
tracking, e.g., drifting, segmentation ?leaking?, optical flow ?bleeding? etc. In
this paper, we make a first attempt towards this goal, and propose a method that
combines dense optical flow tracking, motion trajectory clustering and NRSfM
for 3D reconstruction of objects in videos. For each trajectory cluster, we compute multiple reconstructions by minimizing the reprojection error and the rank
of the 3D shape under different rank bounds of the trajectory matrix. We show
that dense 3D shape is extracted and trajectories are completed across occlusions
and low textured regions, even under mild relative motion between the object and
the camera. We achieve competitive results on a public NRSfM benchmark while
using fixed parameters across all sequences and handling incomplete trajectories,
in contrast to existing approaches. We further test our approach on popular video
segmentation datasets. To the best of our knowledge, our method is the first to
extract dense object models from realistic videos, such as those found in Youtube
or Hollywood movies, without object-specific priors.
1
Introduction
Structure-from-motion is the ability to perceive the 3D shape of objects solely from motion cues. It
is considered the earliest form of depth perception in primates, and is believed to be used by animals
that lack stereopsis, such as insects and fish [1].
In computer vision, non-rigid structure-from-motion (NRSfM) is the extraction of a time-varying
3D point cloud from its 2D point trajectories. The problem is under-constrained since many 3D
time-varying shapes and camera poses give rise to the same 2D image projections. To tackle this
ambiguity, early work of Bregler et al. [2] assumes the per frame 3D shapes lie in a low dimensional
subspace. They recover the 3D shape basis and coefficients, along with camera rotations, using
a 3K factorization of the 2D trajectory matrix, where K the dimension of the shape subspace,
1
Video sequence Trajectory clustering
Missing entries
3D Shape
Depth
NRSfM
Figure 1: Overview. Given a monocular video, we cluster dense flow trajectories using 2D motion
similarities. Each trajectory cluster results in an incomplete trajectory matrix that is the input to our
NRSfM algorithm. Present and missing trajectory entries for the chosen frames are shown in green
and red respectively. The color of the points in the rightmost column represents depth values (red is
close, blue is far). Notice the completion of the occluded trajectories on the belly dancer, that reside
beyond the image border.
extending the rank 3 factorization method for rigid SfM of Tomasi and Kanade [3]. Akhter et al.[4]
observe that the 3D point trajectories admit a similar low-rank decomposition: they can be written
as linear combinations over a 3D trajectory basis. This essentially reflects that 3D (and 2D) point
trajectories are temporally smooth. Temporal smoothness is directly imposed using differentials
over the 3D shape matrix in Dai et al. [5]. Further, rather than recovering the shape or trajectory
basis and coefficients, the authors propose a direct rank minimization of the 3D shape matrix, and
show superior reconstruction results.
Despite such progress, NRSfM has been so far demonstrated only on a limited number of synthetic
or lab acquired video sequences. Factors that limit the application of current approaches to realworld scenarios include:
(i) Missing trajectory data. The aforementioned state-of-the-art NRSfM algorithms assume complete trajectories. This is an unrealistic assumption under object rotations, deformations or occlusions. Work of Torresani et al. [6] relaxes the full-length trajectory assumption. They impose a
Gaussian prior over the 3D shape and use probabilistic PCA within a linear dynamical system for
extracting 3D deformation modes and camera poses; however, their method is sensitive to initialization and degrades with the amount of missing data. Gotardo and Martinez [7] combine the shape
and trajectory low-rank decompositions and can handle missing data; their method is one of our
baselines in Section 3. Park et al. [8] use static background structure to estimate camera poses
and handle missing data using a linear formulation over a predefined trajectory basis. Simon at al.
[9] consider a probabilistic formulation of the bilinear basis model of Akhter et al. [10] over the
non-rigid 3D shape deformations. This results in a matrix normal distribution for the time varying
3D shape with a Kronecker structured covariance matrix over the column and row covariances that
describe shape and temporal correlations respectively. Our work makes no assumptions regarding
temporal smoothness, in contrast to [8, 7, 9].
(ii) Requirement of accurate video segmentation. The low-rank priors typically used in NRSfM
require the object to be segmented from its surroundings. Work of [11] is the only approach that
attempts to combine video segmentation and reconstruction, rather than considering pre-segmented
objects. The authors projectively reconstruct small trajectory clusters assuming they capture rigidly
moving object parts. Reconstruction results are shown in three videos only, making it hard to judge
the success of this locally rigid model.
This paper aims at closing the gap between theory and application in object-agnostic NRSfM from
realistic monocular videos. We build upon recent advances in tracking, video segmentation and
low-rank matrix completion to extract 3D shapes of objects in videos under rigid and non-rigid
motion. We assume a scaled orthographic camera model, as standard in the literature [12, 13], and
low-rank object-independent shape priors for the moving objects. Our goal is a richer representation
of the video segments in terms of rotations and 3D deformations, and temporal completion of their
trajectories through occlusion gaps or tracking failures.
2
An overview of our approach is presented in Figure 1. Given a video sequence, we compute dense
point trajectories and cluster them using 2D motion similarities. For each trajectory cluster, we
first complete the 2D trajectory matrix using standard low-rank matrix completion. We then recover the camera poses through a rank 3 truncation of the trajectory matrix and Euclidean upgrade.
Last, keeping the camera poses fixed, we minimize the reprojection error of the observed trajectory
entries along with the nuclear norm of the 3D shape. A byproduct of affine NRSfM is trajectory
completion. The recovered 3D time-varying shape is backprojected in the image and the resulting
2D trajectories are completed through deformations, occlusions or other tracking ambiguities, such
as lack of texture. In summary, our contributions are:
(i) Joint study of motion segmentation and structure-from-motion. We use as input to reconstruction
dense trajectories from optical flow linking [14], as opposed to a) sparse corner trajectories [15],
used in previous NRSfM works [4, 5], or b) subspace trajectories of [16, 17], that are full-length
but cannot tolerate object occlusions. Reconstruction needs to be robust to segmentation mistakes.
Motion trajectory clusters are inevitably polluted with ?bleeding? trajectories that, although reside
on the background, they anchor on occluding contours. We use morphological operations to discard
such trajectories that do not belong to the shape subspace and confuse reconstruction.
(ii) Multiple hypothesis 3D reconstruction through trajectory matrix completion under various rank
bounds, for tackling the rank ambiguity.
(iii) We show that, under high trajectory density, rank 3 factorization of the trajectory matrix, as
opposed to 3K, is sufficient to recover the camera rotations in NRSfM. This allows the use of an
easy, well-studied Euclidean upgrade for the camera rotations, similar to the one proposed for rigid
SfM [3].
We present competitive results of our method on the recently proposed NRSfM benchmark of [17],
under a fixed set of parameters and while handling incomplete trajectories, in contrast to existing approaches. Further, we present extensive reconstruction results in videos from two popular video segmentation benchmarks, VSB100 [18] and Moseg [19], that contain videos from Hollywood movies
and Youtube. To the best of our knowledge, we are the first to show dense non-rigid reconstructions
of objects from real videos, without employing object-specific shape priors [10, 20]. Our code is
available at: www.eecs.berkeley.edu/?katef/nrsfm.
2
2.1
Low-rank 3D video reconstruction
Video segmentation by multiscale trajectory clustering
Given a video sequence, we want to segment the moving objects in the scene. Brox and Malik
[19] propose spectral clustering of dense point trajectories from 2D motion similarities and achieve
state-of-the-art performance on video segmentation benchmarks. We extend their method to produce multiscale (rather than single scale) trajectory clustering to deal with segmentation ambiguities
caused by scale and motion variations of the objects in the video scene. Specifically, we first compute a spectral embedding from the top eigenvectors of the normalized trajectory affinity matrix. We
then obtain discrete trajectory clusterings using the discretization method of [21], while varying the
number of eigenvectors to be 10, 20, 30 and 40 in each video sequence.
Ideally, each point trajectory corresponds to a sequence of 2D projections of a 3D physical point.
However, each trajectory cluster is spatially surrounded by a thin layer of trajectories that reside
outside the true object mask and do not represent projections of 3D physical points. They are
the result of optical flow ?bleeding ? to untextured surroundings [22], and anchor themselves on
occluding contours of the object. Although ?bleeding? trajectories do not drift across objects, they
are a source of noise for reconstruction since they do not belong to the subspace spanned by the true
object trajectories. We discard them by computing an open operation (erosion followed by dilation)
and an additional erosion of the trajectory cluster mask in each frame.
2.2
Non-rigid structure-from-motion
Given a trajectory cluster that captures an object in space and time, let Xtk ? R3?1 denote the 3D
coordinate [X Y Z]T of the kth object point at the tth frame. We represent 3D object shape with a
3
matrix S that contains the time varying coordinates of K object surface points in F frames:
? ? 1
S1
X1
? ? ?
= ? ... ? = ? ...
?
S3F ?P
SF
XF
1
X12
???
?
X1P
.. ? .
. ?
XF
2
???
XF
P
For the special case of rigid objects, shape coordinates are constant and the shape matrix takes the
simplified form: S3?P = [X1 X2 ? ? ? XP ] .
We adopt a scaled orthographic camera model for reconstruction [3]. Under orthography, the
projection rays are perpendicular to the image plane and the projection equation takes the form:
x = RX + t, where x = [x y]T is the vector of 2D pixel coordinates, R2?3 is a scaled truncated
rotation matrix and t2?1 is the camera translation. Combining the projection equations for all object
points in all fames, we obtain:
? 1
? 1?
1
1 ?
x1
? ..
.
xF
1
???
..
.
???
x2
..
.
xF
2
xP
t
.. ?
? .. ? ? 1P T ,
=
R
?
S
+
.
.
xF
tF
P
where the camera pose matrix R takes the form:
? 1?
Rrigid
2F ?3
R
= ? ... ? ,
RF
R1
= ? ...
0
?
Rnonrigid
2F ?3F
0
..
.
0
???
???
???
(1)
0
.. ?
.
.
F
R
?
(2)
We subtract the camera translation tt from the pixel coordinates xt , t = 1 ? ? ? F , fixing the origin
of the coordinate system on the objects?s center of mass in each frame, and obtain the centered
trajectory matrix W2F ?P for which W = R ? S.
? denote an incomplete trajectory matrix of a cluster obtained from our multiscale trajectory
Let W
clustering. Let H ? {0, 1}2F ?P denote a binary matrix that indicates presence or absence of
? Given W,
? H, we solve for complete trajectories W, shape S and camera pose R
entries in W.
by minimizing the camera reprojection error and 3D shape rank under various rank bounds for the
trajectory matrix. Rather than minimizing the matrix rank which is intractable, we minimize the
matrix nuclear norm instead (denoted by k?k? ), that yields the best convex approximation for the
matrix rank over the unit ball of matrices. Let denote Hadamard product and k?kF denote the
Frobenius matrix norm. Our cost function reads:
NRSfM(K):
min .
? 2 + kW ? R ? Sk2 + 1K>1 ? ?kSv k?
kH (W ? W)k
F
F
subject to
Rank(W) ? 3K, ??t , s.t. Rt (Rt )T = ?t I2?2 , t = 1 ? ? ? F.
W,R,S
(3)
We compute multiple reconstructions with K ? {1 ? ? ? 9}. Sv denotes the re-arranged shape matrix
where each row contains the vectorized 3D shape in that frame:
? 1
?
1
1
1
1
1
X1
SvF ?3P
= ? ..
.
X1F
Y1
..
.
Y1F
Z1
..
.
Z1F
???
???
???
XP
..
.
XPF
YP
..
.
YPF
ZP
..
.
ZPF
? = [PX
PY
PZ ] (I3 ? S),
(4)
where PX , PY , PZ are appropriate row selection matrices. Dai et al. [5] observe that SvF ?3P has
lower rank than the original S3F ?P since it admits a K-rank decomposition, instead of 3K, assuming per frame 3D shapes span a K dimensional subspace. Though S facilitates the writing of
the projection equations, minimizing the rank of the re-arranged matrix Sv avoids spurious degrees
of freedom. Minimization of the nuclear norm of Sv is used only in the non-rigid case (K > 1).
In the rigid case, the shape does not change in time and Sv1?3P has rank 1 by construction. We
approximately solve Eq. 3 in three steps.
Low-rank trajectory matrix completion
rank bound constraint:
min .
W
subject to
We want to complete the 2D trajectory matrix under a
? 2
kH (W ? W)k
F
Rank(W) ? 3K.
4
(5)
Due to its intractability, the rank bound constraint is typically imposed by a factorization, W =
U V T , U2F ?r, VP ?r , for our case r = 3K. Work of [23] empirically shows that the following
regularized problem is less prone to local minima than its non-regularized counterpart (? = 0):
min .
W,U2F ?3K ,VP ?3K
? 2 + ? (kUk2 + kVk2 )
kH (W ? W)k
F
F
F
2
W = UVT .
subject to
(6)
We solve Eq. 6 using the method of Augmented Lagrange multipliers. We want to explicitly search
over different rank bounds for the trajectory matrix W as we vary K. We do not choose to minimize
the nuclear norm instead, despite being convex, since different weights for the nuclear term result in
matrices of different ranks, thus is harder to control explicitly the rank bound. Prior work [24, 23]
shows that the bilinear formulation of Eq. 6, despite being non-convex in comparison to the nuclear
? 2 + kWk? ), it returns the same optimum in cases r >= r?,
regularized objective (kH (W ? W)k
F
where r? denotes the rank obtained by the unconstrained minimization of the nuclear regularized
objective. We use the continuation strategy proposed in [23] over r to avoid local minima for r < r?:
starting from large values of r, we iteratively reduce it till the desired rank bound 3K is achieved.
For details, please see [23, 24].
Euclidean upgrade Given a complete trajectory matrix, minimization of the reprojection error
term of Eq. 3 under the orthonormality constraints is equivalent to a SfM or NRSfM problem in its
standard form, previously studied in the seminal works of [3, 2]:
min .
kW ? R ? Sk2F
subject to
??t , s.t. Rt (Rt )T = ?t I2?2 , t = 1 ? ? ? F.
R,S
(7)
For rigid objects, Tomasi and Kanade [3] recover the camera pose and shape matrix via SVD of W
? ? S.
? The factorization is not unique
truncated to rank 3: W = UDVT = (UD1/2 )(D1/2 VT ) = R
?1
? ?S
?=R
? ? GG S.
? We estimate G so that RG
? satisfies the
since for any invertible matrix G3?3 : R
orthonormality constraints:
orthogonality:
same norm:
? 2t?1 GGT R
? T = 0, t = 1 ? ? ? F
R
2t
? 2t?1 GGT R
?T
? 2t GGT R
? T , t = 1 ? ? ? F.
R
=
R
2t?1
2t
(8)
The constraints of Eq. 8 form an overdetermined homogeneous linear system with respect to the
elements of the gram matrix Q = GGT . We estimate Q using least-squares and factorize it using
SVD to obtain G up to an arbitrary scaling and rotation of its row space [25]. Then, the rigid object
?
shape is obtained by S3?P = G?1 S.
For non-rigid objects, a similar Euclidean upgrade of the rotation matrices has been attempted using
a rank 3K (rather than 3) decomposition of W [26]. In the non-rigid case, the corrective transformation G has size 3K ? 3K. Each column triplet 3K ? 3 is recovered independently since it contains
the rotation information from all frames. For a long time, an overlooked rank 3 constraint on the
Gram matrix Qk = GTk Gk spurred conjectures regarding the ambiguity of shape recovery under
non-rigid motion [26]. This lead researchers to introduce additional priors for further constraining
the problem, such as temporal smoothness [27]. Finally, the work of [4] showed that orthonormality
constraints are sufficient to recover a unique non-rigid 3D shape. Dai et al. [5] proposed a practical
algorithm for Euclidean upgrade using rank 3K decomposition of W that minimizes the nuclear
norm of Qk under the orthonormality constraints.
Surprisingly, we have found that in practice it is not necessary to go beyond rank 3 truncation of W
to obtain the rotation matrices in the case of dense NRSfM. The large majority of trajectories span
the rigid component of the object, and their information suffices to compute the objects? rotations.
This is not the case for synthetic NRSfM datasets, where the number of tracked points on the articulating links is similar to the points spanning the ?torso-like? component, as in the famous ?Dance?
sequence [12]. In Section 3, we show dense face reconstruction results while varying the truncating
rank ?r of W for the Euclidean upgrade step, and verify that ?r = 3 is more stable than ?r > 3 for
NRSfM of faces.
Rank regularized least-squares for 3D shape recovery In the non-rigid case, given the recovered
camera poses R, we minimize the reprojection error of the observed trajectory entries and 3D shape
5
incomplete trajectories
groundtruth 3D shape
frontal view
missing entries
ours
complete trajectories
ours: frontal view
ours
ours: frontal view
sequence 3
99 frames long
mild deform./rot.
sequence 2
10 frames long
abrupt deform./rot.
rotated
Figure 2: Qualitative results in the synthetic benchmark of [17]. High quality reconstructions are
obtained with oracle (full-length) trajectories for both abrupt and smooth motion. For incomplete
trajectories, in the 3rd column we show in red the missing and in green the present trajectory entries.
The reconstruction result for the 2nd video sequence that has 30% missing data, though worse, is
still recognizable.
nuclear norm:
min .
1
2 kH
v
? ? R ? S)k2 + ?kSv k?
(W
F
(9)
PY PZ ] (I3 ? S).
? to constrain the 3D shape estimation; howNotice that we consider only the observed entries in W
ever, information from the complete W has been used for extracting the rotation matrices R. We
solve the convex, non-smooth problem in Eq. 9 using the nuclear minimization algorithm proposed
in [28]. It generalizes the accelerated proximal gradient method of [29] from l1 regularized leastsquares on vectors to nuclear norm regularized least-squares on matrices. It has a better iteration
complexity than the Fixed Point Continuation (FPC) method of [30] and the Singular Value Thresholding (SVT) method [31].
S
subject to
S = [PX
Given camera pose R and shape S, we backproject to obtain complete centered trajectory matrix
W = R ? S. Though we can in principle iterate over the extraction of camera pose and 3D shape, we
observed benefits from such iteration only in the rigid case. This observation agrees with the results
of Marques and Costeira [32] for rigid SfM from incomplete trajectories.
3
Experiments
The only available dense NRSfM benchmark has been recently introduced in Garg et al. [17]. They
propose a dense NRSfM method that minimizes a robust discontinuity term over the recovered 3D
depth along with 3D shape rank. However, their method assumes as input full-length trajectories
obtained via the subspace flow tracking method of [16]. Unfortunately, the tracker of [16] can
tolerate only very mild out-of-plane rotations or occlusions, which is a serious limitation for tracking
in real videos. Our method does not impose the full-length trajectory requirement. Also, we show
that the robust discontinuity term in [17] may not be necessary for high quality reconstructions.
The benchmark contains four synthetic video sequences that depict a deforming face, and three real
sequences that depict a deforming back, face and heart, respectively. Only the synthetic sequences
have ground-truth 3D shapes available, since it is considerably more difficult to obtain ground-truth
for NRSfM in non-synthetic environments. Dense full-length ground-truth 2D trajectories are provided for all sequences. For evaluation, we use the code supplied with the benchmark, that performs
a pre-alignment step at each frame between St and StGT using Procrustes analysis. Reconstruction
performance is measured by mean RMS error across all frames, where the per frame RMS error of
kSt ?St kF
.
a shape St with respect to ground-truth shape StGT is defined as: kSt GT
kF
GT
Figure 2 presents our qualitative results and Table 1 compares our performance against previous
state-of-the-art NRSfM methods: Trajectory Basis (TB) [12], Metric Projections (MP) [33], Variational Reconstruction (VR) [17] and CSF [7]. For CSF, we were not able to complete the experiment
for sequences 3 and 4 due to the non-scalable nature of the algorithm. Next to the error of each
6
Figure 3: Reconstruction results in the ?Back?, ?Face? and ?Heart? sequences of [17]. We show
present and missing trajectory entries, per frame depth maps and retextured depth maps.
method we show in parentheses the rank used, that is, the rank that gave the best error. Our method
uses exactly the same parameters and K = 9 for all four sequences. Baseline VR [17] adapts the
weight for the nuclear norm of S for each sequence. This shows robustness of our method under
varying object deformations. ?r is the truncated rank of W used for the Euclidean upgrade step.
When ?r > 3, we use the Euclidean upgrade proposed in [5]. ?r = 3 gives the most stable face
reconstruction results.
Next, to imitate a more realistic setup, we introduce missing entries to the ground-truth 2D tracks by
?hiding? trajectory entries that are occluded due to face rotations. The occluded points are shown in
red in Figure 2 3rd column. From the ?incomplete trajectories? section of Table 1, we see that the
error increase for our method is small in comparison to the full-length trajectory case.
In the real ?Back?, ?Face? and ?Heart? sequences of the benchmark, the objects are pre-segmented.
We keep all trajectories that are at least five frames long. This results in 29.29%, 30.54% and
? We used K = 8 for all sequences.
52.71% missing data in the corresponding trajectory matrices W.
We show qualitative results in Figure 3. The present and missing entries are shown in green and
red, respectively. The missing points occupy either occluded regions, or regions with ambiguous
correspondence, e.g., under specularities in the Heart sequence.
Next, we test our method on reconstructing objects from videos of two popular video segmentation
datasets: VSB100 [18], that contains videos uploaded to Youtube, and Moseg [19], that contains
videos from Hollywood movies. Each video is between 19 and 121 frames long. For all videos
we use K ? {1 ? ? ? 5}. We keep all trajectories longer than five frames. This results in missing
data varying from 20% to 70% across videos, with an average of 45% missing trajectory entries.
We visualize reconstructions for the best trajectory clusters (the ones closest to the ground-truth
segmentations supplied with the datasets) in Figure 4.
Discussion Our 3D reconstruction results in real videos show that, under high trajectory density,
small object rotations suffice to create the depth perception. We also observe the tracking quality to
be crucial for reconstruction. Optical flow deteriorates as the spatial resolution decreases, and thus
high video resolution is currently important for our method. The most important failure cases for our
Seq.1 (10)
Seq.2 (10)
Seq.3 (99)
Seq.4 (99)
ground-truth full trajectories
TB [12] MP [33] VR [17] ours
ours
?r = 3 ?r = 6
18.38 (2) 19.44 (3) 4.01 (9) 5.16
6.69
7.47 (2) 4.87 (3) 3.45 (9) 3.71
5.20
4.50 (4) 5.13 (6) 2.60 (9) 2.81
2.88
6.61 (4) 5.81 (4) 2.81 (9) 3.19
3.08
ours
?r = 9
21.02
25.6
3.00
3.54
incomplete trajectories
ours ?r = 3
CSF
4.92 (8.93% occl)
9.44 (31.60% occl)
3.40 (14.07% occl)
5.53 ( 13.63% occl)
15.6
36.8
??
??
Table 1: Reconstruction results on the NRSfM benchmark of [17]. We show mean RMS error per
cent (%). Numbers for TB, MP and VR baselines are from [17]. In the first column, we show in
parentheses the number of frames. ?r is the rank of W used for the Euclidean upgrade. The last
two columns shows the performance of our algorithm and CSF baseline when occluded points in the
ground-truth tracks are hidden.
7
K=4
K=2
K=1
K=3
K=3
K=3
K=1
K=1
K=2
K=2
Figure 4: Reconstruction results on the VSB and Moseg video segmentation datasets. For each
example we show a) the trajectory cluster, b) the present and missing entries, and c) the depths of the
visible (as estimated from ray casting) points, where red and blue denote close and far respectively.
method are highly articulated objects, which violates the low-rank assumptions. 3D reconstruction
of articulated bodies is the focus of our current work.
4
Conclusion
We have presented a practical method for extracting dense 3D object models from monocular uncalibrated video without object-specific priors. Our method considers as input trajectory motion
clusters obtained from automatic video segmentation that contain large amounts of missing data
due to object occlusions and rotations. We have applied our NRSfM method on synthetic dense reconstruction benchmarks and on numerous videos from Youtube and Hollywood movies. We have
shown that a richer object representation is achievable from video under mild conditions of camera
motion and object deformation: small object rotations are sufficient to recover 3D shape. ?We see
because we move, we move because we see?, said Gibson in his ?Perception of the Visual World?
[34]. We believe this paper has made a step towards encompassing 3D perception from motion into
general video analysis.
Acknowledgments
The authors would like to thank Philipos Modrohai for useful discussions. M.S. acknowledges
funding from Direcci?on General de Investigaci?on of Spain under project DPI2012-32168 and the
Ministerio de Educaci?on (scholarship FPU-AP2010-2906).
References
[1] Andersen, R.A., Bradley, D.C.: Perception of three-dimensional structure from motion. Trends in cognitive sciences 2 (1998) 222?228
[2] Bregler, C., Hertzmann, A., Biermann, H.: Recovering non-rigid 3d shape from image streams. In:
CVPR. (2000)
[3] Tomasi, C., Kanade, T.: shape and motion from image streams: a factorization method. Technical report,
IJCV (1991)
[4] Akhter, I., Sheikh, Y., Khan, S., Kanade, T.: Trajectory space: A dual representation for nonrigid structure
from motion. IEEE Transactions on Pattern Analysis and Machine Intelligence 33 (2011) 1442?1456
8
[5] Dai, Y.: A simple prior-free method for non-rigid structure-from-motion factorization. In: IJCV. (2012)
[6] Torresani, L., Hertzmann, A., Bregler, C.: Nonrigid structure-from-motion: Estimating shape and motion
with hierarchical priors. TPAMI 30 (2008)
[7] Gotardo, P.F.U., Martinez, A.M.: Computing smooth time trajectories for camera and deformable shape
in structure from motion with occlusion. TPAMI 33 (2011)
[8] Park, H.S., Shiratori, T., Matthews, I., Sheikh, Y.: 3d reconstruction of a moving point from a series of
2d projections. ECCV (2010)
[9] Simon, T., Valmadre, J., Matthews, I., Sheikh, Y.: Separable spatiotemporal priors for convex reconstruction of time-varying 3d point clouds. In: ECCV. (2014)
[10] Akhter, I., Simon, T., Khan, S., Matthews, I., Sheikh, Y.: Bilinear spatiotemporal basis models. In: ACM
Transaction on graphics, Accepted with minor revisions. (2011)
[11] Russell, C., Yu, R., Agapito, L.: Video pop-up: Monocular 3d reconstruction of dynamic scenes. In:
ECCV. (2014)
[12] Akhter, I., Sheikh, Y., Khan, S., Kanade, T.: Trajectory space: A dual representation for nonrigid structure
from motion. IEEE Trans. Pattern Anal. Mach. Intell. 33 (2011)
[13] Torresani, L., Bregler, C.: Space-time tracking. In: ECCV. (2002)
[14] Sundaram, N., Brox, T., Keutzer, K.: Dense point trajectories by GPU-accelerated large displacement
optical flow. In: ECCV. (2010)
[15] Lucas, B.D., Kanade, T.: An iterative image registration technique with an application to stereo vision
(darpa). In: Proceedings of the 1981 DARPA Image Understanding Workshop. (1981)
[16] Garg, R., Roussos, A., de Agapito, L.: A variational approach to video registration with subspace constraints. International Journal of Computer Vision 104 (2013) 286?314
[17] Garg, R., Roussos, A., Agapito, L.: Dense variational reconstruction of non-rigid surfaces from monocular video. (2013)
[18] Galasso, F., Nagaraja, N.S., Cardenas, T.J., Brox, T., Schiele, B.: A unified video segmentation benchmark: Annotation, metrics and analysis. In: ICCV. (2013)
[19] Brox, T., Malik, J.: Object segmentation by long term analysis of point trajectories. In: ECCV. (2010)
[20] Bao, S.Y.Z., Chandraker, M., Lin, Y., Savarese, S.: Dense object reconstruction with semantic priors. In:
CVPR. (2013) 1264?1271
[21] Yu, S., Shi, J.: Multiclass spectral clustering. In: ICCV. (2003)
[22] Thompson, W.B.: Exploiting discontinuities in optical flow. International Journal of Computer Vision 30
(1998) 17?4
[23] Cabral, R., de la Torre, F., Costeira, J., Bernardino, A.: Unifying nuclear norm and bilinear factorization
approaches for low-rank matrix decomposition. In: ICCV. (2013)
[24] Burer, S., Monteiro, R.D.C.: Local minima and convergence in low-rank semidefinite programming.
Math. Program. 103 (2005) 427?444
[25] Brand, M.: A Direct Method for 3D Factorization of Nonrigid Motion Observed in 2D. In: CVPR. (2005)
[26] Xiao, J., Chai, J., Kanade, T.: A closed-form solution to nonrigid shape and motion recovery. Technical
Report CMU-RI-TR-03-16, Robotics Institute, Pittsburgh, PA (2003)
[27] Torresani, L., Yang, D.B., Alexander, E.J., Bregler, C.: Tracking and modeling non-rigid objects with
rank constraints. In: CVPR. (2001)
[28] Toh, K.C., Yun, S.: An accelerated proximal gradient algorithm for nuclear norm regularized linear least
squares problems. Pacific Journal of Optimization (2010)
[29] Beck, A., Teboulle, M.: A fast iterative shrinkage-thresholding algorithm for linear inverse problems.
SIAM Journal on Imaging Sciences 2 (2009) 183?202
[30] Ma, S., Goldfarb, D., Chen, L.: Fixed point and bregman iterative methods for matrix rank minimization.
Math. Program. 128 (2011) 321?353
[31] Cai, J.F., Cand`es, E.J., Shen, Z.: A singular value thresholding algorithm for matrix completion. SIAM
J. on Optimization 20 (2010) 1956?1982
[32] Marques, M., Costeira, J.: Estimating 3d shape from degenerate sequences with missing data. Computer
Vision and Image Understanding 113 (2009) 261 ? 272
[33] Paladini, M., Del Bue, A., Sto?sic, M., Dodig, M., Xavier, J., Agapito, L.: Optimal metric projections for
deformable and articulated structure-from-motion. IJCV 96 (2012)
[34] Gibson, J.J.: The perception of the visual world. The American Journal of Psychology, 64 (1951) 622?625
9
| 5414 |@word mild:4 achievable:1 norm:12 nd:1 open:1 decomposition:6 covariance:2 tr:1 harder:1 contains:6 series:1 ours:8 rightmost:1 existing:2 bradley:1 current:3 recovered:4 discretization:1 toh:1 tackling:1 written:1 gpu:1 realistic:4 occl:4 visible:1 ministerio:1 shape:53 depict:2 sundaram:1 zpf:1 cue:1 intelligence:1 imitate:1 plane:2 math:2 five:2 x1p:1 along:3 kvk2:1 direct:2 differential:1 qualitative:3 ijcv:3 combine:3 ray:2 recognizable:1 introduce:2 acquired:1 mask:2 themselves:1 cand:1 considering:1 hiding:1 spain:2 provided:1 project:1 suffice:1 estimating:2 agnostic:1 mass:1 cabral:1 revision:1 minimizes:2 unified:1 transformation:1 temporal:6 berkeley:5 tackle:1 fpc:1 exactly:1 scaled:3 k2:1 control:1 s3f:2 unit:1 local:3 svt:1 limit:1 mistake:1 despite:3 bilinear:4 mach:1 rigidly:1 solely:1 approximately:1 garg:3 initialization:1 studied:3 co:1 limited:2 factorization:9 perpendicular:1 unique:2 camera:22 practical:2 acknowledgment:1 practice:1 orthographic:2 displacement:1 gibson:2 projection:10 pre:4 ud1:1 cannot:1 close:2 selection:1 writing:1 seminal:1 py:3 www:1 equivalent:1 imposed:2 demonstrated:1 shi:1 missing:19 center:1 map:2 go:1 uploaded:1 starting:1 independently:1 convex:5 truncating:1 resolution:2 thompson:1 recovery:3 abrupt:2 shen:1 perceive:1 colombia:1 d1:1 nuclear:14 spanned:1 his:1 embedding:1 handle:2 variation:1 coordinate:6 marta:1 construction:1 programming:1 homogeneous:1 us:1 hypothesis:1 origin:1 overdetermined:1 pa:2 element:1 trend:1 observed:5 cloud:2 capture:2 region:3 morphological:1 andes:1 decrease:1 russell:1 uncalibrated:1 environment:2 complexity:1 hertzmann:2 schiele:1 ideally:1 leaking:1 occluded:5 belly:1 dynamic:1 segment:3 upon:1 textured:1 basis:7 joint:1 darpa:2 various:2 corrective:1 articulated:3 fast:1 describe:1 outside:1 aez:1 richer:2 solve:4 cvpr:4 reconstruct:2 cardenas:1 ability:1 galasso:1 sequence:23 tpami:2 arbel:1 cai:1 reconstruction:36 propose:4 product:1 combining:1 hadamard:1 till:1 degenerate:1 achieve:2 adapts:1 deformable:2 frobenius:1 kh:5 bao:1 los:1 exploiting:1 convergence:1 cluster:14 reprojection:5 extending:1 requirement:2 produce:1 r1:1 zp:1 optimum:1 rotated:1 object:51 completion:9 fixing:1 pose:11 measured:1 minor:1 progress:1 eq:6 recovering:2 judge:1 csf:4 torre:1 centered:2 public:1 violates:1 require:1 suffices:1 leastsquares:1 bregler:5 tracker:1 considered:1 ground:8 normal:1 visualize:1 matthew:3 vary:1 early:1 adopt:1 estimation:1 currently:1 sensitive:1 agrees:1 hollywood:4 tf:1 create:1 reflects:1 minimization:6 gaussian:1 aim:1 i3:2 rather:5 avoid:1 shrinkage:1 varying:10 casting:1 earliest:1 focus:1 rank:51 indicates:1 contrast:3 baseline:4 rigid:27 typically:3 investigaci:1 spurious:1 hidden:1 pixel:2 monteiro:1 aforementioned:1 dual:2 insect:1 denoted:1 lucas:1 animal:1 constrained:1 art:3 special:1 brox:4 spatial:1 extraction:2 represents:1 park:2 kw:2 yu:2 thin:1 t2:1 report:2 torresani:4 salas:1 serious:1 surroundings:3 intell:1 beck:1 occlusion:9 attempt:2 freedom:1 highly:1 evaluation:1 alignment:1 semidefinite:1 predefined:1 accurate:1 bregman:1 byproduct:1 xpf:1 necessary:2 projectively:1 incomplete:10 euclidean:9 savarese:1 re:2 desired:1 deformation:7 column:7 modeling:1 teboulle:1 cost:1 entry:14 polluted:1 graphic:1 spatiotemporal:2 eec:4 sv:3 proximal:2 synthetic:8 considerably:1 st:3 density:2 sv1:1 international:2 siam:2 probabilistic:2 universidad:2 invertible:1 andersen:1 ambiguity:5 opposed:2 choose:1 worse:1 admit:1 corner:1 cognitive:1 american:1 return:1 yp:1 deform:2 valmadre:1 de:6 coefficient:2 jitendra:1 caused:1 explicitly:2 mp:3 udvt:1 stream:2 view:3 lab:1 closed:1 kwk:1 red:6 competitive:2 recover:6 annotation:1 simon:3 nagaraja:1 contribution:1 minimize:4 square:4 qk:2 yield:1 educaci:1 upgrade:9 vp:2 famous:1 trajectory:96 rx:1 researcher:1 failure:2 against:1 dancer:1 static:1 popular:3 color:1 knowledge:2 torso:1 segmentation:18 back:3 tolerate:2 costeira:3 formulation:3 arranged:2 though:3 furthermore:1 correlation:1 multiscale:3 lack:2 del:1 mode:1 sic:1 quality:3 believe:1 agapito:4 contain:2 normalized:1 true:2 counterpart:1 multiplier:1 orthonormality:4 verify:1 xavier:1 spatially:1 read:1 iteratively:1 goldfarb:1 i2:2 semantic:1 zaragoza:2 deal:1 erosion:2 please:1 ambiguous:1 gg:1 nonrigid:5 yun:1 complete:9 tt:1 performs:1 motion:31 l1:1 image:9 variational:3 recently:2 funding:1 superior:1 rotation:18 physical:2 overview:2 empirically:1 tracked:1 linking:1 belong:2 extend:1 smoothness:3 nrsfm:28 unconstrained:1 rd:2 automatic:1 closing:1 rot:2 moving:4 stable:2 similarity:3 surface:2 longer:1 etc:1 gt:2 closest:1 recent:1 showed:1 discard:2 scenario:1 binary:1 success:1 vt:1 minimum:3 dai:4 additional:2 impose:2 ii:2 full:9 multiple:3 segmented:4 smooth:4 xf:6 technical:2 burer:1 believed:1 long:6 lin:1 ggt:4 akhter:5 controlled:1 parenthesis:2 scalable:1 vision:5 essentially:1 metric:3 cmu:1 iteration:2 represent:2 orthography:1 achieved:1 uvt:1 robotics:1 background:2 want:3 singular:2 source:1 crucial:1 backprojected:1 subject:5 facilitates:1 flow:9 extracting:4 presence:1 svf:2 constraining:1 iii:1 easy:1 relaxes:1 yang:1 iterate:1 gave:1 psychology:1 reduce:1 regarding:2 multiclass:1 pca:1 rms:3 stereo:1 articulating:1 useful:1 fragkiadaki:1 eigenvectors:2 procrustes:1 amount:2 locally:1 tth:1 continuation:2 occupy:1 supplied:2 fish:1 notice:1 s3:2 deteriorates:1 estimated:1 per:5 track:2 blue:2 discrete:1 four:2 registration:2 imaging:1 realworld:1 inverse:1 groundtruth:1 seq:4 keutzer:1 scaling:1 sfm:4 bound:8 layer:1 followed:1 correspondence:1 oracle:1 kronecker:1 constraint:10 orthogonality:1 constrain:1 scene:3 x2:2 ri:1 min:5 span:2 separable:1 optical:7 xtk:1 x12:1 px:3 conjecture:1 structured:1 pacific:1 combination:1 ball:1 across:5 reconstructing:1 sheikh:5 g3:1 primate:1 making:1 s1:1 iccv:3 pipeline:1 heart:4 monocular:6 equation:3 previously:1 r3:1 chai:1 available:3 operation:2 generalizes:1 observe:3 hierarchical:1 spectral:3 appropriate:1 robustness:1 drifting:1 original:1 cent:1 assumes:2 clustering:8 include:1 top:1 completed:2 denotes:2 spurred:1 unifying:1 roussos:2 scholarship:1 build:1 malik:4 objective:2 move:2 degrades:1 strategy:1 rt:4 said:1 exhibit:1 affinity:1 kth:1 subspace:8 gradient:2 link:1 arbelaez:1 thank:1 majority:1 w2f:1 considers:1 spanning:1 assuming:2 length:8 code:2 kst:2 minimizing:4 difficult:1 unfortunately:1 setup:1 gk:1 rise:1 anal:1 perform:1 observation:1 datasets:6 benchmark:12 inevitably:1 truncated:3 marque:2 ever:1 frame:19 y1:1 arbitrary:1 drift:1 overlooked:1 pablo:1 introduced:1 extensive:1 z1:1 khan:3 tomasi:3 california:2 pop:1 discontinuity:3 trans:1 beyond:2 able:1 dynamical:1 perception:6 pattern:2 tb:3 program:2 rf:1 green:3 video:48 unrealistic:1 regularized:8 sk2f:1 movie:4 temporally:1 numerous:1 acknowledges:1 extract:2 prior:12 literature:1 understanding:2 kf:3 relative:1 encompassing:1 sk2:1 limitation:1 integrate:1 degree:1 affine:1 sufficient:3 xp:3 vectorized:1 xiao:1 thresholding:3 principle:1 intractability:1 surrounded:1 translation:2 eccv:6 row:4 prone:1 summary:1 surprisingly:1 last:2 truncation:2 keeping:1 free:1 institute:1 face:8 sparse:1 benefit:1 depth:8 dimension:1 gram:2 avoids:1 contour:2 world:2 reside:3 author:3 made:1 simplified:1 far:4 employing:1 transaction:2 keep:2 chandraker:1 anchor:2 pittsburgh:1 assumed:1 backproject:1 spatio:1 factorize:1 stereopsis:1 search:1 iterative:3 triplet:1 dilation:1 table:3 kanade:7 nature:1 robust:4 ca:2 fpu:1 dense:18 border:1 noise:2 katef:2 martinez:2 gtk:1 x1:4 augmented:1 body:1 direcci:1 sto:1 vr:4 sf:1 lie:1 kuk2:1 specific:3 xt:1 r2:1 pz:3 admits:1 grouping:2 intractable:1 workshop:1 texture:1 fame:1 confuse:1 gap:2 chen:1 subtract:1 rg:1 specularities:1 visual:2 lagrange:1 bernardino:1 bue:1 tracking:11 vsb:1 corresponds:1 truth:8 satisfies:1 extracted:1 acm:1 ma:1 goal:2 towards:2 absence:1 hard:1 youtube:4 change:1 specifically:1 ksv:2 accepted:1 e:2 svd:2 attempted:1 biermann:1 deforming:3 katerina:1 occluding:2 la:1 brand:1 alexander:1 frontal:3 accelerated:3 dance:1 handling:2 |
4,877 | 5,415 | Learning Mixtures of Submodular Functions for
Image Collection Summarization
Rishabh Iyer
Department of Electrical Engineering
University of Washington
[email protected]
Sebastian Tschiatschek
Department of Electrical Engineering
Graz University of Technology
[email protected]
Haochen Wei
LinkedIn & Department of Electrical Engineering
University of Washington
[email protected]
Jeff Bilmes
Department of Electrical Engineering
University of Washington
[email protected]
Abstract
We address the problem of image collection summarization by learning mixtures of
submodular functions. Submodularity is useful for this problem since it naturally
represents characteristics such as fidelity and diversity, desirable for any summary.
Several previously proposed image summarization scoring methodologies, in fact,
instinctively arrived at submodularity. We provide classes of submodular component functions (including some which are instantiated via a deep neural network)
over which mixtures may be learnt. We formulate the learning of such mixtures as a
supervised problem via large-margin structured prediction. As a loss function, and
for automatic summary scoring, we introduce a novel summary evaluation method
called V-ROUGE, and test both submodular and non-submodular optimization
(using the submodular-supermodular procedure) to learn a mixture of submodular
functions. Interestingly, using non-submodular optimization to learn submodular
functions provides the best results. We also provide a new data set consisting of
14 real-world image collections along with many human-generated ground truth
summaries collected using Amazon Mechanical Turk. We compare our method
with previous work on this problem and show that our learning approach outperforms all competitors on this new data set. This paper provides, to our knowledge,
the first systematic approach for quantifying the problem of image collection summarization, along with a new data set of image collections and human summaries.
1
Introduction
The number of photographs being uploaded online is growing at an unprecedented rate. A recent
estimate is that 500 million images are uploaded to the internet every day (just considering Flickr,
Facebook, Instagram and Snapchat), a figure which is expected to double every year [22]. Organizing
this vast amount of data is becoming an increasingly important problem. Moreover, the majority
of this data is in the form of personal image collections, and a natural problem is to summarize
such vast collections. For example, one may have a collection of images taken on a holiday trip,
and want to summarize and arrange this collection to send to a friend or family member or to post
on Facebook. Here the problem is to identify a subset of the images which concisely represents
all the diversity from the holiday trip. Another example is scene summarization [28], where one
wants to concisely represent a scene, like the Vatican or the Colosseum. This is relevant for creating
a visual summary of a particular interest point, where we want to identify a representative set of
views. Another application that is gaining importance is summarizing video collections [26, 13] in
order to enable efficient navigation of videos. This is particularly important in security applications,
where one wishes to quickly identify representative and salient images in massive amounts of video.
1
These problems are closely related and can be unified via the problem of finding the most representative subset of images from an entire image collection. We argue that many formulations of
this problem are naturally instances of submodular maximization, a statement supported by the fact
that a number of scoring functions previously investigated for image summarization are (apparently
unintentionally) submodular [30, 28, 5, 29, 8].
A set function f (?) is said to be submodular if for any element v and sets A ? B ? V \{v}, where
V represents the ground set of elements, f (A ? {v}) ? f (A) ? f (B ? {v}) ? f (B). This is
called the diminishing returns property and states, informally, that adding an element to a smaller
set increases the function value more than adding that element to a larger set. Submodular functions
naturally model notions of coverage and diversity in applications, and therefore, a number of machine
learning problems can be modeled as forms of submodular optimization [11, 20, 18]. In this paper,
we investigate structured prediction methods for learning weighted mixtures of submodular functions
for image collection summarization.
Related Work: Previous work on image summarization can broadly be categorized into (a)
clustering-based approaches, and (b) approaches which directly optimize certain scoring functions.
The clustering papers include [12, 8, 16]. For example, [12] proposes a hierarchical clustering-based
summarization approach, while [8] uses k-medoids-based clustering to generate summaries. Similarly [16] proposes top-down based clustering. A number of other methods attempt to directly
optimize certain scoring functions. For example, [28] focuses on scene summarization and poses an
objective capturing important summarization metrics such as likelihood, coverage, and orthogonality.
While they do not explicitly mention this, their objective function is in fact a submodular function.
Furthermore, they propose a greedy algorithm to optimize their objective. A similar approach was proposed by [30, 29], where a set cover function (which incidentally also is submodular) is used to model
coverage, and a minimum disparity formulation is used to model diversity. Interestingly, they optimize
their objective using the same greedy algorithm. Similarly, [15] models the problem of diverse image
retrieval via determinantal point processes (DPPs). DPPs are closely related to submodularity, and in
fact, the MAP inference problem is an instance of submodular maximization. Another approach for
image summarization was posed by [5], where they define an objective function using a graph-cut function, and attempt to solve it using a semidefinite relaxation. They unintentionally use an objective that
is submodular (and approximately monotone [18]) that can be optimized using the greedy algorithm.
Our Contributions: We introduce a family of submodular function components for image collection
summarization over which a convex mixture can be placed, and we propose a large margin formulation
for learning the mixture. We introduce a novel data set of fourteen personal image collections, along
with ground truth human summaries collected via Amazon mechanical Turk, and then subsequently
cleaned via methods described below. Moreover, in order to automatically evaluate the quality of
novel summaries, we introduce a recall-based evaluation metric, which we call V-ROUGE, to compare
automatically generated summaries to the human ones. We are inspired by ROUGE [17], a wellknown evaluation criterion for evaluating summaries in the document summarization community, but
we are unaware of any similar efforts in the computer vision community for image summarization. We
show evidence that V-ROUGE correlates well with human evaluation. Finally, we extensively validate
our approach on these data sets, and show that it outperforms previously explored methods developed
for similar problems. The resulting learnt objective, moreover, matches human summarization
performance on test data.
2
Image Collection Summarization
Summarization is a task that most humans perform intuitively. Broadly speaking, summarization is
the task of extracting information from a source that is both minimal and most important. The precise
meaning of most important (relevance) is typically subjective and thus will differ from individual
to individual and hence is difficult to precisely quantify. Nevertheless, we can identify two general
properties that characterize good image collection summarizes [19, 28]:
Fidelity: A summary should have good coverage, meaning that all of the distinct ?concepts? in
the collection have at least one representative in the summary. For example, a summary of a photo
collection containing both mountains and beaches should contain images of both scene types.
Diversity: Summaries should be as diverse as possible, i.e., summaries should not contain images
that are similar or identical to each other. Other words for this concept include diversity or dispersion.
In computer vision, this property has been referred to as orthogonality [28].
2
Note that [28] also includes the notion of ?likelihood,? where summary images should have high
similarity to many other images in the collection. We believe, however, that such likelihood is
covered by fidelity. I.e., a summary that only has images similar to many in the collection might miss
certain outlier, or minority, concepts in the collection, while a summary that has high fidelity should
include a representative image for every both majority and minority concept in the collection.Also,
the above properties could be made very high without imposing further size or budget constraints.
I.e., the goal of a summary is to find a small or within-budget subset having the above properties.
2.1
Problem Formulation
We cast the problem of image collection summarization as a subset selection problem: given a
collection of images I = (I1 , I2 , ? ? ? , I|V | ) represented by an index set V and given a budget c, we
aim to find a subset S ? V, |S| ? c, which best summarizes the collection. Though alternative
approaches are possible, we aim to solve this problem by learning a scoring function F : 2V ? R+ ,
such that high quality summaries are mapped to high scores and low quality summaries to low scores.
Then, image collection summarization can be performed by computing:
S ? ? argmaxS?V,|S|?c F (S).
(1)
?
For arbitrary set functions, computing S is intractable, but for monotone submodular functions
we rely on the classic result [25] that the greedy algorithm offers a constant-factor mathematical
quality guarantee. Computational tractability notwithstanding, submodular functions are natural for
measuring fidelity and diversity [19] as we argue in Section 4.
2.2
Evaluation Criteria: V-ROUGE
Before describing practical submodular functions for mixture components, we discuss a crucial element for both summarization evaluation and for the automated learning of mixtures: an objective evaluation criterion for judging the quality of summaries. Our criterion is constructed similar to the popular
ROUGE score used in multi-document summarization [17] and that correlates well with human perception. For document summarization, ROUGE (which in fact, is submodular [19, 20]) is defined as:
P
P
min (cw (A), cw (S))
w?W
P S?S P
rS (A) =
( , r(A) when S is clear from the context), (2)
w?W
S?S cw (S)
where S is a set of human-generated reference summaries, W is a set of features (n-grams), and where
cw (A) is the occurrence-count of w in summary A. We may extend r(?) to handle images by letting W
be a set of visual words, S a set of reference summaries, and cw (A) be the occurrence-counts of visual
word w in summary A. Visual words can for example be computed from SIFT-descriptors [21] as common in the popular bag-of-words framework in computer vision [31]. We call this V-ROUGE (visual
ROUGE). In our experiments, we use visual words extracted from color histograms, from super-pixels,
and also from OverFeat [27], a deep convolutional network ? details are given in Section 5.
3
Learning Framework
We construct our submodular scoring functions Fw (?)
convex combinations of non-negative
Pas
m
submodular
functions
f
,
f
,
.
.
.
,
f
,
i.e.
F
(S)
=
1 2
m
w
i=1 wi fi (S), where w = (w1 , . . . , wm ),
P
wi ? 0, i wi = 1. The functions fi are submodular components and assumed to be normalized:
i.e., fi (?) = 0, and fi (V ) = 1 for polymatroid functions and maxA?V fi (A) ? 1 for non-monotone
functions. This ensures that the components are compatible with each other. Obviously, the merit of
the scoring function Fw (?) depends on the selection of the components. In Section 4, we provide a
large number of natural component choices, mixtures over which span a large diversity of submodular
functions. Many of these component functions have appeared individually in past work and are
unified into a single framework in our approach.
Large-margin Structured Prediction: We optimize the weights w of the scoring function Fw (?)
in a large-margin structured prediction framework, i.e. the weights are optimized such that human
summaries S are separated from competitor summaries by a loss-dependent margin:
Fw (S) ? Fw (S 0 ) + `(S 0 ),
?S ? S, S 0 ? Y \ S,
(3)
where `(?) is the considered loss function, and where Y is a structured output space (for example Y
is the set of summaries that satisfy a certain budget c, i.e. Y = {S 0 ? V : |S 0 | ? c}). We assume
3
the loss to be normalized, 0 ? `(S 0 ) ? 1, ?S 0 ? V , to ensure mixture and loss are calibrated.
Equation (3) can be stated as Fw (S) ? maxS 0 ?Y [Fw (S 0 ) + `(S 0 )] , ?S ? S which is called lossaugmented inference. We introduce slack variables and minimize the regularized sum of slacks [20]:
X
?
0
0
min
max
[F
(S
)
+
`(S
)]
?
F
(S)
+ kwk22 ,
(4)
w
w
S 0 ?Y
2
w?0,kwk1 =1
S?S
where the non-negative orthant constraint, w ? 0, ensures that the final mixture is submodular. Note
a 2-norm regularizer is used on top of a 1-norm constraint kwk1 = 1 which we interpret as a prior to
encourage higher entropy, and thus more diverse mixture, distributions. Tractability depends on the
choice of the loss function. An obvious choice is `(S) = 1 ? r(S), which yields a non-submodular
optimization problem suitable for optimization methods such as [10] (and which we try in Section 7).
We also consider other loss functions that retain submodularity in loss augmented inference. For
now, assume that S? = maxS 0 ?Y [Fw (S 0 ) + `(S 0 )] can be estimated efficiently. The objective in (4)
can then be minimized using standard stochastic gradient descent methods, where the gradient for
sample S with respect to weight wi is given as
?
?
2
?
?
? ? fi (S) + ?wi .
Fw (S) + `(S) ? Fw (S) + kwk2 = fi (S)
(5)
?wi
2
Loss Functions: A natural loss function is `1?R (S) = 1 ? r(S) where r(S) = V-ROUGE(S).
Because r(S) is submodular, 1 ? r(S) is supermodular and hence maximizing Fw (S 0 ) + `(S 0 )
requires difference-of-submodular set function maximization [24] which is NP-hard [10]. We
also consider two alternative loss functions [20], complement V-ROUGE and surrogate V-ROUGE.
Complement V-ROUGE sets `c (S) = r(V \ S) and is still submodular but it is non-monotone.
`c (?) does have the necessary characteristics of a proper loss: summaries S+ with large V-ROUGE
score are mapped to small values and summaries S? with small V-ROUGE score are mapped to
large values. In particular, submodularity means r(S) + r(V \ S) ? r(V ) + r(?) = r(V ) or
r(V \ S) ? r(V ) ? r(S) = 1 ? r(S), so complement rouge is a submodular
upper
P
P bound of the ideal
supermodular loss. We define surrogate V-ROUGE as `surr (A) = Z1 S?S w?W c cw (A), where
S
WSc is the set of visual words that do not appear in reference summary S and Z is a normalization
constant. Here, a summary has a high loss if it contains many visual words that do not occur in
reference summaries and a low loss if it mainly contains visual words that occur in the reference
summaries. Surrogate V-ROUGE is not only monotone submodular, it is modular.
Loss augmented Inference: Depending on the loss function, different algorithms for performing
loss augmented inference, i.e. computation of the maximum in (4), must be used. When using the
surrogate loss lsurr (?), the mixture function together with the loss, i.e. fL (S) = Fw (S) + `(S), is
submodular and monotone. Hence, the greedy algorithm [25] can be used for maximization. This
algorithm is extremely simple to implement, and starting at S 0 = ?, iteratively chooses an element
j?
/ S t that maximizes fL (S t ? j), until the budget constraint is violated. While the complexity of
this simple procedure is O(n2 ) function evaluations, it can be significantly accelerated, thanks again
to submodularity [23], which in practice we find is almost linear time. When using complement rouge
`c (?) as the loss, fL (S) is still submodular but no longer monotone, so we utilize the randomized
greedy algorithm [2] (which is essentially a randomized variant of the greedy algorithm above, and
has approximation guarantees). Finally, when using loss 1-V-ROUGE, Fw (S) + `(S) is neither
submodular nor monotone and approximate maximization is intractable. However, we resort to well
motivated and scalable heuristics, such as the submodular-supermodular procedures that have shown
good performance in various applications [24, 10].
Runtime Inference:
Having learnt the weights for the mixture components, the resulting function
Pm
Fw (S) = i=1 wi fi (S) is monotone submodular, which can be optimized by the accelerated greedy
algorithm [23]. Thanks to submodularity, we can obtain near optimal solutions very efficiently.
4
Submodular Component Functions
In this section, we consider candidate submodular component functions to use in Fw (?). We consider
first functions capturing more of the notion of fidelity, and then next diversity, although the distinction
is not entirely crystal clear in these functions as some have aspects of both. Many of the components
are graph-based. We define a weighted graph G(V, E, s), with V representing a the full set of images
and E is every pair of elements in V . Each edge (i, j) ? E has weight si,j computed from the visual
features as described in Section 7. The weight si,j is a similarity score between images i and j.
4
4.1
Fidelity-like Functions
A function representing the fidelity of a subset to the whole is one that gets a large value when
the subset faithfully represents that whole. An intuitively reasonable property for such a function
is one that scores a summary highly if it is the case that the summary, as a whole, is similar to a
large majority of items in the set V . In this case, if a given summary A has a fidelity of f (A), then
any superset B ? A should, if anything, have higher fidelity, and thus it seems natural to use only
monotone non-decreasing functions as fidelity functions. Submodularity is also a natural property
since as more and more properties of an image collection are covered by a summary, the less chance
any given image not part of the summary would have in offering additional coverage ? in other
words, submodularity is a natural model for measuring the inherent redundancy in any summary.
Given this, we briefly describe some possible choices for coverage functions:
Facility Location. Given a summary S ? V , we can quantify coverage of the whole image collection
V by the similarity between i ? V and its
Pclosest image j ? S. Summing these similarities yields the
facility location function ffac.loc. (S) = i?V maxj?S si,j . The facility location function has been
used for scene summarization in [28] and as one of the components in [20].
Sum Coverage. Here, we compute the average similarity in S rather than the similarity of the best
element in S only. From the graph perspective
P
P (G) we sum over the weights of edges with at least
one vertex in S. Thus, fsum cov. (S) = i?V j?S si,j .
Thresholded sum/truncated graph cut This function has been used in document summarization [20] and is similar to the sum coverage function except
that instead of summing over all elements
P
in S, we threshold the inner sum. Define ?i (S) = j?S si,j , i.e. informally, ?i (S) conveys how
much of image i is covered by S. In order to keep an element i P
from being overly covered by S as the
cause of the objective getting large, we define fthresh.sum (S) = i?V min(?i (S), ? ?i (V )), which is
both monotone and submodular [20]. Under budget constraints, this function avoids summaries that
over-cover any images.
Feature functions. Consider a bag-of-words image model where for i ? V , bi = (bi,w )w?W
is a bag-of-words representation of image i indexed by the set of visual words W (cf. Section 5).
We can then define
function [14], defined using the visual words, as follows:
P a feature
P coverage
ffeat.cov. (S) = w?W g
b
, where g(?) is a monotone non-decreasing concave function.
i,w
i?I
This class is both monotone and submodular, and an added benefit of scalability, since it does not
require computation of a O(n2 ) similarity matrix like the graph-based functions proposed above.
4.2
Diversity
Diversity is another trait of a good summary, and there are a number of ways to quantify it. In this
case, while submodularity is still quite natural, monotonicity sometimes is not.
Penalty based diversity/dispersion Given P
a set S,
Pwe penalize similarity within S by summing
over all pairs as follows: fdissim. (S) = ? i?S j?S,j>i si,j [28] (a variant, also submodular,
P
takes the form ? i,j?S si,j [19]). These functions are submodular, and monotone decreasing, so
when added to other functions can yield non-monotone submodular functions. Such functions have
occurred before in document summarization [19], as a dispersion function [1], and even for scene
summarization [28] (in this last case, the submodularity property was not explicitly mentioned).
Diversity reward based on clusters. As in [20], we define a cluster based function rewarding
diversity. Given clusters P1 , P2 , ? ? ? , Pk obtained by some clustering algorithm. We define diversity
Pk
reward functions fdiv.reward (S) = j=1 g(S ? Pj ), where g(?) is a monotone submodular function
so that fdiv.reward (?) is also monotone and submodular. Given a budget, fdiv.reward (S) is maximized
by selecting S as diverse, over different clusters, as possible because of diminishing credit when
repeatedly choosing an item in a cluster.
5
Visual Words for Evaluation
V-ROUGE (see Section 2.2) depends on a visual ?bag-of-words? vocabulary, and to construct a visual
vocabulary, multitude choices exists. Common choices include SIFT descriptors [21], color descriptors [34], raw image patches [7], etc. For encoding, vector quantization (histogram encoding) [4],
sparse coding [35], kernel codebook encoding [4], etc. can all be used. For the construction of our
5
V-ROUGE metric, we computed three lexical types and used their union as our visual vocabulary. The
different types are intended to capture information about the images at different scales of abstraction.
Color histogram. The goal here is to capture overall image information via color information. We
follow the method proposed in [34]: Firstly, we extract the most frequent colors in RGB color space
from the images in an image collection using 10 ? 10 pixel patches. Secondly, these frequent colors
are clustered by k-means into 128 clusters, resulting in 128 cluster centers. Finally, we quantize the
most frequent colors in every 10 ? 10 pixel image patch using nearest neighbor vector quantization.
For every image, the resulting bag-of-colors is normalized to unit `1 -norm.
Super pixels. Here, we wish to capture information about small objects or image regions that are
identified by segmentation. Images are first segmented using the quick shift algorithm implemented
in VLFeat [33]. For every segment, dense SIFT descriptors are computed and clustered into 200
clusters. Then, a patch-wise intermediate bag of words bpatch is computed by vector quantization
and the RGB color histogram of the corresponding patch cpatch is appended to that set of words.
This results in intermediate features ?patch = [bpatch , cpatch ]. These intermediate features are again
clustered into 200 clusters. Finally, the intermediate features are vector-quantized according to their
`1 -distance. This final bag-of-words representation is normalized to unit `1 -norm.
Deep convolutional neural network. Our deep neural network based words are meant to capture
high-level information from the images. We use OverFeat [27], i.e. an image recognizer and feature
extractor based on a convolutional neural network for extracting medium to high level image features.
A sliding window is moved across an input picture such that every image is divided into 10 ? 10
blocks (using a 50% overlap) and the pixels within the window are presented to OverFeat as input.
The activations on layer 17 are taken as intermediate features ?k and clustered by k-means into 300
clusters. Then, each ?k is encoded by kernel codebook encoding [4]. For every image, the resulting
bag-of-words representation is normalized to the unit `1 -norm.
6
Data Collection
Dataset. One major contribution of our paper is our new data set which we plan soon to publicly
release. Our data set consists of 14 image collections, each comprising 100 images. The image
collections are typical real world personal image collections as they, for the most part, were taken
during holiday trips. For each collection, human-generated summaries were collected using Amazon
mechanical Turk. Workers were asked to select a subset of 10 images from an image collection such
that it summarizes the collection in the best possible way.1 In contrast to previous work on movie
summarization [13], Turkers were not tested for their ability to produce high quality summaries.
Every Turker was rewarded 10 US cents for every summary.
Pruning of poor human-generated summaries. The summaries collected using Amazon
mechanical Turk differ drastically in quality. For example, some of the collected summaries have low
quality because they do not represent an image collection properly, e.g. they consist only of pictures
of the same people but no pictures showing, say, architecture. Even though we went through several
distinct iterations of summary collection via Amazon Turk, improving the quality of our instructions
each time, it was impossible to ensure that all individuals produced meaningful summaries. Such
low quality summaries can drastically degrade performance of the learning algorithm. We thus
developed a strategy to automatically prune away bad summaries, where ?bad? is defined as the
worst V-ROUGE score relative to a current set of human summaries. The strategy is depicted in
Algorithm 1. Each pruning step removes the worst human summary, and then creates a new instance
of V-ROUGE using the updated pruned summaries. Pruning proceeds as long as a significant fraction
(greater than a desired ?p-value?) of null-hypothesis summarizes (generated uniformly at random)
scores better than the worst human summary. We chose a significant value of p = 0.10.
7
Experiments
To validate our approach, we learned mixtures of submodular functions with 594 component
functions using the data set described in Section 6. In this data set, all human generated reference
summaries are size 10, and we evaluated performance of our learnt mixtures also by producing size
10 summaries. The component functions were the monotone submodular functions described in
1
We did not provide explicit instructions on precisely how to summarize an image collection and instead
only asked that they choose a representative subset. We relied on their high-level intuitive understanding that the
gestalt of the image collection should be preserved in the summary.
6
Algorithm 1 Algorithm for pruning poor human-generated summaries.
Require: Confidence level p, human summaries S, number of random summaries N
Sample N uniformly at random size-10 image sets, to be used as summaries R = (R1 , . . . , RN )
Instantiate
PV-ROUGE-score rS (?) instantiated with summaries S
1
o ? |R|
R?R 1{rS (R)>minS?S rS (S)} // fraction of random summaries better than worst human
while o > p do
S ? S \ (argminS?S rS (S))
Re-instantiate V-ROUGE score rS (?) using updated pruned human summaries S.
Recompute overlap o as above, but with updated V-ROUGE score.
end while
return Pruned human summaries S
Figure 1: Three example 10?10 image collections from our new data set.
Section 4 using features described in Section 5. For weight optimization, we used AdaGrad [6], an
adaptive subgradient method allowing for informative gradient-based learning. We do 20 passes
through the samples in the collection.
We considered two types of experiments: 1) cheating experiments to verify that our proposed mixture
components can effectively learn good scoring functions; and 2) a 14-fold cross-validation experiment
to test our approach in real- world scenarios. In the cheating experiments, training and testing is
performed on the same image collection, and this is repeated 14 times. By contrast, for our 14-fold
cross-validation experiments, training is performed on 13 out of 14 image collections and testing is
performed on the held out summary, again repeating this 14 times. In both experiment types, since
our learnt functions are always monotone submodular, we compute summaries S ? of size 10 that
approximately maximize the scoring functions using the greedy algorithm. For these summaries,
we compute the V-ROUGE score r(S ? ). For easy score interpretation, we normalize it according to
sc(S ? ) = (r(S ? ) ? R)/(H ? R), where R is the average V-ROUGE score of random summaries
(computed from 1000 summaries) and where H is the average V-ROUGE score of the collected final
pruned human summaries. The result sc(S ? ) is smaller than zero if S ? scores worse than the average
random summary and larger than one if it scores better than the average human summary.
The best cheating results are shown as Cheat in Table 1, learnt using 1-V-ROUGE as a loss. The
results in column Min are computed by constrainedly minimizing V-ROUGE via the methods of [11],
and the results in column Max are computed by maximizing V-ROUGE using the greedy algorithm.
Therefore, the Max column is an approximate upper bound on our achievable performance. Clearly,
we are able to learn good scoring functions, as on average we significantly exceed average human
performance, i.e., we achieve an average score of 1.42 while the average human score is 1.00.
Results for cross-validation experiments are presented in Table 1. In the columns Our Methods
we present the performance of our mixtures learnt using the proposed loss functions described in
Section 3. We also present a set of baseline comparisons, using similarity scores computed via a
histogram intersection [32] method over the visual words used in the construction of V-ROUGE. We
present baseline results for the following schemes:
the facility location objective ffac.loc. (S) alone;
the facility location objective mixed with a ?-weighted penalty, i.e. ffac.loc. (S) + ?fdissim. (S);
Maximal marginal relevance [3], using ? to tradeoff between relevance and diversity;
Graphcut mixed with a ?-weighted penalty, similar to FLpen but where graphcut is used in
place of facility location;
kM K-Medoids clustering [9, Algorithm 14.2]. Initial cluster centers were selected uniformly at
random. As a dissimilarity score between images i and j, we used 1 ? si,j . Clustering was
run 20 times, and we used the cluster centers of the best clustering as the summary.
FL
FLpen
MMR
GCpen
7
In each of the above cases where a ? weight is used, we take for each image collection the ? ?
{0, 0.1, 0.2, . . . , 0.9, 1.0} that produced a submodular function that when maximized produced the
best average V-ROUGE score on the 13 training image sets. This approach, therefore, selects the best
baseline possible when performing a grid-search on the training sets. Note that both ?-dependent
functions, i.e. FLpen and GCpen , are non-monotone submodular. Therefore, we used the randomized
greedy algorithm [2] for maximization which has a mathematical guarantee (we ran the algorithm 10
times and used the best result).
Table 1 shows that using 1-V-ROUGE as a loss significantly outperforms the other methods. Furthermore, the performance is on average better than human performance, i.e. we achieve an average score
of 1.13 while the average human score is 1.00. This indicates that we can efficiently learn scoring
functions suitable for image collection summarization. For the other two losses, i.e. surrogate and
complement V-ROUGE, performance is significantly worse. Thus, in this case it seems advantageous
to use the proper (supermodular) loss and heuristic optimization (the submodular-supermodular
procedure [24, 10]) for loss-augmented inference during training, compared to using an approximate
(submodular or modular) loss in combination with an optimization algorithm for loss-augmented
inference with strong guarantees. This could, however, perhaps be circumvented by constructing a
more accurate strictly submodular surrogate loss but we leave this to future work.
Table 1: Cross-Validation Experiments (see text for details). Average human performance is 1.00,
average random performance is 0.00. For each image collection, the best result achieved by any of
Our Methods and by any of the Baseline Methods is highlighted in bold.
Limits
Our Methods
Baseline Methods
8
No.
Min
Max
Cheat
`1?R
`c
`surr
FL
FLpen
MMR
GCpen
kM
1
2
3
4
5
6
7
8
9
10
11
12
13
14
Avg.
-2.55
-2.06
-2.07
-3.20
-1.65
-2.83
-2.44
-1.66
-2.32
-1.46
-1.55
-1.74
-0.94
-1.46
-2.00
2.78
2.22
2.24
2.04
1.92
2.40
2.07
2.04
2.59
2.34
1.85
2.39
1.72
1.75
2.17
1.71
1.38
1.64
1.42
1.60
1.81
1.07
1.45
1.73
1.39
1.22
1.57
0.77
1.07
1.42
1.51
1.27
1.46
1.04
1.11
1.47
1.07
1.13
1.21
1.06
0.95
1.11
0.32
1.08
1.13
0.87
1.26
0.95
0.81
1.06
0.65
0.96
0.96
1.13
0.78
0.92
0.58
0.53
0.97
0.89
-0.36
0.44
0.23
-0.18
0.58
0.27
0.15
0.07
0.51
0.14
-0.08
0.12
0.14
0.77
0.20
1.45
0.18
0.47
0.71
0.96
1.26
0.93
0.62
0.81
1.58
0.43
0.78
0.02
0.23
0.75
0.82
0.58
0.94
1.01
0.93
1.16
0.70
0.38
0.94
0.99
0.56
0.54
-0.06
0.14
0.69
-0.51
0.65
0.85
0.51
0.95
-0.08
-0.33
0.57
0.09
-0.26
-0.29
0.02
0.52
0.22
0.21
1.06
0.21
-0.53
-0.02
-1.28
0.20
-0.84
-1.27
-0.59
0.07
0.05
-0.01
-0.04
-0.80
-0.27
1.23
0.89
0.52
1.32
0.70
1.05
0.97
0.91
0.38
0.73
0.26
0.63
0.02
0.29
0.71
Conclusions and Future Work
We have considered the task of automated summarization of image collections. A new data set
together with many human generated ground truth summaries was presented and a novel automated
evaluation metric called V-ROUGE was introduced. Based on large-margin structured prediction,
and either submodular or non-submodular optimization, we proposed a method for learning scoring
functions for image collection summarization and demonstrated its empirical effectiveness. In future
work, we would like to scale our methods to much larger image collections. A key step in this
direction is to consider low complexity and highly scalable classes of submodular functions. Another
challenge for larger image collections is how to collect ground truth, as it would be difficult for a
human to summarize a collection of, say, 10,000 images.
Acknowledgments: This material is based upon work supported by the National Science Foundation
under Grant No. (IIS-1162606), the Austrian Science Fund under Grant No. (P25244-N15), a Google
and a Microsoft award, and by the Intel Science and Technology Center for Pervasive Computing.
Rishabh Iyer is also supported by a Microsoft Research Fellowship award.
References
[1] A. Borodin, H. C. Lee, and Y. Ye. Max-sum diversification, monotone submodular functions and dynamic
updates. In Proc. of the 31st symposium on Principles of Database Systems, pages 155?166. ACM, 2012.
[2] N. Buchbinder, M. Feldman, J. Naor, and R. Schwartz. Submodular maximization with cardinality
constraints. In SODA, 2014.
8
[3] J. Carbonell and J. Goldstein. The use of MMR, diversity-based reranking for reordering documents and
producing summaries. In Research and Development in Information Retrieval, pages 335?336, 1998.
[4] K. Chatfield, V. Lemtexpitsky, A. Vedaldi, and A. Zisserman. The devil is in the details: an evaluation of
recent feature encoding methods. In British Machine Vision Conference (BMVC), 2011.
[5] T. Denton, M. Demirci, J. Abrahamson, A. Shokoufandeh, and S. Dickinson. Selecting canonical views for
view-based 3-d object recognition. In ICPR, volume 2, pages 273?276, Aug 2004.
[6] J. Duchi, E. Hazan, and Y. Singer. Adaptive subgradient methods for online learning and stochastic
optimization. Journal of Machine Learning Research (JMLR), 12:2121?2159, July 2011.
[7] R. Fergus, P. Perona, and A. Zisserman. Object class recognition by unsupervised scale-invariant learning.
In CVPR, volume 2, pages 264?271. IEEE, June 2003.
[8] Y. Hadi, F. Essannouni, and R. O. H. Thami. Video summarization by k-medoid clustering. In Symposium
on Applied Computing (SAC), pages 1400?1401, 2006.
[9] T. Hastie, R. Tibshirani, and J. Friedman. The elements of statistical learning, volume 2. Springer New
York, 2nd edition, 2009.
[10] R. Iyer and J. Bilmes. Algorithms for approximate minimization of the difference between submodular
functions, with applications. In UAI, 2012.
[11] R. Iyer, S. Jegelka, and J. Bilmes. Fast semidifferential-based submodular function optimization. In ICML,
2013.
[12] A. Jaffe, M. Naaman, T. Tassa, and M. Davis. Generating summaries and visualization for large collections
of geo-referenced photographs. In MIR, pages 89?98. ACM, 2006.
[13] A. Khosla, R. Hamid, C.-J. Lin, and N. Sundaresan. Large-scale video summarization using web-image
priors. In Conference on Computer Vision and Pattern Recognition (CVPR), June 2013.
[14] K. Kirchhoff and J. Bilmes. Submodularity for data selection in machine translation. In Empirical Methods
in Natural Language Processing (EMNLP), October 2014.
[15] A. Kulesza and B. Taskar. k-DPPs: Fixed-size determinantal point processes. In Proceedings of the 28th
International Conference on Machine Learning, 2011.
[16] X. Li, L. Chen, L. Zhang, F. Lin, and W.-Y. Ma. Image annotation by large-scale content-based image
retrieval. In Proc. 14th ann. ACM Int. Conf. on Multimedia, pages 607?610. ACM, 2006.
[17] C.-Y. Lin. Rouge: A package for automatic evaluation of summaries. In Text Summarization Branches
Out: Proceedings of the ACL-04 Workshop, pages 74?81, July 2004.
[18] H. Lin and J. Bilmes. Multi-document summarization via budgeted maximization of submodular functions.
In NAACL, 2010.
[19] H. Lin and J. Bilmes. A class of submodular functions for document summarization. In ACL/HLT-2011,
Portland, OR, June 2011.
[20] H. Lin and J. Bilmes. Learning mixtures of submodular shells with application to document summarization.
In Conference on Uncertainty in Artificial Intelligence (UAI), pages 479?490, 2012.
[21] D. Lowe. Object recognition from local scale-invariant features. In International Conference on Computer
Vision (ICCV), volume 2, pages 1150?1157, 1999.
[22] M. Meeker and L. Wu. Internet trends. Technical report, Kleiner Perkins Caufield & Byers, 2013.
[23] M. Minoux. Accelerated greedy algorithms for maximizing submodular set functions. Optimization
Techniques, pages 234?243, 1978.
[24] M. Narasimhan and J. Bilmes. A submodular-supermodular procedure with applications to discriminative
structure learning. In Conference on Uncertainty in Artificial Intelligence (UAI), Edinburgh, Scotland, July
2005. Morgan Kaufmann Publishers.
[25] G. Nemhauser, L. Wolsey, and M. Fisher. An analysis of approximations for maximizing submodular set
functions?i. Mathematical Programming, 14(1):265?294, 1978.
[26] C.-W. Ngo, Y.-F. Ma, and H. Zhang. Automatic video summarization by graph modeling. In International
Conference on Computer Vision (ICCV), pages 104?109 vol.1, Oct 2003.
[27] P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. LeCun. OverFeat: Integrated Recognition,
Localization and Detection using Convolutional Networks. ArXiv e-prints, Dec. 2013.
[28] I. Simon, N. Snavely, and S. M. Seitz. Scene Summarization for Online Image Collections. ICCV, 2007.
[29] P. Sinha and R. Jain. Extractive summarization of personal photos from life events. In International
Conference on Multimedia and Expo (ICME), pages 1?6, 2011.
[30] P. Sinha, S. Mehrotra, and R. Jain. Summarization of personal photologs using multidimensional content
and context. In International Conference on Multimedia Retrieval (ICMR), pages 1?8, 2011.
[31] J. Sivic, B. Russell, A. Efros, A. Zisserman, and W. Freeman. Discovering objects and their location in
images. In International Conference on Computer Vision (ICCV), volume 1, pages 370?377, Oct 2005.
[32] M. J. Swain and D. H. Ballard. Color indexing. Int. journal of computer vision, 7(1):11?32, 1991.
[33] A. Vedaldi and B. Fulkerson. VLFeat: An open and portable library of computer vision algorithms.
http://www.vlfeat.org/, 2008.
[34] C. Wengert, M. Douze, and H. J?egou. Bag-of-colors for improved image search. In International
Conference on Multimedia (MM), pages 1437?1440. ACM, 2011.
[35] J. Yang, K. Yu, Y. Gong, and T. S. Huang. Linear spatial pyramid matching using sparse coding for image
classification. In CVPR, pages 1794?1801, 2009.
9
| 5415 |@word briefly:1 achievable:1 advantageous:1 norm:5 seems:2 nd:1 semidifferential:1 open:1 instruction:2 km:2 seitz:1 r:6 rgb:2 egou:1 mention:1 initial:1 contains:2 disparity:1 score:25 loc:3 selecting:2 offering:1 document:9 interestingly:2 outperforms:3 subjective:1 past:1 current:1 com:1 si:8 gmail:1 activation:1 must:1 determinantal:2 informative:1 remove:1 fund:1 update:1 alone:1 greedy:12 instantiate:2 selected:1 item:2 reranking:1 intelligence:2 discovering:1 scotland:1 provides:2 quantized:1 codebook:2 location:7 recompute:1 firstly:1 org:1 zhang:3 mathematical:3 along:3 constructed:1 symposium:2 consists:1 naor:1 introduce:5 surr:2 expected:1 p1:1 nor:1 growing:1 multi:2 inspired:1 freeman:1 decreasing:3 automatically:3 window:2 considering:1 cardinality:1 moreover:3 maximizes:1 medium:1 null:1 mountain:1 maxa:1 developed:2 narasimhan:1 unified:2 finding:1 guarantee:4 every:11 multidimensional:1 concave:1 runtime:1 schwartz:1 unit:3 vlfeat:3 grant:2 appear:1 producing:2 before:2 engineering:4 referenced:1 local:1 limit:1 rouge:39 encoding:5 cheat:2 becoming:1 approximately:2 might:1 chose:1 acl:2 collect:1 minoux:1 tschiatschek:2 bi:2 practical:1 acknowledgment:1 lecun:1 testing:2 practice:1 union:1 implement:1 block:1 procedure:5 empirical:2 significantly:4 vedaldi:2 matching:1 word:22 confidence:1 get:1 selection:3 context:2 impossible:1 optimize:5 www:1 map:1 quick:1 lexical:1 maximizing:4 send:1 uploaded:2 center:4 starting:1 demonstrated:1 convex:2 formulate:1 amazon:5 sac:1 classic:1 handle:1 notion:3 fulkerson:1 linkedin:1 holiday:3 updated:3 construction:2 massive:1 dickinson:1 programming:1 us:1 hypothesis:1 pa:1 element:11 trend:1 recognition:5 particularly:1 cut:2 database:1 taskar:1 electrical:4 capture:4 worst:4 graz:1 ensures:2 region:1 went:1 russell:1 ran:1 mentioned:1 complexity:2 haochen:1 reward:5 asked:2 dynamic:1 personal:5 segment:1 creates:1 upon:1 localization:1 kirchhoff:1 represented:1 various:1 regularizer:1 separated:1 instantiated:2 distinct:2 describe:1 fast:1 jain:2 artificial:2 sc:2 choosing:1 quite:1 modular:2 larger:4 posed:1 solve:2 heuristic:2 encoded:1 say:2 cvpr:3 ability:1 cov:2 highlighted:1 final:3 online:3 obviously:1 unprecedented:1 propose:2 douze:1 maximal:1 frequent:3 relevant:1 argmaxs:1 organizing:1 achieve:2 intuitive:1 moved:1 validate:2 normalize:1 scalability:1 getting:1 double:1 cluster:12 r1:1 produce:1 incidentally:1 generating:1 leave:1 object:5 depending:1 friend:1 pose:1 gong:1 unintentionally:2 nearest:1 aug:1 strong:1 p2:1 coverage:10 implemented:1 extractive:1 quantify:3 direction:1 submodularity:12 differ:2 closely:2 subsequently:1 stochastic:2 human:30 enable:1 material:1 require:2 clustered:4 hamid:1 secondly:1 strictly:1 mm:1 considered:3 ground:5 credit:1 major:1 arrange:1 efros:1 recognizer:1 proc:2 bag:9 individually:1 faithfully:1 weighted:4 minimization:1 clearly:1 always:1 aim:2 super:2 rather:1 pervasive:1 release:1 focus:1 june:3 properly:1 portland:1 likelihood:3 mainly:1 indicates:1 contrast:2 baseline:5 summarizing:1 inference:8 dependent:2 abstraction:1 entire:1 typically:1 integrated:1 diminishing:2 perona:1 i1:1 comprising:1 selects:1 pixel:5 overall:1 fidelity:11 classification:1 proposes:2 overfeat:4 plan:1 development:1 spatial:1 marginal:1 construct:2 having:2 washington:5 beach:1 identical:1 represents:4 yu:1 denton:1 unsupervised:1 icml:1 future:3 minimized:1 np:1 report:1 inherent:1 national:1 individual:3 maxj:1 intended:1 consisting:1 microsoft:2 attempt:2 friedman:1 detection:1 interest:1 investigate:1 highly:2 evaluation:12 mixture:21 navigation:1 rishabh:2 semidefinite:1 held:1 accurate:1 edge:2 encourage:1 worker:1 necessary:1 fsum:1 indexed:1 icmr:1 desired:1 re:1 minimal:1 sinha:2 instance:3 column:4 modeling:1 cover:2 measuring:2 maximization:8 tractability:2 geo:1 vertex:1 subset:9 swain:1 characterize:1 learnt:7 calibrated:1 chooses:1 thanks:2 st:1 international:7 randomized:3 retain:1 systematic:1 rewarding:1 lee:1 together:2 quickly:1 w1:1 again:3 containing:1 choose:1 huang:1 emnlp:1 worse:2 conf:1 creating:1 resort:1 return:2 li:1 diversity:17 coding:2 bold:1 includes:1 int:2 satisfy:1 explicitly:2 depends:3 performed:4 view:3 try:1 lowe:1 apparently:1 hazan:1 wm:1 relied:1 annotation:1 simon:1 contribution:2 minimize:1 appended:1 publicly:1 convolutional:4 descriptor:4 characteristic:2 efficiently:3 maximized:2 yield:3 identify:4 hadi:1 kaufmann:1 raw:1 produced:3 bilmes:9 flickr:1 sebastian:1 hlt:1 facebook:2 competitor:2 turk:5 obvious:1 conveys:1 naturally:3 dataset:1 popular:2 recall:1 knowledge:1 color:12 segmentation:1 goldstein:1 higher:2 supermodular:7 supervised:1 day:1 methodology:1 follow:1 wei:1 zisserman:3 bmvc:1 formulation:4 evaluated:1 though:2 improved:1 ffac:3 furthermore:2 just:1 until:1 web:1 google:1 icme:1 jaffe:1 quality:10 perhaps:1 believe:1 ye:1 concept:4 contain:2 normalized:5 verify:1 facility:6 hence:3 naacl:1 iteratively:1 i2:1 pwe:1 during:2 davis:1 anything:1 byers:1 criterion:4 n15:1 arrived:1 crystal:1 duchi:1 image:87 meaning:2 wise:1 novel:4 fi:8 common:2 polymatroid:1 fourteen:1 tassa:1 million:1 extend:1 occurred:1 interpretation:1 volume:5 interpret:1 kwk2:1 trait:1 significant:2 imposing:1 feldman:1 dpps:3 automatic:3 grid:1 pm:1 similarly:2 submodular:70 language:1 similarity:9 longer:1 etc:2 recent:2 perspective:1 wellknown:1 rewarded:1 scenario:1 certain:4 diversification:1 buchbinder:1 kwk1:2 life:1 scoring:14 morgan:1 minimum:1 additional:1 greater:1 prune:1 maximize:1 july:3 ii:1 sliding:1 full:1 desirable:1 branch:1 sundaresan:1 segmented:1 technical:1 match:1 offer:1 long:1 retrieval:4 cross:4 divided:1 lin:6 post:1 graphcut:2 award:2 prediction:5 variant:2 scalable:2 austrian:1 vision:10 metric:4 essentially:1 lossaugmented:1 histogram:5 represent:2 normalization:1 sometimes:1 kernel:2 iteration:1 achieved:1 penalize:1 preserved:1 dec:1 want:3 fellowship:1 pyramid:1 source:1 crucial:1 publisher:1 pass:1 mir:1 kwk22:1 member:1 effectiveness:1 call:2 extracting:2 ngo:1 near:1 yang:1 ideal:1 intermediate:5 exceed:1 superset:1 easy:1 automated:3 architecture:1 identified:1 hastie:1 inner:1 tradeoff:1 shift:1 motivated:1 chatfield:1 effort:1 penalty:3 speaking:1 cause:1 york:1 repeatedly:1 deep:4 useful:1 covered:4 informally:2 clear:2 amount:2 repeating:1 extensively:1 generate:1 http:1 canonical:1 judging:1 estimated:1 overly:1 medoid:1 tibshirani:1 broadly:2 diverse:4 vol:1 redundancy:1 salient:1 key:1 nevertheless:1 threshold:1 budgeted:1 neither:1 pj:1 thresholded:1 utilize:1 vast:2 graph:7 relaxation:1 monotone:21 fraction:2 year:1 sum:8 subgradient:2 run:1 package:1 uncertainty:2 soda:1 place:1 family:2 almost:1 reasonable:1 wu:1 patch:6 summarizes:4 capturing:2 bound:2 internet:2 fl:5 entirely:1 layer:1 fold:2 occur:2 orthogonality:2 precisely:2 constraint:6 perkins:1 scene:7 expo:1 aspect:1 min:6 span:1 extremely:1 performing:2 pruned:4 circumvented:1 department:4 structured:6 according:2 icpr:1 combination:2 poor:2 smaller:2 across:1 increasingly:1 wi:7 turker:1 intuitively:2 iccv:4 indexing:1 outlier:1 medoids:2 invariant:2 taken:3 equation:1 visualization:1 previously:3 describing:1 discus:1 count:2 slack:2 singer:1 letting:1 merit:1 ffeat:1 shokoufandeh:1 end:1 photo:2 hierarchical:1 away:1 occurrence:2 alternative:2 eigen:1 cent:1 top:2 clustering:10 include:4 ensure:2 tugraz:1 cf:1 arxiv:1 objective:12 added:2 print:1 strategy:2 snavely:1 surrogate:6 said:1 gradient:3 nemhauser:1 cw:6 distance:1 mapped:3 majority:3 degrade:1 carbonell:1 argue:2 collected:6 portable:1 minority:2 argmins:1 modeled:1 index:1 minimizing:1 sermanet:1 difficult:2 october:1 statement:1 negative:2 stated:1 proper:2 summarization:42 perform:1 allowing:1 upper:2 dispersion:3 descent:1 orthant:1 truncated:1 precise:1 rn:1 arbitrary:1 community:2 introduced:1 complement:5 cast:1 cleaned:1 mechanical:4 trip:3 optimized:3 z1:1 security:1 pair:2 cheating:3 sivic:1 concisely:2 distinction:1 learned:1 address:1 able:1 proceeds:1 below:1 perception:1 pattern:1 borodin:1 appeared:1 kulesza:1 summarize:4 challenge:1 including:1 gaining:1 video:6 max:7 suitable:2 overlap:2 natural:9 rely:1 regularized:1 event:1 representing:2 scheme:1 movie:1 technology:2 library:1 picture:3 mathieu:1 mehrotra:1 extract:1 text:2 prior:2 understanding:1 adagrad:1 relative:1 loss:31 reordering:1 mixed:2 wolsey:1 validation:4 foundation:1 rkiyer:1 jegelka:1 principle:1 translation:1 compatible:1 summary:83 supported:3 placed:1 last:1 soon:1 drastically:2 turkers:1 neighbor:1 sparse:2 benefit:1 edinburgh:1 vocabulary:3 world:3 evaluating:1 unaware:1 gram:1 avoids:1 collection:54 made:1 adaptive:2 avg:1 correlate:2 gestalt:1 approximate:4 pruning:4 mmr:3 keep:1 monotonicity:1 uai:3 summing:3 assumed:1 fergus:2 discriminative:1 search:2 meeker:1 kleiner:1 khosla:1 table:4 learn:5 ballard:1 improving:1 quantize:1 investigated:1 constructing:1 did:1 pk:2 dense:1 whole:4 edition:1 n2:2 repeated:1 categorized:1 wengert:1 augmented:5 representative:6 referred:1 intel:1 pv:1 wish:2 explicit:1 candidate:1 jmlr:1 extractor:1 down:1 british:1 bad:2 sift:3 showing:1 instagram:1 explored:1 multitude:1 evidence:1 intractable:2 exists:1 quantization:3 consist:1 adding:2 effectively:1 importance:1 workshop:1 dissimilarity:1 notwithstanding:1 iyer:4 budget:7 margin:6 chen:1 entropy:1 depicted:1 intersection:1 photograph:2 visual:17 springer:1 truth:4 chance:1 extracted:1 acm:5 ma:2 shell:1 oct:2 goal:2 quantifying:1 ann:1 jeff:1 fisher:1 content:2 fw:15 hard:1 typical:1 except:1 uniformly:3 demirci:1 miss:1 called:4 multimedia:4 meaningful:1 select:1 people:1 meant:1 devil:1 relevance:3 violated:1 accelerated:3 evaluate:1 tested:1 |
4,878 | 5,416 | Deep Learning Face Representation by Joint
Identification-Verification
Yi Sun1
Yuheng Chen2
Xiaogang Wang3,4
Xiaoou Tang1,4
Department of Information Engineering, The Chinese University of Hong Kong
2
SenseTime Group
3
Department of Electronic Engineering, The Chinese University of Hong Kong
4
Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences
1
[email protected] [email protected]
[email protected] [email protected]
Abstract
The key challenge of face recognition is to develop effective feature representations for reducing intra-personal variations while enlarging inter-personal
differences. In this paper, we show that it can be well solved with deep learning
and using both face identification and verification signals as supervision. The
Deep IDentification-verification features (DeepID2) are learned with carefully
designed deep convolutional networks. The face identification task increases the
inter-personal variations by drawing DeepID2 features extracted from different
identities apart, while the face verification task reduces the intra-personal
variations by pulling DeepID2 features extracted from the same identity together,
both of which are essential to face recognition. The learned DeepID2 features
can be well generalized to new identities unseen in the training data. On the
challenging LFW dataset [11], 99.15% face verification accuracy is achieved.
Compared with the best previous deep learning result [20] on LFW, the error rate
has been significantly reduced by 67%.
1
Introduction
Faces of the same identity could look much different when presented in different poses, illuminations, expressions, ages, and occlusions. Such variations within the same identity could overwhelm
the variations due to identity differences and make face recognition challenging, especially in
unconstrained conditions. Therefore, reducing the intra-personal variations while enlarging the
inter-personal differences is a central topic in face recognition. It can be traced back to early
subspace face recognition methods such as LDA [1], Bayesian face [16], and unified subspace
[22, 23]. For example, LDA approximates inter- and intra-personal face variations by using two
scatter matrices and finds the projection directions to maximize the ratio between them. More recent
studies have also targeted the same goal, either explicitly or implicitly. For example, metric learning
[6, 9, 14] maps faces to some feature representation such that faces of the same identity are close
to each other while those of different identities stay apart. However, these models are much limited
by their linear nature or shallow structures, while inter- and intra-personal variations are complex,
highly nonlinear, and observed in high-dimensional image space.
In this work, we show that deep learning provides much more powerful tools to handle the two types
of variations. Thanks to its deep architecture and large learning capacity, effective features for face
recognition can be learned through hierarchical nonlinear mappings. We argue that it is essential
to learn such features by using two supervisory signals simultaneously, i.e. the face identification
and verification signals, and the learned features are referred to as Deep IDentification-verification
features (DeepID2). Identification is to classify an input image into a large number of identity
1
classes, while verification is to classify a pair of images as belonging to the same identity or not
(i.e. binary classification). In the training stage, given an input face image with the identification
signal, its DeepID2 features are extracted in the top hidden layer of the learned hierarchical nonlinear
feature representation, and then mapped to one of a large number of identities through another
function g(DeepID2). In the testing stage, the learned DeepID2 features can be generalized to other
tasks (such as face verification) and new identities unseen in the training data. The identification
supervisory signal tends to pull apart the DeepID2 features of different identities since they have to
be classified into different classes. Therefore, the learned features would have rich identity-related
or inter-personal variations. However, the identification signal has a relatively weak constraint on
DeepID2 features extracted from the same identity, since dissimilar DeepID2 features could be
mapped to the same identity through function g(?). This leads to problems when DeepID2 features
are generalized to new tasks and new identities in test where g is not applicable anymore. We solve
this by using an additional face verification signal, which requires that every two DeepID2 feature
vectors extracted from the same identity are close to each other while those extracted from different
identities are kept away. The strong per-element constraint on DeepID2 features can effectively
reduce the intra-personal variations. On the other hand, using the verification signal alone (i.e. only
distinguishing a pair of DeepID2 feature vectors at a time) is not as effective in extracting identityrelated features as using the identification signal (i.e. distinguishing thousands of identities at a
time). Therefore, the two supervisory signals emphasize different aspects in feature learning and
should be employed together.
To characterize faces from different aspects, complementary DeepID2 features are extracted from
various face regions and resolutions, and are concatenated to form the final feature representation
after PCA dimension reduction. Since the learned DeepID2 features are diverse among different
identities while consistent within the same identity, it makes the following face recognition easier.
Using the learned feature representation and a recently proposed face verification model [3], we
achieved the highest 99.15% face verification accuracy on the challenging and extensively studied
LFW dataset [11]. This is the first time that a machine provided with only the face region achieves an
accuracy on par with the 99.20% accuracy of human to whom the entire LFW face image including
the face region and large background area are presented to verify.
In recent years, a great deal of efforts have been made for face recognition with deep learning
[5, 10, 18, 26, 8, 21, 20, 27]. Among the deep learning works, [5, 18, 8] learned features or
deep metrics with the verification signal, while DeepFace [21] and our previous work DeepID
[20] learned features with the identification signal and achieved accuracies around 97.45% on
LFW. Our approach significantly improves the state-of-the-art. The idea of jointly solving the
classification and verification tasks was applied to general object recognition [15], with the focus on
improving classification accuracy on fixed object classes instead of hidden feature representations.
Our work targets on learning features which can be well generalized to new classes (identities) and
the verification task.
2
Identification-verification guided deep feature learning
We learn features with variations of deep convolutional neural networks (deep ConvNets) [12].
The convolution and pooling operations in deep ConvNets are specially designed to extract visual
features hierarchically, from local low-level features to global high-level ones. Our deep ConvNets
take similar structures as in [20]. It contains four convolutional layers, with local weight sharing
[10] in the third and fourth convolutional layers. The ConvNet extracts a 160-dimensional DeepID2
feature vector at its last layer (DeepID2 layer) of the feature extraction cascade. The DeepID2
layer to be learned are fully-connected to both the third and fourth convolutional layers. We use
rectified linear units (ReLU) [17] for neurons in the convolutional layers and the DeepID2 layer.
An illustration of the ConvNet structure used to extract DeepID2 features is shown in Fig. 1 given
an RGB input of size 55 ? 47. When the size of the input region changes, the map sizes in the
following layers will change accordingly. The DeepID2 feature extraction process is denoted as
f = Conv(x, ?c ), where Conv(?) is the feature extraction function defined by the ConvNet, x is the
input face patch, f is the extracted DeepID2 feature vector, and ?c denotes ConvNet parameters to
be learned.
2
Figure 1: The ConvNet structure for DeepID2 feature extraction.
DeepID2 features are learned with two supervisory signals. The first is face identification signal,
which classifies each face image into one of n (e.g., n = 8192) different identities. Identification is
achieved by following the DeepID2 layer with an n-way softmax layer, which outputs a probability
distribution over the n classes. The network is trained to minimize the cross-entropy loss, which we
call the identification loss. It is denoted as
Ident(f, t, ?id ) = ?
n
X
pi log p?i = ? log p?t ,
(1)
i=1
where f is the DeepID2 feature vector, t is the target class, and ?id denotes the softmax layer
parameters. pi is the target probability distribution, where pi = 0 for all i except pt = 1
for the target class t. p?i is the predicted probability distribution. To correctly classify all
the classes simultaneously, the DeepID2 layer must form discriminative identity-related features
(i.e. features with large inter-personal variations). The second is face verification signal, which
encourages DeepID2 features extracted from faces of the same identity to be similar. The verification
signal directly regularize DeepID2 features and can effectively reduce the intra-personal variations.
Commonly used constraints include the L1/L2 norm and cosine similarity. We adopt the following
loss function based on the L2 norm, which was originally proposed by Hadsell et al.[7] for
dimensionality reduction,
(
Verif(fi , fj , yij , ?ve ) =
1
2
1
2
2
kfi ? fj k2
2
max 0, m ? kfi ? fj k2
if yij = 1
,
if yij = ?1
(2)
where fi and fj are DeepID2 feature vectors extracted from the two face images in comparison.
yij = 1 means that fi and fj are from the same identity. In this case, it minimizes the L2 distance
between the two DeepID2 feature vectors. yij = ?1 means different identities, and Eq. (2) requires
the distance larger than a margin m. ?ve = {m} is the parameter to be learned in the verification loss
function. Loss functions based on the L1 norm could have similar formulations [15]. The cosine
similarity was used in [17] as
Verif(fi , fj , yij , ?ve ) =
1
2
(yij ? ?(wd + b)) ,
2
(3)
f ?f
where d = kfi ki2 kfjj k2 is the cosine similarity between DeepID2 feature vectors, ?ve = {w, b} are
learnable scaling and shifting parameters, ? is the sigmoid function, and yij is the binary target of
whether the two compared face images belong to the same identity. All the three loss functions are
evaluated and compared in our experiments.
Our goal is to learn the parameters ?c in the feature extraction function Conv(?), while ?id and ?ve are
only parameters introduced to propagate the identification and verification signals during training.
In the testing stage, only ?c is used for feature extraction. The parameters are updated by stochastic
gradient descent. The identification and verification gradients are weighted by a hyperparameter ?.
Our learning algorithm is summarized in Tab. 1. The margin m in Eq. (2) is a special case, which
cannot be updated by gradient descent since this will collapse it to zero. Instead, m is fixed and
updated every N training pairs (N ? 200, 000 in our experiments) such that it is the threshold of
3
Table 1: The DeepID2 feature learning algorithm.
input: training set ? = {(xi , li )}, initialized parameters ?c , ?id , and ?ve , hyperparameter ?, learning rate ?(t), t ? 0
while not converge do
t ? t + 1 sample two training samples (xi , li ) and (xj , lj ) from ?
fi = Conv(xi , ?c ) and fj = Conv(xj , ?c )
? Ident(fj ,lj ,?id )
(fi ,li ,?id )
??id = ? Ident??
+
??id
id
? Verif(fi ,fj ,yij ,?ve )
??ve = ? ?
,
where
yij = 1 if li = lj , and yij = ?1 otherwise.
??ve
? Verif(fi ,fj ,yij ,?ve )
? Ident(fi ,li ,?id )
?fi =
+??
?fi
?fi
? Verif(fi ,fj ,yij ,?ve )
? Ident(fj ,lj ,?id )
+
?
?
?fj =
?fj
?fj
? Conv(xj ,?c )
? Conv(xi ,?c )
+ ?fj ?
??c = ?fi ?
??c
??c
update ?id = ?id ? ?(t) ? ??id , ?ve = ?ve ? ?(t) ? ??ve , and ?c = ?c ? ?(t) ? ??c .
end while
output ?c
Figure 2: Patches selected for feature extraction. The Joint Bayesian [3] face verification accuracy
(%) using features extracted from each individual patch is shown below.
the feature distances kfi ? fj k to minimize the verification error of the previous N training pairs.
Updating m is not included in Tab. 1 for simplicity.
3
Face Verification
To evaluate the feature learning algorithm described in Sec. 2, DeepID2 features are embedded into
the conventional face verification pipeline of face alignment, feature extraction, and face verification.
We first use the recently proposed SDM algorithm [24] to detect 21 facial landmarks. Then the face
images are globally aligned by similarity transformation according to the detected landmarks. We
cropped 400 face patches, which vary in positions, scales, color channels, and horizontal flipping,
according to the globally aligned faces and the position of the facial landmarks. Accordingly,
400 DeepID2 feature vectors are extracted by a total of 200 deep ConvNets, each of which is
trained to extract two 160-dimensional DeepID2 feature vectors on one particular face patch and
its horizontally flipped counterpart, respectively, of each face.
To reduce the redundancy among the large number of DeepID2 features and make our system
practical, we use the forward-backward greedy algorithm [25] to select a small number of effective
and complementary DeepID2 feature vectors (25 in our experiment), which saves most of the feature
extraction time during test. Fig. 2 shows all the selected 25 patches, from which 25 160-dimensional
DeepID2 feature vectors are extracted and are concatenated to a 4000-dimensional DeepID2 feature
vector. The 4000-dimensional vector is further compressed to 180 dimensions by PCA for face
verification. We learned the Joint Bayesian model [3] for face verification based on the extracted
DeepID2 features. Joint Bayesian has been successfully used to model the joint probability of two
faces being the same or different persons [3, 4].
4
4
Experiments
We report face verification results on the LFW dataset [11], which is the de facto standard test set
for face verification in unconstrained conditions. It contains 13, 233 face images of 5749 identities
collected from the Internet. For comparison purposes, algorithms typically report the mean face
verification accuracy and the ROC curve on 6000 given face pairs in LFW. Though being sound
as a test set, it is inadequate for training, since the majority of identities in LFW have only one
face image. Therefore, we rely on a larger outside dataset for training, as did by all recent highperformance face verification algorithms [4, 2, 21, 20, 13]. In particular, we use the CelebFaces+
dataset [20] for training, which contains 202, 599 face images of 10, 177 identities (celebrities)
collected from the Internet. People in CelebFaces+ and LFW are mutually exclusive. DeepID2
features are learned from the face images of 8192 identities randomly sampled from CelebFaces+
(referred to as CelebFaces+A), while the remaining face images of 1985 identities (referred to as
CelebFaces+B) are used for the following feature selection and learning the face verification models
(Joint Bayesian). When learning DeepID2 features on CelebFaces+A, CelebFaces+B is used as
a validation set to decide the learning rate, training epochs, and hyperparameter ?. After that,
CelebFaces+B is separated into a training set of 1485 identities and a validation set of 500 identities
for feature selection. Finally, we train the Joint Bayesian model on the entire CelebFaces+B data
and test on LFW using the selected DeepID2 features. We first evaluate various aspect of feature
learning from Sec. 4.1 to Sec. 4.3 by using a single deep ConvNet to extract DeepID2 features
from the entire face region. Then the final system is constructed and compared with existing best
performing methods in Sec. 4.4.
4.1
Balancing the identification and verification signals
We investigates the interactions of identification and verification signals on feature learning, by
varying ? from 0 to +?. At ? = 0, the verification signal vanishes and only the identification signal
takes effect. When ? increases, the verification signal gradually dominates the training process. At
the other extreme of ? ? +?, only the verification signal remains. The L2 norm verification loss
in Eq. (2) is used for training. Figure 3 shows the face verification accuracy on the test set by
comparing the learned DeepID2 features with L2 norm and the Joint Bayesian model, respectively.
It clearly shows that neither the identification nor the verification signal is the optimal one to learn
features. Instead, effective features come from the appropriate combination of the two.
This phenomenon can be explained from the view of inter- and intra-personal variations, which
could be approximated by LDA. According to LDA, the inter-personal scatter matrix is Sinter =
Pc
>
xi ? x
?) (?
xi ? x
?) , where x
?i is the mean feature of the i-th identity, x
? is the mean of the
i=1 ni ? (?
entire dataset, and ni is the number of face images of the i-th identity. The intra-personal scatter
Pc P
>
matrix is Sintra =
?i ) (x ? x
?i ) , where Di is the set of features of the i-th
i=1
x?Di (x ? x
identity, x
?i is the corresponding mean, and c is the number of different identities. The inter- and
intra-personal variances are the eigenvalues of the corresponding scatter matrices, and are shown in
Fig. 5. The corresponding eigenvectors represent different variation patterns. Both the magnitude
and diversity of feature variances matter in recognition. If all the feature variances concentrate on a
small number of eigenvectors, it indicates the diversity of intra- or inter-personal variations is low.
The features are learned with ? = 0, 0.05, and +?, respectively. The feature variances of each
given ? are normalized by the corresponding mean feature variance.
When only the identification signal is used (? = 0), the learned features contain both diverse
inter- and intra-personal variations, as shown by the long tails of the red curves in both figures.
While diverse inter-personal variations help to distinguish different identities, large and diverse
intra-personal variations are disturbing factors and make face verification difficult. When both the
identification and verification signals are used with appropriate weighting (? = 0.05), the diversity
of the inter-personal variations keeps unchanged while the variations in a few main directions
become even larger, as shown by the green curve in the left compared to the red one. At the
same time, the intra-personal variations decrease in both the diversity and magnitude, as shown
by the green curve in the right. Therefore, both the inter- and intra-personal variations changes in
a direction that makes face verification easier. When ? further increases towards infinity, both the
inter- and intra-personal variations collapse to the variations in only a few main directions, since
without the identification signal, diverse features cannot be formed. With low diversity on inter5
Figure 3: Face verification accuracy by varying Figure 4: Face verification accuracy of DeepID2
the weighting parameter ?. ? is plotted in log features learned by both the the face identification
scale.
and verification signals, where the number of
training identities (shown in log scale) used for
face identification varies. The result may be
further improved with more than 8192 identities.
Figure 5: Spectrum of eigenvalues of the inter- and intra-personal scatter matrices. Best viewed in
color.
personal variations, distinguishing different identities becomes difficult. Therefore the performance
degrades significantly.
Figure 6 shows the first two PCA dimensions of features learned with ? = 0, 0.05, and +?,
respectively. These features come from the six identities with the largest numbers of face images in
LFW, and are marked by different colors. The figure further verifies our observations. When ? = 0
(left), different clusters are mixed together due to the large intra-personal variations, although the
cluster centers are actually different. When ? increases to 0.05 (middle), intra-personal variations
are significantly reduced and the clusters become distinguishable. When ? further increases towards
infinity (right), although the intra-personal variations further decrease, the cluster centers also begin
to collapse and some clusters become significantly overlapped (as the red, blue, and cyan clusters in
Fig. 6 right), making it hard to distinguish again.
4.2
Rich identity information improves feature learning
We investigate how would the identity information contained in the identification supervisory signal
influence the learned features. In particular, we experiment with an exponentially increasing number
of identities used for identification during training from 32 to 8192, while the verification signal is
generated from all the 8192 training identities all the time. Fig. 4 shows how the verification
accuracies of the learned DeepID2 features (derived from the L2 norm and Joint Bayesian) vary
on the test set with the number of identities used in the identification signal. It shows that
6
Figure 6: The first two PCA dimensions of DeepID2 features extracted from six identities in LFW.
Table 2: Comparison of different verification signals.
verification signal
L2
L2+
L2-
L1
cosine none
L2 norm (%)
Joint Bayesian (%)
94.95
95.12
94.43
94.87
86.23
92.98
92.92
94.13
87.07
93.38
86.43
92.73
identifying a large number (e.g., 8192) of identities is key to learning effective DeepID2 feature
representation. This observation is consistent with those in Sec. 4.1. The increasing number of
identities provides richer identity information and helps to form DeepID2 features with diverse interpersonal variations, making the class centers of different identities more distinguishable.
4.3
Investigating the verification signals
As shown in Sec. 4.1, the verification signal with moderate intensity mainly takes the effect of
reducing the intra-personal variations. To further verify this, we compare our L2 norm verification
signal on all the sample pairs with those only constrain either the positive or negative sample pairs,
denoted as L2+ and L2-, respectively. That is, the L2+ only decreases the distances between
DeepID2 features of the same identity, while L2- only increases the distances between DeepID2
features of different identities if they are smaller than the margin. The face verification accuracies
of the learned DeepID2 features on the test set, measured by the L2 norm and Joint Bayesian
respectively, are shown in Table 2. It also compares with the L1 norm and cosine verification signals,
as well as no verification signal (none). The identification signal is the same (classifying the 8192
identities) for all the comparisons.
DeepID2 features learned with the L2+ verification signal are only slightly worse than those learned
with L2. In contrast, the L2- verification signal helps little in feature learning and gives almost
the same result as no verification signal is used. This is a strong evidence that the effect of the
verification signal is mainly reducing the intra-personal variations. Another observation is that the
face verification accuracy improves in general whenever the verification signal is added in addition
to the identification signal. However, the L2 norm is better than the other compared verification
metrics. This may be due to that all the other constraints are weaker than L2 and less effective in
reducing the intra-personal variations. For example, the cosine similarity only constrains the angle,
but not the magnitude.
4.4
Final system and comparison with other methods
Before learning Joint Bayesian, DeepID2 features are first projected to 180 dimensions by PCA.
After PCA, the Joint Bayesian model is trained on the entire CelebFaces+B data and tested on the
6000 given face pairs in LFW, where the log-likelihood ratio given by Joint Bayesian is compared
to a threshold optimized on the training data for face verification. Tab. 3 shows the face verification
accuracy with an increasing number of face patches to extract DeepID2 features, as well as the time
used to extract those DeepID2 features from each face with a single Titan GPU. We achieve 98.97%
accuracy with all the 25 selected face patches. The feature extraction process is also efficient and
takes only 35 ms for each face image. The face verification accuracy of each individual face patch
is provided in Fig. 2. The short DeepID2 signature is extremely efficient for face identification and
face image search when matching a query image with a large number of candidates.
7
Table 3: Face verification accuracy with DeepID2 features extracted from an increasing number of
face patches.
# patches
1
2
4
8
16
25
accuracy (%)
time (ms)
95.43
1.7
97.28
3.4
97.75
6.1
98.55
11
98.93
23
98.97
35
Table 4: Accuracy comparison with the previous best results on LFW.
method
accuracy (%)
High-dim LBP [4]
TL Joint Bayesian [2]
DeepFace [21]
DeepID [20]
GaussianFace [13]
DeepID2
95.17 ? 1.13
96.33 ? 1.08
97.35 ? 0.25
97.45 ? 0.26
98.52 ? 0.66
99.15 ? 0.13
Figure 7: ROC comparison with the previous best results on LFW. Best viewed in color.
To further exploit the rich pool of DeepID2 features extracted from the large number of patches, we
repeat the feature selection algorithm for another six times, each time choosing DeepID2 features
from the patches that have not been selected by previous feature selection steps. Then we learn
the Joint Bayesian model on each of the seven groups of selected features, respectively. We fuse the
seven Joint Bayesian scores on each pair of compared faces by further learning an SVM. In this way,
we achieve an even higher 99.15% face verification accuracy. The accuracy and ROC comparison
with previous state-of-the-art methods on LFW are shown in Tab. 4 and Fig. 7, respectively. We
achieve the best results and improve previous results with a large margin.
5
Conclusion
This paper have shown that the effect of the face identification and verification supervisory signals
on deep feature representation coincide with the two aspects of constructing ideal features for face
recognition, i.e., increasing inter-personal variations and reducing intra-personal variations, and the
combination of the two supervisory signals lead to significantly better features than either one of
them. When embedding the learned features to the traditional face verification pipeline, we achieved
an extremely effective system with 99.15% face verification accuracy on LFW. The arXiv report of
this paper was published in June 2014 [19].
8
References
[1] P. N. Belhumeur, J. a. P. Hespanha, and D. J. Kriegman. Eigenfaces vs. Fisherfaces:
Recognition using class specific linear projection. PAMI, 19:711?720, 1997.
[2] X. Cao, D. Wipf, F. Wen, G. Duan, and J. Sun. A practical transfer learning algorithm for face
verification. In Proc. ICCV, 2013.
[3] D. Chen, X. Cao, L. Wang, F. Wen, and J. Sun. Bayesian face revisited: A joint formulation.
In Proc. ECCV, 2012.
[4] D. Chen, X. Cao, F. Wen, and J. Sun. Blessing of dimensionality: High-dimensional feature
and its efficient compression for face verification. In Proc. CVPR, 2013.
[5] S. Chopra, R. Hadsell, and Y. LeCun. Learning a similarity metric discriminatively, with
application to face verification. In Proc. CVPR, 2005.
[6] M. Guillaumin, J. Verbeek, and C. Schmid. Is that you? Metric learning approaches for face
identification. In Proc. ICCV, 2009.
[7] R. Hadsell, S. Chopra, and Y. LeCun. Dimensionality reduction by learning an invariant
mapping. In Proc. CVPR, 2006.
[8] J. Hu, J. Lu, and Y.-P. Tan. Discriminative deep metric learning for face verification in the
wild. In Proc. CVPR, 2014.
[9] C. Huang, S. Zhu, and K. Yu. Large scale strongly supervised ensemble metric learning, with
applications to face verification and retrieval. NEC Technical Report TR115, 2011.
[10] G. B. Huang, H. Lee, and E. Learned-Miller. Learning hierarchical representations for face
verification with convolutional deep belief networks. In Proc. CVPR, 2012.
[11] G. B. Huang, M. Ramesh, T. Berg, and E. Learned-Miller. Labeled Faces in the Wild: A
database for studying face recognition in unconstrained environments. Technical Report 0749, University of Massachusetts, Amherst, 2007.
[12] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document
recognition. Proceedings of the IEEE, 1998.
[13] C. Lu and X. Tang. Surpassing human-level face verification performance on LFW with
GaussianFace. Technical report, arXiv:1404.3840, 2014.
[14] A. Mignon and F. Jurie. PCCA: A new approach for distance learning from sparse pairwise
constraints. In Proc. CVPR, 2012.
[15] H. Mobahi, R. Collobert, and J. Weston. Deep learning from temporal coherence in video. In
Proc. ICML, 2009.
[16] B. Moghaddam, T. Jebara, and A. Pentland. Bayesian face recognition. PR, 33:1771?1782,
2000.
[17] V. Nair and G. E. Hinton. Rectified linear units improve restricted Boltzmann machines. In
Proc. ICML, 2010.
[18] Y. Sun, X. Wang, and X. Tang. Hybrid deep learning for face verification. In Proc. ICCV,
2013.
[19] Y. Sun, X. Wang, and X. Tang. Deep learning face representation by joint identificationverification. Technical report, arXiv:1406.4773, 2014.
[20] Y. Sun, X. Wang, and X. Tang. Deep learning face representation from predicting 10,000
classes. In Proc. CVPR, 2014.
[21] Y. Taigman, M. Yang, M. Ranzato, and L. Wolf. DeepFace: Closing the gap to human-level
performance in face verification. In Proc. CVPR, 2014.
[22] X. Wang and X. Tang. Unified subspace analysis for face recognition. In Proc. ICCV, 2003.
[23] X. Wang and X. Tang. A unified framework for subspace face recognition. PAMI, 26:1222?
1228, 2004.
[24] X. Xiong and F. De la Torre Frade. Supervised descent method and its applications to face
alignment. In Proc. CVPR, 2013.
[25] T. Zhang. Adaptive forward-backward greedy algorithm for learning sparse representations.
IEEE Trans. Inf. Theor., 57:4689?4708, 2011.
[26] Z. Zhu, P. Luo, X. Wang, and X. Tang. Deep learning identity-preserving face space. In Proc.
ICCV, 2013.
[27] Z. Zhu, P. Luo, X. Wang, and X. Tang. Deep learning and disentangling face representation by
multi-view perceptron. In Proc. NIPS, 2014.
9
| 5416 |@word kong:2 middle:1 compression:1 norm:11 hu:1 propagate:1 rgb:1 reduction:3 contains:3 score:1 document:1 existing:1 com:1 wd:1 comparing:1 luo:2 gmail:1 scatter:5 must:1 gpu:1 designed:2 update:1 v:1 alone:1 greedy:2 selected:6 accordingly:2 short:1 provides:2 revisited:1 zhang:1 constructed:1 become:3 wild:2 pairwise:1 inter:18 nor:1 multi:1 globally:2 duan:1 little:1 increasing:5 conv:7 provided:2 classifies:1 becomes:1 begin:1 minimizes:1 unified:3 transformation:1 temporal:1 every:2 k2:3 facto:1 unit:2 positive:1 before:1 engineering:2 local:2 tends:1 id:14 pami:2 studied:1 challenging:3 limited:1 collapse:3 kfi:4 jurie:1 practical:2 lecun:3 testing:2 area:1 significantly:6 cascade:1 projection:2 matching:1 sensetime:1 cannot:2 close:2 selection:4 influence:1 conventional:1 map:2 center:3 resolution:1 hadsell:3 simplicity:1 identifying:1 chen2:1 pull:1 regularize:1 embedding:1 handle:1 variation:36 updated:3 target:5 pt:1 tan:1 distinguishing:3 overlapped:1 element:1 recognition:17 approximated:1 updating:1 labeled:1 database:1 observed:1 solved:1 wang:8 thousand:1 region:5 connected:1 sun:6 ranzato:1 decrease:3 highest:1 vanishes:1 environment:1 constrains:1 kriegman:1 personal:34 signature:1 trained:3 solving:1 joint:19 xiaoou:1 various:2 train:1 separated:1 effective:8 detected:1 query:1 outside:1 choosing:1 richer:1 larger:3 solve:1 cvpr:9 drawing:1 wang3:1 otherwise:1 compressed:1 unseen:2 jointly:1 final:3 sdm:1 eigenvalue:2 interaction:1 aligned:2 cao:3 achieve:3 academy:1 cluster:6 object:2 help:3 develop:1 pose:1 measured:1 eq:3 strong:2 predicted:1 come:2 direction:4 guided:1 concentrate:1 torre:1 stochastic:1 human:3 theor:1 yij:13 around:1 great:1 mapping:2 achieves:1 early:1 adopt:1 vary:2 purpose:1 proc:18 applicable:1 largest:1 successfully:1 tool:1 weighted:1 clearly:1 varying:2 derived:1 focus:1 june:1 indicates:1 mainly:2 likelihood:1 hk:3 contrast:1 detect:1 dim:1 entire:5 lj:4 typically:1 hidden:2 frade:1 classification:3 among:3 denoted:3 art:2 softmax:2 special:1 extraction:10 flipped:1 look:1 yu:1 icml:2 wipf:1 report:7 few:2 wen:3 randomly:1 simultaneously:2 ve:14 individual:2 occlusion:1 highly:1 investigate:1 intra:24 alignment:2 extreme:1 pc:2 moghaddam:1 facial:2 initialized:1 plotted:1 classify:3 interpersonal:1 inadequate:1 characterize:1 tang1:1 varies:1 thanks:1 person:1 amherst:1 ie:2 stay:1 lee:1 pool:1 together:3 again:1 central:1 huang:3 worse:1 highperformance:1 li:5 de:2 diversity:5 summarized:1 sec:6 matter:1 titan:1 explicitly:1 collobert:1 view:2 tab:4 red:3 minimize:2 formed:1 ni:2 accuracy:24 convolutional:7 variance:5 ensemble:1 miller:2 shenzhen:1 weak:1 identification:35 bayesian:18 none:2 lu:2 rectified:2 published:1 classified:1 sharing:1 whenever:1 guillaumin:1 di:2 sampled:1 dataset:6 massachusetts:1 color:4 improves:3 dimensionality:3 carefully:1 actually:1 back:1 originally:1 higher:1 supervised:2 improved:1 formulation:2 evaluated:1 though:1 strongly:1 stage:3 convnets:4 hand:1 horizontal:1 ident:5 nonlinear:3 celebrity:1 lda:4 pulling:1 supervisory:7 effect:4 verify:2 normalized:1 contain:1 counterpart:1 deal:1 during:3 encourages:1 cosine:6 hong:2 generalized:4 m:2 l1:4 fj:17 image:19 recently:2 fi:14 sigmoid:1 exponentially:1 belong:1 tail:1 approximates:1 surpassing:1 unconstrained:3 closing:1 supervision:1 similarity:6 celebfaces:10 recent:3 moderate:1 apart:3 inf:1 binary:2 yi:1 verif:5 preserving:1 additional:1 employed:1 belhumeur:1 converge:1 maximize:1 cuhk:3 signal:47 sound:1 reduces:1 technical:4 cross:1 long:1 retrieval:1 verbeek:1 lfw:18 metric:7 arxiv:3 represent:1 achieved:5 background:1 cropped:1 addition:1 lbp:1 specially:1 pooling:1 call:1 extracting:1 ee:1 chopra:2 yuheng:1 ideal:1 yang:1 bengio:1 xj:3 sun1:1 relu:1 architecture:1 reduce:3 idea:1 haffner:1 whether:1 expression:1 pca:6 six:3 effort:1 deep:27 eigenvectors:2 extensively:1 reduced:2 per:1 correctly:1 blue:1 diverse:6 hyperparameter:3 group:2 key:2 four:1 redundancy:1 threshold:2 traced:1 neither:1 kept:1 backward:2 fuse:1 year:1 taigman:1 angle:1 powerful:1 fourth:2 you:1 almost:1 decide:1 electronic:1 patch:13 coherence:1 scaling:1 investigates:1 layer:14 internet:2 cyan:1 distinguish:2 xiaogang:1 constraint:5 infinity:2 constrain:1 xgwang:1 aspect:4 pcca:1 extremely:2 performing:1 relatively:1 department:2 according:3 combination:2 belonging:1 smaller:1 slightly:1 shallow:1 making:2 explained:1 gradually:1 iccv:5 invariant:1 pr:1 restricted:1 pipeline:2 mutually:1 remains:1 overwhelm:1 end:1 studying:1 operation:1 hierarchical:3 away:1 appropriate:2 anymore:1 save:1 xiong:1 top:1 denotes:2 include:1 remaining:1 exploit:1 concatenated:2 chinese:3 especially:1 unchanged:1 added:1 flipping:1 degrades:1 exclusive:1 traditional:1 gradient:4 subspace:4 convnet:6 distance:6 mapped:2 capacity:1 landmark:3 majority:1 topic:1 whom:1 argue:1 collected:2 seven:2 illustration:1 ratio:2 difficult:2 disentangling:1 hespanha:1 negative:1 boltzmann:1 fisherfaces:1 convolution:1 neuron:1 observation:3 ramesh:1 descent:3 pentland:1 hinton:1 jebara:1 intensity:1 introduced:1 pair:9 optimized:1 learned:30 nip:1 trans:1 below:1 pattern:1 challenge:1 including:1 max:1 green:2 belief:1 shifting:1 video:1 rely:1 hybrid:1 predicting:1 advanced:1 zhu:3 improve:2 technology:1 extract:7 schmid:1 epoch:1 l2:21 embedded:1 fully:1 par:1 loss:7 discriminatively:1 mixed:1 age:1 validation:2 verification:82 consistent:2 classifying:1 pi:3 balancing:1 eccv:1 ki2:1 last:1 repeat:1 weaker:1 perceptron:1 institute:1 eigenfaces:1 face:112 sparse:2 curve:4 dimension:5 rich:3 forward:2 made:1 commonly:1 disturbing:1 projected:1 coincide:1 adaptive:1 emphasize:1 implicitly:1 keep:1 global:1 investigating:1 xtang:1 discriminative:2 xi:6 spectrum:1 search:1 table:5 nature:1 learn:5 channel:1 transfer:1 improving:1 bottou:1 complex:1 constructing:1 did:1 hierarchically:1 main:2 verifies:1 complementary:2 fig:7 referred:3 tl:1 roc:3 position:2 candidate:1 third:2 weighting:2 tang:8 enlarging:2 specific:1 mobahi:1 learnable:1 svm:1 dominates:1 evidence:1 essential:2 effectively:2 inter5:1 magnitude:3 nec:1 illumination:1 margin:4 chen:2 easier:2 gap:1 entropy:1 distinguishable:2 visual:1 horizontally:1 contained:1 deepface:3 wolf:1 extracted:17 nair:1 weston:1 identity:59 targeted:1 goal:2 viewed:2 marked:1 towards:2 change:3 hard:1 included:1 except:1 reducing:6 total:1 blessing:1 la:1 select:1 berg:1 people:1 dissimilar:1 evaluate:2 tested:1 phenomenon:1 |
4,879 | 5,417 | Fast Training of Pose Detectors in the Fourier Domain
Jo?ao F. Henriques
Pedro Martins
Rui Caseiro
Jorge Batista
Institute of Systems and Robotics
University of Coimbra
{henriques,pedromartins,ruicaseiro,batista}@isr.uc.pt
Abstract
In many datasets, the samples are related by a known image transformation, such
as rotation, or a repeatable non-rigid deformation. This applies to both datasets
with the same objects under different viewpoints, and datasets augmented with
virtual samples. Such datasets possess a high degree of redundancy, because
geometrically-induced transformations should preserve intrinsic properties of the
objects. Likewise, ensembles of classifiers used for pose estimation should also
share many characteristics, since they are related by a geometric transformation.
By assuming that this transformation is norm-preserving and cyclic, we propose a
closed-form solution in the Fourier domain that can eliminate most redundancies.
It can leverage off-the-shelf solvers with no modification (e.g. libsvm), and train
several pose classifiers simultaneously at no extra cost. Our experiments show that
training a sliding-window object detector and pose estimator can be sped up by orders of magnitude, for transformations as diverse as planar rotation, the walking
motion of pedestrians, and out-of-plane rotations of cars.
1
Introduction
To cope with the rich variety of transformations in natural images, recognition systems require a
representative sample of possible variations. Some of those variations must be learned from data
(e.g. non-rigid deformations), while others can be virtually generated (e.g. translation or rotation).
Recently, there has been a renewed interest in augmenting datasets with virtual samples, both in the
context of supervised [23, 17] and unsupervised learning [6]. This augmentation has the benefits of
regularizing high-capacity classifiers [6], while learning the natural invariances of the visual world.
Some kinds of virtual samples can actually make learning easier ? for example, with horizontallyflipped virtual samples [7, 4, 17], half of the weights of the template in the Dalal-Triggs detector [4]
become redundant by horizontal symmetry. A number of very recent works [14, 13, 8, 1] have shown
that cyclically translated virtual samples also constrain learning problems, which allows impressive
gains in computational efficiency. The core of this technique relies on approximately diagonalizing
the data matrix with the Discrete Fourier Transform (DFT).
In this work, we show that the ?Fourier trick? is not unique to cyclic translation, but can be generalized to other cyclic transformations. Our model captures a wide range of useful image transformations, yet retains the ability to accelerate training with the DFT. As it is only implicit, we can
accelerate training in both datasets of virtual samples and natural datasets with pose annotations.
Also due to the geometrically-induced structure of the training data, our algorithm can obtain several
transformed pose classifiers simultaneously. Some of the best object detection and pose estimation
systems currently learn classifiers for different poses independently [10, 7, 19], and we show how
joint learning of these classifiers can dramatically reduce training times.
1
(a)
(b)
(c)
Figure 1: (a) The horizontal translation of a 6 ? 6 image, by 1 pixel, can be achieved by a 36 ? 36 permutation
matrix P that reorders elements appropriately (depicted is the reordering of 2 pixels). (b) Rotation by a fixed
angle, with linearly-interpolated pixels, requires a more general matrix Q. By studying its influence on a dataset
of rotated samples, we show how to accelerate learning in the Fourier domain. Our model can also deal with
other transformations, including non-rigid. (c) Example HOG template (a car from the Google Earth dataset)
at 4 rotations learned by our model. Positive weights are on the first and third column, others are negative.
1.1
Contributions
Our contributions are as follows: 1) We generalize a previous successful model for translation
[14, 13] to other transformations, and analyze the properties of datasets with many transformed
images; 2) We present closed-form solutions that fully exploit the known structure of these datasets,
for Ridge Regression and Support Vector Regression, based on the DFT and off-the-shelf solvers;
3) With the same computational cost, we show how to train multiple classifiers for different poses
simultaneously; 4) Since our formulas do not require explicitly estimating or knowing the transformation, we demonstrate applicability to both datasets of virtual samples and structured datasets with
pose annotations. We achieve performance comparable to naive algorithms on 3 widely different
tasks, while being several orders of magnitude faster.
1.2
Related work
There is a vast body of works on image transformations and invariances, of which we can only mention a few. Much of the earlier computer vision literature focused on finding viewpoint-invariant
patterns [22]. They were based on image or scene-space coordinates, on which geometric transformations can be applied directly, however they do not apply to modern appearance-based representations. To relate complex transformations with appearance descriptors, a classic approach is to use
tangent vectors [3, 26, 16], which represent a first-order approximation. However, the desire for
more expressiveness has motivated the search for more general models.
Recent works have begun to approximate transformations as matrix-vector products, and try to estimate the transformation matrix explicitly. Tamaki et al. [27] do so for blur and affine transformations
in the context of LDA, while Miao et al. [21] approximate affine transformations with an E-M algorithm, based on a Lie group formulation. They estimate a basis for the transformation operator or the
transformed images, which is a hard analytical/inference problem in itself. The involved matrices
are extremely large for moderately-sized images, necessitating dimensionality reduction techniques
such as PCA, which may be suboptimal.
Several works focus on rotation alone [25, 18, 28, 2], most of them speeding up computations using
Fourier analysis, but they all explicitly estimate a reduced basis on which to project the data. Another
approach is to learn a transformation from data, using more parsimonious factored or deep models
[20]. In contrast, our method generalizes to other transformations and avoids a potentially costly
transformation model or basis estimation.
2
The cyclic orthogonal model for image transformations
Consider the m ? 1 vector x, obtained by vectorizing an image, i.e. stacking its elements into a
vector. The particular order does not matter, as long as it is consistent. The image may be a 32
dimensional array that contains multiple channels, such as RGB, or the values of a densely-sampled
image descriptor.
We wish to quickly train a classifier or regressor with transformed versions of sample images, to
make it robust to those transformations. The model we will use is an m ? m orthogonal matrix Q,
which will represent an incremental transformation of an image as Qx (for example, a small translation or rotation, see Fig. 1-a and 1-b). We can traverse different poses w.r.t. that transformation,
p ? Z, by repeated application of Q with a matrix power, Qp x.
In order for the number of poses to be finite, we must require the transformation to be cyclic, Qs =
Q0 = I, with some period s. This allows us to store all versions of x transformed to different poses
as the rows of an s ? m matrix,
?
T ?
Q0 x
T ?
?
? Q1 x
?
?
(1)
CQ (x) = ?
..
?
?
?
?
.
T
Qs?1 x
Due to Q being cyclic, any pose p ? Z can be found in the row (p mod s) + 1. Note that the first
row of CQ (x) contains the untransformed image x, since Q0 is the identity I. For the purposes of
training a classifier, CQ (x) can be seen as a data matrix, with one sample per row.
Although conceptually simple, we will show through experiments that this model can accurately
capture a variety of natural transformations (Section 5.2). More importantly, we will show that Q
never has to be created explicitly. The algorithms we develop will be entirely data-driven, using
an implicit description of Q from a structured dataset, either composed of virtual samples (e.g., by
image rotation), or natural samples (e.g. using pose annotations).
2.1
Image translation as a special case
A particular case of Q, and indeed what inspired the generalization that we propose, is the s ? s
cyclic shift matrix
T
0s?1
1
P =
,
(2)
Is?1 0s?1
where 0s?1 is an (s?1)?1 vector of zeros. This matrix cyclically permutes the elements of a vector
x as (x1 , x2 , x3 , . . . , xs ) ? (xs , x1 , x2 , . . . , xs?1 ). If x is a one-dimensional horizontal image, with
a single channel, then it is translated to the right by one pixel. An illustration is shown in Fig. 1-a.
By exploiting its relationship with the Discrete Fourier Transform (DFT), the cyclic shift model has
been used to accelerate a variety of learning algorithms in computer vision [14, 13, 15, 8, 1], with
suitable extensions to 2D and multiple channels.
2.2
Circulant matrices and the Discrete Fourier Transform
The basis for this optimization is the fact that the data matrix CP (x), or C(x) for short, formed by
all cyclic shifts of a sample image x, is circulant [5]. All circulant matrices are diagonalized by the
DFT, which can be expressed as the eigendecomposition
C(x) = U diag (F(x)) U H ,
(3)
where .H is the Hermitian transpose (i.e., transposition and complex-conjugation), F(x) denotes the
DFT of a vector x, and U is the unitary DFT basis. The constant matrix U can be used to compute
the DFT of any vector, since it satisfies U x = ?1s F(x). This is possible due to the linearity of the
DFT, though in practice the Fast Fourier Transform (FFT) algorithm is used instead. Note that U
is symmetric, U T = U , and unitary, U H = U ?1 . When working in Fourier-space, Eq. 3 shows
that circulant matrices in a learning problem become diagonal, which drastically reduces the needed
computations. For multiple channels or more images, they may become block-diagonal, but the
principles remain the same [13].
3
An important open question was whether the same diagonalization trick can be applied to image
transformations other than translation. We will show that this is true, using the model from Eq. 1.
3
Fast training with transformations of a single image
We will now focus on the main derivations of our paper, which allow us to quickly train a classifier
with virtual samples generated from an image x by repeated application of the transformation Q.
This section assumes only a single image x is given for training, which makes the presentation
simpler and we hope will give valuable insight into the core of the technique. Section 4 will expand
it to full generality, with training sets of an arbitrary number of images, all transformed by Q.
The first step is to show that some aspect of the data is diagonalizable by the DFT, which we do in
the following theorem.
Theorem 1. Given an orthogonal cyclic matrix Q, i.e. satisfying QT = Q?1 and Qs = Q0 , then
the s ? m matrix X = CQ (x) (from Eq. 1) verifies the following:
? The data matrix X and the uncentered covariance matrix X H X are not circulant in general,
unless Q = P (from Eq. 2).
? The Gram matrix G = XX H is always circulant.
Proof. See Appendix A.1.
Theorem 1 implies that the learning problem in its original form is not diagonalizable by the DFT
basis. However, the same diagonalization is possible for the dual problem, defined by the Gram
matrix G.
Because G is circulant, it has only s degrees of freedom and is fully specified by its first row g [11],
G = C(g). By direct computation from Eq. 1, we can verify that the elements of the first row g
are given by gp = xT Qp?1 x. One interpretation is that g contains the auto-correlation of x through
pose-space, i.e., the inner-product of x with itself as the transformation Q is applied repeatedly.
3.1
Dual Ridge Regression
For now we will restrict our attention to Ridge Regression (RR), since it has the appealing property
of having a solution in closed form, which we can easily manipulate. Section 4.1 will show how
to extend these results to Support Vector Regression. The goal of RR is to find the linear function
P
2
2
f (x) = wT x that minimizes a regularized squared error: i (f (xi ) ? yi ) + ? kwk .
Since we have s samples in the data matrix under consideration (Eq. 1), there are s dual variables,
?1
stored in a vector ?. The RR solution is given by ? = (G + ?I) y [24], where G = XX H is the
s ? s Gram matrix, y is the vector of s labels (one per pose), and ? is the regularization parameter.
The dual form of RR is usually associated with non-linear kernels [24], but since this is not our case
we can compute the explicit primal solution with w = X T ?, yielding
?1
w = X T (G + ?I)
y.
(4)
Applying the circulant eigendecomposition (Eq. 3) to G, and substituting it in Eq. 4,
?1
?1
w = X T U diag (?
g) U H + ?U U H
y = X T U (diag (?
g + ?)) U H y,
(5)
? = F (g), and similarly y
? = F (y). Since inversion of a
where we introduce the shorthand g
diagonal matrix can be done element-wise, and its multiplication by the vector U H y amounts to an
element-wise product, we obtain
?
y
,
(6)
w = X T F ?1
?+?
g
where F ?1 denotes the inverse DFT, and the division is taken element-wise. This formula allows us
to replace a costly matrix inversion with fast DFT and element-wise operations. We also do not need
to compute and store the full G, as the auto-correlation vector g suffices. As we will see in the next
section, there is a simple modification to Eq. 6 that turns out to be very useful for pose estimation.
4
3.2
Training several components simultaneously
A relatively straightforward way to estimate the object pose in an input image x is to train a classifier
for each pose (which we call components), evaluate all of them and take the maximum, i.e.
fpose (x) = arg max wpT x.
(7)
p
This can also be used as the basis for a pose-invariant classifier, by replacing argmax with max
[10]. Of course, training one component per pose can quickly become expensive. However, we can
exploit the fact that these training problems become tightly related when the training set contains
transformed images.
Recall that y specifies the labels for a training set of s transformed images, one label per pose.
Without any loss of generality, suppose that the label is 1 for a given pose t and 0 for all others, i.e.
y contains a single peak at element t. Then by shifting the peak with P p y, we will train a classifier
for pose t + p. In this manner we can train classifiers for all poses simply by varying the labels P p y,
with p = 0, . . . , s ? 1.
Based on Eq. 6, we can concatenate the solutions for all s components into a single m ? s matrix,
W =
w0
???
ws?1
?1
= X T (G + ?I)
?1
T
= X (G + ?I)
P 0y
???
P s?1 y
T
C (y) .
(8)
(9)
Diagonalization yields
??
y
F (X) ,
(10)
W =F
diag
?+?
g
where .? denotes complex-conjugation. Since their arguments are matrices, the DFT/IDFT operations here work along each column. The product of F (X) by the diagonal matrix simply amounts
to multiplying each of its rows by a scalar factor, which is inexpensive. Eq. 10 has nearly the same
computational cost as Eq. 6, which trains a single classifier.
T
4
?1
Transformation of multiple images
The training method described in the previous section would find little applicability for modern
recognition tasks if it remained limited to transformations of a single image. Naturally, we would
like to use
n images xi . We now
have a dataset of ns samples, which can be divided into n sample
groups Qp?1 xi |p = 1, . . . , s , each containing the transformed versions of one image.
This case becomes somewhat complicated by the fact that the data matrix X now has three dimensions ? the m features, the n sample groups, and the s poses of each sample group. In this m ? n ? s
array, each column vector (along the first dimension) is defined as
X?ip = Qp?1 xi ,
i = 1, . . . , n; p = 1, . . . , s,
(11)
where we have used ? to denote a one-dimensional slice of the three-dimensional array X.1 A twodimensional slice will be denoted by X??p , which yields a m ? n matrix, one for each p = 1, . . . , s.
Through a series of block-diagonalizations and reorderings, we can show (Appendix A.2-A.5) that
the solution W , of size m ? s, describing all s components (similarly to Eq. 10), is obtained with
?1 ?
? ?p = X
? ??p (?
W
g??p + ?I) Y??p
,
p = 1, . . . , s,
(12)
? is the DFT
where a hat ? over an array denotes the DFT along the dimension that has size s (e.g. X
of X along the third dimension), Yip specifies the label of the sample with pose p in group i, and g
is the n ? n ? s array with elements
1
For reference, our slice notation ? works the same way as the slice notation : in Matlab or NumPy.
5
T
gijp = xTi Qp?1 xj = X?i1
X?jp ,
i, j = 1, . . . , n; p = 1, . . . , s.
(13)
It may come as a surprise that, after all these changes, Eq. 12 still essentially looks like a dual Ridge
Regression (RR) problem (compare it to Eq. 4). Eq. 12 can be interpreted as splitting the original
problem into s smaller problems, one for each Fourier frequency, which are independent and can be
solved in parallel. A Matlab implementation is given in Appendix B.2
4.1
Support Vector Regression
Given that we can decompose such a large RR problem into s smaller RR problems, by applying
the DFT and slicing operators (Eq. 12), it is natural to ask whether the same can be done with
other algorithms. Leveraging a recent result [13], where this was done for image translation, the
same steps can be repeated for the dual formulation of other algorithms, such as Support Vector
Regression (SVR). Although RR can deal with complex data, SVR requires an extension to the
complex domain, which we show in Appendix A.6. We give a Matlab implementation in Appendix
B, which can use any off-the-shelf SVR solver without modification.
4.2
Efficiency
Naively training one detector per pose would require solving s large ns?ns systems (either with RR
or SVR). In contrast, our method learns jointly all detectors using s much smaller n?n subproblems.
The computational savings can be several orders of magnitude for large s. Our experiments seem to
validate this conclusion, even in relatively large recognition tasks (Section 6).
5
Orthogonal transformations in practice
Until now, we avoided the question of how to compute a transformation model Q. This may seem
like a computational burden, not to mention a hard estimation problem ? for example, what is the
cyclic orthogonal matrix Q that models planar rotations with period s? Inspecting Eq. 12-13,
however, reveals that we do not need to form Q explicitly, but can work with just a data matrix X
of transformed images. From there on, we exploit the knowledge that this data was obtained from
some matrix Q, and that is enough to allow fast training in the Fourier domain. This allows a great
deal of flexibility in implementation.
5.1
Virtual transformations
One way to obtain a structured data matrix X is with virtual samples. From the original dataset of n
samples, we can generate ns virtual samples using a standard image operator (e.g. planar rotation).
However, we should keep in mind that the accuracy of the proposed method will be affected by how
much the image operator resembles a pure cyclic orthogonal transformation.
Linearity. Many common image transformations, such as rotation or scale, are implemented by
nearest-neighbor or bilinear interpolation. For a fixed amount of rotation or scale, these functions
are linear functions in the input pixels, i.e. each output pixel is a fixed linear combination of some
of the input pixels. As such, they fulfill the linearity requirement.
Orthogonality. For an operator to be orthogonal, it must preserve the L2 norm of its inputs. At the
expense of introducing some non-linearity, we simply renormalize each virtual sample to have the
same norm as the original sample, which seems to work well in practice (Section 6).
Cyclicity. We conducted some experiments with planar rotation on satellite imagery (Section 6.1)
? rotation by 360/s degrees is cyclic with period s. In the future, we plan to experiment with noncyclic operators (similar to how cyclic translation is used to approximate image translation [14]).
2
The supplemental material is available at: www.isr.uc.pt/?henriques/transformations/
6
Figure 2: Example detections and estimated poses in 3 different settings. We can accelerate training with
(a) planar rotations (Google Earth), (b) non-rigid deformations in walking pedestrians (TUD-Campus/TUDCrossing), and (c) out-of-plane rotations (KITTI). Best viewed in color.
5.2
Natural transformations
Another interesting possibility is to use pose annotations to create a structured data matrix. This
data-driven approach allows us to consider more complicated transformations than those associated
with virtual samples. Given s views of n objects under different poses, we can build the m ? n ? s
data matrix X and use the same methodology as before. In Section 6 we describe experiments with
the walk cycle of pedestrians, and out-of-plane rotations of cars in street scenes. These transformations are cyclic, though highly non-linear, and we use the same renormalization as in Section 5.1.
5.3
Negative samples
One subtle aspect is how to obtain a structured data matrix from negative samples. This is simple for
virtual transformations, but not for natural transformations. For example, with planar rotation we
can easily generate rotated negative samples with arbitrary poses. However, the same operation with
walk cycles of pedestrians is not defined. How do we advance the walk cycle of a non-pedestrian? As
a pragmatic solution, we consider that negative samples are unaffected by natural transformations,
so a negative sample is constant for all s poses. Because the DFT of a constant signal is 0, except for
the DC value (the first frequency), we can ignore untransformed negative samples in all subproblems
for p 6= 1 (Eq. 12). This simple observation can result in significant computational savings.
6
Experiments
To demonstrate the generality of the proposed model, we conducted object detection and pose estimation experiments on 3 widely different settings, which will be described shortly. We implemented
a detector based on Histogram of Oriented Gradients (HOG) templates [4] with multiple components
[7]. This framework forms the basis on which several recent advances in object detection are built
[19, 10, 7]. The baseline algorithm independently trains s classifiers (components), one per pose, enabling pose-invariant object detection and pose prediction (Eq. 7). Components are then calibrated,
as usual for detectors with multiple components [7, 19]. The proposed method does not require any
ad-hoc calibration, since the components are jointly trained and related by the orthogonal matrix Q,
which preserves their L2 norm.
For the performance evaluation, ground truth objects are assigned to hypothesis by the widely used
Pascal criterion of bounding box overlap [7]. We then measure average precision (AP) and pose error (as epose /s, where epose is the discretized pose difference, taking wrap-around into account). We
tested two variants of each method, trained with both RR and SVR. Although parallelization is trivial, we report timings for single-core implementations, which more accurately reflect the total CPU
load. As noted in previous work [13], detectors trained with SVR have very similar performance to
those trained with Support Vector Machines.
6.1
Planar rotation in satellite images (Google Earth)
Our first test will be on a car detection task on satellite imagery [12], which has been used in several
works that deal with planar rotation [25, 18]. We annotated the orientations of 697 objects over half
the 30 images of the dataset. The first 7 annotated images were used for training, and the remaining
8 for validation. We created a structured data matrix X by augmenting each sample with 30 virtual
7
Google Earth
Time (s) AP
Pose
Fourier SVR
4.5
73.0
9.4
training RR
3.7
71.4
10.0
SVR 130.7
73.2
9.8
Standard
RR
399.3
72.7
10.3
TUD Campus/Crossing
KITTI
Time (s) AP
Pose Time (s) AP
0.1
81.5
9.3
15.0
53.5
0.08
82.2
8.9
15.5
53.4
40.5
80.2
9.5
454.2
56.5
45.8
81.6
9.4
229.6
54.5
Pose
14.9
15.0
13.8
14.0
Table 1: Results for pose detectors trained with Support Vector Regression (SVR) and Ridge Regression (RR).
We report training time, Average Precision (AP) and pose error (both in percentage).
samples, using 12? rotations. A visualization of trained weights is shown in Fig. 1-c and Appendix
B. Experimental results are presented in Table 1. Recall that our primary goal is to demonstrate
faster training, not to improve detection performance, which is reflected in the results. Nevertheless,
the two proposed fast Fourier algorithms are 29 to 107? faster than the baseline algorithms.
6.2
Walk cycle of pedestrians (TUD-Campus and TUD-Crossing)
We can consider a walking pedestrian to undergo a cyclic non-rigid deformation, with each period
corresponding to one step. Because this transformation is time-dependent, we can learn it from
video data. We used TUD-Campus for training and TUD-Crossing for testing (see Fig. 2). We
annotated a key pose in all 272 frames, so that the images of a pedestrian between two key poses
represent a whole walk cycle. Sampling 10 images per walk cycle (corresponding to 10 poses), we
obtained 10 sample groups for training, for a total of 100 samples.
From Table 1, the proposed algorithms seem to slightly outperform the baseline, showing that these
non-rigid deformations can be accurately accounted for. However, they are over 2 orders of magnitude faster. In addition to the speed benefits observed in Section 6.1, another factor at play is that
for natural transformations we can ignore the negative samples in s ? 1 of the subproblems (Section
5.3), whereas the baseline algorithms must consider them when training each of the s components.
6.3
Out-of-plane rotations of cars in street scenes (KITTI)
For our final experiment, we will attempt to demonstrate that the speed advantage of our method
still holds for difficult out-of-plane rotations. We chose the very recent KITTI benchmark [9], which
includes an object detection set of 7481 images of street scenes. The facing angle of cars (along the
vertical axis) is provided, which we bin into 15 discrete poses. We performed an 80-20% train-test
split of the images, considering cars of ?moderate? difficulty [9], and obtained 73 sample groups for
training with 15 poses each (for a total of 1095 samples).
Table 1 shows that the proposed method achieves competitive performance, but with a dramatically
lower computational cost. The results agree with the intuition that out-of-plane rotations strain the
assumptions of linearity and orthogonality, since they result in large deformations of the object.
Nevertheless, the ability to learn a useful model under such adverse conditions shows great promise.
7
Conclusions and future work
In this work, we derived new closed-form formulas to quickly train several pose classifiers at once,
and take advantage of the structure in datasets with pose annotation or virtual samples. Our implicit
transformation model seems to be surprisingly expressive, and in future work we would like to
experiment with other transformations, including non-cyclic. Other interesting directions include
larger-scale variants and the composition of multiple transformations.
Acknowledgements. The authors would like to thank Jo?ao Carreira for valuable discussions. They also acknowledge support by the FCT project PTDC/EEA-CRO/122812/2010, grants
SFRH/BD75459/2010, SFRH/BD74152/2010, and SFRH/BPD/90200/2012.
8
References
[1] V. N. Boddeti, T. Kanade, and B.V.K. Kumar. Correlation filters for object alignment. In
CVPR, 2013. 1, 2.1
[2] C.-Y. Chang, A. A. Maciejewski, and V. Balakrishnan. Fast eigenspace decomposition of
correlated images. IEEE Transactions on Image Processing, 9(11):1937?1949, 2000. 1.2
[3] O. Chapelle and B. Scholkopf. Incorporating invariances in non-linear support vector machines. In Advances in neural information processing systems, 2002. 1.2
[4] N. Dalal and B. Triggs. Histograms of oriented gradients for human detection. In CVPR, 2005.
1, 6
[5] P. J. Davis. Circulant matrices. American Mathematical Soc., 1994. 2.2
[6] A. Dosovitskiy, J. T. Springenberg, and T. Brox. Unsupervised feature learning by augmenting
single images. In International Conference on Learning Representations, 2014. 1
[7] P.F. Felzenszwalb, R.B. Girshick, D. McAllester, and D. Ramanan. Object detection with
discriminatively trained part-based models. TPAMI, 2010. 1, 6
[8] H. K. Galoogahi, T. Sim, and S. Lucey. Multi-channel correlation filters. In ICCV, 2013. 1,
2.1
[9] A. Geiger, P. Lenz, and R. Urtasun. Are we ready for autonomous driving? The KITTI Vision
Benchmark Suite. In CVPR, 2012. 6.3
[10] A. Geiger, C. Wojek, and R. Urtasun. Joint 3d estimation of objects and scene layout. In NIPS,
2011. 1, 3.2, 6
[11] R. M. Gray. Toeplitz and Circulant Matrices: A Review. Now Publishers, 2006. 3
[12] G. Heitz and D. Koller. Learning spatial context: Using stuff to find things. In ECCV, 2008.
6.1
[13] J. F. Henriques, J. Carreira, R. Caseiro, and J. Batista. Beyond hard negative mining: Efficient
detector learning via block-circulant decomposition. In ICCV, 2013. 1, 1.1, 2.1, 2.2, 4.1, 6
[14] J. F. Henriques, R. Caseiro, P. Martins, and J. Batista. Exploiting the circulant structure of
tracking-by-detection with kernels. In ECCV, 2012. 1, 1.1, 2.1, 5.1
[15] J. F. Henriques, R. Caseiro, P. Martins, and J. Batista. High-speed tracking with kernelized
correlation filters. TPAMI, 2015. 2.1
[16] N. Jojic, P. Simard, B. J. Frey, and D. Heckerman. Separating appearance from deformation.
In ICCV, 2001. 1.2
[17] A. Krizhevsky, I. Sutskever, and G. Hinton. Imagenet classification with deep convolutional
neural networks. In NIPS, 2012. 1
[18] K. Liu, H. Skibbe, T. Schmidt, T. Blein, K. Palme, T. Brox, and O. Ronneberger. Rotationinvariant HOG descriptors using fourier analysis in polar and spherical coordinates. International Journal of Computer Vision, 106(3):342?364, February 2014. 1.2, 6.1
[19] T. Malisiewicz, A. Gupta, and A. A. Efros. Ensemble of exemplar-svms for object detection
and beyond. In ICCV, 2011. 1, 6
[20] R. Memisevic and G. E. Hinton. Learning to represent spatial transformations with factored
higher-order boltzmann machines. Neural Computation, 22(6):1473?1492, 2010. 1.2
[21] X. Miao and R. Rao. Learning the lie groups of visual invariance. Neural computation,
19(10):2665?2693, 2007. 1.2
[22] J. L. Mundy. Object recognition in the geometric era: A retrospective. Lecture Notes in
Computer Science, pages 3?28, 2006. 1.2
[23] M. Paulin, J. Revaud, Z. Harchaoui, F. Perronnin, and C. Schmid. Transformation pursuit for
image classification. In CVPR, 2014. 1
[24] R. Rifkin, G. Yeo, and T. Poggio. Regularized least-squares classification. Nato Science Series
Sub Series III: Computer and Systems Sciences, 190:131?154, 2003. 3.1
[25] U. Schmidt and S. Roth. Learning rotation-aware features: From invariant priors to equivariant
descriptors. In CVPR, 2012. 1.2, 6.1
[26] P. Simard, Y. LeCun, J. Denker, and B. Victorri. Transformation invariance in pattern recognition ? tangent distance and tangent propagation. In LNCS. Springer, 1998. 1.2
[27] B. Tamaki, T.and Yuan, K. Harada, B. Raytchev, and K. Kaneda. Linear discriminative image
processing operator analysis. In CVPR, 2012. 1.2
[28] M. Uenohara and T. Kanade. Optimal approximation of uniformly rotated images. IEEE
Transactions on Image Processing, 7(1):116?119, 1998. 1.2
9
| 5417 |@word version:3 dalal:2 inversion:2 norm:4 seems:2 triggs:2 open:1 rgb:1 covariance:1 decomposition:2 q1:1 mention:2 reduction:1 cyclic:17 contains:5 series:3 liu:1 batista:5 renewed:1 diagonalized:1 yet:1 must:4 concatenate:1 blur:1 alone:1 half:2 plane:6 core:3 short:1 paulin:1 transposition:1 traverse:1 simpler:1 mathematical:1 along:5 direct:1 become:5 scholkopf:1 yuan:1 shorthand:1 hermitian:1 manner:1 introduce:1 indeed:1 equivariant:1 multi:1 discretized:1 inspired:1 spherical:1 little:1 xti:1 window:1 solver:3 cpu:1 becomes:1 project:2 estimating:1 linearity:5 xx:2 notation:2 provided:1 eigenspace:1 campus:4 what:2 kind:1 interpreted:1 minimizes:1 supplemental:1 finding:1 transformation:54 suite:1 stuff:1 classifier:17 ramanan:1 grant:1 positive:1 before:1 timing:1 frey:1 bilinear:1 era:1 untransformed:2 interpolation:1 approximately:1 ap:5 chose:1 resembles:1 limited:1 range:1 malisiewicz:1 unique:1 lecun:1 testing:1 practice:3 block:3 x3:1 lncs:1 ronneberger:1 svr:9 operator:7 twodimensional:1 context:3 influence:1 applying:2 www:1 roth:1 straightforward:1 attention:1 layout:1 independently:2 focused:1 splitting:1 pure:1 permutes:1 slicing:1 factored:2 estimator:1 q:3 array:5 importantly:1 insight:1 classic:1 variation:2 coordinate:2 diagonalizable:2 autonomous:1 pt:2 suppose:1 play:1 hypothesis:1 trick:2 element:10 crossing:3 recognition:5 satisfying:1 walking:3 expensive:1 observed:1 solved:1 capture:2 revaud:1 cycle:6 cro:1 valuable:2 intuition:1 moderately:1 trained:7 solving:1 division:1 efficiency:2 basis:8 translated:2 accelerate:5 joint:2 easily:2 derivation:1 train:11 fast:7 describe:1 widely:3 larger:1 cvpr:6 ability:2 toeplitz:1 gp:1 transform:4 itself:2 jointly:2 ip:1 final:1 hoc:1 advantage:2 rr:13 tpami:2 analytical:1 propose:2 product:4 rifkin:1 flexibility:1 achieve:1 description:1 validate:1 exploiting:2 sutskever:1 requirement:1 satellite:3 incremental:1 rotated:3 object:18 kitti:5 develop:1 pose:52 augmenting:3 exemplar:1 nearest:1 qt:1 sim:1 eq:20 soc:1 implemented:2 implies:1 come:1 direction:1 annotated:3 filter:3 human:1 mcallester:1 virtual:17 material:1 bin:1 require:5 ptdc:1 ao:2 generalization:1 suffices:1 decompose:1 inspecting:1 extension:2 hold:1 around:1 ground:1 great:2 substituting:1 driving:1 efros:1 achieves:1 earth:4 purpose:1 estimation:7 sfrh:3 lenz:1 polar:1 label:6 currently:1 create:1 hope:1 always:1 fulfill:1 shelf:3 varying:1 derived:1 focus:2 contrast:2 baseline:4 inference:1 dependent:1 rigid:6 perronnin:1 eliminate:1 w:1 kernelized:1 koller:1 expand:1 transformed:10 i1:1 pixel:7 arg:1 dual:6 orientation:1 pascal:1 denoted:1 classification:3 plan:1 spatial:2 special:1 yip:1 uc:2 brox:2 once:1 never:1 having:1 aware:1 sampling:1 saving:2 look:1 unsupervised:2 nearly:1 future:3 others:3 tud:6 report:2 dosovitskiy:1 few:1 modern:2 oriented:2 composed:1 preserve:3 simultaneously:4 densely:1 tightly:1 numpy:1 argmax:1 attempt:1 freedom:1 detection:12 interest:1 possibility:1 highly:1 mining:1 evaluation:1 alignment:1 yielding:1 primal:1 poggio:1 orthogonal:8 unless:1 walk:6 renormalize:1 deformation:7 girshick:1 column:3 earlier:1 rao:1 retains:1 cost:4 applicability:2 stacking:1 introducing:1 krizhevsky:1 successful:1 harada:1 conducted:2 stored:1 calibrated:1 caseiro:4 peak:2 international:2 memisevic:1 off:3 regressor:1 quickly:4 jo:2 augmentation:1 squared:1 imagery:2 reflect:1 containing:1 american:1 simard:2 yeo:1 account:1 includes:1 pedestrian:8 matter:1 explicitly:5 ad:1 performed:1 try:1 view:1 closed:4 analyze:1 kwk:1 competitive:1 complicated:2 parallel:1 annotation:5 contribution:2 square:1 formed:1 accuracy:1 convolutional:1 descriptor:4 characteristic:1 likewise:1 ensemble:2 yield:2 conceptually:1 generalize:1 considering:1 accurately:3 multiplying:1 unaffected:1 detector:10 inexpensive:1 frequency:2 involved:1 naturally:1 proof:1 associated:2 gain:1 sampled:1 dataset:6 begun:1 ask:1 recall:2 knowledge:1 car:7 dimensionality:1 color:1 subtle:1 actually:1 miao:2 higher:1 supervised:1 planar:8 methodology:1 reflected:1 formulation:2 done:3 though:2 box:1 generality:3 just:1 implicit:3 correlation:5 until:1 working:1 horizontal:3 replacing:1 expressive:1 propagation:1 google:4 lda:1 gray:1 verify:1 true:1 regularization:1 assigned:1 jojic:1 q0:4 symmetric:1 deal:4 davis:1 noted:1 criterion:1 generalized:1 ridge:5 demonstrate:4 necessitating:1 motion:1 cp:1 image:53 wise:4 consideration:1 regularizing:1 recently:1 common:1 rotation:26 sped:1 qp:5 jp:1 extend:1 interpretation:1 significant:1 composition:1 dft:18 similarly:2 chapelle:1 calibration:1 impressive:1 recent:5 moderate:1 driven:2 store:2 jorge:1 yi:1 preserving:1 seen:1 somewhat:1 wojek:1 redundant:1 period:4 signal:1 sliding:1 multiple:8 full:2 harchaoui:1 reduces:1 boddeti:1 faster:4 long:1 wpt:1 divided:1 kaneda:1 manipulate:1 prediction:1 variant:2 regression:10 vision:4 essentially:1 histogram:2 represent:4 kernel:2 robotics:1 achieved:1 addition:1 whereas:1 victorri:1 publisher:1 appropriately:1 extra:1 parallelization:1 posse:1 induced:2 virtually:1 undergo:1 thing:1 balakrishnan:1 leveraging:1 mod:1 seem:3 call:1 unitary:2 leverage:1 split:1 enough:1 iii:1 fft:1 variety:3 xj:1 restrict:1 suboptimal:1 bpd:1 reduce:1 inner:1 knowing:1 shift:3 whether:2 motivated:1 pca:1 retrospective:1 repeatedly:1 matlab:3 deep:2 dramatically:2 useful:3 amount:3 svms:1 reduced:1 generate:2 specifies:2 outperform:1 percentage:1 estimated:1 per:7 diverse:1 discrete:4 promise:1 affected:1 group:8 redundancy:2 key:2 nevertheless:2 libsvm:1 tamaki:2 vast:1 geometrically:2 angle:2 inverse:1 springenberg:1 parsimonious:1 geiger:2 appendix:6 comparable:1 entirely:1 conjugation:2 orthogonality:2 constrain:1 scene:5 x2:2 interpolated:1 fourier:15 aspect:2 argument:1 extremely:1 speed:3 kumar:1 martin:3 relatively:2 fct:1 structured:6 combination:1 remain:1 smaller:3 slightly:1 heckerman:1 appealing:1 modification:3 invariant:4 iccv:4 taken:1 visualization:1 agree:1 turn:1 describing:1 needed:1 mind:1 studying:1 generalizes:1 operation:3 available:1 pursuit:1 apply:1 denker:1 schmidt:2 shortly:1 hat:1 original:4 denotes:4 assumes:1 remaining:1 include:1 exploit:3 build:1 february:1 question:2 costly:2 primary:1 usual:1 diagonal:4 gradient:2 wrap:1 distance:1 thank:1 separating:1 capacity:1 street:3 w0:1 trivial:1 urtasun:2 assuming:1 relationship:1 cq:4 illustration:1 difficult:1 potentially:1 hog:3 relate:1 subproblems:3 expense:1 negative:9 implementation:4 boltzmann:1 vertical:1 observation:1 mundy:1 datasets:12 benchmark:2 finite:1 enabling:1 acknowledge:1 hinton:2 strain:1 dc:1 frame:1 arbitrary:2 expressiveness:1 eea:1 specified:1 imagenet:1 learned:2 nip:2 beyond:2 usually:1 pattern:2 built:1 including:2 max:2 video:1 shifting:1 power:1 suitable:1 overlap:1 natural:10 difficulty:1 regularized:2 diagonalizing:1 improve:1 axis:1 created:2 ready:1 naive:1 auto:2 schmid:1 speeding:1 review:1 geometric:3 literature:1 tangent:3 vectorizing:1 multiplication:1 l2:2 acknowledgement:1 prior:1 reordering:2 fully:2 permutation:1 loss:1 discriminatively:1 interesting:2 lecture:1 facing:1 validation:1 eigendecomposition:2 degree:3 affine:2 consistent:1 principle:1 viewpoint:2 share:1 translation:10 row:7 eccv:2 course:1 accounted:1 surprisingly:1 transpose:1 henriques:6 drastically:1 allow:2 lucey:1 institute:1 wide:1 template:3 circulant:12 neighbor:1 taking:1 felzenszwalb:1 isr:2 benefit:2 slice:4 dimension:4 heitz:1 world:1 avoids:1 rich:1 gram:3 author:1 avoided:1 fpose:1 cope:1 qx:1 transaction:2 approximate:3 ignore:2 nato:1 keep:1 uncentered:1 reveals:1 reorder:1 xi:4 discriminative:1 search:1 table:4 kanade:2 learn:4 channel:5 robust:1 symmetry:1 complex:5 domain:5 diag:4 main:1 linearly:1 bounding:1 whole:1 verifies:1 repeated:3 body:1 augmented:1 fig:4 representative:1 x1:2 renormalization:1 n:4 precision:2 sub:1 wish:1 explicit:1 lie:2 third:2 learns:1 cyclically:2 formula:3 theorem:3 remained:1 load:1 xt:1 repeatable:1 showing:1 x:3 gupta:1 intrinsic:1 naively:1 burden:1 incorporating:1 diagonalization:4 magnitude:4 rui:1 easier:1 surprise:1 depicted:1 simply:3 appearance:3 visual:2 desire:1 expressed:1 tracking:2 scalar:1 chang:1 applies:1 springer:1 pedro:1 truth:1 satisfies:1 relies:1 sized:1 identity:1 presentation:1 goal:2 viewed:1 replace:1 hard:3 change:1 adverse:1 carreira:2 except:1 uniformly:1 wt:1 total:3 invariance:5 experimental:1 pragmatic:1 coimbra:1 support:8 evaluate:1 tested:1 correlated:1 |
4,880 | 5,418 | LSDA: Large Scale Detection through Adaptation
Judy Hoffman , Sergio Guadarrama , Eric Tzeng , Ronghang Hu? , Jeff Donahue ,
EECS, UC Berkeley, ? EE, Tsinghua University
{jhoffman, sguada, tzeng, jdonahue}@eecs.berkeley.edu
[email protected]
Ross Girshick , Trevor Darrell , Kate Saenko4
EECS, UC Berkeley, 4 CS, UMass Lowell
{rbg, trevor}@eecs.berkeley.edu, [email protected]
Abstract
A major challenge in scaling object detection is the difficulty of obtaining labeled
images for large numbers of categories. Recently, deep convolutional neural networks (CNNs) have emerged as clear winners on object classification benchmarks,
in part due to training with 1.2M+ labeled classification images. Unfortunately,
only a small fraction of those labels are available for the detection task. It is much
cheaper and easier to collect large quantities of image-level labels from search engines than it is to collect detection data and label it with precise bounding boxes.
In this paper, we propose Large Scale Detection through Adaptation (LSDA), an
algorithm which learns the difference between the two tasks and transfers this
knowledge to classifiers for categories without bounding box annotated data, turning them into detectors. Our method has the potential to enable detection for the
tens of thousands of categories that lack bounding box annotations, yet have plenty
of classification data. Evaluation on the ImageNet LSVRC-2013 detection challenge demonstrates the efficacy of our approach. This algorithm enables us to
produce a >7.6K detector by using available classification data from leaf nodes in
the ImageNet tree. We additionally demonstrate how to modify our architecture
to produce a fast detector (running at 2fps for the 7.6K detector). Models and
software are available at lsda.berkeleyvision.org.
1
Introduction
Both classification and detection are key visual recognition challenges, though historically very
different architectures have been deployed for each. Recently, the R-CNN model [1] showed how
to adapt an ImageNet classifier into a detector, but required bounding box data for all categories.
We ask, is there something generic in the transformation from classification to detection that can be
learned on a subset of categories and then transferred to other classifiers?
One of the fundamental challenges in training object detection systems is the need to collect a
large of amount of images with bounding box annotations. The introduction of detection challenge
datasets, such as PASCAL VOC [2], have propelled progress by providing the research community
a dataset with enough fully annotated images to train competitive models although only for 20
classes. Even though the more recent ImageNet detection challenge dataset [3] has extended the set
of annotated images, it only contains data for 200 categories. As we look forward towards the goal
of scaling our systems to human-level category detection, it becomes impractical to collect a large
quantity of bounding box labels for tens or hundreds of thousands of categories.
?
This work was supported in part by DARPA?s MSEE and SMISC programs, by NSF awards IIS-1427425,
and IIS-1212798, IIS-1116411, and by support from Toyota.
1
W
DET
Wdog
?
DET
apple
W
W
WCLASSIFY
cat
DET
Wcat
CLASSIFY
apple
ICLASSIFY
dog
I
DET
apple
dog
apple
CLASSIFY
dog
cat
ICLASSIFY
Detectors
Classifiers
IDET
Figure 1: The core idea is that we can learn detectors (weights) from labeled classification data (left),
for a wide range of classes. For some of these classes (top) we also have detection labels (right), and
can learn detectors. But what can we do about the classes with classification data but no detection
data (bottom)? Can we learn something from the paired relationships for the classes for which we
have both classifiers and detectors, and transfer that to the classifier at the bottom to make it into a
detector?
In contrast, image-level annotation is comparatively easy to acquire. The prevalence of image tags
allows search engines to quickly produce a set of images that have some correspondence to any
particular category. ImageNet [3], for example, has made use of these search results in combination
with manual outlier detection to produce a large classification dataset comprised of over 20,000
categories. While this data can be effectively used to train object classifier models, it lacks the
supervised annotations needed to train state-of-the-art detectors.
In this work, we propose Large Scale Detection through Adaptation (LSDA), an algorithm that
learns to transform an image classifier into an object detector. To accomplish this goal, we use
supervised convolutional neural networks (CNNs), which have recently been shown to perform well
both for image classification [4] and object detection [1, 5]. We cast the task as a domain adaptation
problem, considering the data used to train classifiers (images with category labels) as our source
domain, and the data used to train detectors (images with bounding boxes and category labels) as our
target domain. We then seek to find a general transformation from the source domain to the target
domain, that can be applied to any image classifier to adapt it into a object detector (see Figure 1).
Girshick et al. (R-CNN) [1] demonstrated that adaptation, in the form of fine-tuning, is very important for transferring deep features from classification to detection and partially inspired our approach.
However, the R-CNN algorithm uses classification data only to pre-train a deep network and then
requires a large number of bounding boxes to train each detection category.
Our LSDA algorithm uses image classification data to train strong classifiers and requires detection
bounding box labeled data for only a small subset of the final detection categories and much less
time. It uses the classes labeled with both classification and detection labels to learn a transformation
of the classification network into a detection network. It then applies this transformation to adapt
classifiers for categories without any bounding box annotated data into detectors.
Our experiments on the ImageNet detection task show significant improvement (+50% relative
mAP) over a baseline of just using raw classifier weights on object proposal regions. One can
adapt any ImageNet-trained classifier into a detector using our approach, whether or not there are
corresponding detection labels for that class.
2
Related Work
Recently, Multiple Instance Learning (MIL) has been used for training detectors using weak labels,
i.e. images with category labels but not bounding box labels. The MIL paradigm estimates latent
labels of examples in positive training bags, where each positive bag is known to contain at least one
positive example. Ali et al. [6] constructs positive bags from all object proposal regions in a weakly
labeled image that is known to contain the object, and uses a version of MIL to learn an object
detector. A similar method [7] learns detectors from PASCAL VOC images without bounding box
2
"
fcA"
det"
fc6"
det"
fc7"
?B"
det"
layers"175"
Input"image"
Region"
Proposals"
Warped""
region"
cat:"0.90"
cat?"yes"
dog:"0.45"
dog?"no"
adapt"
fcB"
LSDA"Net"
background"
background:"0.25"
Produce""
Predic=ons"
Figure 2: Detection with the LSDA network. Given an image, extract region proposals, reshape the
regions to fit into the network size and finally produce detection scores per category for the region.
Layers with red dots/fill indicate they have been modified/learned during fine-tuning with available
bounding box annotated data.
labels. MIL-based methods are a promising approach that is complimentary to ours. They have not
yet been evaluated on the large-scale ImageNet detection challenge to allow for direct comparison.
Deep convolutional neural networks (CNNs) have emerged as state of the art on popular object
classification benchmarks (ILSVRC, MNIST) [4]. In fact, ?deep features? extracted from CNNs
trained on the object classification task are also state of the art on other tasks, e.g., subcategory
classification, scene classification, domain adaptation [8] and even image matching [9]. Unlike the
previously dominant features (SIFT [10], HOG [11]), deep CNN features can be learned for each
specific task, but only if sufficient labeled training data are available. R-CNN [1] showed that finetuning deep features on a large amount of bounding box labeled data significantly improves detection
performance.
Domain adaptation methods aim to reduce dataset bias caused by a difference in the statistical distributions between training and test domains. In this paper, we treat the transformation of classifiers
into detectors as a domain adaptation task. Many approaches have been proposed for classifier
adaptation; e.g., feature space transformations [12], model adaptation approaches [13, 14] and joint
feature and model adaptation [15, 16]. However, even the joint learning models are not able to modify the feature extraction process and so are limited to shallow adaptation techniques. Additionally,
these methods only adapt between visual domains, keeping the task fixed, while we adapt both from
a large visual domain to a smaller visual domain and from a classification task to a detection task.
Several supervised domain adaptation models have been proposed for object detection. Given a
detector trained on a source domain, they adjust its parameters on labeled target domain data. These
include variants for linear support vector machines [17, 18, 19], as well as adaptive latent SVMs [20]
and adaptive exemplar SVM [21]. A related recent method [22] proposes a fast adaptation technique
based on Linear Discriminant Analysis. These methods require labeled detection data for all object
categories, both in the source and target domains, which is absent in our scenario. To our knowledge,
ours is the first method to adapt to held-out categories that have no detection data.
3
Large Scale Detection through Adaptation (LSDA)
We propose Large Scale Detection through Adaptation (LSDA), an algorithm for adapting classifiers
to detectors. With our algorithm, we are able to produce a detection network for all categories of
interest, whether or not bounding boxes are available at training time (see Figure 2).
Suppose we have K categories we want to detect, but we only have bounding box annotations for m
categories. We will refer to the set of categories with bounding box annotations as B = {1, ...m},
and the set of categories without bounding box annotations as set A = {m, ..., K}. In practice,
we will likely have m K, as is the case in the ImageNet dataset. We assume availability of
classification data (image-level labels) for all K categories and will use that data to initialize our
network.
3
LSDA transforms image classifiers into object detectors using three key insights:
1. Recognizing background is an important step in adapting a classifier into a detector
2. Category invariant information can be transferred between the classifier and detector feature representations
3. There may be category specific differences between a classifier and a detector
We will next demonstrate how our method accomplishes each of these insights as we describe the
training of LSDA.
3.1
Training LSDA: Category Invariant Adaptation
For our convolutional neural network, we adopt the architecture of Krizhevsky et al. [4], which
achieved state-of-the-art performance on the ImageNet ILSVRC2012 classification challenge. Since
this network requires a large amount of data and time to train its approximately 60 million parameters, we start by pre-training the CNN trained on the ILSVRC2012 classification dataset, which
contains 1.2 million classification-labeled images of 1000 categories. Pre-training on this dataset
has been shown to be a very effective technique [8, 5, 1], both in terms of performance and in terms
of limiting the amount of in-domain labeled data needed to successfully tune the network. Next, we
replace the last weight layer (1000 linear classifiers) with K linear classifiers, one for each category
in our task. This weight layer is randomly initialized and then we fine-tune the whole network on
our classification data. At this point, we have a network that can take an image or a region proposal
as input, and produce a set of scores for each of the K categories. We find that even using the net
trained on classification data in this way produces a strong baseline (see Section 4).
We next transform our classification network into a detection network. We do this by fine-tuning
layers 1-7 using the available labeled detection data for categories in set B. Following the Regionsbased CNN (R-CNN) [1] algorithm, we collect positive bounding boxes for each category in set B
as well as a set of background boxes using a region proposal algorithm, such as selective search [23].
We use each labeled region as a fine-tuning input to the CNN after padding and warping it to the
CNN?s input size. Note that the R-CNN fine-tuning algorithm requires bounding box annotated data
for all categories and so can not directly be applied to train all K detectors. Fine-tuning transforms
all network weights (except for the linear classifiers for set A) and produces a softmax detector for
categories in set B, which includes a weight vector for the new background class.
Layers 1-7 are shared between all categories in set B and we find empirically that fine-tuning induces
a generic, category invariant transformation of the classification network into a detection network.
That is, even though fine-tuning sees no detection data for categories in set A, the network transforms in a way that automatically makes the original set A image classifiers much more effective
at detection (see Figure 3). Fine-tuning for detection also learns a background weight vector that
encodes a generic ?background? category. This background model is important for modeling the
task shift from image classification, which does not include background distractors, to detection,
which is dominated by background patches.
3.2
Training LSDA: Category Specific Adaptation
Finally, we learn a category specific transformation that will change the classifier model parameters
into the detector model parameters that operate on the detection feature representation. The category
specific output layer (f c8) is comprised of f cA, f cB, ?B, and f c ? BG. For categories in set B,
this transformation can be learned through directly fine-tuning the category specific parameters f cB
(Figure 2). This is equivalent to fixing f cB and learning a new layer, zero initialized, ?B, with
equivalent loss to f cB , and adding together the outputs of ?B and f cB .
Let us define the weights of the output layer of the original classification network as W c , and the
weights of the output layer of the adapted detection network as W d . We know that for a category
i ? B, the final detection weights should be computed as Wid = Wic + ?Bi . However, since
there is no detection data for categories in A, we can not directly learn a corresponding ?A layer
during fine-tuning. Instead, we can approximate the fine-tuning that would have occurred to f cA
had detection data been available. We do this by finding the nearest neighbors categories in set B
for each category in set A and applying the average change. Here we define nearest neighbors as
4
those categories with the nearest (minimal Euclidean distance) `2 -normalized f c8 parameters in the
classification network. This corresponds to the classification model being most similar and hence,
we assume, the detection model should be most similar. We denote the k th nearest neighbor in set
B of category j ? A as NB (j, k), then we compute the final output detection weights for categories
in set A as:
?j ? A : Wjd
=
Wjc +
k
1X
?BNB (j,i)
k i=1
(1)
Thus, we adapt the category specific parameters even without bounding boxes for categories in set
A. In the next section we experiment with various values of k, including taking the full average:
k = |B|.
3.3
Detection with LSDA
At test time we use our network to extract K + 1 scores per region proposal in an image (similar to
the R-CNN [1] pipeline). One for each category and an additional score for the background category.
Finally, for a given region, the score for category i is computed by combining the per category score
with the background score: scorei ? scorebackground .
In contrast to the R-CNN [1] model which trains SVMs on the extracted features from layer 7
and bounding box regression on the extracted features from layer 5, we directly use the final score
vector to produce the prediction scores without either of the retraining steps. This choice results in a
small performance loss, but offers the flexibility of being able to directly combine the classification
portion of the network that has no detection labeled data, and reduces the training time from 3 days
to roughly 5.5 hours.
4
Experiments
To demonstrate the effectiveness of our approach we present quantitative results on the ILSVRC2013
detection dataset. The dataset offers a 200-category detection challenge. The training set has ?400K
annotated images and on average 1.534 object classes per image. The validation set has 20K annotated images with ?50K annotated objects. We simulate having access to classification labels for
all 200 categories and having detection annotations for only the first 100 categories (alphabetically
sorted).
4.1
Experiment Setup & Implementation Details
We start by separating our data into classification and detection sets for training and a validation
set for testing. Since the ILSVRC2013 training set has on average fewer objects per image than
the validation set, we use this data as our classification data. To balance the categories we use
?1000 images per class (200,000 total images). Note: for classification data we only have access
to a single image-level annotation that gives a category label. In effect, since the training set may
contain multiple objects, this single full-image label is a weak annotation, even compared to other
classification training data sets. Next, we split the ILSVRC2013 validation set in half as [1] did,
producing two sets: val1 and val2. To construct our detection training set, we take the images
with bounding box labels from val1 for only the first 100 categories (? 5000 images). Since the
validation set is relatively small, we augment our detection set with 1000 bounding box annotated
images per category from the ILSVRC2013 training set (following the protocol of [1]). Finally we
use the second half of the ILSVRC2013 validation set (val2) for our evaluation.
We implemented our CNN architectures and execute all fine-tuning using the open source software
package Caffe [24] and have made our model definitions weights publicly available.
4.2
Quantitative Analysis on Held-out Categories
We evaluate the importance of each component of our algorithm through an ablation study. As
a baseline we consider training the network with only the classification data (no adaptation) and
applying the network to the region proposals. The summary of the importance of our three adaptation
components is shown in Figure 3. Our full LSDA model achieves a 50% relative mAP boost over
5
Detection
Adaptation Layers
Output Layer
Adaptation
mAP Trained
100 Categories
mAP Held-out
100 Categories
mAP All
200 Categories
No Adapt (Classification Network)
fcbgrnd
fcbgrnd ,fc6
fcbgrnd ,fc7
fcbgrnd ,fcB
fcbgrnd ,fc6 ,fc7
fcbgrnd ,fc6 ,fc7 ,fcB
fcbgrnd ,layers1-7,fcB
-
12.63
14.93
24.72
23.41
18.04
25.78
26.33
27.81
10.31
12.22
13.72
14.57
11.74
14.20
14.42
15.85
11.90
13.60
19.20
19.00
14.90
20.00
20.40
21.83
fcbgrnd ,layers1-7,fcB
fcbgrnd ,layers1-7,fcB
fcbgrnd ,layers1-7,fcB
28.12
27.95
27.91
15.97
16.15
15.96
22.05
22.05
21.94
29.72
26.25
28.00
Avg NN (k=5)
Avg NN (k=10)
Avg NN (k=100)
Oracle: Full Detection Network
Table 1: Ablation study for the components of LSDA. We consider removing different pieces of our
algorithm to determine which pieces are essential. We consider training with the first 100 (alphabetically) categories of the ILSVRC2013 detection validation set (on val1) and report mean average
precision (mAP) over the 100 trained on and 100 held out categories (on val2). We find the best
improvement is from fine-tuning all layers and using category specific adaptation.
the classification only network. The most important step of our algorithm proved to be adapting
the feature representation, while the least important was adapting the category specific parameter.
This fits with our intuition that the main benefit of our approach is to transfer category invariant
information from categories with known bounding box annotation to those without the bounding
box annotations.
In Table 1, we present a more detailed analysis of the
different adaptation techniques we could use to train the
network. We find that the best category invariant adaptation approach is to learn the background category layer
and adapt all convolutional and fully connected layers,
bringing mAP on the held-out categories from 10.31% up
to 15.85%. Additionally, using output layer adaptation
(k = 10) further improves performance, bringing mAP
to 16.15% on the held-out categories (statistically significant at p = 0.017 using a paired sample t-test [25]). The
last row shows the performance achievable by our detection network if it had access to detection data for all 200
categories, and serves as a performance upper bound.1
LSDA
16.15
LSDA (bg+ft)
15.85
LSDA (bg only)
12.2
Classification Net
10.31
0
5
10
15
20
Figure 3: Comparison (mAP%) of our
We find that one of the biggest reasons our algorithm im- full system (LSDA) on categories with
proves is from reducing localization error. For example, no bounding boxes at training time.
in Figure 4, we show that while the classification only
trained net tends to focus on the most discriminative part
of an object (ex: face of an animal) after our adaptation, we learn to localize the whole object (ex:
entire body of the animal).
4.3
Error Analysis on Held Out Categories
We next present an analysis of the types of errors that our system (LSDA) makes on the held out
object categories. First, in Figure 5, we consider three types of false positive errors: Loc (localization errors), BG (confusion with background), and Oth (other error types, which is essentially
1
To achieve R-CNN performance requires additionally learning SVMs on the activations of layer 7 and
bounding box regression on the activations of layer 5. Each of these steps adds between 1-2mAP at high
computation cost and using the SVMs removes the adaptation capacity of the system.
6
Figure 4: We show example detections on held out categories, for which we have no detection
training data, where our adapted network (LSDA) (shown with green box) correctly localizes and
labels the object of interest, while the classification network baseline (shown in red) incorrectly
localizes the object. This demonstrates that our algorithm learns to adapt the classifier into a detector
which is sensitive to localization and background rejection.
correctly localizing an object, but misclassifying it). After separating all false positives into one of
these three error types we visually show the percentage of errors found in each type as you look at
the top scoring 25-3200 false positives.2 We consider the baseline of starting with the classification
only network and show the false positive breakdown in Figure 5(b). Note that the majority of false
positive errors are confusion with background and localization errors. In contrast, after adapting
the network using LSDA we find that the errors found in the top false positives are far less due to
localization and background confusion (see Figure 5(c)). Arguably one of the biggest differences between classification and detection is the ability to accurately localize objects and reject background.
Therefore, we show that our method successfully adapts the classification parameters to be more
suitable for detection.
In Figure 5(a) we show examples of the top scoring Oth error types for LSDA on the held-out
categories. This means the detector localizes an incorrect object type. For example, the motorcycle
detector localized and mislabeled bicycle and the lemon detector localized and mislabeled an orange.
In general, we noticed that many of the top false positives from the Oth error type were confusion
with very similar categories.
4.4
Large Scale Detection
To showcase the capabilities of our technique we produced a 7604 category detector. The first
categories correspond to the 200 categories from the ILSVRC2013 challenge dataset which have
bounding box labeled data available. The other 7404 categories correspond to leaf nodes in the
ImageNet database and are trained using the available full image labeled classification data. We
trained a full detection network using the 200 fully annotated categories and trained the other 7404
last layer nodes using only the classification data. Since we lack bounding box annotated data for
the majority of the categories we show example top detections in Figure 6. The results are filtered
using non-max suppression across categories to only show the highest scoring categories.
The main contribution of our algorithm is the adaptation technique for modifying a convolutional
neural network for detection. However, the choice of network and how the net is used at test time
both effect the detection time computation. We have therefore also implemented and released a
version of our algorithm running with fast region proposals [27] on a spatial pyramid pooling network [28], reducing our detection time down to half a second per image (from 4s per image) with
nearly the same performance. We hope that this will allow the use of our 7.6K model on large data
sources such as videos. We have released the 7.6K model and code to run detection (both the way
presented in this paper and our faster version) at lsda.berkeleyvision.org.
2
We modified the analysis software made available by Hoeim et al. [26] to work on ILSVRC-2013 detection
7
mushroom
microphone
motorcycle
microphone (sim): ov=0.00 1?r=?3.00
lemon
laptop
nail
miniskirt
1?r=?6.00
mushroom
(sim): ov=0.00 1?r=?8.00
miniskirt (sim): ov=0.00 motorcycle
1?r=?1.00 (sim): ov=0.00
nail (sim): ov=0.00 1?r=?4.00 laptop (sim): ov=0.00 1?r=?3.00 lemon (sim): ov=0.00 1?r=?5.00
(a) Example Top Scoring False Positives: LSDA correctly localizes but incorrectly labels object
Held?out Categories
Held?out Categories
100
Loc
Oth
BG
80
percentage of each type
percentage of each type
100
60
40
20
0
25
50
60
40
20
0
25
100 200 400 800 1600 3200
total false positives
(b) Classification Network
Loc
Oth
BG
80
50
100 200 400 800 1600 3200
total false positives
(c) LSDA Network
Figure 5: We examine the top scoring false positives from LSDA. Many of our top scoring false
positives come from confusion with other categories (a). (b-c) Comparison of error type breakdown
on the categories which have no training bounding boxes available (held-out categories). After
adapting the network using our algorithm (LSDA), the percentage of false positive errors due to
localization and background confusion is reduced (c) as compared to directly using the classification
network in a detection framework (b).
whippet: 2.0
dog: 4.1
taillight: 0.9
American bison: 7.0
wheel
and axle: 1.0
car: 6.0
sofa: 8.0
Figure 6: Example top detections from our 7604 category detector. Detections from the 200 categories that have bounding box training data available are shown in blue. Detections from the
remaining 7404 categories for which only classification training data is available are shown in red.
5
Conclusion
We have presented an algorithm that is capable of transforming a classifier into a detector. We
use CNN models to train both a classification and a detection network. Our multi-stage algorithm
uses corresponding classification and detection data to learn the change from a classification CNN
network to a detection CNN network, and applies that difference to future classifiers for which there
is no available detection data.
We show quantitatively that without seeing any bounding box annotated data, we can increase performance of a classification network by 50% relative improvement using our adaptation algorithm.
Given the significant improvement on the held out categories, our algorithm has the potential to
enable detection of tens of thousands of categories. All that would be needed is to train a classification layer for the new categories and use our fine-tuned detection model along with our output layer
adaptation techniques to update the classification parameters directly.
Our approach significantly reduces the overhead of producing a high quality detector. We hope that
in doing so we will be able to minimize the gap between having strong large-scale classifiers and
strong large-scale detectors. There is still a large gap to reach oracle (known bounding box labels)
performance. For future work we would like to explore multiple instance learning techniques to
discover and mine patches for the categories that lack bounding box data.
8
References
[1] R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich feature hierarchies for accurate object detection
and semantic segmentation. In In Proc. CVPR, 2014.
[2] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The pascal visual object
classes (voc) challenge. International Journal of Computer Vision, 88(2):303?338, June 2010.
[3] A. Berg, J. Deng, and L. Fei-Fei. ImageNet Large Scale Visual Recognition Challenge. 2012.
[4] A. Krizhevsky, I. Sutskever, and G. E. Hinton. ImageNet classification with deep convolutional neural
networks. In Proc. NIPS, 2012.
[5] P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. LeCun. Overfeat: Integrated recognition,
localization and detection using convolutional networks. CoRR, abs/1312.6229, 2013.
[6] K. Ali and K. Saenko. Confidence-rated multiple instance boosting for object detection. In IEEE Conference on Computer Vision and Pattern Recognition, 2014.
[7] H. Song, R. Girshick, S. Jegelka, J. Mairal, Z. Harchaoui, and T. Darrell. On learning to localize objects
with minimal supervision. In Proceedings of the International Conference on Machine Learning (ICML),
2014.
[8] J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng, and T. Darrell. DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition. In Proc. ICML, 2014.
[9] Philipp Fischer, Alexey Dosovitskiy, and Thomas Brox. Descriptor matching with convolutional neural
networks: a comparison to sift. ArXiv e-prints, abs/1405.5769, 2014.
[10] D. G. Lowe. Distinctive image features from scale-invariant key points. IJCV, 2004.
[11] N. Dalal and B. Triggs. Histograms of oriented gradients for human detection. In In Proc. CVPR, 2005.
[12] B. Kulis, K. Saenko, and T. Darrell. What you saw is not what you get: Domain adaptation using
asymmetric kernel transforms. In Proc. CVPR, 2011.
[13] J. Yang, R. Yan, and A. Hauptmann. Adapting SVM classifiers to data with shifted distributions. In ICDM
Workshops, 2007.
[14] Y. Aytar and A. Zisserman. Tabula rasa: Model transfer for object category detection. In Proc. ICCV,
2011.
[15] J. Hoffman, E. Rodner, J. Donahue, K. Saenko, and T. Darrell. Efficient learning of domain-invariant
image representations. In Proc. ICLR, 2013.
[16] L. Duan, D. Xu, and Ivor W. Tsang. Learning with augmented features for heterogeneous domain adaptation. In Proc. ICML, 2012.
[17] J. Yang, R. Yan, and A. G. Hauptmann. Cross-domain video concept detection using adaptive svms. ACM
Multimedia, 2007.
[18] Y. Aytar and A. Zisserman. Tabula rasa: Model transfer for object category detection. In IEEE International Conference on Computer Vision, 2011.
[19] J. Donahue, J. Hoffman, E. Rodner, K. Saenko, and T. Darrell. Semi-supervised domain adaptation with
instance constraints. In Computer Vision and Pattern Recognition (CVPR), 2013.
[20] J. Xu, S. Ramos, D. V?azquez, and A.M. L?opez. Domain adaptation of deformable part-based models.
IEEE Trans. on Pattern Analysis and Machine Intelligence, In Press, 2014.
[21] Y. Aytar and A. Zisserman. Enhancing exemplar svms using part level transfer regularization. In British
Machine Vision Conference, 2012.
[22] D. Goehring, J. Hoffman, E. Rodner, K. Saenko, and T. Darrell. Interactive adaptation of real-time object
detectors. In International Conference on Robotics and Automation (ICRA), 2014.
[23] J.R.R. Uijlings, K.E.A. van de Sande, T. Gevers, and A.W.M. Smeulders. Selective search for object
recognition. International Journal of Computer Vision, 104(2):154?171, 2013.
[24] Yangqing Jia, Evan Shelhamer, Jeff Donahue, Sergey Karayev, Jonathan Long, Ross Girshick, Sergio
Guadarrama, and Trevor Darrell. Caffe: Convolutional architecture for fast feature embedding. arXiv
preprint arXiv:1408.5093, 2014.
[25] M. D. Smucker, J. Allan, and B. Carterette. A comparison of statistical significance tests for information
retrieval evaluation. In In Conference on Information and Knowledge Management, 2007.
[26] D. Hoeim, Y. Chodpathumwan, and Q. Dai. Diagnosing error in object detectors. In In Proc. ECCV,
2012.
[27] P. Kr?ahenb?uhl and V. Koltun. Geodesic object proposals. In In Proc. ECCV, 2014.
[28] K. He, X. Zhang, S. Ren, and J. Sun. Spatial pyramid pooling in deep convolutional networks for visual
recognition. In In Proc. ECCV, 2014.
9
| 5418 |@word kulis:1 cnn:18 version:3 dalal:1 achievable:1 retraining:1 everingham:1 triggs:1 open:1 hu:1 seek:1 contains:2 uma:1 efficacy:1 score:9 loc:3 tuned:1 ours:2 guadarrama:2 activation:3 yet:2 mushroom:2 enables:1 remove:1 update:1 half:3 leaf:2 fewer:1 intelligence:1 core:1 filtered:1 boosting:1 node:3 philipp:1 org:2 zhang:3 diagnosing:1 along:1 direct:1 koltun:1 fps:1 incorrect:1 ijcv:1 combine:1 overhead:1 allan:1 roughly:1 examine:1 multi:1 inspired:1 voc:3 automatically:1 duan:1 considering:1 becomes:1 discover:1 laptop:2 what:3 msee:1 complimentary:1 finding:1 transformation:9 impractical:1 berkeley:4 quantitative:2 interactive:1 classifier:31 demonstrates:2 producing:2 arguably:1 positive:18 modify:2 tsinghua:2 treat:1 tends:1 approximately:1 alexey:1 collect:5 limited:1 range:1 bi:1 statistically:1 lecun:1 testing:1 practice:1 prevalence:1 evan:1 yan:2 significantly:2 adapting:7 matching:2 reject:1 pre:3 confidence:1 seeing:1 get:1 wheel:1 nb:1 applying:2 equivalent:2 map:10 demonstrated:1 williams:1 starting:1 insight:2 fill:1 embedding:1 limiting:1 target:4 suppose:1 hierarchy:1 us:5 recognition:8 showcase:1 breakdown:2 asymmetric:1 labeled:17 database:1 bottom:2 ft:1 preprint:1 tsang:1 thousand:3 region:14 ilsvrc2012:2 connected:1 sun:1 highest:1 intuition:1 transforming:1 mine:1 geodesic:1 trained:11 weakly:1 ov:7 ali:2 localization:7 distinctive:1 eric:1 mislabeled:2 fca:1 darpa:1 finetuning:1 joint:2 cat:4 various:1 train:14 fast:4 describe:1 effective:2 caffe:2 emerged:2 cvpr:4 ability:1 fischer:1 transform:2 final:4 karayev:1 net:5 propose:3 adaptation:36 combining:1 ablation:2 motorcycle:3 flexibility:1 achieve:1 adapts:1 deformable:1 sutskever:1 darrell:9 produce:11 object:39 fixing:1 exemplar:2 nearest:4 progress:1 sim:7 strong:4 implemented:2 c:2 indicate:1 come:1 annotated:13 cnns:4 modifying:1 human:2 wid:1 enable:2 require:1 im:1 visually:1 cb:5 bicycle:1 major:1 achieves:1 adopt:1 released:2 proc:11 sofa:1 bag:3 label:22 ross:2 sensitive:1 saw:1 successfully:2 hoffman:5 hope:2 aim:1 modified:2 mil:4 focus:1 june:1 improvement:4 contrast:3 suppression:1 baseline:5 detect:1 nn:3 entire:1 transferring:1 integrated:1 fcb:7 selective:2 layers1:4 classification:59 pascal:3 augment:1 proposes:1 animal:2 art:4 softmax:1 tzeng:3 brox:1 uhl:1 uc:2 initialize:1 construct:2 extraction:1 having:3 orange:1 look:2 icml:3 nearly:1 plenty:1 future:2 report:1 quantitatively:1 dosovitskiy:1 randomly:1 oriented:1 cheaper:1 ab:2 detection:93 interest:2 evaluation:3 adjust:1 oth:5 held:14 accurate:1 capable:1 tree:1 euclidean:1 initialized:2 girshick:5 smisc:1 minimal:2 instance:4 classify:2 modeling:1 predic:1 localizing:1 cost:1 subset:2 hundred:1 comprised:2 recognizing:1 krizhevsky:2 eec:4 accomplish:1 fundamental:1 international:5 together:1 quickly:1 management:1 warped:1 american:1 potential:2 de:1 availability:1 includes:1 automation:1 kate:1 caused:1 bg:6 piece:2 lowe:1 doing:1 red:3 competitive:1 start:2 portion:1 capability:1 annotation:12 gevers:1 jia:2 contribution:1 smeulders:1 minimize:1 publicly:1 convolutional:12 descriptor:1 correspond:2 yes:1 weak:2 raw:1 accurately:1 produced:1 ren:1 apple:4 detector:40 reach:1 opez:1 manual:1 trevor:3 definition:1 proved:1 lowell:1 dataset:10 ask:1 popular:1 knowledge:3 distractors:1 improves:2 car:1 segmentation:1 wjd:1 supervised:4 day:1 zisserman:4 evaluated:1 box:37 though:3 execute:1 just:1 stage:1 lack:4 quality:1 effect:2 contain:3 normalized:1 concept:1 hence:1 regularization:1 semantic:1 during:2 berkeleyvision:2 scorei:1 demonstrate:3 confusion:6 image:44 recently:4 propelled:1 empirically:1 winner:1 million:2 occurred:1 he:1 significant:3 refer:1 tuning:14 rasa:2 aytar:3 had:2 dot:1 access:3 supervision:1 fc7:4 sergio:2 something:2 dominant:1 add:1 showed:2 recent:2 scenario:1 sande:1 scoring:6 additional:1 tabula:2 dai:1 deng:1 accomplishes:1 determine:1 paradigm:1 ii:3 semi:1 multiple:4 full:7 harchaoui:1 reduces:2 faster:1 adapt:12 offer:2 cross:1 long:1 retrieval:1 icdm:1 award:1 paired:2 overfeat:1 prediction:1 variant:1 regression:2 heterogeneous:1 essentially:1 vision:6 enhancing:1 arxiv:3 histogram:1 kernel:1 sergey:1 pyramid:2 achieved:1 robotics:1 ahenb:1 proposal:10 background:19 want:1 fine:16 winn:1 source:6 operate:1 unlike:1 bringing:2 pooling:2 effectiveness:1 rodner:3 ee:1 yang:2 split:1 enough:1 easy:1 fit:2 architecture:5 reduce:1 idea:1 cn:1 det:7 absent:1 shift:1 whether:2 padding:1 song:1 deep:10 clear:1 detailed:1 tune:2 amount:4 transforms:4 ten:3 induces:1 svms:6 category:107 reduced:1 percentage:4 nsf:1 misclassifying:1 shifted:1 per:9 correctly:3 blue:1 key:3 yangqing:1 localize:3 fraction:1 run:1 package:1 you:3 nail:2 patch:2 scaling:2 layer:24 bound:1 rbg:1 correspondence:1 oracle:2 adapted:2 lemon:3 constraint:1 fei:2 scene:1 software:3 encodes:1 tag:1 dominated:1 simulate:1 c8:2 relatively:1 uml:1 transferred:2 combination:1 smaller:1 across:1 shallow:1 outlier:1 invariant:7 iccv:1 pipeline:1 previously:1 needed:3 idet:1 know:1 serf:1 available:16 generic:4 reshape:1 eigen:1 original:2 thomas:1 top:10 running:2 include:2 remaining:1 prof:1 comparatively:1 icra:1 warping:1 malik:1 noticed:1 print:1 quantity:2 wjc:1 gradient:1 iclr:1 distance:1 separating:2 capacity:1 majority:2 mail:1 discriminant:1 reason:1 code:1 relationship:1 providing:1 balance:1 acquire:1 sermanet:1 setup:1 unfortunately:1 hog:1 ronghang:1 implementation:1 perform:1 subcategory:1 upper:1 datasets:1 benchmark:2 incorrectly:2 extended:1 hinton:1 precise:1 community:1 dog:6 required:1 cast:1 imagenet:13 engine:2 learned:4 hour:1 boost:1 nip:1 trans:1 able:4 pattern:3 challenge:12 program:1 including:1 green:1 max:1 video:2 gool:1 suitable:1 difficulty:1 turning:1 ramos:1 localizes:4 historically:1 rated:1 mathieu:1 extract:2 relative:3 fully:3 loss:2 localized:2 validation:7 shelhamer:1 jegelka:1 sufficient:1 wic:1 row:1 eccv:3 summary:1 supported:1 last:3 keeping:1 bias:1 allow:2 wide:1 neighbor:3 taking:1 face:1 benefit:1 van:2 rich:1 forward:1 made:3 adaptive:3 avg:3 far:1 alphabetically:2 lsda:29 approximate:1 ons:1 mairal:1 hoeim:2 discriminative:1 fergus:1 search:5 latent:2 table:2 additionally:4 promising:1 learn:10 transfer:6 fc6:4 ca:2 carterette:1 obtaining:1 val1:3 uijlings:1 domain:23 protocol:1 did:1 significance:1 main:2 bounding:35 whole:2 body:1 xu:2 augmented:1 biggest:2 deployed:1 judy:1 precision:1 spatial:2 toyota:1 learns:5 donahue:6 removing:1 down:1 british:1 specific:9 sift:2 svm:2 essential:1 workshop:1 mnist:1 false:13 adding:1 effectively:1 importance:2 corr:1 decaf:1 kr:1 hauptmann:2 gap:2 easier:1 rejection:1 bison:1 likely:1 explore:1 ivor:1 visual:8 vinyals:1 partially:1 applies:2 corresponds:1 extracted:3 acm:1 goal:2 sorted:1 towards:1 jeff:2 replace:1 shared:1 change:3 lsvrc:1 except:1 reducing:2 microphone:2 total:3 multimedia:1 saenko:6 ilsvrc:2 berg:1 support:2 azquez:1 jonathan:1 bnb:1 evaluate:1 ex:2 |
4,881 | 5,419 | Local Decorrelation for Improved Pedestrian Detection
Woonhyun Nam?
StradVision, Inc.
[email protected]
Piotr Doll?ar
Microsoft Research
Joon Hee Han
POSTECH, Republic of Korea
[email protected]
[email protected]
Abstract
Even with the advent of more sophisticated, data-hungry methods, boosted decision trees remain extraordinarily successful for fast rigid object detection, achieving top accuracy on numerous datasets. While effective, most boosted detectors
use decision trees with orthogonal (single feature) splits, and the topology of the
resulting decision boundary may not be well matched to the natural topology of
the data. Given highly correlated data, decision trees with oblique (multiple feature) splits can be effective. Use of oblique splits, however, comes at considerable
computational expense. Inspired by recent work on discriminative decorrelation
of HOG features, we instead propose an efficient feature transform that removes
correlations in local neighborhoods. The result is an overcomplete but locally
decorrelated representation ideally suited for use with orthogonal decision trees.
In fact, orthogonal trees with our locally decorrelated features outperform oblique
trees trained over the original features at a fraction of the computational cost. The
overall improvement in accuracy is dramatic: on the Caltech Pedestrian Dataset,
we reduce false positives nearly tenfold over the previous state-of-the-art.
1
Introduction
In recent years object detectors have undergone an impressive transformation [11, 32, 14]. Nevertheless, boosted detectors remain extraordinarily successful for fast detection of quasi-rigid objects.
Such detectors were first proposed by Viola and Jones in their landmark work on efficient sliding
window detection that made face detection practical and commercially viable [35]. This initial architecture remains largely intact today: boosting [31, 12] is used to train and combine decision trees
and a cascade is employed to allow for fast rejection of negative samples. Details, however, have
evolved considerably; in particular, significant progress has been made on the feature representation [6, 9, 2] and cascade architecture [3, 8]. Recent boosted detectors [1, 7] achieve state-of-the-art
accuracy on modern benchmarks [10, 22] while retaining computational efficiency.
While boosted detectors have evolved considerably over the past decade, decision trees with orthogonal (single feature) splits ? also known as axis-aligned decision trees ? remain popular and predominant. A possible explanation for the persistence of orthogonal splits is their efficiency: oblique
(multiple feature) splits incur considerable computational cost during both training and detection.
Nevertheless, oblique trees can hold considerable advantages. In particular, Menze et al. [23] recently demonstrated that oblique trees used in conjunction with random forests are quite effective
given high dimensional data with heavily correlated features.
To achieve similar advantages while avoiding the computational expense of oblique trees, we instead
take inspiration from recent work by Hariharan et al. [15] and propose to decorrelate features prior to
applying orthogonal trees. To do so we introduce an efficient feature transform that removes correlations in local image neighborhoods (as opposed to decorrelating features globally as in [15]). The
result is an overcomplete but locally decorrelated representation that is ideally suited for use with
orthogonal trees. In fact, orthogonal trees with our locally decorrelated features require estimation
of fewer parameters and actually outperform oblique trees trained over the original features.
?
This research was performed while W.N. was a postdoctoral researcher at POSTECH.
1
Figure 1: A comparison of boosting of orthogonal and oblique trees on highly correlated data while
varying the number (T ) and depth (D) of the trees. Observe that orthogonal trees generalize poorly
as the topology of the decision boundary is not well aligned to the natural topology of the data.
We evaluate boosted decision tree learning with decorrelated features in the context of pedestrian
detection. As our baseline we utilize the aggregated channel features (ACF) detector [7], a popular,
top-performing detector for which source code is available online. Coupled with use of deeper trees
and a denser sampling of the data, the improvement obtained using our locally decorrelated channel
features (LDCF) is substantial. While in the past year the use of deep learning [25], motion features
[27], and multi-resolution models [36] has brought down log-average miss rate (MR) to under 40%
on the Caltech Pedestrian Dataset [10], LDCF reduces MR to under 25%. This translates to a nearly
tenfold reduction in false positives over the (very recent) state-of-the-art.
The paper is organized as follows. In ?2 we review orthogonal and oblique trees and demonstrate
that orthogonal trees trained on decorrelated data may be equally or more effective as oblique trees
trained on the original data. We introduce the baseline in ?3 and in ?4 show that use of oblique
trees improves results but at considerable computational expense. Next, in ?5, we demonstrate that
orthogonal trees trained with locally decorrelated features are efficient and effective. Experiments
and results are presented in ?6. We begin by briefly reviewing related work next.
1.1
Related Work
Pedestrian Detection: Recent work in pedestrian detection includes use of deformable part models
and their extensions [11, 36, 26], convolutional nets and deep learning [33, 37, 25], and approaches
that focus on optimization and learning [20, 18, 34]. Boosted detectors are also widely used. In
particular, the channel features detectors [9, 1, 2, 7] are a family of conceptually straightforward and
efficient detectors based on boosted decision trees computed over multiple feature channels such as
color, gradient magnitude, gradient orientation and others. Current top results on the INRIA [6] and
Caltech [10] Pedestrian Datasets include instances of the channel features detector with additional
mid-level edge features [19] and motion features [27], respectively.
Oblique Decision Trees: Typically, decision trees are trained with orthogonal (single feature) splits;
however, the extension to oblique (multiple feature) splits is fairly intuitive and well known, see
e.g. [24]. In fact, Breiman?s foundational work on random forests [5] experimented with oblique
trees. Recently there has been renewed interest in random forests with oblique splits [23, 30] and
Marin et al. [20] even applied such a technique to pedestrian detection. Likewise, while typically
orthogonal trees are used with boosting [12], oblique trees can easily be used instead. The contribution of this work is not the straightforward coupling of oblique trees with boosting, rather, we
propose a local decorrelation transform that eliminates the necessity of oblique splits altogether.
Decorrelation: Decorrelation is a common pre-processing step for classification [17, 15]. In recent
work, Hariharan et al. [15] proposed an efficient scheme for estimating covariances between HOG
features [6] with the goal of replacing linear SVMs with LDA and thus allowing for fast training.
Hariharan et al. demonstrated that the global covariance matrix for a detection window can be estimated efficiently as the covariance between two features should depend only on their relative offset.
Inspired by [15], we likewise exploit the stationarity of natural image statistics, but instead propose
to estimate a local covariance matrix shared across all image patches. Next, rather than applying
global decorrelation, which would be computationally prohibitive for sliding window detection with
a nonlinear classifier1 , we instead propose to apply an efficient local decorrelation transform. The
result is an overcomplete representation well suited for use with orthogonal trees.
1
Global decorrelation coupled with a linear classifier is efficient as the two linear operations can be merged.
2
Figure 2: A comparison of boosting with orthogonal decision trees (T = 5) on transformed data.
Orthogonal trees with both decorrelated and PCA-whitened features show improved generalization
while ZCA-whitening is ineffective. Decorrelating the features is critical, while scaling is not.
2
Boosted Decision Trees with Correlated Data
Boosting is a simple yet powerful tool for classification and can model complex non-linear functions
[31, 12]. The general idea is to train and combine a number of weak learners into a more powerful
strong classifier. Decision trees are frequently used as the weak learner in conjunction with boosting,
and in particular orthogonal decision trees, that is trees in which every split is a threshold on a single
feature, are especially popular due to their speed and simplicity [35, 7, 1].
The representational power obtained by boosting orthogonal trees is not limited by use of orthogonal
splits; however, the number and depth of the trees necessary to fit the data may be large. This can
lead to complex decision boundaries and poor generalization, especially given highly correlated
features. Figure 1(a)-(c) shows the result of boosted orthogonal trees on correlated data. Observe
that the orthogonal trees generalize poorly even as we vary the number and depth of the trees.
Decision trees with oblique splits can more effectively model data with correlated features as the
topology of the resulting classifier can better match the natural topology of the data [23]. In oblique
trees, every split is based on a linear projection of the data z = w| x followed by thresholding.
The projection w can be sparse (and orthogonal splits are a special case with kwk0 = 1). While
in principle numerous approaches can be used to obtain w, in practice linear discriminant analysis
(LDA) is a natural choice for obtaining discriminative splits efficiently [16]. LDA aims to minimize
within-class scatter while maximizing between-class scatter. w is computed from class-conditional
mean vectors ?+ and ?? and a class-independent covariance matrix ? as follows:
w = ??1 (?+ ? ?? ).
(1)
The covariance may be degenerate if the amount or underlying dimension of the data is low; in this
case LDA can be regularized by using (1 ? )? + I in place of ?. In Figure 1(d) we apply boosted
oblique trees trained with LDA on the same data as before. Observe the resulting decision boundary
better matches the underlying data distribution and shows improved generalization.
The connection between whitening and LDA is well known [15]. Specifically, LDA simplifies to a
trivial classification rule on whitened data (data whose covariance is the identity). Let ? = Q?Q|
be the eigendecomposition1 of ? where
Q is an orthogonal matrix and ? is a diagonal matrix of
1
eigenvalues. W = Q?? 2 Q| = ?? 2 is known as a whitening matrix because the covariance of
x0 = Wx is the identity matrix. Given whitened data and means, LDA can be interpreted as
|
|
learning the trivial projection w0 = ?0+ ? ?0? = W?+ ? W?? since w0 x0 = w0 Wx = w| x.
Can whitening or a related transform likewise simplify learning of boosted decision trees?
Using standard terminology
[17], we define the following
related transforms: decorrelation (Q| ),
? 21 |
? 21 |
PCA-whitening (? Q ), and ZCA-whitening (Q? Q ). Figure 2 shows the result of boosting
orthogonal trees on the variously transformed features, using the same data as before. Observe that
with decorrelated and PCA-whitened features orthogonal trees show improved generalization. In
fact, as each split is invariant to scaling of individual features, orthogonal trees with PCA-whitened
and decorrelated features give identical results. Decorrelating the features is critical, while scaling
is not. The intuition is clear: each split operates on a single feature, which is most effective if
the features are decorrelated. Interestingly, the standard ZCA-whitened transform used by LDA is
ineffective: while the resulting features are not technically correlated, due to the additional rotation
by Q each resulting feature is a linear combination of features obtained by PCA-whitening.
3
3
Baseline Detector (ACF)
We next briefly review our baseline detector and evaluation benchmark. This will allow us to apply
the ideas from ?2 to object detection in subsequent sections. In this work we utilize the channel
features detectors [9, 7, 1, 2], a family of conceptually straightforward and efficient detectors for
which variants have been utilized for diverse tasks such as pedestrian detection [10], sign recognition
[22] and edge detection [19]. Specifically, for our experiments we focus on pedestrian detection and
employ the aggregate channel features (ACF) variant [7] for which code is available online2 .
Given an input image, ACF computes several feature channels, where each channel is a per-pixel
feature map such that output pixels are computed from corresponding patches of input pixels (thus
preserving image layout). We use the same channels as [7]: normalized gradient magnitude (1 channel), histogram of oriented gradients (6 channels), and LUV color channels (3 channels), for a total
of 10 channels. We downsample the channels by 2x and features are single pixel lookups in the
aggregated channels. Thus, given a h ? w detection window, there are h/2 ? w/2 ? 10 candidate
features (channel pixel lookups). We use RealBoost [12] with multiple rounds of bootstrapping to
train and combine 2048 depth-3 decision trees over these features to distinguish object from background. Soft-cascades [3] and an efficient multiscale sliding-window approach are employed. Our
baseline uses slightly altered parameters from [7] (RealBoost, deeper trees, and less downsampling);
this increases model capacity and benefits our final approach as we report in detail in ?6.
Current practice is to use the INRIA Pedestrian Dataset [6] for parameter tuning, with the test set
serving as a validation set, see e.g. [20, 2, 9]. We utilize this dataset in much the same way and report
full results on the more challenging Caltech Pedestrian Dataset [10]. Following the methodology
of [10], we summarize performance using the log-average miss rate (MR) between 10?2 and 100
false positives per image. We repeat all experiments 10 times and report the mean MR and standard
error for every result. Due to the use of a log-log scale, even small improvements in (log-average)
MR correspond to large reductions in false-positives. On INRIA, our (slightly modified) baseline
version of ACF scores at 17.3% MR compared to 17.0% MR for the model reported in [7].
4
Detection with Oblique Splits (ACF-LDA)
In this section we modify the ACF detector to enable oblique splits and report the resulting gains.
Recall that given input x, at each split of an oblique decision tree we need to compute z = w| x for
some projection w and threshold the result. For our baseline pedestrian detector, we use 128 ? 64
windows where each window is represented by a feature vector x of size 128/2 ? 64/2 ? 10 = 20480
(see ?3). Given the high dimensionality of the input x coupled with the use of thousands of trees in
a typical boosted classifier, for efficiency w must be sparse.
Local w: We opt to use w?s that correspond to local m?m blocks of pixels. In other words, we treat
x as a h/2 ? w/2 ? 10 tensor and allow w to operate over any m ? m ? 1 patch in a single channel
of x. Doing so holds multiple advantages. Most importantly, each pixel has strongest correlations
to spatially nearby pixels [15]. Since oblique splits are expected to help most when features are
strongly correlated, operating over local neighborhoods is a natural choice. In addition, using local
w allows for faster lookups due to the locality of adjacent pixels in memory.
Complexity: First, let us consider the complexity of training the oblique splits. Let d = h/2?w/2 be
the window size of a single channel. The number of patches per channel in x is about d, thus naively
training a single split means applying LDA d times ? once per patch ? and keeping w with lowest
error. Instead of computing d independent matrices ? per channel, for efficiency, we compute ?, a
d ? d covariance matrix for the entire window, and reconstruct individual m2 ? m2 ??s by fetching
appropriate entries from ?. A similar trick can be used for the ??s. Computing ? is O(nd2 ) given
n training examples (and could be made faster by omitting unnecessary elements). Inverting each
?, the bottleneck of computing Eq. (1), is O(dm6 ) but independent of n and thus fairly small as
n m. Finally computing z = w| x over all n training examples and d projections is O(ndm2 ).
Given the high complexity of each step, a naive brute-force approach for training is infeasible.
Speedup: While the weights over training examples change at every boosting iteration and after
every tree split, in practice we find it is unnecessary to recompute the projections that frequently.
Table 1, rows 2-4, shows the results of ACF with oblique splits, updated every T boosting iterations
2
http://vision.ucsd.edu/?pdollar/toolbox/doc/
4
Shared ?
T
Miss Rate
Training
-
-
17.3 ? .33
4.93m
ACF-LDA-4
ACF-LDA-16
ACF-LDA-?
No
No
No
4
16
?
14.9 ? .37
15.1 ? .28
17.0 ? .22
303.57m
78.11m
5.82m
ACF-LDA? -4
ACF-LDA? -16
ACF-LDA? -?
Yes
Yes
Yes
4
16
?
14.7 ? .29
15.1 ? .12
16.4 ? .17
194.26m
51.19m
5.79m
LDCF
Yes
-
13.7 ? .15
6.04m
ACF
Table 1: A comparison of boosted trees with orthogonal and oblique splits.
(denoted by ACF-LDA-T ). While more frequent updates improve accuracy, ACF-LDA-16 has negligibly higher MR than ACF-LDA-4 but a nearly fourfold reduction in training time (timed using 12
cores). Training the brute force version of ACF-LDA, updated at every iteration and each tree split
(7 interior nodes per depth-3 tree) would have taken about 5 ? 4 ? 7 = 140 hours. For these results we
used regularization of = .1 and patch size of m = 5 (effect of varying m is explored in ?6).
Shared ?: The crux and computational bottleneck of ACF-LDA is the computation and application
of a separate covariance ? at each local neighborhood. In recent work on training linear object
detectors using LDA, Hariharan et al. [15] exploited the observation that the statistics of natural
images are translationally invariant and therefore the covariance between two features should depend
only on their relative offset. Furthermore, as positives are rare, [15] showed that the covariances can
be precomputed using natural images. Inspired by these observations, we propose to use a single,
fixed covariance ? shared across all local image neighborhoods. We precompute one ? per channel
and do not allow it to vary spatially or with boosting iteration. Table 1, rows 5-7, shows the results
of ACF with oblique splits using fixed ?, denoted by ACF-LDA? . As before, the ??s and resulting
w are updated every T iterations. As expected, training time is reduced relative to ACF-LDA.
Surprisingly, however, accuracy improves as well, presumably due to the implicit regularization
effect of using a fixed ?. This is a powerful result we will exploit further.
Summary: ACF with local oblique splits and a single shared ? (ACF-LDA? -4) achieves 14.7% MR
compared to 17.3% MR for ACF with orthogonal splits. The 2.6% improvement in log-average MR
corresponds to a nearly twofold reduction in false positives but comes at considerable computational
cost. In the next section, we propose an alternative, more efficient approach for exploiting the use of
a single shared ? capturing correlations in local neighborhoods.
5
Locally Decorrelated Channel Features (LDCF)
We now have all the necessary ingredients to introduce our approach. We have made the following
observations: (1) oblique splits learned with LDA over local m ? m patches improve results over
orthogonal splits, (2) a single covariance matrix ? can be shared across all patches per channel, and
(3) orthogonal trees with decorrelated features can potentially be used in place of oblique trees. This
suggests the following approach: for every m ? m patch p in x, we can create a decorrelated representation by computing Q| p, where Q?Q| is the eigendecomposition of ? as before, followed by
use of orthogonal trees. However, such an approach is computationally expensive.
First, due to use of overlapping patches, computing Q| p for every overlapping patch results in an
overcomplete representation with a factor m2 increase in feature dimensionality. To reduce dimensionality, we only utilize the top k eigenvectors in Q, resulting in k < m2 features per pixel. The
intuition is that the top eigenvectors capture the salient neighborhood structure. Our experiments
in ?6 confirm this: using as few as k = 4 eigenvectors per channel for patches of size m = 5 is
sufficient. As our second speedup, we observe that the projection Q| p can be computed by a series
of k convolutions between a channel image and each m ? m filter reshaped from its corresponding
eigenvector (column of Q). This is possible because the covariance matrix ? is shared across all
patches per channel and hence the derived Q is likewise spatially invariant. Decorrelating all 10
channels in an entire feature pyramid for a 640 ? 480 image takes about .5 seconds.
5
Figure 3: Top-left: autocorrelation for each channel. Bottom-left: learned decorrelation filters.
Right: visualization of original and decorrelated channels averaged over positive training examples.
In summary, we modify ACF by taking the original 10 channels and applying k = 4 decorrelating
(linear) filters per channel. The result is a set of 40 locally decorrelated channel features (LDCF).
To further increase efficiency, we downsample the decorrelated channels by a factor of 2x which has
negligible impact on accuracy but reduces feature dimension to the original value. Given the new
locally decorrelated channels, all other steps of ACF training and testing are identical. The extra
implementation effort is likewise minimal: given the decorrelation filters, a few lines of code suffice
to convert ACF into LDCF. To further improve clarity, all source code for LDCF will be released.
Results of the LDCF detector on the INRIA dataset are given in the last row of Table 1. The
LCDF detector (which uses orthogonal splits) improves accuracy over ACF with oblique splits by
an additional 1% MR. Training time is significantly faster, and indeed, is only ?1 minute longer
than for the original ACF detector. More detailed experiments and results are reported in ?6. We
conclude by (1) describing the estimation of ? for each channel, (2) showing various visualizations,
and (3) discussing the filters themselves and connections to known filters.
Estimating ?: We can estimate a spatially constant ? for each channel using any large collection of natural images. ? for each channel is represented by a spatial autocorrelation function
?(x,y),(x+?x,y+?y) = C(?x, ?y). Given a collection of natural images, we first estimate a separate autocorrelation function for each image and then average the results. Naive computation of the
final function is O(np2 ) but the Wiener-Khinchin theorem reduces the complexity to O(np log p)
via the FFT [4], where n and p denote the number of images and pixels per image, respectively.
Visualization: Fig. 3, top-left, illustrates the estimated autocorrelations for each channel. Nearby
features are highly correlated and oriented gradients are spatially correlated along their orientation
due to curvilinear continuity [15]. Fig. 3, bottom-left, shows the decorrelation filters for each channel obtained by reshaping the largest eigenvectors of ?. The largest eigenvectors are smoothing
filters while the smaller ones resemble increasingly higher-frequency filters. The corresponding
eigenvalues decay rapidly and in practice we use the top k = 4 filters. Observe that the decorrelation filters for oriented gradients are aligned to their orientation. Finally, Fig. 3, right, shows original
and decorrelated channels averaged over positive training examples.
Discussion: Our decorrelation filters are closely related to sinusoidal, DCT basis, and Gaussian
derivative filters. Spatial interactions in natural images are often well-described by Markov models [13] and first-order stationary Markov processes are known to have sinusoidal KLT bases [29].
In particular, for the LUV color channels, our filters are similar to the discrete cosine transform
(DCT) bases that are often used to approximate the KLT. For oriented gradients, however, the decorrelation filters are no longer well modeled by the DCT bases (note also that our filters are applied
densely whereas the DCT typically uses block processing). Alternatively, we can interpret our filters
as Gaussian derivative filters. Assume that the autocorrelation is modeled by a squared-exponential
function C(?x) = exp(??x2 /2l2 ), which is fairly reasonable given the estimation results in Fig. 3.
In 1D, the k th largest eigenfunction of such an autocorrelation function is a k ? 1 order Gaussian
derivative filter [28]. It is straightforward to extend the result to an anisotropic multivariate case in
which case the eigenfunctions are Gaussian directional derivative filters similar to our filters.
6
Figure 4: (a-b) Use of k = 4 local decorrelation filters of size m = 5 gives optimal performance.
(c) Increasing tree depth while simultaneously enlarging the quantity of data available for training
can have a large impact on accuracy (blue stars indicate optimal depth at each sampling interval).
description
1.
2.
3.
4.
5.
6.
7.
ACF
LDCF small ?
LDCF random
LDCF LUV only
LDCF grad only
LDCF constant
LDCF
# channels
(modified) baseline
decorrelation w k smallest filters
filtering w k random filters
decorrelation of LUV channels only
decorrelation of grad channels only
decorrelation w constant filters
proposed approach
10
10k
10k
3k + 7
3 + 7k
10k
10k
miss rate
17.3 ? .33
61.7 ? .28
15.6 ? .26
16.2 ? .37
14.9 ? .29
14.2 ? .34
13.7 ? .15
Table 2: Locally decorrelated channels compared to alternate filtering strategies. See text.
6
Experiments
In this section, we demonstrate the effectiveness of locally decorrelated channel features (LDCF) in
the context of pedestrian detection. We: (1) study the effect of parameter settings, (2) test variations
of our approach, and finally (3) compare our results with the state-of-the-art.
Parameters: LDCF has two parameters: the count and size of the decorrelation filters. Fig. 4(a) and
(b) show the results of LDCF on the INRIA dataset while varying the filter count (k) and size (m),
respectively. Use of k = 4 decorrelation filters of size m = 5 improves performance up to ?4% MR
compared to ACF. Inclusion of additional higher-frequency filters or use of larger filters can cause
performance degradation. For all remaining experiments we fix k = 4 and m = 5.
Variations: We test variants of LDCF and report results on INRIA in Table 2. LDCF (row 7) outperforms all variants, including the baseline (1). Filtering the channels with the smallest k eigenvectors
(2) or k random filters (3) performs worse. Local decorrelation of only the color channels (4) or
only the gradient channels (5) is inferior to decorrelation of all channels. Finally, we test constant
decorrelation filters obtained from the intensity channel L that resemble the first k DCT basis filters.
Use of unique filters per channel outperforms use of constant filters across all channels (6).
Model Capacity: Use of locally decorrelated features implicitly allows for richer, more effective
splitting functions, increasing modeling capacity and generalization ability. Inspired by their success, we explore additional strategies for augmenting model capacity. For the following experiments,
we rely solely on the training set of the Caltech Pedestrian Dataset [10]. Of the 71 minute long training videos (?128k images), we use every fourth video as validation data and the rest for training. On
the validation set, LDCF outperforms ACF by a considerable margin, reducing MR from 46.2% to
41.7%. We first augment model capacity by increasing the number of trees twofold (to 4096) and the
sampled negatives fivefold (to 50k). Surprisingly, doing so reduces MR by an additional 4%. Next,
we experiment with increasing maximum tree depth while simultaneously enlarging the amount of
data available for training. Typically, every 30th image in the Caltech dataset is used for training and
testing. Instead, Figure 4(c) shows validation performance of LDCF with different tree depths while
varying the training data sampling interval. The impact of maximum depth on performance is quite
large. At a dense sampling interval of every 4th frame, use of depth-5 trees (up from depth-2 for the
original approach) improves performance by an additional 5% to 32.6% MR. Note that consistent
with the generalization bounds of boosting [31], use of deeper trees requires more data.
7
1
1
72% VJ
46% HOG
21% pAUCBoost
20% FisherBoost
20% LatSvm?V2
20% ConvNet
19% CrossTalk
17% ACF
16% VeryFast
15% RandForest
14% LDCF
14% Franken
14% Roerei
13% SketchTokens
.64
.50
.40
miss rate
.30
.20
.80
.64
.50
.40
.30
miss rate
.80
.10
.10
.05
.05
?2
10
?1
10
0
10
1
95% VJ
68% HOG
48% DBN?Mut
46% MF+Motion+2Ped
46% MOCO
45% MultiSDP
44% ACF?Caltech
43% MultiResC+2Ped
41% MT?DPM
39% JointDeep
38% MT?DPM+Context
37% ACF+SDt
30% ACF?Caltech+
25% LDCF
.20
?3
10
10
false positives per image
?2
10
?1
10
0
10
1
10
false positives per image
(a) INRIA Pedestrian Dataset
(b) Caltech Pedestrian Dataset
Figure 5: A comparison of our LDCF detector with state-of-the-art pedestrian detectors.
INRIA Results: In Figure 5(a) we compare LDCF with state-of-the-art detectors on INRIA [6]
using benchmark code maintained by [10]. Since the INRIA dataset is oft-used as a validation set,
including in this work, we include these results for completeness only. LDCF is essentially tied for
second place with Roerei [2] and Franken [21] and outperformed by ?1% MR by SketchTokens [19].
These approaches all belong to the family of channel features detectors, and as the improvements
proposed in this work are orthogonal, the methods could potentially be combined.
Caltech Results: We present our main result on the Caltech Pedestrian Dataset [10], see Fig. 5(b),
generated using the official evaluation code available online3 . The Caltech dataset has become the
standard for evaluating pedestrian detectors and the latest methods based on deep learning (JointDeep) [25], multi-resolution models (MT-DPM) [36] and motion features (ACF+SDt) [27] achieve
under 40% log-average MR. For a complete comparison, we first present results for an augmented
capacity ACF model which uses more (4096) and deeper (depth-5) trees trained with RealBoost
using dense sampling of the training data (every 4th image). See preceding note on model capacity
for details and motivation. This augmented model (ACF-Caltech+) achieves 29.8% MR, a considerable nearly 10% MR gain over previous methods, including the baseline version of ACF (ACFCaltech). With identical parameters, locally decorrelated channel features (LDCF) further reduce
error to 24.9% MR with substantial gains at higher recall. Overall, this is a massive improvement
and represents a nearly 10x reduction in false positives over the previous state-of-the-art.
7
Conclusion
In this work we have presented a simple, principled approach for improving boosted object detectors.
Our core observation was that effective but expensive oblique splits in decision trees can be replaced
by orthogonal splits over locally decorrelated data. Moreover, due to the stationary statistics of
image features, the local decorrelation can be performed efficiently via convolution with a fixed
filter bank precomputed from natural images. Our approach is general, simple and fast.
Our method showed dramatic improvement over previous state-of-the-art. While some of the gain
was from increasing model capacity, use of local decorrelation gave a clear and significant boost.
Overall, we reduced false-positives tenfold on Caltech. Such large gains are fairly rare.
In the present work we did not decorrelate features across channels (decorrelation was applied independently per channel). This is a clear future direction. Testing local decorrelation in the context of
other classifiers (e.g. convolutional nets or linear classifiers as in [15]) would also be interesting.
While the proposed locally decorrelated channel features (LDCF) require only modest modification
to existing code, we will release all source code used in this work to ease reproducibility.
3
http://www.vision.caltech.edu/Image_Datasets/CaltechPedestrians/
8
References
[1] R. Benenson, M. Mathias, R. Timofte, and L. Van Gool. Pedestrian detection at 100 frames per second.
In CVPR, 2012.
[2] R. Benenson, M. Mathias, T. Tuytelaars, and L. Van Gool. Seeking the strongest rigid detector. In CVPR,
2013.
[3] L. Bourdev and J. Brandt. Robust object detection via soft cascade. In CVPR, 2005.
[4] G. Box, G. Jenkins, and G. Reinsel. Time series analysis: forecasting and control. Prentice Hall, 1994.
[5] L. Breiman. Random forests. Machine Learning, 45(1):5?32, Oct. 2001.
[6] N. Dalal and B. Triggs. Histograms of oriented gradients for human detection. In CVPR, 2005.
[7] P. Doll?ar, R. Appel, S. Belongie, and P. Perona. Fast feature pyramids for object detection. PAMI, 2014.
[8] P. Doll?ar, R. Appel, and W. Kienzle. Crosstalk cascades for frame-rate pedestrian detection. In ECCV,
2012.
[9] P. Doll?ar, Z. Tu, P. Perona, and S. Belongie. Integral channel features. In BMVC, 2009.
[10] P. Doll?ar, C. Wojek, B. Schiele, and P. Perona. Pedestrian detection: An evaluation of the state of the art.
PAMI, 34, 2012.
[11] P. Felzenszwalb, R. Girshick, D. McAllester, and D. Ramanan. Object detection with discriminatively
trained part based models. PAMI, 32(9):1627?1645, 2010.
[12] J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: a statistical view of boosting. The
Annals of Statistics, 38(2):337?374, 2000.
[13] S. Geman and D. Geman. Stochastic relaxation, gibbs distributions, and the bayesian restoration of
images. PAMI, PAMI-6(6):721?741, 1984.
[14] R. B. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich feature hierarchies for accurate object detection
and emantic segmentation. In CVPR, 2014.
[15] B. Hariharan, J. Malik, and D. Ramanan. Discriminative decorrelation for clustering and classification.
In ECCV, 2012.
[16] T. Hastie, R. Tibshirani, and J. Friedman. The elements of statistical learning. Springer, 2009.
[17] A. Krizhevsky and G. Hinton. Learning multiple layers of features from tiny images. Master?s thesis,
Department of Computer Science, University of Toronto, 2009.
[18] D. Levi, S. Silberstein, and A. Bar-Hillel. Fast multiple-part based object detection using kd-ferns. In
CVPR, 2013.
[19] J. Lim, C. L. Zitnick, and P. Doll?ar. Sketch tokens: A learned mid-level representation for contour and
object detection. In CVPR, 2013.
[20] J. Mar??n, D. V?azquez, A. L?opez, J. Amores, and B. Leibe. Random forests of local experts for pedestrian
detection. In ICCV, 2013.
[21] M. Mathias, R. Benenson, R. Timofte, and L. Van Gool. Handling occlusions with franken-classifiers. In
ICCV, 2013.
[22] M. Mathias, R. Timofte, R. Benenson, and L. Van Gool. Traffic sign recognition - how far are we from
the solution? In IJCNN, 2013.
[23] B. H. Menze, B. M. Kelm, D. N. Splitthoff, U. Koethe, and F. A. Hamprecht. On oblique random forests.
In Machine Learning and Knowledge Discovery in Databases, 2011.
[24] S. K. Murthy, S. Kasif, and S. Salzberg. A system for induction of oblique decision trees. Journal of
Artificial Intelligence Research, 1994.
[25] W. Ouyang and X. Wang. Joint deep learning for pedestrian detection. In ICCV, 2013.
[26] D. Park, D. Ramanan, and C. Fowlkes. Multiresolution models for object detection. In ECCV, 2010.
[27] D. Park, C. L. Zitnick, D. Ramanan, and P. Doll?ar. Exploring weak stabilization for motion feature
extraction. In CVPR, 2013.
[28] C. E. Rasmussen and C. K. I. Williams. Gaussian Processes for Machine Learning. The MIT Press, 2006.
[29] W. Ray and R. Driver. Further decomposition of the karhunen-lo`eve series representation of a stationary
random process. IEEE Transactions on Information Theory, 16(6):663?668, Nov 1970.
[30] J. J. Rodriguez, L. I. Kuncheva, and C. J. Alonso. Rotation forest: A new classifier ensemble method.
PAMI, 28(10), 2006.
[31] R. E. Schapire, Y. Freund, P. Bartlett, and W. S. Lee. Boosting the margin: A new explanation for the
effectiveness of voting methods. The Annals of Statistics, 1998.
[32] P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. LeCun. Overfeat: Integrated recognition,
localization and detection using convolutional networks. arXiv:1312.6229, 2013.
[33] P. Sermanet, K. Kavukcuoglu, S. Chintala, and Y. LeCun. Pedestrian detection with unsupervised multistage feature learning. In CVPR, 2013.
[34] C. Shen, P. Wang, S. Paisitkriangkrai, and A. van den Hengel. Training effective node classifiers for
cascade classification. IJCV, 103(3):326?347, July 2013.
[35] P. A. Viola and M. J. Jones. Robust real-time face detection. IJCV, 57(2):137?154, 2004.
[36] J. Yan, X. Zhang, Z. Lei, S. Liao, and S. Z. Li. Robust multi-resolution pedestrian detection in traffic
scenes. In CVPR, 2013.
[37] X. Zeng, W. Ouyang, and X. Wang. Multi-stage contextual deep learning for ped. det. In ICCV, 2013.
9
| 5419 |@word version:3 briefly:2 dalal:1 triggs:1 covariance:15 decomposition:1 decorrelate:2 dramatic:2 reduction:5 necessity:1 initial:1 series:3 score:1 renewed:1 interestingly:1 past:2 outperforms:3 existing:1 current:2 com:2 contextual:1 yet:1 scatter:2 must:1 dct:5 subsequent:1 additive:1 wx:2 remove:2 update:1 stationary:3 intelligence:1 fewer:1 prohibitive:1 paisitkriangkrai:1 oblique:37 core:2 recompute:1 boosting:15 node:2 completeness:1 toronto:1 brandt:1 zhang:2 along:1 become:1 driver:1 viable:1 ijcv:2 combine:3 ray:1 autocorrelation:5 introduce:3 x0:2 indeed:1 expected:2 themselves:1 frequently:2 multi:4 inspired:4 globally:1 tenfold:3 window:9 kwk0:1 increasing:5 begin:1 estimating:2 matched:1 underlying:2 suffice:1 moreover:1 advent:1 lowest:1 evolved:2 interpreted:1 ouyang:2 eigenvector:1 transformation:1 bootstrapping:1 every:14 voting:1 classifier:9 brute:2 control:1 ramanan:4 positive:12 before:4 negligible:1 local:21 modify:2 treat:1 marin:1 solely:1 pami:6 inria:10 luv:4 suggests:1 challenging:1 ease:1 limited:1 averaged:2 practical:1 unique:1 lecun:2 testing:3 crosstalk:2 practice:4 block:2 foundational:1 yan:1 online3:1 cascade:6 significantly:1 projection:7 persistence:1 pre:1 word:1 emantic:1 fetching:1 interior:1 hee:1 prentice:1 context:4 applying:4 www:1 map:1 demonstrated:2 maximizing:1 straightforward:4 layout:1 latest:1 independently:1 williams:1 resolution:3 shen:1 simplicity:1 splitting:1 m2:4 rule:1 importantly:1 nam:2 variation:2 latsvm:1 updated:3 annals:2 hierarchy:1 today:1 heavily:1 massive:1 us:4 trick:1 element:2 recognition:3 expensive:2 utilized:1 amores:1 geman:2 database:1 bottom:2 negligibly:1 wang:3 capture:1 thousand:1 substantial:2 intuition:2 principled:1 complexity:4 schiele:1 ideally:2 multistage:1 trained:9 depend:2 reviewing:1 incur:1 technically:1 localization:1 efficiency:5 learner:2 basis:2 easily:1 joint:1 kasif:1 represented:2 various:1 train:3 fast:7 effective:9 artificial:1 aggregate:1 neighborhood:7 extraordinarily:2 hillel:1 quite:2 whose:1 widely:1 larger:1 denser:1 richer:1 cvpr:10 reconstruct:1 ability:1 statistic:5 sketchtokens:2 tuytelaars:1 reshaped:1 transform:7 final:2 online:1 advantage:3 eigenvalue:2 net:2 koethe:1 propose:7 interaction:1 frequent:1 tu:1 aligned:3 rapidly:1 reproducibility:1 poorly:2 achieve:3 deformable:1 representational:1 degenerate:1 multiresolution:1 intuitive:1 description:1 curvilinear:1 exploiting:1 darrell:1 object:14 help:1 coupling:1 bourdev:1 ac:1 augmenting:1 progress:1 kuncheva:1 eq:1 strong:1 resemble:2 come:2 indicate:1 direction:1 merged:1 closely:1 filter:36 stochastic:1 stabilization:1 human:1 enable:1 mcallester:1 require:2 crux:1 fix:1 generalization:6 opt:1 extension:2 exploring:1 hold:2 hall:1 exp:1 presumably:1 vary:2 achieves:2 smallest:2 released:1 estimation:3 outperformed:1 largest:3 create:1 tool:1 brought:1 mit:1 gaussian:5 aim:1 modified:2 rather:2 reinsel:1 boosted:15 varying:4 breiman:2 conjunction:2 np2:1 derived:1 focus:2 klt:2 release:1 improvement:7 nd2:1 zca:3 baseline:10 rigid:3 downsample:2 typically:4 entire:2 integrated:1 perona:3 quasi:1 transformed:2 pixel:11 overall:3 classification:5 orientation:3 denoted:2 augment:1 retaining:1 overfeat:1 art:9 special:1 fairly:4 spatial:2 smoothing:1 once:1 extraction:1 piotr:1 sampling:5 identical:3 represents:1 park:2 jones:2 unsupervised:1 nearly:6 commercially:1 future:1 others:1 report:5 simplify:1 np:1 employ:1 few:2 modern:1 oriented:5 simultaneously:2 densely:1 individual:2 mut:1 variously:1 replaced:1 translationally:1 occlusion:1 microsoft:2 friedman:2 detection:36 stationarity:1 interest:1 highly:4 evaluation:3 predominant:1 hamprecht:1 accurate:1 edge:2 integral:1 necessary:2 korea:1 orthogonal:36 modest:1 tree:72 timed:1 overcomplete:4 girshick:2 minimal:1 instance:1 column:1 soft:2 modeling:1 salzberg:1 ar:7 restoration:1 cost:3 republic:1 entry:1 rare:2 krizhevsky:1 successful:2 reported:2 considerably:2 combined:1 lee:1 squared:1 thesis:1 opposed:1 silberstein:1 worse:1 expert:1 derivative:4 li:1 sinusoidal:2 sdt:2 lookup:3 star:1 includes:1 pedestrian:27 inc:1 performed:2 view:1 doing:2 traffic:2 contribution:1 minimize:1 hariharan:5 accuracy:8 wiener:1 convolutional:3 largely:1 likewise:5 ensemble:1 efficiently:3 correspond:2 yes:4 conceptually:2 generalize:2 weak:3 directional:1 bayesian:1 kavukcuoglu:1 fern:1 researcher:1 murthy:1 detector:29 strongest:2 opez:1 decorrelated:26 frequency:2 pdollar:2 chintala:1 gain:5 sampled:1 dataset:14 popular:3 ped:3 recall:2 color:4 lim:1 improves:5 dimensionality:3 organized:1 segmentation:1 knowledge:1 sophisticated:1 actually:1 higher:4 methodology:1 improved:4 bmvc:1 decorrelating:5 box:1 strongly:1 mar:1 furthermore:1 implicit:1 stage:1 correlation:4 sketch:1 replacing:1 zeng:1 nonlinear:1 multiscale:1 overlapping:2 rodriguez:1 autocorrelations:1 continuity:1 logistic:1 lda:27 lei:1 omitting:1 effect:3 normalized:1 regularization:2 inspiration:1 hence:1 spatially:5 postech:3 round:1 adjacent:1 during:1 inferior:1 maintained:1 cosine:1 complete:1 demonstrate:3 performs:1 hungry:1 motion:5 image:26 recently:2 common:1 rotation:2 mt:3 anisotropic:1 extend:1 belong:1 interpret:1 significant:2 gibbs:1 tuning:1 dbn:1 inclusion:1 han:1 impressive:1 operating:1 whitening:7 realboost:3 longer:2 base:3 multivariate:1 recent:8 showed:2 menze:2 success:1 discussing:1 exploited:1 caltech:15 preserving:1 additional:7 preceding:1 mr:21 employed:2 aggregated:2 wojek:1 july:1 sliding:3 multiple:8 full:1 reduces:4 match:2 faster:3 long:1 reshaping:1 equally:1 impact:3 variant:4 regression:1 whitened:6 vision:2 essentially:1 liao:1 arxiv:1 histogram:2 iteration:5 pyramid:2 background:1 addition:1 whereas:1 interval:3 source:3 benenson:4 extra:1 eliminates:1 operate:1 rest:1 ineffective:2 eigenfunctions:1 dpm:3 effectiveness:2 eve:1 split:37 fft:1 fit:1 gave:1 architecture:2 topology:6 hastie:2 reduce:3 idea:2 simplifies:1 translates:1 grad:2 det:1 bottleneck:2 pca:5 bartlett:1 effort:1 forecasting:1 cause:1 appel:2 deep:5 clear:3 eigenvectors:6 detailed:1 amount:2 transforms:1 mid:2 locally:15 svms:1 reduced:2 http:2 schapire:1 outperform:2 sign:2 estimated:2 per:18 tibshirani:2 serving:1 diverse:1 blue:1 discrete:1 levi:1 salient:1 terminology:1 nevertheless:2 threshold:2 achieving:1 khinchin:1 clarity:1 utilize:4 relaxation:1 fraction:1 year:2 convert:1 powerful:3 fourth:1 master:1 place:3 family:3 reasonable:1 timofte:3 patch:13 doc:1 decision:25 scaling:3 capturing:1 bound:1 layer:1 followed:2 distinguish:1 franken:3 ijcnn:1 x2:1 scene:1 nearby:2 speed:1 performing:1 speedup:2 department:1 alternate:1 combination:1 poor:1 precompute:1 kd:1 remain:3 across:6 slightly:2 smaller:1 increasingly:1 modification:1 fivefold:1 den:1 invariant:3 iccv:4 taken:1 computationally:2 visualization:3 remains:1 describing:1 precomputed:2 count:2 available:5 operation:1 jenkins:1 doll:7 apply:3 observe:6 leibe:1 v2:1 appropriate:1 online2:1 fowlkes:1 alternative:1 altogether:1 eigen:1 original:9 top:8 remaining:1 include:2 clustering:1 exploit:2 especially:2 tensor:1 seeking:1 malik:2 quantity:1 strategy:2 diagonal:1 gradient:9 convnet:1 separate:2 capacity:8 landmark:1 w0:3 alonso:1 discriminant:1 trivial:2 induction:1 code:8 modeled:2 downsampling:1 sermanet:2 potentially:2 hog:4 expense:3 negative:2 implementation:1 allowing:1 observation:4 convolution:2 datasets:2 markov:2 benchmark:3 viola:2 hinton:1 kelm:1 frame:3 ucsd:1 classifier1:1 intensity:1 inverting:1 toolbox:1 connection:2 learned:3 hour:1 boost:1 eigenfunction:1 bar:1 oft:1 summarize:1 including:3 memory:1 explanation:2 video:2 gool:4 power:1 critical:2 decorrelation:30 natural:12 force:2 regularized:1 rely:1 scheme:1 altered:1 improve:3 numerous:2 mathieu:1 axis:1 coupled:3 naive:2 text:1 prior:1 review:2 l2:1 discovery:1 relative:3 freund:1 discriminatively:1 interesting:1 filtering:3 ingredient:1 validation:5 eigendecomposition:1 sufficient:1 consistent:1 undergone:1 thresholding:1 principle:1 bank:1 tiny:1 lo:1 row:4 eccv:3 summary:2 token:1 repeat:1 surprisingly:2 keeping:1 last:1 infeasible:1 rasmussen:1 allow:4 deeper:4 face:2 taking:1 felzenszwalb:1 sparse:2 benefit:1 van:5 boundary:4 depth:13 dimension:2 evaluating:1 rich:1 computes:1 contour:1 hengel:1 made:4 collection:2 far:1 transaction:1 approximate:1 nov:1 implicitly:1 confirm:1 global:3 unnecessary:2 conclude:1 belongie:2 discriminative:3 fergus:1 alternatively:1 postdoctoral:1 decade:1 table:6 channel:61 robust:3 correlated:11 obtaining:1 forest:7 improving:1 complex:2 zitnick:2 vj:2 official:1 did:1 dense:2 main:1 joon:1 motivation:1 augmented:2 fig:6 acf:42 exponential:1 candidate:1 tied:1 donahue:1 down:1 minute:2 theorem:1 enlarging:2 showing:1 offset:2 experimented:1 explored:1 decay:1 naively:1 false:9 effectively:1 kr:1 magnitude:2 illustrates:1 karhunen:1 margin:2 rejection:1 suited:3 locality:1 mf:1 explore:1 springer:1 corresponds:1 oct:1 conditional:1 goal:1 identity:2 twofold:2 shared:8 considerable:7 change:1 specifically:2 typical:1 operates:1 reducing:1 miss:6 degradation:1 kienzle:1 total:1 mathias:4 intact:1 azquez:1 evaluate:1 avoiding:1 handling:1 |
4,882 | 542 | Learning Global Direct Inverse Kinematics
Kenneth Kreutz-Delgado t
Electrical & Computer Eng.
UC San Diego
La Jolla, CA 92093-0407
David DeMers?
Computer Science & Eng.
UC San Diego
La Jolla, CA 92093-0114
Abstract
We introduce and demonstrate a bootstrap method for construction of an inverse function for the robot kinematic mapping using only sample configurationspace/workspace data. Unsupervised learning (clustering) techniques are used on
pre-image neighborhoods in order to learn to partition the configuration space
into subsets over which the kinematic mapping is invertible. Supervised leaming is then used separately on each of the partitions to approximate the inverse
function. The ill-posed inverse kinematics function is thereby regularized, and
a globa1 inverse kinematics solution for the wristless Puma manipulator is developed.
1 INTRODUCTION
The robot forward kinematics function is a continuous mapping
f : C ~ en -
w ~ Xm
which maps a set of n joint parameters from the configuration space, C, to the mdimensiona1 task space, W. If m S n, the robot has redundant degrees-of-freedom
(dof's). In general, control objectives such as the positioning and orienting of the endeffector are specified with respect to task space co-ordinates; however, the manipulator is
typica1ly controlled only in the configuration space. Therefore, it is important to be able
to find some 0 E C such that f(O) is a particular target va1ue
E W. This is the inverse
kinematics problem.
xo
? e-mail: [email protected]
t e-mail: [email protected]
589
590
DeMers and Kreutz-Delgado
The inverse kinematics problem is ill-posed. If there are redundant doCs then the problem is
locally ill-posed, because the solution is non-unique and consists of a non-trivial manifold 1
in C. With or without redundant dof's, the problem is generally globally ill-posed because
of the existence of a finite set of solution branches - there will typically be multiple
configurations which result in the same task space location. Thus computation of a direct
inverse is problematic due to the many-to--one nature (and therefore non-invertibility) of
the map I .
The inverse problem can be solved explicitly, that is, in closed form, for only certain kinds
of manipulators. E.g. six dof elbow manipulators with separable wrist (where the first three
joints are used for positioning and the last three have a common origin and are used for
orientation), such as the Puma 560, are solvable, see (Craig, 86). The alternative to a closed
form solution is a numerical solution, usually either using the inverse of the Jacobian, which
is a Newton-style approach, or by using gradient descent (also a Jacobian-based method).
These methods are iterative and require expensive Jacobian or gradient computation at each
step, thus they are not well-suited for real-time control.
Neural networks can be used to find an inverse by implementing either direct inverse
modeling (estimating the explicit function 1-1) or differential methods. Implementations
of the direct inverse approach typically fail due to the non-linearity of the solution sef,
or resolve this problem by restriction to a single solution a priori. However, such a prior
restriction of the solutions may not be possible or acceptable in all circumstances, and may
drastically reduce the dexterity and manipulability of the arm.
The differential approaches either find only the nearest local solution, or resolve the multiplicity of solutions at training time, as with Jordan's forward modeling (Jordan & Rumelhart,
1990) or the approach of (Nguyen & Patel, 1990). We seek to regularize the mapping in
such a way that all possible solutions are available at run-time, and can be computed
efficiently as a direct constant-time inverse rather than approximated by slower iterative
differential methods. To achieve the fast run-time solution, a significant cost in training
time must be paid; however, it is not unreasonable to invest resources in off-line learning
in order to attain on-line advantages. Thus we wish to gain the run-time computational
efficiency of a direct inverse solution while also achieving the benefits of the differential
approaches.
This paper introduces a method for performing global regularization; that is, identifying
the complete, finite set of solutions to the inverse kinematics problem for a non-redundant
manipulator. This will provide the ability to choose a particular solution at run time.
Resolving redundancy is beyond the scope of this paper, however, preliminary work on
a method which may be integrated with the work presented here is shown in (DeMers
& Kreutz-Delgado, 1991). In the remainder of this paper it will be assumed that the
manipulator does not have redundant dof's. It will also be assumed that all of the joints are
revolute, thus the configuration space is a subset of the n-torus, Tn.
IGenerically of dimensionality equal to n - m.
Zrhe target values are assumed to be in the range of I, i E W = I(C), so the existence of a
solution is not an issue in this paper.
3Training a network to minimize mean squared error with multiple target values for the same
input value results in a "learned" response of the average of the targets. Since the targets lie on a
number of non-linear manifolds (for the redundant case) or consist of a finite number of points (for
the non-redundant case), the average of multiple targets will typically not be a correct target.
Learning Global Direct Inverse Kinematics
2 TOPOLOGY OF THE KINEMATICS FUNCTION
The kinematics mapping is continuous and smooth and, generically, neighborhoods in
configuration space map to neighborhoods in the task space4 ? The configuration space,
C, is made up of a finite number of disjoint regions or partitions, separated by n - 1
dimensional surfaces where the Jacobian loses rank (called critical surfaces), see (Burdick,
1988, Burdick, 1991).
Let
I : Tn -- Rn
be the kinematic mapping. Then
k
W
= I(C) = UIi (Cd
i=l
=
where Ii is the restriction of I to Ci , Ii : Ci
en /1 -- Rn and the factor space
locally diffeomorphic to Rn. The Ci are each a connected region such that
VOECi ,
en /1 is
det(J(O));tO
where J is the Jacobian of I, J == d(J!. Define Wi as I(Cd. Generically, Ii is one-to-one
and onto open neighborhoods of Wi 5 , thus by the inverse function theorem
:3 gi(X)
=I
j-
1 :
Wi -- Ci, such that I
0
gi(X)
= X,
Vx E Wi
In the general case, with redundant dof's, the kinematics over a single configuration-space
region can be viewed as a fiber bundle, where the fibers are homeomorphic to Tn-m.
The base space is the reachable workspace (the image of Ci under j). Solution branch
resolution can be done by identifying distinct connected open coordinate neighborhoods of
the configuration space which cover the workspace. Redundancy resolution can be done by
a consistent parameterization of the fibers within each neighborhood. In the case at hand,
without redundant dof's, the "fibers" are singleton sets and no resolution is needed.
In the remainder of this paper, we will use input/output data to identify the individual
regions, Ci, of a non-redundant manipulator, over which the mapping Ii : Ci -- Wi is
invertible. The input/output data will then be partitioned modulo the configuration regions
Ci, and each li- 1 approximated individually.
3 SAMPLING APPROACH
If the manipulator can be measured and a large sample of (0, i) pairs taken, stored such
that the samples can be searched efficiently, a rough estimate of the inverse solutions at
a particular target point io may be obtained by finding all of the 0 points whose image lies
within some ( of
The pre-image of this (-ball will generically consist of several distinct
(distorted) balls in the configuration space. If the sampling is adequate then there will be
one such ball for each of the inverse solution branches. If each of the the points in each ball
is given a label for the solution branch, the labeled data may then be used for supervised
x
xo.
4This property fails when the manipulator is in a singular configuration, at which the Jacobian,
deft loses rank.
sSince it is generically true that J is non-singular.
591
592
DeMers and Kreutz-Delgado
learning of a classifier of solution branches in the configuration space. In this way we will
have "bootstrapped" our way to the development of a solution branch classifier.
Taking advantage of the continuous nature of the forward mapping, note that if io is slightly
perturbed by a "jump" to a neighboring target point then the pre-image balls will also be
perturbed. We can assign labels to the new data consistent with labels already assigned
to the previous data, by computing the distances between the new, unlabeled balls and the
previously labeled balls. Continuing in this fashion, io traces a path through the entire
workspace and solution branch labels may be given to all points in C which map to within
f of one of the selected i points along the sweep.
This procedure results in a significant and representative proportion of the data now being
labeled as to solution branch. Thus we now have labeled data (if, i, B(
where
8(
{I, ... , k} indicates which of the k solution branches, Ci , the point is in. We
can now construct a classifier using supervised learning to compute the branches B( e) for
a given B. Once an estimate of B( 0) is developed, we may use it to classify large amounts
of (if, i) data, and partition the data into k sets, one for each of the solution branches, Ci.
en,
e
e) =
4
RESOLUTION OF SOLUTION BRANCHES
We applied the above to the wristIess Puma 560, a 3-R manipulator for end-effector
positioning in R3. We took 40,000 samples of (if, i) points, and examined all points within
lOcm of selected target values ii. The ii formed a grid of 90 locations in the workspace.
3,062 of the samples fell within 10 cm of one of the ii. The configuration space points for
each target ii were clustered into four groups, corresponding to the four possible solution
branches of the wristless Puma 560. About 3% of the points were clustered into the wrong
group, based on the labeling scheme used. These 3,062 points were then used as training
patterns for a feedforward neural network classifier. A point was classified into the group
associated with the output unit of the neural network with maximum activation. The output
values were normalized to sum to 1.0. The network was tested on 50,000 new, previously
unseen (if, i) pairs, and correctly classified more than 98% of them.
All of the erroneous classifications were for points near the critical surfaces. Therefore the
activation levels of the output units can be used to estimate closeness to a critical surface.
Examining the test data and assigning all 0 points for which no output unit has activation
greater than or equal to 0.8 to the "near-a-singularity" class, the remaining points were
100% correctly classified.
Figure 1 shows the true critical manifold separating the regions of configuration space, and
the estimated manifold consisting of points from the test set where the maximum activation
of output units of the trained neural network is less than 0.8. The configuration space is a
subset of the 3-torus, which is shown here "sliced" along three generators and represented
as a cube. Because the Puma 560 has physiCal limits on the range of motion of its joints, the
regions shown are in fact six distinct regions, and there is no wraparound in any direction.
This classifier network is our candidate for an estimate of B( e). With it, the samples can
be separated into groups corresponding to the domains of each of the Ii, thus regularizing
into k 6 one-to-one invertible pieces6 ?
=
6 Although
there are only four inverse solutions for any i. If there were no joint limits, then the
Learning Global Direcr Inverse Kinemarics
Joint 2
Joint 2
Figure 1: The analytically derived critical surfaces, along with J ,000 points/or which no
unit 0/ the neural network classifier has greater than 0.8 activation.
5 DIRECT INVERSE SOLUTIONS
The classifier neural network can now be used to partition the data into four groups, one for
each of the branches, Ci . For each of these data sets, we train a feedforward network to learn
the mapping in the inverse direction. The target vectors were represented as vectors of the
sine of the half-angle (a measure motivated by the quatemion representation of orientation).
MSE under 0.001 were achieved for each of the four. This looks like a very small error,
however, this error is somewhat misleading. The configuration space error is measured in
units which are difficult to interpret. More important is the error in the workspace when
the solution computed is used in the forward kinematics mapping to position the ann. Over
a test set of 4,000 points, the average positioning error was 5.2 cm over the 92 cm radius
workspace.
We have as yet made no attempts to optimize the network or training for the direct inverse;
the thrust of our work is in achieving the regularization. It is clear that substantially better
performance can be developed, for example, by following (Ritter, et al., 1989), and we
expect end-effector positioning errors of less than 1% to be easily achievable.
6
DISCUSSION
We have shown that by exploiting the topological property of continuity of the kinematic
mapping for a non-redundant 3-dof robot we can determine all of the solution regions of the
inverse kinematic mapping. We have mapped out the configuration space critical surfaces
and thus discovered an important topological property of the mapping, corresponding to
an important physical property of the manipulator, by unsupervised learning. We can
boostrap from the original input/output data, unlabeled as to solution branch, and construct
an accurate classifier for the entire configuration space. The data can thereby be partitioned
into sets which are individually one-t()-{)ne and invenible, and the inverse mapping can
be directly approximated for each. Thus a large learning-time investment results in a fast
run-time direct inverse kinematics solution.
cube shown would be a true 3-torus, with opposite faces identified. Thus the small pieces in the
corners would be part of the larger regions by wraparound in the Joint 2 direction.
593
594
DeMers and Kreutz-Delgado
We need a thorough sampling of the configuration space in order to ensure that enough
points will fall within each f-ball, thus the data requirements are clearly exponential in the
number of degrees of freedom of the manipulator. Even with efficient storage and retrieval
in geometric data structures, such as a k-d tree, high dimensional systems may not be
tractable by our methods.
Fortunately practical and useful robotic systems of six and seven degrees of freedom
should be amenable to this method, especially if separable into positioning and orienting
subsystems.
Acknowledgements
This work was supported in part by NSF Presidential Young Investigator award IRI9057631 and a NASA/Netrologic grant. The first author would like to thank NIPS for
providing student travel grants. We thank Gary Cottrell for his many helpful comments and
enthusiastic discussions.
References
Joel Burdick (1991), "A Classification of 3R Regional Manipulator Singularities and Geometries", Proc. 19911?E? Inti. Con! Robotics & Automation, Sacramento.
Joel Burdick (1988), "Kinematics and Design of Redundant Robot Manipulators", Stanford
Ph.D. Thesis, Dept. of Mechanical Engineering.
John Craig (1986), Introduction to Robotics, Addison-Wesley.
David DeMers & Kenneth Kreutz-Delgado (1991), "Learning Global Topological Properties of Robot Kinematic Mappings for Neural Network-Based Configuration Control", in
Bekey, ed. Proc. USC Workshop on Neural Networks in Robotics, (to appear).
Michael I. Jordan (1988), "Supervised Learning and Systems with Excess Degrees of
Freedom", COINS Technical Report 88-27, University of Massachusetts at Amherst.
Michael!. Jordan & David E. Rumelhart (1990), "Forward Models: Supervised Learning
with a Distal Teacher". Submitted to Cognitive Science.
L. Nguyen & R.V. Patel (1990), "A Neural Network Based Strategy for the Inverse Kinematics Problem in Robotics", in Jamshidi and Saif, eds., Robotics and Manufacturing:
recent Trends in Research, Education and Applications, vol. 3, pp. 995-1000 (ASME
Press).
Helge J. Ritter, Thomas M. Martinetz, & Klaus J. Schulten (1989), ''Topology-Conserving
Maps for Learning Visuo-Motor-Coordination", Neural Networks, Vol. 2, pp. 159-168.
| 542 |@word achievable:1 proportion:1 open:2 seek:1 eng:2 paid:1 thereby:2 delgado:6 configuration:21 bootstrapped:1 activation:5 assigning:1 yet:1 must:1 john:1 cottrell:1 numerical:1 partition:5 thrust:1 burdick:4 motor:1 half:1 selected:2 parameterization:1 location:2 along:3 direct:10 differential:4 consists:1 manipulability:1 introduce:1 enthusiastic:1 globally:1 resolve:2 elbow:1 estimating:1 linearity:1 kind:1 cm:3 substantially:1 developed:3 finding:1 thorough:1 classifier:8 wrong:1 control:3 unit:6 grant:2 appear:1 engineering:1 local:1 limit:2 io:3 path:1 examined:1 co:1 range:2 unique:1 practical:1 wrist:1 investment:1 bootstrap:1 procedure:1 attain:1 puma:5 pre:3 onto:1 unlabeled:2 subsystem:1 storage:1 restriction:3 optimize:1 map:5 resolution:4 identifying:2 sacramento:1 regularize:1 his:1 deft:1 coordinate:1 diego:2 construction:1 target:12 modulo:1 origin:1 trend:1 rumelhart:2 expensive:1 approximated:3 labeled:4 electrical:1 solved:1 region:10 connected:2 sef:1 trained:1 efficiency:1 easily:1 joint:8 represented:2 fiber:4 train:1 separated:2 distinct:3 fast:2 labeling:1 klaus:1 neighborhood:6 dof:7 whose:1 posed:4 larger:1 stanford:1 tested:1 presidential:1 ability:1 gi:2 unseen:1 advantage:2 took:1 remainder:2 neighboring:1 achieve:1 conserving:1 jamshidi:1 invest:1 exploiting:1 requirement:1 measured:2 nearest:1 c:1 direction:3 radius:1 correct:1 vx:1 implementing:1 education:1 require:1 assign:1 clustered:2 preliminary:1 singularity:2 mapping:15 scope:1 proc:2 travel:1 label:4 coordination:1 individually:2 rough:1 clearly:1 rather:1 derived:1 rank:2 indicates:1 diffeomorphic:1 helpful:1 typically:3 integrated:1 entire:2 issue:1 classification:2 ill:4 orientation:2 priori:1 development:1 uc:2 cube:2 equal:2 construct:2 once:1 sampling:3 look:1 unsupervised:2 report:1 individual:1 usc:1 geometry:1 consisting:1 bekey:1 attempt:1 freedom:4 kinematic:6 joel:2 introduces:1 generically:4 bundle:1 amenable:1 accurate:1 tree:1 continuing:1 effector:2 classify:1 modeling:2 cover:1 cost:1 subset:3 examining:1 stored:1 perturbed:2 teacher:1 amherst:1 workspace:7 ritter:2 off:1 invertible:3 michael:2 squared:1 thesis:1 choose:1 corner:1 cognitive:1 style:1 li:1 singleton:1 student:1 invertibility:1 automation:1 explicitly:1 piece:1 sine:1 closed:2 minimize:1 formed:1 efficiently:2 identify:1 craig:2 classified:3 submitted:1 ed:2 pp:2 associated:1 visuo:1 con:1 demers:7 gain:1 massachusetts:1 dimensionality:1 nasa:1 wesley:1 supervised:5 response:1 done:2 hand:1 continuity:1 manipulator:14 orienting:2 normalized:1 true:3 regularization:2 assigned:1 analytically:1 distal:1 asme:1 complete:1 demonstrate:1 tn:3 motion:1 image:5 dexterity:1 common:1 physical:2 interpret:1 significant:2 grid:1 reachable:1 robot:6 surface:6 base:1 recent:1 jolla:2 certain:1 greater:2 somewhat:1 fortunately:1 determine:1 redundant:12 ii:9 branch:15 resolving:1 multiple:3 smooth:1 technical:1 positioning:6 retrieval:1 award:1 controlled:1 circumstance:1 achieved:1 robotics:5 separately:1 singular:2 regional:1 fell:1 comment:1 martinetz:1 jordan:4 near:2 feedforward:2 enough:1 topology:2 opposite:1 identified:1 reduce:1 space4:1 det:1 six:3 motivated:1 adequate:1 generally:1 useful:1 clear:1 amount:1 locally:2 ph:1 problematic:1 nsf:1 estimated:1 disjoint:1 correctly:2 vol:2 group:5 redundancy:2 four:5 achieving:2 kenneth:2 sum:1 run:5 inverse:29 angle:1 distorted:1 doc:1 acceptable:1 topological:3 uii:1 performing:1 separable:2 ball:8 slightly:1 wi:5 partitioned:2 helge:1 multiplicity:1 xo:2 inti:1 taken:1 resource:1 previously:2 kinematics:15 fail:1 r3:1 needed:1 addison:1 tractable:1 end:2 available:1 unreasonable:1 alternative:1 coin:1 slower:1 existence:2 original:1 thomas:1 clustering:1 remaining:1 ensure:1 newton:1 homeomorphic:1 especially:1 sweep:1 objective:1 already:1 strategy:1 gradient:2 distance:1 thank:2 mapped:1 separating:1 seven:1 mail:2 manifold:4 trivial:1 providing:1 difficult:1 trace:1 implementation:1 design:1 finite:4 descent:1 rn:3 ucsd:2 discovered:1 wraparound:2 ordinate:1 david:3 pair:2 mechanical:1 specified:1 learned:1 nip:1 able:1 beyond:1 usually:1 pattern:1 xm:1 critical:6 regularized:1 solvable:1 arm:1 scheme:1 misleading:1 ne:1 prior:1 geometric:1 acknowledgement:1 revolute:1 expect:1 generator:1 degree:4 consistent:2 cd:2 supported:1 last:1 drastically:1 fall:1 taking:1 face:1 benefit:1 forward:5 made:2 jump:1 san:2 author:1 nguyen:2 excess:1 approximate:1 patel:2 global:5 robotic:1 kreutz:7 assumed:3 continuous:3 iterative:2 learn:2 nature:2 ca:2 mse:1 domain:1 sliced:1 representative:1 en:4 fashion:1 fails:1 position:1 schulten:1 explicit:1 wish:1 torus:3 exponential:1 lie:2 candidate:1 jacobian:6 young:1 theorem:1 erroneous:1 closeness:1 consist:2 workshop:1 ci:11 suited:1 gary:1 loses:2 viewed:1 ann:1 leaming:1 manufacturing:1 called:1 ece:1 la:2 searched:1 investigator:1 dept:1 regularizing:1 |
4,883 | 5,420 | Do Convnets Learn Correspondence?
Jonathan Long
Ning Zhang
Trevor Darrell
University of California ? Berkeley
{jonlong, nzhang, trevor}@cs.berkeley.edu
Abstract
Convolutional neural nets (convnets) trained from massive labeled datasets [1]
have substantially improved the state-of-the-art in image classification [2] and object detection [3]. However, visual understanding requires establishing correspondence on a finer level than object category. Given their large pooling regions and
training from whole-image labels, it is not clear that convnets derive their success
from an accurate correspondence model which could be used for precise localization. In this paper, we study the effectiveness of convnet activation features for
tasks requiring correspondence. We present evidence that convnet features localize at a much finer scale than their receptive field sizes, that they can be used to
perform intraclass aligment as well as conventional hand-engineered features, and
that they outperform conventional features in keypoint prediction on objects from
PASCAL VOC 2011 [4].
1
Introduction
Recent advances in convolutional neural nets [2] dramatically improved the state-of-the-art in image
classification. Despite the magnitude of these results, many doubted [5] that the resulting features
had the spatial specificity necessary for localization; after all, whole image classification can rely
on context cues and overly large pooling regions to get the job done. For coarse localization, such
doubts were alleviated by record breaking results extending the same features to detection on PASCAL [3].
Now, the same questions loom on a finer scale. Are the modern convnets that excel at classification
and detection also able to find precise correspondences between object parts? Or do large receptive
fields mean that correspondence is effectively pooled away, making this a task better suited for
hand-engineered features?
In this paper, we provide evidence that convnet features perform at least as well as conventional
ones, even in the regime of point-to-point correspondence, and we show considerable performance
improvement in certain settings, including category-level keypoint prediction.
1.1
Related work
Image alignment Image alignment is a key step in many computer vision tasks, including face
verification, motion analysis, stereo matching, and object recognition. Alignment results in correspondence across different images by removing intraclass variability and canonicalizing pose.
Alignment methods exist on a supervision spectrum from requiring manually labeled fiducial points
or landmarks, to requiring class labels, to fully unsupervised joint alignment and clustering models.
Congealing [6] is an unsupervised joint alignment method based on an entropy objective. Deep
congealing [7] builds on this idea by replacing hand-engineered features with unsupervised feature
learning from multiple resolutions. Inspired by optical flow, SIFT flow [8] matches densely sampled
SIFT features for correspondence and has been applied to motion prediction and motion transfer. In
Section 3, we apply SIFT flow using deep features for aligning different instances of the same class.
1
Keypoint localization Semantic parts carry important information for object recognition, object
detection, and pose estimation. In particular, fine-grained categorization, the subject of many recent
works, depends strongly on part localization [9, 10]. Large pose and appearance variation across
examples make part localization for generic object categories a challenging task.
Most of the existing works on part localization or keypoint prediction focus on either facial landmark
localization [11] or human pose estimation. Human pose estimation has been approached using tree
structured methods to model the spatial relationships between parts [12, 13, 14], and also using
poselets [15] as an intermediate step to localize human keypoints [16, 17]. Tree structured models
and poselets may struggle when applied to generic objects with large articulated deformations and
wide shape variance.
Deep learning Convolutional neural networks have gained much recent attention due to their success in image classification [2]. Convnets trained with backpropagation were initially succesful in
digit recognition [18] and OCR [19]. The feature representations learned from large data sets have
been found to generalize well to other image classification tasks [20] and even to object detection
[3, 21]. Recently, Toshev et al. [22] trained a cascade of regression-based convnets for human pose
estimation and Jain et al. [23] combine a weak spatial model with deep learning methods.
The latter work trains multiple small, independent convnets on 64 ? 64 patches for binary bodypart detection. In contrast, we employ a powerful pretained ImageNet model that shares mid-elvel
feature representations among all parts in Section 5.
Several recent works have attempted to analyze and explain this overwhelming success. Zeiler and
Fergus [24] provide several heuristic visualizations suggesting coarse localization ability. Szegedy
et al. [25] show counterintuitive properties of the convnet representation, and suggest that individual
feature channels may not be more semantically meaningful than other bases in feature space. A
concurrent work [26] compares convnet features with SIFT in a standard descriptor matching task.
This work illuminates and extends that comparison by providing visual analysis and by moving
beyond single instance matching to intraclass correspondence and keypoint prediction.
1.2
Preliminaries
We perform experiments using a network architecture almost identical1 to that popularized by
Krizhevsky et al. [2] and trained for classification using the 1.2 million images of the ILSVRC
2012 challenge dataset [1]. All experiments are implemented using caffe [27], and our network
is the publicly available caffe reference model. We use the activations of each layer as features,
referred to as convn, pooln, or fcn for the nth convolutional, pooling, or fully connected layer,
respectively. We will use the term receptive field, abbreviated rf, to refer to the set of input pixels
that are path-connected to a particular unit in the convnet.
2
Feature visualization
In this section and Figures 1 and 2, we provide a
novel visual investigation of the effective pooling regions of convnet features.
Table 1: Convnet receptive field sizes and strides,
for an input of size 227 ? 227.
In Figure 1, we perform a nonparametric reconlayer
rf size
stride
struction of images from features in the spirit
conv1 11 ? 11
4?4
of HOGgles [28]. Rather than paired dictionary
conv2 51 ? 51
8?8
learning, however, we simply replace patches
conv3 99 ? 99
16 ? 16
with averages of their top-k nearest neighbors
conv4 131 ? 131 16 ? 16
in a convnet feature space. To do so, we first
conv5 163 ? 163 16 ? 16
compute all features at a particular layer, repool5 195 ? 195 32 ? 32
sulting in an 2d grid of feature vectors. We associate each feature vector with a patch in the
original image at the center of the corresponding receptive field and with size equal to the receptive
field stride. (Note that the strides of the receptive fields are much smaller than the receptive fields
1
Ours reverses the order of the response normalization and pooling layers.
2
conv4
conv5
uniform rf
5 neighbors
1 neighbor
5 neighbors
1 neighbor
conv3
Figure 1: Even though they have large receptive fields, convnet features carry local information at
a finer scale. Upper left: given an input image, we replaced 16 ? 16 patches with averages over
1 or 5 nearest neighbor patches, computed using convnet features centered at those patches. The
yellow square illustrates one input patch, and the black squares show the corresponding rfs for the
three layers shown. Right: Notice that the features retrieve reasonable matches for the centers of
their receptive fields, even though those rfs extend over large regions of the source image. In the
?uniform rf? column, we show the best that could be expected if convnet features discarded all
spatial information within their rfs, by choosing input patches uniformly at random from conv3sized neighborhoods. (Best viewed electronically.)
themselves, which overlap. Refer to Table 1 above for specific numbers.) We replace each such
patch with an average over k nearest neighbor patches using a database of features densely computed on the images of PASCAL VOC 2011. Our database contains at least one million patches for
every layer. Features are matched by cosine similarity.
Even though the feature rfs cover large regions of the source images, the specific resemblance of
the resulting images shows that information is not spread uniformly throughout those regions. Notable features (e.g., the tires of the bicycle and the facial features of the cat) are replaced in their
corresponding locations. Also note that replacement appears to become more semantic and less
visually specific as the layer deepens: the eyes and nose of the cat get replaced with differently colored or shaped eyes and noses, and the fur gets replaced with various animal furs, with the diversity
increasing with layer number.
Figure 2 gives a feature-centric rather than image-centric view of feature locality. For each column,
we first pick a random seed feature vector (computed from a PASCAL image), and find k nearest
neighbor features, again by cosine similarity. Instead of averaging only the centers, we average
the entire receptive fields of the neighbors. The resulting images show that similar features tend to
respond to similar colors specifically in the centers of their receptive fields.
3
conv4
conv5
500 nbrs 50 nbrs 5 nbrs
conv3
Figure 2: Similar convnet features tend to have similar receptive field centers. Starting from a
randomly selected seed patch occupying one rf in conv3, 4, or 5, we find the nearest k neighbor
features computed on a database of natural images, and average together the corresponding receptive
fields. The contrast of each image has been expanded after averaging. (Note that since each layer
is computed with a stride of 16, there is an upper bound on the quality of alignment that can be
witnessed here.)
3
Intraclass alignment
We conjecture that category learning implicitly aligns instances by pooling over a discriminative
mid-level representation. If this is true, then such features should be useful for post-hoc alignment
in a similar fashion to conventional features. To test this, we use convnet features for the task of
aligning different instances of the same class. We approach this difficult task in the style of SIFT
flow [8]: we retrieve near neighbors using a coarse similarity measure, and then compute dense
correspondences on which we impose an MRF smoothness prior which finally allows all images to
be warped into alignment.
Nearest neighbors are computed using fc7 features. Since we are specifically testing the quality of
alignment, we use the same nearest neighbors for convnet or conventional features, and we compute
both types of features at the same locations, the grid of convnet rf centers in the response to a single
image.
Alignment is determined by solving an MRF formulated on this grid of feature locations. Let p be a
point on this grid, let fs (p) be the feature vector of the source image at that point, and let ft (p) be the
feature vector of the target image at that point. For each feature grid location p of the source image,
there is a vector w(p) giving the displacement of the corresponding feature in the target image. We
use the energy function
X
X
kfs (p) ? ft (p + w(p))k2 + ?
kw(p) ? w(q)k22 ,
E(w) =
p
(p,q)?E
where E are the edges of a 4-neighborhood graph and ? is the regularization parameter. Optimization is performed using belief propagation, with the techniques suggested in [29]. Message passing
is performed efficiently using the squared Euclidean distance transform [30]. (Unlike the L1 regularization originally used by SIFT flow [8], this formulation maintains rotational invariance of w.)
Based on its performance in the next section, we use conv4 as our convnet feature, and SIFT with
descriptor radius 20 as our conventional feature. From validation experiments, we set ? = 3 ? 10?3
for both conv4 and SIFT features (which have a similar scale).
Given the alignment field w, we warp target to source using bivariate spline interpolation (implemented in SciPy [31]). Figure 3 gives examples of alignment quality for a few different seed images,
using both SIFT and convnet features. We show five warped nearest neighbors as well as keypoints
transferred from those neighbors.
We quantitatively assess the alignment by measuring the accuracy of predicted keypoints. To obtain
good predictions, we warp 25 nearest neighbors for each target image, and order them from smallest
to greatest deformation energy (we found this method to outperform ordering using the data term).
We take the predicted keypoints to be the median points (coordinate-wise) of the top five aligned
keypoints according to this ordering.
We assess correctness using mean PCK [32]. We consider a ground truth keypoint to be correctly
predicted if the prediction lies within a Euclidean distance of ? times the maximum of the bounding
4
five nearest neighbors
SIFT flow
conv4 flow
SIFT flow
conv4 flow
target image
Figure 3: Convnet features can bring different instances of the same class into good alignment at
least as well (on average) as traditional features. For each target image (left column), we show
warped versions of five nearest neighbor images aligned with conv4 flow (first row), and warped
versions aligned with SIFT flow [8] (second row). Keypoints from the warped images are shown
copied to the target image. The cat shows a case where convnet features perform better, while the
bicycle shows a case where SIFT features perform better. (Note that each instance is warped to a
square bounding box before alignment. Best viewed in color.)
Table 2: Keypoint transfer accuracy using convnet flow, SIFT flow, and simple copying from nearest
neighbors. Accuracy (PCK) is shown per category using ? = 0.1 (see text) and means are also
shown for the stricter values ? = 0.05 and 0.025. On average, convnet flow performs as well as
SIFT flow, and performs a bit better for stricter tolerances.
aero bike bird
conv4 flow 28.2 34.1 20.4
SIFT flow 27.6 30.8 19.9
NN transfer 18.3 24.8 14.5
boat
17.1
17.5
15.4
bttl
50.6
49.4
48.1
bus
36.7
36.4
27.6
car
20.9
20.7
16.0
mean
conv4 flow
SIFT flow
NN transfer
cat
19.6
16.0
11.1
chair
15.7
16.1
12.0
cow
25.4
25.0
16.8
? = 0.1
24.9
24.7
19.9
table
12.7
16.1
15.7
dog
18.7
16.3
12.7
? = 0.05
11.8
10.9
7.8
horse
25.9
27.7
20.2
mbike
23.1
28.3
18.5
prsn
21.4
20.2
18.7
plant
40.2
36.4
33.4
sheep
21.1
20.5
14.0
sofa
14.5
17.2
15.5
train
18.3
19.9
14.6
tv
33.3
32.9
30.0
mean
24.9
24.7
19.9
? = 0.025
4.08
3.55
2.35
box width and height, picking some ? ? [0, 1]. We compute the overall accuracy for each type of
keypoint, and report the average over keypoint types. We do not penalize predicted keypoints that
are not visible in the target image.
Results are given in Table 2. We show per category results using ? = 0.1, and mean results for
? = 0.1, 0.05, and 0.025. Indeed, convnet learned features are at least as capable as SIFT at
alignment, and better than might have been expected given the size of their receptive fields.
4
Keypoint classification
In this section, we specifically address the ability of convnet features to understand semantic information at the scale of parts. As an initial test, we consider the task of keypoint classification:
given an image and the coordinates of a keypoint on that image, can we train a classifier to label the
keypoint?
5
Table 3: Keypoint classification accuracies, in percent, on the twenty categories of PASCAL 2011
val, trained with SIFT or convnet features. The best SIFT and convnet scores are bolded in each
category.
aero
SIFT 10 36
(radius) 20 37
40 35
80 33
160 27
conv 1 16
(layer) 2 37
3 42
4 44
5 44
bike
42
50
54
43
36
14
43
50
53
51
bird
36
39
37
37
34
15
40
46
49
49
boat
32
35
41
42
38
19
35
41
42
41
bttl
67
74
76
75
72
20
69
76
78
77
bus
64
67
68
66
59
29
63
69
70
68
car
40
47
47
42
35
15
38
46
45
44
cat
37
40
37
30
25
22
44
52
55
53
chair
33
36
39
43
39
16
35
39
41
39
cow
37
43
40
36
30
17
40
45
48
45
table
60
68
69
70
67
29
61
64
68
63
dog
34
38
36
31
27
17
38
47
51
50
horse mbike prsn
39
38
29
42
48
33
42
49
32
36
51
27
32
46
25
14
16
15
40
44
34
48
52
40
51
53
41
49
52
39
(a)
(a) cat left eye
plant sheep sofa
63
37 42
70
44 52
69
39 52
70
35 49
70
29 48
33
18 12
65
39 41
74
46 50
76
49 52
73
47 47
train
64
68
74
69
66
27
63
71
73
71
tv
75
77
78
77
76
29
72
77
76
75
mean
45
50
51
48
44
20
47
54
56
54
(b)
Figure 5: Cross validation scores for cat
keypoint classification as a function of
the SVM parameter C. In (a), we plot
mean accuracy against C for five different convnet features; in (b) we plot
the same for SIFT features of different
sizes. We use C = 10?6 for all experiments in Table 3.
(b) cat nose
Figure 4: Convnet features show fine
localization ability, even beyond their
stride and in cases where SIFT features
do not perform as well. Each plot is
a 2D histogram of the locations of the
maximum responses of a classifer in a
21 by 21 pixel rectangle taken around a
ground truth keypoint.
For this task we use keypoint data [15] on the twenty classes of PASCAL VOC 2011 [4]. We extract
features at each keypoint using SIFT [33] and using the column of each convnet layer whose rf
center lies closest to the keypoint. (Note that the SIFT features will be more precisely placed as a
result of this approximation.) We trained one-vs-all linear SVMs on the train set using SIFT at five
different radii and each of the five convolutional layer activations as features (in general, we found
pooling and normalization layers to have lower performance). We set the SVM parameter C = 10?6
for all experiments based on five-fold cross validation on the training set (see Figure 5).
Table 3 gives the resulting accuracies on the val set. We find features from convnet layers consistently perform at least as well as and often better than SIFT at this task, with the highest performance
coming from layers conv4 and conv5. Note that we are specifically testing convnet features trained
only for classification; the same net could be expected to achieve even higher performance if trained
for this task.
Finally, we study the precise location understanding of our classifiers by computing their responses
with a single-pixel stride around ground truth keypoint locations. For two example keypoints (cat
left eye and nose), we histogram the locations of the maximum responses within a 21 pixel by 21
pixel rectangle around the keypoint, shown in Figure 4. We do not include maximum responses
that lie on the boundary of this rectangle. While the SIFT classifiers do not seem to be sensitive
to the precise locations of the keypoints, in many cases the convnet ones seem to be capable of
localization finer than their strides, not just their receptive field sizes. This observation motivates
our final experiments to consider detection-based localization performance.
6
5
Keypoint prediction
We have seen that despite their large receptive field sizes, convnets work as well as the handengineered feature SIFT for alignment and slightly better than SIFT for keypoint classification.
Keypoint prediction provides a natural follow-up test. As in Section 3, we use keypoint annotations
from PASCAL VOC 2011, and we assume a ground truth bounding box.
Inspired in part by [3, 34, 23], we train sliding window part detectors to predict keypoint locations
independently. R-CNN [3] and OverFeat [34] have both demonstrated the effectiveness of deep convolutional networks on the generic object detection task. However, neither of them have investigated
the application of CNNs for keypoint prediction.2 R-CNN starts from bottom-up region proposal
[35], which tends to overlook the signal from small parts. OverFeat, on the other hand, combines
convnets trained for classification and for regression and runs in multi-scale sliding window fashion.
We rescale each bounding box to 500 ? 500 and compute conv5 (with a stride of 16 pixels). Each
cell of conv5 contains one 256-dimensional descriptor. We concatenate conv5 descriptors from a
local region of 3 ? 3 cells, giving an overall receptive field size of 195 ? 195 and feature dimension
of 2304. For each keypoint, we train a linear SVM with hard negative mining. We consider the ten
closest features to each ground truth keypoint as positive examples, and all the features whose rfs
do not contain the keypoint as negative examples. We also train using dense SIFT descriptors for
comparison. We compute SIFT on a grid of stride eight and bin size of eight using VLFeat [36]. For
SIFT, we consider features within twice the bin size from the ground truth keypoint to be positives,
while samples that are at least four times the bin size away are negatives.
We augment our SVM detectors with a spherical Gaussian prior over candidate locations constructed
by nearest neighbor matching. The mean of each Gaussian is taken to be the location of the keypoint
in the nearest neighbor in the training set found using cosine similarity on pool5 features, and we
use a fixed standard deviation of 22 pixels. Let s(Xi ) be the output score of our local detector for
keypoint Xi , and let p(Xi ) be the prior score. We combine these to yield a final score f (Xi ) =
s(Xi )1?? p(Xi )? , where ? ? [0, 1] is a tradeoff parameter. In our experiments, we set ? = 0.1 by
cross validation. At test time, we predict the keypoint location as the highest scoring candidate over
all feature locations.
We evaluate the predicted keypoints using the measure PCK introduced in Section 3, taking ? = 0.1.
A predicted keypoint is defined as correct if the distance between it and the ground truth keypoint is
less than ? ? max(h, w) where h and w are the height and width of the bounding box. The results
using conv5 and SIFT with and without the prior are shown in Table 4. From the table, we can see
that local part detectors trained on the conv5 feature outperform SIFT by a large margin and that the
prior information is helpful in both cases. To our knowledge, these are the first keypoint prediction
results reported on this dataset. We show example results from five different categories in Figure
6. Each set consists of rescaled bounding box images with ground truth keypoint annotations and
predicted keypoints using SIFT and conv5 features, where each color corresponds to one keypoint.
As the figure shows, conv5 outperforms SIFT, often managing satisfactory outputs despite the
challenge of this task. A small offset can be noticed for some keypoints like eyes and noses, likely
due to the limited stride of our scanning windows. A final regression or finer stride could mitigate
this issue.
6
Conclusion
Through visualization, alignment, and keypoint prediction, we have studied the ability of the intermediate features implicitly learned in a state-of-the-art convnet classifier to understand specific,
local correspondence. Despite their large receptive fields and weak label training, we have found in
all cases that convnet features are at least as useful (and sometimes considerably more useful) than
conventional ones for extracting local visual information.
Acknowledgements This work was supported in part by DARPA?s MSEE and SMISC programs, by NSF
awards IIS-1427425, IIS-1212798, and IIS-1116411, and by support from Toyota.
2
But see works cited in Section 1.1 regarding keypoint localization.
7
Table 4: Keypoint prediction results on PASCAL VOC 2011. The numbers give average accuracy
of keypoint prediction using the criterion described in Section 3, PCK with ? = 0.1.
SIFT
SIFT+prior
conv5
conv5+prior
aero
17.9
33.5
38.5
50.9
Groundtruth
bike
16.5
36.9
37.6
48.8
bird
15.3
22.7
29.6
35.1
boat
15.6
23.1
25.3
32.5
bttl
25.7
44.0
54.5
66.1
SIFT+prior
bus
21.7
42.6
52.1
62.0
car
22.0
39.3
28.6
45.7
cat
12.6
22.1
31.5
34.2
chair
11.3
18.5
8.9
21.4
cow
7.6
23.5
30.5
41.1
table
6.5
11.2
24.1
27.2
dog
12.5
20.6
23.7
29.3
horse
18.3
32.2
35.8
46.8
Groundtruth
conv5+prior
mbike
15.1
33.9
29.9
45.6
prsn
15.9
26.7
39.3
47.1
plant
21.3
30.6
38.2
42.5
sheep
14.7
25.7
30.5
38.8
SIFT+prior
sofa
15.1
26.5
24.5
37.6
train
9.2
21.9
41.5
50.7
tv
19.9
32.4
42.0
45.6
mean
15.7
28.4
33.3
42.5
conv5+prior
Figure 6: Examples of keypoint prediction on five classes of the PASCAL dataset: aeroplane, cat,
cow, potted plant, and horse. Each keypoint is associated with one color. The first column is the
ground truth annotation, the second column is the prediction result of SIFT+prior and the third
column is conv5+prior. (Best viewed in color).
References
[1] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. ImageNet: A Large-Scale Hierarchical
Image Database. In CVPR, 2009.
[2] A. Krizhevsky, I. Sutskever, and G. Hinton. Imagenet classification with deep convolutional neural networks. In NIPS, 2012.
[3] R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich feature hierarchies for accurate object detection
and semantic segmentation. In CVPR, 2014.
[4] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman.
The
PASCAL Visual Object Classes Challenge 2011 (VOC2011) Results.
http://www.pascalnetwork.org/challenges/VOC/voc2011/workshop/index.html.
[5] Debate on Yann LeCun?s Google+ page. https://plus.google.com/+YannLeCunPhD/posts/JBBFfv2XgWM.
Accessed: 2014-5-31.
[6] G. B. Huang, V. Jain, and E. Learned-Miller. Unsupervised joint alignment of complex images. In ICCV,
2007.
8
[7] G. B. Huang, M. A. Mattar, H. Lee, and E. Learned-Miller. Learning to align from scratch. In NIPS,
2012.
[8] C. Liu, J. Yuen, and A. Torralba. Sift flow: Dense correspondence across scenes and its applications.
PAMI, 33(5):978?994, 2011.
[9] J. Liu and P. N. Belhumeur. Bird part localization using exemplar-based models with enforced pose and
subcategory consistenty. In ICCV, 2013.
[10] T. Berg and P. N. Belhumeur. POOF: Part-based one-vs.-one features for fine-grained categorization, face
verification, and attribute estimation. In CVPR, 2013.
[11] P. N. Belhumeur, D. W. Jacobs, D. J. Kriegman, and N. Kumar. Localizing parts of faces using a consensus
of exemplars. In CVPR, 2011.
[12] Y. Yang and D. Ramanan. Articulated pose estimation using flexible mixtures of parts. In CVPR, 2011.
[13] M. Sun and S. Savarese. Articulated part-based model for joint object detection and pose estimation. In
ICCV, 2011.
[14] X. Zhu and D. Ramanan. Face detection, pose estimation, and landmark localization in the wild. In
CVPR, 2012.
[15] L. Bourdev and J. Malik. Poselets: Body part detectors trained using 3d human pose annotations. In
ICCV, 2009.
[16] G. Gkioxari, B. Hariharan, R. Girshick, and J. Malik. Using k-poselets for detecting people and localizing
their keypoints. In CVPR, 2014.
[17] G. Gkioxari, P. Arbelaez, L. Bourdev, and J. Malik. Articulated pose estimation using discriminative
armlet classifiers. In CVPR, 2013.
[18] Y. LeCun, B. Boser, J.S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel. Backpropagation applied to hand-written zip code recognition. In Neural Computation, 1989.
[19] Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition.
In Proceedings of the IEEE, pages 2278?2324, 1998.
[20] J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng, and T. Darrell. DeCAF: A deep convolutional activation feature for generic visual recognition. In ICML, 2014.
[21] P. Sermanet, K. Kavukcuoglu, S. Chintala, and Y. LeCun. Pedestrian detection with unsupervised multistage feature learning. In CVPR, 2013.
[22] A. Toshev and C. Szegedy. DeepPose: Human pose estimation via deep neural networks. In CVPR, 2014.
[23] A. Jain, J. Tompson, M. Andriluka, G. W. Taylor, and C. Bregler. Learning human pose estimation
features with convolutional networks. In ICLR, 2014.
[24] M. D Zeiler and R. Fergus. Visualizing and understanding convolutional neural networks. In ECCV,
2014.
[25] C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I. Goodfellow, and R. Fergus. Intriguing
properties of neural networks. In ICLR, 2014.
[26] P. Fischer, A. Dosovitskiy, and T. Brox. Descriptor Matching with Convolutional Neural Networks: a
Comparison to SIFT. ArXiv e-prints, May 2014.
[27] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell. Caffe:
Convolutional architecture for fast feature embedding. arXiv preprint arXiv:1408.5093, 2014.
[28] C. Vondrick, A. Khosla, T. Malisiewicz, and A. Torralba. HOGgles: Visualizing Object Detection Features. In ICCV, 2013.
[29] P. Felzenszwalb and D. P. Huttenlocher. Efficient belief propagation for early vision. International journal
of computer vision, 70(1):41?54, 2006.
[30] P. Felzenszwalb and D. Huttenlocher. Distance transforms of sampled functions. Technical report, Cornell
University, 2004.
[31] E. Jones, T. Oliphant, P. Peterson, et al. SciPy: Open source scientific tools for Python, 2001.
[32] Y. Yang and D. Ramanan. Articulated human detection with flexible mixtures of parts. In PAMI, 2013.
[33] D.G. Lowe. Object recognition from local scale-invariant features. In ICCV, 1999.
[34] P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. LeCun. Overfeat: Integrated recognition,
localization and detection using convolutional networks. In ICLR, 2014.
[35] J. Uijlings, K. van de Sande, T. Gevers, and A. Smeulders. Selective search for object recognition. IJCV,
2013.
[36] A. Vedaldi and B. Fulkerson. VLFeat: An open and portable library of computer vision algorithms.
http://www.vlfeat.org/, 2008.
9
| 5420 |@word cnn:2 version:2 everingham:1 open:2 jacob:1 pick:1 carry:2 initial:1 liu:2 contains:2 score:5 deepens:1 ours:1 document:1 outperforms:1 existing:1 guadarrama:1 com:1 activation:4 intriguing:1 written:1 visible:1 concatenate:1 shape:1 plot:3 v:2 cue:1 selected:1 record:1 colored:1 coarse:3 provides:1 potted:1 location:14 detecting:1 org:2 zhang:3 five:10 height:2 accessed:1 constructed:1 become:1 consists:1 ijcv:1 combine:3 wild:1 indeed:1 expected:3 themselves:1 hoggles:2 multi:1 inspired:2 voc:6 spherical:1 overwhelming:1 window:3 struction:1 increasing:1 conv:1 matched:1 bike:3 msee:1 substantially:1 berkeley:2 every:1 mitigate:1 stricter:2 zaremba:1 k2:1 classifier:5 unit:1 vlfeat:3 ramanan:3 before:1 positive:2 local:7 struggle:1 tends:1 despite:4 establishing:1 path:1 interpolation:1 deeppose:1 pami:2 black:1 might:1 bird:4 twice:1 studied:1 plus:1 challenging:1 limited:1 malisiewicz:1 lecun:5 testing:2 backpropagation:2 digit:1 displacement:1 cascade:1 vedaldi:1 alleviated:1 matching:5 specificity:1 suggest:1 get:3 context:1 www:2 conventional:7 gkioxari:2 demonstrated:1 center:7 williams:1 attention:1 starting:1 conv4:11 independently:1 resolution:1 scipy:2 counterintuitive:1 retrieve:2 embedding:1 fulkerson:1 variation:1 coordinate:2 target:8 hierarchy:1 massive:1 goodfellow:1 associate:1 recognition:9 labeled:2 database:4 bottom:1 ft:2 huttenlocher:2 preprint:1 aero:3 region:8 connected:2 sun:1 ordering:2 highest:2 rescaled:1 multistage:1 kriegman:1 trained:11 solving:1 classifer:1 localization:16 joint:4 darpa:1 differently:1 cat:11 various:1 articulated:5 train:9 jain:3 fast:1 effective:1 pool5:1 approached:1 congealing:2 horse:4 choosing:1 neighborhood:2 caffe:3 whose:2 heuristic:1 cvpr:10 ability:4 fischer:1 succesful:1 transform:1 prsn:3 final:3 hoc:1 karayev:1 net:3 coming:1 aligned:3 achieve:1 intraclass:4 sutskever:2 darrell:4 extending:1 categorization:2 object:17 derive:1 bourdev:2 pose:14 exemplar:2 rescale:1 nearest:14 conv5:16 job:1 implemented:2 c:1 predicted:7 revers:1 poselets:4 ning:1 radius:3 correct:1 attribute:1 cnns:1 centered:1 human:8 engineered:3 bin:3 preliminary:1 investigation:1 yuen:1 bregler:1 around:3 ground:9 visually:1 seed:3 bicycle:2 predict:2 dictionary:1 torralba:2 smallest:1 early:1 estimation:11 sofa:3 label:4 jackel:1 sensitive:1 hubbard:1 concurrent:1 correctness:1 occupying:1 tool:1 hoffman:1 gaussian:2 rather:2 cornell:1 focus:1 improvement:1 consistently:1 fur:2 contrast:2 helpful:1 nn:2 entire:1 integrated:1 initially:1 selective:1 pixel:7 overall:2 classification:15 flexible:2 among:1 pascal:10 augment:1 issue:1 overfeat:3 animal:1 art:3 spatial:4 tzeng:1 andriluka:1 brox:1 field:20 equal:1 shaped:1 manually:1 kw:1 jones:1 unsupervised:5 icml:1 fcn:1 report:2 spline:1 quantitatively:1 dosovitskiy:1 employ:1 few:1 modern:1 randomly:1 densely:2 individual:1 replaced:4 replacement:1 detection:15 message:1 mining:1 alignment:21 sheep:3 henderson:1 mixture:2 tompson:1 accurate:2 edge:1 capable:2 necessary:1 facial:2 tree:2 kfs:1 euclidean:2 savarese:1 taylor:1 deformation:2 nbrs:3 mbike:3 smisc:1 sulting:1 girshick:3 instance:6 column:7 witnessed:1 cover:1 measuring:1 localizing:2 deviation:1 uniform:2 krizhevsky:2 reported:1 scanning:1 considerably:1 cited:1 international:1 lee:1 dong:1 picking:1 together:1 again:1 squared:1 huang:2 warped:6 style:1 li:2 doubt:1 szegedy:3 suggesting:1 diversity:1 de:1 stride:12 pooled:1 pedestrian:1 notable:1 depends:1 performed:2 view:1 lowe:1 analyze:1 start:1 maintains:1 annotation:4 gevers:1 jia:2 smeulders:1 ass:2 hariharan:1 publicly:1 square:3 convolutional:13 variance:1 descriptor:6 miller:2 efficiently:1 accuracy:8 bolded:1 yield:1 yellow:1 generalize:1 weak:2 handengineered:1 html:1 kavukcuoglu:1 overlook:1 finer:6 explain:1 detector:5 trevor:2 aligns:1 against:1 energy:2 chintala:1 associated:1 sampled:2 dataset:3 color:5 car:3 knowledge:1 segmentation:1 appears:1 centric:2 originally:1 higher:1 follow:1 response:6 improved:2 zisserman:1 formulation:1 done:1 though:3 strongly:1 box:6 just:1 convnets:9 hand:5 replacing:1 propagation:2 google:2 quality:3 resemblance:1 scientific:1 k22:1 requiring:3 true:1 contain:1 pascalnetwork:1 regularization:2 satisfactory:1 semantic:4 visualizing:2 width:2 cosine:3 criterion:1 illuminates:1 pck:4 mattar:1 performs:2 motion:3 l1:1 bring:1 percent:1 vondrick:1 image:41 wise:1 novel:1 recently:1 million:2 extend:1 refer:2 smoothness:1 grid:6 had:1 moving:1 bruna:1 supervision:1 similarity:4 base:1 aligning:2 fc7:1 align:1 closest:2 recent:4 certain:1 sande:1 binary:1 success:3 scoring:1 seen:1 impose:1 zip:1 deng:1 managing:1 belhumeur:3 signal:1 ii:3 sliding:2 multiple:2 keypoints:13 technical:1 match:2 cross:3 long:2 post:2 award:1 paired:1 prediction:16 mrf:2 regression:3 vision:4 arxiv:3 histogram:2 normalization:2 sometimes:1 cell:2 penalize:1 proposal:1 fine:3 winn:1 median:1 source:6 armlet:1 unlike:1 pooling:7 subject:1 tend:2 flow:20 spirit:1 effectiveness:2 seem:2 extracting:1 near:1 yang:2 intermediate:2 bengio:1 architecture:2 cow:4 idea:1 regarding:1 haffner:1 tradeoff:1 aeroplane:1 stereo:1 f:1 passing:1 deep:8 dramatically:1 useful:3 clear:1 transforms:1 nonparametric:1 mid:2 ten:1 svms:1 category:9 http:3 outperform:3 exist:1 nsf:1 notice:1 overly:1 correctly:1 per:2 key:1 four:1 localize:2 neither:1 rectangle:3 graph:1 enforced:1 run:1 powerful:1 respond:1 extends:1 almost:1 reasonable:1 throughout:1 groundtruth:2 yann:1 patch:12 bit:1 layer:15 bound:1 correspondence:13 copied:1 fold:1 precisely:1 fei:2 scene:1 toshev:2 chair:3 kumar:1 expanded:1 optical:1 conjecture:1 transferred:1 structured:2 tv:3 according:1 popularized:1 across:3 smaller:1 slightly:1 making:1 iccv:6 invariant:1 taken:2 visualization:3 bus:3 abbreviated:1 nose:5 available:1 apply:1 eight:2 ocr:1 away:2 generic:4 hierarchical:1 denker:1 eigen:1 original:1 top:2 clustering:1 include:1 zeiler:2 giving:2 build:1 objective:1 noticed:1 question:1 malik:4 print:1 receptive:19 fiducial:1 traditional:1 gradient:1 iclr:3 convnet:34 distance:4 arbelaez:1 landmark:3 oliphant:1 portable:1 consensus:1 code:1 copying:1 relationship:1 index:1 providing:1 rotational:1 sermanet:2 difficult:1 debate:1 negative:3 motivates:1 conv2:1 perform:8 twenty:2 upper:2 subcategory:1 observation:1 datasets:1 discarded:1 howard:1 hinton:1 variability:1 precise:4 introduced:1 dog:3 imagenet:3 california:1 learned:5 boser:1 nip:2 address:1 able:1 beyond:2 suggested:1 regime:1 doubted:1 challenge:4 program:1 rf:12 including:2 max:1 belief:2 gool:1 greatest:1 overlap:1 natural:2 rely:1 boat:3 nth:1 zhu:1 loom:1 keypoint:45 eye:5 mathieu:1 library:1 excel:1 extract:1 text:1 prior:13 understanding:3 acknowledgement:1 python:1 val:2 fully:2 plant:4 validation:4 shelhamer:1 verification:2 conv1:1 share:1 row:2 eccv:1 placed:1 supported:1 electronically:1 tire:1 understand:2 warp:2 wide:1 conv3:4 face:4 neighbor:21 taking:1 felzenszwalb:2 peterson:1 tolerance:1 van:2 boundary:1 dimension:1 rich:1 erhan:1 implicitly:2 fergus:4 discriminative:2 xi:6 spectrum:1 search:1 khosla:1 table:13 learn:1 transfer:4 channel:1 investigated:1 complex:1 bottou:1 uijlings:1 voc2011:2 spread:1 dense:3 whole:2 bounding:6 body:1 referred:1 fashion:2 lie:3 candidate:2 breaking:1 toyota:1 third:1 grained:2 donahue:3 removing:1 specific:4 sift:44 offset:1 svm:4 evidence:2 bivariate:1 workshop:1 socher:1 effectively:1 gained:1 decaf:1 magnitude:1 illustrates:1 margin:1 suited:1 entropy:1 locality:1 simply:1 appearance:1 likely:1 visual:6 vinyals:1 corresponds:1 truth:9 viewed:3 formulated:1 replace:2 considerable:1 hard:1 specifically:4 determined:1 uniformly:2 semantically:1 averaging:2 invariance:1 attempted:1 meaningful:1 ilsvrc:1 berg:1 support:1 people:1 latter:1 jonathan:1 evaluate:1 scratch:1 |
4,884 | 5,421 | Deep Learning for Real-Time Atari Game Play
Using Offline Monte-Carlo Tree Search Planning
Satinder Singh
Computer Science and Eng.
University of Michigan
[email protected]
Xiaoxiao Guo
Computer Science and Eng.
University of Michigan
[email protected]
Honglak Lee
Computer Science and Eng.
University of Michigan
[email protected]
Richard Lewis
Department of Psychology
University of Michigan
[email protected]
Xiaoshi Wang
Computer Science and Eng.
University of Michigan
[email protected]
Abstract
The combination of modern Reinforcement Learning and Deep Learning approaches holds the promise of making significant progress on challenging applications requiring both rich perception and policy-selection. The Arcade Learning
Environment (ALE) provides a set of Atari games that represent a useful benchmark set of such applications. A recent breakthrough in combining model-free
reinforcement learning with deep learning, called DQN, achieves the best realtime agents thus far. Planning-based approaches achieve far higher scores than the
best model-free approaches, but they exploit information that is not available to
human players, and they are orders of magnitude slower than needed for real-time
play. Our main goal in this work is to build a better real-time Atari game playing
agent than DQN. The central idea is to use the slow planning-based agents to provide training data for a deep-learning architecture capable of real-time play. We
proposed new agents based on this idea and show that they outperform DQN.
1
Introduction
Many real-world Reinforcement Learning (RL) problems combine the challenges of closed-loop
action (or policy) selection with the already significant challenges of high-dimensional perception
(shared with many Supervised Learning problems). RL has made substantial progress on theory
and algorithms for policy selection (the distinguishing problem of RL), but these contributions have
not directly addressed problems of perception. Deep learning (DL) approaches have made remarkable progress on the perception problem (e.g., [11, 17]) but do not directly address policy selection.
RL and DL methods share the aim of generality, in that they both intend to minimize or eliminate
domain-specific engineering, while providing ?off-the-shelf? performance that competes with or exceeds systems that exploit control heuristics and hand-coded features. Combining modern RL and
DL approaches therefore offers the potential for general methods that address challenging applications requiring both rich perception and policy-selection.
The Arcade Learning Environment (ALE) is a relatively new and widely accessible class of benchmark RL problems that provide a particularly challenging combination of policy selection and perception. ALE includes an emulator and a large number of Atari 2600 (a 1970s?80s home video
console) games. The complexity and diversity of the games?both in terms of perceptual challenges
in mapping pixels to useful features for control and in terms of the control policies needed?make
1
ALE a useful set of benchmark RL problems, especially for evaluating general methods intended to
achieve success without hand-engineered features.
Since the introduction of ALE, there have been a number of attempts to build general-purpose Atari
game playing agents. The departure point for this paper is a recent and significant breakthrough [16]
that combines RL and DL to build agents for multiple Atari Games. It achieved the best machineagent real-time game play to date (in some games close to or better than human-level play), does not
require feature engineering, and indeed reuses the same perception architecture and RL algorithm
across all the games. We believe that continued progress on the ALE environment that preserves
these advantages will extend to broad advances in other domains with significant perception and
policy selection challenges. Thus, our immediate goal in the work reported here is to build even
better performing general-purpose Atari Game playing agents. We achieve this by introducing new
methods for combining RL and DL that use slow, off-line Monte Carlo tree search planning methods
to generate training data for a deep-learned classifier capable of state-of-the-art real-time play.
2
Brief background on RL and DL and challenges of perception
RL and more broadly decision-theoretic planning has a suite of methods that address the challenge of
selecting/learning good policies, including value function approximation, policy search, and MonteCarlo Tree Search [9, 10] (MCTS). These methods have different strengths and weaknesses and there
is increasing understanding of how to match them to different types of RL-environments. Indeed,
an accumulating number of applications attest to this success. But it is still not the case that there
are reasonably off-the-shelf approaches to solving complex RL problems of interest to Artificial
Intelligence (AI) such as the games in ALE. One reason for this is that despite major advances there
hasn?t been an off-the-shelf approach to significant perception problems. The perception problem
itself has two components: 1) the sensors at any time step do not capture all the information in
the history of observations, leading to partial observability, and 2) the sensors provide very highdimensional observations that introduce computational and sample-complexity challenges for policy
selection.
One way to handle the perception challenges when a model of the RL environment is available is
to avoid the perception problem entirely by eschewing the building of an explicit policy and instead
using repeated incremental planning via MCTS methods such as UCT [10] (discussed below). Either
when a model is not available, or when an explicit representation of the policy is required, the usual
approach to applied RL success has been to use expert-developed task-specific features of a short
history of observations in combination with function approximation methods and some trial-anderror on the part of the application developer (on small enough problems this can be augmented with
some automated feature selection methods). Eliminating the dependence of applied RL success on
engineered features motivates our interest in combining RL and DL (though see [20] for early work
in this direction).
Over the past decades, deep learning (see [3, 19] for a survey) has emerged as a powerful technique
for learning feature representations from data (again, this is in a stark contrast to the conventional
way of hand-crafting features by domain experts). For example, DL has achieved state-of-the-art results in image classification [11, 4], speech recognition [15, 17, 6], and activity recognition [12, 8].
In DL, features are learned in a compositional hierarchy. Specifically, low-level features are learned
to encode low-level statistical dependencies (e.g., ?edges? in images), and higher-level features encode higher-order dependencies of the lower-level features (e.g., ?object parts?) [14]. In particular,
for data that has strong spatial or temporal dependencies, convolutional neural networks [13] have
been shown to learn invariant high-level features that are informative for supervised tasks. Such
convolutional neural networks were used in the recent successful combination of DL and RL for
Atari Game playing [16] that forms the departure point of our work. We describe this work in more
detail below.
3
Existing Work on Atari Games and a Performance Gap
While the games in ALE are simpler than many modern games, they still pose significant challenges
to human players. In RL terms, for a human player these games are Partially-Observable Markov
Decision Processes (POMDPs). The true state of each game at any given point is captured by the
2
contents of the limited random-access memory (RAM). A human player does not observe the state
and instead perceives the game screen (frame) which is a 2D array of 7-bit pixels, 160 pixels wide by
210 pixels high. The action space available to the player depends on the game but maximally consists
of the 18 discrete actions defined by the joystick controller. The next state is a deterministic function
of the previous state and the player?s action choice. Stochasticity in these games is limited to the
choice of the initial state of the game (which can include a random number seed stored in RAM).
So even though the state transitions are deterministic, the transitions from history of observations
and actions to next observation can be stochastic (because of the stochastic initial hidden state). The
immediate reward at any given step is defined by the game and made available by the ALE; it is
usually a function of the current frame or the difference between current and previous frames. When
running in real-time, the simulator generates 60 frames per second. All the games we consider
terminate in a finite number of time-steps (and so are episodic). The goal in these games is to select
an optimal policy, i.e., to select actions in such a way so as to maximize the expected value of the
cumulative sum of rewards until termination.
Model-Free RL Agents for Atari Games. Here we discuss work that does not access the state
in the games and thus solves the game as a POMDP. In principle one could learn a state representation and infer an associated MDP model using frame-observation and action trajectories, but
these games are so complex that this is rarely done. Instead, partial observability is dealt with by
hand-engineering features of short histories of frames observed so far and model-free RL methods
are used to learn good policies as a function of those feature representations. For example, the paper that introduced ALE [1], used SARSA with several different hand-engineered features sets. The
contingency awareness approach [4] improved performance of the SARSA algorithm by augmenting
the feature sets with a learned representation of the parts of the screen that are under the agent?s control. The sketch-based approach [2] further improves performance by using the tug-of-war sketch
features. HyperNEAT-GGP [7] introduces an evolutionary policy search based Atari game player.
Most recently Deep Q-Network (hereafter DQN) [16] uses a modified version of Q-Learning with a
convolutional neural network (CNN) with three hidden layers for function approximation. This last
approach is the state of the art in this class of methods for Atari games and is the basis for our work;
we present the relevant details in Section 5. It does not use hand-engineered features but instead
provides the last four raw frames as input (four instead of one to alleviate partial observability).
Planning Agents for Atari Games based on UCT. These approaches access the state of the game
from the emulator and hence face a deterministic MDP (other than the random choice of initial state).
They incrementally plan the action to take in the current state using UCT, an algorithm widely used
for games. UCT has three parameters, the number of trajectories, the maximum-depth (uniform for
each trajectory), and a exploration parameter (a scalar set to 1 in all our experiments). In general,
the larger the trajectory & depth parameters are, the slower UCT is but the better it is. UCT uses the
emulator as a model to simulate trajectories as follows. Suppose it is generating the k th trajectory
and the current node is at depth d and the current state is s. It computes a score for each possible
action a in state-depth pair (s, d) as the sum of two terms, an exploitation term that is the MonteCarlo average of the discounted sum of rewards obtained from experiences
p with state-depth pair
(s, d) in the previous k ? 1 trajectories, and an exploration term that is log (n(s, d))/n(s, a, d)
where n(s, a, d) and n(s, d) are the number of experiences of action a with state-depth pair (s, d)
and with state-depth pair (s, d) respectively in the previous k ? 1 trajectories. UCT selects the
action to simulate in order to extend the trajectory greedily with respect to this summed score. Once
the input-parameter number of trajectories are generated each to maximum depth, UCT returns the
exploitation term for each action at the root node (which is the current state it is planning an action
for) as its estimate of the utility of taking that action in the current state of the game. UCT has the
nice theoretical property that the number of simulation steps (number of trajectories ? maximumdepth) needed to ensure any bound on the loss of following the UCT-based policy is independent of
the size of the state space; this result expresses the fact that the use of UCT avoids the perception
problem, but at the cost of requiring substantial computation for every time step of action selection
because it never builds an explicit policy.
Performance Gap & our Opportunity. The opportunity for this paper arises from the following
observations. The model-free RL agents for Atari games are fast (indeed faster than real-time, e.g.,
the CNN-based approach from our paper takes 10?4 seconds to select an action on our computer)
while the UCT-based planning agents are several orders of magnitude slower (much slower than
real-time, e.g., they take seconds to select an action on the same computer). On the other hand,
3
the performance of UCT-based planning agents is much better than the performance of model-free
RL agents (this will be evident in our results below). Our goal is to develop methods that retain
the DL advantage of not needing hand crafted features and the online real-time play ability of the
model-free RL agents by exploiting data generated by UCT-planning agents.
4
Methods for Combining UCT-based RL with DL
We first describe the baseline UCT agent, and then three agents that instantiate different methods of
combining the UCT agent with DL. Recall that in keeping with the goal of building general-purpose
methods as in the DQN work we impose the constraint of reusing the same input representations,
the same function approximation architecture, and the same planning method for all the games.
4.1
Baseline UCT agent that provides training data
This agent requires no training. It does, however, require specification of its two parameters, the
number of trajectories and the maximum-depth. Recall that our proposed new agents will all use
data from this UCT-agent to train a CNN-based policy and so it is reasonable that the resulting
performance of our proposed agents will be worse than that of the UCT-agent. Therefore, in our
experiments we set these two parameters large enough to ensure that they outscore the published
DQN scores, but not so large that they make our computational experiments unreasonably slow.
Specifically, we elected to use 300 as maximum-depth and 10000 as number of trajectories for all
games but two. Pong turns out to be a much simpler game and we could reduce the number of
trajectories to 500, and Enduro turned out to have more distal rewards than the other games and so
we used a maximum-depth of 400. As will be evident from the results in Section 5 this allowed the
UCT agent to significantly outperform DQN in all games but Pong in which DQN already performs
perfectly. We emphasize that the UCT agent does not meet our goal of real-time play. For example,
to play a game just 800 times with the UCT agent (we do this to collect training data for our agent?s
below) takes a few days on a recent multicore computer for each game.
4.2
Our three methods and their corresponding agents
Method 1: UCTtoRegression (for UCT to CNN via Regression). The key idea is to use the action
values computed by the UCT-agent to train a regression-based CNN. The following is done for each
game. Collect 800 UCT-agent runs by playing the game 800 times from start to finish using the UCT
agent above. Build a dataset (table) from these runs as follows. Map the last four frames of each state
along each trajectory into the action-values of all the actions as computed by UCT. This training data
is used to train the CNN via regression (see below for CNN details). The UCTtoRegression-agent
uses the CNN learned by this training procedure to select actions during evaluation.
Method 2: UCTtoClassification (for UCT to CNN via Classification). The key idea is to use the
action choice computed by the UCT-agent (selected greedily from action-values) to train a classifierbased CNN. The following is done for each game. Collect 800 UCT-agent runs as above. These runs
yield a table in which the rows correspond to the last four frames at each state along each trajectory
and the single column is the choice of action that is best according to the UCT-agent at that state of
the trajectory. This training data is used to train the CNN via multinomial classification (see below
for CNN details). The UCTtoClassification-agent uses the CNN-classifier learned by this training
procedure to select actions during evaluation.
One potential issue with the above two agents is that the training data?s input distribution is generated
by the UCT-agent while during testing the UCTtoRegression and UCTtoClassification agents will
perform differently from the UCT-agent and thus could experience an input distribution quite difference from that of the UCT-agent?s. This could limit the testing performance of the UCTtoRegression
and UCTtoClassification agents. Thus, it might be desirable to somehow bias the distribution over
inputs to those likely to be encountered by these agents; this observation motivates our next method.
Method 3: UCTtoClassification-Interleaved (for UCT to CNN via Classification-Interleaved).
The key idea is to focus UCT planning on that part of the state space experienced by the (partially
trained) CNN player. The method accomplishes this by interleaving training and data collection as
4
84
84
fully-connectedlayer (max(0,x))
conv-layer
(tanh)
conv-layer
(tanh)
20
4
20
9
9
16
32
fully-connectedlayer (linear)
256
Figure 1: The CNN architecture from DQN [6] that we adopt in our agents. See text for details.
follows1 . Collect 200 UCT-agent runs as above; these will obviously have the same input distribution
concern raised above. The data from these runs is used to train the CNN via multinomial classification just as in the UCTtoClassification-agent?s method (we do not do this for the UCTtoRegressionagent because as we show below it performs worse than the UCTtoClassification-agent). The trained
CNN is then used to decide action choices in collecting a further 200 runs (though 5% of the time
a random action is chosen to ensure some exploration). At each state of the game along each trajectory, UCT is asked to compute its choice of action and the original data set is augmented with
the last four frames for each state as the rows and the column as UCT?s action choice. This 400
trajectory dataset?s input distribution is now potentially different from that of the UCT-agent. This
dataset is used to train the CNN again via multinomial classification. This interleaved procedure is
repeated until there are a total of 800 runs worth of data in the dataset for the final round of training
of the CNN. The UCTtoClassification-Interleaved agent uses the final CNN-classifier learned by this
training procedure to select actions during testing.
In order to focus our empirical evaluation on the contribution of the non-DL part of our three new
agents, we reused exactly the same convolutional neural network architecture as used in the DQN
work (we describe this architecture in brief detail below). The DQN work modified the reward
functions for some of the games (by saturating them at +1 and ?1) while we use unmodified reward
functions (these only play a role in the UCT-agent components of our methods and not in the CNN
component). We also follow DQN?s frame-skipping techniques: the agent sees and selects actions
on every k th frame instead of every frame (k = 3 for Space Invaders and k = 4 for all other games),
and the latest chosen-action is repeated on subsequently-skipped frames.
4.3
Details of Data Preprocessing and CNN Architecture
Preprocessing (identical to DQN to the best of our understanding). Raw Atari game frames are
160 ? 210 pixel images with a 128-color palette. We convert the RGB representation to gray-scale
and crop an 160 ? 160 region of the image that captures the playing area, and then the cropped
image is down-sampled to 84 ? 84 in order to reuse DQN?s CNN architecture. This procedure is
applied to the last 4 frames associated with a state and stacked to produce a 84 ? 84 ? 4 preprocessed
input representation for each state. We subtracted the pixel-level means and scale the inputs to lie in
the range [-1, 1]. We shuffle the training data to break the strong correlations between consecutive
samples, which therefore reduces the variance of the updates.
CNN Architecture. We use the same deep neural network architecture as DQN [16] for our agents.
As depicted in Figure 1, our network consists of three hidden layers. The input to the neural network
is an 84 ? 84 ? 4 image produced by the preprocessing procedure above. The first hidden layer
convolves 16, 8 ? 8, filters with stride 4 with the input image and applies a rectifier nonlinearity
(tanh). The second hidden layer convolves 32, 4 ? 4, filters with stride 2 again followed by a
rectifier nonlinearity (tanh). The final hidden layer is fully connected and consists of 256 rectifier
(max) units. In the multi-regression-based agent (UCTtoRegression), the output layer is a fully
connected linear layer with a single output for each valid action. In the classification-based agents
(UCTtoClassification, UCTtoClassification-Interleaved), a softmax (instead of linear) function is
applied to the final output layer. We refer the reader to the DQN paper for further detail.
1
Our UCTtoClassification-Interleaved method is a special case of DAgger [18] (in the use of a CNNclassifier and in the use of specific choices of parameters ?1 = 1, and for i > 1, ?i = 0). As a small
point of difference, we note that our emphasis in this paper was in the use of CNNs to avoid the use of handcrafted domain specific features, while the empirical work for DAgger did not have the same emphasis and so
used handcrafted features.
5
Table 1: Performance (game scores) of the four real-time game playing agents, where UCR is short for UCTtoRegression, UCC is short for UCTtoClassification, and UCC-I is short for UCTtoClassification-Interleaved.
Agent
B.Rider
Breakout
Enduro
Pong
Q*bert
Seaquest
S.Invaders
DQN
-best
4092
5184
168
225
470
661
20
21
1952
4500
1705
1740
581
1075
UCC
-best
-greedy
5342 (20)
10514
5676
175(5.63)
351
269
558(14)
942
692
19(0.3)
21
21
11574(44)
29725
19890
2273(23)
5100
2760
672(5.3)
1200
680
UCC-I
-best
-greedy
5388(4.6)
10732
5702
215(6.69)
413
380
601(11)
1026
741
19(0.14)
21
21
13189(35.3)
29900
20025
2701(6.09)
6100
2995
670(4.24)
910
692
UCR
2405(12)
143(6.7)
566(10.2)
19(0.3)
12755(40.7)
1024 (13.8)
441(8.1)
Table 2: Performance (game scores) of the off-line UCT game playing agent.
5
Agent
B.Rider
Breakout
Enduro
Pong
Q*bert
Seaquest
S.Invaders
UCT
7233
406
788
21
18850
3257
2354
Experimental Results
First we present our main performance results and then present some visualizations to help understand the performance of our agents. In Table 1 we compare and contrast the performance of the
four real-time game playing agents, three of which (UCTtoRegression, UCTtoClassification, and
UCTtoClassification-Interleaved) we implemented and evaluated; the performance of the DQN was
obtained from [16].
The columns correspond to the seven games named in the header, and the rows correspond to different assessments of the four agents. Throughout the numbers in parentheses are standard-errors.
The DQN row reports the average performance (game score) of the DQN agent (a random action is
chosen 5% of the time during testing). The DQN-best row is the best performance of the DQN over
all the attempts at each game incorporated in the row corresponding to DQN. Comparing the performance of the UCTtoClassification and UCTtoRegression agents (both use 5% exploration), we see
that the UCTtoClassification agent either competes well with or significantly outperforms the UCTtoRegression agent. More importantly the UCTtoClassification agent outperforms the DQN agent
in all games but Pong (in which both agents do nearly perfectly because the maximum score in this
game is 21). In some games (B.Rider, Enduro, Q*Bert, Sequest and S.Invaders) the percentageperformance gain of UCTtoClassification over DQN is quite large. Similar gains are obtained in the
comparison of UCTtoClassification-best to DQN-best.
We used 5% exploration in our agents to match what the DQN agent does, but it is not clear why
one should consider random action selection during testing. In any case, the effect of this randomness in action-selection will differ across games (based, e.g., on whether a wrong action can
be terminal). Thus, we also present results for the UCTtoClassification-greedy agent in which we
don?t do any exploration. As seen by comparing the rows corresponding to UCTtoClassification
and UCTtoClassification-greedy, the latter agent always outperforms the former and in four games
(Breakout, Enduro, Q*Bert, and Seaquest) achieves further large-percentage improvements.
Table 2 gives the performance of our non-realtime UCT agent (again, with 5% exploration). As
discussed above we selected UCT-agent?s parameters to ensure that this agent outperforms the DQN
agent allowing room for our agents to perform in the middle.
Finally, recall that the UCTtoClassification-Interleaved agent was designed so that its input distribution during training is more likely to match its input distribution during evaluation and we hypothesized that this would improve performance relative to UCTtoClassification. Indeed, in all games but
B. Rider, Pong and S.Invaders in which the two agents perform similarly, UCTtoClassificationInterleaved significantly outperforms UCTtoClassification. The same holds when comparing
6
frame: t-3
t-2
t-1
t
?submarine?
?diver?
?enemy?
?enemy+diver?
Figure 2: Visualization of the first-layer features learned from Seaquest. (Left) visualization of four first-layer
filters; each filter covers four frames, showing the spatio-temporal template. (Middle) a captured screen. (Right)
gray-scale version of the input screen which is fed into the CNN. Four filters were color-coded and visualized
as dotted bounding boxes at the locations where they get activated. This figure is best viewed in color.
UCTtoClassification-Interleaved-best and UCTtoClassification-best as well as UCTtoClassificationInterleaved-greedy and UCTtoClassification-greedy.
In a further preliminary exploration of the effectiveness of the UCTtoClassification-Interleaved
in exploiting additional computational resources for generating UCT runs, on the game Enduro
we compared UCTtoClassification and UCTtoClassification-Interleaved where we allowed each of
them twice the number of UCT runs used in producing the Table 1 above, i.e., 1600 runs while
keeping a batch size of 200. The performance of UCTtoClassification improves from 558 to 581
while the performance of UCTtoClassification-Interleaved improves from 601 to 670, i.e., the interleaved method improved more in absolute and percentage terms as we increased the amount of
training data. This is encouraging and is further confirmation of the hypothesis that motivated the
interleaved method, because the interleaved input distribution would be even more like that of the
final agent with the larger data set.
Learned Features from Convolutional Layers. We
provide visualizations of the learned filters in order to
gain insights on what the CNN learns. Specifically, we
apply the ?optimal stimuli? method [5] to visualize the
features CNN learned after training. The method picks
the input image patches that generate the greatest responsive after convolution with the trained filters. We
select 8*8*4 input patches to visualize the first convolutional layer features and 20*20*4 to visualize the second convolutional layer filters. Note that these patch
sizes correspond to receptive field sizes of the learned
features in each layer.
In Figure 2, we show four first-layer filters of the CNN
trained from Seaquest for UCTtoClassification-agent.
Specifically, each filter covers four frames of 8*8 pixels, which can be viewed as a spatio-temporal tem- Figure 3: Visualization of the second-layer
plate that captures specific patterns and their temporal features learned from Seaquest.
changes. We also show an example screen capture and
visualize where the filters get activated in the gray-scale version of the image (which is the actual
input to the CNN model). The visualization suggests that the first-layer filters capture ?object-part?
patterns and their temporal movements.
Figure 3 visualizes the four second-layer features via the optimal stimulus method, where each row
corresponds to a filter. We can see that the second-layer features capture bigger spatial patterns
(often covering beyond the size of individual objects), while encoding interactions between objects,
such as two enemies moving together, and submarine moving along a direction. Overall, these
qualitative results suggest that the CNN learns relevant patterns useful for game playing.
7
Step 69: FIRE
Step 70: DOWN+FIRE
Step 74:DOWN+FIRE
Step 75:RIGHT+FIRE
Step 76:RIGHT+FIRE
Step 78: RIGHT+FIRE
Step 79:DOWN+FIRE
Figure 4: A visualization of the UCTtoClassification agent?s policy as it kills an enemy agent.
Visualization of Learned Policy. Here we present visualizations of the policy learned by the UCTtoClassification agent with the aim of illustrating both what it does well and what it does not.
Figure 4 shows the policy learned by UCTtoClassification to destroy nearby enemies. The CNN
changes the action from ?Fire? to ?Down+Fire? at time step 70 when the enemies first show up at
the right columns of the screen, which will move the submarine to the same horizontal position of the
closest enemy. At time step 75, the submarine is at the horizontal position of the closest enemy and
the action changes to ?Right+Fire?. The ?Right+Fire? action is repeated until the enemy is destroyed
at time step 79. At time step 79, the predicted action is changed to ?Down+Fire? again to move the
submarine to the horizontal position of the next closest enemy. This shows the UCTtoClassification
agent?s ability to deal with delayed reward as it learns to take a sequence of unrewarded actions
before it obtains any reward when it finally destroys an enemy.
Figure 4 also shows a shortcoming in the UCTtoClassification agent?s policy, namely it does not
purposefully take actions to save a diver (saving a diver can lead to a large reward). For example, at
time step 69, even though there are two divers below and to the right of the submarine (our agent), the
learned policy does not move the submarine downward. This phenomenon was observed frequently.
The reason for this shortcoming is that it can take a large number of time steps to capture 6 divers
and bring them to surface (bringing fewer divers to the surface does not yield a reward); this takes
longer than the planning depth of UCT. Thus, it is UCT that does not purposefully save divers and
thus the training data collected via UCT reflects that defect which is then also present in the play of
the UCTtoClassification (and UCTtoClassification-Interleaved) agent.
6
Conclusion
UCT-based planning agents are unrealistic for Atari game play in at least two ways. First, to play
the game they require access to the state of the game which is unavailable to human players, and
second they are orders of magnitude slower than realtime. On the other hand, by slowing the game
down enough to allow UCT to play leads to the highest scores on the games they have been tried on.
Indeed, by allowing UCT more and more time (and thus allowing for larger number of trajectories
and larger maximum-depth) between moves one can presumably raise the score more and more. We
identified a gap between the UCT-based planning agents performance and the best realtime player
DQN?s performance and developed new agents to partially fill this gap. Our main applied result is
that at the time of the writing of this paper we have the best realtime Atari game playing agents
on the same 7 games that were used to evaluate DQN. Indeed, in most of the 7 games our best
agent beats DQN significantly. Another result is that at least in our experiments training the CNN to
learn a classifier that maps game observations to actions was better than training the CNN to learn
a regression function that maps game observations to action-values (we intend to do further work
to confirm how general this result is on ALE). Finally, we hypothesized that the difference in input
distribution between the UCT agent that generates the training data and the input distribution experienced by our learned agents would diminish performance. The UCTtoClassification-Interleaved
agent we developed to deal with this issue indeed performed better than the UCTtoClassification
agent indirectly confirming our hypothesis and solving the underlying issue.
Acknowledgments. This work was supported in part by NSF grant IIS-1148668. Any opinions,
findings, conclusions, or recommendations expressed here are those of the authors and do not necessarily reflect the views of the sponsors.
8
References
[1] M. G. Bellemare, Y. Naddaf, J. Veness, and M. Bowling. The arcade learning environment: an
evaluation platform for general agents. Journal of Artificial Intelligence Research, 47(1):253?
279, 2013.
[2] M. G. Bellemare, J. Veness, and M. Bowling. Sketch-based linear value function approximation. In Advances in Neural Information Processing Systems, pages 2222?2230, 2012.
[3] Y. Bengio. Learning deep architectures for AI. Foundations and trends in Machine Learning,
2(1):1?127, 2009.
[4] D. Ciresan, U. Meier, and J. Schmidhuber. Multi-column deep neural networks for image
classification. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR),
2012, pages 3642?3649. IEEE, 2012.
[5] D. Erhan, Y. Bengio, A. Courville, and P. Vincent. Visualizing higher-layer features of a deep
network. Technical report, University of Montreal, 2009.
[6] A. Graves, A. Mohamed, and G. Hinton. Speech recognition with deep recurrent neural
networks. In IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), 2013, pages 6645?6649. IEEE, 2013.
[7] M. Hausknecht, P. Khandelwal, R. Miikkulainen, and P. Stone. HyperNEAT-GGP: A
hyperNEAT-based Atari general game player. In Proceedings of the fourteenth international
conference on Genetic and evolutionary computation conference, pages 217?224. ACM, 2012.
[8] A. Karpathy, G. Toderici, S. Shetty, T. Leung, R. Sukthankar, and L. Fei-Fei. Large-scale
video classification with convolutional neural networks. In IEEE Conference on Computer
Vision and Pattern Recognition (CVPR), 2014.
[9] M. Kearns, Y. Mansour, and A. Y. Ng. A sparse sampling algorithm for near-optimal planning
in large Markov decision processes. Machine Learning, 49(2-3):193?208, 2002.
[10] L. Kocsis and C. Szepesv?ari. Bandit based Monte-Carlo planning. In Machine Learning:
European Conference on Machine Learning 2006, pages 282?293. Springer, 2006.
[11] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional
neural networks. In Advances in Neural Information Processing Systems, 2012.
[12] Q. V. Le, W. Y. Zou, S. Y. Yeung, and A. Y. Ng. Learning hierarchical invariant spatio-temporal
features for action recognition with independent subspace analysis. In IEEE Conference on
Computer Vision and Pattern Recognition (CVPR), 2011, pages 3361?3368. IEEE, 2011.
[13] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document
recognition. Proceedings of the IEEE, 86(11):2278?2324, 1998.
[14] H. Lee, R. Grosse, R. Ranganath, and A. Y. Ng. Convolutional deep belief networks for
scalable unsupervised learning of hierarchical representations. In Proceedings of the 26th
International Conference on Machine Learning, pages 609?616, 2009.
[15] H. Lee, P. Pham, Y. Largman, and A. Y. Ng. Unsupervised feature learning for audio classification using convolutional deep belief networks. In Advances in Neural Information Processing
Systems, pages 1096?1104, 2009.
[16] V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, and M. Riedmiller. Playing Atari with deep reinforcement learning. In Deep Learning, Neural Information
Processing Systems Workshop, 2013.
[17] A. Mohamed, G. E. Dahl, and G. Hinton. Acoustic modeling using deep belief networks. IEEE
Transactions on Audio, Speech, and Language Processing, 20(1):14?22, 2012.
[18] S. Ross, G. J. Gordon, and J. A. Bagnell. A reduction of imitation learning and structured
prediction to no-regret online learning. In Proceedings of the 14th International Conference
on Artificial Intelligence and Statistics (AISTATS), 2011.
[19] J. Schmidhuber. Deep learning in neural networks: An overview. Neural Networks, 2014.
[20] G. Tesauro. Temporal difference learning and TD-gammon. Communications of the ACM,
38(3):58?68, 1995.
9
| 5421 |@word trial:1 cnn:34 version:3 eliminating:1 exploitation:2 middle:2 illustrating:1 reused:1 termination:1 simulation:1 tried:1 rgb:1 eng:4 pick:1 reduction:1 initial:3 score:10 selecting:1 hereafter:1 genetic:1 document:1 past:1 existing:1 outperforms:5 current:7 comparing:3 skipping:1 informative:1 confirming:1 designed:1 update:1 intelligence:3 instantiate:1 selected:2 greedy:6 fewer:1 slowing:1 short:5 provides:3 node:2 location:1 simpler:2 wierstra:1 along:4 qualitative:1 consists:3 combine:2 introduce:1 expected:1 indeed:7 planning:18 frequently:1 simulator:1 multi:2 terminal:1 discounted:1 td:1 encouraging:1 actual:1 toderici:1 guoxiao:1 increasing:1 perceives:1 conv:2 competes:2 underlying:1 what:4 atari:19 hyperneat:3 follows1:1 developer:1 developed:3 finding:1 suite:1 temporal:7 every:3 collecting:1 exactly:1 classifier:4 wrong:1 control:4 unit:1 grant:1 reuses:1 producing:1 before:1 engineering:3 limit:1 despite:1 encoding:1 meet:1 might:1 emphasis:2 twice:1 collect:4 challenging:3 suggests:1 limited:2 range:1 acknowledgment:1 lecun:1 testing:5 regret:1 procedure:6 episodic:1 area:1 riedmiller:1 empirical:2 significantly:4 gammon:1 arcade:3 suggest:1 get:2 close:1 selection:12 writing:1 bellemare:2 accumulating:1 sukthankar:1 conventional:1 deterministic:3 map:3 latest:1 survey:1 pomdp:1 insight:1 continued:1 array:1 importantly:1 fill:1 handle:1 hierarchy:1 play:14 suppose:1 distinguishing:1 us:5 hypothesis:2 invader:5 trend:1 recognition:8 particularly:1 observed:2 role:1 wang:1 capture:7 region:1 connected:2 shuffle:1 movement:1 highest:1 substantial:2 environment:6 pong:6 complexity:2 reward:10 asked:1 trained:4 singh:1 solving:2 raise:1 basis:1 icassp:1 convolves:2 differently:1 joystick:1 train:7 stacked:1 fast:1 describe:3 shortcoming:2 monte:3 eschewing:1 artificial:3 header:1 quite:2 heuristic:1 widely:2 emerged:1 larger:4 enemy:11 cvpr:3 ability:2 statistic:1 itself:1 final:5 online:2 obviously:1 kocsis:1 advantage:2 sequence:1 interaction:1 relevant:2 combining:6 loop:1 date:1 turned:1 achieve:3 anderror:1 breakout:3 exploiting:2 sutskever:1 produce:1 generating:2 incremental:1 silver:1 object:4 help:1 develop:1 montreal:1 augmenting:1 pose:1 recurrent:1 multicore:1 progress:4 solves:1 strong:2 implemented:1 predicted:1 differ:1 direction:2 filter:13 stochastic:2 subsequently:1 exploration:8 human:6 engineered:4 cnns:1 opinion:1 require:3 alleviate:1 preliminary:1 sarsa:2 hold:2 pham:1 diminish:1 presumably:1 seed:1 mapping:1 visualize:4 major:1 achieves:2 early:1 adopt:1 consecutive:1 purpose:3 tanh:4 ross:1 reflects:1 destroys:1 sensor:2 always:1 aim:2 modified:2 avoid:2 shelf:3 encode:2 focus:2 improvement:1 contrast:2 skipped:1 greedily:2 baseline:2 leung:1 eliminate:1 hidden:6 bandit:1 selects:2 pixel:7 issue:3 classification:11 overall:1 seaquest:6 plan:1 art:3 breakthrough:2 spatial:2 summed:1 raised:1 softmax:1 once:1 never:1 special:1 veness:2 sampling:1 ng:4 identical:1 field:1 broad:1 saving:1 unsupervised:2 nearly:1 tem:1 report:2 stimulus:2 gordon:1 richard:1 few:1 modern:3 preserve:1 individual:1 delayed:1 intended:1 fire:12 attempt:2 interest:2 mnih:1 evaluation:5 weakness:1 introduces:1 activated:2 edge:1 capable:2 partial:3 experience:3 hausknecht:1 tree:3 theoretical:1 increased:1 column:5 modeling:1 cover:2 unmodified:1 cost:1 introducing:1 uniform:1 krizhevsky:1 successful:1 reported:1 stored:1 dependency:3 international:4 accessible:1 retain:1 lee:3 off:5 together:1 again:5 central:1 reflect:1 worse:2 expert:2 leading:1 return:1 stark:1 reusing:1 potential:2 diversity:1 stride:2 includes:1 hasn:1 depends:1 performed:1 root:1 break:1 closed:1 view:1 attest:1 tug:1 start:1 dagger:2 contribution:2 minimize:1 convolutional:11 variance:1 yield:2 correspond:4 dealt:1 raw:2 vincent:1 kavukcuoglu:1 produced:1 carlo:3 trajectory:20 pomdps:1 worth:1 published:1 randomness:1 history:4 visualizes:1 xiaoxiao:1 mohamed:2 associated:2 sampled:1 gain:3 dataset:4 recall:3 color:3 improves:3 enduro:6 higher:4 supervised:2 day:1 follow:1 maximally:1 improved:2 done:3 though:4 evaluated:1 generality:1 box:1 just:2 uct:56 until:3 correlation:1 hand:9 sketch:3 horizontal:3 assessment:1 incrementally:1 somehow:1 gray:3 believe:1 mdp:2 dqn:31 building:2 effect:1 hypothesized:2 requiring:3 true:1 former:1 hence:1 deal:2 distal:1 visualizing:1 round:1 game:77 during:8 bowling:2 covering:1 stone:1 plate:1 evident:2 theoretic:1 performs:2 bring:1 largman:1 elected:1 image:10 recently:1 ari:1 console:1 multinomial:3 rl:26 overview:1 handcrafted:2 extend:2 discussed:2 significant:6 refer:1 honglak:2 ai:2 similarly:1 stochasticity:1 nonlinearity:2 language:1 baveja:1 moving:2 access:4 specification:1 longer:1 surface:2 closest:3 recent:4 tesauro:1 schmidhuber:2 success:4 captured:2 seen:1 additional:1 impose:1 accomplishes:1 maximize:1 signal:1 ale:11 ii:1 multiple:1 desirable:1 needing:1 infer:1 reduces:1 exceeds:1 technical:1 match:3 faster:1 offer:1 coded:2 bigger:1 parenthesis:1 sponsor:1 prediction:1 scalable:1 regression:5 crop:1 controller:1 vision:3 yeung:1 represent:1 achieved:2 background:1 cropped:1 szepesv:1 addressed:1 bringing:1 effectiveness:1 near:1 unrewarded:1 bengio:3 enough:3 destroyed:1 automated:1 finish:1 psychology:1 architecture:11 perfectly:2 identified:1 ciresan:1 observability:3 idea:5 reduce:1 haffner:1 whether:1 motivated:1 war:1 utility:1 reuse:1 speech:4 compositional:1 action:46 deep:20 useful:4 clear:1 karpathy:1 amount:1 visualized:1 generate:2 outperform:2 percentage:2 nsf:1 dotted:1 per:1 platform:1 kill:1 broadly:1 naddaf:1 discrete:1 promise:1 express:1 key:3 four:15 preprocessed:1 dahl:1 ram:2 destroy:1 defect:1 sum:3 convert:1 run:11 fourteenth:1 powerful:1 named:1 throughout:1 reasonable:1 decide:1 reader:1 submarine:7 patch:3 realtime:5 home:1 decision:3 ucr:2 bit:1 entirely:1 layer:22 bound:1 interleaved:18 followed:1 courville:1 encountered:1 activity:1 strength:1 constraint:1 fei:2 nearby:1 generates:2 simulate:2 performing:1 relatively:1 department:1 structured:1 according:1 combination:4 across:2 rider:4 making:1 invariant:2 resource:1 visualization:9 discus:1 montecarlo:2 turn:1 needed:3 fed:1 antonoglou:1 umich:5 available:5 apply:1 observe:1 hierarchical:2 indirectly:1 responsive:1 subtracted:1 batch:1 save:2 shetty:1 slower:5 original:1 unreasonably:1 running:1 include:1 ensure:4 opportunity:2 exploit:2 build:6 especially:1 crafting:1 move:4 intend:2 already:2 receptive:1 dependence:1 usual:1 bagnell:1 evolutionary:2 gradient:1 subspace:1 seven:1 collected:1 reason:2 providing:1 rickl:1 ggp:2 potentially:1 motivates:2 policy:25 perform:3 allowing:3 observation:10 convolution:1 markov:2 benchmark:3 finite:1 beat:1 immediate:2 hinton:3 incorporated:1 communication:1 frame:19 mansour:1 bert:4 introduced:1 pair:4 required:1 palette:1 namely:1 meier:1 imagenet:1 acoustic:2 learned:18 purposefully:2 address:3 beyond:1 below:9 perception:14 usually:1 departure:2 pattern:7 challenge:9 including:1 memory:1 video:2 max:2 belief:3 greatest:1 unrealistic:1 improve:1 brief:2 mcts:2 text:1 nice:1 understanding:2 relative:1 graf:2 loss:1 fully:4 remarkable:1 foundation:1 contingency:1 awareness:1 agent:101 principle:1 emulator:3 playing:12 share:1 row:8 changed:1 supported:1 last:6 free:7 keeping:2 offline:1 bias:1 allow:1 understand:1 wide:1 template:1 face:1 taking:1 absolute:1 sparse:1 diver:8 depth:13 valid:1 world:1 evaluating:1 rich:2 transition:2 cumulative:1 computes:1 made:3 reinforcement:4 avoids:1 collection:1 preprocessing:3 author:1 far:3 erhan:1 miikkulainen:1 transaction:1 ranganath:1 observable:1 emphasize:1 obtains:1 satinder:1 confirm:1 spatio:3 imitation:1 don:1 search:5 decade:1 why:1 table:7 learn:5 reasonably:1 terminate:1 confirmation:1 unavailable:1 bottou:1 complex:2 necessarily:1 european:1 domain:4 zou:1 did:1 aistats:1 main:3 bounding:1 repeated:4 allowed:2 augmented:2 crafted:1 screen:6 grosse:1 slow:3 experienced:2 position:3 explicit:3 lie:1 perceptual:1 learns:3 interleaving:1 down:7 specific:5 rectifier:3 showing:1 concern:1 dl:14 workshop:1 magnitude:3 downward:1 gap:4 depicted:1 michigan:5 likely:2 expressed:1 saturating:1 partially:3 scalar:1 recommendation:1 applies:1 springer:1 corresponds:1 lewis:1 acm:2 goal:6 viewed:2 room:1 shared:1 content:1 change:3 specifically:4 kearns:1 called:1 total:1 experimental:1 player:11 rarely:1 select:8 highdimensional:1 guo:1 latter:1 arises:1 evaluate:1 audio:2 phenomenon:1 |
4,885 | 5,422 | On the Number of Linear Regions of
Deep Neural Networks
Guido Mont?ufar
Max Planck Institute for Mathematics in the Sciences
[email protected]
Razvan Pascanu
Universit?e de Montr?eal
[email protected]
Yoshua Bengio
Universit?e de Montr?eal, CIFAR Fellow
[email protected]
Kyunghyun Cho
Universit?e de Montr?eal
[email protected]
Abstract
We study the complexity of functions computable by deep feedforward neural networks
with piecewise linear activations in terms of the symmetries and the number of linear
regions that they have. Deep networks are able to sequentially map portions of each
layer?s input-space to the same output. In this way, deep models compute functions
that react equally to complicated patterns of different inputs. The compositional
structure of these functions enables them to re-use pieces of computation exponentially
often in terms of the network?s depth. This paper investigates the complexity of such
compositional maps and contributes new theoretical results regarding the advantage
of depth for neural networks with piecewise linear activation functions. In particular,
our analysis is not specific to a single family of models, and as an example, we employ
it for rectifier and maxout networks. We improve complexity bounds from pre-existing
work and investigate the behavior of units in higher layers.
Keywords: Deep learning, neural network, input space partition, rectifier, maxout
1 Introduction
Artificial neural networks with several hidden layers, called deep neural networks, have become popular
due to their unprecedented success in a variety of machine learning tasks (see, e.g., Krizhevsky et al.
2012, Ciresan et al. 2012, Goodfellow et al. 2013, Hinton et al. 2012). In view of this empirical evidence,
deep neural networks are becoming increasingly favored over shallow networks (i.e., with a single layer
of hidden units), and are often implemented with more than five layers. At the time being, however, the
theory of deep networks still poses many questions. Recently, Delalleau and Bengio (2011) showed that
a shallow network requires exponentially many more sum-product hidden units1 than a deep sum-product
network in order to compute certain families of polynomials. We are interested in extending this kind
of analysis to more popular neural networks, such as those with maxout and rectifier units.
There is a wealth of literature discussing approximation, estimation, and complexity of artificial neural
networks (see, e.g., Anthony and Bartlett 1999). A well-known result states that a feedforward neural
network with a single, huge, hidden layer is a universal approximator of Borel measurable functions (see
Hornik et al. 1989, Cybenko 1989). Other works have investigated universal approximation of probability
distributions by deep belief networks (Le Roux and Bengio 2010, Mont?ufar and Ay 2011), as well as
their approximation properties (Mont?ufar 2014, Krause et al. 2013).
These previous theoretical results, however, do not trivially apply to the types of deep neural networks
that have seen success in recent years. Conventional neural networks often employ either hidden units
1
A single sum-product hidden layer summarizes a layer of product units followed by a layer of sum units.
1
Figure 1: Binary classification using a shallow model with 20 hidden units (solid line) and a deep model
with two layers of 10 units each (dashed line). The right panel shows a close-up of the left panel. Filled
markers indicate errors made by the shallow model.
with a bounded smooth activation function, or Boolean hidden units. On the other hand, recently it has
become more common to use piecewise linear functions, such as the rectifier activation g(a) = max{0, a}
(Glorot et al. 2011, Nair and Hinton 2010) or the maxout activation g(a1, . . . , ak ) = max{a1, . . . , ak }
(Goodfellow et al. 2013). The practical success of deep neural networks with piecewise linear units calls
for the theoretical analysis specific for this type of neural networks.
In this respect, Pascanu et al. (2013) reported a theoretical result on the complexity of functions computable
by deep feedforward networks with rectifier units. They showed that, in the asymptotic limit of many
hidden layers, deep networks are able to separate their input space into exponentially more linear response
regions than their shallow counterparts, despite using the same number of computational units.
Building on the ideas from Pascanu et al. (2013), we develop a general framework for analyzing deep
models with piecewise linear activations. We describe how the intermediary layers of these models
are able to map several pieces of their inputs into the same output. The layer-wise composition of the
functions computed in this way re-uses low-level computations exponentially often as the number of
layers increases. This key property enables deep networks to compute highly complex and structured
functions. We underpin this idea by estimating the number of linear regions of functions computable by
two important types of piecewise linear networks: with rectifier units and with maxout units. Our results
for the complexity of deep rectifier networks yield a significant improvement over the previous results
on rectifier networks mentioned above, showing a favorable behavior of deep over shallow networks even
with a moderate number of hidden layers. Furthermore, our analysis of deep rectifier and maxout networks
provides a platform to study a broad variety of related networks, such as convolutional networks.
The number of linear regions of the functions that can be computed by a given model is a measure of the
model?s flexibility. An example of this is given in Fig. 1, which compares the learned decision boundary of a
single-layer and a two-layer model with the same number of hidden units (see details in the Supplementary
Material). This illustrates the advantage of depth; the deep model captures the desired boundary more
accurately, approximating it with a larger number of linear pieces. As noted earlier, deep networks are able
to identify an exponential number of input neighborhoods by mapping them to a common output of some
intermediary hidden layer. The computations carried out on the activations of this intermediary layer are
replicated many times, once in each of the identified neighborhoods. This allows the networks to compute
very complex looking functions even when they are defined with relatively few parameters. The number
of parameters is an upper bound for the dimension of the set of functions computable by a network, and
a small number of parameters means that the class of computable functions has a low dimension. The
set of functions computable by a deep feedforward piecewise linear network, although low dimensional,
achieves exponential complexity by re-using and composing features from layer to layer.
2 Feedforward Neural Networks and their Compositional Properties
In this section we discuss the ability of deep feedforward networks to re-map their input-space to create
complex symmetries by using only relatively few computational units. The key observation of our analysis
is that each layer of a deep model is able to map different regions of its input to a common output. This
leads to a compositional structure, where computations on higher layers are effectively replicated in all
input regions that produced the same output at a given layer. The capacity to replicate computations over
the input-space grows exponentially with the number of network layers. Before expanding these ideas, we
introduce basic definitions needed in the rest of the paper. At the end of this section, we give an intuitive
perspective for reasoning about the replicative capacity of deep models.
2
2.1 Definitions
A feedforward neural network is a composition of layers of computational units which defines a function
F : Rn0 ? Rout of the form
F (x; ?) = fout ? gL ? fL ? ? ? ? ? g1 ? f1(x),
(1)
where fl is a linear preactivation function and gl is a nonlinear activation function. The parameter ? is
composed of input weight matrices Wl ? Rk?nl ?nl?1 and bias vectors bl ? Rk?nl for each layer l ? [L].
>
The output of the l-th layer is a vector xl = [xl,1, . . . , xl,nl ] of activations xl,i of the units i ? [nl ] in
that layer. This is computed from the activations of the preceding layer by xl = gl (fl (xl?1)). Given the
activations xl?1 of the units in the (l ? 1)-th layer, the preactivation of layer l is given by
fl (xl?1) = Wlxl?1 + bl ,
>
where fl = [fl,1, . . . , fl,nl ] is an array composed of nl preactivation vectors fl,i ? Rk , and the activation
of the i-th unit in the l-th layer is given by
xl,i = gl,i(fl,i(xl?1)).
We will abbreviate gl ? fl by hl . When the layer index l is clear, we will drop the corresponding subscript.
We are interested in piecewise linear activations, and will consider the following two important types.
? Rectifier unit:
gi(fi) = max {0, fi}, where fi ? R and k = 1.
? Rank-k maxout unit: gi(fi) = max{fi,1, . . . , fi,k }, where fi = [fi,1, . . . , fi,k ] ? Rk .
The structure of the network refers to the way its units are arranged. It is specified by the number n0 of
input dimensions, the number of layers L, and the number of units or width nl of each layer.
We will classify the functions computed by different network structures, for different choices of parameters,
in terms of their number of linear regions. A linear region of a piecewise linear function F : Rn0 ? Rm
is a maximal connected subset of the input-space Rn0 , on which F is linear. For the functions that we
consider, each linear region has full dimension, n0.
2.2 Shallow Neural Networks
Rectifier units have two types of behavior; they can be either constant 0 or linear, depending on their
inputs. The boundary between these two behaviors is given by a hyperplane, and the collection of all
the hyperplanes coming from all units in a rectifier layer forms a hyperplane arrangement. In general,
if the activation function g : R ? R has a distinguished (i.e., irregular) behavior at zero (e.g., an inflection
point or non-linearity), then the function Rn0 ? Rn1 ; x ?
7 g(Wx + b) has a distinguished behavior at
all inputs from any of the hyperplanes Hi := {x ? Rn0 : Wi,:x + bi = 0} for i ? [n1]. The hyperplanes
capturing this distinguished behavior also form a hyperplane arrangement (see, e.g., Pascanu et al. 2013).
The hyperplanes in the arrangement split the input-space into several regions. Formally, a region of a
hyperplane arrangement {H1, . . . , Hn1 } is a connected component of the complement Rn0 \ (?iHi),
i.e., a set of points delimited by these hyperplanes (possibly open towards infinity). The number of regions
of an arrangement can be given in terms of a characteristic function of the arrangement, as P
shown in a
n0
n1
well-known result by Zaslavsky (1975). An arrangement of n1 hyperplanes in Rn0 has at most j=0
j
regions. Furthermore, this number of regions is attained if and only if the hyperplanes are in general
position. This implies that the maximal number of linear
of functions computed by a shallow
Pn0 regions
n1
rectifier network with n0 inputs and n1 hidden units is j=0
j (see Pascanu et al. 2013; Proposition 5).
2.3 Deep Neural Networks
We start by defining the identification of input neighborhoods mentioned in the introduction more formally:
Definition 1. A map F identifies two neighborhoods S and T of its input domain if it maps them to a common subset F (S) = F (T ) of its output domain. In this case we also say that S and T are identified by F .
Example 2. The four quadrants of 2-D Euclidean space are regions that are identified by the absolute
>
value function g : R2 ? R2; (x1, x2) ?
7 [|x1|, |x2|] .
3
3.
2. Fold along the
horizontal axis
1. Fold along the
vertical axis
(a)
Input Space
First Layer Space
S40 S10
S30 S20
S10 S40
S20 S30
S40 S10
S4 S1
S3 S2
S20
S10
S30
S40
S30 S20
S30
S40
S20
S10
Second Layer
Space
(b)
(c)
Figure 2: (a) Space folding of 2-D Euclidean space along the two coordinate axes. (b) An illustration of
how the top-level partitioning (on the right) is replicated to the original input space (left). (c) Identification
of regions across the layers of a deep model.
The computation carried out by the l-th layer of a feedforward network on a set of activations from the
(l ? 1)-th layer is effectively carried out for all regions of the input space that lead to the same activations
of the (l ? 1)-th layer. One can choose the input weights and biases of a given layer in such a way that
the computed function behaves most interestingly on those activation values of the preceding layer which
have the largest number of preimages in the input space, thus replicating the interesting computation many
times in the input space and generating an overall complicated-looking function.
For any given choice of the network parameters, each hidden layer l computes a function hl = gl ? fl on
the output activations of the preceding layer. We consider the function Fl : Rn0 ? Rnl ; Fl := hl ? ? ? ? ? h1
that computes the activations of the l-th hidden layer. We denote the image of Fl by Sl ? Rnl , i.e., the
set of (vector valued) activations reachable by the l-th layer for all possible inputs. Given a subset R ? Sl ,
? 1, . . . , R
? k ? Sl?1 that are mapped by hl onto R; that is, subsets
we denote by PRl the set of subsets R
? 1) = ? ? ? = hl (R
? k ) = R. See Fig. 2 for an illustration.
that satisfy hl (R
The number of separate input-space neighborhoods that are mapped to a common neighborhood
R ? Sl ? Rnl can be given recursively as
X
NRl =
NRl?1
NR0 = 1, for each region R ? Rn0 .
(2)
0 ,
l
R0 ?PR
For example, PR1 is the set of all disjoint input-space neighborhoods whose image by the function
computed by the first layer, h1 : x ?
7 g(Wx + b), equals R ? S1 ? Rn1 .
The recursive formula (2) counts the number of identified sets by moving along the branches of a tree
rooted at the set R of the j-th layer?s output-space (see Fig. 2 (c)). Based on these observations, we can
estimate the maximal number of linear regions as follows.
Lemma 3. The maximal number of linear regionsP
of the functions computed by an L-layer neural network
with piecewise linear activations is at least N = R?P L NRL?1, where NRL?1 is defined by Eq. (2), and
P L is a set of neighborhoods in distinct linear regions of the function computed by the last hidden layer.
Here, the idea to construct a function with many linear regions is to use the first L ? 1 hidden layers to
identify many input-space neighborhoods, mapping all of them to the activation neighborhoods P L of
the (L ? 1)-th hidden layer, each of which belongs to a distinct linear region of the last hidden layer. We
will follow this strategy in Secs. 3 and 4, where we analyze rectifier and maxout networks in detail.
2.4 Identification of Inputs as Space Foldings
In this section, we discuss an intuition behind Lemma 3 in terms of space folding. A map F that identifies
two subsets S and S 0 can be considered as an operator that folds its domain in such a way that the two
4
Figure 3: Space folding of 2-D space in a non-trivial way. Note how the folding can potentially identify
symmetries in the boundary that it needs to learn.
subsets S and S 0 coincide and are mapped to the same output. For instance, the absolute value function
g : R2 ? R2 from Example 2 folds its domain twice (once along each coordinate axis), as illustrated
in Fig. 2 (a). This folding identifies the four quadrants of 2-D Euclidean space. By composing such
operations, the same kind of map can be applied again to the output, in order to re-fold the first folding.
Each hidden layer of a deep neural network can be associated with a folding operator. Each hidden layer
folds the space of activations of the previous layer. In turn, a deep neural network effectively folds its
input-space recursively, starting with the first layer. The consequence of this recursive folding is that
any function computed on the final folded space will apply to all the collapsed subsets identified by the
map corresponding to the succession of foldings. This means that in a deep model any partitioning of
the last layer?s image-space is replicated in all input-space regions which are identified by the succession
of foldings. Fig. 2 (b) offers an illustration of this replication property.
Space foldings are not restricted to foldings along coordinate axes and they do not have to preserve lengths.
Instead, the space is folded depending on the orientations and shifts encoded in the input weights W and
biases b and on the nonlinear activation function used at each hidden layer. In particular, this means that the
sizes and orientations of identified input-space regions may differ from each other. See Fig. 3. In the case
of activation functions which are not piece-wise linear, the folding operations may be even more complex.
2.5 Stability to Perturbation
Our bounds on the complexity attainable by deep models (Secs. 3 and 4) are based on suitable choices
of the network weights. However, this does not mean that the indicated complexity is only attainable
in singular cases. The parametrization of the functions computed by a neural network is continuous.
More precisely, the map ? : RN ? C(Rn0 ; RnL ); ? ?
7 F? , which maps input weights and biases
n0
? = {Wi, bi}L
? RnL computed by the network, is continuous.
i=1 to the continuous functions F? : R
Our analysis considers the number of linear regions of the functions F? . By definition, each linear region
contains an open neighborhood of the input-space Rn0 . Given any function F? with a finite number
of linear regions, there is an > 0 such that for each -perturbation of the parameter ?, the resulting
function F?+ has at least as many linear regions as F? . The linear regions of F? are preserved under
small perturbations of the parameters, because they have a finite volume.
If we define a probability density on the space of parameters, what is the probability of the event that
the function represented by the network has a given number of linear regions? By the above discussion,
the probability of getting a number of regions at least as large as the number resulting from any particular
choice of parameters (for a uniform measure within a bounded domain) is nonzero, even though it may be
very small. This is because there exists an epsilon-ball of non-zero volume around that particular choice of
parameters, for which at least the same number of linear regions is attained. For example, shallow rectifier
networks generically attain the maximal number of regions, even if in close vicinity of any parameter
choice there may be parameters corresponding to functions with very few regions.
For future work it would be interesting to study the partitions of parameter space RN into pieces where
the resulting functions partition their input-spaces into isomorphic linear regions, and to investigate how
many of these pieces of parameter space correspond to functions with a given number of linear regions.
2.6 Empirical Evaluation of Folding in Rectifier MLPs
We empirically examined the behavior of a trained MLP to see if it folds the input-space in the way described
above. First, we note that tracing the activation of each hidden unit in this model gives a piecewise linear
map Rn0 ? R (from inputs to activation values of that unit). Hence, we can analyze the behavior of each
5
h2
?
h(x)
3
2
1
h1
0
1
h3
h1 ? h2 + h3
x
2
h1 ? h2
Figure 4: Folding of the real line into equal-length segments by a sum of rectifiers.
unit by visualizing the different weight matrices corresponding to the different linear pieces of this map. The
weight matrix of one piece of this map can be found by tracking the linear piece used in each intermediary
layer, starting from an input example. This visualization technique, a byproduct of our theoretical analysis,
is similar to the one proposed by Zeiler and Fergus (2013), but is motivated by a different perspective.
After computing the activations of an intermediary hidden unit for each training example, we can, for
instance, inspect two examples that result in similar levels of activation for a hidden unit. With the linear
maps of the hidden unit corresponding to the two examples we perturb one of the examples until it results
in exactly the same activation. These two inputs then can be safely considered as points in two regions
identified by the hidden unit. In the Supplementary Material we provide details and examples of this
visualization technique. We also show inputs identified by a deep MLP.
3 Deep Rectifier Networks
In this section we analyze deep neural networks with rectifier units, based on the general observations
from Sec. 2. We improve upon the results by Pascanu et al. (2013), with a tighter lower-bound on the
maximal number of linear regions of functions computable by deep rectifier networks. First, let us note the
following upper-bound, which follows directly from the fact that each linear region of a rectifier network
corresponds to a pattern of hidden units being active:
Proposition 4. The maximal number of linear regions of the functions computed by any rectifier network
with a total of N hidden units is bounded from above by 2N .
3.1 Illustration of the Construction
Consider a layer of n rectifiers with n0 input variables, where n ? n0. We partition the set of rectifier
units into n0 (non-overlapping) subsets of cardinality p = b n/n0 c and ignore the remainder units. Consider
the units in the j-th subset. We can choose their input weights and biases such that
h1(x) = max {0, wx} ,
h2(x) = max {0, 2wx ? 1} ,
h3(x) = max {0, 2wx ? 2} ,
..
.
hp(x) = max {0, 2wx ? (p ? 1)} ,
where w is a row vector with j-th entry equal to 1 and all other entries set to 0. The product wx selects
the j-th coordinate of x. Adding these rectifiers with alternating signs, we obtain following scalar function:
?j (x) = 1, ?1, 1, . . . , (?1)p?1 [h1(x), h2(x), h3(x), . . . , hp(x)]> .
h
(3)
?j acts only on the j-th input coordinate, we may redefine it to take a scalar input, namely the
Since h
j-th coordinate of x. This function has p linear regions given by the intervals (??, 0], [0, 1], [1, 2],
?j onto the interval (0, 1), as
. . . , [p ? 1, ?). Each of these intervals has a subset that is mapped by h
?
illustrated in Fig. 4. The function hj identifies the input-space strips with j-th coordinate xj restricted to
?=
the intervals (0, 1), (1, 2), . . . , (p ? 1, p). Consider now all the n0 subsets of rectifiers and the function h
>
?
?
?
h1, h2, . . . , hp . This function is locally symmetric about each hyperplane with a fixed j-th coordinate
6
equal to xj = 1, . . . , xj = p ? 1 (vertical lines in Fig. 4), for all j = 1, . . . , n0. Note the periodic pattern
? identifies a total of pn0 hypercubes delimited by these hyperplanes.
that emerges. In fact, the function h
? arises from h by composition with a linear function (alternating sums). This linear
Now, note that h
? as
function can be effectively absorbed in the preactivation function of the next layer. Hence we can treat h
being the function computed by the current layer. Computations by deeper layers, as functions of the unit
hypercube output of this rectifier layer, are replicated on each of the pn0 identified input-space hypercubes.
3.2 Formal Result
We can generalize the construction described above to the case of a deep rectifier network with n0 inputs
and L hidden layers of widths ni ? n0 for all i ? [L]. We obtain the following lower bound for the
maximal number of linear regions of deep rectifier networks:
Theorem 5. The maximal number of linear regions of the functions computed by a neural network with
n0 input units and L hidden layers, with ni ? n0 rectifiers at the i-th layer, is lower bounded by
! n
L?1
0
Y ni n0 X
nL
.
n
j
0
j=0
i=1
The next corollary gives an expression for the asymptotic behavior of these bounds. Assuming that
n0 = O(1) and ni = n for all i ? 1, the number of regions of a single layer model with Ln hidden units
behaves as O(Ln0 nn0 ) (see Pascanu et al. 2013; Proposition 10). For a deep model, Theorem 5 implies:
Corollary 6. A rectifier neural network with n0 input
units and L hidden layers of width n ? n0 can
(L?1)n0 n0
n
compute functions that have ? ( /n0 )
n
linear regions.
Thus we see that the number of linear regions of deep models grows exponentially in L and polynomially
in n, which is much faster than that
with nL hidden units. Our result is a significant
of shallow models
L?1 n0
improvement over the bound ? ( n/n0 )
n
obtained by Pascanu et al. (2013). In particular, our
result demonstrates that even for small values of L and n, deep rectifier models are able to produce
substantially more linear regions than shallow rectifier models. Additionally, using the same strategy
as Pascanu et al. (2013), our result can be reformulated in terms of the number of linear regions per
parameter. This results in a similar behavior, with deep models being exponentially more efficient than
shallow models (see the Supplementary Material).
4 Deep Maxout Networks
A maxout network is a feedforward network with layers defined as follows:
Definition 7. A rank-k maxout layer with n input and m output units is defined by a preactivation function
of the form f : Rn ? Rm?k ; f(x) = Wx+b, with input and bias weights W ? Rm?k?n, b ? Rm?k , and
activations of the form gj (z) = max{z(j?1)k+1, . . . , zjk } for all j ? [m]. The layer computes a function
?
?
max{f1(x), . . . , fk (x)}
?
?
..
g ? f : Rn ? Rm; x ?
7 ?
(4)
?.
.
max{f(m?1)k+1(x), . . . , fmk (x)}
Since the maximum of two convex functions is convex, maxout units and maxout layers compute convex
functions. The maximum of a collection of functions is called their upper envelope. We can view the graph
of each linear function fi : Rn ? R as a supporting hyperplane of a convex set in (n + 1)-dimensional
space. In particular, if each fi, i ? [k] is the unique maximizer fi = max{fi0 : i0 ? [k]} at some input
neighborhood, then the number of linear regions of the upper envelope g1 ? f = max{fi : i ? [k]} is
exactly k. This shows that the maximal number of linear regions of a maxout unit is equal to its rank.
The linear regions of the maxout layer are the intersections of the linear regions of the individual maxout
units. In order to obtain the number of linear regions for the layer, we need to describe the structure of
the linear regions of each maxout unit, and study their possible intersections. Voronoi diagrams can be
7
lifted to upper envelopes of linear functions, and hence they describe input-space partitions generated
by maxout units. Now, how many regions do we obtain by intersecting the regions of m Voronoi diagrams
with k regions each? Computing the intersections of Voronoi diagrams is not easy, in general. A trivial
upper bound for the number of linear regions is km, which corresponds to the case where all intersections
of regions of different units are different from each other. We will give a better bound in Proposition 8.
Now, for the purpose of computing lower bounds, here it will be sufficient to consider certain well-behaved
special cases. One simple example is the division of input-space by k?1 parallel hyperplanes. If m ? n, we
can consider the arrangement of hyperplanes Hi = {x ? Rn : xj = i} for i = 1, . . . , k ? 1, for each maxout unit j ? [m]. In this case, the number of regions is km. If m > n, the same arguments yield kn regions.
Proposition 8. The maximal number of regions of a single layer maxout network with n inputs and m
Pn
2
outputs of rank k is lower bounded by kmin{n,m} and upper bounded by min{ j=0 k jm , km}.
Now we take a look at the deep maxout model. Note that a rank-2 maxout layer can be simulated by a
rectifier layer with twice as many units. Then, by the results from the last section, a rank-2 maxout network
with L ? 1 hidden layers of width n = n0 can identify 2n0 (L?1) input-space regions, and, in turn, it can
compute functions with 2n0 (L?1)2n0 = 2n0 L linear regions. For the rank-k case, we note that a rank-k
maxout unit can identify k cones from its input-domain, whereby each cone is a neighborhood of the
positive half-ray {rWi ? Rn : r ? R+} corresponding to the gradient Wi of the linear function fi for
all i ? [k]. Elaborating this observation, we obtain:
Theorem 9. A maxout network with L layers of width n0 and rank k can compute functions with at least
kL?1kn0 linear regions.
Theorem 9 and Proposition 8 show that deep maxout networks can compute functions with a number of
linear regions that grows exponentially with the number of layers, and exponentially faster than the maximal
number of regions of shallow models with the same number of units. Similarly to the rectifier model, this
exponential behavior can also be established with respect to the number of network parameters. We note
that although certain functions that can be computed by maxout layers can also be computed by rectifier
layers, the rectifier construction from last section leads to functions that are not computable by maxout
networks (except in the rank-2 case). The proof of Theorem 9 is based on the same general arguments
from Sec. 2, but uses a different construction than Theorem 5 (details in the Supplementary Material).
5 Conclusions and Outlook
We studied the complexity of functions computable by deep feedforward neural networks in terms of their
number of linear regions. We specifically focused on deep neural networks having piecewise linear hidden
units which have been found to provide superior performance in many machine learning applications
recently. We discussed the idea that each layer of a deep model is able to identify pieces of its input in
such a way that the composition of layers identifies an exponential number of input regions. This results
in exponentially replicating the complexity of the functions computed in the higher layers. The functions
computed in this way by deep models are complicated, but still they have an intrinsic rigidity caused by
the replications, which may help deep models generalize to unseen samples better than shallow models.
This framework is applicable to any neural network that has a piecewise linear activation function. For
example, if we consider a convolutional network with rectifier units, as the one used in (Krizhevsky et al.
2012), we can see that the convolution followed by max pooling at each layer identifies all patches of the
input within a pooling region. This will let such a deep convolutional neural network recursively identify
patches of the images of lower layers, resulting in exponentially many linear regions of the input space.
The structure of the linear regions depends on the type of units, e.g., hyperplane arrangements for shallow
rectifier vs. Voronoi diagrams for shallow maxout networks. The pros and cons of each type of constraint
will likely depend on the task and are not easily quantifiable at this point. As for the number of regions,
in both maxout and rectifier networks we obtain an exponential increase with depth. However, our bounds
are not conclusive about which model is more powerful in this respect. This is an interesting question
that would be worth investigating in more detail.
The parameter space of a given network is partitioned into the regions where the resulting functions have
corresponding linear regions. The combinatorics of such structures is in general hard to compute, even for
simple hyperplane arrangements. One interesting question for future analysis is whether many regions of the
parameter space of a given network correspond to functions which have a given number of linear regions.
8
References
M. Anthony and P. Bartlett. Neural Network Learning: Theoretical Foundations. Cambridge University
Press, 1999.
D. Ciresan, U. Meier, J. Masci, and J. Schmidhuber. Multi column deep neural network for traffic sign
classification. Neural Networks, 32:333?338, 2012.
G. Cybenko. Approximation by superpositions of a sigmoidal function. Mathematics of Control, Signals
and Systems, 2(4):303?314, 1989.
O. Delalleau and Y. Bengio. Shallow vs. deep sum-product networks. In NIPS, 2011.
X. Glorot, A. Bordes, and Y. Bengio. Deep sparse rectifier neural networks. In AISTATS, 2011.
I. Goodfellow, D. Warde-Farley, M. Mirza, A. Courville, and Y. Bengio. Maxout networks. In Proc. 30th
International Conference on Machine Learning, pages 1319?1327, 2013.
G. Hinton, L. Deng, G. E. Dahl, A. Mohamed, N. Jaitly, A. Senior, V. Vanhoucke, P. Nguyen, T. Sainath,
and B. Kingsbury. Deep neural networks for acoustic modeling in speech recognition. IEEE Signal
Processing Magazine, 29(6):82?97, Nov. 2012.
K. Hornik, M. Stinchcombe, and H. White. Multilayer feedforward networks are universal approximators.
Neural Networks, 2:359?366, 1989.
O. Krause, A. Fischer, T. Glasmachers, and C. Igel. Approximation properties of DBNs with binary hidden
units and real-valued visible units. In Proc. 30th International Conference on Machine Learning, pages
419?426, 2013.
A. Krizhevsky, I. Sutskever, and G. Hinton. ImageNet classification with deep convolutional neural
networks. In NIPS, 2012.
N. Le Roux and Y. Bengio. Deep belief networks are compact universal approximators. Neural
Computation, 22(8):2192?2207, Aug. 2010.
G. Mont?ufar. Universal approximation depth and errors of narrow belief networks with discrete units.
Neural Computation, 26, July 2014.
G. Mont?ufar and N. Ay. Refinements of universal approximation results for deep belief networks and
restricted Boltzmann machines. Neural Computation, 23(5):1306?1319, May 2011.
V. Nair and G. E. Hinton. Rectified linear units improve restricted Boltzmann machines. In Proc. 27th
International Conference on Machine Learning, pages 807?814, 2010.
R. Pascanu and Y. Bengio. Revisiting natural gradient for deep networks. In International Conference
on Learning Representations, 2014.
R. Pascanu, G. Mont?ufar, and Y. Bengio. On the number of response regions of deep feed forward
networks with piece-wise linear activations. arXiv:1312.6098, Dec. 2013.
R. Stanley. An introduction to hyperplane arrangements. In Lect. notes, IAS/Park City Math. Inst., 2004.
J. Susskind, A. Anderson, and G. E. Hinton. The Toronto face dataset. Technical Report UTML TR
2010-001, U. Toronto, 2010.
T. Zaslavsky. Facing Up to Arrangements: Face-Count Formulas for Partitions of Space by Hyperplanes.
Number 154 in Memoirs of the American Mathematical Society. American Mathematical Society,
Providence, RI, 1975.
M. D. Zeiler and R. Fergus. Visualizing and understanding convolutional networks. arXiv:1311.2901,
2013.
9
| 5422 |@word polynomial:1 replicate:1 open:2 km:3 attainable:2 tr:1 outlook:1 solid:1 recursively:3 contains:1 interestingly:1 elaborating:1 existing:1 current:1 activation:33 visible:1 partition:6 wx:8 enables:2 utml:1 drop:1 n0:30 v:2 half:1 parametrization:1 provides:1 pascanu:11 math:1 toronto:2 hyperplanes:11 sigmoidal:1 five:1 kingsbury:1 along:6 mathematical:2 become:2 rnl:5 replication:2 redefine:1 ray:1 introduce:1 behavior:12 mpg:1 pr1:1 multi:1 jm:1 cardinality:1 estimating:1 bounded:6 linearity:1 panel:2 rn0:12 what:1 kind:2 substantially:1 safely:1 fellow:1 act:1 exactly:2 universit:3 rm:5 demonstrates:1 partitioning:2 unit:63 control:1 planck:1 before:1 positive:1 treat:1 limit:1 consequence:1 despite:1 ak:2 analyzing:1 subscript:1 becoming:1 twice:2 studied:1 examined:1 bi:2 igel:1 ihi:1 practical:1 unique:1 recursive:2 razvan:1 susskind:1 empirical:2 universal:6 attain:1 pre:1 refers:1 quadrant:2 onto:2 close:2 operator:2 collapsed:1 measurable:1 map:16 conventional:1 starting:2 sainath:1 convex:4 focused:1 roux:2 react:1 array:1 s40:5 stability:1 nn0:1 coordinate:8 construction:4 dbns:1 magazine:1 guido:1 us:2 goodfellow:3 jaitly:1 fout:1 recognition:1 capture:1 revisiting:1 region:80 connected:2 montufar:1 mentioned:2 intuition:1 complexity:11 warde:1 trained:1 depend:1 segment:1 upon:1 division:1 easily:1 represented:1 distinct:2 describe:3 lect:1 artificial:2 neighborhood:13 whose:1 encoded:1 supplementary:4 larger:1 valued:2 delalleau:2 say:1 ability:1 gi:2 g1:2 unseen:1 fischer:1 final:1 advantage:2 unprecedented:1 product:6 maximal:12 coming:1 remainder:1 flexibility:1 fi0:1 intuitive:1 getting:1 quantifiable:1 sutskever:1 extending:1 produce:1 generating:1 help:1 depending:2 develop:1 pose:1 h3:4 keywords:1 aug:1 eq:1 implemented:1 indicate:1 implies:2 differ:1 material:4 fmk:1 glasmachers:1 f1:2 cybenko:2 proposition:6 tighter:1 around:1 considered:2 mapping:2 achieves:1 purpose:1 estimation:1 favorable:1 intermediary:5 applicable:1 proc:3 superposition:1 largest:1 wl:1 create:1 city:1 rwi:1 pn:1 hj:1 lifted:1 corollary:2 ax:2 improvement:2 rank:10 inflection:1 inst:1 voronoi:4 i0:1 hidden:37 interested:2 selects:1 overall:1 classification:3 orientation:2 favored:1 platform:1 special:1 equal:5 once:2 construct:1 having:1 broad:1 look:1 park:1 future:2 report:1 yoshua:2 mirza:1 piecewise:13 employ:2 few:3 composed:2 preserve:1 individual:1 n1:5 montr:3 huge:1 mlp:2 investigate:2 highly:1 evaluation:1 generically:1 nl:10 farley:1 behind:1 byproduct:1 filled:1 tree:1 euclidean:3 re:5 desired:1 theoretical:6 instance:2 eal:3 earlier:1 boolean:1 classify:1 column:1 modeling:1 subset:12 entry:2 uniform:1 krizhevsky:3 reported:1 kn:1 providence:1 periodic:1 cho:2 hypercubes:2 pn0:3 density:1 international:4 intersecting:1 again:1 hn1:1 rn1:2 choose:2 possibly:1 american:2 de:4 sec:4 satisfy:1 combinatorics:1 caused:1 depends:1 piece:11 view:2 h1:9 analyze:3 traffic:1 portion:1 start:1 s30:5 complicated:3 parallel:1 mlps:1 ni:4 convolutional:5 characteristic:1 succession:2 yield:2 identify:7 correspond:2 generalize:2 identification:3 accurately:1 produced:1 worth:1 rectified:1 strip:1 nr0:1 definition:5 mohamed:1 associated:1 mi:1 proof:1 con:1 dataset:1 popular:2 emerges:1 stanley:1 feed:1 higher:3 delimited:2 attained:2 follow:1 response:2 arranged:1 though:1 anderson:1 furthermore:2 until:1 hand:1 horizontal:1 nonlinear:2 marker:1 overlapping:1 maximizer:1 defines:1 indicated:1 behaved:1 grows:3 zjk:1 building:1 counterpart:1 vicinity:1 kyunghyun:2 hence:3 alternating:2 symmetric:1 nonzero:1 illustrated:2 white:1 visualizing:2 zaslavsky:2 width:5 rooted:1 noted:1 whereby:1 ay:2 pro:1 reasoning:1 image:4 wise:3 recently:3 umontreal:3 fi:14 common:5 superior:1 behaves:2 empirically:1 exponentially:11 volume:2 discussed:1 significant:2 composition:4 cambridge:1 trivially:1 mathematics:2 hp:3 fk:1 similarly:1 replicating:2 reachable:1 moving:1 gj:1 showed:2 recent:1 perspective:2 moderate:1 belongs:1 schmidhuber:1 certain:3 binary:2 success:3 discussing:1 preimages:1 approximators:2 seen:1 preceding:3 deng:1 r0:1 dashed:1 signal:2 branch:1 full:1 july:1 smooth:1 technical:1 faster:2 offer:1 cifar:1 equally:1 a1:2 basic:1 multilayer:1 arxiv:2 dec:1 irregular:1 folding:16 preserved:1 krause:2 kmin:1 interval:4 wealth:1 singular:1 diagram:4 envelope:3 rest:1 pooling:2 call:1 prl:1 feedforward:11 bengio:10 split:1 easy:1 variety:2 xj:4 nrl:4 ciresan:2 identified:10 regarding:1 idea:5 computable:9 shift:1 whether:1 motivated:1 expression:1 bartlett:2 ufar:6 reformulated:1 speech:1 compositional:4 deep:60 clear:1 s4:1 locally:1 sl:4 s3:1 sign:2 disjoint:1 per:1 discrete:1 key:2 four:2 dahl:1 graph:1 sum:7 year:1 rout:1 cone:2 powerful:1 family:2 patch:2 decision:1 summarizes:1 investigates:1 capturing:1 layer:92 bound:12 fl:14 followed:2 hi:2 courville:1 fold:8 infinity:1 s10:5 precisely:1 constraint:1 x2:2 ri:1 argument:2 min:1 relatively:2 structured:1 ball:1 across:1 increasingly:1 wi:3 partitioned:1 shallow:17 s1:2 hl:6 restricted:4 pr:1 ln:1 visualization:2 discus:2 count:2 turn:2 needed:1 end:1 operation:2 apply:2 distinguished:3 original:1 top:1 zeiler:2 epsilon:1 perturb:1 approximating:1 hypercube:1 society:2 bl:2 question:3 arrangement:12 strategy:2 gradient:2 separate:2 mapped:4 simulated:1 capacity:2 considers:1 trivial:2 iro:1 assuming:1 length:2 index:1 illustration:4 potentially:1 underpin:1 boltzmann:2 upper:7 vertical:2 observation:4 inspect:1 convolution:1 finite:2 supporting:1 defining:1 hinton:6 looking:2 rn:7 perturbation:3 complement:1 namely:1 meier:1 specified:1 kl:1 conclusive:1 imagenet:1 s20:5 acoustic:1 learned:1 narrow:1 established:1 nip:2 able:7 pattern:3 max:15 belief:4 stinchcombe:1 ia:1 suitable:1 event:1 natural:1 abbreviate:1 improve:3 identifies:7 axis:3 carried:3 literature:1 understanding:1 asymptotic:2 interesting:4 approximator:1 facing:1 h2:6 foundation:1 vanhoucke:1 sufficient:1 bordes:1 row:1 gl:6 mont:6 last:5 preactivation:5 bias:6 formal:1 deeper:1 senior:1 institute:1 face:2 absolute:2 sparse:1 tracing:1 boundary:4 depth:5 dimension:4 computes:3 forward:1 made:1 collection:2 replicated:5 coincide:1 refinement:1 nguyen:1 polynomially:1 nov:1 compact:1 ignore:1 sequentially:1 active:1 investigating:1 fergus:2 continuous:3 additionally:1 learn:1 ca:3 composing:2 expanding:1 symmetry:3 contributes:1 hornik:2 investigated:1 complex:4 anthony:2 domain:6 aistats:1 s2:1 x1:2 kn0:1 fig:8 borel:1 position:1 exponential:5 xl:10 masci:1 rk:4 formula:2 theorem:6 specific:2 rectifier:41 showing:1 r2:4 evidence:1 glorot:2 exists:1 intrinsic:1 adding:1 effectively:4 illustrates:1 intersection:4 likely:1 absorbed:1 tracking:1 scalar:2 corresponds:2 nair:2 towards:1 maxout:31 hard:1 folded:2 except:1 specifically:1 hyperplane:9 lemma:2 called:2 total:2 isomorphic:1 formally:2 arises:1 rigidity:1 |
4,886 | 5,423 | Generative Adversarial Nets
Ian J. Goodfellow?, Jean Pouget-Abadie?, Mehdi Mirza, Bing Xu, David Warde-Farley,
Sherjil Ozair?, Aaron Courville, Yoshua Bengio?
D?epartement d?informatique et de recherche op?erationnelle
Universit?e de Montr?eal
Montr?eal, QC H3C 3J7
Abstract
We propose a new framework for estimating generative models via an adversarial process, in which we simultaneously train two models: a generative model G
that captures the data distribution, and a discriminative model D that estimates
the probability that a sample came from the training data rather than G. The training procedure for G is to maximize the probability of D making a mistake. This
framework corresponds to a minimax two-player game. In the space of arbitrary
functions G and D, a unique solution exists, with G recovering the training data
distribution and D equal to 21 everywhere. In the case where G and D are defined
by multilayer perceptrons, the entire system can be trained with backpropagation.
There is no need for any Markov chains or unrolled approximate inference networks during either training or generation of samples. Experiments demonstrate
the potential of the framework through qualitative and quantitative evaluation of
the generated samples.
1
Introduction
The promise of deep learning is to discover rich, hierarchical models [2] that represent probability
distributions over the kinds of data encountered in artificial intelligence applications, such as natural
images, audio waveforms containing speech, and symbols in natural language corpora. So far, the
most striking successes in deep learning have involved discriminative models, usually those that
map a high-dimensional, rich sensory input to a class label [14, 20]. These striking successes have
primarily been based on the backpropagation and dropout algorithms, using piecewise linear units
[17, 8, 9] which have a particularly well-behaved gradient . Deep generative models have had less
of an impact, due to the difficulty of approximating many intractable probabilistic computations that
arise in maximum likelihood estimation and related strategies, and due to difficulty of leveraging
the benefits of piecewise linear units in the generative context. We propose a new generative model
estimation procedure that sidesteps these difficulties. 1
In the proposed adversarial nets framework, the generative model is pitted against an adversary: a
discriminative model that learns to determine whether a sample is from the model distribution or the
data distribution. The generative model can be thought of as analogous to a team of counterfeiters,
trying to produce fake currency and use it without detection, while the discriminative model is
analogous to the police, trying to detect the counterfeit currency. Competition in this game drives
both teams to improve their methods until the counterfeits are indistiguishable from the genuine
articles.
?
Ian Goodfellow is now a research scientist at Google, but did this work earlier as a UdeM student
Jean Pouget-Abadie did this work while visiting Universit?e de Montr?eal from Ecole Polytechnique.
?
Sherjil Ozair is visiting Universit?e de Montr?eal from Indian Institute of Technology Delhi
?
Yoshua Bengio is a CIFAR Senior Fellow.
1
All code and hyperparameters available at http://www.github.com/goodfeli/adversarial
?
1
This framework can yield specific training algorithms for many kinds of model and optimization
algorithm. In this article, we explore the special case when the generative model generates samples
by passing random noise through a multilayer perceptron, and the discriminative model is also a
multilayer perceptron. We refer to this special case as adversarial nets. In this case, we can train
both models using only the highly successful backpropagation and dropout algorithms [16] and
sample from the generative model using only forward propagation. No approximate inference or
Markov chains are necessary.
2
Related work
Until recently, most work on deep generative models focused on models that provided a parametric
specification of a probability distribution function. The model can then be trained by maximizing the log likelihood. In this family of model, perhaps the most succesful is the deep Boltzmann
machine [25]. Such models generally have intractable likelihood functions and therefore require
numerous approximations to the likelihood gradient. These difficulties motivated the development
of ?generative machines??models that do not explicitly represent the likelihood, yet are able to generate samples from the desired distribution. Generative stochastic networks [4] are an example of
a generative machine that can be trained with exact backpropagation rather than the numerous approximations required for Boltzmann machines. This work extends the idea of a generative machine
by eliminating the Markov chains used in generative stochastic networks.
Our work backpropagates derivatives through generative processes by using the observation that
lim ?x E?N (0,?2 I) f (x + ) = ?x f (x).
??0
We were unaware at the time we developed this work that Kingma and Welling [18] and Rezende
et al. [23] had developed more general stochastic backpropagation rules, allowing one to backpropagate through Gaussian distributions with finite variance, and to backpropagate to the covariance
parameter as well as the mean. These backpropagation rules could allow one to learn the conditional variance of the generator, which we treated as a hyperparameter in this work. Kingma and
Welling [18] and Rezende et al. [23] use stochastic backpropagation to train variational autoencoders (VAEs). Like generative adversarial networks, variational autoencoders pair a differentiable
generator network with a second neural network. Unlike generative adversarial networks, the second network in a VAE is a recognition model that performs approximate inference. GANs require
differentiation through the visible units, and thus cannot model discrete data, while VAEs require
differentiation through the hidden units, and thus cannot have discrete latent variables. Other VAElike approaches exist [12, 22] but are less closely related to our method.
Previous work has also taken the approach of using a discriminative criterion to train a generative
model [29, 13]. These approaches use criteria that are intractable for deep generative models. These
methods are difficult even to approximate for deep models because they involve ratios of probabilities which cannot be approximated using variational approximations that lower bound the probability. Noise-contrastive estimation (NCE) [13] involves training a generative model by learning the
weights that make the model useful for discriminating data from a fixed noise distribution. Using a
previously trained model as the noise distribution allows training a sequence of models of increasing
quality. This can be seen as an informal competition mechanism similar in spirit to the formal competition used in the adversarial networks game. The key limitation of NCE is that its ?discriminator?
is defined by the ratio of the probability densities of the noise distribution and the model distribution,
and thus requires the ability to evaluate and backpropagate through both densities.
Some previous work has used the general concept of having two neural networks compete. The most
relevant work is predictability minimization [26]. In predictability minimization, each hidden unit
in a neural network is trained to be different from the output of a second network, which predicts
the value of that hidden unit given the value of all of the other hidden units. This work differs from
predictability minimization in three important ways: 1) in this work, the competition between the
networks is the sole training criterion, and is sufficient on its own to train the network. Predictability
minimization is only a regularizer that encourages the hidden units of a neural network to be statistically independent while they accomplish some other task; it is not a primary training criterion.
2) The nature of the competition is different. In predictability minimization, two networks? outputs
are compared, with one network trying to make the outputs similar and the other trying to make the
2
outputs different. The output in question is a single scalar. In GANs, one network produces a rich,
high dimensional vector that is used as the input to another network, and attempts to choose an input
that the other network does not know how to process. 3) The specification of the learning process
is different. Predictability minimization is described as an optimization problem with an objective
function to be minimized, and learning approaches the minimum of the objective function. GANs
are based on a minimax game rather than an optimization problem, and have a value function that
one agent seeks to maximize and the other seeks to minimize. The game terminates at a saddle point
that is a minimum with respect to one player?s strategy and a maximum with respect to the other
player?s strategy.
Generative adversarial networks has been sometimes confused with the related concept of ?adversarial examples? [28]. Adversarial examples are examples found by using gradient-based optimization
directly on the input to a classification network, in order to find examples that are similar to the
data yet misclassified. This is different from the present work because adversarial examples are
not a mechanism for training a generative model. Instead, adversarial examples are primarily an
analysis tool for showing that neural networks behave in intriguing ways, often confidently classifying two images differently with high confidence even though the difference between them is
imperceptible to a human observer. The existence of such adversarial examples does suggest that
generative adversarial network training could be inefficient, because they show that it is possible to
make modern discriminative networks confidently recognize a class without emulating any of the
human-perceptible attributes of that class.
3
Adversarial nets
The adversarial modeling framework is most straightforward to apply when the models are both
multilayer perceptrons. To learn the generator?s distribution pg over data x, we define a prior on
input noise variables pz (z), then represent a mapping to data space as G(z; ?g ), where G is a
differentiable function represented by a multilayer perceptron with parameters ?g . We also define a
second multilayer perceptron D(x; ?d ) that outputs a single scalar. D(x) represents the probability
that x came from the data rather than pg . We train D to maximize the probability of assigning the
correct label to both training examples and samples from G. We simultaneously train G to minimize
log(1 ? D(G(z))). In other words, D and G play the following two-player minimax game with
value function V (G, D):
min max V (D, G) = Ex?pdata (x) [log D(x)] + Ez?pz (z) [log(1 ? D(G(z)))].
(1)
G
D
In the next section, we present a theoretical analysis of adversarial nets, essentially showing that
the training criterion allows one to recover the data generating distribution as G and D are given
enough capacity, i.e., in the non-parametric limit. See Figure 1 for a less formal, more pedagogical
explanation of the approach. In practice, we must implement the game using an iterative, numerical
approach. Optimizing D to completion in the inner loop of training is computationally prohibitive,
and on finite datasets would result in overfitting. Instead, we alternate between k steps of optimizing
D and one step of optimizing G. This results in D being maintained near its optimal solution, so
long as G changes slowly enough. The procedure is formally presented in Algorithm 1.
In practice, equation 1 may not provide sufficient gradient for G to learn well. Early in learning,
when G is poor, D can reject samples with high confidence because they are clearly different from
the training data. In this case, log(1 ? D(G(z))) saturates. Rather than training G to minimize
log(1 ? D(G(z))) we can train G to maximize log D(G(z)). This objective function results in the
same fixed point of the dynamics of G and D but provides much stronger gradients early in learning.
4
Theoretical Results
The generator G implicitly defines a probability distribution pg as the distribution of the samples
G(z) obtained when z ? pz . Therefore, we would like Algorithm 1 to converge to a good estimator
of pdata , if given enough capacity and training time. The results of this section are done in a nonparametric setting, e.g. we represent a model with infinite capacity by studying convergence in the
space of probability density functions.
We will show in section 4.1 that this minimax game has a global optimum for pg = pdata . We will
then show in section 4.2 that Algorithm 1 optimizes Eq 1, thus obtaining the desired result.
3
...
x
XXX
z
Z Z Z
(a)
(b)
(c)
(d)
Figure 1: Generative adversarial nets are trained by simultaneously updating the discriminative distribution
(D, blue, dashed line) so that it discriminates between samples from the data generating distribution (black,
dotted line) px from those of the generative distribution pg (G) (green, solid line). The lower horizontal line is
the domain from which z is sampled, in this case uniformly. The horizontal line above is part of the domain
of x. The upward arrows show how the mapping x = G(z) imposes the non-uniform distribution pg on
transformed samples. G contracts in regions of high density and expands in regions of low density of pg . (a)
Consider an adversarial pair near convergence: pg is similar to pdata and D is a partially accurate classifier.
(b) In the inner loop of the algorithm D is trained to discriminate samples from data, converging to D? (x) =
pdata (x)
. (c) After an update to G, gradient of D has guided G(z) to flow to regions that are more likely
pdata (x)+pg (x)
to be classified as data. (d) After several steps of training, if G and D have enough capacity, they will reach a
point at which both cannot improve because pg = pdata . The discriminator is unable to differentiate between
the two distributions, i.e. D(x) = 12 .
Algorithm 1 Minibatch stochastic gradient descent training of generative adversarial nets. The number of
steps to apply to the discriminator, k, is a hyperparameter. We used k = 1, the least expensive option, in our
experiments.
for number of training iterations do
for k steps do
? Sample minibatch of m noise samples {z (1) , . . . , z (m) } from noise prior pg (z).
? Sample minibatch of m examples {x(1) , . . . , x(m) } from data generating distribution
pdata (x).
? Update the discriminator by ascending its stochastic gradient:
m
? ?d
i
1 Xh
log D x(i) + log 1 ? D G z (i)
.
m i=1
end for
? Sample minibatch of m noise samples {z (1) , . . . , z (m) } from noise prior pg (z).
? Update the generator by descending its stochastic gradient:
m
?? g
1 X
log 1 ? D G z (i)
.
m i=1
end for
The gradient-based updates can use any standard gradient-based learning rule. We used momentum in our experiments.
4.1
Global Optimality of pg = pdata
We first consider the optimal discriminator D for any given generator G.
Proposition 1. For G fixed, the optimal discriminator D is
?
DG
(x) =
pdata (x)
pdata (x) + pg (x)
4
(2)
Proof. The training criterion for the discriminator D, given any generator G, is to maximize the
quantity V (G, D)
Z
Z
V (G, D) = pdata (x) log(D(x))dx + pz (z) log(1 ? D(g(z)))dz
z
Zx
= pdata (x) log(D(x)) + pg (x) log(1 ? D(x))dx
(3)
x
For any (a, b) ? R2 \ {0, 0}, the function y ? a log(y) + b log(1 ? y) achieves its maximum in
a
[0, 1] at a+b
. The discriminator does not need to be defined outside of Supp(pdata ) ? Supp(pg ),
concluding the proof.
Note that the training objective for D can be interpreted as maximizing the log-likelihood for estimating the conditional probability P (Y = y|x), where Y indicates whether x comes from pdata
(with y = 1) or from pg (with y = 0). The minimax game in Eq. 1 can now be reformulated as:
C(G) = max V (G, D)
D
?
?
=Ex?pdata [log DG
(x)] + Ez?pz [log(1 ? DG
(G(z)))]
(4)
?
?
=Ex?pdata [log DG (x)] + Ex?pg [log(1 ? DG
(x))]
pg (x)
pdata (x)
=Ex?pdata log
+ Ex?pg log
Pdata (x) + pg (x)
pdata (x) + pg (x)
Theorem 1. The global minimum of the virtual training criterion C(G) is achieved if and only if
pg = pdata . At that point, C(G) achieves the value ? log 4.
?
?
Proof. For pg = pdata , DG
(x) = 12 , (consider Eq. 2). Hence, by inspecting Eq. 4 at DG
(x) = 21 , we
1
1
find C(G) = log 2 + log 2 = ? log 4. To see that this is the best possible value of C(G), reached
only for pg = pdata , observe that
Ex?pdata [? log 2] + Ex?pg [? log 2] = ? log 4
?
, G), we obtain:
and that by subtracting this expression from C(G) = V (DG
pdata + pg
pdata + pg
C(G) = ? log(4) + KL pdata
+
KL
p
(5)
g
2
2
where KL is the Kullback?Leibler divergence. We recognize in the previous expression the Jensen?
Shannon divergence between the model?s distribution and the data generating process:
C(G) = ? log(4) + 2 ? JSD (pdata kpg )
(6)
Since the Jensen?Shannon divergence between two distributions is always non-negative, and zero
iff they are equal, we have shown that C ? = ? log(4) is the global minimum of C(G) and that the
only solution is pg = pdata , i.e., the generative model perfectly replicating the data distribution.
4.2
Convergence of Algorithm 1
Proposition 2. If G and D have enough capacity, and at each step of Algorithm 1, the discriminator
is allowed to reach its optimum given G, and pg is updated so as to improve the criterion
?
?
Ex?pdata [log DG
(x)] + Ex?pg [log(1 ? DG
(x))]
then pg converges to pdata
Proof. Consider V (G, D) = U (pg , D) as a function of pg as done in the above criterion. Note
that U (pg , D) is convex in pg . The subderivatives of a supremum of convex functions include the
derivative of the function at the point where the maximum is attained. In other words, if f (x) =
sup??A f? (x) and f? (x) is convex in x for every ?, then ?f? (x) ? ?f if ? = arg sup??A f? (x).
This is equivalent to computing a gradient descent update for pg at the optimal D given the corresponding G. supD U (pg , D) is convex in pg with a unique global optima as proven in Thm 1,
therefore with sufficiently small updates of pg , pg converges to px , concluding the proof.
In practice, adversarial nets represent a limited family of pg distributions via the function G(z; ?g ),
and we optimize ?g rather than pg itself, so the proofs do not apply. However, the excellent performance of multilayer perceptrons in practice suggests that they are a reasonable model to use despite
their lack of theoretical guarantees.
5
Model
DBN [3]
Stacked CAE [3]
Deep GSN [5]
Adversarial nets
MNIST
138 ? 2
121 ? 1.6
214 ? 1.1
225 ? 2
TFD
1909 ? 66
2110 ? 50
1890 ? 29
2057 ? 26
Table 1: Parzen window-based log-likelihood estimates. The reported numbers on MNIST are the mean loglikelihood of samples on test set, with the standard error of the mean computed across examples. On TFD, we
computed the standard error across folds of the dataset, with a different ? chosen using the validation set of
each fold. On TFD, ? was cross validated on each fold and mean log-likelihood on each fold were computed.
For MNIST we compare against other models of the real-valued (rather than binary) version of dataset.
5
Experiments
We trained adversarial nets an a range of datasets including MNIST[21], the Toronto Face Database
(TFD) [27], and CIFAR-10 [19]. The generator nets used a mixture of rectifier linear activations [17,
8] and sigmoid activations, while the discriminator net used maxout [9] activations. Dropout [16]
was applied in training the discriminator net. While our theoretical framework permits the use of
dropout and other noise at intermediate layers of the generator, we used noise as the input to only
the bottommost layer of the generator network.
We estimate probability of the test set data under pg by fitting a Gaussian Parzen window to the
samples generated with G and reporting the log-likelihood under this distribution. The ? parameter
of the Gaussians was obtained by cross validation on the validation set. This procedure was introduced in Breuleux et al. [7] and used for various generative models for which the exact likelihood
is not tractable [24, 3, 4]. Results are reported in Table 1. This method of estimating the likelihood
has somewhat high variance and does not perform well in high dimensional spaces but it is the best
method available to our knowledge. Advances in generative models that can sample but not estimate
likelihood directly motivate further research into how to evaluate such models. In Figures 2 and 3
we show samples drawn from the generator net after training. While we make no claim that these
samples are better than samples generated by existing methods, we believe that these samples are at
least competitive with the better generative models in the literature and highlight the potential of the
adversarial framework.
6
Advantages and disadvantages
This new framework comes with advantages and disadvantages relative to previous modeling frameworks. The disadvantages are primarily that there is no explicit representation of pg (x), and that D
must be synchronized well with G during training (in particular, G must not be trained too much
without updating D, in order to avoid ?the Helvetica scenario? in which G collapses too many values
of z to the same value of x to have enough diversity to model pdata ), much as the negative chains of a
Boltzmann machine must be kept up to date between learning steps. The advantages are that Markov
chains are never needed, only backprop is used to obtain gradients, no inference is needed during
learning, and a wide variety of functions can be incorporated into the model. Table 2 summarizes
the comparison of generative adversarial nets with other generative modeling approaches.
The aforementioned advantages are primarily computational. Adversarial models may also gain
some statistical advantage from the generator network not being updated directly with data examples, but only with gradients flowing through the discriminator. This means that components of the
input are not copied directly into the generator?s parameters. Another advantage of adversarial networks is that they can represent very sharp, even degenerate distributions, while methods based on
Markov chains require that the distribution be somewhat blurry in order for the chains to be able to
mix between modes.
7
Conclusions and future work
This framework admits many straightforward extensions:
6
a)
b)
c)
d)
Figure 2: Visualization of samples from the model. Rightmost column shows the nearest training example of
the neighboring sample, in order to demonstrate that the model has not memorized the training set. Samples
are fair random draws, not cherry-picked. Unlike most other visualizations of deep generative models, these
images show actual samples from the model distributions, not conditional means given samples of hidden units.
Moreover, these samples are uncorrelated because the sampling process does not depend on Markov chain
mixing. a) MNIST b) TFD c) CIFAR-10 (fully connected model) d) CIFAR-10 (convolutional discriminator
and ?deconvolutional? generator)
Figure 3: Digits obtained by linearly interpolating between coordinates in z space of the full model.
1. A conditional generative model p(x | c) can be obtained by adding c as input to both G and D.
2. Learned approximate inference can be performed by training an auxiliary network to predict z
given x. This is similar to the inference net trained by the wake-sleep algorithm [15] but with
the advantage that the inference net may be trained for a fixed generator net after the generator
net has finished training.
3. One can approximately model all conditionals p(xS | x6S ) where S is a subset of the indices
of x by training a family of conditional models that share parameters. Essentially, one can use
adversarial nets to implement a stochastic extension of the deterministic MP-DBM [10].
4. Semi-supervised learning: features from the discriminator or inference net could improve performance of classifiers when limited labeled data is available.
5. Efficiency improvements: training could be accelerated greatly by devising better methods for
coordinating G and D or determining better distributions to sample z from during training.
This paper has demonstrated the viability of the adversarial modeling framework, suggesting that
these research directions could prove useful.
7
Deep directed
graphical models
Deep undirected
graphical models
Inference needed
during training.
MCMC needed to
approximate
partition function
gradient.
Generative
autoencoders
Adversarial models
Enforced tradeoff
between mixing
and power of
reconstruction
generation
Synchronizing the
discriminator with
the generator.
Helvetica.
Learned
approximate
inference
Training
Inference needed
during training.
Inference
Learned
approximate
inference
Variational
inference
MCMC-based
inference
Sampling
No difficulties
Requires Markov
chain
Evaluating p(x)
Intractable, may be
approximated with
AIS
Intractable, may be
approximated with
AIS
Requires Markov
chain
Not explicitly
represented, may be
approximated with
Parzen density
estimation
Model design
Models need to be
designed to work
with the desired
inference scheme
? some inference
schemes support
similar model
families as GANs
Careful design
needed to ensure
multiple properties
Any differentiable
function is
theoretically
permitted
No difficulties
Not explicitly
represented, may be
approximated with
Parzen density
estimation
Any differentiable
function is
theoretically
permitted
Table 2: Challenges in generative modeling: a summary of the difficulties encountered by different approaches
to deep generative modeling for each of the major operations involving a model.
Acknowledgments
We would like to acknowledge Patrice Marcotte, Olivier Delalleau, Kyunghyun Cho, Guillaume
Alain and Jason Yosinski for helpful discussions. Yann Dauphin shared his Parzen window evaluation code with us. We would like to thank the developers of Pylearn2 [11] and Theano [6, 1],
particularly Fr?ed?eric Bastien who rushed a Theano feature specifically to benefit this project. Arnaud Bergeron provided much-needed support with LATEX typesetting. We would also like to thank
CIFAR, and Canada Research Chairs for funding, and Compute Canada, and Calcul Qu?ebec for
providing computational resources. Ian Goodfellow is supported by the 2013 Google Fellowship in
Deep Learning. Finally, we would like to thank Les Trois Brasseurs for stimulating our creativity.
References
[1] Bastien, F., Lamblin, P., Pascanu, R., Bergstra, J., Goodfellow, I. J., Bergeron, A., Bouchard, N., and
Bengio, Y. (2012). Theano: new features and speed improvements. Deep Learning and Unsupervised
Feature Learning NIPS 2012 Workshop.
[2] Bengio, Y. (2009). Learning deep architectures for AI. Now Publishers.
[3] Bengio, Y., Mesnil, G., Dauphin, Y., and Rifai, S. (2013). Better mixing via deep representations. In
ICML?13.
[4] Bengio, Y., Thibodeau-Laufer, E., and Yosinski, J. (2014a). Deep generative stochastic networks trainable
by backprop. In ICML?14.
[5] Bengio, Y., Thibodeau-Laufer, E., Alain, G., and Yosinski, J. (2014b). Deep generative stochastic networks trainable by backprop. In Proceedings of the 30th International Conference on Machine Learning
(ICML?14).
[6] Bergstra, J., Breuleux, O., Bastien, F., Lamblin, P., Pascanu, R., Desjardins, G., Turian, J., Warde-Farley,
D., and Bengio, Y. (2010). Theano: a CPU and GPU math expression compiler. In Proceedings of the
Python for Scientific Computing Conference (SciPy). Oral Presentation.
[7] Breuleux, O., Bengio, Y., and Vincent, P. (2011). Quickly generating representative samples from an
RBM-derived process. Neural Computation, 23(8), 2053?2073.
[8] Glorot, X., Bordes, A., and Bengio, Y. (2011). Deep sparse rectifier neural networks. In AISTATS?2011.
8
[9] Goodfellow, I. J., Warde-Farley, D., Mirza, M., Courville, A., and Bengio, Y. (2013a). Maxout networks.
In ICML?2013.
[10] Goodfellow, I. J., Mirza, M., Courville, A., and Bengio, Y. (2013b). Multi-prediction deep Boltzmann
machines. In NIPS?2013.
[11] Goodfellow, I. J., Warde-Farley, D., Lamblin, P., Dumoulin, V., Mirza, M., Pascanu, R., Bergstra,
J., Bastien, F., and Bengio, Y. (2013c). Pylearn2: a machine learning research library. arXiv preprint
arXiv:1308.4214.
[12] Gregor, K., Danihelka, I., Mnih, A., Blundell, C., and Wierstra, D. (2014). Deep autoregressive networks.
In ICML?2014.
[13] Gutmann, M. and Hyvarinen, A. (2010). Noise-contrastive estimation: A new estimation principle for
unnormalized statistical models. In Proceedings of The Thirteenth International Conference on Artificial
Intelligence and Statistics (AISTATS?10).
[14] Hinton, G., Deng, L., Dahl, G. E., Mohamed, A., Jaitly, N., Senior, A., Vanhoucke, V., Nguyen, P.,
Sainath, T., and Kingsbury, B. (2012a). Deep neural networks for acoustic modeling in speech recognition.
IEEE Signal Processing Magazine, 29(6), 82?97.
[15] Hinton, G. E., Dayan, P., Frey, B. J., and Neal, R. M. (1995). The wake-sleep algorithm for unsupervised
neural networks. Science, 268, 1558?1161.
[16] Hinton, G. E., Srivastava, N., Krizhevsky, A., Sutskever, I., and Salakhutdinov, R. (2012b). Improving
neural networks by preventing co-adaptation of feature detectors. Technical report, arXiv:1207.0580.
[17] Jarrett, K., Kavukcuoglu, K., Ranzato, M., and LeCun, Y. (2009). What is the best multi-stage architecture
for object recognition? In Proc. International Conference on Computer Vision (ICCV?09), pages 2146?2153.
IEEE.
[18] Kingma, D. P. and Welling, M. (2014). Auto-encoding variational bayes. In Proceedings of the International Conference on Learning Representations (ICLR).
[19] Krizhevsky, A. and Hinton, G. (2009). Learning multiple layers of features from tiny images. Technical
report, University of Toronto.
[20] Krizhevsky, A., Sutskever, I., and Hinton, G. (2012). ImageNet classification with deep convolutional
neural networks. In NIPS?2012.
[21] LeCun, Y., Bottou, L., Bengio, Y., and Haffner, P. (1998). Gradient-based learning applied to document
recognition. Proceedings of the IEEE, 86(11), 2278?2324.
[22] Mnih, A. and Gregor, K. (2014). Neural variational inference and learning in belief networks. Technical
report, arXiv preprint arXiv:1402.0030.
[23] Rezende, D. J., Mohamed, S., and Wierstra, D. (2014). Stochastic backpropagation and approximate
inference in deep generative models. Technical report, arXiv:1401.4082.
[24] Rifai, S., Bengio, Y., Dauphin, Y., and Vincent, P. (2012). A generative process for sampling contractive
auto-encoders. In ICML?12.
[25] Salakhutdinov, R. and Hinton, G. E. (2009). Deep Boltzmann machines. In AISTATS?2009, pages 448?
455.
[26] Schmidhuber, J. (1992). Learning factorial codes by predictability minimization. Neural Computation,
4(6), 863?879.
[27] Susskind, J., Anderson, A., and Hinton, G. E. (2010). The Toronto face dataset. Technical Report UTML
TR 2010-001, U. Toronto.
[28] Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I. J., and Fergus, R. (2014).
Intriguing properties of neural networks. ICLR, abs/1312.6199.
[29] Tu, Z. (2007). Learning generative models via discriminative approaches. In Computer Vision and Pattern
Recognition, 2007. CVPR?07. IEEE Conference on, pages 1?8. IEEE.
9
| 5423 |@word version:1 eliminating:1 stronger:1 seek:2 covariance:1 contrastive:2 pg:45 tr:1 solid:1 epartement:1 ecole:1 document:1 deconvolutional:1 rightmost:1 existing:1 com:1 activation:3 yet:2 intriguing:2 assigning:1 must:4 dx:2 gpu:1 visible:1 numerical:1 partition:1 utml:1 designed:1 update:6 generative:44 intelligence:2 prohibitive:1 devising:1 recherche:1 provides:1 pascanu:3 math:1 toronto:4 wierstra:2 kingsbury:1 qualitative:1 prove:1 fitting:1 theoretically:2 multi:2 salakhutdinov:2 actual:1 cpu:1 window:3 increasing:1 provided:2 estimating:3 discover:1 confused:1 moreover:1 project:1 what:1 kind:2 interpreted:1 developer:1 developed:2 differentiation:2 guarantee:1 quantitative:1 fellow:1 every:1 expands:1 ebec:1 zaremba:1 universit:3 classifier:2 sherjil:2 unit:9 danihelka:1 scientist:1 frey:1 laufer:2 mistake:1 limit:1 despite:1 encoding:1 approximately:1 black:1 suggests:1 co:1 limited:2 collapse:1 range:1 statistically:1 jarrett:1 contractive:1 directed:1 unique:2 acknowledgment:1 lecun:2 practice:4 implement:2 differs:1 backpropagation:8 digit:1 procedure:4 susskind:1 thought:1 reject:1 confidence:2 word:2 bergeron:2 suggest:1 cannot:4 context:1 descending:1 www:1 equivalent:1 map:1 optimize:1 dz:1 maximizing:2 deterministic:1 straightforward:2 demonstrated:1 sainath:1 convex:4 focused:1 qc:1 pouget:2 scipy:1 rule:3 estimator:1 x6s:1 lamblin:3 gsn:1 his:1 coordinate:1 analogous:2 updated:2 trois:1 play:1 magazine:1 exact:2 olivier:1 goodfellow:8 jaitly:1 recognition:5 particularly:2 approximated:5 updating:2 expensive:1 predicts:1 database:1 labeled:1 preprint:2 capture:1 region:3 connected:1 gutmann:1 ranzato:1 mesnil:1 discriminates:1 warde:4 dynamic:1 trained:11 motivate:1 bottommost:1 depend:1 oral:1 efficiency:1 eric:1 cae:1 pitted:1 differently:1 represented:3 various:1 regularizer:1 train:8 stacked:1 informatique:1 artificial:2 outside:1 jean:2 valued:1 cvpr:1 loglikelihood:1 delalleau:1 ability:1 statistic:1 succesful:1 h3c:1 itself:1 patrice:1 differentiate:1 sequence:1 differentiable:4 advantage:7 net:21 propose:2 subtracting:1 reconstruction:1 fr:1 adaptation:1 neighboring:1 relevant:1 loop:2 tu:1 date:1 iff:1 degenerate:1 mixing:3 competition:5 sutskever:3 convergence:3 optimum:3 produce:2 generating:5 converges:2 object:1 completion:1 nearest:1 sole:1 op:1 eq:4 recovering:1 auxiliary:1 involves:1 come:2 synchronized:1 direction:1 waveform:1 guided:1 closely:1 correct:1 attribute:1 stochastic:11 human:2 virtual:1 memorized:1 backprop:3 require:4 creativity:1 proposition:2 inspecting:1 extension:2 sufficiently:1 mapping:2 predict:1 dbm:1 claim:1 major:1 achieves:2 early:2 desjardins:1 estimation:7 proc:1 label:2 tool:1 minimization:7 clearly:1 j7:1 gaussian:2 always:1 rather:7 avoid:1 vae:1 rezende:3 jsd:1 validated:1 derived:1 improvement:2 likelihood:12 indicates:1 greatly:1 adversarial:31 detect:1 helpful:1 inference:19 dayan:1 entire:1 hidden:6 misclassified:1 transformed:1 upward:1 arg:1 classification:2 aforementioned:1 dauphin:3 development:1 special:2 equal:2 genuine:1 never:1 having:1 sampling:3 represents:1 synchronizing:1 unsupervised:2 icml:6 pdata:33 future:1 minimized:1 report:5 mirza:4 yoshua:2 piecewise:2 primarily:4 modern:1 dg:10 simultaneously:3 recognize:2 divergence:3 attempt:1 ab:1 detection:1 montr:4 highly:1 mnih:2 evaluation:2 mixture:1 farley:4 goodfeli:1 chain:10 cherry:1 accurate:1 necessary:1 desired:3 theoretical:4 eal:4 earlier:1 modeling:7 column:1 disadvantage:3 subset:1 uniform:1 krizhevsky:3 successful:1 too:2 reported:2 encoders:1 accomplish:1 thibodeau:2 cho:1 density:7 international:4 discriminating:1 probabilistic:1 contract:1 parzen:5 quickly:1 gans:4 containing:1 choose:1 slowly:1 sidestep:1 derivative:2 inefficient:1 supp:2 suggesting:1 potential:2 szegedy:1 de:4 diversity:1 bergstra:3 student:1 explicitly:3 mp:1 performed:1 observer:1 picked:1 jason:1 dumoulin:1 sup:2 reached:1 competitive:1 recover:1 option:1 compiler:1 bouchard:1 bayes:1 minimize:3 convolutional:2 variance:3 who:1 yield:1 vincent:2 kavukcuoglu:1 drive:1 zx:1 classified:1 detector:1 reach:2 ed:1 against:2 involved:1 mohamed:2 proof:6 rbm:1 latex:1 sampled:1 gain:1 dataset:3 lim:1 knowledge:1 attained:1 supervised:1 xxx:1 permitted:2 flowing:1 done:2 though:1 anderson:1 stage:1 until:2 autoencoders:3 horizontal:2 mehdi:1 propagation:1 google:2 minibatch:4 lack:1 defines:1 mode:1 quality:1 perhaps:1 behaved:1 scientific:1 believe:1 concept:2 hence:1 kyunghyun:1 arnaud:1 leibler:1 neal:1 game:9 during:6 encourages:1 maintained:1 backpropagates:1 unnormalized:1 criterion:9 trying:4 pedagogical:1 demonstrate:2 polytechnique:1 performs:1 image:4 variational:6 recently:1 funding:1 sigmoid:1 yosinski:3 refer:1 ai:3 dbn:1 language:1 had:2 replicating:1 bruna:1 specification:2 counterfeit:2 own:1 optimizing:3 optimizes:1 scenario:1 schmidhuber:1 binary:1 came:2 success:2 seen:1 minimum:4 somewhat:2 deng:1 determine:1 maximize:5 converge:1 signal:1 dashed:1 semi:1 currency:2 mix:1 full:1 multiple:2 imperceptible:1 technical:5 cross:2 long:1 cifar:5 impact:1 converging:1 involving:1 prediction:1 multilayer:7 essentially:2 vision:2 arxiv:6 iteration:1 represent:6 sometimes:1 achieved:1 conditionals:1 fellowship:1 thirteenth:1 wake:2 publisher:1 breuleux:3 unlike:2 undirected:1 leveraging:1 spirit:1 flow:1 marcotte:1 near:2 intermediate:1 bengio:15 enough:6 viability:1 variety:1 architecture:2 perfectly:1 inner:2 idea:1 rifai:2 haffner:1 tradeoff:1 blundell:1 whether:2 motivated:1 expression:3 reformulated:1 speech:2 passing:1 deep:25 generally:1 fake:1 useful:2 involve:1 typesetting:1 factorial:1 nonparametric:1 http:1 generate:1 exist:1 dotted:1 coordinating:1 blue:1 discrete:2 hyperparameter:2 promise:1 key:1 drawn:1 dahl:1 kept:1 nce:2 enforced:1 compete:1 everywhere:1 striking:2 extends:1 family:4 reasonable:1 reporting:1 yann:1 draw:1 summarizes:1 dropout:4 bound:1 layer:3 courville:3 copied:1 fold:4 sleep:2 encountered:2 generates:1 speed:1 min:1 optimality:1 concluding:2 chair:1 px:2 alternate:1 poor:1 terminates:1 across:2 perceptible:1 qu:1 making:1 iccv:1 theano:4 taken:1 computationally:1 equation:1 visualization:2 bing:1 previously:1 resource:1 mechanism:2 needed:7 know:1 ascending:1 tractable:1 end:2 informal:1 studying:1 available:3 operation:1 gaussians:1 permit:1 apply:3 observe:1 hierarchical:1 blurry:1 existence:1 include:1 ensure:1 graphical:2 approximating:1 gregor:2 objective:4 question:1 quantity:1 erationnelle:1 strategy:3 parametric:2 primary:1 visiting:2 gradient:16 iclr:2 unable:1 thank:3 capacity:5 kpg:1 ozair:2 code:3 index:1 ratio:2 providing:1 unrolled:1 difficult:1 negative:2 design:2 boltzmann:5 perform:1 allowing:1 observation:1 markov:8 datasets:2 finite:2 acknowledge:1 descent:2 behave:1 emulating:1 saturates:1 team:2 incorporated:1 hinton:7 arbitrary:1 police:1 thm:1 sharp:1 canada:2 david:1 introduced:1 pair:2 required:1 kl:3 discriminator:15 imagenet:1 acoustic:1 delhi:1 learned:3 kingma:3 pylearn2:2 nip:3 able:2 adversary:1 usually:1 pattern:1 confidently:2 challenge:1 max:2 green:1 explanation:1 including:1 belief:1 power:1 natural:2 difficulty:7 treated:1 tfd:5 minimax:5 scheme:2 improve:4 github:1 technology:1 library:1 numerous:2 finished:1 auto:2 prior:3 literature:1 calcul:1 python:1 determining:1 relative:1 fully:1 highlight:1 generation:2 limitation:1 proven:1 generator:17 validation:3 agent:1 vanhoucke:1 sufficient:2 imposes:1 article:2 principle:1 classifying:1 uncorrelated:1 share:1 bordes:1 tiny:1 summary:1 supported:1 alain:2 formal:2 senior:2 allow:1 perceptron:4 institute:1 wide:1 face:2 sparse:1 benefit:2 evaluating:1 rich:3 unaware:1 sensory:1 forward:1 autoregressive:1 preventing:1 nguyen:1 far:1 erhan:1 welling:3 hyvarinen:1 approximate:9 implicitly:1 kullback:1 supremum:1 global:5 overfitting:1 corpus:1 discriminative:9 fergus:1 latent:1 iterative:1 table:4 learn:3 nature:1 obtaining:1 improving:1 excellent:1 interpolating:1 bottou:1 domain:2 did:2 aistats:3 linearly:1 arrow:1 noise:13 arise:1 hyperparameters:1 turian:1 allowed:1 fair:1 xu:1 representative:1 predictability:7 supd:1 momentum:1 explicit:1 xh:1 learns:1 ian:3 theorem:1 specific:1 rectifier:2 bastien:4 showing:2 jensen:2 symbol:1 r2:1 abadie:2 pz:5 admits:1 x:1 glorot:1 exists:1 intractable:5 mnist:5 workshop:1 adding:1 backpropagate:3 explore:1 saddle:1 likely:1 ez:2 partially:1 scalar:2 srivastava:1 corresponds:1 rushed:1 stimulating:1 conditional:5 presentation:1 careful:1 maxout:2 shared:1 change:1 infinite:1 specifically:1 uniformly:1 discriminate:1 player:4 shannon:2 perceptrons:3 aaron:1 vaes:2 formally:1 guillaume:1 support:2 indian:1 accelerated:1 evaluate:2 mcmc:2 audio:1 trainable:2 ex:10 |
4,887 | 5,424 | Deep Symmetry Networks
Robert Gens
Pedro Domingos
Department of Computer Science and Engineering
University of Washington
Seattle, WA 98195-2350, U.S.A.
{rcg,pedrod}@cs.washington.edu
Abstract
The chief difficulty in object recognition is that objects? classes are obscured by
a large number of extraneous sources of variability, such as pose and part deformation. These sources of variation can be represented by symmetry groups,
sets of composable transformations that preserve object identity. Convolutional
neural networks (convnets) achieve a degree of translational invariance by computing feature maps over the translation group, but cannot handle other groups.
As a result, these groups? effects have to be approximated by small translations,
which often requires augmenting datasets and leads to high sample complexity.
In this paper, we introduce deep symmetry networks (symnets), a generalization
of convnets that forms feature maps over arbitrary symmetry groups. Symnets
use kernel-based interpolation to tractably tie parameters and pool over symmetry
spaces of any dimension. Like convnets, they are trained with backpropagation.
The composition of feature transformations through the layers of a symnet provides a new approach to deep learning. Experiments on NORB and MNIST-rot
show that symnets over the affine group greatly reduce sample complexity relative
to convnets by better capturing the symmetries in the data.
1
Introduction
Object recognition is a central problem in vision. What makes it challenging are all the nuisance
factors such as pose, lighting, part deformation, and occlusion. It has been shown that if we could
remove these factors, recognition would be much easier [2, 17]. Convolutional neural networks
(convnets), the current state-of-the-art method for object recognition, capture only one type of invariance (translation); the rest have to be approximated via it and standard features. In practice,
the best networks require enormous datasets which are further expanded by affine transformations
[7, 13] yet are sensitive to imperceptible image perturbations [23]. We propose deep symmetry networks, a generalization of convnets based on symmetry group theory [20] that makes it possible to
capture a broad variety of invariances, and correspondingly improves generalization.
A symmetry group is a set of transformations that preserve the identity of an object and obey the
group axioms. Most of the visual nuisance factors are symmetry groups themselves, and by incorporating them into our model we are able to reduce the sample complexity of learning from data
transformed by these groups. Deep symmetry networks (symnets) form feature maps over any symmetry group, rather than just the translation group. A feature map in a deep symmetry network
is defined analogously to convnets as a filter that is applied at all points in the symmetry space.
Each layer in our general architecture is constructed by applying every symmetry in the group to
the input, computing features on the transformed input, and pooling over neighborhoods. The entire
architecture is then trained by backpropagation. In this paper, we instantiate the architecture with the
affine group, resulting in deep affine networks. In addition to translation, the affine group includes
rotation, scaling and shear. The affine group of the two-dimensional plane is six-dimensional (i.e.,
an affine transformation can be represented by a point in 6D affine space). The key challenge with
1
extending convnets to affine spaces is that it is intractable to explicitly represent and compute with
a high-dimensional feature map. We address this by approximating the map using kernel functions,
which not only interpolate but also control pooling in the feature maps. Compared to convnets,
this architecture substantially reduces sample complexity on image datasets involving 2D and 3D
transformations.
We share with other researchers the hypothesis that explanatory factors cannot be disentangled unless they are represented in an appropriate symmetry space [4, 11]. Our adaptation of a representation to work in symmetry space is similar in some respects to the use of tangent distance in
nearest-neighbor classifiers [22]. Symnets, however, are deep networks that compute features in
symmetry space at every level. Whereas the tangent distance approximation is only locally accurate,
symnet feature maps can represent large displacements in symmetry space. There are other deep
networks that reinterpret the invariance of convolutional networks. Scattering networks [6] are cascades of wavelet decompositions designed to be invariant to particular Lie groups, where translation
and rotation invariance have been demonstrated so far. The M-theory of Anselmi et al. [2] constructs features invariant to a symmetry group by using statistics of dot products with group orbits.
We differ from these networks in that we model multiple symmetries jointly in each layer, we do not
completely pool out a symmetry, and we discriminatively train our entire architecture. The first two
differences are important because objects and their subparts may have relative flexibility but not total
invariance along certain dimensions of symmetry space. For example, a leg of a person can be seen
in some but not all combinations of rotation and scale relative to the torso. Without discriminative
training, scattering networks and M-theory are limited to representing features whose invariances
may be inappropriate for a target concept because they are fixed ahead of time, either by the wavelet
hierarchy of the former or unsupervised training of the latter. The discriminative training of symnets
yields features with task-oriented invariance to their sub-features. In the context of digit recognition
this might mean learning the concept of a ?0? with more rotation invariance than a ?6?, which would
incur loss if it had positive weights in the region of symmetry space where a ?9? would also fire.
Much of the vision literature is devoted to features that reduce or remove the effects of certain symmetry groups, e.g., [18, 17]. Each feature by itself is not discriminative for object recognition, so
structure is modeled separately, usually with a representation that does not generalize to novel viewpoints (e.g., bags-of-features) or with a rigid alignment algorithm that cannot represent uncertainty
over geometry (e.g. [9, 19]). Compared to symnets, these features are not learned, have invariance
limited to a small set of symmetries, and destroy information that could be used to model object
sub-structure. Like deformable part models [10], symnets can model and penalize relative transformations that compose up the hierarchy, but can also capture additional symmetries.
Symmetry group theory has made a limited number of appearances in machine learning [8]. A few
applications are discussed by Kondor [12], and they are also used in determinantal point processes
[14]. Methods for learning transformations from examples [24, 11] could potentially benefit from
being embedded in a deep symmetry network. Symmetries in graphical models [21] lead to effective
lifted probabilistic inference algorithms. Deep symmetry networks may be applicable to these and
other areas.
In this paper, we first review symmetry group theory and its relation to sample complexity. We then
describe symnets and their affine instance, and develop new methods to scale to high-dimensional
symmetry spaces. Experiments on NORB and MNIST-rot show that affine symnets can reduce by a
large factor the amount of data required to achieve a given accuracy level.
2
Symmetry Group Theory
A symmetry of an object is a transformation that leaves certain properties of that object intact [20].
A group is a set S with an operator ? on it with the four properties of closure, associativity, an
identity element, and an inverse element. A symmetry group is a type of group where the group
elements are functions and the operator is function composition. A simple geometric example is
the symmetry group of a square, which consists of four reflections and {0, 1, 2, 3} multiples of 90degree rotations. These transformations can be composed together to yield one of the original eight
symmetries. The identity element is the 0-degree rotation. Each symmetry has a corresponding
inverse element. Composition of these symmetries is associative.
2
Lie groups are continuous symmetry groups whose elements form a smooth differentiable manifold.
For example, the symmetries of a circle include reflections and rotations about the center. The affine
group is a set of transformations that preserves collinearity and parallel lines. The Euclidean group
is a subgroup of the affine group that preserves distances, and includes the set of rigid body motions
(translations and rotations) in three-dimensional space.
The elements of a symmetry group can be represented as matrices. In this form, function composition can be performed via matrix multiplication. The transformation P followed by Q (also denoted
Q ? P) is computed as R = QP. In this paper we treat the transformation matrix P as a point
in D-dimensional space, where D depends on the particular representation of the symmetry group
(e.g., D = 6 for affine transformations in the plane).
A generating set of a group is a subset of the group such that any group element can be expressed
through combinations of generating set elements and their inverses. For example, a generating set
of the translation symmetry group is {x ? x + , y ? y + } for infinitesimal . We define the
k-neighborhood of element f in group S under generating set G as the subset of S that can be
expressed as f composed with elements of G or their inverses at most k times. With the previous
example, the k-neighborhood of a translation vector f would take the shape of a diamond centered
at f in the xy-plane.
The orbit of an object x is the set of objects obtained by applying each element of a symmetry group
to x. Formally, a symmetry group S acting on a set of objects X defines an orbit for each x ? X:
Ox = {s ? x : s ? S}. For example, the orbit of an image I(u) whose points are transformed by the
rotation symmetry group s ? I(u) = I(s?1 ? u) is the set of images resulting from all rotations of
that image. If two orbits share an element, they are the same
S orbit. In this way, a symmetry group
S partitions the set of objects into unique orbits X = a Oa . If a data distribution D(x, y) has
the property that all the elements of an orbit share the same label y, S imposes a constraint on the
hypothesis class of a learner, effectively lowering its VC-dimension and sample complexity [1].
3
Deep Symmetry Networks
Deep symmetry networks represent rich compositional structure that incorporates invariance to highdimensional symmetries. The ideas behind these networks are applicable to any symmetry group, be
it rigid-body transformations in 3D or permutation groups over strings. The architecture of a symnet
consists of several layers of feature maps. Like convnets, these feature maps benefit from weight
tying and pooling, and the whole network is trained with backpropagation. The maps and the filters
they apply are in the dimension D of the chosen symmetry group S.
A deep symmetry network has L layers l ? {1, ..., L} each with Il features and corresponding
feature maps. A feature is the dot-product of a set of weights with a corresponding set of values from
a local region of a lower layer followed by a nonlinearity. A feature map represents the application
of a filter at all points in symmetry space. A feature at point P is computed from the feature maps
of the lower layer at points in the k-neighborhood of P. As P moves in the symmetry space of a
feature map, so does its neighborhood of inputs in the lower layer. Feature map i of layer l is denoted
M [l, i] : RD ? R, a scalar function of the D-dimensional symmetry space. Given a generating set
G ? S, the points in the k-neighborhood of the identity element are stored in an array T[ ]. Each
filter i of layer l defines a weight vector w[l, i, j] for each point T[j] in the k-neighborhood. The
vector w[l, i, j] is the size of Il?1 , the number of features in the underlying layer. For example, a
feature in an affine symnet that detects a person would have positive weight for an arm sub-feature in
the region of the k-neighborhood that would transform the arm relative to the person (e.g., smaller,
rotated, and translated relative to the torso). The value of feature map i in layer l at point P is
the dot-product of weights and underlying feature values in the neighborhood of P followed by a
nonlinearity:
M [l, i](P) =
v(P, l, i) =
? (v(P, l, i))
P|T|
x(P0 ) =
w[l, i, j] ? x(P ? T[j])
+
S(M [l ? 1, 0])(P0 )
...
S(M [l ? 1, Il?1 ])(P0 )
j
*
3
(1)
(2)
(3)
Layer l
Feature map i
Layer l-1
Pooled feature maps 0,1,2
Layer l-1
Feature maps 0,1,2
Kernels
Figure 1: The evaluation of point P in map M [l, i]. The elements of the k-neighborhood of P are
computed P ? T[j]. Each point in the neighborhood is evaluated in the pooled feature maps of the
lower layer l ? 1. The pooled maps are computed with kernels on the underlying feature maps. The
dashed line intersects the points in the pooled map whose values form x(P ? T[j]) in Equation 3; it
also intersects the contours of kernels used to compute those pooled values. The value of the feature
is the sum of the dot-products w[l, i, j] ? x(P ? T[j]) over all j, followed by a nonlinearity.
where ? is the nonlinearity (e.g., tanh(x) or max(x, 0)), v(P, l, i) is the dot product, P ? T[j] represents element j in the k-neighborhood of P, and x(P0 ) is the vector of values from the underlying
pooled maps at point P0 . This definition is a generalization of feature maps in convnets1 . Similarly,
the same filter weights w[l, i, j] are tied across all points P in feature map M [l, i]. The evaluation
of a point in a feature map is visualized in Figure 1.
Feature maps M [l, i] are pooledR via kernel convolution to become S(M [l, i]). In the case of
sum-pooling, S(M [l, i])(P) = M [l, i](P ? Q)K(Q) dQ; for max-pooling, S(M [l, i])(P) =
maxQ M [l, i](P ? Q)K(Q). The kernel K(Q) is also a scalar function of the D-dimensional symmetry space. In the previous example of a person feature, the arm feature map could be pooled over
a wide range of rotations but narrow range of translations and scales so that the person feature allows
for moveable but not unrealistic arms. Each filter can specify the kernels it uses to pool lower layers,
but for the sake of brevity and analogy to convnets we assume that the feature maps of a layer are
pooled by the same kernel. Note that convnets discretize these operations, subsample the pooled
map, and use a uniform kernel K(Q) = 1{kQk? < r}.
As with convnets, the values of points in a symnet feature map are used by higher symnet layers,
layers of fully connected hidden units, and ultimately softmax classification. Hidden units take the
familiar form o = ?(Wx + b), with input x, output o, weight
P matrix W, and bias b. The log-loss
of the softmax L on an instance is ?wi ? x ? bi + log ( c exp (wc ? x + bc )), where Y = i is
the true label, wc and bc are the weight vector and bias for class c, and the summation is over the
classes. The input image is treated as a feature map (or maps, if color or stereo) with values in the
translation symmetry space.
Deep symmetry networks are trained with backpropagation and are amenable to the same best practices as convnets. Though feature maps are defined as continuous, in practice the maps and their
gradients are evaluated on a finite set of points P ? M [l, i]. We provide the partial derivative of the
loss L with respect to a weight vector.
?L
=
?w[l, i, j]
?M [l, i](P)
=
?w[l, i, j]
?M [l,i](P)
?L
P?M [l,i] ?M [l,i](P) ?w[l,i,j]
(4)
? 0 (v(P, l, i)) x(P ? T[j])
(5)
P
1
The neighborhood that defines a square filter in convnets is the reference point translated by up to k times
in x and k times in y.
4
A
B1
B2
B3
B4
B5
A
B5
C2
B1
A
B2
C2
B3
C1
C2
C3
C4
B2
B4
B3
C3
C4
B1
B2
B3
B4
B5
C1
C3
B5
A
C2
C1
C4
B1
B1
B1
A
B2
C1
C1
B3
C1
C2
C3
C4
B1
B4
C1
B5
C3
B4
C4
Figure 2: The feature hierarchy of a three-layer deep affine net is visualized with and without
pooling. From top to bottom, the layers (A,B,C) contain one, five, and four feature maps, each corresponding to a labeled part of the cartoon figure. Each horizontal line represents a six-dimensional
affine feature map, and bold circles denote six-dimensional points in the map. The dashed lines
represent the affine transformation from a feature to the location of one of its filter points. For clarity, only a subset of filter points are shown. Left: Without pooling, the hierarchy represents a rigid
affine transformation among all maps. Another point on feature map A is visualized in grey. Right:
Feature maps B1 and C1 are pooled with a kernel that gives those features flexibility in rotation.
The partial derivative of the loss L with respect to the value of a point in a lower layer is
?L
=
?M [l ? 1, i](P)
?M [l, i0 ](P0 )
=
?M [l ? 1, i](P)
?M [l,i0 ](P0 )
?L
P0 ?M [l,i0 ] ?M [l,i0 ](P0 ) ?M [l?1,i](P)
PIl P
i0
? 0 (v(P0 , l, i0 ))
where the gradient of the pooled feature map
P|T|
j
0
[l?1,i])(P ?T[j])
w[l, i0 , j][i] ?S(M?M
[l?1,i](P)
?S(M [l,i])(P)
?M [l,i](Q)
(6)
(7)
equals K(P ? Q) for sum-pooling.
None of this treatment depends explicitly on the dimensionality of the space except for the kernel
and transformation composition which have polynomial dependence on D. In the next section we
apply this architecture to the affine group in 2D, but it could also be applied to the affine group in
3D or any other symmetry group.
4
Deep Affine Networks
We instantiate a deep symmetry network with the
affine symmetry group in the plane. The affine
symmetry group contains transformations capable of rotating, scaling, shearing, and translating
two-dimensional points. The transformation is described by six coordinates:
0
x
a b
x
e
=
+
y0
c d
y
f
1
0
This means that each of the feature maps M [l, i]
?1
and elements T[j] of the k-neighborhood is represented in six dimensions. The identity transformation is a = d = 1, b = c = e = f = 0. The generating
?1
0
1
set of the affine symmetry group contains six el- Figure 3: The six transformations in the generements, each of which is obtained by adding to ating set of the affine group applied to a square
one of the six coordinates in the identity transform. (exaggerated = 0.2, identity is black square).
This generating set is visualized in Figure 3.
A deep affine network can represent a rich part hierarchy where each weight of a feature modulates
the response to a subpart at a point in the affine neighborhood. The geometry of a deep affine network
is best understood by tracing a point on a feature map through its filter point transforms into lower
layers. Figure 2 visualizes this structure without and with pooling on the left and right sides of
the diagram, respectively. Without pooling, the feature hierarchy defines a rigid affine relationship
between the point of evaluation on a map and the location of its sub-features. In contrast, a pooled
value on a sub-feature map is computed from a neighborhood defined by the kernel of points in
affine space; this can represent model flexibility along certain dimensions of affine space.
5
5
Scaling to High-Dimensional Symmetry Spaces
It would be intractable to explicitly represent the high-dimensional feature maps of symnets. Even a
subsampled grid becomes unwieldy at modest dimensions (e.g., a grid in affine space with ten steps
per axis has 106 points). Instead, each feature map is evaluated at N control points. The control
points are local maxima of the feature in symmetry space, found by Gauss-Newton optimization,
each initialized from a prior. This can be seen as a form of non-maximum suppression. Since the
goal is recognition, there is no need to approximate the many points in symmetry space where the
feature is not present. The map is then interpolated with kernel functions; the shape of the function
also controls pooling.
5.1 Transformation Optimization
Convnets max-pool a neighborhood of translation space by exhaustive evaluation of feature locations. There are a number of algorithms that solve for a maximal feature location in symmetry space
but they are not efficient when the feature weights are frequently adjusted [9, 19]. We adopt an
iterative approach that dovetails with the definition of our features.
If a symnet is based on a Lie group, gradient based optimization can be used to find a point P?
that locally maximizes the feature value (Equation 1) initialized at point P. In our experiments
with deep affine nets, we follow the forward compositional (FC) warp [3] to align filters with the
image. An extension of Lucas-Kanade, FC solves for an image alignment. We adapt this procedure
P|T|
to our filters and weight vectors: min?P j kw[l, i, j] ? x(P ? ?P ? T[j])k2 . We run an FC
alignment for each of the N control points in feature map M [l, i], each initialized from a prior.
P|T|
Assuming j kx(P ? ?P ? T[j])k2 is constant, this procedure locally maximizes the dot product
between the filter and the map in Equation 2. Each iteration of FC takes a Gauss-Newton step to
solve for a transformation of the neighborhood of the feature in the underlying map ?P, which is
then composed with the control point: P ? P ? ?P.
5.2 Kernels
Given a set of N local optima O?
=
{(P1 , v1 ), . . . , (PN , vN )} in D-dimensional feature
map M [l, i], we use kernel-based interpolation to compute
a pooled map S(M [l, i]). The kernel performs three
functions: penalizing relative locations of sub-features
in symmetry space (cf. [10]), interpolating the map, and
pooling a region of the map. These roles could be split
into separate filter-specific kernels that are then convolved
appropriately. The choice of these kernels will vary
with the application. In our experiments, we lump these
functions into a single kernel for a layer. We use a Gaussian
T ?1
kernel K(Q) = e?q ? q where q is the D-dimensional
vector representation of Q and the D?D covariance matrix
? controls the shape and extent of the kernel. Several
instances of this kernel are shown in Figure 4. Max-pooling
produced the best results on our tests.
6
Figure 4: Contours of three 6D Gaussian kernels visualized on a surface
in affine space. Points are visualized
by an oriented square transformed by
the affine transformation at that point.
Each kernel has a different covariance
matrix ?.
Experiments
In our experiments we test the hypothesis that a deep network with access to a larger symmetry group
will generalize better from fewer examples, provided those symmetries are present in the data. In
particular, theory suggests that a symnet will have better sample complexity than another classifier
on a dataset if it is based on a symmetry group that generates variations present in that dataset [1].
We compare deep affine symnets to convnets on the MNIST-rot and NORB image classification
datasets, which finely sample their respective symmetry spaces such that learning curves measure
the amount of augmentation that would be required to achieve similar performance. On both datasets
affine symnets achieve a substantial reduction in sample complexity. This is particularly remarkable
on NORB because its images are generated by a symmetry space in 3D. Symnet running time was
within an order of magnitude of convnets, and could be greatly optimized.
6.1 MNIST-rot
MNIST-rot [15] consists of 28x28 pixel greyscale images: 104 for training, 2 ? 103 for validation,
and 5 ? 104 for testing. The images are sampled from the MNIST digit recognition dataset and each
6
Figure 5: Impact of training set size on MNIST-rot test performance for architectures that use either
one convolutional layer or one affine symnet layer.
is rotated by a random angle in the uniform distribution [0, 2?]. With transformations that apply
to the whole image, MNIST-rot is a good testbed for comparing the performance of a single affine
layer to a single convnet layer.
We modified the Theano [5] implementation of convolutional networks so that the network consisted
of a single layer of convolution and maxpooling followed by a hidden layer of 500 units and then
softmax classification. The affine net layer was directly substituted for the convolutional layer. The
control points of the affine net were initialized at uniformly random positions with rotations oriented
around the image center, and each control point was locally optimized with four iterations of GaussNewton updates. The filter points of the affine net were arranged in a square grid. Both the affine
net and the convnet compute a dot-product and use the sigmoid nonlinearity. Both networks were
trained with 50 epochs of mini-batch gradient descent with momentum, and test results are reported
on the network with lowest error on the validation set2 . The convnet did best with small 5 ? 5
filters and the symnet with large 20 ? 20 filters. This is not surprising because the convnet must
approximate the large rotations of the dataset with translations of small patches. The affine net can
pool directly in this space of rotations with large filters.
Learning curves for the two networks are presented in Figure 5. We observe that the affine symnet
roughly halves the error of the convnet. With small sample sizes, the symnet achieves an accuracy
for which the convnet requires about eight times as many samples.
6.2 NORB
MNIST-rot is a synthetic dataset with symmetries that are not necessarily representative of real
images. The NORB dataset [16] contains 2 ? 108 ? 108 pixel stereoscopic images of 50 toys in five
categories: quadrupeds, human figures, airplanes, trucks, and cars. Five of the ten instances of each
category are reserved for the test set. Each toy is photographed on a turntable from an exhaustive
set of angles and lighting conditions. Each image is then perturbed by a random translation shift,
planar rotation, luminance change, contrast change, scaling, distractor object, and natural image
background. A sixth blank category containing just the distractor and background is also used. As
in other papers, we downsample the images to 2?48?48. To compensate for the effect of distractors
in smaller training sets, we also train and test on a version of the dataset that is centrally-cropped to
2 ? 24 ? 24. We report results for whichever version had lower validation error. In our experiments
we train on a variable subset of the first training fold, using the first 2 ? 103 images of the second
fold for validation. Our results use both of the testing folds.
We compare architectures that use two convolutional layers or two affine ones, which performed
better than single-layer ones. As with the MNIST-rot experiments, the symnet and convnet layers
are followed by a layer of 500 hidden units and softmax classification. The symnet control points in
the first layer were arranged in three concentric rings in translation space, with 8 points spaced across
rotation (200 total points). Control points in the second layer were fixed at the center of translation
2
Grid search over learning rate {.1, .2}, mini-batch size {10, 50, 100}, filter size {5, 10, 15, 20, 25}, number of filters {20, 50, 80}, pooling size (convnet) {2, 3, 4}, and number of control points (symnet) {5, 10, 20}.
7
Figure 6: Impact of training set size on NORB test performance for architectures with two convolutional or affine symnet layers followed by a fully connected layer and then softmax classification.
space arranged over 8 rotations and up to 2 vertical scalings (16 total points) to approximate the
effects of elevation change. Control points were not iteratively optimized due to the small size of
object parts in downsampled images. The filter points of the first layer of the affine net were arranged
in a square grid. The second layer filter points were arranged in a circle in translation space at a 3
or 4 pixel radius, with 8 filter points evenly spaced across rotation at each translation. We report the
test results of the networks with lowest validation error on a range of hyperparameters3 .
The learning curves for convnets and affine symnets are shown in Figure 6. Even though the primary
variability in NORB is due to rigid 3D transformations, we find that our affine networks still have
an advantage over convnets. A 3D rotation can be locally approximated with 2D scales, shears, and
rotations. The affine net can represent these transformations and so it benefited from larger filter
patches. The translation approximation of the convnet is unable to properly align larger features to
the true symmetries, and so it performed better with smaller filters. The convnet requires about four
times as much data to reach the accuracy of the symnet with the smallest training set. Larger filters
capture more structure than smaller ones, allowing symnets to generalize better than convnets, and
effectively giving each symnet layer the power of more than one convnet layer.
The left side of the graph may be more indicative of the types of gains symnets may have over
convnets in more realistic datasets that do not have thousands of images of identical 3D shapes.
With the ability to apply more realistic transformations to sub-parts, symnets may also be better able
to reuse substructure on datasets with many interrelated or fine-grained categories. Since symnets
are a clean generalization of convnets, they should benefit from the learning, regularization, and
efficiency techniques used by state-of-the-art networks [13].
7
Conclusion
Symmetry groups underlie the hardest challenges in computer vision. In this paper we introduced
deep symmetry networks, the first deep architecture that can compute features over any symmetry
group. It is a natural generalization of convolutional neural networks that uses kernel interpolation
and transformation optimization to address the difficulties in representing high-dimensional feature
maps. In experiments on two image datasets with 2D and 3D variability, affine symnets achieved
higher accuracy than convnets while using significantly less data.
Directions for future work include extending to other symmetry groups (e.g., lighting, 3D space),
modeling richer distortions, incorporating probabilistic inference, and scaling to larger datasets.
Acknowledgments
This research was partly funded by ARO grant W911NF-08-1-0242, ONR grants N00014-13-10720 and N00014-12-1-0312, and AFRL contract FA8750-13-2-0019. The views and conclusions
contained in this document are those of the authors and should not be interpreted as necessarily
representing the official policies, either expressed or implied, of ARO, ONR, AFRL, or the United
States Government.
3
Grid search over filter size in each layer {6, 9}, pooling size in each layer (convnet) {2, 3, 4}, first layer
control point translation spacing (symnet) {2, 3}, momentum {0, 0.5, 0.9}, others as in MNIST-rot.
8
References
[1] Y. S. Abu-Mostafa. Hints and the VC dimension. Neural Computation, 5(2):278?288, 1993.
[2] F. Anselmi, J. Z. Leibo, L. Rosasco, J. Mutch, A. Tacchetti, and T. Poggio. Unsupervised learning of
invariant representations in hierarchical architectures. ArXiv preprint 1311.4158, 2013.
[3] S. Baker and I. Matthews. Lucas-Kanade 20 years on: A unifying framework. International Journal of
Computer Vision, 56(3):221?255, 2004.
[4] Y. Bengio, A. Courville, and P. Vincent. Representation learning: A review and new perspectives. IEEE
Transactions on Pattern Analysis and Machine Intelligence, 35(8):1798?1828, 2013.
[5] J. Bergstra, O. Breuleux, F. Bastien, P. Lamblin, R. Pascanu, G. Desjardins, J. Turian, D. Warde-Farley,
and Y. Bengio. Theano: a CPU and GPU math expression compiler. In Proceedings of the Python for
Scientific Computing Conference, 2010.
[6] J. Bruna and S. Mallat. Invariant scattering convolution networks. IEEE Transactions on Pattern Analysis
and Machine Intelligence, 35(8):1872?1886, 2013.
[7] D. Ciresan, U. Meier, and J. Schmidhuber. Multi-column deep neural networks for image classification.
In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2012.
[8] P. Diaconis. Group representations in probability and statistics. Institute of Mathematical Statistics, 1988.
[9] B. Drost, M. Ulrich, N. Navab, and S. Ilic. Model globally, match locally: Efficient and robust 3D object
recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2010.
[10] P. Felzenszwalb, D. McAllester, and D. Ramanan. A discriminatively trained, multiscale, deformable part
model. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2008.
[11] G. E. Hinton, A. Krizhevsky, and S. D. Wang. Transforming auto-encoders. In Proceedings of the TwentyFirst International Conference on Artificial Neural Networks, 2011.
[12] I. R. Kondor. Group theoretical methods in machine learning. Columbia University, 2008.
[13] A. Krizhevsky, I. Sutskever, and G. E. Hinton. ImageNet classification with deep convolutional neural
networks. In Advances in Neural Information Processing Systems 25, 2012.
[14] A. Kulesza and B. Taskar. Determinantal point processes for machine learning. ArXiv preprint 1207.6083,
2012.
[15] H. Larochelle, D. Erhan, A. Courville, J. Bergstra, and Y. Bengio. An empirical evaluation of deep architectures on problems with many factors of variation. In Proceedings of the Twenty-Fourth International
Conference on Machine Learning, 2007.
[16] Y. LeCun, F. J. Huang, and L. Bottou. Learning methods for generic object recognition with invariance to
pose and lighting. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
2004.
[17] T. Lee and S. Soatto. Video-based descriptors for object recognition. Image and Vision Computing,
29(10):639?652, 2011.
[18] D. G. Lowe. Object recognition from local scale-invariant features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1999.
[19] F. Lu and E. Milios. Robot pose estimation in unknown environments by matching 2D range scans.
Journal of Intelligent and Robotic Systems, 18(3):249?275, 1997.
[20] W. Miller. Symmetry groups and their applications. Academic Press, 1972.
[21] M. Niepert. Markov chains on orbits of permutation groups. In Proceedings of the Twenty-Eight Conference on Uncertainty in Artificial Intelligence, 2012.
[22] P. Simard, Y. LeCun, and J. S. Denker. Efficient pattern recognition using a new transformation distance.
In Advances in Neural Information Processing Systems 5, 1992.
[23] C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I. Goodfellow, and R. Fergus. Intriguing
properties of neural networks. International Conference on Learning Representations, 2014.
[24] L. Wiskott and T. J. Sejnowski. Slow feature analysis: Unsupervised learning of invariances. Neural
Computation, 14(4):715?770, 2002.
9
| 5424 |@word collinearity:1 version:2 kondor:2 polynomial:1 closure:1 grey:1 decomposition:1 p0:10 covariance:2 reduction:1 contains:3 united:1 bc:2 document:1 fa8750:1 current:1 comparing:1 blank:1 surprising:1 yet:1 intriguing:1 must:1 gpu:1 determinantal:2 realistic:2 partition:1 wx:1 shape:4 remove:2 designed:1 update:1 half:1 instantiate:2 leaf:1 fewer:1 intelligence:3 indicative:1 plane:4 provides:1 pascanu:1 math:1 location:5 five:3 mathematical:1 along:2 constructed:1 c2:5 become:1 consists:3 compose:1 introduce:1 shearing:1 roughly:1 themselves:1 frequently:1 p1:1 distractor:2 multi:1 detects:1 globally:1 cpu:1 inappropriate:1 becomes:1 provided:1 underlying:5 baker:1 maximizes:2 lowest:2 what:1 tying:1 interpreted:1 substantially:1 string:1 transformation:31 every:2 reinterpret:1 tie:1 zaremba:1 classifier:2 k2:2 control:14 unit:4 underlie:1 grant:2 ramanan:1 positive:2 engineering:1 local:4 treat:1 understood:1 interpolation:3 might:1 black:1 suggests:1 challenging:1 limited:3 range:4 bi:1 unique:1 acknowledgment:1 lecun:2 testing:2 practice:3 backpropagation:4 digit:2 procedure:2 displacement:1 area:1 axiom:1 empirical:1 cascade:1 significantly:1 matching:1 downsampled:1 cannot:3 operator:2 context:1 applying:2 map:60 demonstrated:1 center:3 array:1 lamblin:1 disentangled:1 handle:1 variation:3 coordinate:2 target:1 hierarchy:6 mallat:1 us:2 domingo:1 hypothesis:3 goodfellow:1 element:18 recognition:18 approximated:3 particularly:1 labeled:1 bottom:1 role:1 taskar:1 preprint:2 wang:1 capture:4 thousand:1 moveable:1 region:4 connected:2 substantial:1 transforming:1 environment:1 complexity:8 warde:1 ultimately:1 trained:6 incur:1 efficiency:1 learner:1 completely:1 translated:2 represented:5 intersects:2 train:3 effective:1 describe:1 quadruped:1 sejnowski:1 artificial:2 gaussnewton:1 neighborhood:18 exhaustive:2 whose:4 richer:1 larger:5 solve:2 distortion:1 ability:1 statistic:3 jointly:1 itself:1 transform:2 associative:1 advantage:1 differentiable:1 net:9 propose:1 aro:2 product:7 maximal:1 adaptation:1 gen:1 flexibility:3 achieve:4 deformable:2 seattle:1 sutskever:2 optimum:1 extending:2 generating:7 ring:1 rotated:2 object:21 develop:1 augmenting:1 pose:4 nearest:1 solves:1 c:1 larochelle:1 differ:1 direction:1 radius:1 filter:27 vc:2 centered:1 human:1 mcallester:1 translating:1 require:1 government:1 generalization:6 elevation:1 summation:1 adjusted:1 extension:1 around:1 exp:1 mostafa:1 matthew:1 desjardins:1 vary:1 adopt:1 achieves:1 smallest:1 estimation:1 applicable:2 bag:1 label:2 tanh:1 sensitive:1 navab:1 gaussian:2 modified:1 rather:1 pn:1 lifted:1 properly:1 greatly:2 contrast:2 suppression:1 inference:2 downsample:1 rigid:6 el:1 i0:7 entire:2 explanatory:1 associativity:1 hidden:4 relation:1 transformed:4 pixel:3 translational:1 classification:7 among:1 denoted:2 extraneous:1 lucas:2 art:2 softmax:5 equal:1 construct:1 washington:2 cartoon:1 identical:1 represents:4 broad:1 kw:1 unsupervised:3 hardest:1 future:1 report:2 others:1 intelligent:1 hint:1 few:1 oriented:3 composed:3 preserve:4 diaconis:1 interpolate:1 familiar:1 subsampled:1 geometry:2 occlusion:1 fire:1 evaluation:5 alignment:3 farley:1 behind:1 devoted:1 chain:1 amenable:1 accurate:1 capable:1 partial:2 xy:1 poggio:1 respective:1 modest:1 unless:1 euclidean:1 rotating:1 orbit:9 circle:3 initialized:4 obscured:1 deformation:2 theoretical:1 instance:4 column:1 modeling:1 w911nf:1 ating:1 subset:4 uniform:2 krizhevsky:2 stored:1 reported:1 encoders:1 perturbed:1 synthetic:1 person:5 international:4 probabilistic:2 contract:1 lee:1 pool:5 analogously:1 together:1 augmentation:1 central:1 containing:1 rosasco:1 huang:1 derivative:2 simard:1 toy:2 szegedy:1 bergstra:2 pooled:13 b2:5 includes:2 bold:1 explicitly:3 depends:2 performed:3 view:1 lowe:1 compiler:1 parallel:1 pil:1 substructure:1 square:7 il:3 accuracy:4 convolutional:10 descriptor:1 reserved:1 miller:1 yield:2 spaced:2 generalize:3 vincent:1 produced:1 none:1 lu:1 lighting:4 researcher:1 visualizes:1 reach:1 definition:2 infinitesimal:1 sixth:1 milios:1 sampled:1 gain:1 dataset:7 treatment:1 color:1 car:1 improves:1 torso:2 dimensionality:1 distractors:1 scattering:3 higher:2 afrl:2 follow:1 planar:1 specify:1 response:1 mutch:1 arranged:5 evaluated:3 ox:1 though:2 niepert:1 just:2 convnets:24 twentyfirst:1 horizontal:1 multiscale:1 defines:4 scientific:1 b3:5 effect:4 concept:2 true:2 contain:1 consisted:1 former:1 regularization:1 soatto:1 iteratively:1 nuisance:2 performs:1 motion:1 reflection:2 image:25 novel:1 sigmoid:1 rotation:21 shear:2 qp:1 b4:5 discussed:1 composition:5 rd:1 grid:6 similarly:1 nonlinearity:5 had:2 rot:10 dot:7 funded:1 access:1 bruna:2 robot:1 surface:1 maxpooling:1 align:2 exaggerated:1 perspective:1 schmidhuber:1 certain:4 n00014:2 onr:2 seen:2 additional:1 dashed:2 multiple:2 reduces:1 smooth:1 imperceptible:1 match:1 adapt:1 x28:1 academic:1 compensate:1 impact:2 involving:1 vision:10 arxiv:2 iteration:2 kernel:26 represent:9 achieved:1 dovetail:1 penalize:1 c1:8 addition:1 whereas:1 separately:1 background:2 cropped:1 fine:1 diagram:1 spacing:1 source:2 appropriately:1 breuleux:1 rest:1 finely:1 pooling:15 incorporates:1 lump:1 split:1 bengio:3 variety:1 architecture:13 ciresan:1 reduce:4 idea:1 airplane:1 shift:1 six:8 expression:1 b5:5 reuse:1 stereo:1 compositional:2 deep:28 turntable:1 amount:2 transforms:1 locally:6 ten:2 visualized:6 category:4 stereoscopic:1 per:1 subpart:2 abu:1 group:65 key:1 four:5 enormous:1 clarity:1 penalizing:1 clean:1 leibo:1 kqk:1 lowering:1 v1:1 destroy:1 luminance:1 graph:1 sum:3 year:1 run:1 inverse:4 angle:2 uncertainty:2 fourth:1 vn:1 patch:2 scaling:6 capturing:1 layer:48 followed:7 centrally:1 courville:2 fold:3 truck:1 ahead:1 constraint:1 sake:1 wc:2 interpolated:1 generates:1 min:1 expanded:1 photographed:1 department:1 combination:2 smaller:4 across:3 y0:1 wi:1 leg:1 invariant:5 theano:2 equation:3 whichever:1 operation:1 denker:1 eight:3 obey:1 apply:4 appropriate:1 observe:1 hierarchical:1 generic:1 batch:2 convolved:1 anselmi:2 original:1 top:1 running:1 include:2 cf:1 graphical:1 newton:2 unifying:1 giving:1 approximating:1 implied:1 move:1 primary:1 dependence:1 gradient:4 distance:4 separate:1 convnet:12 unable:1 oa:1 evenly:1 manifold:1 extent:1 assuming:1 modeled:1 relationship:1 mini:2 robert:1 potentially:1 greyscale:1 implementation:1 policy:1 twenty:2 diamond:1 allowing:1 discretize:1 vertical:1 convolution:3 unknown:1 datasets:9 markov:1 finite:1 descent:1 hinton:2 variability:3 perturbation:1 arbitrary:1 tacchetti:1 concentric:1 introduced:1 meier:1 required:2 c3:5 optimized:3 imagenet:1 c4:5 learned:1 narrow:1 testbed:1 subgroup:1 maxq:1 tractably:1 address:2 able:2 usually:1 pattern:8 kulesza:1 challenge:2 max:4 video:1 unrealistic:1 power:1 difficulty:2 treated:1 natural:2 arm:4 representing:3 pedrod:1 axis:1 auto:1 columbia:1 review:2 literature:1 geometric:1 python:1 tangent:2 multiplication:1 prior:2 relative:7 epoch:1 embedded:1 loss:4 fully:2 discriminatively:2 permutation:2 analogy:1 composable:1 remarkable:1 validation:5 degree:3 affine:53 imposes:1 wiskott:1 dq:1 viewpoint:1 ulrich:1 share:3 translation:20 bias:2 side:2 warp:1 institute:1 neighbor:1 wide:1 felzenszwalb:1 correspondingly:1 tracing:1 benefit:3 curve:3 dimension:8 rich:2 contour:2 forward:1 made:1 author:1 far:1 erhan:2 transaction:2 approximate:3 robotic:1 b1:8 norb:8 discriminative:3 fergus:1 continuous:2 iterative:1 search:2 chief:1 kanade:2 robust:1 symmetry:83 bottou:1 interpolating:1 necessarily:2 substituted:1 official:1 did:1 whole:2 subsample:1 turian:1 body:2 benefited:1 representative:1 slow:1 sub:7 position:1 momentum:2 lie:3 tied:1 wavelet:2 grained:1 unwieldy:1 specific:1 bastien:1 incorporating:2 intractable:2 mnist:11 adding:1 effectively:2 modulates:1 magnitude:1 kx:1 easier:1 fc:4 interrelated:1 appearance:1 visual:1 set2:1 expressed:3 contained:1 scalar:2 pedro:1 identity:8 goal:1 change:3 except:1 uniformly:1 acting:1 total:3 invariance:13 gauss:2 partly:1 intact:1 formally:1 highdimensional:1 latter:1 scan:1 brevity:1 |
4,888 | 5,425 | A Multiplicative Model for Learning Distributed
Text-Based Attribute Representations
Ryan Kiros, Richard S. Zemel, Ruslan Salakhutdinov
University of Toronto
Canadian Institute for Advanced Research
{rkiros, zemel, rsalakhu}@cs.toronto.edu
Abstract
In this paper we propose a general framework for learning distributed representations of attributes: characteristics of text whose representations can be jointly
learned with word embeddings. Attributes can correspond to a wide variety of
concepts, such as document indicators (to learn sentence vectors), language indicators (to learn distributed language representations), meta-data and side information (such as the age, gender and industry of a blogger) or representations of
authors. We describe a third-order model where word context and attribute vectors
interact multiplicatively to predict the next word in a sequence. This leads to the
notion of conditional word similarity: how meanings of words change when conditioned on different attributes. We perform several experimental tasks including
sentiment classification, cross-lingual document classification, and blog authorship attribution. We also qualitatively evaluate conditional word neighbours and
attribute-conditioned text generation.
1
Introduction
Distributed word representations have enjoyed success in several NLP tasks [1, 2]. More recently,
the use of distributed representations have been extended to model concepts beyond the word level,
such as sentences, phrases and paragraphs [3, 4, 5, 6], entities and relationships [7, 8] and embeddings of semantic categories [9, 10].
In this paper we propose a general framework for learning distributed representations of attributes:
characteristics of text whose representations can be jointly learned with word embeddings. The use
of the word attribute in this context is general. Table 1 illustrates several of the experiments we
perform along with the corresponding notion of attribute. For example, an attribute can represent
an indicator of the current sentence or language being processed. This allows us to learn sentence
and language vectors, similar to the proposed model of [6]. Attributes can also correspond to side
information, or metadata associated with text. For instance, a collection of blogs may come with
information about the age, gender or industry of the author. This allows us to learn vectors that can
capture similarities across metadata based on the associated body of text. The goal of this work
is to show that our notion of attribute vectors can achieve strong performance on a wide variety of
NLP related tasks. In particular, we demonstrate strong quantitative performance on three highly
diverse tasks: sentiment classification, cross-lingual document classification, and blog authorship
attribution.
To capture these kinds of interactions between attributes and text, we propose to use a third-order
model where attribute vectors act as gating units to a word embedding tensor. That is, words are
represented as a tensor consisting of several prototype vectors. Given an attribute vector, a word
embedding matrix can be computed as a linear combination of word prototypes weighted by the
attribute representation. During training, attribute vectors reside in a separate lookup table which
can be jointly learned along with word features and the model parameters. This type of three-way
1
Table 1: Summary of tasks and attribute types used in our experiments. The first three are quantitative while the second three are qualitative.
Task
Sentiment Classification
Cross-Lingual Classification
Authorship Attribution
Conditional Text Generation
Structured Text Generation
Conditional Word Similarity
Dataset
Sentiment Treebank
RCV1/RCV2
Blog Corpus
Gutenberg Corpus
Gutenberg Corpus
Blogs & Europarl
Attribute type
Sentence Vector
Language Vector
Author Metadata
Book Vector
Part of Speech Tags
Author Metadata / Language
interaction can be embedded into a neural language model, where the three-way interaction consists
of the previous context, the attribute and the score (or distribution) of the next word after the context.
Using a word embedding tensor gives rise to the notion of conditional word similarity. More specifically, the neighbours of word embeddings can change depending on which attribute is being conditioned on. For example, the word ?joy? when conditioned on an author with the industry attribute
?religion? appears near ?rapture? and ?god? but near ?delight? and ?comfort? when conditioned on
an author with the industry attribute ?science?. Another way of thinking of our model would be
the language analogue of [11]. They used a factored conditional restricted Boltzmann machine for
modelling motion style defined by real or continuous valued style variables. When our factorization
is embedded into a neural language model, it allows us to generate text conditioned on different
attributes in the same manner as [11] could generate motions from different styles. As we show in
our experiments, if attributes are represented by different books, samples generated from the model
learn to capture associated writing styles from the author. Furthermore, we demonstrate a strong
performance gain for authorship attribution when conditional word representations are used.
Multiplicative interactions have also been previously incorporated into neural language models. [12]
introduced a multiplicative model where images are used for gating word representations. Our
framework can be seen as a generalization of [12] and in the context of their work an attribute would
correspond to a fixed representation of an image. [13] introduced a multiplicative recurrent neural
network for generating text at the character level. In their model, the character at the current timestep
is used to gate the network?s recurrent matrix. This led to a substantial improvement in the ability to
generate text at the character level as opposed to a non-multiplicative recurrent network.
2
Methods
In this section we describe the proposed models. We first review the log-bilinear neural language
model of [14] as it forms the basis for much of our work. Next, we describe a word embedding
tensor and show how it can be factored and introduced into a multiplicative neural language model.
This is concluded by detailing how our attribute vectors are learned.
2.1
Log-bilinear neural language models
The log-bilinear language model (LBL) [14] is a deterministic model that may be viewed as a feedforward neural network with a single linear hidden layer. Each word w in the vocabulary is represented as a K-dimensional real-valued vector rw ? RK . Let R denote the V ? K matrix of
word representation vectors where V is the vocabulary size. Let (w1 , . . . wn?1 ) be a tuple of n ? 1
words where n ? 1 is the context size. The LBL model makes a linear prediction of the next word
representation as
n?1
X
?
r=
C(i) rwi ,
(1)
i=1
(i)
where C , i = 1, . . . , n ? 1 are K ? K context parameter matrices. Thus, ?
r is the predicted
representation of rwn . The conditional probability P (wn = i|w1:n?1 ) of wn given w1 , . . . , wn?1 is
exp(?
rT ri + bi )
P (wn = i|w1:n?1 ) = PV
,
rT rj + bj )
j=1 exp(?
where b ? RV is a bias vector. Learning can be done using backpropagation.
2
(2)
(a) NLM
(b) Multiplicative NLM
(c) Multiplicative NLM with language switch
Figure 1: Three different formulations for predicting the next word in a neural language model. Left:
A standard neural language model (NLM). Middle: The context and attribute vectors interact via
a multiplicative interaction. Right: When words are unshared across attributes, a one-hot attribute
vector gates the factors-to-vocabulary matrix.
2.2
A word embedding tensor
Traditionally, word representation matrices are represented as a matrix R ? RV ?K , such as in
the case of the log-bilinear model. Throughout this work, we instead represent words as a tensor
T ? RV ?K?D where D corresponds to the number of tensor slices. Given an attribute vector
PD
x ? RD , we can compute attribute-gated word representations as T x = i=1 xi T (i) i.e. word
representations with respect to x are computed as a linear combination of slices weighted by each
component xi of x.
It is often unnecessary to use a fully unfactored tensor. Following [15, 16], we re-represent T in
terms of three matrices Wf k ? RF ?K , Wf d ? RF ?D and Wf v ? RF ?V , such that
T x = (Wf v )> ? diag(Wf d x) ? Wf k ,
(3)
where diag(?) denotes the matrix with its argument on the diagonal. These matrices are parametrized
by a pre-chosen number of factors F .
2.3
Multiplicative neural language models
We now show how to embed our word representation tensor T into the log-bilinear neural language
model. Let E = (Wf k )> Wf v denote a ?folded? K ? V matrix of word embeddings. Given the
context w1 , . . . , wn?1 , the predicted next word representation ?
r is given by
?
r=
n?1
X
C(i) E(:, wi ),
(4)
i=1
where E(:, wi ) denotes the column of E for the word representation of wi and C(i) , i = 1, . . . , n?1
are K ? K context matrices. Given a predicted next word representation ?
r, the factor outputs are
f = (Wf k?
r) ? (Wf d x),
(5)
where ? is a component-wise product. The conditional probability P (wn = i|w1:n?1 , x) of wn
given w1 , . . . , wn?1 and x can be written as
exp (Wf v (:, i))> f + bi
P (wn = i|w1:n?1 , x) = PV
.
fv
>
j=1 exp (W (:, j)) f + bj
Here, Wf v (:, i) denotes the column of Wf v corresponding to word i. In contrast to the log-bilinear
model, the matrix of word representations R from before is replaced with the factored tensor T , as
shown in Fig. 1.
2.4
Unshared vocabularies across attributes
Our formulation for T assumes that word representations are shared across all attributes. In some
cases, words may only be specific to certain attributes and not others. An example of this is crosslingual modelling, where it is necessary to have language specific vocabularies. As a running example, consider the case where each attribute corresponds to a language representation vector. Let
3
Table 2: Samples generated from the model when conditioning on various attributes. For the last
example, we condition on the average of the two vectors (symbol <#> corresponds to a number).
Attribute
Bible
Caesar
1
2
(Bible +
Caesar)
Sample
<#> : <#> for thus i enquired unto thee , saying , the lord had not come unto
him . <#> : <#> when i see them shall see me greater am that under the name
of the king on israel .
to tell vs pindarus : shortly pray , now hence , a word . comes hither , and
let vs exclaim once by him fear till loved against caesar . till you are now which
have kept what proper deed there is an ant ? for caesar not wise cassi
let our spring tiger as with less ; for tucking great fellowes at ghosts of broth .
industrious time with golden glory employments . <#> : <#> but are far in men
soft from bones , assur too , set and blood of smelling , and there they cost ,
i learned : love no guile his word downe the mystery of possession
x denote the attribute vector for language ` and x0 for language `0 (e.g. English and French). We
can then compute language-specific word representations T ` by breaking up our decomposition into
language dependent and independent components (see Fig. 1c):
T ` = (W`f v )> ? diag(Wf d x) ? Wf k ,
(6)
where (W`f v )> is a V` ? F language specific matrix. The matrices Wf d and Wf k do not depend
on the language or the vocabulary, whereas (W`f v )> is language specific. Moreover, since each language may have a different sized vocabulary, we use V` to denote the vocabulary size of language `.
Observe that this model has an interesting property in that it allows us to share statistical strength
across word representations of different languages. In particular, we show in our experiments how
we can improve cross-lingual classification performance between English and German when a large
amount of parallel data exists between English and French and only a small amount of parallel data
exists between English and German.
2.5
Learning attribute representations
We now discuss how to learn representation vectors x. Recall that when training neural language
models, the word representations of w1 , . . . , wn?1 are updated by backpropagating through the
word embedding matrix. We can think of this as being a linear layer, where the input to this layer
is a one-hot vector with the i-th position active for word wi . Then multiplying this vector by the
embedding matrix results in the word vector for wi . Thus the columns of the word representations
matrix consisting of words from w1 , . . . , wn?1 will have non-zero gradients with respect to the loss.
This allows us to consistently modify the word representations throughout training.
We construct attribute representations in a similar way. Suppose that L is an attribute lookup table,
where x = f (L(:, x)) and f is an optional non-linearity. We often use a rectifier non-linearity in
order to keep x sparse and positive, which we found made training much more stable. Initially, the
entries of L are generated randomly. During training, we treat L in the same way as the word embedding matrix. This way of learning language representations allows us to measure how ?similar?
attributes are as opposed to using a one-hot encoding of attributes for which no such similarity could
be computed.
In some cases, attributes that are available during training may not also be available at test time.
An example of this is when attributes are used as sentence indicators for learning representations
of sentences. To accommodate for this, we use an inference step similar to that proposed by [6].
That is, at test time all the network parameters are fixed and stochastic gradient descent is used for
inferring the representation of an unseen attribute vector.
3
Experiments
In this section we describe our experimental evaluation and results. Throughout this section we refer
to our model as Attribute Tensor Decomposition (ATD). All models are trained using stochastic gradient descent with an exponential learning rate decay and linear (per epoch) increase in momentum.
We first demonstrate initial qualitative results to get a sense of the tasks our model can perform. For
these, we use the small project Gutenberg corpus which consists of 18 books, some of which have
the same author. We first trained a multiplicative neural language model with a context size of 5,
4
Table 3: A modified version of the game Mad Libs. Given an initialization, the model is to generate
the next 5 words according to the part-of-speech sequence (note that these are not hard constraints).
[DT, NN, IN, DT, JJ]
the meaning of life is...
the cure of the bad
the truth of the good
a penny for the fourth
the globe of those modern
all man upon the same
[TO, VB, VBD, JJS, NNS]
my greatest accomplishment is...
to keep sold most wishes
to make manned most magnificent
to keep wounded best nations
to be allowed best arguments
to be mentioned most people
[PRP, NN, ?,? , JJ, NN]
i could not live without...
his regard , willing tenderness
her french , serious friend
her father , good voice
her heart , likely beauty
her sister , such character
Table 4: Classification accuracies on various tasks. Left: Sentiment classification on the treebank dataset. Competing methods include the Neural Bag of words (NBoW) [5], Recursive Network (RNN) [17], Matrix-Vector Recursive Network (MV-RNN) [18], Recursive Tensor Network
(RTNN) [3], Dynamic Convolutional Network (DCNN) [5] and Paragraph Vector (PV) [6]. Right:
Cross-lingual classification on RCV2. Methods include statistical machine translation (SMT), IMatrix [19], Bag-of-words autoencoders (BAE-*) [20] and BiCVM, BiCVM+ [21]. The use of ?+?
on cross-lingual tasks indicate the use of a third language (French) for learning embeddings.
Method
SVM
BiNB
NBoW
RNN
MVRNN
RTNN
DCNN
PV
ATD
Fine-grained
40.7%
41.9%
42.4%
43.2%
44.4%
45.7%
48.5%
48.7%
45.9%
Positive / Negative
79.4%
83.1%
80.5%
82.4%
82.9%
85.4%
86.8%
87.8%
83.3%
Method
SMT
I-Matrix
BAE-cr
BAE-tree
BiCVM
BiCVM+
BAE-corr
ATD
ATD+
EN ? DE
68.1%
77.6%
78.2%
80.2%
83.7%
86.2%
91.8%
80.8%
83.4%
DE ? EN
67.4%
71.1%
63.6%
68.2%
71.4%
76.9%
72.8%
71.8%
72.9%
where each attribute is represented as a book. This results in 18 learned attribute vectors, one for
each book. After training, we can condition on a book vector and generate samples from the model.
Table 2 illustrates some the generated samples. Our model learns to capture the ?style? associated
with different books. Furthermore, by conditioning on the average of book representations, the
model can generate reasonable samples that represent a hybrid of both attributes, even though such
attribute combinations were not observed during training.
Next, we computed POS sequences from sentences that occur in the training corpus. We trained
a multiplicative neural language model with a context size of 5 to predict the next word from its
context, given knowledge of the POS tag for the next word. That is, we model P (wn = i|w1:n?1 , x)
where x denotes the POS tag for word wn . After training, we gave the model an initial input and
a POS sequence and proceeded to generate samples. Table 3 shows some results for this task.
Interestingly, the model can generate rather funny and poetic completions to the initial context.
3.1
Sentiment classification
Our first quantitative experiments are performed on the sentiment treebank of [3]. A common challenge for sentiment classification tasks is that the global sentiment of a sentence need not correspond
to local sentiments exhibited in sub-phrases of the sentence. To address this issue, [3] collected annotations from the movie reviews corpus of [22] of all subphrases extracted from a sentence parser.
By incorporating local sentiment into their recursive architectures, [3] was able to obtain significant
performance gains with recursive networks over bag of words baselines.
We follow the same experimental procedure proposed by [3] for which evaluation is reported on
two tasks: fine-grained classification of categories {very negative, negative, neutral, positive, very
positive } and binary classification {positive, negative }. We extracted all subphrases of sentences
that occur in the training set and used these to train a multiplicative neural language model. Here,
each attribute is represented as a sentence vector, as in [6]. In order to compute subphrases for
unseen sentences, we apply an inference procedure similar to [6], where the weights of the network
are frozen and gradient descent is used to infer representations for each unseen vector. We trained a
logistic regression classifier using all training subphrases in the training set. At test time, we infer a
representation for a new sentence which is used for making a review prediction. We used a context
5
size of 8, 100 dimensional word vectors initialized from [2] and 100 dimensional sentence vectors
initialized by averaging vectors of words from the corresponding sentence.
Table 4, left panel, illustrates our results on this task in comparison to all other proposed approaches.
Our results are on par with the highest performing recursive network on the fine-grained task and
outperforms all bag-of-words baselines and recursive networks with the exception of the RTNN on
the binary task. Our method is outperformed by the two recently proposed approaches of [5] (a
convolutional network trained on sentences) and Paragraph Vector [6].
3.2
Cross-lingual document classification
We follow the experimental procedure of [19], for which several existing baselines are available to
compare our results. The experiment proceeds as follows. We first use the Europarl corpus [23] for
inducing word representations across languages. Let S be a sentence with words w in language `
and let x be the corresponding language vector. Let
X
X
v` (S) =
T ` (:, w) =
(W`f v (:, w))> ? diag(Wf d x) ? Wf k
(7)
w?S
w?S
denote the sentence representation of S, defined as the sum of language conditioned word representations for each w ? S. Equivalently we define a sentence representation for the translation S 0 of S
denoted as v`0 (S 0 ). We then optimize the following ranking objective:
XX
2
2
2
minimize
max 0, ? +
v` (S) ? v`0 (S 0 )
2 ?
v` (S) ? v`0 (Ck )
2 + ?
?
2
?
S
k
subject to the constraints that each sentence vector has unit norm. Each Ck is a constrastive (nontranslation) sentence of S and ? denotes all model parameters. This type of cross-language ranking
loss was first used by [21] but without the norm constraint which we found significantly improved
the stability of training. The Europarl corpus contains roughly 2 million parallel sentence pairs
between English and German as well as English and French, for which we induce 40 dimensional
word representations. Evaluation is then performed on English and German sections of the Reuters
RCV1/RCV2 corpora. Note that these documents are not parallel. The Reuters dataset contains
multiple labels for each document. Following [19], we only consider documents which have been
assigned to one of the top 4 categories in the label hierarchy. These are CCAT (Corporate/Industrial),
ECAT (Economics), GCAT (Government/Social) and MCAT (Markets). There are a total of 34,000
English documents and 42,753 German documents with vocabulary sizes of 43614 English words
and 50,110 German words. We consider both training on English and evaluating on German and
vice versa. To represent a document, we sum over the word representations of words in that document followed by a unit-ball projection. Following [19] we use an averaged perceptron classifier.
Classification accuracy is then evaluated on a held-out test set in the other language. We used a
monolingual validation set for tuning the margin ?, which was set to ? = 1. Five contrastive terms
were used per example which were randomly assigned per epoch.
Table 4, right panel, shows our results compared to all proposed methods thus far. We are competitive with the current state-of-the-art approaches, being outperformed only by BiCVM+ [21] and
BAE-corr [20] on EN ? DE. The BAE-corr method combines both a reconstruction term and a
correlation regularizer to match sentences, while our method does not consider reconstruction. We
also performed experimentation on a low resource task, where we assume the same conditions as
above with the exception that we only use 10,000 parallel sentence pairs between English and German while still incorporating all English and French parallel sentences. For this task, we compare
against a separation baseline, which is the same as our model but with no parameter sharing across
languages (and thus resembles [21]). Here we achieve 74.7% and 69.7% accuracies (EN?DE and
DE?EN) while the separation baseline obtains 63.8% and 67.1%. This indicates that parameter sharing across languages can be useful when only a small amount of parallel data is available.
Figure 2 further shows t-SNE embeddings of English-German word pairs.1
Another interesting consideration is whether or not the learned language vectors can capture any
interesting properties of various languages. To look into this, we trained a multiplicative neural
language model simultaneously on 5 languages: English, French, German, Czech and Slovak. To
our knowledge, this is the most languages word representations have been jointly learned on. We
1
We note that Germany and Deutschland are nearest neighbours in the original space.
6
(a) Months
(b) Countries
5
4
3
2
1
0.3
0.2
0.1
0.0
? 0.1
uncondit ioned ATD
? 0.2
0
(a) Correlation matrix
uncondit ioned ATD
LBL
condit ioned ATD
6
Inferred at t ribut es difference
Im provem ent over init ial m odel
Figure 2: t-SNE embeddings of English-German word pairs learned from Europarl.
5
10
25
50
100
# Docum ent s (t housands)
382
5
10
25
50
100
# Docum ent s (t housands)
382
(b) Effect of conditional embeddings (c) Effect of inferring attribute vectors
Figure 3: Results on the Blog classification corpus. For the middle and right plots, each pair of same
coloured bars corresponds to the non-inclusion or inclusion of inferred attribute vectors, respectively.
computed a correlation matrix from the language vectors, illustrated in Fig. 3a. Interestingly, we
observe high correlation between Czech and Slovak representations, indicating that the model may
have learned some notion of lexical similarity. That being said, additional experimentation for future
work is necessary to better understand the similarities exhibited through language vectors.
3.3
Blog authorship attribution
For our final task, we use the Blog corpus of [24] which contains 681,288 blog posts from 19,320
authors. For our experiments, we break the corpus into two separate datasets: one containing the
1000 most prolific authors (most blog posts) and the other containing all the rest. Each author comes
with an attribute tag corresponding to a tuple (age, gender, industry) indicating the age range of the
author (10s, 20s or 30s), whether the author is male or female, and what industry the author works
in. Note that industry does not necessary correspond to the topic of blog posts. We use the dataset
of non-prolific authors to train a multiplicative language model conditioned on an attribute tuple
of which there are 234 unique tuples in total. We used 100 dimensional word vectors initialized
from [2], 100 dimensional attribute vectors with random initialization and a context size of 5. A
1000-way classification task is then performed on the prolific author subset and evaluation is done
using 10-fold cross-validation. Our initial experimentation with baselines found that tf-idf performs
well on this dataset (45.9% accuracy). Thus, we consider how much we can improve on the tf-idf
baseline by augmenting word and attribute features.
For the first experiment, we determine the effect conditional word embeddings have on classification
performance, assuming attributes are available at test time. For this, we compute two embedding
matrices from a trained ATD model, one without and with attribute knowledge:
unconditioned ATD :
conditioned ATD :
(Wf v )> Wf k
(8)
(Wf v )> ? diag(Wf d x) ? Wf k .
(9)
We represent a blog post as the sum of word vectors projected to unit norm and augment these with
tf-idf features. As an additional baseline we include a log-bilinear language model [14]. 2 Figure
3b illustrates the results from which we observe that conditioned word embeddings are significantly
more discriminative over word embeddings computed without knowledge of attribute vectors.
2
The log-bilinear model has no concept of attributes.
7
Table 5: Results from a conditional word similarity task using Blog attributes and language vectors.
Query,A,B
school
f/10/student
m/20/tech
journal
f/10/student
m/30/adv.
create
f/30/arts
f/30/internet
joy
m/30/religion
m/20/science
cool
m/10/student
f/10/student
Common
work
church
college
diary
blog
webpage
build
develop
maintain
happiness
sadness
pain
nice
funny
awesome
Unique to A
choir
prom
skool
project
book
yearbook
provide
acquire
generate
rapture
god
heartbreak
beautiful
amazing
neat
Unique to B
therapy
tech
job
zine
app
referral
compile
follow
analyse
delight
comfort
soul
sexy
hott
lame
English
january
june
october
market
markets
internal
war
weapons
global
said
stated
told
two
two-thirds
both
French
janvier
decembre
juin
marche
marches
interne
guerre
terrorisme
mondaile
dit
disait
declare
deux
deuxieme
seconde
German
januar
dezember
juni
markt
binnenmarktes
marktes
krieg
globale
krieges
sagte
gesagt
sagten
zwei
beiden
zweier
For the second experiment, we determine the effect of inferring attribute vectors at test time if they
are not assumed to be available. To do this, we train a logistic regression classifier within each fold
for predicting attributes. We compute an inferred vector by averaging each of the attribute vectors
weighted by the log-probabilities of the classifier. In Fig. 3c we plot the difference in performance
when an inferred vector is augmented vs. when it is not. These results show consistent, albeit small
improvement gains when attribute vectors are inferred at test time.
To get a better sense of the attribute features learned from the model, the supplementary material
contains a t-SNE embedding of the learned attribute vectors. Interestingly, the model learns features
which largely isolate the vectors of all teenage bloggers independent of gender and topic.
3.4
Conditional word similarity
One of the key properties of our tensor formulation is the notion of conditional word similarity,
namely how neighbours of word representations change depending on the attributes that are conditioned on. In order to explore the effects of this, we performed two qualitative comparisons: one
using blog attribute vectors and the other with language vectors. These results are illustrated in
Table 5. For the first comparison on the left, we chose two attributes from the blog corpus and a
query word. We identify each of these attribute pairs as A and B. Next, we computed a ranked list of
the nearest neighbours (by cosine similarity) of words conditioned on each attribute and identified
the top 15 words in each. Out of these 15 words, we display the top 3 words which are common
to both ranked lists, as well as 3 words that are unique to a specific attribute. Our results illustrate
that the model can capture distinctive notions of word similarities depending on which attributes
are being conditioned. On the right of Table 5, we chose a query word in English (italicized) and
computed the nearest neighbours when conditioned on each language vector. This results in neighbours that are either direct translations of the query word or words that are semantically similar. The
supplementary material includes additional examples with nearest neighbours of collocations.
4
Conclusion
There are several future directions from which this work can be extended. One application area
of interest is in learning representations of authors from papers they choose to review as a way of
improving automating reviewer-paper matching [25]. Since authors contribute to different research
topics, it might be more useful to instead consider a mixture of attribute vectors that can allow for
distinctive representations of the same author across research areas. Another interesting application
is learning representations of graphs. Recently, [26] proposed an approach for learning embeddings
of nodes in social networks. Introducing network indicator vectors could allow us to potentially
learn representations of full graphs. Finally, it would be interesting to train a multiplicative neural
language model simultaneously across dozens of languages.
Acknowledgments
We would also like to thank the anonymous reviewers for their valuable comments and suggestions.
This work was supported by NSERC, Google, Samsung, and ONR Grant N00014-14-1-0232.
8
References
[1] Ronan Collobert and Jason Weston. A unified architecture for natural language processing: Deep neural
networks with multitask learning. In ICML, pages 160?167, 2008.
[2] Joseph Turian, Lev Ratinov, and Yoshua Bengio. Word representations: a simple and general method for
semi-supervised learning. In ACL, pages 384?394, 2010.
[3] Richard Socher, Alex Perelygin, Jean Y Wu, Jason Chuang, Christopher D Manning, Andrew Y Ng, and
Christopher Potts. Recursive deep models for semantic compositionality over a sentiment treebank. In
EMNLP, pages 1631?1642, 2013.
[4] Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. Distributed representations of
words and phrases and their compositionality. In NIPS, pages 3111?3119, 2013.
[5] Phil Blunsom, Edward Grefenstette, Nal Kalchbrenner, et al. A convolutional neural network for modelling sentences. In ACL, 2014.
[6] Quoc V Le and Tomas Mikolov. Distributed representations of sentences and documents. ICML, 2014.
[7] Antoine Bordes, Nicolas Usunier, Alberto Garcia-Duran, Jason Weston, and Oksana Yakhnenko. Translating embeddings for modeling multi-relational data. In NIPS, pages 2787?2795, 2013.
[8] Richard Socher, Danqi Chen, Christopher D Manning, and Andrew Ng. Reasoning with neural tensor
networks for knowledge base completion. In NIPS, pages 926?934, 2013.
[9] Yann N Dauphin, Gokhan Tur, Dilek Hakkani-Tur, and Larry Heck. Zero-shot learning for semantic
utterance classification. ICLR, 2014.
[10] Andrea Frome, Greg S Corrado, Jon Shlens, Samy Bengio, Jeffrey Dean, and Tomas Mikolov MarcAurelio Ranzato. Devise: A deep visual-semantic embedding model. NIPS, 2013.
[11] Graham W Taylor and Geoffrey E Hinton. Factored conditional restricted boltzmann machines for modeling motion style. In ICML, pages 1025?1032, 2009.
[12] Ryan Kiros, Richard S Zemel, and Ruslan Salakhutdinov. Multimodal neural language models. ICML,
2014.
[13] Ilya Sutskever, James Martens, and Geoffrey E Hinton. Generating text with recurrent neural networks.
In ICML, pages 1017?1024, 2011.
[14] Andriy Mnih and Geoffrey Hinton. Three new graphical models for statistical language modelling. In
ICML, pages 641?648, 2007.
[15] Roland Memisevic and Geoffrey Hinton. Unsupervised learning of image transformations. In CVPR,
pages 1?8, 2007.
[16] Alex Krizhevsky, Geoffrey E Hinton, et al. Factored 3-way restricted boltzmann machines for modeling
natural images. In AISTATS, pages 621?628, 2010.
[17] Richard Socher, Jeffrey Pennington, Eric H Huang, Andrew Y Ng, and Christopher D Manning. Semisupervised recursive autoencoders for predicting sentiment distributions. In EMNLP, pages 151?161,
2011.
[18] Richard Socher, Brody Huval, Christopher D Manning, and Andrew Y Ng. Semantic compositionality
through recursive matrix-vector spaces. In EMNLP, pages 1201?1211, 2012.
[19] Alexandre Klementiev, Ivan Titov, and Binod Bhattarai. Inducing crosslingual distributed representations
of words. In COLING, pages 1459?1474, 2012.
[20] Sarath Chandar A P, Stanislas Lauly, Hugo Larochelle, Mitesh M Khapra, Balaraman Ravindran, Vikas
Raykar, and Amrita Saha. An autoencoder approach to learning bilingual word representations. NIPS,
2014.
[21] Karl Moritz Hermann and Phil Blunsom. Multilingual distributed representations without word alignment.
ICLR, 2014.
[22] Bo Pang and Lillian Lee. Seeing stars: Exploiting class relationships for sentiment categorization with
respect to rating scales. In ACL, pages 115?124, 2005.
[23] Philipp Koehn. Europarl: A parallel corpus for statistical machine translation. In MT summit, volume 5,
pages 79?86, 2005.
[24] Jonathan Schler, Moshe Koppel, Shlomo Argamon, and James W Pennebaker. Effects of age and gender
on blogging. In AAAI Spring Symposium: Computational Approaches to Analyzing Weblogs, volume 6,
pages 199?205, 2006.
[25] Laurent Charlin, Richard S Zemel, and Craig Boutilier. A framework for optimizing paper matching.
UAI, 2011.
[26] Bryan Perozzi, Rami Al-Rfou, and Steven Skiena. Deepwalk: Online learning of social representations.
KDD, 2014.
9
| 5425 |@word multitask:1 proceeded:1 version:1 middle:2 norm:3 duran:1 willing:1 decomposition:2 contrastive:1 shot:1 accommodate:1 initial:4 contains:4 score:1 document:12 interestingly:3 outperforms:1 existing:1 current:3 written:1 lauly:1 ronan:1 shlomo:1 kdd:1 plot:2 joy:2 v:3 ial:1 contribute:1 toronto:2 node:1 philipp:1 five:1 along:2 direct:1 symposium:1 qualitative:3 consists:2 combine:1 paragraph:3 manner:1 x0:1 ravindran:1 market:3 andrea:1 roughly:1 love:1 kiros:2 multi:1 salakhutdinov:2 project:2 xx:1 moreover:1 linearity:2 panel:2 israel:1 what:2 kind:1 lame:1 teenage:1 unified:1 possession:1 transformation:1 quantitative:3 act:1 golden:1 nation:1 classifier:4 klementiev:1 unit:4 grant:1 before:1 positive:5 declare:1 local:2 modify:1 treat:1 bilinear:8 encoding:1 analyzing:1 lev:1 laurent:1 ioned:3 might:1 chose:2 acl:3 initialization:2 resembles:1 blunsom:2 sadness:1 compile:1 factorization:1 bi:2 range:1 averaged:1 unique:4 acknowledgment:1 recursive:10 backpropagation:1 procedure:3 area:2 rnn:3 significantly:2 yakhnenko:1 deed:1 projection:1 word:102 pre:1 induce:1 matching:2 seeing:1 get:2 deepwalk:1 context:16 live:1 writing:1 rcv2:3 optimize:1 deterministic:1 reviewer:2 lexical:1 dean:2 phil:2 attribution:5 economics:1 marten:1 tomas:3 factored:5 sarath:1 deux:1 shlens:1 his:2 docum:2 embedding:11 stability:1 notion:7 traditionally:1 updated:1 hierarchy:1 suppose:1 parser:1 unshared:2 samy:1 summit:1 observed:1 steven:1 capture:6 adv:1 ranzato:1 highest:1 tur:2 valuable:1 thee:1 substantial:1 mentioned:1 pd:1 slovak:2 ramus:1 dynamic:1 employment:1 trained:7 depend:1 lord:1 upon:1 distinctive:2 eric:1 basis:1 po:4 samsung:1 multimodal:1 represented:6 various:3 regularizer:1 train:4 describe:4 query:4 zemel:4 tell:1 kalchbrenner:1 whose:2 jean:1 supplementary:2 valued:2 kai:1 cvpr:1 koehn:1 ability:1 god:2 unseen:3 think:1 jointly:4 analyse:1 final:1 unconditioned:1 unfactored:1 online:1 sequence:4 frozen:1 propose:3 blogger:2 interaction:5 product:1 reconstruction:2 till:2 achieve:2 inducing:2 ent:3 webpage:1 sutskever:2 exploiting:1 generating:2 categorization:1 illustrate:1 andrew:4 depending:3 augmenting:1 amazing:1 recurrent:4 nearest:4 friend:1 completion:2 develop:1 school:1 edward:1 job:1 strong:3 c:1 predicted:3 come:4 indicate:1 cool:1 frome:1 direction:1 larochelle:1 hermann:1 attribute:78 stochastic:2 nlm:4 translating:1 material:2 larry:1 government:1 generalization:1 anonymous:1 ryan:2 im:1 weblogs:1 therapy:1 exp:4 great:1 rfou:1 predict:2 bj:2 ruslan:2 outperformed:2 bag:4 label:2 him:2 vice:1 tf:3 create:1 weighted:3 dcnn:2 rwi:1 modified:1 rather:1 ck:2 schler:1 cr:1 beauty:1 june:1 koppel:1 improvement:2 consistently:1 modelling:4 indicates:1 potts:1 prp:1 contrast:1 industrial:1 balaraman:1 baseline:8 sense:2 wf:24 am:1 inference:2 tech:2 dependent:1 nn:3 collocation:1 initially:1 hidden:1 her:4 germany:1 issue:1 classification:20 prom:1 dauphin:1 denoted:1 augment:1 art:2 oksana:1 once:1 construct:1 ng:4 ribut:1 look:1 icml:6 unsupervised:1 jon:1 thinking:1 caesar:4 future:2 stanislas:1 others:1 yoshua:1 richard:7 serious:1 prolific:3 saha:1 modern:1 neighbour:8 imatrix:1 simultaneously:2 randomly:2 replaced:1 consisting:2 skiena:1 jeffrey:2 maintain:1 interest:1 highly:1 mnih:1 evaluation:4 alignment:1 male:1 sexy:1 mixture:1 bible:2 argamon:1 atd:10 held:1 tuple:3 necessary:3 mcat:1 tree:1 taylor:1 detailing:1 initialized:3 lbl:3 re:1 instance:1 industry:7 column:3 soft:1 modeling:3 phrase:3 cost:1 happiness:1 introducing:1 entry:1 neutral:1 subset:1 father:1 krizhevsky:1 gutenberg:3 too:1 reported:1 my:1 nns:1 automating:1 memisevic:1 told:1 lee:1 ilya:2 w1:11 aaai:1 opposed:2 vbd:1 containing:2 choose:1 emnlp:3 huang:1 book:9 style:6 de:5 lookup:2 huval:1 heck:1 student:4 star:1 includes:1 chandar:1 mv:1 ranking:2 collobert:1 multiplicative:16 bone:1 performed:5 break:1 jason:3 bhattarai:1 competitive:1 parallel:8 odel:1 annotation:1 minimize:1 pang:1 accuracy:4 convolutional:3 greg:2 characteristic:2 largely:1 correspond:5 identify:1 ant:1 ccat:1 craig:1 multiplying:1 app:1 sharing:2 against:2 james:2 associated:4 gain:3 dataset:5 recall:1 knowledge:5 gcat:1 appears:1 alexandre:1 dt:2 supervised:1 follow:3 improved:1 delight:2 formulation:3 done:2 though:1 evaluated:1 charlin:1 furthermore:2 autoencoders:2 correlation:4 christopher:5 google:1 french:8 logistic:2 semisupervised:1 name:1 effect:6 concept:3 hence:1 assigned:2 moritz:1 semantic:5 illustrated:2 during:4 game:1 raykar:1 backpropagating:1 authorship:5 cosine:1 juni:1 demonstrate:3 performs:1 motion:3 reasoning:1 meaning:2 image:4 wise:2 consideration:1 recently:3 common:3 mt:1 hugo:1 conditioning:2 volume:2 crosslingual:2 million:1 refer:1 significant:1 versa:1 enjoyed:1 rd:1 tuning:1 inclusion:2 language:62 had:1 stable:1 similarity:12 base:1 female:1 optimizing:1 diary:1 certain:1 n00014:1 meta:1 blog:16 success:1 binary:2 life:1 onr:1 devise:1 seen:1 greater:1 additional:3 determine:2 corrado:2 semi:1 rv:3 multiple:1 corporate:1 rj:1 infer:2 full:1 match:1 cross:9 alberto:1 post:4 roland:1 hakkani:1 prediction:2 regression:2 globe:1 represent:6 whereas:1 fine:3 hott:1 country:1 concluded:1 weapon:1 rest:1 perozzi:1 exhibited:2 comment:1 smt:2 subject:1 isolate:1 near:2 canadian:1 feedforward:1 embeddings:14 wn:14 bengio:2 ivan:1 variety:2 switch:1 ecat:1 gave:1 loved:1 architecture:2 competing:1 awesome:1 identified:1 andriy:1 prototype:2 accomplishment:1 whether:2 war:1 sentiment:14 speech:2 jj:2 deep:3 boutilier:1 useful:2 amount:3 processed:1 category:3 rw:1 dit:1 generate:9 per:3 sister:1 bryan:1 diverse:1 shall:1 key:1 blood:1 nal:1 kept:1 timestep:1 bae:6 graph:2 sum:3 ratinov:1 mystery:1 you:1 fourth:1 throughout:3 saying:1 reasonable:1 wu:1 yann:1 funny:2 separation:2 vb:1 graham:1 brody:1 layer:3 internet:1 followed:1 display:1 fold:2 strength:1 occur:2 constraint:3 idf:3 alex:2 ri:1 tag:4 argument:2 spring:2 rcv1:2 performing:1 mikolov:3 structured:1 according:1 combination:3 ball:1 march:1 lingual:7 manning:4 across:10 character:4 wi:5 joseph:1 rsalakhu:1 making:1 quoc:1 restricted:3 heart:1 resource:1 previously:1 discus:1 german:12 usunier:1 available:6 constrastive:1 experimentation:3 apply:1 observe:3 titov:1 rapture:2 voice:1 shortly:1 gate:2 marcaurelio:1 vikas:1 original:1 chuang:1 denotes:5 assumes:1 nlp:2 running:1 include:3 top:3 graphical:1 build:1 tensor:14 objective:1 moshe:1 rt:2 diagonal:1 antoine:1 said:2 gradient:4 pain:1 iclr:2 separate:2 thank:1 entity:1 parametrized:1 me:1 topic:3 italicized:1 collected:1 mad:1 assuming:1 relationship:2 multiplicatively:1 condit:1 acquire:1 equivalently:1 october:1 sne:3 potentially:1 perelygin:1 negative:4 rise:1 stated:1 proper:1 boltzmann:3 subphrases:4 perform:3 gated:1 yearbook:1 datasets:1 sold:1 lillian:1 descent:3 mitesh:1 optional:1 january:1 extended:2 incorporated:1 relational:1 tenderness:1 hinton:5 inferred:5 compositionality:3 introduced:3 rating:1 pair:6 namely:1 sentence:29 fv:1 learned:12 czech:2 nip:5 address:1 beyond:1 able:1 proceeds:1 bar:1 comfort:2 soul:1 ghost:1 challenge:1 rf:3 including:1 max:1 analogue:1 hot:3 greatest:1 ranked:2 hybrid:1 beautiful:1 predicting:3 indicator:5 natural:2 advanced:1 improve:2 movie:1 church:1 metadata:4 utterance:1 autoencoder:1 text:13 review:4 epoch:2 coloured:1 nice:1 embedded:2 fully:1 loss:2 par:1 monolingual:1 generation:3 men:1 interesting:5 suggestion:1 geoffrey:5 age:5 validation:2 consistent:1 blogging:1 treebank:4 share:1 bordes:1 translation:4 karl:1 summary:1 supported:1 last:1 english:17 neat:1 side:2 bias:1 understand:1 perceptron:1 institute:1 wide:2 allow:2 sparse:1 penny:1 distributed:10 slice:2 regard:1 vocabulary:9 evaluating:1 cure:1 unto:2 author:19 qualitatively:1 collection:1 reside:1 made:1 projected:1 far:2 social:3 obtains:1 multilingual:1 keep:3 global:2 active:1 uai:1 corpus:14 unnecessary:1 assumed:1 tuples:1 xi:2 discriminative:1 continuous:1 table:14 learn:7 nicolas:1 init:1 improving:1 interact:2 diag:5 aistats:1 reuters:2 bilingual:1 turian:1 allowed:1 body:1 augmented:1 fig:4 en:5 sub:1 position:1 pv:4 inferring:3 momentum:1 exponential:1 wish:1 breaking:1 third:4 learns:2 grained:3 coling:1 dozen:1 rk:1 embed:1 bad:1 specific:6 rectifier:1 gating:2 symbol:1 list:2 decay:1 svm:1 exists:2 incorporating:2 socher:4 albeit:1 corr:3 pennington:1 conditioned:14 illustrates:4 margin:1 chen:2 danqi:1 led:1 garcia:1 likely:1 explore:1 visual:1 religion:2 nserc:1 bo:1 fear:1 poetic:1 gender:5 corresponds:4 truth:1 extracted:2 weston:2 conditional:15 jjs:1 goal:1 viewed:1 king:1 sized:1 month:1 grefenstette:1 jeff:1 shared:1 man:1 markt:1 change:3 tiger:1 hard:1 specifically:1 folded:1 semantically:1 averaging:2 rkiros:1 total:2 experimental:4 e:1 exception:2 indicating:2 college:1 internal:1 people:1 jonathan:1 evaluate:1 europarl:5 |
4,889 | 5,426 | Sparse Polynomial Learning and Graph Sketching
Murat Kocaoglu1? , Karthikeyan Shanmugam1? , Alexandros G.Dimakis1? , Adam Klivans2?
1
Department of Electrical and Computer Engineering, 2 Department of Computer Science
The University of Texas at Austin, USA
?
[email protected], ? [email protected]
?
[email protected], ? [email protected]
Abstract
Let f : {?1, 1}n ? R be a polynomial with at most s non-zero real coefficients.
We give an algorithm for exactly reconstructing f given random examples from
the uniform distribution on {?1, 1}n that runs in time polynomial in n and 2s
and succeeds if the function satisfies the unique sign property: there is one output
value which corresponds to a unique set of values of the participating parities. This
sufficient condition is satisfied when every coefficient of f is perturbed by a small
random noise, or satisfied with high probability when s parity functions are chosen
randomly or when all the coefficients are positive. Learning sparse polynomials
over the Boolean domain in time polynomial in n and 2s is considered notoriously
hard in the worst-case. Our result shows that the problem is tractable for almost
all sparse polynomials.
Then, we show an application of this result to hypergraph sketching which is the
problem of learning a sparse (both in the number of hyperedges and the size of
the hyperedges) hypergraph from uniformly drawn random cuts. We also provide
experimental results on a real world dataset.
1
Introduction
Learning sparse polynomials over the Boolean domain is one of the fundamental problems from
computational learning theory and has been studied extensively over the last twenty-five years [1?
6]. In almost all cases, known algorithms for learning or interpolating sparse polynomials require
query access to the unknown polynomial. An outstanding open problem is to find an algorithm
for learning s-sparse polynomials with respect to the uniform distribution on {?1, 1}n that runs in
time polynomial in n and g(s) (where g is any fixed function independent of n) and requires only
randomly chosen examples to succeed. In particular, such an algorithm would imply a breakthrough
result for the problem of learning k-juntas (functions that depend on only k n input variables; it
is not known how to learn ?(1)-juntas in polynomial time).
We present an algorithm and a set of natural conditions such that any sparse polynomial f satisfying these conditions can be learned from random examples in time polynomial in n and 2s . In
particular, any f whose coefficients have been subjected to a small perturbation (smoothed analysis
setting) satisfies these conditions (for example, if a Gaussian with arbitrarily small variance is added
independently to each coefficient, f satisfies these conditions with probability 1). We state our main
result here:
Theorem 1. Let f be an s-sparse function that satisfies at least one of the following properties:
a) (smoothed analysis setting)The coefficients {ci }si=1 are in general position or all of them are
perturbed by a small random noise. b) The s parity functions are linearly independent. c) All the
coefficients are positive. Then we learn f with high probability in time poly(n, 2s ).
1
We note that smoothed-analysis, pioneered in [7], has now become a common alternative for problems that seem intractable in the worst-case.
Our algorithm also succeeds in the presence of noise:
Theorem 2. Let f = f1 + f2 be a polynomial such that f1 and f2 depend on mutually disjoint set
of parity functions. f1 is s-sparse and the values of f1 are ?well separated?. Further, kf2 k1 ? ?,
(i.e., f is approximately sparse). If observations are corrupted by additive noise bounded by , then
there exists an algorithm which takes + ? as an input, that gives g in time polynomial in n and 2s
such that kf ? gk2 ? O(? + ) with high probability.
The treatment of the noisy case, i.e., the formal statement of this theorem, the corresponding algorithm, and the related proofs are relegated to the supplementary material. All these results are
based on what we call as the unique sign property: If there is one value that f takes which uniquely
specifies the signs of the parity functions involved, then the function is efficiently learnable. Note
that our results cannot be used for learning juntas or other Boolean-valued sparse polynomials, since
the unique sign property does not hold in these settings.
We show that this property holds for the complement of the cut function on a hypergraph (no. of
hyperedges ? cut value). This fact can be used to learn the cut complement function and eventually
infer the structure of a sparse hypergraph from random cuts. Sparsity implies that the number of
hyperedges and the size of each hyperedge is of constant size. Hypergraphs can be used to represent
relations in many real world data sets. For example, one can represent the relation between the books
and the readers (users) on the Amazon dataset with a hypergraph. Book titles and Amazon users
can be mapped to nodes and hyperedges, respectively ([8]). Then a node belongs to a hyperedge, if
the corresponding book is read by the user represented by that hyperedge. When such graphs evolve
over time (and space), the difference graph filtered by time and space is often sparse. To locate
and learn the few hyperedges from random cuts in such difference graphs constitutes hypergraph
sketching. We test our algorithms on hypergraphs generated from the dataset that contain the time
stamped record of messages between Yahoo! messenger users marked with the user locations (zip
codes).
1.1
Approach and Related Work
The problem of recovering the sparsest solution of a set of underdetermined linear equations has received significant recent attention in the context of compressed sensing [9?11]. In compressed sensing, one tries to recover an unknown sparse vector using few linear observations (measurements),
possibly in the presence of noise.
The recent papers [12,13] are of particular relevance to us since they establish a connection between
learning sparse polynomials and compressed sensing. The authors show that the problem of learning
a sparse polynomial is equivalent to recovering the unknown sparse coefficient vector using linear
measurements. By applying techniques from compressed sensing theory, namely Restricted Isometry Property (see [12]) and incoherence (see [13]), the authors independently established results for
reconstructing sparse polynomials using convex optimization. The results have near-optimal sample
complexity. However, the running time of these algorithms is exponential in the underlying dimension, n. This is because the measurement matrix of the equivalent compressed sensing problem
requires one column for every possible non-zero monomial.
In this paper, we show how to solve this problem in time polynomial in n and 2s under the assumption of unique sign property on the sparse polynomial. Our key contribution is a novel identification
procedure that can reduce the list of potentially non-zero coefficients from the naive bound of 2n to
2s when the function has this property.
On the theoretical side, there has been interesting recent work of [14] that approximately learns
sparse polynomial functions when the underlying domain is Gaussian. Their results do not seem to
translate to the Boolean domain. We also note the work of [15] that gives an algorithm for learning
sparse Boolean functions with respect to a randomly chosen product distribution on {?1, 1}n . Their
work does not apply to the uniform distribution on {?1, 1}n .
On the practical side, we give an application of the theory to the problem of hypergraph sketching.
We generalize a prior work [12] that applied the compressed sensing approach discussed before to
2
graph sketching on evolving social network graphs. In our algorithm, while the sample complexity
requirements are higher, the time complexity is greatly reduced in comparison. We test our algorithms on a real dataset and show that the algorithm is able to scale well on sparse hypergraphs
created out of Yahoo! messenger dataset by filtering through time and location stamps.
2
Definitions
Consider a real-valued function over the Boolean hypercube f : {?1, 1}n ? R. Given a sequence
of labeled samples of the form hf (x), xi, where x is sampled from the uniform distribution U over
the hypercube {?1, 1}n , we are interested in an efficient algorithm that learns the function f with
high probability. Through Fourier expansion, f can be written as a linear combination of monomials:
X
f (x) =
cS ?S (x), ? x ? {?1, 1}n
(1)
S?[n]
where [n] is the set of integers from 1 to n, ?S (x) =
Q
xi and cS ? R. Let c be the vector of
i?S
coefficients cS . A monomial ?S (x) is also called a parity function. More background on Boolean
functions and the Fourier expansion can be found in [16].
In this work, we restrict ourselves to sparse polynomials f with sparsity s in the Fourier domain, i.e.,
f is a linear combination of unknown parity functions ?S1 (x), ?S2 (x), . . . ?Ss (x) with s unknown
real coefficients given by {cSi }si=1 such that cSi 6= 0, ?1 ? i ? s; all other coefficients are 0. Let
the subsets corresponding to the s parity functions form a family of sets I = {Si }si=1 . Finding I is
equivalent to finding the s parity functions.
Note: In certain places, where the context makes it clear, we slightly abuse the notation such that
the set Si identifying a specific parity function is replaced by just the index i. The coefficients may
be denoted simply by ci and the parity functions by ?i (?).
Let F2 denote the binary field. Every parity function ?i (?) can be represented by a vector pi ? Fn?1
.
2
The j-th entry pi (j) in the vector pi is 1, if j ? Si and is 0 otherwise.
Definition 1. A set of s parity functions {?i (?)}si=1 are said to be linearly independent if the corresponding set of vectors {pi }si=1 are linearly independent over F2 .
Similarly, they are said to have rank r if the dimension of the subspace spanned by {pi }si=1 is r.
Definition 2. The coefficients {ci }si=1 are said to be in general position if for all possible set of
s
P
values bi ? {0, 1, ?1}, ? 1 ? i ? s, with at least one nonzero bi ,
ci bi 6= 0
i=1
Definition 3. The coefficients {ci }si=1 are said to be ?-separated
if for all possible set of values
s
P
bi ? {0, 1, ?1}, ? 1 ? i ? s with at least one nonzero bi , ci bi > ?.
i=1
Definition 4. A sign pattern is a distinct vector of signs a = [?1 (?) , ?2 (?) , . . . ?s (?))] ?
{?1, 1}1?s assumed by the set of s parity functions.
Since this work involves switching representations between the real and the binary field, we define
a function q that does the switch.
Definition 5. q : {?1, 1}a?b ? F2a?b is a function that converts a sign matrix X to a matrix Y
over F2 such that Yij = q(Xij ) = 1 ? F2 , if Xij = ?1 and Yij = q(Xij ) = 0 ? F2 , if Xij = 1.
Clearly, it has an inverse function q ?1 such that q ?1 (Y) = X.
We also present some definitions to deal with the case when the polynomial f is not exactly s-sparse
and observations are noisy. Let 2[n] denote the power set of [n].
n
Definition 6. A polynomial f : {?1,
P 1} ? R is called approximately (s, ?)-sparse if there exists
[n]
I ? 2 with |I| = s such that
|cS | < ?, where {cS } are the Fourier coefficients as in (1).
S?I c
In other words, the sum of the absolute values of all the coefficients except the ones corresponding
to I are rather small.
3
3
Problem Setting
Suppose m labeled samples hf (x) , xim
drawn from the uniform distribution U on the Boolean
i=1 are
n
hypercube. For any B ? 2[n] , let cB ? R2 ?1 be the vector of real coefficients such that cB (S) =
n
cS , ?S ? B and cB (S) = 0, ?S ?
/ B. Let A ? Rm?2 be such that every row of A corresponds
to one random input sample x ? U . Let x also denote the row index and S ? [n] denote the
column index of A. A(x, S) = ?S (x). Let AS denote the sub matrix formed by the columns
corresponding to the subsets in S. Let I be the set consisting of the s parity functions of interest
in both the sparse and the approximately sparse cases. A sparse representation of an approximately
(s, ?)-sparse function f is fI = A(x) cI , where cI is as defined above.
We review the compressed sensing framework used in [12] and [13]. Specifically, for the remainder
of the paper, we rely on [13] as a point of reference. We review their framework and explain how
we use it to obtain our results, particularly for the noisy case.
n
Let y ? Rm and ?S ? R2 , such that ?S = 0, ?S ? S c . Note that, here S is a subset of the power
set 2[n] . Now, consider the following convex program for noisy compressed sensing in this setting:
r
1
mink?S k1 subject to
kA?S ? yk2 ? .
(2)
m
Let ?Sopt be an optimum for the program (2). Note that only the columns of A in S are used in the
program. The convex program runs in time poly (m, |S|). The incoherence property of the matrix
A in [13] implies the following.
Theorem 3. ( [13]) For any family of subsets I ? 2[n] such that |I| = s, m = 4096ns2 and
c1 = 4, c2 = 8, for any feasible point ?S of program 2, we have:
n 1/4
k?I c T S k1
(3)
k?S ? ?Sopt k2 ? c1 + c2
m
with probability at least 1 ? O 41n
When S is set to the power set 2[n] , = 0 and y is the vector of observed values for an s-sparse
polynomial, the s-sparse vector cI is a feasible point to program (2). By Theorem 3, the program
recovers the sparse vector cI and hence learns the function. The only caveat is that the complexity
is exponential in n.
The main idea behind our algorithms for noiseless and noisy sparse function learning is to ?capture?
the actual s-sparse set I of interest in a small set S : |S| = O (2s ) of coefficients by a separate
algorithm that runs in time poly(n, 2s ). Using the restricted set of coefficients S, we search for the
sparse solution under the noisy and noiseless cases using program (2).
Lemma 1. Given an algorithm that runs in time poly(n, 2s ) and generates a set of parities S such
that |S| = O (2s ) , I ? S with |I| = s, program (2) with S and m = 4096ns2 random samples
as
inputs runs in time poly(n, 2s ) and learns the correct function with probability 1 ? O 41n .
Unique Sign Pattern Property: The key property that lets us find a small S efficiently is the
unique sign pattern property. Observe that an s-sparse function can produce at most 2s different real
values. If the maximum value obtained always corresponds to a unique pattern of signs of parities,
by looking only at the random samples x corresponding to the subsequent O(n) occurrences of this
maximum value, we show that all the parity functions needed to learn f are captured in a small set
of size 2s+1 (see Lemma 2 and its proof). The unique sign property again plays an important role,
along with Theorem 3 with more technicalities added, in the noisy case, which we visit in Section 2
of the supplementary material.
In the next section, we provide an algorithm to generate the bounded set S for the noiseless case for
an s-sparse function f and provide guarantees for the algorithm formally.
4
Algorithm and Guarantees: Noiseless case
Let I be the family of s subsets {Si }si=1 each corresponding to the s parity functions ?Si (?) in an
s-sparse function f . In this section, we provide an algorithm, named LearnBool, that finds a small
4
subset S of the power set 2[n] that contains elements of I first and then uses program (2) with S.
We show that the algorithm learns f in time poly (n, 2s ) from uniformly randomly drawn labeled
samples from the Boolean hypercube with high probability under some natural conditions.
Recall that if the function is such that f (x) attains its maximum value only if
[?1 (x), ?2 (x) . . . ?s (x)] = amax ? {?1, 1}s for some unique sign pattern amax , then the function
is said to possess the unique sign property. Now we state the main technical lemma for the unique
sign property.
Lemma 2. If an s-sparse function f has the unique
sign property then, in Algorithm 1, S is such
that I ? S, |S| ? 2s+1 with probability 1 ? O n1 and runs in time poly(n, 2s ).
Proof. See the supplementary material.
The proof of the above lemma involves showing that the random matrix Ymax (see Algorithm 1) has
rank at least n ? s, leading to at most 2s solutions for each equation in (4). The feasible solutions
can be obtained by Gaussian elimination in the binary field.
Theorem 4. Let f be an s-sparse function that satisfies at least one of the following properties:
(a) The coefficients {ci }si=1 are in general position.
(b) The s parity functions are linearly independent.
(c) All the coefficients are positive.
Given labeled samples,
Algorithm 1 learns f exactly (or vopt = c) in time poly (n, 2s ) with proba
1
bility 1 ? O n .
Proof. See the supplementary material.
Smoothed Analysis Setting: Perturbing ci ?s with Gaussian random variables of standard deviation
? > 0 or by random variables drawn from any set of reasonable continuous distributions ensures
that the perturbed function satisfies property (a) with probability 1.
Random Parity Functions: When ci ?s are arbitrary and the set of s parity functions are drawn uniformly randomly from 2[n] , then property (b) holds with high probability if s is a constant.
1
Input: Sparsity parameter s, m1 = 2n2s random labeled samples {hf (xi ) , xi i}m
i=1 .
max
Pick samples {xij }nj=1
corresponding to the maximum value of f observed in all the m samples.
Stack all xij row wise into a matrix Xmax of dimensions nmax ? n.
Initialise S = ?. Let Ymax = q (Xmax ).
Find all feasible solutions p ? Fn?1
such that:
2
1nmax ?1 = Ymax p or 0nmax ?1 = Ymax p
(4)
F2n?1 .
Collect all feasible solutions p to either of the above equations in the set P ?
S = {{j ? [n] : p(j) = 1}|p ? P }.
Using m = 4096ns2 more samples (number of rows of A is m corresponding to these new
samples), solve:
?Sopt = mink?S k1 such that A?S = y,
(5)
where y is the vector of m observed values.
Set vopt = ?Sopt .
Output: vopt .
Algorithm 1: LearnBool
5
A Sparse Polynomial Learning Application: Hypergraph Sketching
Hypergraphs can be used to model the relations in real world data sets (e.g., books read by users in
Amazon). We show that the cut functions on hypergraphs satisfy the unique sign property. Learning a cut function of a sparse hypergraph from random cuts is a special case of learning a sparse
5
polynomial from samples drawn uniformly from the Boolean hypercube. To track the evolution of
large hypergraphs over a small time interval, it is enough to learn the cut function of the difference
graph which is often sparse. This is called the graph sketching problem. Previously, graph sketching
was applied to social network evolution [12]. We generalize this to hypergraphs showing that they
satisfy the unique sign property, which enable faster algorithms, and provide experimental results
on real data sets.
5.1
Graph Sketching
A hypergraph G = (V, E) is a set of vertices V along with a set E of subsets of V called the
hyperedges. The size of a hyperedge is the number of variables that the hyperedge connects. Let d
be the maximum hyperedge size of graph G. Let |V | = n and |E| = s.
A random cut S ? V is aT
set of vertices
T selected uniformly at random. Define the value of the cut S
to be c(S) = |{e ? E : e S 6= ?, e V ? S 6= ?}|. Graph sketching is the problem of identifying
the graph structure from random queries that evaluate the value of a random cut, where s n
(sparse setting). Hypergraphs naturally specify relations among a set of objects through hyperedges.
For example, Amazon users can form the set E and Amazon books can form the set V . Each user
may read a subset of books which represents the hyperedge. Learning the hypergraph corresponds
to identifying the sets of books bought by each user. For more examples of hypergraphs in real data
sets, we refer the reader to [8]. Such hypergraphs evolve over time. The difference graph between
two consecutive time instants is expected to be sparse (number of edges s and maximum hyperedge
size d are small). We are interested in learning such hypergraphs from random cut queries.
For simplicity and convenience, we consider the cut complement query, i.e., c?cut, which returns
s ? c(S). One can easily represent the c?cut query with a sparse polynomial as follows: Let node
i correspond to variable xi ? {?1, +1}. A random cut involves choosing xi uniformly randomly
from {?1, +1}. The variables assigned to +1 belong to the random cut S. The value is given by
the polynomial
?
?
!
X Y (1 + xi ) Y (1 ? xi )
X 1 ? X
Y
?
fc?cut (x) =
+
=
(1 +
xi )? . (6)
?
|I|?1
2
2
2
J ?I,
I?E i?I
i?I
I?E
i?J
|J |is even
Hence, the c?cut function is a sparse polynomial where the sparsity is at most s2d?1 . The variables
corresponding to the nodes that belong to some hyperedge appear in the polynomial. We call these
the relevant variables and the number of relevant variables is denoted by k. Note that, in our sparse
setting k ? sd. We note that for a hypergraph with no singleton hyperedge, given the c?cut function,
it is easy to recover the hyper edges from (6). Therefore, we focus on learning the c?cut function to
sketch the hypergraph.
When G is a graph with edges (of cardinality 2), the compressed sensing approach (using program
2) using the cut (or c?cut) values as measurements is shown to be very efficient in [12] in terms
of the sample complexity, i.e., the required number of queries. The run time is efficient because
total number of candidate parities is O(n2 ). However when we consider hypergraphs, i.e., when
d is a large constant, the compressed sensing approach cannot scale computationally (poly(nd )
runtime). Here, based on the theory developed, we give a faster algorithm based on the unique
sign property with sample complexity m1 = O(2k d log n + 22d+1 s2 (log n + k)) and run time of
O(m1 2k , n2 log n)).
We observe that the c?cut polynomial satisfies the unique sign property. From (6), it is evident
that the polynomial has only positive coefficients. Therefore, by Theorem 4, algorithm LearnBool
succeeds. The maximum value of the c?cut function is the number of edges. Notice that the
maximum value is definitely observed in two configurations of the relevant variables: If either all
relevant variables are +1 or all are ?1. Therefore, the maximum value is observed in every 2k?1 ?
2sd samples. Thus, a direct application of LearnBool yields poly(n, 2k?1 ) time complexity, which
improves the O(nd ) bound for small s and d.
Improving further, we provide a more efficient algorithm tailored for the hypergraph sketching problem, which makes use of the unique sign property and some other properties of the cut function.
Algorithm LearnGraph (Algorithm 4) is provided in the supplementary material.
6
4
10
Error Probability vs. ?
Runtime of LearnGraph vs. standard compressed sensing
3
0.25
Prob. of Error
Runtime (seconds)
10
2
10
1
0
0
0.2
0.15
10
10
Setting 1
Setting 3
Setting 2
Setting 4
LearnGraph
Comp. Sensing
200
400
600
No. of variables, n
800
0.1
1000
(a) Runtime vs. # of variables, d = 3 and s = 1.
1
2
3
4
5
6
7
? (# of samples/n)
8
9
10
(b) Probability of error vs. ?.
Figure 1: Performance figures comparing LearnGraph and Compressed Sensing approach.
Theorem 5. Algorithm 4 exactly learns the c?cut function with probability 1 ? O( n1 )with sample
complexity m1 = O(2k d log n + 22d+1 s2 (log n + k)) and time complexity O(2k m1 + n2 d log n)) .
Proof. See the supplementary material.
5.2
Yahoo! Messenger User Communication Pattern Dataset
We performed simulations using MATLAB on an Intel(R) Xeon(R) quad-core 3.6 GHz machine
with 16 GB RAM and 10M cache. We run our algorithm on the Yahoo! Messenger User Communication Pattern Dataset [17]. This dataset contains the timestamped user communication data, i.e.,
information about a large number of messages sent over Yahoo! Messenger, for a duration of 28
days.
Dataset: Each row represents a message. The first two columns show the day and time (time
stamp) of the message respectively. The third and fifth columns show the ID of the transmitting and
receiving users, respectively. The fourth column shows the zipcode (spatial stamp) from which this
particular message is transmitted. The sixth column shows if the transmitter was in the contact list
of the reciver user (y) or not (n). If a transmitter sends the same receiver more than one message
from the same zipcode, only the first message is shown in the dataset. In total, there are 100000
unique users and 5649 unique zipcodes.
We form a hypergraph from the dataset as follows: The transmitting users form the hyperedges and
the receiving users form the nodes of the hypergraph. A hyperedge connects a set T of users if
there is a transmitting user that sends a message to all the users in T . In any given time interval ?t
(short time interval) and small set of locations ?x specified by the number of zip codes, there are
few users who transmit (s) and they transmit to very few users (d). The complete set of nodes in the
hypergraph (n) is taken to be those receiving users who are active during m consecutive intervals
of length ?t and in a set of ?x zipcodes. This gives rise to a sparse graph. We identify the active
set of transmitting users (hyperedges) and their corresponding receivers (nodes in these hyperedges)
during a short time interval ?t and a randomly selected space interval (?x, i.e., zip codes) from a
large pool of receivers (nodes) that are observed during m intervals of length ?t. Details of ?t, m
and ?x chosen for experiments are given in Table 1. We note that n is in the order of 1000 usually.
Remark: Our task is to learn the c?cut function from the random queries, i.e., random +/-1 assignment of variables and corresponding c?cut values. The generated sparse graph contains only
hyperedges that have more than 1 node. Other hyperedges (transmitting users) with just one node in
the sparse hypergraph are not taken into account. This is because a singleton hyperedge i is always
counted in the c?cut function thereby effectively its presence is masked. First, we identify the relevant variables that participate in the sparse graph. After identifying this set of candidates, correlating
the corresponding candidate parities with the function output yields the Fourier coefficient of that
parity (see Algorithm 4).
7
Table 1: Runtime for different graphs. LG: LearnGraph, CS: Compressed sensing based alg.
(b) Runtime for d = 4 and s = 3 graph.
(a) Runtime for d = 4 and s = 1 graph.
HnH
Alg. H
LG
CS
88
159
288
556
1221
1.96
265.63
2.13
-
2.23
-
2.79
-
4.94
-
n
HH
Alg. H
LG
CS
52
104
246
412
1399
1.91
39.89
2.08
> 10823
2.08
-
2.30
-
4.98
-
(c) Simulation parameters for Fig. 1b
5.2.1
Setting No.
Interval
# of Int.
n
max(d)
max(s)
Zip. Set Size
Setting 1
Setting 2
Setting 3
Setting 4
5 min.
20 sec.
10 min.
2 min.
20
200
10
50
6822
5730
6822
6822
10
22
11
30
19
4
13
21
20
200
2
50
Performance Comparison with Compressed Sensing Approach
First, we compare the runtime of our implementation LearnGraph with the compressed sensing
based algorithm from [12]. Both algorithms correctly identify the relevant variables in all the considered range of parameters. The last step of finding the corresponding Fourier coefficients is omitted
and can be easily implemented (Algorithm 4) without significantly affecting the running time. As
can be seen in Tables 1a, 1b and Fig. 1a, LearnGraph scales well to graphs on thousands of nodes.
On the contrary, the compressed sensing approach must handle a measurement matrix of size O(nd ),
which becomes prohibitively large on graphs involving more than a few hundred nodes.
5.2.2
Error Performance of LearnGraph
Error probability (probability that the correct c?cut function is not recovered) versus the number
of samples used is plotted for four different experimental settings of ?t, ?x and m in Fig. 1b. For
each time interval, the error probability is calculated by averaging the number of errors among 100
different trials. For each value of ? (number of samples), the error probability is averaged over time
intervals to illustrate the error performance. We only keep the intervals for which the graph filtered
with the considered zipcodes contains at least one user with more than one neighbor. We find that
for the first 3 settings, the error probability decreases with more samples. For the fourth setting, d
and s are very large and hence a large number of samples are required. For that reason, the error
probability does not improve significantly. The probability of error can be reduced by repeating the
experiment multiple times and taking a majority, at the cost of significantly more samples. Our plot
shows only the probability of error without such a majority amplification.
6
Conclusions
We presented a novel algorithm for learning sparse polynomials by random samples on the Boolean
hypercube. While the general problem of learning all sparse polynomials is notoriously hard, we
show that almost all sparse polynomials can be efficiently learned using our algorithm. This is
because our unique sign property holds for randomly perturbed coefficients, in addition to several
other natural settings. As an application, we show that graph and hypergraph sketching lead to sparse
polynomial learning problems that always satisfy the unique sign property. This allows us to obtain
efficient reconstruction algorthms that outperform the previous state of the art for these problems.
An important open problem is to achieve the sample complexity of [12] while keeping the computational complexity polynomial in n.
Acknowledgments
M.K, K.S. and A.D. acknowledge the support of NSF via CCF 1422549, 1344364, 1344179 and
DARPA STTR and a ARO YIP award.
8
References
[1] E. Kushilevitz and Y. Mansour, ?Learning decision trees using the Fourier spectrum,? in SIAM
J. Comput., vol. 22, no. 6, 1993, pp. 1331?1348.
[2] Y. Mansour, ?Randomized interpolation and approximation of sparse polynomials,? in SIAM
J. Comput., vol. 24, no. 2. Philadelphia, PA: Society for Industrial and Applied Mathematics,
1995, pp. 357?368.
[3] R. Schapire and R. Sellie, ?Learning sparse multivariate polynomials over a field with queries
and counterexamples,? in JCSS: Journal of Computer and System Sciences, vol. 52, 1996.
[4] A. C. Gilbert, S. Guha, P. Indyk, S. Muthukrishnan, and M. Strauss, ?Near-optimal sparse
Fourier representations via sampling,? in Proceedings of STOC, 2002, pp. 152?161.
[5] P. Gopalan, A. Kalai, and A. Klivans, ?Agnostically learning decision trees,? in Proceedings of
STOC, 2008, pp. 527?536.
[6] A. Akavia, ?Deterministic sparse Fourier approximation via fooling arithmetic progressions,?
in Proceedings of COLT, 2010, pp. 381?393.
[7] D. Spielman and S. Teng, ?Smoothed analysis of algorithms: Why the simplex algorithm
usually takes polynomial time,? in JACM: Journal of the ACM, vol. 51, 2004.
?
[8] P. Li, ?Relational learning with hypergraphs,? Ph.D. dissertation, Ecole
Polytechnique F?ed?erale
de Lausanne, 2013.
[9] E. J. Cand`es, J. Romberg, and T. Tao, ?Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information,? Information Theory, IEEE Transactions
on, vol. 52, no. 2, pp. 489?509, 2006.
[10] E. J. Cand`es and T. Tao, ?Decoding by linear programming,? Information Theory, IEEE Transactions on, vol. 51, no. 12, pp. 4203?4215, 2005.
[11] D. L. Donoho, ?Compressed sensing,? Information Theory, IEEE Transactions on, vol. 52,
no. 4, pp. 1289?1306, 2006.
[12] P. Stobbe and A. Krause, ?Learning Fourier sparse set functions,? in Proceedings of the International Conference on Artificial Intelligence and Statistics, 2012, pp. 1125?1133.
[13] S. Negahban and D. Shah, ?Learning sparse boolean polynomials,? in Proceedings of the Communication, Control, and Computing (Allerton), 2012 50th Annual Allerton Conference on.
IEEE, 2012, pp. 2032?2036.
[14] A. Andoni, R. Panigrahy, G. Valiant, and L. Zhang, ?Learning sparse polynomial functions,?
in Proceedings of SODA, 2014.
[15] A. T. Kalai, A. Samorodnitsky, and S.-H. Teng, ?Learning and smoothed analysis,? in Proceedings of FOCS. IEEE Computer Society, 2009, pp. 395?404.
[16] R. O?Donnell, Analysis of Boolean Functions.
Cambridge University Press, 2014.
[17] Yahoo, ?Yahoo! webscope dataset ydata-ymessenger-user-communication-pattern-v1 0,? http:
//research.yahoo.com/Academic Relations.
9
| 5426 |@word trial:1 polynomial:45 nd:3 kf2:1 open:2 simulation:2 pick:1 thereby:1 configuration:1 contains:4 ecole:1 ka:1 comparing:1 recovered:1 com:1 si:15 written:1 must:1 fn:2 subsequent:1 additive:1 plot:1 v:4 intelligence:1 selected:2 core:1 short:2 junta:3 record:1 alexandros:1 filtered:2 caveat:1 dissertation:1 node:12 location:3 allerton:2 zhang:1 five:1 along:2 c2:2 direct:1 become:1 ns2:3 focs:1 expected:1 cand:2 bility:1 actual:1 quad:1 cache:1 cardinality:1 becomes:1 provided:1 bounded:2 underlying:2 notation:1 what:1 dimakis:1 developed:1 finding:3 nj:1 guarantee:2 every:5 runtime:8 exactly:4 prohibitively:1 rm:2 k2:1 control:1 appear:1 positive:4 before:1 engineering:1 sd:2 switching:1 id:1 incoherence:2 interpolation:1 approximately:5 abuse:1 studied:1 collect:1 lausanne:1 bi:6 range:1 averaged:1 unique:22 practical:1 acknowledgment:1 procedure:1 evolving:1 significantly:3 word:1 nmax:3 cannot:2 vopt:3 f2n:1 convenience:1 romberg:1 context:2 applying:1 gilbert:1 equivalent:3 deterministic:1 attention:1 independently:2 convex:3 duration:1 amazon:5 identifying:4 simplicity:1 kushilevitz:1 amax:2 spanned:1 initialise:1 handle:1 transmit:2 suppose:1 play:1 pioneered:1 user:27 exact:1 programming:1 us:1 pa:1 element:1 satisfying:1 particularly:1 cut:33 labeled:5 observed:6 role:1 jcss:1 electrical:1 capture:1 worst:2 thousand:1 ensures:1 decrease:1 csi:2 complexity:11 hypergraph:19 depend:2 f2a:1 f2:7 easily:2 darpa:1 represented:2 muthukrishnan:1 separated:2 distinct:1 query:8 artificial:1 hyper:1 choosing:1 whose:1 supplementary:6 valued:2 solve:2 s:1 mkocaoglu:1 compressed:17 otherwise:1 statistic:1 noisy:7 indyk:1 zipcode:2 sequence:1 reconstruction:2 aro:1 product:1 remainder:1 relevant:6 erale:1 translate:1 ymax:4 achieve:1 amplification:1 participating:1 xim:1 requirement:1 optimum:1 produce:1 adam:1 object:1 illustrate:1 received:1 recovering:2 c:10 involves:3 implies:2 implemented:1 correct:2 enable:1 material:6 elimination:1 learngraph:8 require:1 f1:4 underdetermined:1 yij:2 hold:4 considered:3 cb:3 gk2:1 consecutive:2 omitted:1 utexas:4 title:1 clearly:1 gaussian:4 always:3 rather:1 kalai:2 focus:1 rank:2 transmitter:2 greatly:1 industrial:1 attains:1 relation:5 relegated:1 interested:2 tao:2 among:2 colt:1 denoted:2 yahoo:8 spatial:1 breakthrough:1 special:1 art:1 yip:1 field:4 sampling:1 represents:2 constitutes:1 simplex:1 few:5 randomly:8 replaced:1 ourselves:1 consisting:1 connects:2 n1:2 proba:1 interest:2 message:8 highly:1 behind:1 edge:4 tree:2 incomplete:1 plotted:1 theoretical:1 column:8 xeon:1 boolean:13 assignment:1 cost:1 deviation:1 subset:8 monomials:1 entry:1 uniform:5 vertex:2 masked:1 hundred:1 guha:1 n2s:1 perturbed:4 corrupted:1 fundamental:1 definitely:1 siam:2 randomized:1 international:1 negahban:1 donnell:1 receiving:3 decoding:1 pool:1 sketching:12 s2d:1 transmitting:5 again:1 satisfied:2 possibly:1 book:7 leading:1 return:1 li:1 account:1 singleton:2 de:1 sec:1 coefficient:26 int:1 satisfy:3 performed:1 try:1 recover:2 hf:3 contribution:1 formed:1 variance:1 who:2 efficiently:3 correspond:1 yield:2 identify:3 generalize:2 identification:1 notoriously:2 comp:1 explain:1 messenger:5 ed:1 stobbe:1 definition:8 sixth:1 pp:11 involved:1 frequency:1 naturally:1 proof:6 recovers:1 sampled:1 dataset:12 treatment:1 recall:1 improves:1 dimakis1:1 higher:1 day:2 specify:1 just:2 sketch:1 usa:1 contain:1 ccf:1 evolution:2 hence:3 assigned:1 read:3 nonzero:2 deal:1 during:3 uniquely:1 evident:1 complete:1 polytechnique:1 wise:1 novel:2 fi:1 common:1 sopt:4 perturbing:1 discussed:1 hypergraphs:13 m1:5 belong:2 significant:1 measurement:5 refer:1 cambridge:1 counterexample:1 mathematics:1 similarly:1 access:1 yk2:1 multivariate:1 isometry:1 recent:3 belongs:1 certain:1 hyperedge:12 arbitrarily:1 binary:3 captured:1 transmitted:1 seen:1 zip:4 signal:1 arithmetic:1 multiple:1 timestamped:1 infer:1 technical:1 faster:2 academic:1 visit:1 award:1 involving:1 noiseless:4 represent:3 tailored:1 xmax:2 c1:2 background:1 affecting:1 addition:1 krause:1 interval:11 hyperedges:13 sends:2 posse:1 webscope:1 subject:1 sent:1 contrary:1 sttr:1 seem:2 call:2 integer:1 bought:1 near:2 presence:3 enough:1 easy:1 switch:1 zipcodes:3 restrict:1 agnostically:1 reduce:1 idea:1 texas:1 gb:1 remark:1 matlab:1 clear:1 gopalan:1 repeating:1 extensively:1 ph:1 reduced:2 generate:1 specifies:1 outperform:1 xij:6 schapire:1 nsf:1 http:1 notice:1 sign:23 disjoint:1 track:1 correctly:1 sellie:1 vol:7 key:2 four:1 drawn:6 stamped:1 v1:1 ram:1 graph:25 year:1 fooling:1 convert:1 run:10 inverse:1 sum:1 uncertainty:1 prob:1 fourth:2 soda:1 named:1 place:1 almost:3 reader:2 family:3 reasonable:1 decision:2 bound:2 annual:1 karthiksh:1 generates:1 fourier:10 klivans:2 min:3 department:2 combination:2 slightly:1 reconstructing:2 s1:1 restricted:2 taken:2 computationally:1 equation:3 mutually:1 previously:1 eventually:1 hh:1 needed:1 tractable:1 subjected:1 apply:1 observe:2 progression:1 occurrence:1 alternative:1 shah:1 running:2 instant:1 k1:4 establish:1 hypercube:6 society:2 contact:1 added:2 said:5 subspace:1 separate:1 mapped:1 majority:2 participate:1 reason:1 panigrahy:1 code:3 length:2 index:3 lg:3 statement:1 potentially:1 stoc:2 mink:2 rise:1 implementation:1 murat:1 twenty:1 unknown:5 observation:3 acknowledge:1 relational:1 looking:1 communication:5 locate:1 mansour:2 perturbation:1 stack:1 smoothed:6 arbitrary:1 complement:3 namely:1 required:2 specified:1 connection:1 learned:2 established:1 able:1 usually:2 pattern:8 sparsity:4 program:11 max:3 power:4 natural:3 rely:1 improve:1 imply:1 created:1 naive:1 philadelphia:1 prior:1 review:2 kf:1 evolve:2 interesting:1 filtering:1 versus:1 sufficient:1 principle:1 pi:5 austin:2 row:5 parity:25 last:2 keeping:1 monomial:2 formal:1 side:2 neighbor:1 taking:1 absolute:1 sparse:65 fifth:1 ghz:1 dimension:3 calculated:1 world:3 author:2 counted:1 social:2 transaction:3 technicality:1 keep:1 active:2 correlating:1 receiver:3 assumed:1 xi:9 spectrum:1 search:1 continuous:1 why:1 table:3 learn:7 robust:1 improving:1 alg:3 expansion:2 interpolating:1 poly:10 domain:5 main:3 linearly:4 s2:3 noise:5 karthikeyan:1 n2:3 fig:3 intel:1 sub:1 position:3 sparsest:1 exponential:2 comput:2 candidate:3 stamp:3 third:1 learns:7 theorem:9 specific:1 showing:2 sensing:18 learnable:1 list:2 r2:2 intractable:1 exists:2 andoni:1 strauss:1 effectively:1 valiant:1 ci:13 fc:1 simply:1 jacm:1 corresponds:4 satisfies:7 acm:1 succeed:1 marked:1 donoho:1 feasible:5 hard:2 samorodnitsky:1 specifically:1 except:1 uniformly:6 averaging:1 lemma:5 called:4 total:2 teng:2 experimental:3 e:2 succeeds:3 shanmugam1:1 formally:1 support:1 relevance:1 outstanding:1 spielman:1 evaluate:1 |
4,890 | 5,427 | A Residual Bootstrap for High-Dimensional
Regression with Near Low-Rank Designs
Miles E. Lopes
Department of Statistics
University of California, Berkeley
Berkeley, CA 94720
[email protected]
Abstract
We study the residual bootstrap (RB) method in the context of high-dimensional
linear regression. Specifically, we analyze the distributional approximation of linear contrasts c> (?b? ? ?), where ?b? is a ridge-regression estimator. When regression coefficients are estimated via least squares, classical results show that RB
consistently approximates the laws of contrasts, provided that p n, where the
design matrix is of size n ? p. Up to now, relatively little work has considered
how additional structure in the linear model may extend the validity of RB to the
setting where p/n 1. In this setting, we propose a version of RB that resamples
residuals obtained from ridge regression. Our main structural assumption on the
design matrix is that it is nearly low rank ? in the sense that its singular values
decay according to a power-law profile. Under a few extra technical assumptions, we derive a simple criterion for ensuring that RB consistently approximates
the law of a given contrast. We then specialize this result to study confidence
intervals for mean response values Xi> ?, where Xi> is the ith row of the design. More precisely, we show that conditionally on a Gaussian design with near
low-rank structure, RB simultaneously approximates all of the laws Xi> (?b? ? ?),
i = 1, . . . , n. This result is also notable as it imposes no sparsity assumptions on
?. Furthermore, since our consistency results are formulated in terms of the Mallows (Kantorovich) metric, the existence of a limiting distribution is not required.
1
Introduction
Until recently, much of the emphasis in the theory of high-dimensional statistics has been on ?first
order? problems, such as estimation and prediction. As the understanding of these problems has
become more complete, attention has begun to shift increasingly towards ?second order? problems,
dealing with hypothesis tests, confidence intervals, and uncertainty quantification [1?6]. In this
direction, much less is understood about the effects of structure, regularization, and dimensionality
? leaving many questions open. One collection of such questions that has attracted growing interest
deals with the operating characteristics of the bootstrap in high dimensions [7?9] . Due to the fact
that bootstrap is among the most widely used tools for approximating the sampling distributions of
test statistics and estimators, there is much practical value in understanding what factors allow for
the bootstrap to succeed in the high-dimensional regime.
The regression model and linear contrasts. In this paper, we focus our attention on highdimensional linear regression, and our aim is to know when the residual bootstrap (RB) method
consistently approximates the laws of linear contrasts. (A review of RB is given in Section 2.)
1
To specify the model, suppose that we observe a response vector Y ? Rn , generated according to
Y = X? + ?,
n?p
(1)
p
where X ? R
is the observed design matrix, ? ? R is an unknown vector of coefficients, and
the error variables ? = (?1 , . . . , ?n ) are drawn i.i.d. from an unknown distribution F0 , with mean
0 and unknown variance ? 2 < ?. As is conventional in high-dimensional statistics, we assume
the model (1) is embedded in a sequence of models indexed by n. Hence, we allow X, ?, and p to
vary implicitly with n. We will leave p/n unconstrained until Section 3.3, where we will assume
p/n 1 in Theorem 3, and then in Section 3.4, we will assume further that p/n is bounded strictly
between 0 and 1. The distribution F0 is fixed with respect to n, and none of our results require F0
to have more than four moments.
Although we are primarily interested in cases where the design matrix X is deterministic, we will
also study the performance of the bootstrap conditionally on a Gaussian design. For this reason,
we will use the symbol E[. . . |X] even when the design is non-random so that confusion does not
arise in relating different sections of the paper. Likewise, the symbol E[. . . ] refers to unconditional
expectation over all sources of randomness. Whenever the design is random, we will assume X ?
? ?,
denoting the distribution of X by PX , and the distribution of ? by P? .
Within the context of the regression, we will be focused on linear contrasts c> (?b ??), where c ? Rp
is a fixed vector and ?b ? Rp is an estimate of ?. The importance of contrasts arises from the fact
that they unify many questions about a linear model. For instance, testing the significance of the ith
coefficient ?i may be addressed by choosing c to be the standard basis vector c> = e>
i . Another
important problem is quantifying the uncertainty of point predictions, which may be addressed by
choosing c> = Xi> , i.e. the ith row of the design matrix. In this case, an approximation to the law
of the contrast leads to a confidence interval for the mean response value E[Yi ] = Xi> ?. Further
applications of contrasts occur in the broad topic of ANOVA [10].
Intuition for structure and regularization in RB. The following two paragraphs explain the core
conceptual aspects of the paper. To understand the role of regularization in applying RB to highdimensional regression, it is helpful to think of RB in terms of two ideas. First, if ?bLS denotes the
ordinary least squares estimator, then it is a simple but important fact that contrasts can be written
as c> (?bLS ? ?) = a> ? where a>:= c> (X > X)?1 X > . Hence, if it were possible to sample directly
from F0 , then the law of any such contrast could be easily determined. Since F0 is unknown, the
second key idea is to use the residuals of some estimator ?b as a proxy for samples from F0 . When
p n, the least-squares residuals are a good proxy [11, 12]. However, it is well-known that leastsquares tends to overfit when p/n 1. When ?bLS fits ?too well?, this means that its residuals are
b
?too small?, and hence they give a poor proxy for F0 . Therefore, by using a regularized estimator ?,
b
overfitting can be avoided, and the residuals of ? may offer a better way of obtaining ?approximate
samples? from F0 .
The form of regularized regression we will focus on is ridge regression:
?b? := (X > X + ?Ip?p )?1 X > Y,
(2)
where ? > 0 is a user-specificed regularization parameter. As will be seen in Sections 3.2 and 3.3,
the residuals obtained from ridge regression lead to a particularly good approximation of F0 when
the design matrix X is nearly low-rank, in the sense that most of its singular values are close to
0. In essence, this condition is a form of sparsity, since it implies that the rows of X nearly lie
in a low-dimensional subspace of Rp . However, this type of structural condition has a significant
advantage over the the more well-studied assumption that ? is sparse. Namely, the assumption that
X is nearly low-rank can be inspected directly in practice ? whereas sparsity in ? is typically
unverifiable. In fact, our results will impose no conditions on ?, other than that k?k2 remains
bounded as (n, p) ? ?. Finally, it is worth noting that the occurrence of near low-rank design
matrices is actually very common in applications, and is often referred to as collinearity [13, ch.
17].
Contributions and outline. The primary contribution of this paper is a complement to the work of
Bickel and Freedman [12] (hereafter B&F 1983) ? who showed that in general, the RB method fails
2
to approximate the laws of least-squares contrasts c> (?bLS ? ?) when p/n 1. Instead, we develop
an alternative set of results, proving that even when p/n 1, RB can successfully approximate the
laws of ?ridged contrasts? c> (?b? ? ?) for many choices of c ? Rp , provided that the design matrix
X is nearly low rank. A particularly interesting consequence of our work is that RB successfully
approximates the law c> (?b? ? ?) for a certain choice of c that was shown in B&F 1983 to ?break?
RB when applied to least-squares. Specifically, such a c can be chosen as one of the rows of X with
a high leverage score (see Section 4). This example corresponds to the practical problem of setting
confidence intervals for mean response values E[Yi ] = Xi> ?. (See [12, p. 41], as well as Lemma 2
and Theorem 4 in Section 3.4). Lastly, from a technical point of view, a third notable aspect of our
results is that they are formulated in terms of the Mallows-`2 metric, which frees us from having to
impose conditions that force a limiting distribution to exist.
Apart from B&F 1983, the most closely related works we are aware of are the recent papers [7]
and [8], which also consider RB in the high-dimensional setting. However, these works focus on
role of sparsity in ? and do not make use of low-rank structure in the design, whereas our work deals
only with structure in the design and imposes no sparsity assumptions on ?.
The remainder of the paper is organized as follows. In Section 2, we formulate the problem of
approximating the laws of contrasts, and describe our proposed methodology for RB based on ridge
regression. Then, in Section 3 we state several results that lay the groundwork for Theorem 4, which
shows that that RB can successfully approximate all of the laws L(Xi> (?b? ? ?)|X), i = 1, . . . , n,
conditionally on a Gaussian design. Due to space constraints, all proofs are deferred to material that
will appear in a separate work.
Notation and conventions. If U and V are random variables, then L(U |V ) denotes the law of U ,
conditionally on V . If an and bn are two sequences of real numbers, then the notation an . bn
means that there is an absolute constant ?0 > 0 and an integer n0 ? 1 such that an ? ?0 bn for all
n ? n0 . The notation an bn means that an . bn and bn . an . For a square matrix A ? Rk?k
whose eigenvalues are real, we denote them by ?min (A) = ?k (A) ? ? ? ? ? ?1 (A) = ?max (A).
2
Problem setup and methodology
Problem setup. For any c ? Rp , it is clear that conditionally on X, the law of c> (?b? ? ?) is
completely determined by F0 , and hence it makes sense to use the notation
?
?
(3)
?? (F0 ; c) := L c> (?b? ? ?)?
?X .
The problem we aim to solve is to approximate the distribution ?? (F0 ; c) for suitable choices of c.
Review of the residual bootstrap (RB) procedure. We briefly explain the steps involved in the
residual bootstrap procedure, applied to the ridge estimator ?b? of ?. To proceed somewhat indirectly,
consider the following ?bias-variance? decomposition of ?? (F0 ; c), conditionally on X,
?
?
?? (F0 ; c) = L c> ?b? ? E[?b? |X] ?
(4)
?X + c> E[?b? |X] ? ? .
|
{z
}
|
{z
}
=: ?? (F0 ;c)
=: bias(?? (F0 ;c))
Note that the distribution ?(F0 ; c) has mean zero, and so that the second term on the right side is
the bias of ?? (F0 ; c) as an estimator of ?? (F0 ; c). Furthermore, the distribution ?? (F0 ; c) may be
viewed as the ?variance component? of ?? (F0 ; c). We will be interested in situations where the
regularization parameter ? may be chosen small enough so that the bias component is small. In this
case, one has ?? (F0 ; c) ? ?? (F0 ; c), and then it is enough to find an approximation to the law
?? (F0 ; c), which is unknown. To this end, a simple manipulation of c> (?b? ? E[?b? ]) leads to
?
?
?? (F0 ; c) = L(c> (X > X + ?Ip?p )?1 X > ??
(5)
?X).
Now, to approximate ?? (F0 ; c), let Fb be any centered estimate of F0 . (Typically, Fb is obtained by
using the centered residuals of some estimator of ?, but this is not necessary in general.) Also, let
?? = (??1 , . . . , ??n ) ? Rn be an i.i.d. sample from Fb. Then, replacing ? with ?? in line (5) yields
?
?
?? (Fb; c) = L(c> (X > X + ?Ip?p )?1 X > ?? ?
(6)
?X).
3
At this point, we define the (random) measure ?? (Fb; c) to be the RB approximation to ?? (F0 ; c).
Hence, it is clear that the RB approximation is simply a ?plug-in rule?.
A two-stage approach. An important feature of the procedure just described is that we are free
to use any centered estimator Fb of F0 . This fact offers substantial flexibility in approximating
?? (F0 ; c). One way of exploiting this flexibility is to consider a two-stage approach, where a ?pilot?
ridge estimator ?b% is used to first compute residuals whose centered empirical distribution function
is Fb% , say. Then, in the second stage, the distribution Fb% is used to approximate ?? (F0 ; c) via the
relation (6). To be more detailed, if (b
e1 (%), . . . , ebn (%)) = eb(%) := Y ? X ?b% are the residuals of ?b% ,
b
then we define F% to be the distribution that places mass 1/n at each of the values ebi (%) ? e?(%) with
Pn
e?(%) := n1 i=1 ebi (%). Here, it is important to note that the value % is chosen to optimize Fb% as an
approximation to F0 . By contrast, the choice of ? depends on the relative importance of width and
coverage probability for confidence intervals based on ?? (Fb% ; c). Theorems 1, 3, and 4 will offer
some guidance in selecting % and ?.
Resampling algorithm. To summarize the discussion above, if B is user-specified number of
bootstrap replicates, our proposed method for approximating ?? (F0 ; c) is given below.
1. Select ? and %, and compute the residuals eb(%) = Y ? X ?b% .
2. Compute the centered distribution function Fb% , putting mass 1/n at each ebi (%) ? e?(%).
3. For j = 1, . . . , B:
? Draw a vector ?? ? Rn of n i.i.d. samples from Fb% .
? Compute zj := c> (X > X + ?Ip?p )?1 X > ?? .
4. Return the empirical distribution of z1 , . . . , zB .
Clearly, as B ? ?, the empirical distribution of z1 , . . . , zB converges weakly to ?? (Fb% ; c), with
probability 1. As is conventional, our theoretical analysis in the next section will ignore Monte Carlo
issues, and address only the performance of ?? (Fb% ; c) as an approximation to ?? (F0 ; c).
3
Main results
The following metric will be central to our theoretical results, and has been a standard tool in the
analysis of the bootstrap, beginning with the work of Bickel and Freedman [14].
The Mallows (Kantorovich) metric. For two random vectors U and V in a Euclidean space, the
Mallows-`2 metric is defined by
n h
i
o
d22 (L(U ), L(V )) := inf E kU ? V k22 : (U, V ) ? ?
(7)
???
where the infimum is over the class ? of joint distributions ? whose marginals are L(U ) and L(V ).
It is worth noting that convergence in d2 is strictly stronger than weak convergence, since it also
requires convergence of second moments. Additional details may be found in the paper [14].
3.1
A bias-variance decomposition for bootstrap approximation
To give some notation for analyzing the bias-variance decomposition of ?? (F0 ; c) in line (4), we
define the following quantities based upon the ridge estimator ?b? . Namely, the variance is
v? = v? (X; c) := var(?? (F0 ; c)|X) = ? 2 kc> (X > X + ?Ip?p )?1 X > k22 .
To express the bias of ?? (F0 ; c), we define the vector ?(X) ? Rp according to
?(X) := ? ? E[?b? ] = Ip?p ? (X > X + ?Ip?p )?1 X > X ?,
4
(8)
and then put
b2? = b2? (X; c) := bias2 (?? (F0 ; c)) = (c> ?(X))2 .
(9)
We will sometimes omit the arguments of v? and b2? to lighten notation. Note that v? (X; c) does not
depend on ?, and b2? (X; c) only depends on ? through ?(X).
The following result gives a regularized and high-dimensional extension of some lemmas in Freedman?s early work [11] on RB for least squares. The result does not require any structural conditions
on the design matrix, or on the true parameter ?.
Theorem 1 (consistency criterion). Suppose X ? Rn?p is fixed. Let Fb be any estimator of F0 , and
let c ? Rp be any vector such that v? = v? (X; c) 6= 0. Then with P? -probability 1, the following
inequality holds for every n ? 1, and every ? > 0,
b2
d22 ?1v? ?? (F0 ; c), ?1v? ?? (Fb; c) ? ?12 d22 (F0 , Fb) + v?? .
(10)
?
Remarks. Observe that the normalization 1/ v? ensures that the bound is non-trivial, since the
?
distribution ?? (F0 ; c)/ v? has variance equal to 1 for all n (and hence does not become degenerate
for large n). To consider the choice of ?, it is simple to verify that the ratio b2? /v? decreases monotonically as ? decreases. Note also that as ? becomes small, the variance v? becomes large, and
likewise, confidence intervals based on ?? (Fb; c) become wider. In other words, there is a trade-off
between the width of the confidence interval and the size of the bound (10).
Sufficient conditions for consistency of RB. An important practical aspect of Theorem 1 is that
for any given contrast c, the variance v? (X; c) can be easily estimated, since it only requires an
estimate of ? 2 , which can be obtained from Fb. Consequently, whenever theoretical bounds on
d22 (F0 , Fb) and b2? (X; c) are available, the right side of line (10) can be controlled. In this way,
Theorem 1 offers a simple route for guaranteeing that RB is consistent. In Sections 3.2 and 3.3 to
follow, we derive a bound on E[d22 (F0 , Fb)|X] in the case where Fb is chosen to be Fb% . Later on in
Section 3.4, we study RB consistency in the context of prediction with a Gaussian design, and there
we derive high probability bounds on both v? (X; c) and b2? (X; c) where c is a particular row of X.
3.2
A link between bootstrap consistency and MSPE
If ?b is an estimator of ?, its mean-squared prediction error (MSPE), conditionally on X, is defined
as
?
?
mspe(?b |X) := n1 E kX(?b ? ?)k22 ?
(11)
?X .
The previous subsection showed that in-law approximation of contrasts is closely tied to the approximation of F0 . We now take a second step of showing that if the centered residuals of an estimator
?b are used to approximate F0 , then the quality of this approximation can be bounded naturally in
terms of mspe(?b |X). This result applies to any estimator ?b computed from the observations (1).
Theorem 2. Suppose X ? Rn?p is fixed. Let ?b be any estimator of ?, and let Fb be the empirical
b Also, let Fn denote the empirical distribution of n i.i.d.
distribution of the centered residuals of ?.
samples from F0 . Then for every n ? 1,
?
2
?
?
E d22 (Fb, F0 )?X
? 2 mspe(?b |X) + 2 E[d22 (Fn , F0 )] + 2?n .
(12)
Remarks. As we will see in the next section, the MSPE of ridge regression can be bounded in a
sharp way when the design matrix is approximately low rank, and there we will analyze mspe(?b% |X)
for the pilot estimator. Consequently, when near low-rank structure is available, the only remaining
issue in controlling the right side of line (12) is to bound the quantity E[d22 (Fn , F0 )|X]. The very
recent work of Bobkov and Ledoux [15] provides an in-depth study of this question, and they derive
a variety bounds under different tail conditions on F0 . We summarize one of their results below.
Lemma 1 (Bobkov and Ledoux, 2014). If F0 has a finite fourth moment, then
E[d22 (Fn , F0 )] . log(n)n?1/2 .
5
(13)
Remarks. The fact that the squared distance is bounded at the rate of log(n)n?1/2 is an indication that d2 is a rather strong metric on distributions. For a detailed discussion of this result, see
Corollaries 7.17 and 7.18 in the paper [15]. Although it is possible to obtain faster rates when more
b
stringent tail conditions are placed on F0 , we will only need a fourth moment, since the mspe(?|X)
term in Theorem 2 will often have a slower rate than log(n)n?1/2 , as discussed in the next section.
3.3
Consistency of ridge regression in MSPE for near low rank designs
In this subsection, we show that when the tuning parameter % is set at a suitable rate, the pilot ridge
estimator ?b% is consistent in MSPE when the design matrix is near low-rank ? even when p/n is
large, and without any sparsity constraints on ?. We now state some assumptions.
A1. There is a number ? > 0, and absolute constants ?1 , ?2 > 0, such that
b ? ?2 i??
?1 i?? ? ?i (?)
for all i = 1, . . . , n ? p.
A2. There are absolute constants ?, ? > 0, such that for every n ? 1,
%
n
= n?? and
?
n
= n?? .
A3. The vector ? ? Rp satisfies k?k2 . 1.
Due to Theorem 2, the following bound shows that the residuals of ?b% may be used to extract a
consistent approximation to F0 . Two other notable features of the bound are that it is non-asymptotic
and dimension-free.
Theorem 3. Suppose that X ? Rn?p is fixed and that Assumptions 1?3 hold, with p/n 1. Assume
1
?
1
further that ? is chosen as ? = 2?
3 when ? ? (0, 2 ), and ? = ?+1 when ? > 2 . Then,
(
2?
if ? ? (0, 12 ),
n? 3
?
(14)
mspe(?b% |X) .
? ?+1
n
if ? > 21 .
Also, both bounds in (14) are tight in the sense that ? can be chosen so that ?b% attains either rate.
b are observable, they may be used to estimate ? and guide
Remarks. Since the eigenvalues ?i (?)
the selection of %/n = n?? . However, from a practical point of view, we found it easier to select %
via cross-validation in numerical experiments, rather than via an estimate of ?.
A link with Pinsker?s Theorem. In the particular case when F0 is a centered Gaussian distribution, the ?prediction problem? of estimating X? is very similar to estimating the mean parameters of
a Gaussian sequence model, with error measured in the `2 norm. In the alternative sequence-model
format, the decay condition on the eigenvalues of n1 X > X translates into an ellipsoid constraint on
the mean parameter sequence [16, 17]. For this reason, Theorem 3 may be viewed as ?regression
version? of `2 error bounds for the sequence model under an ellipsoid constraint (cf. Pinsker?s Theorem, [16, 17]). Due to the fact that the latter problem has a very well developed literature, there
may be various ?neighboring results? elsewhere. Nevertheless, we could not find a direct reference
for our stated MSPE bound in the current setup. For the purposes of our work in this paper, the more
important point to take away from Theorem 3 is that it can be coupled with Theorem 2 for proving
consistency of RB.
3.4
Confidence intervals for mean responses, conditionally on a Gaussian design
In this section, we consider the situation where the design matrix X has rows Xi> ? Rp drawn
i.i.d. from a multivariate normal distribution N (0, ?), with X ?
? ?. (The covariance matrix ? may
vary with n.) Conditionally on a realization of X, we analyze the RB approximation of the laws
?? (F0 ; Xi ) = L(Xi> (?b? ? ?)|X). As discussed in Section 1, this corresponds to the problem of
setting confidence intervals for the mean responses E[Yi ] = Xi> ?. Assuming that the population
eigenvalues ?i (?) obey a decay condition, we show below in Theorem 4 that RB succeeds with high
PX -probability. Moreover, this consistency statement holds for all of the laws ?? (F0 ; Xi ) simultaneously. That is, among the n distinct laws ?? (F0 ; Xi ), i = 1, . . . , n, even the worst bootstrap
approximation is still consistent. We now state some population-level assumptions.
6
A4. The operator norm of ? ? Rp?p satisfies k?kop . 1.
Next, we impose a decay condition on the eigenvalues of ?. This condition also ensures that ? is
invertible for each fixed p ? even though the bottom eigenvalue may become arbitrarily small as p
becomes large. It is important to notice that we now use ? for the decay exponent of the population
eigenvalues, whereas we used ? when describing the sample eigenvalues in the previous section.
A5. There is a number ? > 0, and absolute constants k1 , k2 > 0, such that for all i = 1, . . . , p,
k1 i?? ? ?i (?) ? k2 i?? .
A6. There are absolute constants k3 , k4 ? (0, 1) such that for all n ? 3, we have the bounds
k3 ? np ? k4 and p ? n ? 2.
The following lemma collects most of the effort needed in proving our final result in Theorem 4.
Here it is also helpful to recall the notation ?/n = n?? and %/n = n?? from Assumption 2.
Lemma 2. Suppose that the matrix X ? Rn?p has rows Xi> drawn i.i.d. from N (0, ?), and that
Assumptions 2?6 hold. Furthermore, assume that ? chosen so that 0 < ? < min{?, 1}. Then, the
statements below are true.
(i) (bias inequality)
Fix any ? > 0. Then, there is an absolute constant ?0 > 0, such that for all large n, the following
event holds with PX -probability at least 1 ? n?? ? ne?n/16 ,
max b2? (X; Xi ) ? ?0 ? n?? ? (? + 1) log(n + 2).
1?i?n
(15)
(ii) (variance inequality)
There are absolute constants ?1 , ?2 > 0 such that for all large n, the following event holds with
?
PX -probability at least 1 ? 4n exp(??1 n ? ),
1
1?i?n v? (X;Xi )
max
?
? ?2 n1? ? .
(16)
(iii) (mspe inequalities)
?
when
Suppose that ? is chosen as ? = 2?/3 when ? ? (0, 21 ), and that ? is chosen as ? = 1+?
1
? > 2 . Then, there are absolute constants ?3 , ?4 , ?5 , ?6 > 0 such that for all large n,
(
2?
with PX -probability at least 1 ? exp(??3 n2?4?/3 ), if ? ? (0, 12 )
?4 n? 3
b
mspe(?% |X) ?
?
2
? ?+1
?6 n
with PX -probability at least 1 ? exp(??5 n 1+? ),
if ? > 21 .
Remarks. Note that the two rates in part (iii) coincide as ? approaches 1/2. At a conceptual level,
the entire lemma may be explained in relatively simple terms. Viewing the quantities mspe(?b% |X),
b2? (X; Xi ) and v? (X; Xi ) as functionals of a Gaussian matrix, the proof involves deriving concentration bounds for each of them. Indeed, this is plausible given that these quantities are smooth
functionals of X. However, the difficulty of the proof arises from the fact that they are also highly
non-linear functionals of X. We now combine Lemmas 1 and 2 with Theorems 1 and 2 to show that
all of the laws ?? (F0 ; Xi ) can be simultaneously approximated via our two-stage RB method.
Theorem 4. Suppose that F0 has a finite fourth moment, Assumptions 2?6 hold, and ? is chosen
?
so that 1+?
< ? < min{?, 1}. Also suppose that ? is chosen as ? = 2?/3 when ? ? (0, 12 ), and
?
? = ?+1
when ? > 12 . Then, there is a sequence of positive numbers ?n with limn?? ?n = 0, such
that the event
?
h
i
?
?
E max d22 ?1v? ?? (F0 ; Xi ), ?1v? ?? (Fb% ; Xi ) ?
X
? ?n
(17)
?
?
1?i?n
has PX -probability tending to 1 as n ? ?.
Remark. Lemma 2 gives explicit bounds on the numbers ?n , as well as the probabilities of the
corresponding events, but we have stated the result in this way for the sake of readability.
7
4
Simulations
In four different settings of n, p, and the decay parameter ?, we compared the nominal 90% confidence intervals (CIs) of four methods: ?oracle?, ?ridge?, ?normal?, and ?OLS?, to be described
below. In each setting, we generated N1 := 100 random designs X with i.i.d. rows drawn from
N (0, ?), where ?j (?) = j ?? , j = 1, . . . , p, and the eigenvectors of ? were drawn randomly by
setting them to be the Q factor in a QR decomposition of a standard p ? p Gaussian matrix. Then,
for each realization of X, we generated N2 := 1000 realizations of Y according to the model (1),
where ? = 1/k1k2 ? Rp , and F0 is the centered t distribution on 5 degrees of freedom, rescaled to
have standard deviation ? = 0.1. For each X, and each corresponding Y , we considered the problem of setting a 90% CI for the mean response value Xi>? ?, where Xi>? is the row with the highest
leverage score, i.e. i? = argmax1?i?n Hii and H := X(X > X)?1 X > . This problem was shown in
B&F 1983 to be a case where the standard RB method based on least-squares fails when p/n 1.
Below, we refer to this method as ?OLS?.
To describe the other three methods, ?ridge? refers to the interval [Xi>? ?b? ? qb0.95 , Xi>? ?b? ? qb0.05 ],
where qb? is the ?% quantile of the numbers z1 , . . . , zB computed in the proposed algorithm in
Section 2, with B = 1000 and c> = Xi>? . To choose the parameters ? and % for a given X and Y ,
we first computed rb as the value that optimized the MSPE error of a ridge estimator ?br with respect
to 5-fold cross validation; i.e. cross validation was performed for every distinct pair (X, Y ). We then
put % = 5b
r and ? = 0.1b
r, as we found the prefactors 5 and 0.1 to work adequately across various
settings. (Optimizing % with respect to MSPE is motivated by Theorems 1, 2, and 3. Also, choosing ?
to be somewhat smaller than % conforms with the constraints on ? and ? in Theorem 4.) The method
?normal? refers to the CI based on the normal approximation L(Xi>? (?b? ??)|X) ? N (0, ?b2 ), where
?b2 = ?
b2 kXi>? (X > X +?Ip?p )?1 X > k22 , ? = 0.1b
r, and ?
b2 is the usual unbiased estimate of ? 2 based
on OLS residuals. The ?oracle? method refers to the interval [Xi>? ?b? ? q?0.95 , Xi>? ?b? ? q?0.05 ], with
? = 0.1b
r, and q?? being the empirical ?% quantile of Xi> (?b? ? ?) over all 1000 realizations of Y
based on a given X. (This accounts for the randomness in ? = 0.1b
r.)
Within a given setting of the triplet (n, p, ?), we refer to the ?coverage? of a method as the fraction of
the N1 ?N2 = 105 instances where the method?s CI contained the parameter Xi>? ?. Also, we refer to
?width? as the average width of a method?s intervals over all of the 105 instances. The four settings of
(n, p, ?) correspond to moderate/high dimension and moderate/fast decay of the eigenvalues ?i (?).
Even in the moderate case of p/n = 0.45, the results show that the OLS intervals are too narrow
and have coverage noticeably less than 90%. As expected, this effect becomes more pronounced
when p/n = 0.95. The ridge and normal intervals perform reasonably well across settings, with
both performing much better than OLS. However, it should be emphasized that our study of RB
is motivated by the desire to gain insight into the behavior of the bootstrap in high dimensions
? rather than trying to outperform particular methods. In future work, we plan to investigate the
relative merits of the ridge and normal intervals in greater detail.
Table 1: Comparison of nominal 90% confidence intervals
setting 1
n = 100, p = 45, ? = 0.5
setting 2
n = 100, p = 95, ? = 0.5
setting 3
n = 100, p = 45, ? = 1
setting 4
n = 100, p = 95, ? = 1
width
coverage
width
coverage
width
coverage
width
coverage
oracle
0.21
0.90
0.22
0.90
0.20
0.90
0.21
0.90
ridge
0.20
0.87
0.26
0.88
0.21
0.90
0.26
0.92
normal
0.23
0.91
0.26
0.88
0.22
0.91
0.23
0.87
OLS
0.16
0.81
0.06
0.42
0.16
0.81
0.06
0.42
Acknowledgements. MEL thanks Prof. Peter J. Bickel for many helpful discussions, and gratefully acknowledges the DOE CSGF under grant DE-FG02-97ER25308, as well as the NSF-GRFP.
8
References
[1] C.-H. Zhang and S. S. Zhang. Confidence intervals for low dimensional parameters in high
dimensional linear models. Journal of the Royal Statistical Society: Series B, 76(1):217?242,
2014.
[2] A. Javanmard and A. Montanari. Hypothesis testing in high-dimensional regression under the
Gaussian random design model: Asymptotic theory. arXiv preprint arXiv:1301.4240, 2013.
[3] A. Javanmard and A. Montanari. Confidence intervals and hypothesis testing for highdimensional regression. arXiv preprint arXiv:1306.3171, 2013.
[4] P. B?uhlmann. Statistical significance in high-dimensional linear models. Bernoulli,
19(4):1212?1242, 2013.
[5] S. van de Geer, P. B?uhlmann, and Y. Ritov. On asymptotically optimal confidence regions and
tests for high-dimensional models. arXiv preprint arXiv:1303.0518, 2013.
[6] J. D. Lee, D. L. Sun, Y. Sun, and J. E. Taylor. Exact inference after model selection via the
lasso. arXiv preprint arXiv:1311.6238, 2013.
[7] A. Chatterjee and S. N. Lahiri. Rates of convergence of the adaptive lasso estimators to the
oracle distribution and higher order refinements by the bootstrap. The Annals of Statistics,
41(3):1232?1259, 2013.
[8] H. Liu and B. Yu. Asymptotic properties of lasso+mls and lasso+ridge in sparse highdimensional linear regression. Electronic Journal of Statistics, 7:3124?3169, 2013.
[9] V. Chernozhukov, D. Chetverikov, and K. Kato. Gaussian approximations and multiplier bootstrap for maxima of sums of high-dimensional random vectors. The Annals of Statistics,
41(6):2786?2819, 2013.
[10] E. L. Lehmann and J. P. Romano. Testing statistical hypotheses. Springer, 2005.
[11] D. A. Freedman. Bootstrapping regression models. The Annals of Statistics, 9(6):1218?1228,
1981.
[12] P. J. Bickel and D. A. Freedman. Bootstrapping regression models with many parameters. In
Festschrift for Erich L. Lehmann, pages 28?48. Wadsworth, 1983.
[13] N. R. Draper and H. Smith. Applied regression analysis. Wiley-Interscience, 1998.
[14] P. J. Bickel and D. A. Freedman. Some asymptotic theory for the bootstrap. The Annals of
Statistics, pages 1196?1217, 1981.
[15] S. Bobkov and M. Ledoux. One-dimensional empirical measures, order statistics, and Kantorovich transport distances. preprint, 2014.
[16] A. B. Tsybakov. Introduction to nonparametric estimation. Springer, 2009.
[17] L. Wasserman. All of nonparametric statistics. Springer, 2006.
9
| 5427 |@word collinearity:1 briefly:1 version:2 stronger:1 norm:2 open:1 d2:2 simulation:1 bn:6 decomposition:4 covariance:1 moment:5 liu:1 series:1 score:2 hereafter:1 selecting:1 denoting:1 groundwork:1 current:1 attracted:1 written:1 fn:4 numerical:1 n0:2 resampling:1 beginning:1 ith:3 smith:1 core:1 grfp:1 provides:1 argmax1:1 readability:1 zhang:2 direct:1 become:4 specialize:1 combine:1 interscience:1 paragraph:1 javanmard:2 expected:1 indeed:1 behavior:1 growing:1 little:1 becomes:4 provided:2 estimating:2 bounded:5 notation:7 moreover:1 mass:2 what:1 developed:1 bootstrapping:2 berkeley:3 every:5 k2:4 grant:1 omit:1 appear:1 positive:1 understood:1 tends:1 consequence:1 analyzing:1 approximately:1 emphasis:1 eb:2 studied:1 collect:1 practical:4 testing:4 mallow:4 practice:1 bootstrap:18 procedure:3 empirical:7 confidence:14 word:1 refers:4 er25308:1 close:1 selection:2 operator:1 put:2 context:3 applying:1 optimize:1 conventional:2 deterministic:1 attention:2 focused:1 formulate:1 unify:1 d22:10 wasserman:1 estimator:20 rule:1 insight:1 deriving:1 proving:3 population:3 limiting:2 annals:4 controlling:1 suppose:8 inspected:1 user:2 nominal:2 exact:1 hypothesis:4 approximated:1 particularly:2 lay:1 distributional:1 observed:1 role:2 bottom:1 preprint:5 worst:1 region:1 ensures:2 sun:2 decrease:2 trade:1 rescaled:1 highest:1 substantial:1 intuition:1 pinsker:2 weakly:1 depend:1 tight:1 upon:1 basis:1 completely:1 easily:2 joint:1 various:2 distinct:2 fast:1 describe:2 monte:1 choosing:3 whose:3 widely:1 solve:1 plausible:1 say:1 statistic:11 think:1 ip:8 final:1 sequence:7 advantage:1 eigenvalue:9 ledoux:3 indication:1 propose:1 remainder:1 neighboring:1 kato:1 realization:4 flexibility:2 degenerate:1 pronounced:1 ebi:3 qr:1 exploiting:1 convergence:4 guaranteeing:1 leave:1 converges:1 wider:1 derive:4 develop:1 stat:1 measured:1 strong:1 coverage:7 involves:1 implies:1 convention:1 direction:1 closely:2 centered:9 stringent:1 viewing:1 material:1 noticeably:1 require:2 fix:1 leastsquares:1 strictly:2 extension:1 hold:7 considered:2 normal:7 exp:3 k3:2 vary:2 bickel:5 early:1 a2:1 purpose:1 estimation:2 chernozhukov:1 uhlmann:2 successfully:3 tool:2 clearly:1 gaussian:11 aim:2 rather:3 pn:1 corollary:1 focus:3 consistently:3 rank:12 bernoulli:1 contrast:17 attains:1 sense:4 helpful:3 inference:1 typically:2 entire:1 relation:1 kc:1 interested:2 issue:2 among:2 exponent:1 plan:1 bobkov:3 wadsworth:1 equal:1 aware:1 having:1 sampling:1 broad:1 yu:1 nearly:5 future:1 np:1 lighten:1 few:1 primarily:1 randomly:1 simultaneously:3 festschrift:1 n1:6 freedom:1 interest:1 a5:1 highly:1 investigate:1 replicates:1 deferred:1 unconditional:1 necessary:1 conforms:1 indexed:1 euclidean:1 taylor:1 guidance:1 theoretical:3 instance:3 a6:1 ordinary:1 deviation:1 too:3 kxi:1 thanks:1 lee:1 off:1 invertible:1 squared:2 central:1 choose:1 return:1 account:1 de:2 b2:14 coefficient:3 notable:3 depends:2 later:1 break:1 view:2 performed:1 analyze:3 contribution:2 square:8 variance:10 characteristic:1 likewise:2 who:1 yield:1 correspond:1 weak:1 none:1 carlo:1 worth:2 randomness:2 explain:2 whenever:2 involved:1 naturally:1 proof:3 gain:1 pilot:3 begun:1 recall:1 subsection:2 dimensionality:1 organized:1 actually:1 higher:1 ridged:1 follow:1 methodology:2 response:7 specify:1 ritov:1 though:1 furthermore:3 just:1 stage:4 lastly:1 until:2 overfit:1 replacing:1 lahiri:1 transport:1 infimum:1 quality:1 effect:2 validity:1 k22:4 true:2 verify:1 unbiased:1 adequately:1 regularization:5 hence:6 multiplier:1 mile:1 deal:2 conditionally:9 width:8 essence:1 mel:1 criterion:2 trying:1 outline:1 ridge:18 complete:1 confusion:1 resamples:1 recently:1 common:1 ols:6 tending:1 extend:1 tail:2 approximates:5 relating:1 marginals:1 discussed:2 significant:1 refer:3 tuning:1 unconstrained:1 consistency:8 erich:1 gratefully:1 f0:63 operating:1 prefactors:1 multivariate:1 showed:2 recent:2 optimizing:1 inf:1 apart:1 moderate:3 manipulation:1 route:1 certain:1 mspe:17 inequality:4 arbitrarily:1 yi:3 seen:1 additional:2 somewhat:2 impose:3 greater:1 monotonically:1 ii:1 smooth:1 technical:2 faster:1 plug:1 offer:4 cross:3 e1:1 a1:1 controlled:1 ensuring:1 prediction:5 regression:22 metric:6 expectation:1 arxiv:8 sometimes:1 normalization:1 whereas:3 interval:19 addressed:2 singular:2 leaving:1 source:1 limn:1 extra:1 integer:1 structural:3 near:6 noting:2 leverage:2 iii:2 enough:2 variety:1 fit:1 lasso:4 idea:2 br:1 translates:1 shift:1 motivated:2 effort:1 peter:1 proceed:1 romano:1 remark:6 clear:2 detailed:2 eigenvectors:1 nonparametric:2 tsybakov:1 outperform:1 exist:1 zj:1 nsf:1 notice:1 estimated:2 rb:32 bls:4 express:1 key:1 four:4 putting:1 nevertheless:1 drawn:5 k4:2 anova:1 draper:1 asymptotically:1 fraction:1 sum:1 uncertainty:2 bias2:1 fourth:3 lope:1 lehmann:2 place:1 electronic:1 draw:1 bound:15 fold:1 oracle:4 occur:1 precisely:1 constraint:5 unverifiable:1 sake:1 aspect:3 argument:1 min:3 qb:1 performing:1 relatively:2 px:7 format:1 department:1 according:4 poor:1 across:2 smaller:1 increasingly:1 explained:1 remains:1 describing:1 needed:1 know:1 merit:1 end:1 available:2 observe:2 obey:1 away:1 indirectly:1 hii:1 occurrence:1 alternative:2 slower:1 rp:11 existence:1 denotes:2 remaining:1 cf:1 a4:1 k1:2 quantile:2 prof:1 approximating:4 classical:1 society:1 question:4 quantity:4 primary:1 concentration:1 usual:1 kantorovich:3 subspace:1 distance:2 separate:1 link:2 topic:1 trivial:1 reason:2 assuming:1 ellipsoid:2 ratio:1 setup:3 statement:2 stated:2 design:26 unknown:5 perform:1 observation:1 finite:2 situation:2 rn:7 sharp:1 complement:1 namely:2 required:1 specified:1 pair:1 z1:3 optimized:1 california:1 narrow:1 address:1 below:6 regime:1 sparsity:6 summarize:2 max:4 royal:1 power:1 suitable:2 event:4 difficulty:1 quantification:1 regularized:3 force:1 residual:19 ne:1 acknowledges:1 extract:1 coupled:1 review:2 understanding:2 literature:1 acknowledgement:1 relative:2 law:20 embedded:1 asymptotic:4 interesting:1 var:1 validation:3 degree:1 sufficient:1 proxy:3 imposes:2 consistent:4 row:9 elsewhere:1 placed:1 free:3 bias:8 allow:2 understand:1 side:3 guide:1 absolute:8 sparse:2 van:1 dimension:4 depth:1 fb:26 collection:1 adaptive:1 coincide:1 avoided:1 refinement:1 functionals:3 approximate:8 observable:1 ignore:1 implicitly:1 dealing:1 ml:1 overfitting:1 conceptual:2 xi:31 triplet:1 table:1 ku:1 reasonably:1 ca:1 obtaining:1 significance:2 main:2 montanari:2 arise:1 profile:1 freedman:6 n2:3 referred:1 wiley:1 fails:2 chetverikov:1 explicit:1 lie:1 tied:1 third:1 theorem:22 rk:1 kop:1 emphasized:1 showing:1 symbol:2 decay:7 a3:1 importance:2 ci:4 fg02:1 chatterjee:1 kx:1 easier:1 simply:1 desire:1 contained:1 applies:1 springer:3 ch:1 corresponds:2 satisfies:2 succeed:1 viewed:2 formulated:2 quantifying:1 consequently:2 towards:1 specifically:2 determined:2 lemma:8 zb:3 geer:1 succeeds:1 select:2 highdimensional:4 latter:1 arises:2 |
4,891 | 5,428 | Fast and Robust Least Squares Estimation in
Corrupted Linear Models
Brian McWilliams?
Gabriel Krummenacher? Mario Lucic Joachim M. Buhmann
Department of Computer Science
ETH Z?urich, Switzerland
{mcbrian,gabriel.krummenacher,lucic,jbuhmann}@inf.ethz.ch
Abstract
Subsampling methods have been recently proposed to speed up least squares estimation in large scale settings. However, these algorithms are typically not robust
to outliers or corruptions in the observed covariates.
The concept of influence that was developed for regression diagnostics can be
used to detect such corrupted observations as shown in this paper. This property
of influence ? for which we also develop a randomized approximation ? motivates
our proposed subsampling algorithm for large scale corrupted linear regression
which limits the influence of data points since highly influential points contribute
most to the residual error. Under a general model of corrupted observations, we
show theoretically and empirically on a variety of simulated and real datasets that
our algorithm improves over the current state-of-the-art approximation schemes
for ordinary least squares.
1
Introduction
To improve scalability of the widely used ordinary least squares algorithm, a number of randomized
approximation algorithms have recently been proposed. These methods, based on subsampling the
dataset, reduce the computational time from O np2 to o(np2 )1 [14]. Most of these algorithms
are concerned with the classical fixed design setting or the case where the data is assumed to be
sampled i.i.d. typically from a sub-Gaussian distribution [7]. This is known to be an unrealistic
modelling assumption since real-world data are rarely well-behaved in the sense of the underlying
distributions.
We relax this limiting assumption by considering the setting where with some probability, the observed covariates are corrupted with additive noise. This scenario corresponds to a generalised
version of the classical problem of ?errors-in-variables? in regression analysis which has recently
been considered in the context of sparse estimation [12]. This corrupted observation model poses a
more realistic model of real data which may be subject to many different sources of measurement
noise or heterogeneity in the dataset.
A key consideration for sampling is to ensure that the points used for estimation are typical of the
full dataset. Typicality requires the sampling distribution to be robust against outliers and corrupted
points. In the i.i.d. sub-Gaussian setting, outliers are rare and can often easily be identified by
examining the statistical leverage scores of the datapoints.
Crucially, in the corrupted observation setting described in ?2, the concept of an outlying point
concerns the relationship between the observed predictors and the response. Now, leverage alone
cannot detect the presence of corruptions. Consequently, without using additional knowledge about
?
1
Authors contributed equally.
Informally: f (n) = o(g(n)) means f (n) grows more slowly than g(n).
1
the corrupted points, the OLS estimator (and its subsampled approximations) are biased. This also
rules out stochastic gradient descent (SGD) ? which is often used for large scale regression ? since
convex cost functions and regularizers which are typically used for noisy data are not robust with
respect to measurement corruptions.
This setting motivates our use of influence ? the effective impact of an individual datapoint exerts on
the overall estimate ? in order to detect and therefore avoid sampling corrupted points. We propose
an algorithm which is robust to corrupted observations and exhibits reduced bias compared with
other subsampling estimators.
Outline and Contributions. In ?2 we introduce our corrupted observation model before reviewing
the basic concepts of statistical leverage and influence in ?3. In ?4 we briefly review two subsampling
approaches to approximating least squares based on structured random projections and leverage
weighted importance sampling. Based on these ideas we present influence weighted subsampling
(IWS-LS), a novel randomized least squares algorithm based on subsampling points with small
influence in ?5.
In ?6 we analyse IWS-LS in the general setting where the observed predictors can be corrupted
with additive sub-Gaussian noise. Comparing the IWS-LS estimate with that of OLS and other
randomized least squares approaches we show a reduction in both bias and variance. It is important
to note that the simultaneous reduction in bias and variance is relative to OLS and randomized
approximations which are only unbiased in the non-corrupted setting. Our results rely on novel
finite sample characteristics of leverage and influence which we defer to ?SI.3. Additionally, in
?SI.4 we prove an estimation error bound for IWS-LS in the standard sub-Gaussian model.
Computing influence exactly is not practical in large-scale applications and so we propose two randomized approximation algorithms based on the randomized leverage approximation of [8]. Both
of these algorithms run in o(np2 ) time which improve scalability in large problems. Finally, in ?7
we present extensive experimental evaluation which compares the performance of our algorithms
against several randomized least squares methods on a variety of simulated and real datasets.
2
Statistical model
In this work we consider a variant of the standard linear model
(1)
y = X + ?,
where ? 2 R is a noise term independent of X 2 R
. However, rather than directly observing
X we instead observe Z where
Z = X + U W.
(2)
U = diag(u1 , . . . , un ) and ui is a Bernoulli random variable with probability ? of being 1.
W 2 Rn?p is a matrix of measurement corruptions. The rows of Z therefore are corrupted with
probability ? and not corrupted with probability (1 ?).
Definition 1 (Sub-gaussian matrix). A zero-mean matrix X is called sub-Gaussian with parameter
1
p
>
( n1 x2 , n1 ?x ) if (a) Each row x>
i 2 R is sampled independently and has E[xi xi ] = n ?x . (b) For
p
>
any unit vector v 2 R , v xi is a sub-Gaussian random variable with parameter at most p1p x .
n
n?p
We consider the specific instance of the linear corrupted observation model in Eqs. (1), (2) where
? X, W 2 Rn?p are sub-Gaussian with parameters ( n1 x2 , n1 ?x ) and ( n1
tively,
? ? 2 Rn is sub-Gaussian with parameters ( n1 ?2 , n1 ?2 In ),
2 1
w , n ?w )
respec-
and all are independent of each other.
The key challenge is that even when ? and the magnitude of the corruptions, w are relatively small,
the standard linear regression estimate is biased and can perform poorly (see ?6). Sampling methods
which are not sensitive to corruptions in the observations can perform even worse if they somehow
subsample a proportion rn > ?n of corrupted points. Furthermore, the corruptions may not be large
enough to be detected via leverage based techniques alone.
The model described in this section generalises the ?errors-in-variables? model from classical least
squares modelling. Recently, similar models have been studied in the high dimensional (p
n)
2
setting in [4?6, 12] in the context of robust sparse estimation. The ?low-dimensional? (n > p)
setting is investigated in [4], but the ?big data? setting (n
p) has not been considered so far.2
In the high-dimensional problem, knowledge of the corruption covariance, ?w [12], or the data
covariance ?x [5], is required to obtain a consistent estimate. This assumption may be unrealistic in
many settings. We aim to reduce the bias in our estimates without requiring knowledge of the true
covariance of the data or the corruptions, and instead sub-sample only non-corrupted points.
3
Diagnostics for linear regression
In practice, the sub-Gaussian linear model assumption is often violated either by heterogeneous
noise or by a corruption model as in ?2. In such scenarios, fitting a least squares model to the full
dataset is unwise since the outlying or corrupted points can have a large adverse effect on the model
fit. Regression diagnostics have been developed in the statistics literature to detect such points (see
e.g. [2] for a comprehensive overview). Recently, [14] proposed subsampling points for least squares
based on their leverage scores. Other recent works suggest related influence measures that identify
subspace [16] and multi-view [15] clusters in high dimensional data.
3.1
Statistical leverage
For the standard linear model in Eq. (1), the well known least squares solution is
b = arg min ky
X k2 = X > X
1
X> y.
(3)
The projection matrix I L with L := X(X> X) 1 X> specifies the subspace in which the residual
lies. The diagonal elements of the ?hat matrix? L, li := Lii , i = 1, . . . , n are the statistical leverage
scores of the ith sample. Leverage scores quantify to what extent a particular sample is an outlier
with respect to the distribution of X.
An equivalent definition from [14] which will be useful later concerns any matrix U 2 Rn?p which
spans the column space of X (for example, the matrix whose columns are the left singular vectors of
X). The statistical leverage scores of the rows of X are the squared row norms of U, i.e. li = kUi k2 .
Although the use of leverage can be motivated from the least squares solution in Eq. (3), the leverage scores do not take into account the relationship between the predictor variables and the response
variable y. Therefore, low-leverage points may have a weak predictive relationship with the response and vice-versa. In other words, it is possible for such points to be outliers with respect to the
conditional distribution P (y|X) but not the marginal distribution on X.
3.2
Influence
A concept that captures the predictive relationship between covariates and response is influence.
Influential points are those that might not be outliers in the geometric sense, but instead adversely
affect the estimated coefficients.
One way to assess the influence of a point is to compute the change in the learned model when
the point is removed from the estimation step. [2]. We can compute a leave-one-out least squares
estimator by straightforward application of the Sherman-Morrison-Woodbury formula (see Prop. 3
in ?SI.3):
where ei = yi
we have
2
3
b
i
= X> X
x>
i xi
1
X> y
b
x>
i yi =
? 1 x>
i ei
1 li
xi b OLS . Defining the influence3 , di as the change in expected mean squared error
?
di = b
b
i
?>
?
X> X b
b
i
?
=
e2i li
(1
li )
Unlike [5, 12] and others we do not consider sparsity in our solution since n
The expression we use is also called Cook?s distance [2].
3
2.
p.
Points with large values of di are those which, if added to the model, have the largest adverse effect
on the resulting estimate. Since influence only depends on the OLS residual error and the leverage
scores, it can be seen that the influence of every point can be computed at the cost of a least squares
fit. In the next section we will see how to approximate both quantities using random projections.
4
Fast randomized least squares algorithms
We briefly review two randomized approaches to least squares approximation: the importance
weighted subsampling approach of [9] and the dimensionality reduction approach [14]. The former proposes an importance sampling probability distribution according to which, a small number
of rows of X and y are drawn and used to compute the regression coefficients. If the sampling probabilities are proportional to the statistical leverages, the resulting estimator is close to the optimal
estimator [9]. We refer to this as LEV-LS.
The dimensionality reduction approach can be viewed as a random projection step followed by a
uniform subsampling. The class of Johnson-Lindenstrauss projections ? e.g. the SRHT ? has been
shown to approximately uniformize leverage scores in the projected space. Uniformly subsampling
the rows of the projected matrix proves to be equivalent to leverage weighted sampling on the original dataset [14]. We refer to this as SRHT-LS. It is analysed in the statistical setting by [7] who also
propose ULURU, a two step fitting procedure which aims to correct for the subsampling bias and
consequently converges to the OLS estimate at a rate independent of the number of subsamples [7].
Subsampled Randomized Hadamard Transform (SRHT) The SHRT consists of a preconditioning step after
q which nsubs rows of the new matrix are subsampled uniformly at random in the
n
following way nsubs
SHD ? X = ?X with the definitions [3]:
? S is a subsampling matrix.
? D is a diagonal matrix whose entries are drawn independently from { 1, 1}.
? H 2 Rn?n is a normalized Walsh-Hadamard matrix4 which is defined recursively as
?
?
Hn/2 Hn/2
+1 +1
Hn =
, H2 =
.
Hn/2
Hn/2
+1
1
We set H =
p1 Hn
n
so it has orthonormal columns.
As a result, the rows of the transformed matrix ?X have approximately uniform leverage scores.
(see [17] for detailed analysis of the SRHT). Due to the recursive nature of H, the cost of applying
the SRHT is O (pn log nsubs ) operations, where nsubs is the number of rows sampled from X [1].
The SRHT-LS algorithm solves b SRHT = arg min k?y ?X k2 which for an appropriate
2
? which satisfies
subsampling ratio, r = ?( ?p2 ) results in a residual error, e
(4)
k?
ek ? (1 + ?)kek
where e = y
X b OLS is the vector of OLS residual errors [14].
Randomized leverage computation Recently, a method based on random projections has been
proposed to approximate the leverage scores based on first reducing the dimensionality of the data
using the SRHT followed by computing the leverage scores using this low-dimensional approximation [8?10, 13].
The leverage approximation algorithm of [8] uses a SRHT, ?1 2 Rr1 ?n to first compute the approximate SVD of X,
>
?1 X = U?X ??X V?X
. Followed by a second SHRT ?2 2 Rp?r2 to compute an approximate
orthogonal basis for X
R
1
1
? = XR
= V?X ??X
2 Rp?p , U
1
?2 2 Rn?r2 .
(5)
4
For the Hadamard transform, n must be a power of two but other transforms exist (e.g. DCT, DFT) for
which similar theoretical guarantees hold and there is no restriction on n.
4
? ?li = kU
? i k2 .
The approximate leverage scores are now the squared row norms of U,
From [14] we derive the following result relating to randomized approximation of the leverage
?li ? (1 + ?l )li ,
(6)
where the approximation error, ?l depends on the choice of projection dimensions r1 and r2 .
The leverage weighted least squares (LEV-LS) algorithm samples rows of X and y with probability
proportional to li (or ?li in the approximate case) and performs least squares on this subsample. The
residual error resulting from the leverage weighted least squares is bounded by Eq. (4) implying
that LEV-LS and SRHT-LS are equivalent [14]. It is important to note that under the corrupted
observation model these approximations will be biased.
5
Influence weighted subsampling
In the corrupted observation model, OLS and therefore the random approximations to OLS described in ?4 obtain poor predictions. To remedy this, we propose influence weighted subsampling
(IWS-LS) which is described in Algorithm 1. IWS-LS subsamples
points according to the distriPn
bution, Pi = c/di where c is a normalizing constant so that i=1 Pi = 1. OLS is then estimated on
the subsampled points. The sampling procedure ensures that points with high influence are selected
infrequently and so the resulting estimate is less biased than the full OLS solution. Several approaches similar in spirit have previously been proposed based on identifying and down-weighting
the effect of highly influential observations [19].
Obviously, IWS-LS is impractical in the scenarios we consider since it requires the OLS residuals
and full leverage scores. However, we use this as a baseline and to simplify the analysis. In the next
section, we propose an approximate influence weighted subsampling algorithm which combines the
approximate leverage computation of [8] and the randomized least squares approach of [14].
Algorithm 1 Influence weighted subsampling
(IWS-LS).
Input: Data: Z, y
1: Solve b OLS = arg min ky Z k2
2: for i = 1 . . . n do
3:
ei = yi zi b OLS
>
1
4:
l i = z>
zi
i (Z Z)
5:
di = e2i li /(1 li )2
6: end for
? y
? ) of (Z, y) proportional to
7: Sample rows (Z,
8: Solve b IWS = arg min k?
y
Output: b IWS
? k2
Z
1
di
Algorithm 2 Residual weighted subsampling
(aRWS-LS)
Input: Data: Z, y
1: Solve b SRHT = arg min k? ? (y Z )k2
? = y Z b SRHT
2: Estimate residuals: e
?
? ) of (Z, y) proportional to
3: Sample rows (Z, y
1
e?2i
4: Solve b RW S = arg min k?
y
Output: b RW S
? k2
Z
Randomized approximation algorithms. Using the ideas from ?4 and ?4 we obtain the following
randomized approximation to the influence scores
d?i =
e?2i ?li
,
(1 ?li )2
(7)
where e?i is the ith residual error computed using the SRHT-LS estimator. Since the approximation errors of e?i and ?li are bounded (inequalities (4) and (6)), this suggests that our randomized
approximation to influence is close to the true influence.
Basic approximation. The first approximation algorithm is identical to Algorithm 1 except that
leverage and residuals are replaced by their randomized approximations as in Eq. (7). We refer to
this algorithm as Approximate influence weighted subsampling (aIWS-LS). Full details are given
in Algorithm 3 in ?SI.2.
5
Residual Weighted Sampling. Leverage scores are typically uniform [7, 13] for sub-Gaussian
data. Even in the corrupted setting, the difference in leverage scores between corrupted and noncorrupted points is small (see ?6). Therefore, the main contribution to the influence for each point
will originate from the residual error, e2i . Consequently, we propose sampling with probability
inversely proportional to the approximate residual, e?12 . The resulting algorithm Residual Weighted
i
Subsampling (aRWS-LS) is detailed in Algorithm 2. Although aRWS-LS is not guaranteed to be
a good approximation to IWS-LS, empirical results suggests that it works well in practise and is
faster to compute than aIWS-LS.
Computational complexity. Clearly, the computational complexity of IWS-LS is O np2 . The
computation complexity of aIWS-LS is O np log nsubs + npr2 + nsubs p2 , where the first term
is the cost of SRHT-LS, the second term is the cost of approximate leverage computation and the
last term solves OLS on the subsampled dataset. Here, r2 is the dimension of the random projection detailed in Eq. (5). The cost of aRWS-LS is O np log nsubs + np + nsubs p2 where
the first term is the cost of SRHT-LS, the second term is the cost of computing the residuals
e, and the last term solves OLS on the subsampled dataset. This computation can be reduced to
O np log nsubs + nsubs p2 . Therefore the cost of both aIWS-LS and aRWS-LS is o(np2 ).
6
Estimation error
In this section we will prove an upper bound on the estimation error of IWS-LS in the corrupted
model. First, we show that the OLS error consists of two additional variance terms that depend on the
size and proportion of the corruptions and an additional bias term. We then show that IWS-LS can
significantly reduce the relative variance and bias in this setting, so that it no longer depends on the
magnitude of the corruptions but only on their proportion. We compare these results to recent results
from [4, 12] suggesting that consistent estimation requires knowledge about ?w . More recently, [5]
show that incomplete knowledge about this quantity results in a biased estimator where the bias is
proportional to the uncertainty about ?w . We see that the form of our bound matches these results.
Inequalities are said to hold with high probability (w.h.p.) if the probability of failure is not more
than C1 exp( C2 log p) where C1 , C2 are positive constants that do not depend on the scaling quantities n, p, w . The symbol . means that we ignore constants that do not depend on these scaling
quantities. Proofs are provided in the supplement. Unless otherwise stated, k?k denotes the `2 norm
for vectors and the spectral norm for matrices.
Corrupted observation model. As a baseline, we first investigate the behaviour of the OLS estimator in the corrupted model.
2 2
Theorem 1 (A bound on k b OLS
k). If n & minx (?wx ) p log p then w.h.p.
!
r
p log p
1
2
2p
b
k OLS
k.
+ ? w pk k ?
(8)
? x+? ? w +?
w + w x k k
n
where 0 < ? min (?x ) + ? min (?w ).
Remark 1 (No corruptions case). Notice for a fixed w , taking lim?!0 or for a fixed ? taking
lim w !0 (i.e. there are no corruptions) the above error reduces to the least squares result (see for
example [4]).
p
Remark 2 (Variance and Bias). The first three terms inp
(8) scale with 1/n so as n ! 1, these
terms tend towards 0. The last term does not depend on 1/n and so for some non-zero ? the least
squares estimate will incur some bias depending on the fraction and magnitude of corruptions.
We are now ready to state our theorem characterising the mean squared error of the influence
weighted subsampling estimator.
2
2
Theorem 2 (Influence sampling in the corrupted model). For n & minx(?w?x ) p log p we have
!
?
?r
? ?
p log p
1
p
b
k IWS
k.
+ ?k k
+ ? pk k .
? x+
( w + 1)
nsubs
where 0 <
?
min (??x )
and ??x is the covariance of the influence weighted subsampled data.
6
(b) Leverage (0.1)
(a) Influence (1.1)
Figure 1: Comparison of the distribution of the influence and leverage for corrupted and noncorrupted points. The `1 distance between the histograms is shown in brackets.
Remark 3. Theorem 2 states that the influence weighted subsampling estimator removes p
the proportional dependance
of
the
error
on
so
the
additional
variance
terms
scale
as
O(?/
?
p/nsubs )
w
w
p
p
and O(? p/nsubs ). The relative contribution of the bias term is ? pk k compared with
2p
? w
pk k for the OLS or non-influence-based subsampling methods.
Comparison with fully corrupted setting. We note that the bound in Theorem 1 is similar to the
bound in [5] for an estimator where all data points are corrupted (i.e. ? = 1) and where incomplete
knowledge of the covariance matrix of the corruptions, ?w is used. The additional bias in the
estimator is proportional to the uncertainty in the estimate of ?w ? in Theorem 1 this corresponds to
2
w . Unbiased estimation is possible if ?w is known. See the Supplementary Information for further
discussion, where the relevant results from [5] are provided in Section SI.6.1 as Lemma 16.
7
Experimental results
We compare IWS-LS against the methods SRHT-LS [14], ULURU [7]. These competing methods
represent current state-of-the-art in fast randomized least squares. Since SRHT-LS is equivalent to
LEV-LS [9] the comparison will highlight the difference between importance sampling according
to the two difference types of regression diagnostic in the corrupted model. Similar to IWS-LS,
ULURU is also a two-step procedure where the first is equivalent to SRHT-LS. The second reduces
bias by subtracting the result of regressing onto the residual. The experiments with the corrupted
data model will demonstrate the difference in robustness of IWS-LS and ULURU to corruptions in
the observations. Note that we do not compare with SGD. Although SGD has excellent properties
for large-scale linear regression, we are not aware of a convex loss function which is robust to the
corruption model we propose.
We assess the empirical performance of our method compared with standard and state-of-the-art
randomized approaches to linear regression in several difference scenarios. We evaluate these methods on the basis of the estimation error: the `2 norm of the difference between the true weights and
the learned weights, k b
k. We present additional results for root mean squared prediction error
(RMSE) on the test set in ?SI.7.
For all the experiments on simulated data sets we use ntrain = 100, 000, ntest = 1000, p = 500.
For datasets of this size, computing exact leverage is impractical and so we report on results for
IWS-LS in ?SI.7. For aIWS-LS and aRWS-LS we used the same number of sub-samples to
approximate the leverage scores and residuals as for solving the regression. For aIWS-LS we set
r2 = p/2 (see Eq. (5)). The results are averaged over 100 runs.
Corrupted data. We investigate the corrupted data noise model described in Eqs. (1)-(2). We
show three scenarios where ? = {0.05, 0.1, 0.3}. X and W were sampled from independent, zeromean Gaussians with standard deviation x = 1 and w = 0.4 respectively. The true regression
coefficients, were sampled from a standard Gaussian. We added i.i.d. zero-mean Gaussian noise
with standard deviation e = 0.1.
Figure 1 shows the difference in distribution of influence and leverage between non-corrupted points
(top) and corrupted points (bottom) for a dataset with 30% corrupted points. The distribution of
leverage is very similar between the corrupted and non-corrupted points, as quantified by the `1
difference. This suggests that leverage alone cannot be used to identify corrupted points.
7
(a) 5% Corruptions
(b) 30% Corruptions
(c) Airline delay
Figure 2: Comparison of mean estimation error and standard deviation on two corrupted simulated
datasets and the airline delay dataset.
On the other hand, although there are some corrupted points with small influence, they typically
have a much larger influence than non-corrupted points. We give a theoretical explanation of this
phenomenon in ?SI.3 (remarks 4 and 5).
Figure 2(a) and (b) shows the estimation error and the mean squared prediction error for different
subsample sizes. In this setting, computing IWS-LS is impractical (due to the exact leverage computation) so we omit the results but we notice that aIWS-LS and aRWS-LS quickly improve over
the full least squares solution and the other randomized approximations in all simulation settings. In
all cases, influence based methods also achieve lower-variance estimates.
For 30% corruptions for a small number of samples ULURU outperforms the other subsampling
methods. However, as the number of samples increases, influence based methods start to outperform
OLS. Here, ULURU converges quickly to the OLS solution but is not able to overcome the bias
introduced by the corrupted datapoints. Results for 10% corruptions are shown in Figs. 5 and 6 and
we provide results on smaller corrupted datasets (to show the performance of IWS-LS) as well as
non-corrupted data simulated according to [13] in ?SI.7.
Airline delay dataset The dataset consists of details of all commercial flights in the USA over 20
years. Dataset along with visualisations available from http://stat-computing.org/dataexpo/2009/.
Selecting the first ntrain = 13, 000 US Airways flights from January 2000 (corresponding to approximately 1.5 weeks) our goal is to predict the delay time of the next ntest = 5, 000 US Airways
flights. The features in this dataset consist of a binary vector representing origin-destination pairs
and a real value representing distance (p = 170).
The dataset might be expected to violate the usual i.i.d. sub-Gaussian design assumption of standard
linear regression since the length of delays are often very different depending on the day. For
example, delays may be longer due to public holidays or on weekends. Of course, such regular
events could be accounted for in the modelling step, but some unpredictable outliers such as weather
delay may also occur. Results are presented in Figure 2(c), the RMSE is the error in predicted delay
time in minutes. Since the dataset is smaller, we can run IWS-LS to observe the accuracy of
aIWS-LS and aRWS-LS in comparison. For more than 3000 samples, these algorithm outperform
OLS and quickly approach IWS-LS. The result suggests that the corrupted observation model is a
good model for this dataset. Furthermore, ULURU is unable to achieve the full accuracy of the OLS
solution.
8
Conclusions
We have demonstrated theoretically and empirically under the generalised corrupted observation
model that influence weighted subsampling is able to significantly reduce both the bias and variance
compared with the OLS estimator and other randomized approximations which do not take influence
into account. Importantly our fast approximation, aRWS-LS performs similarly to IWS-LS. We
find ULURU quickly converges to the OLS estimate, although it is not able to overcome the bias
induced by the corrupted datapoints despite its two-step procedure. The performance of IWS-LS
relative to OLS in the airline delay problem suggests that the corrupted observation model is a more
realistic modelling scenario than the standard sub-Gaussian design model for some tasks. Software
is available at http://people.inf.ethz.ch/kgabriel/software.html.
Acknowledgements. We thank David Balduzzi, Cheng Soon Ong and the anonymous reviewers
for invaluable discussions, suggestions and comments.
8
References
[1] Nir Ailon and Edo Liberty. Fast dimension reduction using rademacher series on dual bch
codes. In 19th Annual ACM-SIAM Symposium on Discrete Algorithms, pages 1?9, 2008.
[2] David A Belsley, Edwin Kuh, and Roy E Welsch. Regression Diagnostics. Identifying Influential Data and Sources of Collinearity. Wiley, 1981.
[3] Christos Boutsidis and Alex Gittens. Improved matrix algorithms via the Subsampled Randomized Hadamard Transform. 2012. arXiv:1204.0062v4 [cs.DS].
[4] Yudong Chen and Constantine Caramanis. Orthogonal Matching Pursuit with Noisy and Missing Data: Low and High Dimensional Results. June 2012. arXiv:1206.0823.
[5] Yudong Chen and Constantine Caramanis. Noisy and Missing Data Regression: DistributionOblivious Support Recovery. In International Conference on Machine Learning, 2013.
[6] Yudong Chen, Constantine Caramanis, and Shie Mannor. Robust Sparse Regression under
Adversarial Corruption. In International Conference on Machine Learning, 2013.
[7] P Dhillon, Y Lu, D P Foster, and L Ungar. New Subsampling Algorithms for Fast Least
Squares Regression. In Advances in Neural Information Processing Systems, 2013.
[8] Petros Drineas, Malik Magdon-Ismail, Michael W Mahoney, and David P Woodruff. Fast approximation of matrix coherence and statistical leverage. September 2011. arXiv:1109.3843v2
[cs.DS].
[9] Petros Drineas, Michael W. Mahoney, and S. Muthukrishnan. Sampling algorithms for l2
regression and applications. In Proceedings of the Seventeenth Annual ACM-SIAM Symposium
on Discrete Algorithm, SODA ?06, pages 1127?1136, New York, NY, USA, 2006. ACM.
[10] Petros Drineas, Michael W Mahoney, S Muthukrishnan, and Tam?as Sarl?os. Faster least squares
approximation. Numerische Mathematik, 117(2):219?249, 2011.
[11] Daniel Hsu, Sham Kakade, and Tong Zhang. A tail inequality for quadratic forms of subgaussian random vectors. Electron. Commun. Probab., 17:no. 52, 1?6, 2012.
[12] Po-Ling Loh and Martin J Wainwright. High-dimensional regression with noisy and missing
data: Provable guarantees with nonconvexity. The Annals of Statistics, 40(3):1637?1664, June
2012.
[13] Ping Ma, Michael W Mahoney, and Bin Yu. A Statistical Perspective on Algorithmic Leveraging. In proceedings of the International Conference on Machine Learning, 2014.
[14] Michael W Mahoney. Randomized algorithms for matrices and data.
arXiv:1104.5557v3 [cs.DS].
April 2011.
[15] Brian McWilliams and Giovanni Montana. Multi-view predictive partitioning in high dimensions. Statistical Analysis and Data Mining, 5(4):304?321, 2012.
[16] Brian McWilliams and Giovanni Montana. Subspace clustering of high-dimensional data: a
predictive approach. Data Mining and Knowledge Discovery, 28:736?772, 2014.
[17] Joel A Tropp. Improved analysis of the subsampled randomized Hadamard transform. November 2010. arXiv:1011.1595v4 [math.NA].
[18] Roman Vershynin. Introduction to the non-asymptotic analysis of random matrices. November
2010. arXiv:1011.3027.
[19] Roy E Welsch. Regression sensitivity analysis and bounded-influence estimation. In Evaluation of econometric models, pages 153?167. Academic Press, 1980.
9
| 5428 |@word collinearity:1 briefly:2 version:1 proportion:3 norm:5 simulation:1 crucially:1 covariance:5 sgd:3 recursively:1 reduction:5 series:1 score:17 selecting:1 woodruff:1 daniel:1 outperforms:1 current:2 comparing:1 analysed:1 si:9 must:1 dct:1 additive:2 realistic:2 wx:1 remove:1 alone:3 implying:1 selected:1 cook:1 ntrain:2 ith:2 mannor:1 contribute:1 math:1 org:1 zhang:1 along:1 c2:2 symposium:2 prove:2 consists:3 fitting:2 combine:1 introduce:1 theoretically:2 expected:2 p1:1 multi:2 unpredictable:1 considering:1 provided:2 underlying:1 bounded:3 what:1 developed:2 impractical:3 guarantee:2 every:1 exactly:1 k2:8 partitioning:1 mcwilliams:3 unit:1 omit:1 generalised:2 before:1 positive:1 limit:1 despite:1 lev:4 approximately:3 might:2 studied:1 quantified:1 montana:2 suggests:5 walsh:1 averaged:1 seventeenth:1 practical:1 woodbury:1 practice:1 recursive:1 xr:1 procedure:4 empirical:2 eth:1 significantly:2 weather:1 projection:8 matching:1 word:1 inp:1 regular:1 suggest:1 cannot:2 close:2 onto:1 context:2 influence:40 applying:1 restriction:1 equivalent:5 demonstrated:1 reviewer:1 missing:3 straightforward:1 urich:1 independently:2 l:52 typicality:1 convex:2 numerische:1 identifying:2 recovery:1 estimator:13 rule:1 importantly:1 orthonormal:1 datapoints:3 srht:18 holiday:1 e2i:3 limiting:1 annals:1 commercial:1 exact:2 us:1 origin:1 element:1 infrequently:1 roy:2 observed:4 bottom:1 capture:1 ensures:1 removed:1 ui:1 covariates:3 complexity:3 practise:1 ong:1 depend:4 reviewing:1 solving:1 predictive:4 incur:1 basis:2 preconditioning:1 edwin:1 drineas:3 easily:1 po:1 caramanis:3 muthukrishnan:2 weekend:1 fast:7 effective:1 detected:1 sarl:1 whose:2 widely:1 solve:4 supplementary:1 larger:1 relax:1 otherwise:1 statistic:2 analyse:1 noisy:4 transform:4 obviously:1 subsamples:2 propose:7 subtracting:1 relevant:1 hadamard:5 uniformize:1 poorly:1 achieve:2 ismail:1 scalability:2 ky:2 cluster:1 r1:1 rademacher:1 leave:1 converges:3 derive:1 develop:1 depending:2 stat:1 pose:1 eq:8 p2:4 solves:3 predicted:1 c:3 quantify:1 switzerland:1 liberty:1 correct:1 stochastic:1 public:1 bin:1 behaviour:1 ungar:1 anonymous:1 brian:3 hold:2 considered:2 exp:1 algorithmic:1 week:1 predict:1 electron:1 bch:1 estimation:15 iw:25 sensitive:1 largest:1 vice:1 weighted:18 clearly:1 gaussian:15 aim:2 rather:1 avoid:1 pn:1 np2:5 june:2 joachim:1 modelling:4 bernoulli:1 adversarial:1 baseline:2 detect:4 sense:2 typically:5 visualisation:1 transformed:1 overall:1 arg:6 html:1 dual:1 proposes:1 art:3 marginal:1 aware:1 sampling:14 identical:1 yu:1 others:1 np:4 simplify:1 report:1 roman:1 comprehensive:1 individual:1 subsampled:9 replaced:1 n1:7 highly:2 investigate:2 mining:2 evaluation:2 regressing:1 joel:1 mahoney:5 bracket:1 diagnostics:4 regularizers:1 orthogonal:2 unless:1 incomplete:2 theoretical:2 instance:1 column:3 ordinary:2 cost:9 deviation:3 entry:1 rare:1 uniform:3 predictor:3 delay:9 examining:1 johnson:1 corrupted:51 vershynin:1 international:3 randomized:25 siam:2 sensitivity:1 destination:1 v4:2 michael:5 quickly:4 na:1 squared:6 hn:6 slowly:1 worse:1 adversely:1 lii:1 ek:1 tam:1 li:15 account:2 suggesting:1 coefficient:3 depends:3 later:1 view:2 root:1 mario:1 observing:1 bution:1 start:1 defer:1 rmse:2 contribution:3 ass:2 square:27 accuracy:2 kek:1 variance:8 characteristic:1 who:1 identify:2 weak:1 lu:1 corruption:23 datapoint:1 simultaneous:1 ping:1 edo:1 definition:3 against:3 failure:1 boutsidis:1 proof:1 di:6 petros:3 sampled:5 hsu:1 dataset:16 knowledge:7 lim:2 improves:1 dimensionality:3 day:1 response:4 improved:2 april:1 zeromean:1 furthermore:2 d:3 hand:1 flight:3 tropp:1 ei:3 o:1 somehow:1 behaved:1 grows:1 usa:2 effect:3 concept:4 unbiased:2 requiring:1 true:4 former:1 normalized:1 remedy:1 dhillon:1 outline:1 demonstrate:1 performs:2 invaluable:1 characterising:1 lucic:2 consideration:1 novel:2 recently:7 ols:29 empirically:2 tively:1 overview:1 tail:1 relating:1 measurement:3 refer:3 versa:1 rr1:1 dft:1 similarly:1 sherman:1 longer:2 recent:2 perspective:1 constantine:3 inf:2 commun:1 scenario:6 inequality:3 binary:1 yi:3 seen:1 additional:6 v3:1 morrison:1 full:7 violate:1 sham:1 reduces:2 generalises:1 faster:2 match:1 unwise:1 academic:1 equally:1 impact:1 prediction:3 variant:1 regression:21 basic:2 heterogeneous:1 exerts:1 arxiv:6 histogram:1 represent:1 c1:2 singular:1 source:2 biased:5 unlike:1 airline:4 comment:1 subject:1 tend:1 induced:1 shie:1 leveraging:1 spirit:1 subgaussian:1 leverage:43 presence:1 enough:1 concerned:1 variety:2 affect:1 fit:2 zi:2 identified:1 competing:1 reduce:4 idea:2 motivated:1 expression:1 loh:1 york:1 remark:4 gabriel:2 useful:1 detailed:3 informally:1 transforms:1 rw:2 reduced:2 http:2 specifies:1 outperform:2 exist:1 notice:2 estimated:2 diagnostic:1 discrete:2 key:2 drawn:2 nonconvexity:1 shd:1 econometric:1 fraction:1 year:1 run:3 uncertainty:2 soda:1 coherence:1 scaling:2 bound:6 followed:3 guaranteed:1 cheng:1 quadratic:1 annual:2 occur:1 krummenacher:2 alex:1 x2:2 software:2 u1:1 speed:1 min:9 span:1 relatively:1 martin:1 department:1 influential:4 structured:1 according:4 ailon:1 poor:1 smaller:2 gittens:1 kakade:1 outlier:7 previously:1 mathematik:1 end:1 available:2 operation:1 gaussians:1 pursuit:1 magdon:1 observe:2 v2:1 appropriate:1 spectral:1 shrt:2 robustness:1 hat:1 rp:2 original:1 denotes:1 top:1 subsampling:27 ensure:1 clustering:1 balduzzi:1 prof:1 approximating:1 classical:3 malik:1 added:2 quantity:4 usual:1 diagonal:2 said:1 september:1 exhibit:1 gradient:1 minx:2 subspace:3 distance:3 unable:1 thank:1 simulated:5 nsubs:13 originate:1 extent:1 provable:1 length:1 code:1 relationship:4 ratio:1 matrix4:1 stated:1 design:3 motivates:2 contributed:1 perform:2 upper:1 observation:16 datasets:5 finite:1 descent:1 november:2 january:1 heterogeneity:1 defining:1 rn:7 introduced:1 david:3 pair:1 required:1 extensive:1 learned:2 able:3 sparsity:1 challenge:1 explanation:1 wainwright:1 unrealistic:2 power:1 event:1 rely:1 buhmann:1 residual:18 representing:2 scheme:1 improve:3 inversely:1 ready:1 nir:1 review:2 literature:1 geometric:1 uluru:8 acknowledgement:1 l2:1 probab:1 relative:4 discovery:1 asymptotic:1 fully:1 loss:1 highlight:1 suggestion:1 proportional:8 h2:1 consistent:2 jbuhmann:1 foster:1 pi:2 row:13 course:1 accounted:1 last:3 soon:1 bias:16 taking:2 sparse:3 overcome:2 dimension:4 yudong:3 world:1 lindenstrauss:1 giovanni:2 author:1 projected:2 outlying:2 far:1 approximate:12 ignore:1 kuh:1 assumed:1 xi:5 un:1 additionally:1 nature:1 ku:1 robust:8 kui:1 investigated:1 excellent:1 diag:1 pk:4 main:1 big:1 noise:7 subsample:3 ling:1 fig:1 ny:1 wiley:1 tong:1 christos:1 sub:15 airway:2 lie:1 weighting:1 formula:1 down:1 theorem:6 minute:1 specific:1 symbol:1 r2:5 concern:2 normalizing:1 consist:1 dependance:1 importance:4 supplement:1 magnitude:3 chen:3 welsch:2 ch:2 corresponds:2 satisfies:1 acm:3 ma:1 prop:1 conditional:1 viewed:1 goal:1 consequently:3 towards:1 adverse:2 change:2 typical:1 respec:1 uniformly:2 reducing:1 except:1 lemma:1 called:2 belsley:1 ntest:2 experimental:2 svd:1 rarely:1 people:1 support:1 ethz:2 violated:1 evaluate:1 phenomenon:1 |
4,892 | 5,429 | Fast Multivariate Spatio-temporal Analysis
via Low Rank Tensor Learning
Mohammad Taha Bahadori?
Dept. of Electrical Engineering
Univ. of Southern California
Los Angeles, CA 90089
[email protected]
Qi (Rose) Yu?
Dept. of Computer Science
Univ. of Southern California
Los Angeles, CA 90089
[email protected]
Yan Liu
Dept. of Computer Science
Univ. of Southern California
Los Angeles, CA 90089
[email protected]
Abstract
Accurate and efficient analysis of multivariate spatio-temporal data is critical in
climatology, geology, and sociology applications. Existing models usually assume
simple inter-dependence among variables, space, and time, and are computationally expensive. We propose a unified low rank tensor learning framework for multivariate spatio-temporal analysis, which can conveniently incorporate different
properties in spatio-temporal data, such as spatial clustering and shared structure
among variables. We demonstrate how the general framework can be applied to
cokriging and forecasting tasks, and develop an efficient greedy algorithm to solve
the resulting optimization problem with convergence guarantee. We conduct experiments on both synthetic datasets and real application datasets to demonstrate
that our method is not only significantly faster than existing methods but also
achieves lower estimation error.
1
Introduction
Spatio-temporal data provide unique information regarding ?where? and ?when?, which is essential
to answer many important questions in scientific studies from geology, climatology to sociology. In
the context of big data, we are confronted with a series of new challenges when analyzing spatiotemporal data because of the complex spatial and temporal dependencies involved.
A plethora of excellent work has been conducted to address the challenge and achieved successes to
a certain extent [8, 13]. Often times, geostatistical models use cross variogram and cross covariance
functions to describe the intrinsic dependency structure. However, the parametric form of cross
variogram and cross covariance functions impose strong assumptions on the spatial and temporal
correlation, which requires domain knowledge and manual work. Furthermore, parameter learning
of those statistical models is computationally expensive, making them infeasible for large-scale
applications.
Cokriging and forecasting are two central tasks in multivariate spatio-temporal analysis. Cokriging
utilizes the spatial correlations to predict the value of the variables for new locations. One widely
adopted method is multitask Gaussian process (MTGP) [4], which assumes a Gaussian process prior
over latent functions to directly induce correlations between tasks. However, for a cokriging task
with M variables of P locations for T time stamps, the time complexity of MTGP is O(M 3 P 3 T )
[4]. For forecasting, popular methods in multivariate time series analysis include vector autoregressive (VAR) models, autoregressive integrated moving average (ARIMA) models, and cointegration
models. An alternative method for spatio-temporal analysis is Bayesian hierarchical spatio-temporal
models with either separable and non-separable space-time covariance functions [6]. Rank reduced
?
Authors have equal contributions.
1
models have been proposed to capture the inter-dependency among variables [1]. However, very
few models can directly handle the correlations among variables, space and time simultaneously in
a scalable way. In this paper, we aim to address this problem by presenting a unified framework for
many spatio-temporal analysis tasks that are scalable for large-scale applications.
Tensor representation provides a convenient way to capture inter-dependencies along multiple dimensions. Therefore it is natural to represent the multivariate spatio-temporal data in tensor. Recent
advances in low rank learning have led to simple models that can capture the commonalities among
each mode of the tensor [15, 20]. Similar argument can be found in the literature of spatial data recovery [11], neuroimaging analysis [26], and multi-task learning [20]. Our work builds upon recent
advances in low rank tensor learning [15, 11, 26] and further considers the scenario where additional
side information of data is available. For example, in geo-spatial applications, apart from measurements of multiple variables, geographical information is available to infer location adjacency; in
social network applications, friendship network structure is collected to obtain preference similarity.
To utilize the side information, we can construct a Laplacian regularizer from the similarity matrices,
which favors locally smooth solutions.
We develop a fast greedy algorithm for learning low rank tensors based on the greedy structure
learning framework [2, 24, 21]. Greedy low rank tensor learning is efficient, as it does not require
full singular value decomposition of large matrices as opposed to other alternating direction methods
[11]. We also provide a bound on the difference between the loss function at our greedy solution
and the one at the globally optimal solution. Finally, we present experiment results on simulation
datasets as well as application datasets in climate and social network analysis, which show that our
algorithm is faster and achieves higher prediction accuracy than state-of-art approaches in cokriging
and forecasting tasks.
2
Tensor formulation for multivariate spatio-temporal analysis
The critical element in multivariate spatio-temporal analysis is an efficient way to incorporate the
spatial temporal correlations into modeling and automatically capture the shared structures across
variables, locations, and time. In this section, we present a unified low rank tensor learning framework that can perform various types of spatio-temporal analysis. We will use two important applications, i.e., cokriging and forecasting, to motivate and describe the framework.
2.1
Cokriging
In geostatistics, cokriging is the task of interpolating the data of one variable for unknown locations
by taking advantage of the observations of variables from known locations. For example, by making
use of the correlations between precipitation and temperature, we can obtain more precise estimate
of temperature in unknown locations than univariate kriging. Formally, denote the complete data
for P locations over T time stamps with M variables as X ? RP ?T ?M . We only observe the
measurements for a subset of locations ? ? {1, . . . , P } and their side information such as longitude
and latitude. Given the measurements X? and the side information, the goal is to estimate a tensor
W ? RP ?T ?M that satisfies W? = X? . Here X? represents the outcome of applying the index
operator I? to X:,:,m for all variables m = 1, . . . , M . The index operator I? is a diagonal matrix
whose entries are one for the locations included in ? and zero otherwise.
Two key consistency principles have been identified for effective cokriging [9, Chapter 6.2]: (1)
Global consistency: the data on the same structure are likely to be similar. (2) Local consistency: the
data in close locations are likely to be similar. The former principle is akin to the cluster assumption
in semi-supervised learning [25]. We incorporate these principles in a concise and computationally
efficient low-rank tensor learning framework.
To achieve global consistency, we constrain the tensor W to be low rank. The low rank assumption
is based on the belief that high correlations exist within variables, locations and time, which leads to
natural clustering of the data. Existing literature have explored the low rank structure among these
three dimensions separately, e.g., multi-task learning [19] for variable correlation, fixed rank kriging
[7] for spatial correlations. Low rankness assumes that the observed data can be described with a
few latent factors. It enforces the commonalities along three dimensions without an explicit form
for the shared structures in each dimension.
2
For local consistency, we construct a regularizer via the spatial Laplacian matrix. The Laplacian
matrix is defined as L = D ?
PA, where A is a kernel matrix constructed by pairwise similarity
and diagonal matrix Di,i =
j (Ai,j ). Similar ideas have been explored in matrix completion
[16]. In cokriging literature, the local consistency is enforced via the spatial covariance matrix. The
Bayesian models often impose the Gaussian process prior on the observations with the covariance
matrix K = Kv ? Kx where Kv is the covariance between variables and Kx is that for locations.
The Laplacian regularization term corresponds to the relational Gaussian process [5] where the
covariance matrix is approximated by the spatial Laplacian.
In summary, we can perform cokriging and find the value of tensor W by solving the following
optimization problem:
(
)
M
X
2
>
c
W = argmin kW? ? X? kF + ?
tr(W:,:,m LW:,:,m )
s.t.
rank(W) ? ?,
(1)
W
m=1
qP
2
where the Frobenius norm of a tensor A is defined as kAkF =
i,j,k Ai,j,k and ?, ? > 0
are the parameters that make tradeoff between the local and global consistency, respectively. The
low rank constraint finds the principal components of the tensor and reduces the complexity of
the model while the Laplacian regularizer clusters the data using the relational information among
the locations. By learning the right tradeoff between these two techniques, our method is able to
benefit from both. Due to the various definitions of tensor rank, we use rank as supposition for rank
complexity, which will be specified in later section.
2.2
Forecasting
Forecasting estimates the future value of multivariate time series given historical observations.
For ease of presentation, we use the classical VAR model with K lags and coefficient tensor
W ? RP ?KP ?M as an example. Using the matrix representation, the VAR(K) process defines
the following data generation process:
X:,t,m = W:,:,m Xt,m + E:,t,m , for m = 1, . . . , M and t = K + 1, . . . , T,
(2)
>
>
]> denotes the concatenation of K-lag historical data before
where Xt,m = [X:,t?1,m
, . . . , X:,t?K,m
time t. The noise tensor E is a multivariate Gaussian with zero mean and unit variance .
Existing multivariate regression methods designed to capture the complex correlations, such as
Tucker decomposition [20], are computationally expensive. A scalable solution requires a simpler
model that also efficiently accounts for the shared structures in variables, space, and time. Similar
global and local consistency principles still hold in forecasting. For global consistency, we can use
low rank constraint to capture the commonalities of the variables as well as the spatial correlations
on the model parameter tensor, as in [8]. For local consistency, we enforce the predicted value
for close locations to be similar via spatial Laplacian regularization. Thus, we can formulate the
forecasting task as the following optimization problem over the model coefficient tensor W:
(
)
M
X
2
>
c
b
b
b
W = argmin kX ? X kF + ?
tr(X:,:,m LX:,:,m )
s.t. rank(W) ? ?, Xb:,t,m = W:,:,m Xt,m
W
m=1
(3)
Though cokriging and forecasting are two different tasks, we can easily see that both formulations
follow the global and local consistency principles and can capture the inter-correlations from spatialtemporal data.
2.3
Unified Framework
We now show that both cokriging and forecasting can be formulated into the same tensor learning
framework. Let us rewrite the loss function in Eq. (1) and Eq. (3) in the form of multitask regression
and complete the quadratic form for the loss function. The cokriging task can be reformulated as
follows:
( M
)
X
> ?1
2
c
W = argmin
kW:,:,m H ? (H ) X?,m kF
s.t. rank(W) ? ?
(4)
W
m=1
3
where we define HH > = I? + ?L.1 For the forecasting problem, HH > = IP + ?L and we have:
( M
)
T
X X
c = argmin
W
kHW:,:,m Xt,m ? (H ?1 )X:,t,m k2F
s.t. rank(W) ? ?,
(5)
W
m=1 t=K+1
By slight change of notation (cf. Appendix D), we can easily see that the optimal solution of both
problems can be obtained by the following optimization problem with appropriate choice of tensors
Y and V:
( M
)
X
c = argmin
W
kW:,:,m Y:,:,m ? V:,:,m k2F
s.t. rank(W) ? ?.
(6)
W
m=1
After unifying the objective function, we note that tensor rank has different notions such as CP
rank, Tucker rank and mode n-rank [15, 11]. In this paper, we choose the mode-n rank, which is
computationally more tractable [11, 23]. The mode-n rank of a tensor W is the rank of its mode-n
unfolding W(n) .2 In particular, for a tensor W with N mode, we have the following definition:
mode-n rank(W) =
N
X
rank(W(n) ).
(7)
n=1
A common practice to solve this formulation with mode n-rank constraint is to relax the rank constraint to a convex nuclear norm constraint [11, 23]. However, those methods are computationally
expensive since they need full singular value decomposition of large matrices. In the next section,
we present a fast greedy algorithm to tackle the problem.
3
Fast greedy low rank tensor learning
To solve the non-convex problem in Eq. (6) and find its optimal solution, we propose a greedy
learning algorithm by successively adding rank-1 estimation of the mode-n unfolding. The main
idea of the algorithm is to unfold the tensor into a matrix, seek for its rank-1 approximation and
then fold back into a tensor with same dimensionality. We describe this algorithm in three steps:
(i) First, we show that we can learn rank-1 matrix estimations efficiently by solving a generalized
eigenvalue problem, (ii) We use the rank-1 matrix estimation to greedily solve the original tensor
rank constrained problem, and (iii) We propose an enhancement via orthogonal projections after
each greedy step.
Optimal rank-1 Matrix Learning The following lemma enables us to find such optimal rank-1
estimation of the matrices.
Lemma 1. Consider the following rank constrained problem:
n
o
2
b1 = argmin
A
kY ? AXkF ,
(8)
A:rank(A)=1
q?n
p?n
b1 can be written as
where Y ? R
,X ? R
, and A ? Rq?p . The optimal solution of A
>
b
b
b
b
A1 = uv , kb
vk2 = 1 where v is the dominant eigenvector of the following generalized eigenvalue
problem:
(XY > Y X > )v = ?(XX > )v
(9)
b
and u can be computed as
1
b.
b= >
Y X >v
(10)
u
b XX > v
b
v
Proof is deferred to Appendix A. Eq. (9) is a generalized eigenvalue problem whose dominant
eigenvector can be found efficiently [12]. If XX > is full rank, as assumed in Theorem 2, the
problem is simplified to a regular eigenvalue problem whose dominant eigenvector can be efficiently
computed.
1
We can use Cholesky decomposition to obtain H. In the rare cases that I? + ?L is not full rank, IP is
added where is a very small positive value.
2
The mode-n unfolding of a tensor is the matrix resulting from treating n as the first mode of the matrix,
and cyclically concatenating other modes. Tensor refolding is the reverse direction operation [15].
4
Algorithm 1 Greedy Low-rank Tensor Learning
1: Input: transformed data Y, V of M variables, stopping criteria ?
2: Output: N mode tensor W
3: Initialize W ? 0
4: repeat
5:
for n = 1 to N do
6:
Bn ? argmin L(refold(W(n) + B); Y, V)
B: rank(B)=1
7:
8:
9:
?n ? L(W; Y, V) ? L(refold(W(n) + Bn ); Y, V)
end for
n? ? argmax{?n }
n
10:
11:
12:
13:
if ?n? > ? then
W ? W + refold(Bn? , n? )
end if
W ? argminrow(A(1) )?row(W(1) ) L(A; Y, V)
# Optional Orthogonal Projection Step.
col(A(1) )?col(W(1) )
14: until ?n? < ?
Greedy Low n-rank Tensor Learning The optimal rank-1 matrix learning serves as a basic element in our greedy algorithm. Using Lemma 1, we can solve the problem in Eq. (6) in the Forward
Greedy Selection framework as follows: at each iteration of the greedy algorithm, it searches for the
mode that gives the largest decrease in the objective function. It does so by unfolding the tensor in
that mode and finding the best rank-1 estimation of the unfolded tensor. After finding the optimal
mode, it adds the rank-1 estimate in that mode to the current estimation of the tensor. Algorithm
PM
1 shows the details of this approach, where L(W; Y, V) = m=1 kW:,:,m Y:,:,m ? V:,:,m k2F . Note
that we can find the optimal rank-1 solution in only one of the modes, but it is enough to guarantee
the convergence of our greedy algorithm.
Theorem 2 bounds the difference between the loss function evaluated at each iteration of the greedy
algorithm and the one at the globally optimal solution.
>
Theorem 2. Suppose in Eq. (6) the matrices Y:,:,m
Y:,:,m for m = 1, . . . , M are positive definite.
The solution of Algo. 1 at its kth iteration step satisfies the following inequality:
L(Wk ; Y, V) ? L(W ? ; Y, V) ?
?
(kYk2 kW(1)
k? )2
(k + 1)
,
(11)
where W ? is the global minimizer of the problem in Eq. (6) and kYk2 is the largest singular value
of a block diagonal matrix created by placing the matrices Y(:, :, m) on its diagonal blocks.
The detailed proof is given in Appendix B. The key idea of the proof is that the amount of decrease
in the loss function by each step in the selected mode is not smaller than the amount of decrease if we
had selected the first mode. The theorem shows that we can obtain the same rate of convergence for
learning low rank tensors as achieved in [22] for learning low rank matrices. The greedy algorithm
in Algorithm 1 is also connected to mixture regularization in [23]: the mixture approach decomposes
the solution into a set of low rank structures while the greedy algorithm successively learns a set of
rank one components.
Greedy Algorithm with Orthogonal Projections It is well-known that the forward greedy algorithm may make steps in sub-optimal directions because of noise. A common solution to alleviate the
effect of noise is to make orthogonal projections after each greedy step [2, 21]. Thus, we enhance the
forward greedy algorithm by projecting the solution into the space spanned by the singular vectors
of its mode-1 unfolding. The greedy algorithm with orthogonal projections performs an extra step in
line 13 of Algorithm 1: It finds the top k singular vectors of the solution: [U, S, V ] ? svd(W(1) , k)
where k is the iteration number. Then it finds the best solution in the space spanned by U and V by
solving Sb ? minS L(U SV > , Y, V) which has a closed form solution. Finally, it reconstructs the
b > , 1). Note that the projection only needs to find top k singular vectors
solution: W ? refold(U SV
which can be computed efficiently for small values of k.
5
1
0.9
0.8
0.7
0.6
Forward
Orthogonal
ADMM
Trace
1000
15
Run Time (Sec)
1.1
20
Mixture Rank Complexity
Parameter Estimation RMSE
1200
Forward
Orthogonal
ADMM
Trace
MTL?L1
MTL?L21
MTL?Dirty
1.2
10
5
0
50
100
150
# of Samples
(a) RMSE
200
250
?5
0
600
400
200
0.5
0.4
0
800
Forward Greedy
Orthogonal Greedy
ADMM
50
100
150
# of Samples
(b) Rank
200
0 1
10
2
# of Variables
10
(c) Scalability
Figure 1: Tensor estimation performance comparison on the synthetic dataset over 10 random runs.
(a) parameter Estimation RMSE with training time series length, (b) Mixture Rank Complexity with
training time series length, (c) running time for one single round with respect to number of variables.
4
Experiments
We evaluate the efficacy of our algorithms on synthetic datasets and real-world application datasets.
4.1
Low rank tensor learning on synthetic data
For empirical evaluation, we compare our method with multitask learning (MTL) algorithms, which
also utilize the commonalities between different prediction tasks for better performance. We use the
following baselines: (1) Trace norm regularized MTL (Trace), which seeks the low rank structure
only on the task dimension; (2) Multilinear MTL [20], which adapts the convex relaxation of low
rank tensor learning solved with Alternating Direction Methods of Multiplier (ADMM) [10] and
Tucker decomposition to describe the low rankness in multiple dimensions; (3) MTL-L1 , MTL-L21
[19], and MTL-LDirty [14], which investigate joint sparsity of the tasks with Lp norm regularization.
For MTL-L1 , MTL-L21 [19] and MTL-LDirty , we use MALSAR Version 1.1 [27].
We construct a model coefficient tensor W of size 20 ? 20 ? 10 with CP rank equals to 1.
Then, we generate the observations Y and V according to multivariate regression model V:,:,m =
W:,:,m Y:,:,m + E:,:,m for m = 1, . . . , M , where E is tensor with zero mean Gaussian noise elements.
We split the synthesized data into training and testing time series and vary the length of the training
time series from 10 to 200. For each training length setting, we repeat the experiments for 10 times
and select the model parameters via 5-fold cross validation. We measure the prediction performance
via two criteria: parameter estimation accuracy and rank complexity. For accuracy, we calculate the
RMSE of the estimation versus the true model coefficient tensor. For rank complexity, we calculate
PN
the mixture rank complexity [23] as M RC = n1 n=1 rank(W(n) ).
The results are shown in Figure 1(a) and 1(b). We omit the Tucker decomposition as the results are
not comparable. We can clearly see that the proposed greedy algorithm with orthogonal projections
achieves the most accurate tensor estimation. In terms of rank complexity, we make two observations: (i) Given that the tensor CP rank is 1, greedy algorithm with orthogonal projections produces
the estimate with the lowest rank complexity. This can be attributed to the fact that the orthogonal
projections eliminate the redundant rank-1 components that fall in the same spanned space. (ii) The
rank complexity of the forward greedy algorithm increases as we enlarge the sample size. We believe that when there is a limited number of observations, most of the new rank-1 elements added
to the estimate are not accurate and the cross-validation steps prevent them from being added to the
model. However, as the sample size grows, the rank-1 estimates become more accurate and they are
preserved during the cross-validation.
To showcase the scalability of our algorithm, we vary the number of variables and generate a series
of tensor W ? R20?20?M for M from 10 to 100 and record the running time (in seconds) for three
tensor learning algorithms, i.e, forward greedy, greedy with orthogonal projections and ADMM. We
measure the run time on a machine with a 6-core 12-thread Intel Xenon 2.67GHz processor and
12GB memory. The results are shown in Figure 1(c). The running time of ADMM increase rapidly
with the data size while the greedy algorithm stays steady, which confirms the speedup advantage
of the greedy algorithm.
6
Table 1: Cokriging RMSE of 6 methods averaged over 10 runs. In each run, 10% of the locations
are assumed missing.
DATASET
USHCN
CCDS
Y ELP
F OURSQUARE
4.2
ADMM
0.8051
0.8292
0.7730
0.1373
F ORWARD
0.7594
0.5555
0.6993
0.1338
O RTHOGONAL
0.7210
0.4532
0.6958
0.1334
S IMPLE
0.8760
0.7634
NA
NA
O RDINARY
0.7803
0.7312
NA
NA
MTGP
1.0007
1.0296
NA
NA
Spatio-temporal analysis on real world data
We conduct cokriging and forecasting experiments on four real-world datasets:
USHCN The U.S. Historical Climatology Network Monthly (USHCN)3 dataset consists of
monthly climatological data of 108 stations spanning from year 1915 to 2000. It has three climate variables: (1) daily maximum, (2) minimum temperature averaged over month, and (3) total
monthly precipitation.
CCDS The Comprehensive Climate Dataset (CCDS)4 is a collection of climate records of North
America from [18]. The dataset was collected and pre-processed by five federal agencies. It contains
monthly observations of 17 variables such as Carbon dioxide and temperature spanning from 1990 to
2001. The observations were interpolated on a 2.5 ? 2.5 degree grid, with 125 observation locations.
Yelp The Yelp dataset5 contains the user rating records for 22 categories of businesses on Yelp
over ten years. The processed dataset includes the rating values (1-5) binned into 500 time intervals
and the corresponding social graph for 137 active users. The dataset is used for the spatio-temporal
recommendation task to predict the missing user ratings across all business categories.
Foursquare The Foursquare dataset [17] contains the users? check-in records in Pittsburgh area
from Feb 24 to May 23, 2012, categorized by different venue types such as Art & Entertainment,
College & University, and Food. The dataset records the number of check-ins by 121 users in each
of the 15 category of venues over 1200 time intervals, as well as their friendship network.
4.2.1
Cokriging
We compare the cokriging performance of our proposed method with the classical cokriging approaches including simple kriging and ordinary cokriging with nonbias condition [13] which are
applied to each variables separately. We further compare with multitask Gaussian process (MTGP)
[4] which also considers the correlation among variables. We also adapt ADMM for solving the
nuclear norm relaxed formulation of the cokriging formulation as a baseline (see Appendix C for
more details). For USHCN and CCDS, we construct a Laplacian matrix by calculating the pairwise
Haversine distance of locations. For Foursquare and Yelp, we construct the graph Laplacian from
the user friendship network.
For each dataset, we first normalize it by removing the trend and diving by the standard deviation.
Then we randomly pick 10% of locations (or users for Foursquare) and eliminate the measurements
of all variables over the whole time span. Then, we produce the estimates for all variables of each
timestamp. We repeat the procedure for 10 times and report the average prediction RMSE for all
timestamps and 10 random sets of missing locations. We use the MATLAB Kriging Toolbox6 for
the classical cokriging algorithms and the MTGP code provided by [4].
Table 1 shows the results for the cokriging task. The greedy algorithm with orthogonal projections is
significantly more accurate in all three datasets. The baseline cokriging methods can only handle the
two dimensional longitude and latitude information, thus are not applicable to the Foursquare and
Yelp dataset with additional friendship information. The superior performance of the greedy algorithm can be attributed to two of its properties: (1) It can obtain low rank models and achieve global
consistency; (2) It usually has lower estimation bias compared to nuclear norm relaxed methods.
3
http://www.ncdc.noaa.gov/oa/climate/research/ushcn
http://www-bcf.usc.edu/?liu32/data/NA-1990-2002-Monthly.csv
5
http://www.yelp.com/dataset_challenge
6
http://globec.whoi.edu/software/kriging/V3/english.html
4
7
Table 2: Forecasting RMSE for VAR process with 3 lags, trained with 90% of the time series.
DATASET
USHCN
CCDS
FSQ
T UCKER
0.8975
0.9438
0.1492
ADMM F ORWARD
0.9227 0.9171
0.8448 0.8810
0.1407 0.1241
O RTHO O RTHO NL
0.9069 0.9175
0.8325 0.8555
0.1223 0.1234
T RACE
0.9273
0.8632
0.1245
MTLl1
0.9528
0.9105
0.1495
MTLl21 MTLdirty
0.9543 0.9735
0.9171 1.0950
0.1495 0.1504
Table 3: Running time (in seconds) for cokriging and forecasting.
DATASET
ORTHO
ADMM
4.2.2
USHCN
93.03
791.25
C OKRIGING
CCDS
YELP
16.98
78.47
320.77 2928.37
FSQ
91.51
720.40
F ORECASTING
USHCN CCDS
FSQ
75.47
21.38
37.70
235.73
45.62
33.83
Forecasting
We present the empirical evaluation on the forecasting task by comparing with multitask regression
algorithms. We split the data along the temporal dimension into 90% training set and 10% testing
set. We choose VAR(3) model and during the training phase, we use 5-fold cross-validation.
As shown in Table 2, the greedy algorithm with orthogonal projections again achieves the best prediction accuracy. Different from the cokriging task, forecasting does not necessarily need the correlations of locations for prediction. One might raise the question as to whether the Laplacian regularizer helps. Therefore, we report the results for our formulation without Laplacian (ORTHONL)
for comparison. For efficiency, we report the running time (in seconds) in Table 3 for both tasks of
cokriging and forecasting. Compared with ADMM, which is a competitive baseline also capturing
the commonalities among variables, space, and time, our greedy algorithm is much faster for most
datasets.
As a qualitative study, we plot the map of most predictive regions analyzed by the greedy algorithm using CCDS dataset in Fig. 2. Based on the concept
of how informative the past values of the climate
measurements in a specific location are in predicting future values of other time series, we define the
aggregate strength of predictiveness of each region
PP PM
as w(t) = p=1 m=1 |Wp,t,m |. We can see that
two regions are identified as the most predictive regions: (1) The southwest region, which reflects the
impact of the Pacific ocean and (2) The southeast region, which frequently experiences relative sea level
rise, hurricanes, and storm surge in Gulf of Mexico.
Another interesting region lies in the center of Colorado, where the Rocky mountain valleys act as a
funnel for the winds from the west, providing locally
divergent wind patterns.
5
Figure 2: Map of most predictive regions
analyzed by the greedy algorithm using 17
variables of the CCDS dataset. Red color
means high predictiveness whereas blue denotes low predictiveness.
Conclusion
In this paper, we study the problem of multivariate spatio-temporal data analysis with an emphasis
on two tasks: cokriging and forecasting. We formulate the problem into a general low rank tensor
learning framework which captures both the global consistency and the local consistency principle.
We develop a fast and accurate greedy solver with theoretical guarantees for its convergence. We
validate the correctness and efficiency of our proposed method on both the synthetic dataset and realapplication datasets. For future work, we are interested in investigating different forms of shared
structure and extending the framework to capture non-linear correlations in the data.
Acknowledgment
We thank the anonymous reviewers for their helpful feedback and comments. The research was
sponsored by the NSF research grants IIS-1134990, IIS- 1254206 and Okawa Foundation Research
Award. The views and conclusions are those of the authors and should not be interpreted as representing the official policies of the funding agency, or the U.S. Government.
8
References
[1] T. Anderson. Estimating linear restrictions on regression coefficients for multivariate normal
distributions. The Annals of Mathematical Statistics, pages 327?351, 1951.
[2] A. Barron, A. Cohen, W. Dahmen, and R. DeVore. Approximation and learning by greedy
algorithms. The Annals of Statistics, 2008.
[3] D. Bertsekas and J. Tsitsiklis. Parallel and Distributed Computation: Numerical Methods.
Prentice Hall Inc, 1989.
[4] E. Bonilla, K. Chai, and C. Williams. Multi-task Gaussian Process Prediction. In NIPS, 2007.
[5] W. Chu, V. Sindhwani, Z. Ghahramani, and S. Keerthi. Relational learning with Gaussian
processes. In NIPS, 2006.
[6] N. Cressie and H. Huang. Classes of nonseparable, spatio-temporal stationary covariance
functions. JASA, 1999.
[7] N. Cressie and G. Johannesson. Fixed rank kriging for very large spatial data sets. JRSS B
(Statistical Methodology), 70(1):209?226, 2008.
[8] N. Cressie, T. Shi, and E. Kang. Fixed rank filtering for spatio-temporal data. J. Comp. Graph.
Stat., 2010.
[9] N. Cressie and C. Wikle. Statistics for spatio-temporal data. John Wiley & Sons, 2011.
[10] D. Gabay and B. Mercier. A dual algorithm for the solution of nonlinear variational problems
via finite element approximation. Computers & Mathematics with Applications, 2(1):17?40,
1976.
[11] S. Gandy, B. Recht, and I. Yamada. Tensor completion and low-n-rank tensor recovery via
convex optimization. Inverse Problems, 2011.
[12] M. Gu, A. Ruhe, G. Sleijpen, H. van der Vorst, Z. Bai, and R. Li. 5. Generalized Hermitian
Eigenvalue Problems. Society for Industrial and Applied Mathematics, 2000.
[13] E. Isaaks and R. Srivastava. Applied geostatistics. London: Oxford University, 2011.
[14] A. Jalali, S. Sanghavi, C. Ruan, and P. Ravikumar. A dirty model for multi-task learning. In
NIPS, 2010.
[15] T. Kolda and B. Bader. Tensor decompositions and applications. SIAM review, 2009.
[16] W.-J. Li and D.-Y. Yeung. Relation regularized matrix factorization. In IJCAI, 2009.
[17] X. Long, L. Jin, and J. Joshi. Exploring trajectory-driven local geographic topics in foursquare.
In UbiComp, 2012.
[18] A. Lozano, H. Li, A. Niculescu-Mizil, Y. Liu, C. Perlich, J. Hosking, and N. Abe. Spatialtemporal causal modeling for climate change attribution. In KDD, 2009.
[19] F. Nie, H. Huang, X. Cai, and C. H. Ding. Efficient and robust feature selection via joint
`2,1 -norms minimization. In NIPS, 2010.
[20] B. Romera-Paredes, H. Aung, N. Bianchi-Berthouze, and M. Pontil. Multilinear multitask
learning. In ICML, 2013.
[21] S. Shalev-Shwartz, A. Gonen, and O. Shamir. Large-scale convex minimization with a lowrank constraint. In ICML, 2011.
[22] S. Shalev-Shwartz, N. Srebro, and T. Zhang. Trading Accuracy for Sparsity in Optimization
Problems with Sparsity Constraints. SIAM Journal on Optimization, 2010.
[23] R. Tomioka, K. Hayashi, and H. Kashima. Convex Tensor Decomposition via Structured
Schatten Norm Regularization. NIPS, 2013.
[24] T. Zhang. Adaptive Forward-Backward Greedy Algorithm for Learning Sparse Representations. IEEE Trans Inf Theory, pages 4689?4708, 2011.
[25] D. Zhou, O. Bousquet, T. Lal, J. Weston, and B. Sch?olkopf. Learning with local and global
consistency. In NIPS, 2003.
[26] H. Zhou, L. Li, and H. Zhu. Tensor regression with applications in neuroimaging data analysis.
JASA, 2013.
[27] J. Zhou, J. Chen, and J. Ye. MALSAR: Multi-tAsk Learning via StructurAl Regularization.
http://www.public.asu.edu/?jye02/Software/MALSAR/, 2011.
9
| 5429 |@word multitask:6 version:1 norm:8 paredes:1 confirms:1 simulation:1 seek:2 bn:3 covariance:8 decomposition:8 pick:1 concise:1 tr:2 bai:1 liu:2 series:10 efficacy:1 contains:3 romera:1 past:1 existing:4 current:1 com:1 comparing:1 chu:1 written:1 john:1 timestamps:1 numerical:1 informative:1 kdd:1 enables:1 designed:1 treating:1 plot:1 sponsored:1 stationary:1 greedy:41 selected:2 asu:1 core:1 yamada:1 record:5 provides:1 location:22 preference:1 lx:1 simpler:1 zhang:2 five:1 rc:1 along:3 constructed:1 mathematical:1 become:1 qualitative:1 consists:1 hermitian:1 pairwise:2 inter:4 frequently:1 surge:1 multi:5 nonseparable:1 globally:2 automatically:1 unfolded:1 food:1 gov:1 solver:1 precipitation:2 provided:1 xx:3 notation:1 estimating:1 lowest:1 mountain:1 argmin:7 interpreted:1 eigenvector:3 unified:4 finding:2 guarantee:3 temporal:23 act:1 tackle:1 unit:1 grant:1 omit:1 bertsekas:1 before:1 positive:2 engineering:1 local:10 yelp:7 analyzing:1 oxford:1 might:1 emphasis:1 ease:1 limited:1 factorization:1 averaged:2 unique:1 acknowledgment:1 enforces:1 testing:2 practice:1 block:2 definite:1 vorst:1 procedure:1 pontil:1 unfold:1 area:1 empirical:2 yan:1 significantly:2 convenient:1 projection:12 pre:1 induce:1 regular:1 ncdc:1 close:2 selection:2 operator:2 valley:1 prentice:1 context:1 applying:1 www:4 restriction:1 map:2 reviewer:1 missing:3 center:1 shi:1 williams:1 attribution:1 convex:6 formulate:2 recovery:2 nuclear:3 spanned:3 dahmen:1 handle:2 notion:1 ortho:1 annals:2 kolda:1 suppose:1 colorado:1 user:7 shamir:1 cressie:4 pa:1 element:5 trend:1 expensive:4 approximated:1 showcase:1 observed:1 ding:1 electrical:1 capture:9 solved:1 calculate:2 region:8 connected:1 decrease:3 malsar:3 kriging:6 rose:1 rq:1 agency:2 complexity:11 nie:1 imple:1 motivate:1 trained:1 solving:4 rewrite:1 algo:1 raise:1 predictive:3 upon:1 efficiency:2 gu:1 easily:2 joint:2 various:2 chapter:1 america:1 regularizer:4 univ:3 fast:5 describe:4 effective:1 london:1 kp:1 ubicomp:1 cokriging:28 aggregate:1 outcome:1 shalev:2 whose:3 lag:3 widely:1 solve:5 relax:1 otherwise:1 favor:1 statistic:3 ip:2 confronted:1 advantage:2 eigenvalue:5 cai:1 propose:3 rapidly:1 achieve:2 adapts:1 frobenius:1 kv:2 normalize:1 ky:1 scalability:2 los:3 olkopf:1 chai:1 convergence:4 cluster:2 enhancement:1 plethora:1 sea:1 produce:2 extending:1 ijcai:1 help:1 develop:3 completion:2 stat:1 geology:2 lowrank:1 eq:7 strong:1 longitude:2 c:1 predicted:1 trading:1 direction:4 bader:1 kb:1 public:1 adjacency:1 require:1 southwest:1 government:1 alleviate:1 anonymous:1 multilinear:2 exploring:1 hold:1 predictiveness:3 hall:1 normal:1 predict:2 achieves:4 commonality:5 vary:2 estimation:14 applicable:1 southeast:1 largest:2 correctness:1 reflects:1 unfolding:5 federal:1 minimization:2 clearly:1 gaussian:9 aim:1 pn:1 zhou:3 rank:80 check:2 industrial:1 perlich:1 greedily:1 baseline:4 vk2:1 helpful:1 stopping:1 niculescu:1 sb:1 integrated:1 eliminate:2 gandy:1 relation:1 transformed:1 interested:1 among:9 html:1 dual:1 spatial:14 art:2 constrained:2 initialize:1 timestamp:1 equal:2 construct:5 ruan:1 enlarge:1 represents:1 kw:5 yu:1 k2f:3 placing:1 icml:2 future:3 report:3 sanghavi:1 few:2 randomly:1 simultaneously:1 comprehensive:1 hurricane:1 usc:4 argmax:1 phase:1 keerthi:1 n1:1 investigate:1 evaluation:2 deferred:1 mixture:5 analyzed:2 nl:1 xb:1 accurate:6 cointegration:1 daily:1 xy:1 experience:1 orthogonal:14 conduct:2 causal:1 theoretical:1 sociology:2 modeling:2 csv:1 ordinary:1 geo:1 deviation:1 subset:1 entry:1 rare:1 conducted:1 dependency:4 answer:1 spatiotemporal:1 sv:2 synthetic:5 hosking:1 recht:1 geographical:1 venue:2 siam:2 stay:1 enhance:1 na:7 again:1 central:1 successively:2 opposed:1 choose:2 reconstructs:1 huang:2 mtgp:5 li:4 account:1 sec:1 wk:1 north:1 coefficient:5 includes:1 inc:1 bonilla:1 race:1 later:1 view:1 wind:2 closed:1 red:1 competitive:1 parallel:1 rmse:7 contribution:1 accuracy:5 variance:1 efficiently:5 bayesian:2 trajectory:1 comp:1 processor:1 l21:3 manual:1 definition:2 pp:1 involved:1 tucker:4 storm:1 ruhe:1 proof:3 di:1 attributed:2 ushcn:8 dataset:16 popular:1 knowledge:1 color:1 dimensionality:1 variogram:2 noaa:1 back:1 higher:1 supervised:1 follow:1 mtl:12 devore:1 methodology:1 formulation:6 evaluated:1 though:1 anderson:1 furthermore:1 correlation:15 until:1 nonlinear:1 elp:1 defines:1 mode:21 aung:1 scientific:1 believe:1 grows:1 effect:1 ye:1 concept:1 multiplier:1 true:1 geographic:1 former:1 regularization:6 lozano:1 alternating:2 wp:1 climate:7 round:1 during:2 kyk2:2 climatological:1 steady:1 criterion:2 generalized:4 presenting:1 complete:2 mohammad:1 demonstrate:2 climatology:3 performs:1 cp:3 temperature:4 l1:3 variational:1 funding:1 common:2 superior:1 qp:1 cohen:1 slight:1 synthesized:1 measurement:5 monthly:5 ai:2 uv:1 consistency:15 pm:2 grid:1 mathematics:2 had:1 moving:1 similarity:3 add:1 feb:1 dominant:3 multivariate:14 recent:2 diving:1 apart:1 reverse:1 scenario:1 driven:1 certain:1 inf:1 inequality:1 success:1 der:1 minimum:1 additional:2 relaxed:2 impose:2 v3:1 redundant:1 semi:1 ii:4 multiple:3 full:4 infer:1 reduces:1 smooth:1 faster:3 adapt:1 cross:8 long:1 dept:3 ravikumar:1 award:1 a1:1 laplacian:11 qi:1 prediction:7 scalable:3 regression:6 basic:1 impact:1 yeung:1 iteration:4 represent:1 kernel:1 achieved:2 preserved:1 whereas:1 separately:2 interval:2 singular:6 sch:1 extra:1 comment:1 joshi:1 structural:1 iii:1 enough:1 split:2 identified:2 regarding:1 idea:3 okawa:1 tradeoff:2 angeles:3 thread:1 whether:1 gb:1 forecasting:20 akin:1 reformulated:1 gulf:1 matlab:1 detailed:1 amount:2 locally:2 ten:1 processed:2 category:3 reduced:1 generate:2 http:5 exist:1 nsf:1 blue:1 arima:1 key:2 four:1 prevent:1 utilize:2 backward:1 graph:3 relaxation:1 geostatistical:1 year:2 enforced:1 run:5 inverse:1 utilizes:1 appendix:4 comparable:1 capturing:1 bound:2 fsq:3 fold:3 quadratic:1 binned:1 strength:1 constraint:7 constrain:1 software:2 bousquet:1 interpolated:1 argument:1 min:1 span:1 separable:2 speedup:1 pacific:1 structured:1 according:1 jr:1 across:2 smaller:1 son:1 lp:1 making:2 projecting:1 computationally:6 dioxide:1 hh:2 tractable:1 end:2 serf:1 mercier:1 adopted:1 available:2 operation:1 observe:1 hierarchical:1 barron:1 enforce:1 appropriate:1 bahadori:1 ocean:1 kashima:1 alternative:1 rp:3 original:1 assumes:2 clustering:2 include:1 denotes:2 cf:1 top:2 dirty:2 running:5 dataset5:1 unifying:1 entertainment:1 calculating:1 ghahramani:1 build:1 classical:3 society:1 tensor:55 objective:2 question:2 added:3 parametric:1 dependence:1 diagonal:4 jalali:1 southern:3 kth:1 distance:1 thank:1 schatten:1 concatenation:1 oa:1 topic:1 extent:1 considers:2 collected:2 spanning:2 length:4 code:1 index:2 providing:1 mexico:1 neuroimaging:2 carbon:1 trace:4 rise:1 policy:1 unknown:2 perform:2 bianchi:1 observation:9 datasets:10 finite:1 jin:1 optional:1 relational:3 precise:1 station:1 ccds:9 abe:1 rating:3 specified:1 lal:1 california:3 kang:1 geostatistics:2 nip:6 address:2 able:1 trans:1 usually:2 pattern:1 latitude:2 gonen:1 sparsity:3 challenge:2 including:1 memory:1 belief:1 critical:2 natural:2 business:2 regularized:2 taha:1 predicting:1 mizil:1 representing:1 zhu:1 rocky:1 created:1 prior:2 literature:3 yanliu:1 review:1 kf:3 relative:1 loss:5 kakf:1 generation:1 interesting:1 filtering:1 srebro:1 var:5 versus:1 validation:4 funnel:1 foundation:1 degree:1 jasa:2 principle:6 berthouze:1 row:1 summary:1 repeat:3 english:1 infeasible:1 tsitsiklis:1 side:4 bias:1 fall:1 taking:1 sparse:1 benefit:1 ghz:1 feedback:1 dimension:7 distributed:1 world:3 van:1 autoregressive:2 author:2 forward:9 collection:1 adaptive:1 simplified:1 historical:3 social:3 refold:4 r20:1 global:10 active:1 investigating:1 b1:2 pittsburgh:1 assumed:2 spatio:19 shwartz:2 foursquare:6 search:1 latent:2 decomposes:1 table:6 validate:1 learn:1 robust:1 ca:3 wikle:1 excellent:1 complex:2 interpolating:1 necessarily:1 domain:1 official:1 main:1 big:1 noise:4 whole:1 gabay:1 categorized:1 fig:1 intel:1 west:1 wiley:1 tomioka:1 sub:1 explicit:1 concatenating:1 col:2 lie:1 stamp:2 lw:1 learns:1 cyclically:1 theorem:4 removing:1 friendship:4 xt:4 specific:1 explored:2 supposition:1 divergent:1 jye02:1 essential:1 intrinsic:1 adding:1 johannesson:1 kx:3 rankness:2 spatialtemporal:2 chen:1 led:1 univariate:1 likely:2 orward:2 conveniently:1 recommendation:1 sindhwani:1 hayashi:1 corresponds:1 minimizer:1 satisfies:2 weston:1 goal:1 presentation:1 formulated:1 month:1 shared:5 admm:11 change:2 included:1 principal:1 lemma:3 total:1 svd:1 formally:1 select:1 college:1 cholesky:1 incorporate:3 evaluate:1 srivastava:1 |
4,893 | 543 | Recognition of Manipulated Objects
by Motor Learning
Hiroaki Gomi
Mitsuo Kawato
ATR Auditory and Visual Perception Research Laboratories,
Inui-dani, Sanpei-dani, Seika-cho, Soraku-gun, Kyoto 619-02, Japan
Abstract
We present two neural network controller learning schemes based on feedbackerror-learning and modular architecture for recognition and control of multiple
manipulated objects. In the first scheme, a Gating Network is trained to acquire
object-specific representations for recognition of a number of objects (or sets of
objects). In the second scheme, an Estimation Network is trained to acquire
function-specific, rather than object-specific, representations which directly estimate
physical parameters. Both recognition networks are trained to identify manipulated
objects using somatic and/or visual information. After learning, appropriate
motor commands for manipulation of each object are issued by the control
networks.
1 INTRODUCTION
Conventional feedforward neural-network controllers (Barto et aI., 1983; Psaltis et al.,
1987; Kawato et aI., 1987, 1990; Jordan, 1988; Katayama & Kawato, 1991) can not cope
with multiple or changeable manipulated objects or disturbances because they cannot
change immediately the control law corresponding to the object. In interaction with
manipulated objects or, in more general terms, in interaction with an environment which
contains unpredictable factor, feedback information is essential for control and object
recognition. From these considerations, Gomi & Kawato (1990) have examined the
adaptive feedback controller learning schemes using feedback-error-Iearning, from which
impedance control (Hogan, 1985) can be obtained automatically. However, in that scheme,
some higher system needs to supervise the setting of the appropriate mechanical impedance
for each manipulated object or environment.
In this paper, we introduce semi-feedforward control schemes using neural networks which
receive feedback and/or feedforward information for recognition of multiple manipulated
objects based on feedback-error-learning and modular network architecture. These schemes
have two advantages over previous ones as follows. (1) Learning is achieved without the
547
548
Gomi and Kawato
exact target motor command vector, which is unavailable during supervised motor learning.
(2) Although somatic information alone was found to be sufficient to recognize objects,
object identification is predictive and more reliable when both somatic and visual information
are used.
2 RECOGNITION OF MANIPULATED OBJECTS
The most important issues in object manipulation are (l) how to recognize the manipulated
object and (2) how to achieve uniform performance for different objects. There are several
ways to acquire helpful information for recognizing manipulated objects. Visual information
and somatic information (performance by motion) are most informative for object recognition
for manipulation.
The physical characteristics useful for object manipulation such as mass, softness and
slipperiness, can not be predicted without the experience of manipulating similar objects.
In this respect, object recognition for manipulation should be learned through object
manipulation.
3 MODULAR ARCHITECTURE USING GATING NETWORK
Jacobs et al. (1990, 1991) and Nowlan & Hinton (1990, 1991) have proposed a competitive
modular network architecture which is applied to the task decomposition problem or
classification problems. Jacobs (1991) applied this network architecture to the multi-payload
robotics task in which each expert network controller is trained for each category of
manipulated objects in terms of the object's mass. In his scheme, the payload's identity is
fed to the gating network to select a suitable expert network which acts as a feedforward
controller.
We examined modular network architecture using feedback-e"or-learning for simultaneous
learning of object recognition and control task as shown in Fig.l.
M1Tt rU."nbmmOOi--._,",So
v
"'::" ,:f,
t:~~ ~~t~ t:~
Gatmg
Network ....+----.
Expert Network 1
Expert Network 2
t--=-.c::u-~
Expert Network 3
+
u~
u
Controlled
t -.......eo{4~"""""-"l~ object
...-_..
1--. . . .
Fig.1 Configuration of the modular architecture using Gating Network
for object manipulation based on feedback-error-learning
In this learning scheme, the quasi-target vector for combined output of expert networks is
employed instead of the exact target vector. This is because it is unlikely that the exact
target motor command vector can be provided in learning. The quasi-target vector of
feedforward motor command, u' is produced by :
(1)
U '- U + Ufo '
Recognition of Manipulated Objects by Motor Learning
Here, U denotes the previous final motor command and ufo denotes the feedback motor
command. Using this quasi-target vector, the gating and expert networks are trained to
maximize the log-likelihood function, In L, by using backpropagation.
In L
=In i
gje -IU'-u,r /2(1,2
(2)
j=!
Here, uj is the i th expert network output, (Ij is a variance scaling parameter of the i th
expert network and gj' the i th output of gating network, is calculated by
gj
e
S,
(3)
= - 1 1- ,
Le
sJ
j=!
where Sj denotes the weighted input received by the i th output unit. The total output of
the modular network is
11
uff
=~gjUj'
j=l
(4)
By maximizing Eq.2 using steepest ascent method, the gating network learns to choose
the expert network whose output is closest to the quasi-target command, and each expert
network is tuned correctly when it is chosen by the gating network. The desired trajectory
is fed to the expert networks so as to make them work as feedforward controllers.
4 SIMULATION OF OBJECT MANIPULATION
BY MODULAR ARCHITECTURE WITH GATING NETWORK
We show the advantage of the learning schemes presented above by simulation results
below. The configuration of the controlled object and manipulated object is shown in
Fig.2 in which M, B, K respectively denote the mass, viscosity and stiffness of the
coupled object (controlled- and manipulated-object). The manipulated object is changed
every epoch (l [sec]) while the coupled object is controlled to track the desired trajectory.
Fig.3 shows the selected object, the feedforward and feedback motor commands, and the
desired and actual trajectories before learning.
----------
l~__
~
a
x
-4--j
M
t:~
.24,------~1------r-----"--~~1
o
Fig.2 Configuration of the controlled
object and the manipulated object
5
20
time [??c]
Fig.3 Temporal patterns of the selected
object, the motor commands, the desired
and actual trajectories before learning
The desired trajectory, x d ' was produced by Ornstein-Uhlenbeck random process. As
shown in Fig.3, the error between the desired trajectory and the actual trajectory remained
because the feedback controller in which the gains were fixed, was employed in this
condition. (Physical characteristics of the objects used are listed in Fig.4a)
549
550
Gomi and Kawato
4.1 SOMATIC INFORMATION FOR GATING NETWORK
We call the actual trajectory vector, x, and the final motor command,
"somatic
infonnation". Somatic infonnation should be most useful for on-line (feedback) recognition
of the dynamical characteristics of manipulated objects. The latest four times data of
somatic information were used as the gating network inputs for identification of the
coupled object in this simulation. s ofEq.3 is expressed as:
s(t) = '1'1 (x(t), x(t -1), x(t - 2), x(t - 3), u(t), u(t -1), u(t - 2), u(t - 3?).
(5)
The dynamical characteristics of coupled objects are shown in Fig.4a. The object was
changed in every epoch (l [secD. The variance scaling parameter was (Jj = 0.8 and the
learning rates were 77g a,e =1. 0 x 10-3 and 77 expert i = 1. 0 x 10-5 ? The three-layered feedforward
neural network (input 16, hidden 30, output 3) was employed for the gating network and
the two-layered linear networks (input 3, output 1) were used for the expert networks.
Comparing the expert's weights after learning and the coupled object characteristics in
Fig.4a, we realize that expert networks No.1, No.2, No.3 obtained the inverse dynamics
of coupled objects y, (3, a, respectively. The time variation of object, the gating network
outputs, motor commands and trajectories after learning are shown in Fig.4b. The gating
network outputs for the objects responded correctly in the most of the time and the
feedback motor command, ufo' was almost zero. As a consequence of adaptation, the
actual trajectory almost perfectly corresponded with the desired trajectory.
b.
a.
-
'Y
Gating Net Outputs v.s. Objects
U ,
-- -- ---
relinal
chara::loristiicsl ,mago D-u::-;-""""::'::"::-;:;':-:=-r-"7.:""':...--i
M B K
a
1.0 2.0 8.0
none
f3
5.0 7.0 4.0
none
- 20 --\---'_ _---"'.-_ _----'-..L"--_ _
~~---.:,;
!1~actC:al
none
1~L_8.03.01
____.0 L:::Lll?[ili?~
....~.... =::O~
~d.::
L.
IS.
_
o
5
_
10
tlmo
15
20
[,.cl
Fig.4 Somatic information for gating network, a. Statistical analysis of the
correspondence of the expert networks with each object after learning (averaged
gating outputs), b. Temporal patterns of objects, gating outputs, motor commands
and trajectories after learning
4.2 VISUAL INFORMATION FOR GATING NETWORK
We usually assume the manipulated object's characteristics by using visual infonnation.
Visual information might be helpful for feedforward recognition. In this case, s of Eq.3 is
expressed as:
s(t) = 'l'2(V(t?) .
(6)
We used three visual cues corresponding to each coupled object in this simulation as
shown in Fig.5a. At each epoch in this simulation, one of three visual cues selected
randomly is randomly placed at one of four possible locations on a 4 x 4 retinal matrix.
Recognition of Manipulated Objects by Motor Learning
The visual cues of each object are different, but object ex and ex* have the same dynamical
characteristics as shown in Fig.5a. The gating network should identify the object and
select a suitable expert network for feedforward control by using this visual information.
The learning coefficients were O"j 0.7, 17gate 1. 0 X 10-3 , 17eXpert j = 1. 0 X 10-5 . The same
networks used in above experiment were used in this simulation.
=
=
After learning, the expert network No.2 acquired the inverse dynamics of object ex and ex * ,
and expert network No.3 accomplished this for object y. It is recognized from Fig.5b that
the gating network almost perfectly selected expert network No.2 for object ex and ex*,
and almost perfectly selected expert network No.3 for object y. Expert network No.1
which did not acquire inverse dynamics corresponding to any of the three objects, was not
selected in the test period after learning. The actual trajectory in the test period corresponded
almost perfectly to the desired trajectory.
b.
a.
Gating Net. Outputs V.s. Objects
------ - -
tlma [sac]
Fig. 5 Visual information for gating network, a. Statistical analysis of the
correspondence of the expert networks with each Object after learning (averaged
gating outputs), b. Temporal patterns of objects, gating outputs, motor commands
and trajectories after learning
4.3 SOMATIC & VISUAL INFORMATION FOR GATING NETWORK
We show here the simulation results by using both of somatic and visual information as
the gating network inputs. In this case, s ofEq.3 is represented as:
s(t)= 'l'3(x(t),?? ?,x(t-3),u(t),???,u(t-3),V(t)).
(7)
In this simulation, the object ex and ~* had different dynamical characteristics, but shared
same visual cue as listed in Fig.6a. Thus, to identify the coupled object one by one, it is
necessary for the gating network to utilize not only visual information but also somatic
information. The learning coefficients were O"j =1. 0, 17gale = 1. 0 X 10-3 and
17expert j = 1. 0 X 10-5 . The gating network had 32 input units, 50 hidden units, and 1 output
unit, and the expert networks were the same as in the above experiment.
After learning, expert networks No.1, No.2, No.3 acquired the inverse dynamics of
objects y, ~*, ex respectively. As shown in Fig.6b, the gating network identified the
object almost correctly.
551
552
Gomi and Kawato
--
b.
a.
-
- ---
Gating Net. Outputs v.s. Objects
Objac1
physical
charactonstics
M
B K
- 20~------~----~
____________
~
8
:2j~
1
LL__.J?~2Lill1iiliJill___? ? ?"l
-
actual
~
0
j
o
5
10
15
20
time [.ocJ
Fig. 6 Somatic & Visual information for gating network, a. Statistical analysis of the
correspondence of the expert networks with each object after learning (averaged gating
outputs), b. Temporal patterns of objects, gating outputs, motor commands and
trajectories after learning
4.4 UNKNOWN OBJECT RECOGNITION
BY USING SOMATIC INFORMATION
Fig.7b shows the responses for unknown objects whose physical characteristics were
slightly different from known objects (see Fig.7a and Fig.4a) in the case using somatic
information as the gating network inputs. Even if each tested object was not the same as
any of the known (learned) objects, the closest expert network was selected. (compare
Fig.4a and Fig.7a) During some period in the test phase, the feedback command increased
because of an inappropriate feedforward command.
-- --- - --- -----
b.
a.
Gating Net. Outputs v.s. Objects
object
physical
charactorisbCS
M
a'
rotinal II---..:-..---.-----..-:~_=_r____._;:_;;__t
Imago
B K
II--'--'----'--+---'--"---'---+--'---'--"-t
2.0 3.0 7.0
none
4.0 6 .0 5.0
none
20
__III
~
0
~
?2
....,..".~~,y 'i~~~k--~ ~
O~--....lIt,.------':'--i:..i,-_ _~--'-_~
9.0 2.0 2.0 none
tim. [secJ
Fig. 7 Unknown objects recognition by using Somatic information, a. Statistical
analysis of the correspondence of the expert networks with each object after learning
(averaged gating outputs), b. Temporal patterns of objects, gating outputs, motor
commands and trajectories after learning
Recognition of Manipulated Obj ects by Motor Learning
5 MODULAR ARCHITECTURE
USING ESTIMATION NETWORK
The previous modular architecture is competitive in the sense that expert networks
compete with each other to occupy its niche in the input space. We here propose a new
cooperative modular architecture where expert networks specified for different functions
cooperate to produce the required output. In this scheme, estimation networks are trained
to recognize physical characteristics of manipulated objects by using feedback information.
Using this method, an infinite number of manipulated objects in the limited domain can
be treated by using a small number of estimation networks. We applied this method to
recognizing the mass of the manipulated objects. (see Fig.8)
Fig.9a shows the output of the estimation network compared to actual masses. The
realized trajectory almost coincided with the desired trajectory as shown in Fig.9b. This
learning scheme can be applied not only to estimating mass but also to other physical
characteristics such as softness or slipperiness.
a.
~
8
6
~ 4
~
2
0'r--_"""T""'_""""''''-_-'--_--'
0.0
0.5
1.0
~me
b. _
~
2.0
I f\.~
desired traj9Ctory
2
1
actual traJecklry
.~ a -"'--__ ~ I
8.
1.5
(sec]
'..-1
-1
,1"\
\
\
-2
-3
o
j ,i.x
Fig. 8 Confaguration of the modular architecture using
mass estimation network for object manipulation by
feedback-error-Iearning
5
10
15
20
time (sec]
Fig. 9 a. Comparison of actual &
estimated mass, b. desired & actual
trajectory
6 DISCUSSION
In the first scheme, the internal models for object manipulation (in this case, inverse
dynamics) were represented not in terms of visual information but rather, of somatic
information (see 4.2). Although the current simulation is primitive, it indicates the very
important issue that functional internal-representations of objects (or environments),
rather than declarative ones, were acquired by motor learning.
The quasi-target motor command in the first scheme and the motor command error in the
second scheme are not always exactly correct in each time step because the proposed
learning schemes are based on the feedback-error-learning method. Thus, the learning rates
in the proposed schemes should be slower than those schemes in which exact target
commands are employed. In our preliminary simulation, it was about five times slower.
However, we emphasize that exact target motor commands are not available in supervised
motor learning.
The limited number of controlled objects which can be dealt with by the modular network
with a gating network is a considerable problem (Jacobs, 1991; Nowlan, 1990, 1991).
This problem depends on choosing an appropriate number of expert networks and value of
the variance scaling parameter, (J' . Once this is done, the expert networks can interpolate
553
554
Gomi and Kawato
the appropriate output for a number of unknown objects. Our second scheme provides a
more satisfactory solution to this problem.
On the other hand, one possible drawback of the second scheme is that it may be difficult
to estimate many physical parameters for complicated objects, even though the learning
scheme which directly estimates the physical parameters can handle any number of
objects.
We showed here basic examinations of two types of neural networks - a gating network
and a direct estimation network. Both networks use feedback and/or feedforward information
for recognition of multiple manipulated objects. In future. we will attempt to integrate
these two architectures in order to model tasks involving skilled motor coordination and
high level recognition.
Ack nowledgmen t
We would like to thank Drs. E. Yodogawa and K. Nakane of AlR Auditory and Visual
Perception Research Laboratories for their continuing encouragement. Supported by HFSP
Grant to M.K.
References
Barto, A.G., Sutton. R.S., Anderson, C.W. (1983) Neuronlike adaptive elements that can
solve difficult learning control problems; IEEE Trans. on Sys. Man and Cybern.
SMC-13, pp.834-846
Gomi, H., Kawato, M. (1990) Learning control for a closed loop system using feedbackerror-learning. Proc. the 29th IEEE Conference on Decision and Control, Hawaii.
Dec., pp.3289-3294
Hogan, N. (1985) Impedance control: An approach to manipulation: Part I - Theory, Part
II - Implementation, Part III - Applications, ASME Journal of Dynamic Systems,
Measurement, and Control, Vol. 107, pp.1-24
Jacobs, R.A., Jordan, M.I., Barto, A.G. (1990) Task decomposition through competition
in a modular connectionist architecture: The what and where vision tasks, COINS
Technical Report 90-27, pp.1-49
Jacobs, R.A., Jordan, M.I. (1991) A competitive modular connectionist architecture. In
Lippmann, R.P. et al., (Eds.) NIPS 3, pp.767-773
Jordan. M.I. (1988) Supervised learning and systems with excess degrees of freedom,
COINS Technical Report 88-27, pp.1-41
Kawato, M., Furukawa, K., Suzuki, R. (1987) A hierarchical neural-network model for
control and learning of voluntary movement; Bioi. Cybern. 57, pp.169-185
Kawato, M. (1990) Computational schemes and neural network models for formation and
control of multijoint arm trajectory. In: Miller, T., Sutton, R.S., Werbos, P.J.(Eds.)
Neural Networksfor Control, The MIT Press, Cambridge, Massachusetts, pp.197-228
Katayama, M., Kawato, M. (1991) Learning trajectory and force control of an artificial
muscle arm by parallel-hierarchical neural network model. In Lippmann, R.P. et al.,
(Eds.) NIPS 3, pp.436-442
Nowlan, S.J. (1990) Competing experts: An experimental investigation of associative
mixture models, Univ. Toronto Tech. Rep. CRG-TR-90-5, pp.I-77
Nowlan, S.1., Hinton, G.E. (1991) Evaluation of adaptive mixtures of competing experts.
In Lippmann, R.P. et al., (Eds.) NIPS 3, pp.774-780
Psaltis, D., Sideris, A., Yamamura, A. (1987) Neural controllers, Proc. IEEE Int. Con!
Neural Networks, Vol.4, pp.551-557
| 543 |@word simulation:10 jacob:5 decomposition:2 tr:1 configuration:3 contains:1 tuned:1 current:1 comparing:1 nowlan:4 realize:1 informative:1 motor:26 alone:1 cue:4 selected:7 sys:1 steepest:1 provides:1 toronto:1 location:1 five:1 skilled:1 direct:1 ect:1 introduce:1 acquired:3 seika:1 multi:1 automatically:1 actual:11 unpredictable:1 lll:1 inappropriate:1 provided:1 estimating:1 mass:8 what:1 temporal:5 every:2 act:1 iearning:2 softness:2 exactly:1 control:17 unit:4 grant:1 before:2 yodogawa:1 consequence:1 sutton:2 might:1 examined:2 limited:2 smc:1 averaged:4 backpropagation:1 cannot:1 layered:2 cybern:2 conventional:1 maximizing:1 latest:1 primitive:1 immediately:1 sac:1 his:1 handle:1 variation:1 tlma:1 target:10 exact:5 element:1 recognition:19 werbos:1 cooperative:1 movement:1 environment:3 dynamic:6 hogan:2 trained:6 predictive:1 represented:2 univ:1 artificial:1 corresponded:2 formation:1 choosing:1 whose:2 modular:15 solve:1 final:2 associative:1 advantage:2 net:4 propose:1 interaction:2 adaptation:1 loop:1 achieve:1 competition:1 produce:1 object:97 tim:1 ij:1 received:1 eq:2 predicted:1 payload:2 drawback:1 correct:1 preliminary:1 investigation:1 crg:1 estimation:7 proc:2 multijoint:1 psaltis:2 infonnation:3 coordination:1 weighted:1 dani:2 mit:1 always:1 rather:3 command:22 barto:3 likelihood:1 indicates:1 tech:1 sense:1 helpful:2 unlikely:1 hidden:2 manipulating:1 quasi:5 iu:1 issue:2 classification:1 once:1 f3:1 lit:1 future:1 connectionist:2 report:2 randomly:2 manipulated:24 recognize:3 interpolate:1 phase:1 attempt:1 freedom:1 mitsuo:1 neuronlike:1 evaluation:1 mixture:2 necessary:1 experience:1 continuing:1 desired:11 increased:1 alr:1 uniform:1 recognizing:2 yamamura:1 cho:1 combined:1 choose:1 gale:1 hawaii:1 expert:37 japan:1 gje:1 retinal:1 sec:3 coefficient:2 int:1 ornstein:1 depends:1 closed:1 competitive:3 complicated:1 parallel:1 ofeq:2 responded:1 variance:3 characteristic:11 ufo:3 miller:1 identify:3 dealt:1 identification:2 produced:2 none:6 trajectory:22 gomi:7 simultaneous:1 ed:4 pp:12 con:1 gain:1 auditory:2 massachusetts:1 higher:1 supervised:3 response:1 done:1 though:1 anderson:1 hand:1 laboratory:2 satisfactory:1 during:2 ack:1 asme:1 motion:1 cooperate:1 consideration:1 kawato:12 functional:1 physical:10 measurement:1 cambridge:1 ai:2 encouragement:1 had:2 gj:2 closest:2 showed:1 manipulation:11 inui:1 issued:1 sideris:1 rep:1 accomplished:1 muscle:1 furukawa:1 eo:1 employed:4 recognized:1 maximize:1 period:3 semi:1 ii:3 multiple:4 kyoto:1 technical:2 controlled:6 involving:1 basic:1 controller:8 vision:1 uhlenbeck:1 achieved:1 robotics:1 dec:1 receive:1 ascent:1 jordan:4 call:1 obj:1 feedforward:12 iii:2 architecture:15 perfectly:4 identified:1 competing:2 soraku:1 jj:1 useful:2 listed:2 viscosity:1 category:1 occupy:1 estimated:1 correctly:3 track:1 vol:2 four:2 utilize:1 compete:1 inverse:5 almost:7 decision:1 networksfor:1 scaling:3 uff:1 correspondence:4 hiroaki:1 slightly:1 supervise:1 fed:2 drs:1 available:1 stiffness:1 hierarchical:2 appropriate:4 coin:2 gate:1 slower:2 changeable:1 denotes:3 uj:1 realized:1 thank:1 atr:1 gun:1 me:1 declarative:1 ru:1 acquire:4 difficult:2 implementation:1 unknown:4 voluntary:1 hinton:2 somatic:17 mechanical:1 specified:1 required:1 learned:2 nip:3 trans:1 usually:1 perception:2 below:1 dynamical:4 pattern:5 ocj:1 reliable:1 suitable:2 treated:1 examination:1 disturbance:1 force:1 arm:2 scheme:22 coupled:8 katayama:2 epoch:3 law:1 integrate:1 degree:1 sufficient:1 changed:2 placed:1 supported:1 feedback:17 calculated:1 suzuki:1 adaptive:3 cope:1 sj:2 excess:1 emphasize:1 lippmann:3 impedance:3 unavailable:1 cl:1 domain:1 did:1 fig:30 learns:1 coincided:1 niche:1 remained:1 specific:3 gating:40 essential:1 imago:1 visual:19 expressed:2 bioi:1 identity:1 nakane:1 shared:1 man:1 considerable:1 change:1 infinite:1 total:1 hfsp:1 ili:1 experimental:1 select:2 internal:2 tested:1 ex:8 |
4,894 | 5,430 | Provable Non-convex Robust PCA
Praneeth Netrapalli 1? U N Niranjan2?
1
Sujay Sanghavi3
Animashree Anandkumar2
Prateek Jain4
Microsoft Research, Cambridge MA. 2 The University of California at Irvine.
3
The University of Texas at Austin. 4 Microsoft Research, India.
Abstract
We propose a new method for robust PCA ? the task of recovering a low-rank matrix from sparse corruptions that are of unknown value and support. Our method
involves alternating between projecting appropriate residuals onto the set of lowrank matrices, and the set of sparse matrices; each projection is non-convex but
easy to compute. In spite of this non-convexity, we establish exact recovery of the
low-rank matrix, under the same conditions that are required by existing methods
(which are based on convex optimization).
For an m?n input matrix (m ? n), our
method has a running time of O r2 mn per iteration, and needs O (log(1/)) iterations to reach an accuracy of . This is close to the running times of simple PCA
via the power method, which requires O (rmn) per iteration, and O (log(1/)) iterations. In contrast, the existing methods
for robust PCA, which are based on
convex optimization, have O m2 n complexity per iteration, and take O (1/)
iterations, i.e., exponentially more iterations for the same accuracy.
Experiments on both synthetic and real data establishes the improved speed and
accuracy of our method over existing convex implementations.
Keywords:
tions.
1
Robust PCA, matrix decomposition, non-convex methods, alternating projec-
Introduction
Principal component analysis (PCA) is a common procedure for preprocessing and denoising, where
a low rank approximation to the input matrix (such as the covariance matrix) is carried out. Although
PCA is simple to implement via eigen-decomposition, it is sensitive to the presence of outliers,
since it attempts to ?force fit? the outliers to the low rank approximation. To overcome this, the
notion of robust PCA is employed, where the goal is to remove sparse corruptions from an input
matrix and obtain a low rank approximation. Robust PCA has been employed in a wide range
of applications, including background modeling [LHGT04], 3d reconstruction [MZYM11], robust
topic modeling [Shi13], and community detection [CSX12], and so on.
Concretely, robust PCA refers to the following problem: given an input matrix M = L? + S ? , the
goal is to decompose it into sparse S ? and low rank L? matrices. The seminal works of [CSPW11,
CLMW11] showed that this problem can be provably solved via convex relaxation methods, under
some natural conditions on the low rank and sparse components. While the theory is elegant, in
practice, convex techniques are expensive to run on a large scale and have poor convergence rates.
Concretely, for decomposing an m?n matrix, say with m ? n, the best specialized
implementations
(typically first-order methods) have a per-iteration complexity of O m2 n , and require O(1/)
number of iterations to achieve an error of . In contrast, the usual PCA, which carries out a rankr approximation of the input matrix, has O(rmn) complexity per iteration ? drastically smaller
?
Part of the work done while interning at Microsoft Research, India
1
when r is much smaller than m, n. Moreover, PCA requires exponentially fewer iterations for
convergence: an accuracy is achieved with only O (log(1/)) iterations (assuming constant gap in
singular values).
In this paper, we design a non-convex algorithm which is ?best of both the worlds? and bridges the
gap between (the usual) PCA and convex methods for robust PCA. Our method has low computational complexity similar to PCA (i.e. scaling costs and convergence rates), and at the same time,
has provable global convergence guarantees, similar to the convex methods. Proving global convergence for non-convex methods is an exciting recent development in machine learning. Non-convex
alternating minimization techniques have recently shown success in many settings such as matrix
completion [Kes12, JNS13, Har13], phase retrieval [NJS13], dictionary learning [AAJ+ 13], tensor
decompositions for unsupervised learning [AGH+ 12], and so on. Our current work on the analysis
of non-convex methods for robust PCA is an important addition to this growing list.
1.1
Summary of Contributions
We propose a simple intuitive algorithm for robust PCA with low per-iteration cost and a fast convergence rate. We prove tight guarantees for recovery of sparse and low rank components, which
match those for the convex methods. In the process, we derive novel matrix perturbation bounds,
when subject to sparse perturbations. Our experiments reveal significant gains in terms of speed-ups
over the convex relaxation techniques, especially as we scale the size of the input matrices.
Our method consists of simple alternating (non-convex) projections onto low-rank and sparse matrices. For an m?n matrix, our method has a running time of O(r2 mn log(1/)), where r is the rank of
the low rank component. Thus, our method has a linear convergence rate, i.e. it requires O(log(1/))
iterations to achieve an error of , where r is the rank of the low rank component L? . When the rank
r is small, this nearly matches the complexity of PCA, (which is O(rmn log(1/))).
We prove recovery of the sparse and low rank components under a set of requirements which are
tight and match those for the convex techniques (up to constant factors). In particular, under the
deterministic sparsity model, where each row and each column
of the sparse matrix S ? has at most
2
? fraction of non-zeros, we require that ? = O 1/(? r) , where ? is the incoherence factor (see
Section 3).
In addition to strong theoretical guarantees, in practice, our method enjoys significant advantages over the state-of-art solver for (1), viz., the inexact augmented Lagrange multiplier (IALM)
method [CLMW11]. Our method outperforms IALM in all instances, as we vary the sparsity levels,
incoherence, and rank, in terms of running time to achieve a fixed level of accuracy. In addition,
on a real dataset involving the standard task of foreground-background separation [CLMW11], our
method is significantly faster and provides visually better separation.
Overview of our techniques: Our proof technique involves establishing error contraction with
each projection onto the sets of low rank and sparse matrices. We first describe the proof ideas when
L? is rank one. The first projection step is a hard thresholding procedure on the input matrix M to
remove large entries and then we perform rank-1 projection of the residual to obtain L(1) . Standard
matrix perturbation results (such as Davis-Kahan) provide `2 error bounds between the singular
vectors of L(1) and L? . However, these bounds do not suffice for establishing the correctness of our
method. Since the next step in our method involves hard thresholding of the residual M ? L(1) ,
we require element-wise error bounds on our low rank estimate. Inspired by the approach of Erd?os
et al. [EKYY13], where they obtain similar element-wise bounds for the eigenvectors of sparse
Erd?os?R?enyi graphs, we derive these bounds by exploiting the fixed point characterization of the
eigenvectors1 . A Taylor?s series expansion reveals that the perturbation between the estimated and
the true eigenvectors consists of bounding the walks in a graph whose adjacency matrix corresponds
to (a subgraph of) the sparse component S ? . We then show that if the graph is sparse enough,
then this perturbation can be controlled, and thus, the next thresholding step results in further error
contraction. We use an induction argument to show that the sparse estimate is always contained in
the true support of S ? , and that there is an error contraction in each step. For the case, where L? has
rank r > 1, our algorithm proceeds in several stages, where we progressively compute higher rank
1
If the input matrix M is not symmetric, we embed it in a symmetric matrix and consider the eigenvectors
of the corresponding matrix.
2
projections which alternate with the hard thresholding steps. In stage k = [1, 2, . . . , r], we compute
rank-k projections, and show that after a sufficient number of alternating projections, we reduce the
error to the level of (k + 1)th singular value of L? , using similar arguments as in the rank-1 case.
We then proceed to performing rank-(k + 1) projections which alternate with hard thresholding.
This stage-wise procedure is needed for ill-conditioned matrices, since we cannot hope to recover
lower eigenvectors in the beginning when there are large perturbations. Thus, we establish global
convergence guarantees for our proposed non-convex robust PCA method.
1.2
Related Work
Guaranteed methods for robust PCA have received a lot of attention in the past few years, starting
from the seminal works of [CSPW11, CLMW11], where they showed recovery of an incoherent low
rank matrix L? through the following convex relaxation method:
Conv-RPCA :
min kLk? + ?kSk1 ,
s.t., M = L + S,
(1)
L,S
where kLk? denotes the nuclear norm of L (nuclear norm is the sum of singular values). A typical
solver for this convex program involves projection on to `1 and nuclear norm balls (which are convex
sets). Note that the convex method can be viewed as ?soft? thresholding in the standard and spectral
domains, while our method involves hard thresholding in these domains.
[CSPW11] and [CLMW11] consider two different models of sparsity for S ? . Chandrasekaran et
al. [CSPW11] consider a deterministic sparsity model, where each row and column of the m ? n
matrix, S, ?
has at
most ? fraction of non-zero entries. For guaranteed recovery, they require ? =
O 1/(?2 r n) , where ? is the incoherence level of L? , and r is its rank. Hsu et al. [HKZ11]
improve upon this result to obtain guarantees for an optimal sparsity level of ? = O 1/(?2 r) .
This matches the requirements of our non-convex method for exact recovery. Note that when the
rank r = O(1), this allows for a constant fraction of corrupted entries. Cand`es et al. [CLMW11]
consider a different model with random sparsity and additional incoherence constraints, viz.,
p they
?
require kU V > k? < ? r/n. Note that our assumption of incoherence, viz., kU (i) k < ? r/n,
only yields kU V > k? < ?2 r/n. The additional assumption enables [CLMW11] to prove exact
recovery with a constant fraction of corrupted entries, even when L? is nearly full-rank. We note that
removing the kU V > k? condition
? for robust PCA would imply solving the planted clique problem
when the clique size is less than n [Che13]. Thus, our recovery guarantees are tight upto constants
without these additional assumptions.
A number of works have considered modified models under the robust PCA framework,
e.g. [ANW12, XCS12]. For instance, Agarwal et al. [ANW12] relax the incoherence assumption to
a weaker ?diffusivity? assumption, which bounds the magnitude of the entries in the low rank part,
but incurs an additional approximation error. Xu et al.[XCS12] impose special sparsity structure
where a column can either be non-zero or fully zero.
In terms of state-of-art specialized solvers, [CLMW11] implements the in-exact augmented Lagrangian multipliers (IALM) method and provides guidelines for parameter tuning. Other related
methods such as multi-block alternating directions method of multipliers (ADMM) have also been
considered for robust PCA, e.g. [WHML13]. Recently, a multi-step multi-block stochastic ADMM
method was analyzed for this problem [SAJ14], and this requires 1/ iterations to achieve an error
of . In addition, the convergence rate is tight in terms of scaling with respect to problem size (m, n)
and sparsity and rank parameters, under random noise models.
There is only one other work which considers a non-convex method for robust PCA [KC12]. However, their result holds only for significantly more restrictive settings and does not cover the deterministic sparsity assumption that we study. Moreover, the projection step in their method can
have an arbitrarily large rank, so the running time is still O(m2 n), which is the same as the convex
methods. In contrast, we have an improved running time of O(r2 mn).
2
Algorithm
In this section, we present our algorithm for the robust PCA problem. The robust PCA problem can
be formulated as the following optimization problem: find L, S s.t. kM ? L ? SkF ? 2 and
2
is the desired reconstruction error
3
Figure 1: Illustration of alternating projections. The goal is to find a matrix L? which lies in the
intersection of two sets: L = { set of rank-r matrices} and SM = {M ? S, where S is a sparse
matrix}. Intuitively, our algorithm alternately projects onto the above two non-convex sets, while
appropriately relaxing the rank and the sparsity levels.
1. L lies in the set of low-rank matrices,
2. S lies in the set of sparse matrices.
A natural algorithm for the above problem is to iteratively project M ? L onto the set of sparse
matrices to update S, and then to project M ? S onto the set of low-rank matrices to update L. Alternatively, one can view the problem as that of finding a matrix L in the intersection of the following
two sets: a) L = { set of rank-r matrices}, b) SM = {M ?S, where S is a sparse matrix}. Note that
these projections can be done efficiently, even though the sets are non-convex. Hard thresholding
(HT) is employed for projections on to sparse matrices, and singular value decomposition (SVD) is
used for projections on to low rank matrices.
Rank-1 case: We first describe our algorithm for the special case when L? is rank 1. Our algorithm performs an initial hard thresholding to remove very large entries from input M . Note that if
we performed the projection on to rank-1 matrices without the initial hard thresholding, we would
not make any progress since it is subject to large perturbations. We alternate between computing
the rank-1 projection of M ? S, and performing hard thresholding on M ? L to remove entries
exceeding a certain threshold. This threshold is gradually decreased as the iterations proceed, and
the algorithm is run for a certain number of iterations (which depends on the desired reconstruction
error).
General rank case: When L? has rank r > 1, a naive extension of our algorithm consists of alternating projections on to rank-r matrices and sparse matrices. However, such a method has poor
performance on ill-conditioned matrices. This is because after the initial thresholding of the input
matrix M , the sparse corruptions in the residual are of the order of the top singular value (with the
choice of threshold as specified in the algorithm). When the lower singular values are much smaller,
the corresponding singular vectors are subject to relatively large perturbations and thus, we cannot
make progress in improving the reconstruction error. To alleviate the dependence on the condition
number, we propose an algorithm that proceeds in stages. In the k th stage, the algorithm alternates
between rank-k projections and hard thresholding for a certain number of iterations. We run the
algorithm for r stages, where r is the rank of L? . Intuitively, through this procedure, we recover
the lower singular values only after the input matrix is sufficiently denoised, i.e. sparse corruptions
at the desired level have been removed. Figure 1 shows a pictorial representation of the alternating
projections in different stages.
Parameters: As can be seen, the only real parameter to the algorithm is ?, used in thresholding,
which represents ?spikiness? of L? . That is if the user expects L? to be ?spiky? and the sparse part
to be heavily diffused, then higher value of ? can be provided. In our implementation, we found
that selecting
? aggressively helped speed up recovery of our algorithm. In particular, we selected
?
? = 1/ n.
Complexity: The complexity of each iteration within a single stage is O(kmn), since it involves
calculating the rank-k approximation3 of an m?n matrix (done e.g. via vanilla PCA). The number of
iterations in each stage is O (log (1/)) and there are at most r stages. Thus the overall complexity
2
of the entire algorithm
is then O(r mn log(1/)). This is drastically lower than the best known
2
bound of O m n/ on the number of iterations required by convex methods, and just a factor r
away from the complexity of vanilla PCA.
3
Note that we only require a rank-k approximation of the matrix rather than the actual singular vectors.
Thus, the computational complexity has no dependence on the gap between the singular values.
4
b S)
b = AltProj(M, , r, ?): Non-convex Alternating Projections based Robust PCA
Algorithm 1 (L,
1: Input: Matrix M ? Rm?n , convergence criterion , target rank r, thresholding parameter ?.
2: Pk (A) denotes the best rank-k approximation of matrix A. HT? (A) denotes hard-thresholding,
i.e. (HT? (A))ij = Aij if |Aij | ? ? and 0 otherwise.
3: Set initial threshold ?0 ? ??1 (M ).
4: L(0) = 0, S (0) = HT?0 (M ? L(0) )
5: for Stage k = 1 to r do
6:
for Iteration t = 0 to T = 10 log n?
M ? S (0)
2 / do
7:
Set threshold ? as
!
t
1
(t)
(t)
? = ? ?k+1 (M ? S ) +
?k (M ? S )
(2)
2
8:
L(t+1) = Pk (M ? S (t) )
9:
S (t+1) = HT? (M ? L(t+1) )
10:
end for
then
11:
if ??k+1 (L(t+1) ) < 2n
(T )
(T )
12:
Return: L , S
/* Return rank-k estimate if remaining part has small norm */
13:
else
14:
S (0) = S (T )
/* Continue to the next stage */
15:
end if
16: end for
17: Return: L(T ) , S (T )
3
Analysis
In this section, we present our main result on the correctness of AltProj. We assume the following
conditions:
(L1) Rank of L? is at most r.
(L2) L? is ?-incoherent, i.e., if L? = U ? ?? (V ? )> is the SVD of L? , then k(U ? )i k2 ?
?
?
?
? r,
m
th
?1 ? i ? m and k(V ? )i k2 ? ??nr , ?1 ? i ? n, where (U ? )i and (V ? )i denote the i
rows of U ? and V ? respectively.
(S1) Each row and column of S have at most ? fraction of non-zero entries such that ? ?
1
512?2 r .
Note that in general, it is not possible to have a unique recovery of low-rank and sparse components.
For example, if the input matrix M is both sparse and low rank, then there is no unique decomposition (e.g. M = e1 e>
1 ). The above conditions ensure uniqueness of the matrix decomposition
problem.
Additionally, we set the parameter ? in Algorithm 1 be set as ? =
4?2 r
?
.
mn
We now establish that our proposed algorithm recovers the low rank and sparse components under
the above conditions.
Theorem 1 (Noiseless Recovery). Under conditions (L1), (L2) and S ? , and choice of ? as above,
b and Sb of Algorithm 1 satisfy:
the outputs L
b
, and Supp Sb ? Supp (S ? ) .
L ? L?
? ,
Sb ? S ?
? ?
mn
F
?
Remark (tight recovery conditions): Our result is tight up to constants, in terms of allowable
sparsity level under the deterministic sparsity model. In other words, if we exceed the sparsity limit
imposed in S1, it is possible to construct instances where there is no unique decomposition4 . Our
4
For instance, consider the n ? n matrix which has r copies of the all ones matrix, each of size nr , placed
across the diagonal. We see that this matrix has rank r and is incoherent with parameter ? = 1. Note that
5
conditions L1, L2 and S1 also match the conditions required by the convex method for recovery, as
established in [HKZ11].
Remark (convergence rate): Our method has a linear rate of convergence, i.e. O(log(1/))
to achieve an error of , and hence we provide a strongly polynomial method for robust PCA. In
contrast, the best known bound for convex methods for robust PCA is O(1/) iterations to converge
to an -approximate solution.
Theorem 1 provides recovery guarantees assuming that L? is exactly rank-r. However, in several
real-world scenarios, L? can be nearly rank-r. Our algorithm can handle such situations, where
M = L? + N ? + S ? , with N ? being an additive noise. Theorem 1 is a special case of the following
theorem which provides recovery guarantees when N ? has small `? norm.
Theorem 2 (Noisy Recovery). Under conditions (L1), (L2) and S ? , and choice of ? as in Theo?
r (L )
b Sb of Algorithm 1 satisfy:
,the outputs L,
rem 1, when the noise kN ? k? ? ?100n
?
8 mn
b
2
?
?
?
?
?
+
2?
r
7
kN
k
+
L
?
L
kN
k
2
? ,
r
F
?
2?2 r
8 mn
b
?
?
?
?
?
?
+
7 kN k2 +
kN k? , and Supp Sb ? Supp (S ? ) .
S ? S
?
mn
mn
r
?
3.1
Proof Sketch
We now present the key steps in the proof of Theorem 1. A detailed proof is provided in the appendix.
Step I: Reduce to the symmetric case, while maintaining incoherence of L? and sparsity of S ? .
Using standard symmetrization arguments, we can reduce the problem to the symmetric case, where
all the matrices involved are symmetric. See appendix for details on this step.
Step II: Show decay in kL ? L? k? after projection onto the set of rank-k matrices. The
t-th iterate L(t+1) of the k-th stage is given by L(t+1) = Pk (L? + S ? ? S (t) ). Hence, L(t+1) is
obtained by using the top principal components of a perturbation of L? given by L? + (S ? ? S (t) ).
The key step in our analysis is to show that when an incoherent and low-rank L? is perturbed by a
sparse matrix S ? ? S (t) , then kL(t+1) ? L? k? is small and is much smaller than |S ? ? S (t) |? . The
following lemma formalizes the intuition; see the appendix for a detailed proof.
Lemma 1. Let L? , S ? be symmetric and satisfy the assumptions of Theorem 1 and let S (t) and L(t)
be the tth iterates of the k th stage of Algorithm 1. Let ?1? , . . . , ?n? be the eigenvalues of L? , s.t.,
|?1? | ? ? ? ? ? |?r? |. Then, the following holds:
!
t
1
2?2 r ?
(t+1)
?
?
?k+1 +
?L
?
|?k | ,
L
n
2
?
!
t
8?2 r ?
1
?
(t+1)
?
?k+1 +
|?k | , and Supp S (t+1) ? Supp (S ? ) .
S ? S
?
n
2
?
b and Sb of Algorithm 1 satisfy:
Moreover, the outputs L
b
L ? L?
? ,
Sb ? S ?
? , and Supp Sb ? Supp (S ? ) .
n
F
?
Step III: Show decay in kS ? S ? k? after projection onto the set of sparse matrices. We next
show that if kL(t+1) ? L? k? is much smaller than kS (t) ? S ? k? then the iterate S (t+1) also has
a much smaller error (w.r.t. S ? ) than S (t) . The above given lemma formally provides the error
bound.
Step IV: Recurse the argument. We have now reduced the `? norm of the sparse part by a factor of
half, while maintaining its sparsity. We can now go back to steps II and III and repeat the arguments
for subsequent iterations.
a fraction of ? = O (1/r) sparse perturbations suffice to erase one of these blocks making it impossible to
recover the matrix.
6
500
600
n?
700
800
IALM
2
400
200
500
600
n?
700
800
10
1
n = 2000, ? = 1, n ? = 3r
AltProj
IALM
1.5
AltProj
IALM
2
10
Time(s)
600
Time(s)
2
10
AltProj
IALM
n = 2000, r = 10, n ? = 100
n = 2000, r = 5, ? = 1
Max. Rank
Time(s)
n = 2000, r = 5, ? = 1
2
?
2.5
3
50
100
r
150
200
(a)
(b)
(c)
(d)
Figure 2: Comparison of AltProj and IALM on synthetic datasets. (a) Running time of AltProj and
IALM with varying ?. (b) Maximum rank of the intermediate iterates of IALM. (c) Running time
of AltProj and IALM with varying ?. (d) Running time of AltProj and IALM with varying r.
4
Experiments
We now present an empirical study of our AltProj method. The goal of this study is two-fold: a)
establish that our method indeed recovers the low-rank and sparse part exactly, without significant
parameter tuning, b) demonstrate that AltProj is significantly faster than Conv-RPCA (see (1)); we
solve Conv-RPCA using the IALM method [CLMW11], a state-of-the-art solver [LCM10]. We
implemented our method in Matlab and used a Matlab implementation of the IALM method by
[LCM10].
We consider both synthetic experiments and experiments on real data involving the problem of
foreground-background separation in a video. Each of our results for synthetic datasets is averaged
over 5 runs.
Parameter Setting: Our pseudo-code (Algorithm 1) prescribes the threshold ? in Step 4, which
depends on the knowledge of the singular values of the low rank component L? . Instead, in the
?S (t) )
?
experiments, we set the threshold at the (t + 1)-th step of k-th stage as ? = ??k+1 (M
. For
n
synthetic experiments, we employ the ? used for data generation, and for real-world datasets, we
tune ? through cross-validation. We found that the above thresholding provides exact recovery
while speeding up the computation significantly. We would
also like to note that [CLMW11] sets
?
the regularization parameter ? in Conv-RPCA (1) as 1/ n (assuming m ? n). However, we found
that for problems with ?
large incoherence such a parameter setting does not provide exact recovery.
Instead, we set ? = ?/ n in our experiments.
Synthetic datasets: Following the experimental setup of [CLMW11], the low-rank part L? = U V T
is generated using normally distributed U ? Rm?r , V ? Rn?r . Similarly, supp(S ? ) is generated
?
by sampling a uniformly random subset of [m] ?[n] with size kS ? k0 and each non-zero Sij
is drawn
?
?
i.i.d. from the uniform distribution over [r/(2 mn), r/ mn]. For increasing incoherence of L? ,
we randomly zero-out rows of U, V and then re-normalize them.
There are three key problem parameters for RPCA with a fixed matrix size: a) sparsity of S ? ,
b) incoherence of L? , c) rank of L? . We investigate performance of both AltProj and IALM by
varying each of the three parameters while fixing the others. In our plots (see Figure 2), we report
computational time required by each of the two methods for decomposing M into L + S up to a
relative error (kM ? L ? SkF /kM kF ) of 10?3 . Figure 2 shows that AltProj scales significantly
better than IALM for increasingly dense S ? . We attribute this observation to the fact that as kS ? k0
increases, the problem is ?harder? and the intermediate iterates of IALM have ranks significantly
larger than r. Our intuition is confirmed by Figure 2 (b), which shows that when density (?) of S ?
is 0.4 then the intermediate iterates of IALM can be of rank over 500 while the rank of L? is only
5. We observe a similar trend for the other parameters, i.e., AltProj scales significantly better than
IALM with increasing incoherence parameter ? (Figure 2 (c)) and increasing rank (Figure 2 (d)).
See Appendix C for additional plots.
Real-world datasets: Next, we apply our method to the problem of foreground-background (F-B)
separation in a video [LHGT04]. The observed matrix M is formed by vectorizing each frame and
stacking them column-wise. Intuitively, the background in a video is the static part and hence forms
a low-rank component while the foreground is a dynamic but sparse perturbation.
Here, we used two benchmark datasets named Escalator and Restaurant dataset. The Escalator
dataset has 3417 frames at a resolution of 160 ? 130. We first applied the standard PCA method for
extracting low-rank part. Figure 3 (b) shows the extracted background from the video. There are
7
(a)
(b)
(c)
(d)
Figure 3: Foreground-background separation in the Escalator video. (a): Original image frame. (b):
Best rank-10 approximation; time taken is 3.1s. (c): Low-rank frame obtained using AltProj; time
taken is 63.2s. (d): Low-rank frame obtained using IALM; time taken is 1688.9s.
(a)
(b)
(c)
(d)
Figure 4: Foreground-background separation in the Restaurant video. (a): Original frame from the
video. (b): Best rank-10 approximation (using PCA) of the original frame; 2.8s were required to
compute the solution (c): Low-rank part obtained using AltProj; computational time required by
AltProj was 34.9s. (d): Low-rank part obtained using IALM; 693.2s required by IALM to compute
the low-rank+sparse decomposition.
several artifacts (shadows of people near the escalator) that are not desirable. In contrast, both IALM
and AltProj obtain significantly better F-B separation (see Figure 3(c), (d)). Interestingly, AltProj removes the steps of the escalator which are moving and arguably are part of the dynamic foreground,
while IALM keeps the steps in the background part. Also, our method is significantly faster, i.e.,
our method, which takes 63.2s is about 26 times faster than IALM, which takes 1688.9s.
Restaurant dataset: Figure 4 shows the comparison of AltProj and IALM on a subset of the ?Restaurant? dataset where we consider the last 2055 frames at a resolution of 120?160. AltProj was around
19 times faster than IALM. Moreover, visually, the background extraction seems to be of better quality (for example, notice the blur near top corner counter in the IALM solution). Plot(b) shows the
PCA solution and that also suffers from a similar blur at the top corner of the image, while the
background frame extracted by AltProj does not have any noticeable artifacts.
5
Conclusion
In this work, we proposed a non-convex method for robust PCA, which consists of alternating projections on to low rank and sparse matrices. We established global convergence of our method under
conditions which match those for convex methods. At the same time, our method has much faster
running times, and has superior experimental performance. This work opens up a number of interesting questions for future investigation. While we match the convex methods, under the deterministic
sparsity model, studying the random sparsity model is of interest. Our noisy recovery results assume deterministic noise; improving the results under random noise needs to be investigated. There
are many decomposition problems beyond the robust PCA setting, e.g. structured sparsity models,
robust tensor PCA problem, and so on. It is interesting to see if we can establish global convergence
for non-convex methods in these settings.
Acknowledgements
AA and UN would like to acknowledge NSF grant CCF-1219234, ONR N00014-14-1-0665, and Microsoft faculty fellowship. SS would like to acknowledge NSF grants 1302435, 0954059, 1017525
and DTRA grant HDTRA1-13-1-0024. PJ would like to acknowledge Nikhil Srivastava and
Deeparnab Chakrabarty for several insightful discussions during the course of the project.
8
References
[AAJ+ 13]
A. Agarwal, A. Anandkumar, P. Jain, P. Netrapalli, and R. Tandon. Learning
Sparsely Used Overcomplete Dictionaries via Alternating Minimization. Available
on arXiv:1310.7991, Oct. 2013.
[AGH+ 12] A. Anandkumar, R. Ge, D. Hsu, S. M. Kakade, and M. Telgarsky. Tensor Methods for
Learning Latent Variable Models. Available at arXiv:1210.7559, Oct. 2012.
[ANW12]
A. Agarwal, S. Negahban, and M. Wainwright. Noisy matrix decomposition via convex
relaxation: Optimal rates in high dimensions. The Annals of Statistics, 40(2):1171?
1197, 2012.
[Bha97]
Rajendra Bhatia. Matrix Analysis. Springer, 1997.
[Che13]
Y. Chen. Incoherence-Optimal Matrix Completion. ArXiv e-prints, October 2013.
[CLMW11] Emmanuel J. Cand`es, Xiaodong Li, Yi Ma, and John Wright. Robust principal component analysis? J. ACM, 58(3):11, 2011.
[CSPW11] Venkat Chandrasekaran, Sujay Sanghavi, Pablo A. Parrilo, and Alan S. Willsky.
Rank-sparsity incoherence for matrix decomposition. SIAM Journal on Optimization,
21(2):572?596, 2011.
[CSX12]
Yudong Chen, Sujay Sanghavi, and Huan Xu. Clustering sparse graphs. In Advances
in neural information processing systems, pages 2204?2212, 2012.
[EKYY13] L?aszl?o Erd?os, Antti Knowles, Horng-Tzer Yau, and Jun Yin. Spectral statistics of
Erd?os?R?enyi graphs I: Local semicircle law. The Annals of Probability, 2013.
[Har13]
Moritz Hardt. On the provable convergence of alternating minimization for matrix
completion. arXiv:1312.0925, 2013.
[HKZ11]
Daniel Hsu, Sham M Kakade, and Tong Zhang. Robust matrix decomposition with
sparse corruptions. ITIT, 2011.
[JNS13]
Prateek Jain, Praneeth Netrapalli, and Sujay Sanghavi. Low-rank matrix completion
using alternating minimization. In STOC, 2013.
[KC12]
Anastasios Kyrillidis and Volkan Cevher. Matrix alps: Accelerated low rank and sparse
matrix reconstruction. In SSP Workshop, 2012.
[Kes12]
Raghunandan H. Keshavan. Efficient algorithms for collaborative filtering. Phd Thesis,
Stanford University, 2012.
[LCM10]
Zhouchen Lin, Minming Chen, and Yi Ma. The augmented lagrange multiplier method
for exact recovery of corrupted low-rank matrices. arXiv:1009.5055, 2010.
[LHGT04]
Liyuan Li, Weimin Huang, IY-H Gu, and Qi Tian. Statistical modeling of complex
backgrounds for foreground object detection. ITIP, 2004.
[MZYM11] Hossein Mobahi, Zihan Zhou, Allen Y. Yang, and Yi Ma. Holistic 3d reconstruction of
urban structures from low-rank textures. In ICCV Workshops, pages 593?600, 2011.
[NJS13]
Praneeth Netrapalli, Prateek Jain, and Sujay Sanghavi. Phase retrieval using alternating
minimization. In NIPS, pages 2796?2804, 2013.
[SAJ14]
H. Sedghi, A. Anandkumar, and E. Jonckheere. Guarantees for Stochastic ADMM in
High Dimensions. Preprint., Feb. 2014.
[Shi13]
Lei Shi. Sparse additive text models with low rank background. In Advances in Neural
Information Processing Systems, pages 172?180, 2013.
[WHML13] X. Wang, M. Hong, S. Ma, and Z. Luo. Solving multiple-block separable convex
minimization problems using two-block alternating direction method of multipliers.
arXiv:1308.5294, 2013.
[XCS12]
Huan Xu, Constantine Caramanis, and Sujay Sanghavi. Robust pca via outlier pursuit.
IEEE Transactions on Information Theory, 58(5):3047?3064, 2012.
9
| 5430 |@word faculty:1 polynomial:1 norm:6 seems:1 open:1 km:3 decomposition:11 covariance:1 contraction:3 minming:1 incurs:1 harder:1 klk:2 carry:1 initial:4 series:1 selecting:1 daniel:1 interestingly:1 outperforms:1 existing:3 past:1 current:1 ksk1:1 luo:1 jns13:2 john:1 additive:2 subsequent:1 blur:2 enables:1 remove:5 plot:3 progressively:1 update:2 half:1 fewer:1 selected:1 beginning:1 volkan:1 provides:6 characterization:1 iterates:4 projec:1 zhang:1 prove:3 consists:4 indeed:1 cand:2 growing:1 multi:3 inspired:1 rem:1 actual:1 solver:4 increasing:3 conv:4 project:4 provided:2 moreover:4 suffice:2 erase:1 prateek:3 finding:1 guarantee:9 formalizes:1 pseudo:1 exactly:2 rm:2 k2:3 normally:1 grant:3 arguably:1 local:1 limit:1 establishing:2 incoherence:13 k:4 relaxing:1 range:1 tian:1 averaged:1 unique:3 practice:2 block:5 implement:2 procedure:4 empirical:1 semicircle:1 significantly:9 projection:24 ups:1 word:1 refers:1 spite:1 onto:8 close:1 cannot:2 impossible:1 seminal:2 deterministic:6 lagrangian:1 imposed:1 shi:1 go:1 attention:1 starting:1 zihan:1 convex:38 agh:2 resolution:2 recovery:20 m2:3 nuclear:3 proving:1 handle:1 notion:1 annals:2 target:1 heavily:1 user:1 exact:7 tandon:1 element:2 trend:1 expensive:1 sparsely:1 observed:1 aszl:1 preprint:1 solved:1 wang:1 counter:1 removed:1 intuition:2 convexity:1 complexity:10 dynamic:2 prescribes:1 tight:6 solving:2 upon:1 gu:1 k0:2 caramanis:1 escalator:5 enyi:2 fast:1 describe:2 jain:3 bhatia:1 aaj:2 whose:1 larger:1 solve:1 stanford:1 say:1 relax:1 otherwise:1 s:1 nikhil:1 statistic:2 kahan:1 noisy:3 advantage:1 eigenvalue:1 propose:3 reconstruction:6 holistic:1 subgraph:1 achieve:5 intuitive:1 normalize:1 exploiting:1 convergence:15 requirement:2 telgarsky:1 object:1 tions:1 derive:2 completion:4 fixing:1 ij:1 lowrank:1 noticeable:1 keywords:1 received:1 progress:2 netrapalli:4 recovering:1 implemented:1 involves:6 shadow:1 strong:1 direction:2 attribute:1 stochastic:2 alp:1 adjacency:1 require:6 decompose:1 alleviate:1 investigation:1 extension:1 hold:2 sufficiently:1 considered:2 around:1 wright:1 visually:2 dictionary:2 vary:1 uniqueness:1 rpca:5 sensitive:1 bridge:1 symmetrization:1 correctness:2 establishes:1 minimization:6 hope:1 always:1 modified:1 rather:1 bha97:1 zhou:1 varying:4 viz:3 rank:87 contrast:5 sb:8 typically:1 entire:1 provably:1 overall:1 hossein:1 ill:2 development:1 art:3 special:3 construct:1 extraction:1 sampling:1 represents:1 unsupervised:1 nearly:3 foreground:8 future:1 hdtra1:1 others:1 report:1 sanghavi:5 few:1 employ:1 randomly:1 pictorial:1 phase:2 raghunandan:1 microsoft:4 attempt:1 detection:2 interest:1 investigate:1 analyzed:1 recurse:1 huan:2 skf:2 iv:1 taylor:1 walk:1 desired:3 re:1 overcomplete:1 theoretical:1 cevher:1 instance:4 column:5 modeling:3 soft:1 cover:1 cost:2 stacking:1 entry:8 expects:1 subset:2 uniform:1 interning:1 kn:5 perturbed:1 corrupted:3 synthetic:6 density:1 negahban:1 siam:1 iy:1 thesis:1 huang:1 corner:2 yau:1 return:3 li:2 supp:9 parrilo:1 ialm:28 satisfy:4 depends:2 performed:1 view:1 lot:1 helped:1 recover:3 denoised:1 contribution:1 collaborative:1 formed:1 accuracy:5 efficiently:1 yield:1 confirmed:1 corruption:5 rajendra:1 reach:1 suffers:1 inexact:1 involved:1 chakrabarty:1 proof:6 recovers:2 static:1 irvine:1 gain:1 dataset:5 hsu:3 animashree:1 hardt:1 knowledge:1 back:1 higher:2 improved:2 erd:4 done:3 though:1 strongly:1 just:1 stage:15 spiky:1 sketch:1 keshavan:1 o:4 artifact:2 reveal:1 quality:1 lei:1 xiaodong:1 multiplier:5 true:2 ccf:1 hence:3 regularization:1 aggressively:1 alternating:16 symmetric:6 iteratively:1 moritz:1 during:1 davis:1 criterion:1 hong:1 allowable:1 demonstrate:1 performs:1 l1:4 allen:1 image:2 wise:4 novel:1 recently:2 common:1 superior:1 rmn:3 specialized:2 overview:1 exponentially:2 significant:3 cambridge:1 jonckheere:1 tuning:2 sujay:6 vanilla:2 zhouchen:1 similarly:1 moving:1 feb:1 showed:2 recent:1 constantine:1 scenario:1 certain:3 n00014:1 onr:1 success:1 arbitrarily:1 continue:1 yi:3 seen:1 additional:5 impose:1 dtra:1 employed:3 converge:1 ii:2 multiple:1 full:1 desirable:1 sham:1 anastasios:1 alan:1 match:7 faster:6 cross:1 retrieval:2 lin:1 e1:1 controlled:1 qi:1 involving:2 noiseless:1 arxiv:6 iteration:24 agarwal:3 achieved:1 background:13 addition:4 fellowship:1 decreased:1 spikiness:1 else:1 singular:12 horng:1 appropriately:1 subject:3 elegant:1 anandkumar:3 extracting:1 near:2 presence:1 yang:1 exceed:1 iii:2 easy:1 enough:1 intermediate:3 iterate:2 fit:1 restaurant:4 reduce:3 praneeth:3 idea:1 kyrillidis:1 texas:1 pca:38 proceed:2 remark:2 matlab:2 detailed:2 eigenvectors:4 tune:1 tth:1 reduced:1 kc12:2 nsf:2 notice:1 estimated:1 per:6 key:3 threshold:7 drawn:1 urban:1 pj:1 ht:5 graph:5 relaxation:4 fraction:6 year:1 sum:1 run:4 named:1 chandrasekaran:2 knowles:1 separation:7 appendix:4 scaling:2 bound:10 guaranteed:2 fold:1 constraint:1 speed:3 argument:5 min:1 performing:2 separable:1 relatively:1 structured:1 alternate:4 ball:1 poor:2 smaller:6 across:1 increasingly:1 kakade:2 making:1 s1:3 projecting:1 iccv:1 outlier:3 intuitively:3 gradually:1 sij:1 taken:3 needed:1 ge:1 end:3 studying:1 available:2 decomposing:2 pursuit:1 apply:1 observe:1 away:1 appropriate:1 spectral:2 upto:1 eigen:1 original:3 denotes:3 running:10 top:4 remaining:1 ensure:1 clustering:1 maintaining:2 calculating:1 restrictive:1 emmanuel:1 especially:1 establish:5 tensor:3 diffused:1 question:1 print:1 planted:1 dependence:2 usual:2 nr:2 diagonal:1 ssp:1 topic:1 considers:1 provable:3 induction:1 willsky:1 assuming:3 sedghi:1 code:1 illustration:1 setup:1 october:1 stoc:1 implementation:4 design:1 guideline:1 unknown:1 perform:1 observation:1 datasets:6 sm:2 benchmark:1 acknowledge:3 situation:1 frame:9 rn:1 perturbation:11 community:1 pablo:1 required:7 specified:1 kl:3 california:1 established:2 alternately:1 nip:1 beyond:1 proceeds:2 sparsity:20 program:1 including:1 max:1 video:7 wainwright:1 power:1 natural:2 force:1 residual:4 mn:12 improve:1 imply:1 carried:1 incoherent:4 jun:1 naive:1 speeding:1 text:1 l2:4 vectorizing:1 kf:1 acknowledgement:1 relative:1 law:1 fully:1 generation:1 interesting:2 filtering:1 validation:1 sufficient:1 exciting:1 thresholding:17 austin:1 row:5 course:1 summary:1 placed:1 last:1 antti:1 repeat:1 copy:1 theo:1 enjoys:1 drastically:2 clmw11:12 weaker:1 aij:2 india:2 wide:1 sparse:39 distributed:1 overcome:1 dimension:2 yudong:1 world:4 concretely:2 preprocessing:1 transaction:1 approximate:1 keep:1 clique:2 global:5 reveals:1 alternatively:1 un:1 latent:1 additionally:1 ku:4 robust:28 improving:2 expansion:1 investigated:1 complex:1 domain:2 pk:3 main:1 dense:1 bounding:1 noise:5 kmn:1 xu:3 augmented:3 venkat:1 tong:1 diffusivity:1 exceeding:1 lie:3 removing:1 theorem:7 embed:1 altproj:22 insightful:1 mobahi:1 r2:3 list:1 decay:2 workshop:2 phd:1 magnitude:1 texture:1 conditioned:2 gap:3 chen:3 intersection:2 yin:1 lagrange:2 contained:1 deeparnab:1 springer:1 aa:1 corresponds:1 extracted:2 ma:5 acm:1 oct:2 goal:4 viewed:1 formulated:1 admm:3 hard:11 typical:1 uniformly:1 denoising:1 principal:3 lemma:3 e:2 svd:2 experimental:2 formally:1 support:2 people:1 accelerated:1 srivastava:1 |
4,895 | 5,431 | Spectral Methods Meet EM: A Provably Optimal
Algorithm for Crowdsourcing
Yuchen Zhang?
Xi Chen]
Dengyong Zhou?
Michael I. Jordan?
?
University of California, Berkeley, Berkeley, CA 94720
{yuczhang,jordan}@berkeley.edu
]
New York University, New York, NY 10012
[email protected]
?
Microsoft Research, 1 Microsoft Way, Redmond, WA 98052
[email protected]
Abstract
The Dawid-Skene estimator has been widely used for inferring the true labels
from the noisy labels provided by non-expert crowdsourcing workers. However,
since the estimator maximizes a non-convex log-likelihood function, it is hard to
theoretically justify its performance. In this paper, we propose a two-stage efficient algorithm for multi-class crowd labeling problems. The first stage uses the
spectral method to obtain an initial estimate of parameters. Then the second stage
refines the estimation by optimizing the objective function of the Dawid-Skene
estimator via the EM algorithm. We show that our algorithm achieves the optimal
convergence rate up to a logarithmic factor. We conduct extensive experiments on
synthetic and real datasets. Experimental results demonstrate that the proposed
algorithm is comparable to the most accurate empirical approach, while outperforming several other recently proposed methods.
1
Introduction
With the advent of online crowdsourcing services such as Amazon Mechanical Turk, crowdsourcing
has become an appealing way to collect labels for large-scale data. Although this approach has
virtues in terms of scalability and immediate availability, labels collected from the crowd can be of
low quality since crowdsourcing workers are often non-experts and can be unreliable. As a remedy,
most crowdsourcing services resort to labeling redundancy, collecting multiple labels from different
workers for each item. Such a strategy raises a fundamental problem in crowdsourcing: how to infer
true labels from noisy but redundant worker labels?
For labeling tasks with k different categories, Dawid and Skene [8] propose a maximum likelihood
approach based on the Expectation-Maximization (EM) algorithm. They assume that each worker
is associated with a k ? k confusion matrix, where the (l, c)-th entry represents the probability that
a randomly chosen item in class l is labeled as class c by the worker. The true labels and worker confusion matrices are jointly estimated by maximizing the likelihood of the observed worker
labels, where the unobserved true labels are treated as latent variables. Although this EM-based
approach has had empirical success [21, 20, 19, 26, 6, 25], there is as yet no theoretical guarantee
for its performance. A recent theoretical study [10] shows that the global optimal solutions of the
Dawid-Skene estimator can achieve minimax rates of convergence in a simplified scenario, where
the labeling task is binary and each worker has a single parameter to represent her labeling accuracy (referred to as a ?one-coin model? in what follows). However, since the likelihood function is
non-convex, this guarantee is not operational because the EM algorithm may get trapped in a local
optimum. Several alternative approaches have been developed that aim to circumvent the theoretical
deficiencies of the EM algorithm, still in the context of the one-coin model [14, 15, 11, 7]. Unfor1
tunately, they either fail to achieve the optimal rates or depend on restrictive assumptions which are
hard to justify in practice.
We propose a computationally efficient and provably optimal algorithm to simultaneously estimate
true labels and worker confusion matrices for multi-class labeling problems. Our approach is a
two-stage procedure, in which we first compute an initial estimate of worker confusion matrices
using the spectral method, and then in the second stage we turn to the EM algorithm. Under some
mild conditions, we show that this two-stage procedure achieves minimax rates of convergence up
to a logarithmic factor, even after only one iteration of EM. In particular, given any ? ? (0, 1),
we provide the bounds on the number of workers and the number of items so that our method can
correctly estimate labels for all items with probability at least 1 ? ?. We also establish a lower bound
to demonstrate the optimality of this approach. Further, we provide both upper and lower bounds for
estimating the confusion matrix of each worker and show that our algorithm achieves the optimal
accuracy.
This work not only provides an optimal algorithm for crowdsourcing but sheds light on understanding the general method of moments. Empirical studies show that when the spectral method is used
as an initialization for the EM algorithm, it outperforms EM with random initialization [18, 5]. This
work provides a concrete way to theoretically justify such observations. It is also known that starting
from a root-n consistent estimator obtained by the spectral method, one Newton-Raphson step leads
to an asymptotically optimal estimator [17]. However, obtaining a root-n consistent estimator and
performing a Newton-Raphson step can be demanding computationally. In contrast, our initialization doesn?t need to be root-n consistent, thus a small portion of data suffices to initialize. Moreover,
performing one iteration of EM is computationally more attractive and numerically more robust than
a Newton-Raphson step especially for high-dimensional problems.
2
Related Work
Many methods have been proposed to address the problem of estimating true labels in crowdsourcing
[23, 20, 22, 11, 19, 26, 7, 15, 14, 25]. The methods in [20, 11, 15, 19, 14, 7] are based on the
generative model proposed by Dawid and Skene [8]. In particular, Ghosh et al. [11] propose a
method based on Singular Value Decomposition (SVD) which addresses binary labeling problems
under the one-coin model. The analysis in [11] assumes that the labeling matrix is full, that is,
each worker labels all items. To relax this assumption, Dalvi et al. [7] propose another SVD-based
algorithm which explicitly considers the sparsity of the labeling matrix in both algorithm design
and theoretical analysis. Karger et al. propose an iterative algorithm for binary labeling problems
under the one-coin model [15] and extend it to multi-class labeling tasks by converting a k-class
problem into k ? 1 binary problems [14]. This line of work assumes that tasks are assigned to
workers according to a random regular graph, thus imposing specific constraints on the number
of workers and the number of items. In Section 5, we compare our theoretical results with that
of existing approaches [11, 7, 15, 14]. The methods in [20, 19, 6] incorporate Bayesian inference
into the Dawid-Skene estimator by assuming a prior over confusion matrices. Zhou et al. [26,
25] propose a minimax entropy principle for crowdsourcing which leads to an exponential family
model parameterized with worker ability and item difficulty. When all items have zero difficulty, the
exponential family model reduces to the generative model suggested by Dawid and Skene [8].
Our method for initializing the EM algorithm in crowdsourcing is inspired by recent work using
spectral methods to estimate latent variable models [3, 1, 4, 2, 5, 27, 12, 13]. The basic idea in this
line of work is to compute third-order empirical moments from the data and then to estimate parameters by computing a certain orthogonal decomposition of a tensor derived from the moments. Given
the special symmetric structure of the moments, the tensor factorization can be computed efficiently using the robust tensor power method [3]. A problem with this approach is that the estimation
error can have a poor dependence on the condition number of the second-order moment matrix and
thus empirically it sometimes performs worse than EM with multiple random initializations. Our
method, by contrast, requires only a rough initialization from the moment of moments; we show that
the estimation error does not depend on the condition number (see Theorem 2 (b)).
3
Problem Setup
Throughout this paper, [a] denotes the integer set {1, 2, . . . , a} and ?b (A) denotes the b-th largest
singular value of the matrix A. Suppose that there are m workers, n items and k classes. The true
2
Algorithm 1: Estimating confusion matrices
Input: integer k, observed labels zij ? Rk for i ? [m] and j ? [n].
bi ? Rk?k for i ? [m].
Output: confusion matrix estimates C
(1) Partition the workers into three disjoint and non-empty group G1 , G2 and G3 . Compute the
group aggregated labels Zgj by Eq. (1).
(2) For (a, b, c) ? {(2, 3, 1), (3, 1, 2), (1, 2, 3)}, compute the second and the third order moments
c2 ? Rk?k , M
c3 ? Rk?k?k by Eq. (2a)-(2d), then compute C
bc ? Rk?k and W
c ? Rk?k by
M
tensor decomposition:
b ? Rk?k (such that Q
bT M
c2 Q
b = I) using SVD.
(a) Compute whitening matrix Q
c3 (Q,
b Q,
b Q)
b
(b) Compute eigenvalue-eigenvector pairs {(b
?h , vbh )}kh=1 of the whitened tensor M
by using the robust tensor power method [3]. Then compute w
bh = ?
bh?2 and
b T )?1 (b
?
bh = (Q
?h vbh ).
bc by some ?
(c) For l = 1, . . . , k, set the l-th column of C
bh whose l-th coordinate has the
c by w
greatest component, then set the l-th diagonal entry of W
bh .
bi by Eq. (3).
(3) Compute C
label yj of item j ? [n] is assumed to be sampled from a probability distribution P[yj = l] = wl
Pk
where {wl : l ? [k]} are positive values satisfying l=1 wl = 1. Denote by a vector zij ? Rk
the label that worker i assigns to item j. When the assigned label is c, we write zij = ec , where ec
represents the c-th canonical basis vector in Rk in which the c-th entry is 1 and all other entries are
0. A worker may not label every item. Let ?i indicate the probability that worker i labels a randomly
chosen item. If item j is not labeled by worker i, we write zij = 0. Our goal is to estimate the true
labels {yj : j ? [n]} from the observed labels {zij : i ? [m], j ? [n]}.
In order to obtain an estimator, we need to make assumptions on the process of generating observed
labels. Following the work of Dawid and Skene [8], we assume that the probability that worker i
labels an item in class l as class c is independent of any particular chosen item, that is, it is a constant
over j ? [n]. Let us denote the constant probability by ?ilc . Let ?il = [?il1 ?il2 ? ? ? ?ilk ]T . The
matrix Ci = [?i1 ?i2 . . . ?ik ] ? Rk?k is called the confusion matrix of worker i. Besides estimating
the true labels, we also want to estimate the confusion matrix for each worker.
4
Our Algorithm
In this section, we present an algorithm to estimate confusion matrices and true labels. Our algorithm
consists of two stages. In the first stage, we compute an initial estimate of confusion matrices via
the method of moments. In the second stage, we perform the standard EM algorithm by taking the
result of the Stage 1 as an initialization.
4.1
Stage 1: Estimating Confusion Matrices
Partitioning the workers into three disjoint and non-empty groups G1 , G2 and G3 , the outline of
this stage is the following: we use the spectral method to estimate the averaged confusion matrices
for the three groups, then utilize this intermediate estimate to obtain the confusion matrix of each
individual worker. In particular, for g ? {1, 2, 3} and j ? [n], we calculate the averaged labeling
within each group by
1 X
Zgj :=
zij .
(1)
|Gg |
i?Gg
P
Denoting the aggregated confusion matrix columns by ?gl := E(Zgj |yj = l) = |G1g | i?Gg ?i ?il ,
our first step is to estimate Cg := [?g1 , ?g2 , . . . , ?gk ] and to estimate the distribution of true labels
3
W := diag(w1 , w2 , . . . , wk ). The following proposition shows that we can solve for Cg and W
from the moments of {Zgj }.
Proposition 1 (Anandkumar et al. [3]). Assume that the vectors {?g1 , ?g2 , . . . , ?gk } are linearly
independent for each g ? {1, 2, 3}. Let (a, b, c) be a permutation of {1, 2, 3}. Define
?1
Zaj ,
?1
Zbj ,
0
Zaj
:= E[Zcj ? Zbj ] (E[Zaj ? Zbj ])
0
Zbj
:= E[Zcj ? Zaj ] (E[Zbj ? Zaj ])
0
E[Zaj
0
Zbj
]
0
0
M2 :=
?
and M3 := E[Zaj
? Zbj
? Zcj ];
Pk
P
k
then we have M2 = l=1 wl ?cl ? ?cl and M3 = l=1 wl ?cl ? ?cl ? ?cl .
Since we only have finite samples, the expectations in Proposition 1 have to be approximated by
empirical moments. In particular, they are computed by averaging over indices j = 1, 2, . . . , n. For
each permutation (a, b, c) ? {(2, 3, 1), (3, 1, 2), (1, 2, 3)}, we compute
0
baj
Z
:=
0
bbj
Z
:=
n
1 X
n
n
1 X
n
j=1
n
1 X
n
Zcj ? Zbj
Zcj ? Zaj
n
?1
Zaj ,
(2a)
Zbj ,
(2b)
j=1
n
1 X
j=1
Zaj ? Zbj
Zbj ? Zaj
?1
j=1
n
X
0
0
c2 := 1
baj
bbj
M
Z
?Z
,
n j=1
(2c)
n
X
0
0
c3 := 1
baj
bbj
M
Z
?Z
? Zcj .
n j=1
(2d)
The statement of Proposition 1 suggests that we can recover the columns of Cc and the diagonal
c2 and M
c3 . This is implemented by the tensor facentries of W by operating on the moments M
torization method in Algorithm 1. In particular, the tensor factorization algorithm returns a set of
vectors {(b
?h , w
bh ) : h = 1, . . . , k}, where each (b
?h , w
bh ) estimates a particular column of Cc (for
some ?cl ) and a particular diagonal entry of W (for some wl ). It is important to note that the tensor
factorization algorithm doesn?t provide a one-to-one correspondence between the recovered column and the true columns of Cc . Thus, ?
b1 , . . . , ?
bk represents an arbitrary permutation of the true
columns.
To discover the index correspondence, we take each ?
bh and examine its greatest component. We
assume that within each group, the probability of assigning a correct label is always greater than
the probability of assigning any specific incorrect label. This assumption will be made precise
in the next section. As a consequence, if ?
bh corresponds to the l-th column of Cc , then its l-th
bc to
coordinate is expected to be greater than other coordinates. Thus, we set the l-th column of C
some vector ?
bh whose l-th coordinate has the greatest component (if there are multiple such vectors,
then randomly select one of them; if there is no such vector, then randomly select a ?
bh ). Then, we
c to the scalar w
set the l-th diagonal entry of W
bh associated with ?
bh . Note that by iterating over
b for c = 1, 2, 3 respectively. There will be
(a, b, c) ? {(2, 3, 1), (3, 1, 2), (1, 2, 3)}, we obtain C
c
c
three copies of W estimating the same matrix W ?we average them for the best accuracy.
In the second step, we estimate each individual confusion matrix Ci . The following proposition
shows that we can recover Ci from the moments of {zij }. See [24] for the proof.
Proposition 2. For any g ? {1, 2, 3} and any i ? Gg , let a ? {1, 2, 3}\{g} be one of the remaining
group index. Then
T
?i Ci W (Ca )T = E[zij Zaj
].
bi using the empirical approximation
Proposition 2 suggests a plug-in estimator for Ci . We compute C
T
b c
b
of E[zij Zaj ] and using the matrices Ca , Cb , W obtained in the first step. Concretely, we calculate
?
?
n
? 1 X
?1 ?
T
bi := normalize
c (C
ba )T
zij Zaj
W
,
(3)
C
? n
?
j=1
4
where the normalization operator rescales the matrix columns, making sure that each column sums
to one. The overall procedure for Stage 1 is summarized in Algorithm 1.
4.2
Stage 2: EM algorithm
The second stage is devoted to refining the initial estimate provided by Stage 1. The joint likelihood
of true label yj and observed labels zij , as a function of confusion matrices ?i , can be written as
L(?; y, z) :=
n Y
m Y
k
Y
(?iyj c )I(zij =ec ) .
j=1 i=1 c=1
By assuming
a uniform prior over y, we maximize the marginal log-likelihood function `(?) :=
P
log( y?[k]n L(?; y, z)). We refine the initial estimate of Stage 1 by maximizing the objective function, which is implemented by the Expectation Maximization (EM) algorithm. The EM algorithm
takes the values {b
?ilc } provided as output by Stage 1 as initialization, then executes the following
E-step and M-step for at least one round.
E-step Calculate the expected value of the log-likelihood function, with respect to the conditional
distribution of y given z under the current estimate of ?:
( k
!)
n
m Y
k
X
X
Y
I(zij =ec )
qbjl log
(?ilc )
,
Q(?) := Ey|zf,b? [log(L(?; y, z))] =
j=1
where
qbjl ? Pk
exp
l0 =1
i=1 c=1
l=1
Pm Pk
exp
?ilc )
i=1
c=1 I(zij = ec ) log(b
Pm Pk
?il0 c )
i=1
c=1 I(zij = ec ) log(b
for j ? [n], l ? [k].
(4)
M-step Find the estimate ?
b that maximizes the function Q(?):
Pn
bjl I(zij = ec )
j=1 q
for i ? [m], l ? [k], c ? [k].
?
bilc ? Pk Pn
bjl I(zij = ec0 )
c0 =1
j=1 q
(5)
In practice, we alternatively execute the updates (4) and (5), for one iteration or until convergence.
Each update increases the objective function `(?). Since `(?) is not concave, the EM update doesn?t
guarantee converging to the global maximum. It may converge to distinct local stationary points for
different initializations. Nevertheless, as we prove in the next section, it is guaranteed that the EM
algorithm will output statistically optimal estimates of true labels and worker confusion matrices if
it is initialized by Algorithm 1.
5
Convergence Analysis
To state our main theoretical results, we first need to introduce some notation and assumptions. Let
wmin := min{wl }kl=1 and ?min := min{?i }m
i=1
be the smallest portion of true labels and the most extreme sparsity level of workers. Our first
assumption assumes that both wmin and ?min are strictly positive, that is, every class and every
worker contributes to the dataset.
Our second assumption assumes that the confusion matrices for each of the three groups, namely
C1 , C2 and C3 , are nonsingular. As a consequence, if we define matrices Sab and tensors Tabc for
any a, b, c ? {1, 2, 3} as
Sab :=
k
X
wl ?al ? ?bl = Ca W (Cb )T
and
l=1
Tabc :=
k
X
wl ?al ? ?bl ? ?cl ,
l=1
then there will be a positive scalar ?L such that ?k (Sab ) ? ?L > 0.
Our third assumption assumes that within each group, the average probability of assigning a correct
label is always higher than the average probability of assigning any incorrect label. To make this
5
statement rigorous, we define a quantity
? := min
min min {?gll ? ?glc }
g?{1,2,3} l?[k] c?[k]\{l}
indicating the smallest gap between diagonal entries and non-diagonal entries in the same confusion
matrix column. The assumption requires ? being strictly positive. Note that this assumption is
group-based, thus does not assume the accuracy of any individual worker.
Finally, we introduce a quantity that measures the average ability ofP
workers in identifying distinct
labels. For two discrete distributions P and Q, let DKL (P, Q) := i P (i) log(P (i)/Q(i)) represent the KL-divergence between P and Q. Since each column of the confusion matrix represents a
discrete distribution, we can define the following quantity:
m
1 X
D = min0
?i DKL (?il , ?il0 ) .
(6)
l6=l m
i=1
The quantity D lower bounds the averaged KL-divergence between two columns. If D is strictly
positive, it means that every pair of labels can be distinguished by at least one subset of workers. As
the last assumption, we assume that D is strictly positive.
The following two theorems characterize the performance of our algorithm. We split the convergence analysis into two parts. Theorem 1 characterizes the performance of Algorithm 1, providing
sufficient conditions for achieving an arbitrarily accurate initialization. We provide the proof of
Theorem 1 in the long version of this paper [24].
n
o
Theorem 1. For any scalar ? > 0 and any scalar satisfying ? min ?min36?k
,
2
, if the
wmin ?L
number of items n satisfies
5
k log((k + m)/?)
n=?
,
2 w 2 ? 13
2 ?min
min L
then the confusion matrices returned by Algorithm 1 are bounded as
bi ? Ci k? ?
kC
for all i ? [m],
with probability at least 1 ? ?. Here, k ? k? denotes the element-wise `? -norm of a matrix.
Theorem 2 characterizes the error rate in Stage 2. It states that when a sufficiently accurate
initialization is taken, the updates (4) and (5) refine the estimates ?
b and yb to the optimal accuracy.
See the long version of this paper [24] for the proof.
Theorem 2. Assume that there is a positive scalar ? such that ?ilc ? ? for all (i, l, c) ? [m] ? [k]2 .
bi are initialized in a manner such that
For any scalar ? > 0, if confusion matrices C
bi ? Ci k? ? ? := min ? , ?D
kC
for all i ? [m],
(7)
2 16
and the number of workers m and the number of items n satisfy
log(mk/?)
log(1/?) log(kn/?) + log(mn)
and n = ?
,
m=?
?min wmin ?2
D
then, for ?
b and qb obtained by iterating (4) and (5) (for at least one round), with probability at least
1 ? ?,
(a) Letting ybj = arg maxl?[k] qbjl , we have that ybj = yj holds for all j ? [n].
(b) kb
?il ? ?il k22 ?
48 log(2mk/?)
?i wl n
holds for all (i, l) ? [m] ? [k].
In Theorem 2, the assumption that all confusion matrix entries are lower bounded by ? > 0 is
somewhat restrictive. For datasets violating this assumption, we enforce positive confusion matrix
entries by adding random noise: Given any observed label zij , we replace it by a random label in
{1, ..., k} with probability k?. In this modified model, every entry of the confusion matrix is lower
bounded by ?, so that Theorem 2 holds. The random noise makes the constant D smaller than its
original value, but the change is minor for small ?.
6
Dataset name
Bird
RTE
TREC
Dog
Web
# classes
2
2
2
4
5
# items
108
800
19,033
807
2,665
# workers
39
164
762
52
177
# worker labels
4,212
8,000
88,385
7,354
15,567
Table 1: Summary of datasets used in the real data experiment.
To see the consequence of the convergence analysis, we take error rate in Theorem 1 equal to the
constant ? defined in Theorem 2. Then we combine the statements of the two theorems. This shows
that if we choose the number
n such that
of workers m andthe number of items
5
k
1
e
e
and n = ?
;
(8)
m=?
2 w 2 ? 13 min{?2 , (?D)2 }
D
?min
min L
that is, if both m and n are lower bounded by a problem-specific constant and logarithmic terms,
then with high probability, the predictor yb will be perfectly accurate, and the estimator ?
b will be
e
bounded as kb
?il ? ?il k22 ? O(1/(?
w
n)).
To
show
the
optimality
of
this
convergence
rate, we
i l
present the following minimax lower bounds. Again, see [24] for the proof.
Theorem 3. There are universal constants c1 > 0 and c2 > 0 such that:
(a) For any {?ilc }, {?i } and any number of items n, if the number of workers m ? 1/(4D), then
n
i
hX
I(b
yj 6= yj ){?ilc }, {?i }, y = v ? c1 n.
inf sup E
y
b v?[k]n
j=1
(b) For any {wl }, {?i }, any worker-item pair (m, n) and any pair of indices (i, l) ? [m] ? [k], we
have
h
i
1
2
inf
sup E kb
?il ? ?il k2 {wl }, {?i } ? c2 min 1,
.
?
b ??Rm?k?k
?i w l n
In part (a) of Theorem 3, we see that the number of workers should be at least 1/(4D), otherwise
any predictor will make many mistakes. This lower bound matches our sufficient condition on the
number of workers m (see Eq. (8)). In part (b), we see that the best possible estimate for ?il has
?(1/(?i wl n)) mean-squared error. It verifies the optimality of our estimator ?
bil . It is worth noting
that the constraint on the number of items n (see Eq. (8)) might be improvable. In real datasets we
usually have n m so that the optimality for m is more important than for n.
It is worth contrasting our convergence rate with existing algorithms. Ghosh et al. [11] and Dalvi et
al. [7] proposed consistent estimators for the binary one-coin model. To attain an error rate ?, their
algorithms require m and n scaling with 1/? 2 , while our algorithm only requires m and n scaling
with log(1/?). Karger et al. [15, 14] proposed algorithms for both binary and multi-class problems.
Their algorithm assumes that workers are assigned by a random regular graph. Moreover, their
analysis assumes that the limit of number of items goes to infinity, or that the number of workers is
many times the number of items. Our algorithm no longer requires these assumptions.
We also compare our algorithm with the majority voting estimator, where the true label is simply
estimated by a majority vote among workers. Gao and Zhou [10] showed that if there are many
spammers and few experts, the majority voting estimator gives almost a random guess. In cone
to guarantee good performance. Since mD is the
trast, our algorithm only requires mD = ?(1)
aggregated KL-divergence, a small number of experts are sufficient to ensure it is large enough.
6
Experiments
In this section, we report the results of empirical studies comparing the algorithm we propose in
Section 4 (referred to as Opt-D&S) with a variety of existing methods which are also based on the
generative model of Dawid and Skene. Specifically, we compare to the Dawid & Skene estimator
7
0.35
Opt?D&S: 1st iteration
Opt?D&S: 50th iteration
MV?D&S: 1st iteration
MV?D&S: 50th iteration
0.2
Label prediction error
Label prediction error
0.2
Opt?D&S: 1st iteration
Opt?D&S: 50th iteration
MV?D&S: 1st iteration
MV?D&S: 50th iteration
0.21
0.18
0.16
0.14
0.12
0.1
Opt?D&S: 1st iteration
Opt?D&S: 50th iteration
MV?D&S: 1st iteration
MV?D&S: 50th iteration
0.3
Label prediction error
0.22
0.19
0.18
0.17
0.25
0.2
0.16
0.15
0.08
0.15
?6
10
?5
10
?4
10
?3
10
?2
10
?1
10
10
?6
?5
10
?4
10
?3
10
?2
10
10
?1
?6
10
?5
10
?4
10
?3
10
Threshold
Threshold
Threshold
(a) RTE
(b) Dog
(c) Web
?2
10
?1
10
Figure 1: Comparing MV-D&S and Opt-D&S with different thresholding parameter ?. The label
prediction error is plotted after the 1st EM update and after convergence.
Bird
RTE
TREC
Dog
Web
Opt-D&S
10.09
7.12
29.80
16.89
15.86
MV-D&S
11.11
7.12
30.02
16.66
15.74
Majority Voting
24.07
10.31
34.86
19.58
26.93
KOS
11.11
39.75
51.96
31.72
42.93
Ghosh-SVD
27.78
49.13
42.99
?
?
EigenRatio
27.78
9.00
43.96
?
?
Table 2: Error rate (%) in predicting true labels on real data.
initialized by majority voting (referred to as MV-D&S), the pure majority voting estimator, the
multi-class labeling algorithm proposed by Karger et al. [14] (referred to as KOS), the SVD-based
algorithm proposed by Ghosh et al. [11] (referred to as Ghost-SVD) and the ?Eigenvalues of Ratio?
algorithm proposed by Dalvi et al. [7] (referred to as EigenRatio). The evaluation is made on five
real datasets.
We compare the crowdsourcing algorithms on three binary tasks and two multi-class tasks. Binary
tasks include labeling bird species [22] (Bird dataset), recognizing textual entailment [21] (RTE
dataset) and assessing the quality of documents in the TREC 2011 crowdsourcing track [16] (TREC
dataset). Multi-class tasks include labeling the breed of dogs from ImageNet [9] (Dog dataset) and
judging the relevance of web search results [26] (Web dataset). The statistics for the five datasets
are summarized in Table 1. Since the Ghost-SVD algorithm and the EigenRatio algorithm work on
binary tasks, they are evaluated only on the Bird, RTE and TREC datasets. For the MV-D&S and
the Opt-D&S methods, we iterate their EM steps until convergence.
Since entries of the confusion matrix are positive, we find it helpful to incorporate this prior knowledge into the initialization stage of the Opt-D&S algorithm. In particular, when estimating the confusion matrix entries by Eq. (3), we add an extra checking step before the normalization, examining
if the matrix components are greater than or equal to a small threshold ?. For components that are
smaller than ?, they are reset to ?. The default choice of the thresholding parameter is ? = 10?6 .
Later, we will compare the Opt-D&S algorithm with respect to different choices of ?. It is important to note that this modification doesn?t change our theoretical result, since the thresholding is not
needed in case that the initialization error is bounded by Theorem 1.
Table 2 summarizes the performance of each method. The MV-D&S and the Opt-D&S algorithms
consistently outperform the other methods in predicting the true label of items. The KOS algorithm,
the Ghost-SVD algorithm and the EigenRatio algorithm yield poorer performance, presumably due
to the fact that they rely on idealized assumptions that are not met by the real data. In Figure 1, we
compare the Opt-D&S algorithm with respect to different thresholding parameters ? ? {10?i }6i=1 .
We plot results for three datasets (RET, Dog, Web), where the performance of MV-D&S is equal to or
slightly better than that of Opt-D&S. The plot shows that the performance of the Opt-D&S algorithm
is stable after convergence. But at the first EM iterate, the error rates are more sensitive to the choice
of ?. A proper choice of ? makes Opt-D&S outperform MV-D&S. The result suggests that a
proper initialization combined with one EM iterate is good enough for the purposes of prediction.
In practice, the best choice of ? can be obtained by cross validation.
8
References
[1] A. Anandkumar, D. P. Foster, D. Hsu, S. M. Kakade, and Y.-K. Liu. A spectral algorithm for latent
Dirichlet allocation. arXiv preprint: 1204.6703, 2012.
[2] A. Anandkumar, R. Ge, D. Hsu, and S. M. Kakade. A tensor spectral approach to learning mixed membership community models. In Annual Conference on Learning Theory, 2013.
[3] A. Anandkumar, R. Ge, D. Hsu, S. M. Kakade, and M. Telgarsky. Tensor decompositions for learning
latent variable models. arXiv preprint:1210.7559, 2012.
[4] A. Anandkumar, D. Hsu, and S. M. Kakade. A method of moments for mixture models and hidden
Markov models. In Annual Conference on Learning Theory, 2012.
[5] A. T. Chaganty and P. Liang. Spectral experts for estimating mixtures of linear regressions. arXiv preprint:
1306.3729, 2013.
[6] X. Chen, Q. Lin, and D. Zhou. Optimistic knowledge gradient policy for optimal budget allocation in
crowdsourcing. In Proceedings of ICML, 2013.
[7] N. Dalvi, A. Dasgupta, R. Kumar, and V. Rastogi. Aggregating crowdsourced binary ratings. In Proceedings of World Wide Web Conference, 2013.
[8] A. P. Dawid and A. M. Skene. Maximum likelihood estimation of observer error-rates using the EM
algorithm. Journal of the Royal Statistical Society, Series C, pages 20?28, 1979.
[9] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. Imagenet: A large-scale hierarchical image
database. In IEEE CVPR, 2009.
[10] C. Gao and D. Zhou. Minimax optimal convergence rates for estimating ground truth from crowdsourced
labels. arXiv preprint arXiv:1310.5764, 2014.
[11] A. Ghosh, S. Kale, and P. McAfee. Who moderates the moderators? crowdsourcing abuse detection in
user-generated content. In Proceedings of the ACM Conference on Electronic Commerce, 2011.
[12] D. Hsu, S. M. Kakade, and T. Zhang. A spectral algorithm for learning hidden Markov models. Journal
of Computer and System Sciences, 78(5):1460?1480, 2012.
[13] P. Jain and S. Oh. Learning mixtures of discrete product distributions using spectral decompositions.
arXiv preprint:1311.2972, 2013.
[14] D. R. Karger, S. Oh, and D. Shah. Efficient crowdsourcing for multi-class labeling. In ACM SIGMETRICS, 2013.
[15] D. R. Karger, S. Oh, and D. Shah. Budget-optimal task allocation for reliable crowdsourcing systems.
Operations Research, 62(1):1?24, 2014.
[16] M. Lease and G. Kazai. Overview of the TREC 2011 crowdsourcing track. In Proceedings of TREC 2011,
2011.
[17] E. Lehmann and G. Casella. Theory of Point Estimation. Springer, 2nd edition, 2003.
[18] P. Liang. Partial information from spectral methods. NIPS Spectral Learning Workshop, 2013.
[19] Q. Liu, J. Peng, and A. T. Ihler. Variational inference for crowdsourcing. In NIPS, 2012.
[20] V. C. Raykar, S. Yu, L. H. Zhao, G. H. Valadez, C. Florin, L. Bogoni, and L. Moy. Learning from crowds.
Journal of Machine Learning Research, 11:1297?1322, 2010.
[21] R. Snow, B. O?Connor, D. Jurafsky, and A. Y. Ng. Cheap and fast?but is it good? evaluating non-expert
annotations for natural language tasks. In Proceedings of EMNLP, 2008.
[22] P. Welinder, S. Branson, S. Belongie, and P. Perona. The multidimensional wisdom of crowds. In NIPS,
2010.
[23] J. Whitehill, P. Ruvolo, T. Wu, J. Bergsma, and J. R. Movellan. Whose vote should count more: Optimal
integration of labels from labelers of unknown expertise. In NIPS, 2009.
[24] Y. Zhang, X. Chen, D. Zhou, and M. I. Jordan. Spectral methods meet EM: A provably optimal algorithm
for crowdsourcing. arXiv preprint arXiv:1406.3824, 2014.
[25] D. Zhou, Q. Liu, J. C. Platt, and C. Meek. Aggregating ordinal labels from crowds by minimax conditional
entropy. In Proceedings of ICML, 2014.
[26] D. Zhou, J. C. Platt, S. Basu, and Y. Mao. Learning from the wisdom of crowds by minimax entropy. In
NIPS, 2012.
[27] J. Zou, D. Hsu, D. Parkes, and R. Adams. Contrastive learning using spectral methods. In NIPS, 2013.
9
| 5431 |@word mild:1 version:2 norm:1 nd:1 c0:1 decomposition:5 contrastive:1 moment:14 initial:5 liu:3 series:1 karger:5 zij:18 denoting:1 bc:3 document:1 outperforms:1 existing:3 recovered:1 com:1 current:1 comparing:2 yet:1 assigning:4 written:1 refines:1 partition:1 cheap:1 plot:2 update:5 stationary:1 generative:3 guess:1 item:26 ruvolo:1 parkes:1 provides:2 zhang:3 five:2 c2:7 become:1 ik:1 incorrect:2 consists:1 prove:1 combine:1 dalvi:4 manner:1 introduce:2 theoretically:2 peng:1 expected:2 examine:1 multi:8 inspired:1 provided:3 estimating:9 moreover:2 discover:1 maximizes:2 notation:1 advent:1 bounded:6 what:1 ec0:1 eigenvector:1 developed:1 contrasting:1 ret:1 ghosh:5 unobserved:1 guarantee:4 berkeley:3 every:5 collecting:1 voting:5 concave:1 multidimensional:1 shed:1 k2:1 rm:1 platt:2 partitioning:1 positive:9 service:2 before:1 local:2 aggregating:2 mistake:1 consequence:3 limit:1 meet:2 abuse:1 might:1 bird:5 initialization:13 collect:1 suggests:3 yuczhang:1 jurafsky:1 gll:1 factorization:3 branson:1 bi:7 statistically:1 averaged:3 commerce:1 yj:8 practice:3 rte:5 movellan:1 procedure:3 empirical:7 universal:1 attain:1 regular:2 get:1 operator:1 bh:7 context:1 maximizing:2 go:1 kale:1 starting:1 convex:2 amazon:1 identifying:1 assigns:1 pure:1 m2:2 estimator:17 oh:3 glc:1 coordinate:4 suppose:1 user:1 us:1 dawid:11 element:1 satisfying:2 approximated:1 labeled:2 database:1 observed:6 preprint:6 initializing:1 calculate:3 raise:1 depend:2 basis:1 joint:1 distinct:2 jain:1 fast:1 labeling:16 crowd:6 whose:3 widely:1 solve:1 cvpr:1 relax:1 otherwise:1 ability:2 statistic:1 g1:4 breed:1 jointly:1 noisy:2 online:1 eigenvalue:2 propose:8 product:1 reset:1 achieve:2 kh:1 normalize:1 scalability:1 convergence:13 empty:2 optimum:1 assessing:1 generating:1 telgarsky:1 adam:1 dengyong:2 minor:1 eq:6 implemented:2 indicate:1 met:1 snow:1 min0:1 correct:2 kb:3 require:1 hx:1 suffices:1 proposition:7 opt:17 strictly:4 hold:3 sufficiently:1 ground:1 exp:2 presumably:1 cb:2 achieves:3 smallest:2 purpose:1 estimation:5 label:51 sensitive:1 largest:1 wl:13 rough:1 always:2 sigmetrics:1 aim:1 modified:1 zhou:9 pn:2 sab:3 derived:1 l0:1 refining:1 consistently:1 likelihood:8 contrast:2 zcj:6 cg:2 rigorous:1 helpful:1 inference:2 membership:1 bt:1 her:1 kc:2 hidden:2 perona:1 i1:1 provably:3 arg:1 overall:1 among:1 special:1 initialize:1 integration:1 marginal:1 equal:3 ng:1 represents:4 yu:1 icml:2 report:1 few:1 bil:1 randomly:4 simultaneously:1 divergence:3 individual:3 microsoft:3 detection:1 evaluation:1 mixture:3 extreme:1 light:1 devoted:1 accurate:4 poorer:1 worker:45 partial:1 il2:1 orthogonal:1 conduct:1 yuchen:1 initialized:3 plotted:1 tabc:2 theoretical:7 mk:2 column:14 maximization:2 entry:13 subset:1 uniform:1 predictor:2 recognizing:1 examining:1 welinder:1 characterize:1 kn:1 synthetic:1 combined:1 st:7 fundamental:1 dong:1 michael:1 concrete:1 w1:1 again:1 squared:1 choose:1 emnlp:1 worse:1 expert:6 resort:1 iyj:1 return:1 zhao:1 li:2 valadez:1 summarized:2 wk:1 availability:1 rescales:1 satisfy:1 explicitly:1 mv:13 idealized:1 later:1 root:3 observer:1 optimistic:1 characterizes:2 portion:2 sup:2 recover:2 crowdsourced:2 annotation:1 il:10 accuracy:5 improvable:1 who:1 efficiently:1 yield:1 nonsingular:1 rastogi:1 wisdom:2 bayesian:1 worth:2 cc:4 expertise:1 zaj:14 executes:1 moderator:1 baj:3 casella:1 turk:1 associated:2 proof:4 ihler:1 sampled:1 hsu:6 dataset:7 knowledge:2 higher:1 violating:1 yb:2 entailment:1 execute:1 evaluated:1 kazai:1 stage:20 until:2 web:7 quality:2 name:1 k22:2 true:19 remedy:1 assigned:3 symmetric:1 i2:1 attractive:1 round:2 raykar:1 gg:4 outline:1 demonstrate:2 confusion:29 performs:1 ilc:7 image:1 wise:1 variational:1 recently:1 empirically:1 overview:1 extend:1 numerically:1 connor:1 imposing:1 chaganty:1 pm:2 language:1 had:1 stable:1 longer:1 operating:1 whitening:1 add:1 labelers:1 bergsma:1 recent:2 showed:1 optimizing:1 inf:2 moderate:1 scenario:1 certain:1 outperforming:1 success:1 binary:10 arbitrarily:1 greater:3 somewhat:1 ey:1 converting:1 deng:1 aggregated:3 maximize:1 redundant:1 converge:1 multiple:3 full:1 infer:1 reduces:1 match:1 plug:1 cross:1 raphson:3 long:2 ofp:1 lin:1 dkl:2 converging:1 prediction:5 basic:1 ko:3 whitened:1 regression:1 expectation:3 arxiv:8 iteration:15 represent:2 sometimes:1 normalization:2 c1:3 want:1 singular:2 w2:1 extra:1 sure:1 jordan:3 integer:2 anandkumar:5 noting:1 intermediate:1 split:1 enough:2 mcafee:1 variety:1 iterate:3 florin:1 xichen:1 perfectly:1 idea:1 moy:1 returned:1 spammer:1 york:2 iterating:2 category:1 outperform:2 canonical:1 judging:1 estimated:2 trapped:1 correctly:1 disjoint:2 track:2 write:2 discrete:3 dasgupta:1 group:10 redundancy:1 nevertheless:1 threshold:4 achieving:1 utilize:1 asymptotically:1 graph:2 sum:1 cone:1 parameterized:1 lehmann:1 family:2 throughout:1 almost:1 electronic:1 wu:1 summarizes:1 scaling:2 comparable:1 bound:6 guaranteed:1 meek:1 correspondence:2 refine:2 ilk:1 annual:2 constraint:2 deficiency:1 infinity:1 fei:2 optimality:4 min:16 qb:1 performing:2 kumar:1 skene:11 according:1 poor:1 smaller:2 slightly:1 em:25 appealing:1 g3:2 kakade:5 making:1 modification:1 taken:1 computationally:3 bbj:3 turn:1 count:1 fail:1 needed:1 ordinal:1 letting:1 ge:2 operation:1 hierarchical:1 spectral:16 enforce:1 distinguished:1 alternative:1 coin:5 shah:2 original:1 assumes:7 denotes:3 remaining:1 ensure:1 include:2 dirichlet:1 newton:3 l6:1 restrictive:2 especially:1 establish:1 society:1 bl:2 tensor:12 objective:3 quantity:4 strategy:1 dependence:1 md:2 diagonal:6 gradient:1 majority:6 collected:1 considers:1 assuming:2 besides:1 index:4 providing:1 ratio:1 liang:2 setup:1 statement:3 gk:2 whitehill:1 ba:1 design:1 proper:2 policy:1 unknown:1 perform:1 upper:1 zf:1 observation:1 datasets:8 markov:2 finite:1 immediate:1 incorporate:2 precise:1 trec:7 arbitrary:1 community:1 rating:1 pair:4 mechanical:1 kl:4 extensive:1 c3:5 namely:1 dog:6 imagenet:2 california:1 textual:1 nip:6 address:2 redmond:1 suggested:1 usually:1 ghost:3 sparsity:2 royal:1 reliable:1 power:2 greatest:3 demanding:1 treated:1 difficulty:2 circumvent:1 predicting:2 rely:1 natural:1 mn:1 minimax:7 torization:1 prior:3 understanding:1 checking:1 bjl:2 permutation:3 mixed:1 allocation:3 validation:1 sufficient:3 consistent:4 principle:1 thresholding:4 foster:1 summary:1 gl:1 last:1 copy:1 wide:1 basu:1 taking:1 wmin:4 maxl:1 default:1 world:1 evaluating:1 doesn:4 concretely:1 made:2 simplified:1 ec:7 unreliable:1 global:2 assumed:1 belongie:1 xi:1 alternatively:1 search:1 latent:4 iterative:1 table:4 robust:3 ca:4 operational:1 obtaining:1 contributes:1 cl:7 zou:1 diag:1 pk:6 main:1 ybj:2 linearly:1 noise:2 edition:1 verifies:1 referred:6 ny:1 il1:1 inferring:1 mao:1 exponential:2 third:3 theorem:15 rk:10 specific:3 nyu:1 virtue:1 workshop:1 socher:1 adding:1 ci:7 budget:2 chen:3 gap:1 entropy:3 logarithmic:3 simply:1 gao:2 bogoni:1 g2:4 scalar:6 il0:2 springer:1 corresponds:1 truth:1 satisfies:1 acm:2 conditional:2 goal:1 replace:1 content:1 hard:2 change:2 specifically:1 justify:3 averaging:1 called:1 specie:1 experimental:1 svd:8 m3:2 vote:2 indicating:1 select:2 relevance:1 zbj:11 crowdsourcing:20 |
4,896 | 5,432 | Unsupervised Transcription of Piano Music
Taylor Berg-Kirkpatrick Jacob Andreas Dan Klein
Computer Science Division
University of California, Berkeley
{tberg,jda,klein}@cs.berkeley.edu
Abstract
We present a new probabilistic model for transcribing piano music from audio to
a symbolic form. Our model reflects the process by which discrete musical events
give rise to acoustic signals that are then superimposed to produce the observed
data. As a result, the inference procedure for our model naturally resolves the
source separation problem introduced by the the piano?s polyphony. In order to
adapt to the properties of a new instrument or acoustic environment being transcribed, we learn recording-specific spectral profiles and temporal envelopes in an
unsupervised fashion. Our system outperforms the best published approaches on
a standard piano transcription task, achieving a 10.6% relative gain in note onset
F1 on real piano audio.
1
Introduction
Automatic music transcription is the task of transcribing a musical audio signal into a symbolic representation (for example MIDI or sheet music). We focus on the task of transcribing piano music,
which is potentially useful for a variety of applications ranging from information retrieval to musicology. This task is extremely difficult for multiple reasons. First, even individual piano notes are
quite rich. A single note is not simply a fixed-duration sine wave at an appropriate frequency, but
rather a full spectrum of harmonics that rises and falls in intensity. These profiles vary from piano
to piano, and therefore must be learned in a recording-specific way. Second, piano music is generally polyphonic, i.e. multiple notes are played simultaneously. As a result, the harmonics of the
individual notes can and do collide. In fact, combinations of notes that exhibit ambiguous harmonic
collisions are particularly common in music, because consonances sound pleasing to listeners. This
polyphony creates a source-separation problem at the heart of the transcription task.
In our approach, we learn the timbral properties of the piano being transcribed (i.e. the spectral and
temporal shapes of each note) in an unsupervised fashion, directly from the input acoustic signal. We
present a new probabilistic model that describes the process by which discrete musical events give
rise to (separate) acoustic signals for each keyboard note, and the process by which these signals are
superimposed to produce the observed data. Inference over the latent variables in the model yields
transcriptions that satisfy an informative prior distribution on the discrete musical structure and at
the same time resolve the source-separation problem.
For the problem of unsupervised piano transcription where the test instrument is not seen during
training, the classic starting point is a non-negative factorization of the acoustic signal?s spectrogram. Most previous work improves on this baseline in one of two ways: either by better modeling
the discrete musical structure of the piece being transcribed [1, 2] or by better adapting to the timbral properties of the source instrument [3, 4]. Combining these two kinds of approaches has proven
challenging. The standard approach to modeling discrete musical structures?using hidden Markov
or semi-Markov models?relies on the availability of fast dynamic programs for inference. Here,
coupling these discrete models with timbral adaptation and source separation breaks the conditional
independence assumptions that the dynamic programs rely on. In order to avoid this inference
problem, past approaches typically defer detailed modeling of discrete structure or timbre to a postprocessing step [5, 6, 7].
1
Event Params
PLAY
?(n)
M (nr)
velocity
Note Events
REST
velocity
duration
time
Envelope Params
Note Activation
?
(n)
A
(nr)
time
time
Component Spectrogram
(n)
S (nr)
freq
freq
Spectral Params
time
N
freq
Spectrogram
X (r)
R
time
Figure 1: We transcribe a dataset consisting of R songs produced by a single piano with N notes. For each
keyboard note, n, and each song, r, we generate a sequence of musical events, M (nr) , parameterized by ?(n) .
Then, conditioned on M (nr) , we generate an activation time series, A(nr) , parameterized by ?(n) . Next,
conditioned on A(nr) , we generate a component spectrogram for note n in song r, S (nr) , parameterized by
? (n) . The observed total spectrogram for song r is produced by superimposing component spectrograms:
P
X (r) = n S (nr) .
We present the first approach that tackles these discrete and timbral modeling problems jointly. We
have two primary contributions: first, a new generative model that reflects the causal process underlying piano sound generation in an articulated way, starting with discrete musical structure; second,
a tractable approximation to the inference problem over transcriptions and timbral parameters. Our
approach achieves state-of-the-art results on the task of polyphonic piano music transcription. On
a standard dataset consisting of real piano audio data, annotated with ground-truth onsets, our approach outperforms the best published models for this task on multiple metrics, achieving a 10.6%
relative gain in our primary measure of note onset F1 .
2
Model
Our model is laid out in Figure 1. It has parallel random variables for each note on the piano keyboard. For now, we illustrate these variables for a single concrete note?say C] in the 4th octave?
and in Section 2.4 describe how the parallel components combine to produce the observed audio
signal. Consider a single song, divided into T time steps. The transcription will be I musical events
long. The component model for C] consists of three primary random variables:
M , a sequence of I symbolic musical events, analogous to the locations
and values of symbols along the C] staff line in sheet music,
A, a time series of T activations, analogous to the loudness of sound emitted by the C] piano string over time as it peaks and attenuates during each
event in M ,
S, a spectrogram of T frames, specifying the spectrum of frequencies over
time in the acoustic signal produced by the C] string.
2
M (nr)
Envelope Params
?(n)
Event State
E2
E1
E3
Truncate to duration Di
Scale to velocity Vi
Add noise
Duration
Velocity
D3
D2
D1
V1
V2
V3
A(nr)
Figure 2: Joint distribution on musical events, M (nr) , and activations, A(nr) , for note n in song r, conditioned
on event parameters, ?(n) , and envelope parameters, ?(n) . The dependence of Ei , Di , and Vi on n and r is
suppressed for simplicity.
The parameters that generate each of these random variables are depicted in Figure 1. First, musical
]
events, M , are generated from a distribution parameterized by ?(C ) , which specifies the probability
]
that the C key is played, how long it is likely to be held for (duration), and how hard it is likely to
be pressed (velocity). Next, the activation of the C] string over time, A, is generated conditioned on
]
M from a distribution parameterized by a vector, ?(C ) (see Figure 1), which specifies the shape of
the rise and fall of the string?s activation each time the note is played. Finally, the spectrogram, S,
]
is generated conditioned on A from a distribution parameterized by a vector, ? (C ) (see Figure 1),
which specifies the frequency distribution of sounds produced by the C] string. As depicted in
]
Figure 3, S is produced from the outer product of ? (C ) and A. The joint distribution for the note1 is:
]
]
]
]
P (S, A, M |? (C ) , ?(C ) , ?(C ) ) = P (M |?(C ) )
? P (A|M, ?
(C] )
]
[Event Model, Section 2.1]
)
? P (S|A, ? (C ) )
[Activation Model, Section 2.2]
[Spectrogram Model, Section 2.3]
In the next three sections we give detailed descriptions for each of the component distributions.
2.1
Event Model
Our symbolic representation of musical structure (see Figure 2) is similar to the MIDI format used by
musical synthesizers. M consists of a sequence of I random variables representing musical events
for the C] piano key: M = (M1 , M2 , . . . , MI ). Each event Mi , is a tuple consisting of a state,
Ei , which is either PLAY or REST, a duration Di , which is a length in time steps, and a velocity Vi ,
which specifies how hard the key was pressed (assuming Ei is PLAY).
The graphical model for the process that generates M is depicted in Figure 2. The sequence of
states, (E1 , E2 , . . . , EI ), is generated from a Markov model. The transition probabilities, ?TRANS ,
control how frequently the note is played (some notes are more frequent than others). An event?s
duration, Di , is generated conditioned on Ei from a distribution parameterized by ?DUR . The durations of PLAY events have a multinomial parameterization, while the durations of REST events are
distributed geometrically. An event?s velocity, Vi , is a real number on the unit interval and is generated conditioned on Ei from a distribution parameterized by ?VEL . If Ei = REST, deterministically
]
Vi = 0. The complete event parameters for keyboard note C] are ?(C ) = (?TRANS , ?DUR , ?VEL ).
1
For notational convenience, we suppress the C] superscripts on M , A, and S until Section 2.4.
3
2.2
Activation Model
In an actual piano, when a key is pressed, a hammer strikes a string and a sound with sharply
rising volume is emitted. The string continues to emit sound as long as the key remains depressed, but the volume decays since no new energy is being transferred. When the key is
released, a damper falls back onto the string, truncating the decay. Examples of this trajectory are depicted in Figure 1 in the graph of activation values. The graph depicts the note being played softly and held, and then being played more loudly, but held for a shorter time.
In our model, PLAY events represent hammer strikes on a piano string with raised damper,
while REST events represent the lowered damper. In our parameterization, the shape of the
rise and decay is shared by all PLAY events for a given note, regardless of their
?
duration and velocity. We call this shape an envelope and describe it using a posi]
tive vector of parameters. For our running example of C , this parameter vector is
]
?(C ) (depicted to the right).
The time series of activations for the C] string, A, is a positive vector of length T , where T denotes
the total length of the song in time steps. Let [A]t be the activation at time step t. As mentioned in
Section 2, A may be thought of as roughly representing the loudness of sound emitted by the piano
string as a function of time. The process that generates A is depicted in Figure 2. We generate A
conditioned on the musical events, M . Each musical event, Mi = (Ei , Di , Vi ), produces a segment
of activations, Ai , of length Di . For PLAY events, Ai will exhibit an increase in activation. For
REST events, the activation will remain low. The segments are appended together to make A. The
activation values in each segment are generated in a way loosely inspired by piano mechanics. Given
]
]
?(C ) , we generate the values in segment Ai as follows: ?(C ) is first truncated to duration Di , then
is scaled by the velocity of the strike, Vi , and, finally, is used to parameterize an activation noise
distribution which generates the segment Ai . Specifically, we add independent Gaussian noise to
]
each dimension after ?(C ) is truncated and scaled. In principle, this choice of noise distribution
gives a formally deficient model, since activations are positive, but in practice performs well and has
the benefit of making inference mathematically simple (see Section 3.1).
2.3
Component Spectrogram Model
Piano sounds have a harmonic structure; when played, each piano string primarily emits
energy at a fundamental frequency determined by the string?s length, but also at all
integer multiples of that frequency (called partials) with diminishing strength (see the depiction to the right). For example, the audio signal produced by the C] string will vary in
loudness, but its frequency distribution will remain mostly fixed. We call this frequency
distribution a spectral profile. In our parameterization, the spectral profile of C] is speci]
]
fied by a positive spectral parameter vector, ? (C ) (depicted to the right). ? (C ) is a vector
]
of length F , where [? (C ) ]f represents the weight of frequency bin f .
In our model, the audio signal produced by the C] string over the course of the song is represented
as a spectrogram, S, which is a positive matrix with F rows, one for each frequency bin, f , and T
columns, one for each time step, t (see Figures 1 and 3 for examples). We denote the magnitude
of frequency f at time step t as [S]f t . In order to generate the spectrogram (see Figure 3), we
first produce a matrix of intermediate values by taking the outer product of the spectral profile,
]
? (C ) , and the activations, A. These intermediate values are used to parameterize a spectrogram
noise distribution that generates S. Specifically, for each frequency bin f and each time step t, the
corresponding value of the spectrogram, [S]f t , is generated from a noise distribution parameterized
]
by [? (C ) ]f ? [A]t . In practice, the choice of noise distribution is very important. After examining
residuals resulting from fitting mean parameters to actual musical spectrograms, we experimented
with various noising assumptions, including multiplicative gamma noise, additive Gaussian noise,
log-normal noise, and Poisson noise. Poisson noise performed best. This is consistent with previous
findings in the literature, where non-negative matrix factorization using KL divergence (which has
a generative interpretation as maximum likelihood inference in a Poisson model [8]) is commonly
chosen [7, 2]. Under the Poisson noise assumption, the spectrogram is interpreted as a matrix of
(large) integer counts.
4
A(1r)
(N )
freq
...
freq
(1)
A(N r)
time
time
S (1r)
S (N r)
freq
X (r)
2.4
Figure 3:
Conditional distribution for song r on the observed
total spectrogram, X (r) , and
the component spectrograms for
each note, (S (1r) , . . . , S (N r) ),
given the activations for each
note, (A(1r) , . . . , A(N r) ), and
spectral parameters for each note,
(? (1) , . . . , ? (N ) ).
X (r) is the
superposition of the component
P
spectrograms: X (r) = n S (nr) .
time
Full Model
So far we have only looked at the component of the model corresponding to a single note?s contribution to a single song. Our full model describes the generation of a collection of many songs, from
a complete instrument with many notes. This full model is diagrammed in Figures 1 and 3. Let a
piano keyboard consist of N notes (on a standard piano, N is 88), where n indexes the particular
note. Each note, n, has its own set of musical event parameters, ?(n) , envelope parameters, ?(n) ,
and spectral parameters, ? (n) . Our corpus consists of R songs (?recordings?), where r indexes a
particular song. Each pair of note n and song r has it?s own musical events variable, M (nr) , activations variable, A(nr) , and component spectrogram S (nr) . The full spectrogram for song r, which is
the observed input, is denoted
as X (r) . Our model generates X (r) by superimposing the component
P (nr)
(r)
spectrograms: X = n S
. Going forward, we will need notation to group together variables
across all N notes: let ? = (?(1) , . . . , ?(N ) ), ? = (?(1) , . . . , ?(N ) ), and ? = (? (1) , . . . , ? (N ) ).
Also let M (r) = (M (1r) , . . . , M (N r) ), A(r) = (A(1r) , . . . , A(N r) ), and S (r) = (S (1r) , . . . , S (N r) ).
3
Learning and Inference
Our goal is to estimate the unobserved musical events for each song, M (r) , as well as the unknown
envelope and spectral parameters of the piano that generated the data, ? and ?. Our inference will
estimate both, though our output is only the musical events, which specify the final transcription.
Because MIDI sample banks (piano synthesizers) are readily available, we are able to provide the
system with samples from generic pianos (but not from the piano being transcribed). We also give
the model information about the distribution of notes in real musical scores by providing it with an
external corpus of symbolic music data. Specifically, the following information is available to the
model during inference: 1) the total spectrogram for each song, X (r) , which is the input, 2) the event
parameters, ?, which we estimate by collecting statistics on note occurrences in the external corpus
of symbolic music, and 3) truncated normal priors on the envelopes and spectral profiles, ? and ?,
which we extract from the MIDI samples.
? = (M (1) , . . . , M (R) ), A? = (A(1) , . . . , A(R) ), and S? = (S (1) , . . . , S (R) ), the tuples of event,
Let M
activation, and spectrogram variables across all notes n and songs r. We would like to compute
? , ?, and ?. However, marginalizing over the activations A? couples
the posterior distribution on M
the discrete musical structure with the superposition process of the component spectrograms in an
? , A,
? ?, and ? via iterated
intractable way. We instead approximate the joint MAP estimates of M
? Specifically, we
conditional modes [9], only marginalizing over the component spectrograms, S.
perform the following optimization via block-coordinate ascent:
"
#
Y X
max
P (X (r) , S (r) , A(r) , M (r) |?, ?, ?) ? P (?, ?)
? ,A,?,?
?
M
r
S (r)
? in Section 3.1,
The updates for each group of variables are described in the following sections: M
? in Section 3.2, A? in Section 3.3, and ? in Section 3.4.
5
3.1
Updating Events
? to maximize the objective while the other variables are held fixed. The joint disWe update M
? and A? is a hidden semi-Markov model [10]. Given the optimal velocity for each
tribution on M
segment of activations, the computation of the emission potentials for the semi-Markov dynamic
? can be performed exactly and efficiently. We let
program is straightforward and the update over M
the distribution of velocities for PLAY events be uniform. This choice, together with the choice of
Gaussian activation noise, yields a simple closed-form solution for the optimal setting of the velocity
(nr)
(nr)
variable Vi
. Let [?(n) ]j denote the jth value of the envelope vector ?(n) . Let [Ai ]j be the jth
(nr)
entry of the segment of A(nr) generated by event Mi . The velocity that maximizes the activation
segment?s probability is given by:
PDi(nr) (n)
(nr)
[?
]
?
[A
]
j
j
j=1
i
(nr)
Vi
=
PDi(nr) (n) 2
]j
j=1 [?
3.2
Updating Envelope Parameters
Given settings of the other variables, we update the envelope parameters, ?, to optimize the objec(nr)
tive. The truncated normal prior on ? admits a closed-form update. Let I(j, n, r) = {i : Di
? j},
(n)
the set of event indices for note n in song r with durations no longer than j time steps. Let [?0 ]j
(n)
be the location parameter for the prior on [? ]j , and let ? be the scale parameter (which is shared
across all n and j). The update for [?(n) ]j is given by:
P
P
(nr)
(nr)
(n)
[Ai ]j + 2?1 2 [?0 ]j
(n,r)
i?I(j,n,r) Vi
(n)
[? ]j =
P
P
(nr) 2
]j + 2?1 2
(n,r)
i?I(j,n,r) [Ai
3.3
Updating Activations
? we optimize the objective with respect to A,
? with the other
In order to update the activations, A,
variables held fixed. The choice of Poisson noise for generating each of the component spectrograms,P
S (nr) , means that the conditional distribution of the total spectrogram for each song,
(r)
(nr)
X
=
, with S (r) marginalized out, is also
nS
Poisson. Specifically, the distribution of
P
(r)
[X ]f t is Poisson with mean n [? (n) ]f ? [A(nr) ]t . Optimizing the probability of X (r) under
this conditional distribution with respect to A(r) corresponds to computing the supervised NMF
using KL divergence [7]. However, to perform the correct update in our model, we must also incorporate the distribution of A(r) , and so, instead of using the standard multiplicative NMF updates,
we use exponentiated gradient ascent [11] on the log objective. Let L denote the log objective, let
?
? (n, r, t) denote the velocity-scaled envelope value used to generate the activation value [A(nr) ]t ,
and let ? 2 denote the variance parameter for the Gaussian activation noise. The gradient of the log
objective with respect to [A(nr) ]t is:
"
#
X
[X (r) ]f t ? [? (n) ]f
1 (nr)
?L
(n)
P
=
?
[?
]
?
[A
]
?
?
?
(n,
r,
t)
f
t
0
0
(n ) ] ? [A(n r) ]
?2
?[A(nr) ]t
f
t
n0 [?
f
3.4
Updating Spectral Parameters
The update for the spectral parameters, ?, is similar to that of the activations. Like the activations, ?
is part of the parameterization of the Poisson distribution on each X (r) . We again use exponentiated
(n)
gradient ascent. Let [?0 ]f be the location parameter of the prior on [? (n) ]f , and let ? be the scale
parameter (which is shared across all n and f ). The gradient of the the log objective with respect to
[? (n) ]f is given by:
"
#
X
?L
[X (r) ]f t ? [A(nr) ]t
1 (n)
(n)
(nr)
P
=
?
[A
]
?
[?
]
?
[?
]
t
f
f
0
(n0 ) ] ? [A(n0 r) ]
?2
?[? (n) ]f
f
t
n0 [?
(r,t)
6
4
Experiments
Because polyphonic transcription is so challenging, much of the existing literature has either worked
with synthetic data [12] or assumed access to the test instrument during training [5, 6, 13, 7]. As
our ultimate goal is the transcription of arbitrary recordings from real, previously-unseen pianos, we
evaluate in an unsupervised setting, on recordings from an acoustic piano not observed in training.
Data We evaluate on the MIDI-Aligned Piano Sounds (MAPS) corpus [14]. This corpus includes
a collection of piano recordings from a variety of time periods and styles, performed by a human
player on an acoustic ?Disklavier? piano equipped with electromechanical sensors under the keys.
The sensors make it possible to transcribe directly into MIDI while the instrument is in use, providing
a ground-truth transcript to accompany the audio for the purpose of evaluation. In keeping with much
of the existing music transcription literature, we use the first 30 seconds of each of the 30 ENSTDkAm
recordings as a development set, and the first 30 seconds of each of the 30 ENSTDkCl recordings
as a test set. We also assume access to a collection of synthesized piano sounds for parameter
initialization, which we take from the MIDI portion of the MAPS corpus, and a large collection of
symbolic music data from the IMSLP library [15, 16], used to estimate the event parameters in our
model.
Preprocessing We represent the input audio as a magnitude spectrum short-time Fourier transform
with a 4096-frame window and a hop size of 512 frames, similar to the approach used by Weninger
et al. [7]. We temporally downsample the resulting spectrogram by a factor of 2, taking the maximum magnitude over collapsed bins. The input audio is recorded at 44.1 kHz and the resulting
spectrogram has 23ms frames.
Initialization and Learning We estimate initializers and priors for the spectral parameters, ?,
and envelope parameters, ?, by fitting isolated, synthesized, piano sounds. We collect these isolated
sounds from the MIDI portion of MAPS, and average the parameter values across several synthesized
pianos. We estimate the event parameters ? by counting note occurrences in the IMSLP data. At
decode time, to fit the spectral and envelope parameters and predict transcriptions, we run 5 iterations
of the block-coordinate ascent procedure described in Section 3.
Evaluation We report two standard measures of performance: an onset evaluation, in which a
predicted note is considered correct if it falls within 50ms of a note in the true transcription, and
a frame-level evaluation, in which each transcription is converted to a boolean matrix specifying
which notes are active at each time step, discretized to 10ms frames. Each entry is compared to
the corresponding entry in the true matrix. Frame-level evaluation is sensitive to offsets as well
as onsets, but does not capture the fact that note onsets have greater musical significance than do
offsets. As is standard, we report precision (P), recall (R), and F1 -measure (F1 ) for each of these
metrics.
4.1
Comparisons
We compare our system to three state-of-the-art unsupervised systems: the hidden semi-Markov
model described by Benetos and Weyde [2] and the spectrally-constrained factorization models described by Vincent et al. [3] and O?Hanlon and Plumbley [4]. To our knowledge, Benetos and Weyde
[2] report the best published onset results for this dataset, and O?Hanlon and Plumbley [4] report the
best frame-level results.
The literature also includes a number of supervised approaches to this task. In these approaches,
a model is trained on annotated recordings from a known instrument. While best performance
is achieved when testing on the same instrument used for training, these models can also achieve
reasonable performance when applied to new instruments. Thus, we also compare to a discriminative
baseline, a simplified reimplementation of a state-of-the-art supervised approach [7] which achieves
slightly better performance than the original on this task. This system only produces note onsets,
and therefore is not evaluated at a frame-level. We train the discriminative baseline on synthesized
audio with ground-truth MIDI annotations, and apply it directly to our test instrument, which the
system has never seen before.
7
System
Discriminative [7]
Benetos [2]
Vincent [3]
O?Hanlon [4]
This work
P
Onsets
R
F1
P
Frames
R
F1
76.8
62.7
48.6
78.1
65.1
76.8
73.0
74.7
70.4
68.6
69.0
58.3
76.4
79.6
73.4
69.1
63.6
72.8
80.7
68.0
70.7
73.2
74.4
Table 1: Unsupervised transcription results on the MAPS corpus. ?Onsets? columns show scores for identification (within ?50ms) of note start times. ?Frames? columns show scores for 10ms frame-level evaluation. Our
system achieves state-of-the-art results on both metrics.2
4.2
Results
Our model achieves the best published numbers on this task: as shown in Table 1, it achieves an
onset F1 of 76.4, which corresponds to a 10.6% relative gain over the onset F1 achieved by the
system of Vincent et al. [3], the top-performing unsupervised baseline on this metric. Surprisingly,
the discriminative baseline [7], which was not developed for the unsupervised task, outperforms
all the unsupervised baselines in terms of onset evaluation, achieving an F1 of 70.4. Evaluated on
frames, our system achieves an F1 of 74.4, corresponding to a more modest 1.6% relative gain over
the system of O?Hanlon and Plumbley [4], which is the best performing baseline on this metric.
The surprisingly competitive discriminative baseline shows that it is possible to achieve high onset
accuracy on this task without adapting to the test instrument. Thus, it is reasonable to ask how much
of the gain our model achieves is due to its ability to learn instrument timbre. If we skip the blockcoordinate ascent updates (Section 3) for the envelope and spectral parameters, and thus prevent our
system from adapting to the test instrument, onset F1 drops from 76.4 to 72.6. This result indicates
that learning instrument timbre does indeed help performance.
As a short example of our system?s behavior, Figure 4 shows our system?s output passed through a
commercially-available MIDI-to-sheet-music converter. This example was chosen because its onset
F1 of 75.5 and error types are broadly representative of the system?s performance on our data. The
resulting score has musically plausible errors.
Predicted score
Reference score
Figure 4: Result of passing our system?s prediction and the reference transcription MIDI through the GarageBand MIDI-to-sheet-music converter. This is a transcription of the first three bars of Schumann?s Hobgoblin.
A careful inspection of the system?s output suggests that a large fraction of errors are either off by
an octave (i.e. the frequency of the predicted note is half or double the correct frequency) or are
segmentation errors (in which a single key press is transcribed as several consecutive key presses).
While these are tricky errors to correct, they may also be relatively harmless for some applications
because they are not detrimental to musical perception: converting the transcriptions back to audio
using a synthesizer yields music that is qualitatively quite similar to the original recordings.
5
Conclusion
We have shown that combining unsupervised timbral adaptation with a detailed model of the generative relationship between piano sounds and their transcriptions can yield state-of-the-art performance. We hope that these results will motivate further joint approaches to unsupervised music
transcription. Paths forward include exploring more nuanced timbral parameterizations and developing more sophisticated models of discrete musical structure.
2
For consistency we re-ran all systems in this table with our own evaluation code (except for the system
of Benetos and Weyde [2], for which numbers are taken from the paper). For O?Hanlon and Plumbley [4]
scores are higher than the authors themselves report; this is due to an extra post-processing step suggested by
O?Hanlon in personal correspondence.
8
References
[1] Masahiro Nakano, Yasunori Ohishi, Hirokazu Kameoka, Ryo Mukai, and Kunio Kashino. Bayesian nonparametric music parser. In IEEE International Conference on Acoustics, Speech and Signal Processing,
2012.
[2] Emmanouil Benetos and Tillman Weyde. Explicit duration hidden markov models for multiple-instrument
polyphonic music transcription. In International Society for Information Music Retrieval, 2013.
[3] Emmanuel Vincent, Nancy Bertin, and Roland Badeau. Adaptive harmonic spectral decomposition for
multiple pitch estimation. IEEE Transactions on Audio, Speech, and Language Processing, 2010.
[4] Ken O?Hanlon and Mark D. Plumbley. Polyphonic piano transcription using non-negative matrix factorisation with group sparsity. In IEEE International Conference on Acoustics, Speech, and Signal Processing, 2014.
[5] Graham E. Poliner and Daniel P.W. Ellis. A discriminative model for polyphonic piano transcription.
EURASIP Journal on Advances in Signal Processing, 2007.
[6] R. Lienhart C. G. van de Boogaart. Note onset detection for the transcription of polyphonic piano music.
In Multimedia and Expo ICME. IEEE, 2009.
[7] Felix Weninger, Christian Kirst, Bjorn Schuller, and Hans-Joachim Bungartz. A discriminative approach
to polyphonic piano note transcription using supervised non-negative matrix factorization. In IEEE International Conference on Acoustics, Speech and Signal Processing, 2013.
[8] Paul H. Peeling, Ali Taylan Cemgil, and Simon J. Godsill. Generative spectrogram factorization models
for polyphonic piano transcription. IEEE Transactions on Audio, Speech, and Language Processing,
2010.
[9] Julian Besag. On the statistical analysis of dirty pictures. Journal of the Royal Statistical Society, 1986.
[10] Stephen Levinson. Continuously variable duration hidden Markov models for automatic speech recognition. Computer Speech & Language, 1986.
[11] Jyrki Kivinen and Manfred K. Warmuth. Exponentiated gradient versus gradient descent for linear predictors. Information and Computation, 1997.
[12] Matti P. Ryynanen and Anssi Klapuri. Polyphonic music transcription using note event modeling. In
IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, 2005.
[13] Sebastian B?ock and Markus Schedl. Polyphonic piano note transcription with recurrent neural networks.
In IEEE International Conference on Acoustics, Speech and Signal Processing, 2012.
[14] Valentin Emiya, Roland Badeau, and Bertrand David. Multipitch estimation of piano sounds using a
new probabilistic spectral smoothness principle. IEEE Transactions on Audio, Speech, and Language
Processing, 2010.
[15] The international music score library project, June 2014. URL http://imslp.org.
[16] Vladimir Viro. Peachnote: Music score search and analysis platform. In The International Society for
Music Information Retrieval, 2011.
9
| 5432 |@word rising:1 d2:1 jacob:1 decomposition:1 pressed:3 series:3 score:9 daniel:1 outperforms:3 past:1 existing:2 activation:31 synthesizer:3 must:2 readily:1 additive:1 informative:1 shape:4 christian:1 drop:1 update:11 polyphonic:11 n0:4 generative:4 half:1 tillman:1 parameterization:4 warmuth:1 inspection:1 short:2 manfred:1 parameterizations:1 location:3 org:1 plumbley:5 along:1 consists:3 dan:1 combine:1 fitting:2 indeed:1 behavior:1 themselves:1 frequently:1 mechanic:1 roughly:1 discretized:1 inspired:1 bertrand:1 resolve:2 actual:2 equipped:1 window:1 schumann:1 project:1 underlying:1 notation:1 maximizes:1 kind:1 interpreted:1 string:15 spectrally:1 developed:1 finding:1 unobserved:1 temporal:2 berkeley:2 collecting:1 tackle:1 exactly:1 scaled:3 tricky:1 control:1 unit:1 positive:4 before:1 felix:1 cemgil:1 path:1 initialization:2 specifying:2 challenging:2 collect:1 suggests:1 factorization:5 testing:1 practice:2 block:2 tribution:1 reimplementation:1 procedure:2 adapting:3 thought:1 symbolic:7 convenience:1 onto:1 sheet:4 noising:1 collapsed:1 optimize:2 map:5 straightforward:1 regardless:1 starting:2 duration:14 truncating:1 simplicity:1 factorisation:1 m2:1 classic:1 harmless:1 coordinate:2 analogous:2 play:8 parser:1 decode:1 poliner:1 velocity:14 recognition:1 particularly:1 updating:4 continues:1 blockcoordinate:1 observed:7 capture:1 parameterize:2 ran:1 mentioned:1 environment:1 kameoka:1 ock:1 dynamic:3 personal:1 diagrammed:1 trained:1 motivate:1 segment:8 ali:1 creates:1 division:1 bungartz:1 collide:1 joint:5 represented:1 listener:1 various:1 articulated:1 train:1 fast:1 describe:2 quite:2 plausible:1 say:1 ability:1 statistic:1 unseen:1 jointly:1 transform:1 superscript:1 final:1 sequence:4 product:2 adaptation:2 frequent:1 aligned:1 combining:2 achieve:2 description:1 double:1 produce:6 generating:1 help:1 coupling:1 illustrate:1 recurrent:1 transcript:1 c:1 predicted:3 skip:1 annotated:2 hammer:2 correct:4 human:1 bin:4 musically:1 f1:12 mathematically:1 exploring:1 considered:1 ground:3 normal:3 taylan:1 predict:1 vary:2 achieves:7 consecutive:1 released:1 purpose:1 estimation:2 superposition:2 sensitive:1 reflects:2 hope:1 sensor:2 gaussian:4 rather:1 avoid:1 focus:1 emission:1 joachim:1 notational:1 june:1 superimposed:2 likelihood:1 indicates:1 besag:1 pdi:2 baseline:8 inference:10 downsample:1 softly:1 typically:1 diminishing:1 hidden:5 going:1 denoted:1 development:1 art:5 raised:1 constrained:1 platform:1 never:1 hop:1 represents:1 unsupervised:12 commercially:1 others:1 report:5 primarily:1 simultaneously:1 gamma:1 divergence:2 individual:2 consisting:3 pleasing:1 initializers:1 detection:1 evaluation:8 kirkpatrick:1 held:5 emit:1 tuple:1 partial:1 shorter:1 modest:1 taylor:1 loosely:1 re:1 causal:1 isolated:2 column:3 modeling:5 boolean:1 elli:1 entry:3 uniform:1 predictor:1 examining:1 valentin:1 musicology:1 params:4 synthetic:1 objec:1 peak:1 fundamental:1 international:7 probabilistic:3 off:1 together:3 continuously:1 concrete:1 posi:1 again:1 recorded:1 transcribed:5 external:2 style:1 potential:1 converted:1 de:1 dur:2 availability:1 includes:2 satisfy:1 onset:17 vi:10 piece:1 sine:1 break:1 multiplicative:2 performed:3 closed:2 portion:2 wave:1 start:1 competitive:1 parallel:2 annotation:1 defer:1 simon:1 contribution:2 appended:1 benetos:5 accuracy:1 musical:27 variance:1 efficiently:1 yield:4 identification:1 vincent:4 iterated:1 bayesian:1 produced:7 weninger:2 trajectory:1 published:4 sebastian:1 energy:2 frequency:13 e2:2 naturally:1 di:8 mi:4 couple:1 gain:5 emits:1 dataset:3 ask:1 nancy:1 recall:1 knowledge:1 improves:1 electromechanical:1 segmentation:1 sophisticated:1 back:2 higher:1 supervised:4 specify:1 vel:2 though:1 evaluated:2 emmanouil:1 until:1 ei:8 icme:1 mode:1 nuanced:1 klapuri:1 true:2 freq:6 during:4 ambiguous:1 m:5 octave:2 complete:2 performs:1 postprocessing:1 ranging:1 harmonic:5 common:1 multinomial:1 khz:1 volume:2 interpretation:1 m1:1 synthesized:4 ai:7 smoothness:1 automatic:2 consistency:1 depressed:1 badeau:2 language:4 lowered:1 access:2 han:1 longer:1 depiction:1 add:2 posterior:1 own:3 optimizing:1 keyboard:5 seen:2 greater:1 staff:1 spectrogram:31 speci:1 converting:1 v3:1 strike:3 maximize:1 signal:16 semi:4 period:1 multiple:6 full:5 sound:14 stephen:1 levinson:1 adapt:1 long:3 retrieval:3 divided:1 post:1 e1:2 roland:2 prediction:1 pitch:1 metric:5 poisson:8 iteration:1 represent:3 achieved:2 interval:1 source:5 envelope:15 rest:6 extra:1 ascent:5 recording:10 accompany:1 deficient:1 emitted:3 call:2 integer:2 counting:1 intermediate:2 variety:2 independence:1 fit:1 converter:2 andreas:1 ultimate:1 passed:1 url:1 song:20 transcribing:3 e3:1 passing:1 speech:9 useful:1 generally:1 collision:1 detailed:3 nonparametric:1 multipitch:1 ken:1 generate:8 specifies:4 http:1 klein:2 broadly:1 discrete:11 group:3 key:9 achieving:3 d3:1 prevent:1 v1:1 graph:2 geometrically:1 fraction:1 run:1 parameterized:9 laid:1 reasonable:2 separation:4 graham:1 jda:1 played:7 correspondence:1 strength:1 sharply:1 worked:1 expo:1 markus:1 damper:3 generates:5 fourier:1 extremely:1 performing:2 format:1 relatively:1 transferred:1 developing:1 truncate:1 combination:1 describes:2 remain:2 across:5 suppressed:1 slightly:1 making:1 heart:1 taken:1 remains:1 previously:1 count:1 tractable:1 instrument:15 available:3 apply:1 v2:1 spectral:18 appropriate:1 generic:1 occurrence:2 original:2 denotes:1 running:1 top:1 include:1 dirty:1 graphical:1 marginalized:1 nakano:1 music:25 emmanuel:1 society:3 objective:6 looked:1 primary:3 dependence:1 nr:39 exhibit:2 loudness:3 gradient:6 detrimental:1 separate:1 outer:2 reason:1 assuming:1 length:6 code:1 index:3 relationship:1 providing:2 julian:1 vladimir:1 difficult:1 mostly:1 potentially:1 negative:4 rise:5 godsill:1 suppress:1 attenuates:1 unknown:1 perform:2 markov:8 descent:1 truncated:4 frame:13 arbitrary:1 intensity:1 nmf:2 introduced:1 tive:2 pair:1 david:1 kl:2 acoustic:13 california:1 learned:1 trans:2 able:1 bar:1 suggested:1 perception:1 masahiro:1 sparsity:1 program:3 including:1 max:1 royal:1 event:41 rely:1 kivinen:1 residual:1 schuller:1 representing:2 library:2 temporally:1 picture:1 extract:1 prior:6 piano:47 literature:4 marginalizing:2 relative:4 generation:2 proven:1 versus:1 bertin:1 consistent:1 principle:2 bank:1 row:1 course:1 surprisingly:2 keeping:1 jth:2 exponentiated:3 fall:4 taking:2 lienhart:1 distributed:1 benefit:1 van:1 dimension:1 transition:1 rich:1 forward:2 commonly:1 collection:4 preprocessing:1 simplified:1 qualitatively:1 author:1 far:1 adaptive:1 transaction:3 approximate:1 midi:12 transcription:30 emiya:1 active:1 yasunori:1 corpus:7 assumed:1 tuples:1 discriminative:7 spectrum:3 search:1 latent:1 ryo:1 table:3 polyphony:2 learn:3 matti:1 timbral:7 significance:1 noise:16 paul:1 profile:6 fied:1 representative:1 depicts:1 fashion:2 n:1 precision:1 deterministically:1 explicit:1 peeling:1 specific:2 symbol:1 offset:2 timbre:3 decay:3 experimented:1 admits:1 consist:1 intractable:1 workshop:1 hanlon:7 magnitude:3 conditioned:8 depicted:7 simply:1 likely:2 bjorn:1 corresponds:2 truth:3 relies:1 transcribe:2 conditional:5 goal:2 jyrki:1 careful:1 shared:3 hard:2 eurasip:1 specifically:5 determined:1 except:1 total:5 called:1 multimedia:1 superimposing:2 player:1 formally:1 berg:1 mark:1 incorporate:1 evaluate:2 audio:16 d1:1 |
4,897 | 5,433 | Combinatorial Pure Exploration of
Multi-Armed Bandits
Shouyuan Chen1? Tian Lin2
Irwin King1
Michael R. Lyu1
Wei Chen3
2
3
The Chinese University of Hong Kong
Tsinghua University
Microsoft Research Asia
1
2
{sychen,king,lyu}@cse.cuhk.edu.hk
[email protected] 3 [email protected]
1
Abstract
We study the combinatorial pure exploration (CPE) problem in the stochastic multi-armed
bandit setting, where a learner explores a set of arms with the objective of identifying
the optimal member of a decision class, which is a collection of subsets of arms with
certain combinatorial structures such as size-K subsets, matchings, spanning trees or paths,
etc. The CPE problem represents a rich class of pure exploration tasks which covers not
only many existing models but also novel cases where the object of interest has a nontrivial combinatorial structure. In this paper, we provide a series of results for the general
CPE problem. We present general learning algorithms which work for all decision classes
that admit offline maximization oracles in both fixed confidence and fixed budget settings.
We prove problem-dependent upper bounds of our algorithms. Our analysis exploits the
combinatorial structures of the decision classes and introduces a new analytic tool. We also
establish a general problem-dependent lower bound for the CPE problem. Our results show
that the proposed algorithms achieve the optimal sample complexity (within logarithmic
factors) for many decision classes. In addition, applying our results back to the problems
of top-K arms identification and multiple bandit best arms identification, we recover the
best available upper bounds up to constant factors and partially resolve a conjecture on the
lower bounds.
1
Introduction
Multi-armed bandit (MAB) is a predominant model for characterizing the tradeoff between exploration and exploitation in decision-making problems. Although this is an intrinsic tradeoff in many
tasks, some application domains prefer a dedicated exploration procedure in which the goal is to
identify an optimal object among a collection of candidates and the reward or loss incurred during
exploration is irrelevant. In light of these applications, the related learning problem, called pure exploration in MABs, has received much attention. Recent advances in pure exploration MABs have
found potential applications in many domains including crowdsourcing, communication network
and online advertising.
In many of these application domains, a recurring problem is to identify the optimal object with
certain combinatorial structure. For example, a crowdsourcing application may want to find the best
assignment from workers to tasks such that overall productivity of workers is maximized. A network
routing system during the initialization phase may try to build a spanning tree that minimizes the
delay of links, or attempts to identify the shortest path between two sites. An online advertising
system may be interested in finding the best matching between ads and display slots. The literature
of pure exploration MAB problems lacks a framework that encompasses these kinds of problems
where the object of interest has a non-trivial combinatorial structure. Our paper contributes such
a framework which accounts for general combinatorial structures, and develops a series of results,
including algorithms, upper bounds and lower bounds for the framework.
In this paper, we formulate the combinatorial pure exploration (CPE) problem for stochastic multiarmed bandits. In the CPE problem, a learner has a fixed set of arms and each arm is associated with
an unknown reward distribution. The learner is also given a collection of sets of arms called decision
class, which corresponds to a collection of certain combinatorial structures. During the exploration
period, in each round the learner chooses an arm to play and observes a random reward sampled from
?
This work was done when the first two authors were interns at Microsoft Research Asia.
1
the associated distribution. The objective is when the exploration period ends, the learner outputs a
member of the decision class that she believes to be optimal, in the sense that the sum of expected
rewards of all arms in the output set is maximized among all members in the decision class.
The CPE framework represents a rich class of pure exploration problems. The conventional pure exploration problem in MAB, whose objective is to find the single best arm, clearly fits into this framework, in which the decision class is the collection of all singletons. This framework also naturally
encompasses several recent extensions, including the problem of finding the top K arms (henceforth
T OP K) [18, 19, 8, 20, 31] and the multi-bandit problem of finding the best arms simultaneously
from several disjoint sets of arms (henceforth MB) [12, 8]. Further, this framework covers many
more interesting cases where the decision classes correspond to collections of non-trivial combinatorial structures. For example, suppose that the arms represent the edges in a graph. Then a decision
class could be the set of all paths between two vertices, all spanning trees or all matchings of the
graph. And, in these cases, the objectives of CPE become identifying the optimal paths, spanning
trees and matchings through bandit explorations, respectively. To our knowledge, there are no results
available in the literature for these pure exploration tasks.
The CPE framework raises several interesting challenges to the design and analysis of pure exploration algorithms. One challenge is that, instead of solving each type of CPE task in an ad-hoc way,
one requires a unified algorithm and analysis that support different decision classes. Another challenge stems from the combinatorial nature of CPE, namely that the optimal set may contain some
arms with very small expected rewards (e.g., it is possible that a maximum matching contains the
edge with the smallest weight); hence, arms cannot be eliminated simply based on their own rewards in the learning algorithm or ignored in the analysis. This differs from many existing approach
of pure exploration MABs. Therefore, the design and analysis of algorithms for CPE demands novel
techniques which take both rewards and combinatorial structures into account.
Our results. In this paper, we propose two novel learning algorithms for general CPE problem: one
for the fixed confidence setting and one for the fixed budget setting. Both algorithms support a wide
range of decision classes in a unified way. In the fixed confidence setting, we present Combinatorial
Lower-Upper Confidence Bound (CLUCB) algorithm. The CLUCB algorithm does not need to know
the definition of the decision class, as long as it has access to the decision class through a maximization oracle. We upper bound the number of samples used by CLUCB. This sample complexity bound
depends on both the expected rewards and the structure of decision class. Our analysis relies on a
novel combinatorial construction called exchange class, which may be of independent interest for
other combinatorial optimization problems. Specializing our result to T OP K and MB, we recover
the best available sample complexity bounds [19, 13, 20] up to constant factors. While for other decision classes in general, our result establishes the first sample complexity upper bound. We further
show that CLUCB can be easily extended to the fixed budget setting and PAC learning setting and
we provide related theoretical guarantees in the supplementary material.
Moreover, we establish a problem-dependent sample complexity lower bound for the CPE problem.
Our lower bound shows that the sample complexity of the proposed CLUCB algorithm is optimal
(to within logarithmic factors) for many decision classes, including T OP K, MB and the decision
classes derived from matroids (e.g., spanning tree). Therefore our upper and lower bounds provide
a nearly full characterization of the sample complexity of these CPE problems. For more general
decision classes, our results show that the upper and lower bounds are within a relatively benign
factor. To the best of our knowledge, there are no problem-dependent lower bounds known for pure
exploration MABs besides the case of identifying the single best arm [24, 1]. We also notice that
our result resolves the conjecture of Bubeck et al. [8] on the problem-dependent sample complexity
lower bounds of T OP K and MB problems, for the cases of Gaussian reward distributions.
In the fixed budget setting, we present a parameter-free algorithm called Combinatorial Successive
Accept Reject (CSAR) algorithm. We prove a probability of error bound of the CSAR algorithm. This
bound can be shown to be equivalent to the sample complexity bound of CLUCB within logarithmic
factors, although the two algorithms are based on quite different techniques. Our analysis of CSAR
re-uses exchange classes as tools. This suggests that exchange classes may be useful for analyzing
similar problems. In addition, when applying the algorithm to back T OP K and MB, our bound
recovers the best known result in the fixed budget setting due to Bubeck et al. [8] up to constant
factors.
2
2
Problem Formulation
In this section, we formally define the CPE problem. Suppose that there are n arms and the arms
are numbered 1, 2, . . . , n. Assume that each arm e 2 [n] is associated with a reward distribution
T
'e . Let w = w(1), . . . , w(n) denote the vector of expected rewards, where each entry w(e) =
EX?'e [X] denotes the expected reward of arm e. Following standard assumptions of stochastic
MABs, we assume that all reward distributions have R-sub-Gaussian tails for some known constant
R > 0. Formally,
if X is a random
variable drawn from 'e for some e 2 [n], then, for all t 2 R,
?
?
one has E exp(tX tE[X]) ? exp(R2 t2 /2). It is known that the family of R-sub-Gaussian tail
distributions encompasses all distributions that are supported on [0, R] as well as many unbounded
distributions such as Gaussian distributions with variance R2 (see e.g., [27, 28]).
We define a decision class M ? 2[n] as a collection of sets of arms. Let M? = arg maxM 2M w(M )
denote the optimal member of the decision class M which maximizes the sum of expected rewards1 .
A learner?s objective is to identify M? from M by playing the following game with the stochastic
environment. At the beginning of the game, the decision class M is revealed to the learner while
the reward distributions {'e }e2[n] are unknown to her. Then, the learner plays the game over a
sequence of rounds; in each round t, she pulls an arm pt 2 [n] and observes a reward sampled
from the associated reward distribution 'pt . The game continues until certain stopping condition is
satisfied. After the game finishes, the learner need to output a set Out 2 M.
We consider two different stopping conditions of the game, which are known as fixed confidence
setting and fixed budget setting in the literature. In the fixed confidence setting, the learner can stop
the game at any round. She need to guarantee that Pr[Out = M? ] 1
for a given confidence
parameter . The learner?s performance is evaluated by her sample complexity, i.e., the number of
pulls used by the learner. In the fixed budget setting, the game stops after a fixed number T of rounds,
where T is given before the game starts. The learner tries to minimize the probability of error, which
is formally Pr[Out 6= M? ], within T rounds. In this setting, her performance is measured by the
probability of error.
3
Algorithm, Exchange Class and Sample Complexity
In this section, we present Combinatorial Lower-Upper Confidence Bound (CLUCB) algorithm, a
learning algorithm for the CPE problem in the fixed confidence setting, and analyze its sample complexity. En route to our sample complexity bound, we introduce the notions of exchange classes and
the widths of decision classes, which play an important role in the analysis and sample complexity
bound. Furthermore, the CLUCB algorithm can be extended to the fixed budget and PAC learning
settings, the discussion of which is included in the supplementary material (Appendix B).
Oracle. We allow the CLUCB algorithm to access a maximization oracle. A maximization oracle
takes a weight vector v 2 Rn as input and finds an optimal set from a given decision class M with
respect to the weight vector v. Formally, we call a function Oracle: Rn ! M a maximization oracle
for M if, for all v 2 Rn , we have Oracle(v) 2 arg maxM 2M v(M ). It is clear that a wide range
of decision classes admit such maximization oracles, including decision classes corresponding to
collections of matchings, paths or bases of matroids (see later for concrete examples). Besides the
access to the oracle, CLUCB does not need any additional knowledge of the decision class M.
Algorithm. Now we describe the details of CLUCB, as shown in Algorithm 1. During its execution,
the CLUCB algorithm maintains empirical mean w
?t (e) and confidence radius radt (e) for each arm
e 2 [n] and each round t. The construction of confidence radius ensures that |w(e) w
?t (e)| ?
radt (e) holds with high probability for each arm e 2 [n] and each round t > 0. CLUCB begins
with an initialization phase in which each arm is pulled once. Then, at round t
n, CLUCB uses
the following procedure to choose an arm to play. First, CLUCB calls the oracle which finds the
set Mt = Oracle(w
? t ). The set Mt is the ?best? set with respect to the empirical means w
? t . Then,
CLUCB explores possible refinements of Mt . In particular, CLUCB uses the confidence radius to
compute an adjusted expectation vector w
? t in the following way: for each arm e 2 Mt , w
?t (e) is
equal to to the lower confidence bound w
?t (e) = w
?t (e) radt (e); and for each arm e 62 Mt , w
?t (e) is
equal to the upper confidence bound w
?t (e) = w
?t (e) + radt (e). Intuitively, the adjusted expectation
vector w
? t penalizes arms belonging to the current set Mt and encourages exploring arms out of
P
We define v(S) , i2S v(i) for any vector v 2 Rn and any set S ? [n]. In addition, for convenience,
we will assume that M? is unique.
1
3
Algorithm 1 CLUCB: Combinatorial Lower-Upper Confidence Bound
Require: Confidence 2 (0, 1); Maximization oracle: Oracle(?) : Rn ! M
Initialize: Play each arm e 2 [n] once. Initialize empirical means w
? n and set Tn (e)
1 for all e.
1: for t = n, n + 1, . . . do
2:
Mt
Oracle(w
?t)
3:
Compute confidence radius radt (e) for all e 2 [n]
. radt (e) is defined later in Theorem 1
4:
for e = 1, . . . , n do
5:
if e 2 Mt then w
?t (e)
w
?t (e) radt (e)
6:
else w
?t (e)
w
?t (e) + radt (e)
?t
7:
M
Oracle(w
?t)
? t) = w
8:
if w
? t (M
?t (Mt ) then
9:
Out
Mt
10:
return Out
11:
pt
arg maxe2(M? t \Mt )[(Mt \M? t ) radt (e)
. break ties arbitrarily
12:
Pull arm pt and observe the reward
13:
Update empirical means w
? t+1 using the observed reward
14:
Update number of pulls: Tt+1 (pt )
Tt (pt ) + 1 and Tt+1 (e)
Tt (e) for all e 6= pt
Mt . CLUCB then calls the oracle using the adjusted expectation vector w
? t as input to compute a
? t = Oracle(w
? t) = w
refined set M
? t ). If w
? t (M
?t (Mt ) then CLUCB stops and returns Out = Mt .
? t and
Otherwise, CLUCB pulls the arm that belongs to the symmetric difference between Mt and M
has the largest confidence radius (intuitively the largest uncertainty). This ends the t-th round of
CLUCB. We note that CLUCB generalizes and unifies the ideas of several different fixed confidence
algorithms dedicated to the T OP K and MB problems in the literature [19, 13, 20].
3.1 Sample complexity
Now we establish a problem-dependent sample complexity bound of the CLUCB algorithm. To formally state our result, we need to introduce several notions.
Gap. We begin with defining a natural hardness measure of the CPE problem. For each arm e 2 [n],
we define its gap e as
?
w(M? ) maxM 2M:e2M w(M ) if e 62 M? ,
(1)
e =
w(M? ) maxM 2M:e62M w(M ) if e 2 M? ,
where we adopt the convention that the maximum value of an empty set is
hardness H as the sum of inverse squared gaps
X
2
H=
e .
1. We also define the
(2)
e2[n]
We see that, for each arm e 62 M? , the gap e represents the sub-optimality of the best set that
includes arm e; and, for each arm e 2 M? , the gap e is the sub-optimality of the best set that does
not include arm e. This naturally generalizes and unifies previous definitions of gaps [1, 12, 18, 8].
Exchange class and the width of a decision class. A notable challenge of our analysis stems from
the generality of CLUCB which, as we have seen, supports a wide range of decision classes M.
Indeed, previous algorithms for special cases including T OP K and MB require a separate analysis
for each individual type of problem. Such strategy is intractable for our setting and we need a unified
analysis for all decision classes. Our solution to this challenge is a novel combinatorial construction
called exchange class, which is used as a proxy for the structure of the decision class. Intuitively,
an exchange class B for a decision class M can be seen as a collection of ?patches? (borrowing
concepts from source code management) such that, for any two different sets M, M 0 2 M, one can
transform M to M 0 by applying a series of patches of B; and each application of a patch yields a
valid member of M. These patches are later used by our analysis to build gadgets that interpolate
between different members of the decision class and serve to bridge key quantities. Furthermore, the
maximum patch size of B will play an important role in our sample complexity bound.
Now we formally define the exchange class. We begin with the definition of exchange sets, which
formalize the aforementioned ?patches?. We define an exchange set b as an ordered pair of disjoint
sets b = (b+ , b ) where b+ \ b = ; and b+ , b ? [n]. Then, we define operator such that, for
any set M ? [n] and any exchange set b = (b+ , b ), we have M b , M \b [ b+ . Similarly, we
also define operator such that M b , M \b+ [ b .
4
We call a collection of exchange sets B an exchange class for M if B satisfies the following property.
For any M, M 0 2 M such that M 6= M 0 and for any e 2 (M \M 0 ), there exists an exchange set
(b+ , b ) 2 B which satisfies five constraints: (a) e 2 b , (b) b+ ? M 0 \M , (c) b ? M \M 0 , (d)
(M b) 2 M and (e) (M 0 b) 2 M.
Intuitively, constraints (b) and (c) resemble the concept of patches in the sense that b+ contains
only the ?new? elements from M 0 and b contains only the ?old? elements of M ; constraints (d)
and (e) allow one to transform M one step closer to M 0 by applying a patch b 2 B to yield (M
b) 2 M (and similarly for M 0 b). These transformations are the basic building blocks in our
analysis. Furthermore, as we will see later in our examples, for many decision classes, there are
exchange classes representing natural combinatorial structures, e.g., augmenting paths and cycles of
matchings.
In our analysis, the key quantity of exchange class is called width, which is defined as the size of the
largest exchange set as follows
width(B) =
max
(b+ ,b )2B
(3)
|b+ | + |b |.
Let Exchange(M) denote the family of all possible exchange classes for M. We define the width
of a decision class M as the width of the thinnest exchange class
width(M) =
min
B2Exchange(M)
(4)
width(B).
Sample complexity. Our main result of this section is a problem-dependent sample complexity
bound of the CLUCB algorithm which show that, with high probability, CLUCB returns the optimal
? width(M)2 H samples.
set M? and uses at most O
Theorem 1. Given any 2 (0, 1), any decision class M ? 2[n] and any expected rewards w 2 Rn .
Assume that the reward distribution 'e for each arm e 2 [n] has mean w(e) with
qan R-sub-Gaussian
3
tail. Let M? = arg maxM 2M w(M ) denote the optimal set. Set radt (e) = R 2 log 4nt /Tt (e)
for all t > 0 and e 2 [n]. Then, with probability at least 1
, the CLUCB algorithm (Algorithm 1)
returns the optimal set Out = M? and
T ? O R2 width(M)2 H log nR2 H/
,
(5)
where T denotes the number of samples used by Algorithm 1, H is defined in Eq. (2) and width(M)
is defined in Eq. (4).
3.2 Examples of decision classes
Now we investigate several concrete types of decision classes, which correspond to different CPE
tasks. We analyze the width of these decision classes and apply Theorem 1 to obtain the sample
complexity bounds. A detailed analysis and the constructions of exchange classes can be found in
the supplementary material (Appendix F). We begin with the problems of top-K arm identification
(T OP K) and multi-bandit best arms identification (MB).
Example 1 (T OP K and MB). For any K 2 [n], the problem of finding the top K arms with the
largest expected reward can be modeled by decision class MTOP K(K) = {M ? [n] | M = K}.
Let A = {A1 , . . . , Am } be a partition of [n]. The problem of identifying the best arms from each
group of arms A1 , . . . , Am can be modeled by decision class MMB(A) = {M ? [n] | 8i 2
[m], |M \ Ai | = 1}. Note that maximization oracles for these two decision classes are trivially the
functions of returning the top k arms or the best arms of each group.
Then we have width(MTOP K(K) ) ? 2 and width(MMB(A) ) ? 2 (see Fact 2 and 3 in the supplementary material) and therefore the sample complexity of CLUCB for solving T OP K and MB is
O H log(nH/ ) , which matches previous results in the fixed confidence setting [19, 13, 20] up to
constant factors.
Next we consider the problem of identifying the maximum matching and the problem of finding
the shortest path (by negating the rewards), in a setting where arms correspond to edges. For these
problems, Theorem 1 establishes the first known sample complexity bound.
5
Example 2 (Matchings and Paths). Let G(V, E) be a graph with n edges and assume there is a oneto-one mapping between edges E and arms [n]. Suppose that G is a bipartite graph. Let MM ATCH(G)
correspond to the set of all matchings in G. Then we have width(MM ATCH(G) ) ? |V | (In fact, we
construct an exchange class corresponding to the collection of augmenting cycles and augmenting
paths of G; see Fact 4).
Next suppose that G is a directed acyclic graph and let s, t 2 V be two vertices. Let MPATH(G,s,t)
correspond to the set of all paths from s to t. Then we have width(MPATH(G,s,t) ) ? |V | (In fact,
we construct an exchange class corresponding to the collection of disjoint pairs of paths; see
Fact 5). Therefore the sample complexity bounds of CLUCB for decision classes MM ATCH(G) and
MPATH(G,s,t) are O |V |2 H log(nH/ ) .
Last, we investigate the general problem of identifying the maximum-weight basis of a matroid.
Again, Theorem 1 is the first sample complexity upper bound for this type of pure exploration tasks.
Example 3 (Matroids). Let T = (E, I) be a finite matroid, where E is a set of size n (called
ground set) and I is a family of subsets of E (called independent sets) which satisfies the axioms of
matroids (see Footnote 3 in Appendix F). Assume that there is a one-to-one mapping between E and
[n]. Recall that a basis of matroid T is a maximal independent set. Let MM ATROID(T ) correspond
to the set of all bases of T . Then we have width(MM ATROID(T ) ) ? 2 (derived from strong basis
exchange property of matroids; see Fact 1) and the sample complexity of CLUCB for MM ATROID(T )
is O H log(nH/ ) .
The last example MM ATROID(T ) is a general type of decision class which encompasses many pure
exploration tasks including T OP K and MB as special cases, where T OP K corresponds to uniform
matroids of rank K and MB corresponds to partition matroids. It is easy to see that MM ATROID(T )
also covers the decision class that contains all spanning trees of a graph. On the other hand, it has
been established that matchings and paths cannot be formulated as matroids since they are matroid
intersections [26].
4
Lower Bound
In this section, we present a problem-dependent lower bound on the sample complexity of the CPE
problem. To state our results, we first define the notion of -correct algorithm as follows. For any
2 (0, 1), we call an algorithm A a -correct algorithm if, for any expected reward w 2 Rn , the
probability of error of A is at most , i.e., Pr[M? 6= Out] ? , where Out is the output of A.
We show that, for any decision class M and any expected rewards w, a -correct algorithm A must
use at least ? H log(1/ ) samples in expectation.
Theorem 2. Fix any decision class M ? 2[n] and any vector w 2 Rn . Suppose that, for each
arm e 2 [n], the reward distribution 'e is given by 'e = N (w(e), 1), where we let N (?, 2 )
denote Gaussian distribution with mean ? and variance 2 . Then, for any 2 (0, e 16 /4) and any
-correct algorithm A, we have
? ?
1
1
E[T ]
H log
,
(6)
16
4
where T denote the number of total samples used by algorithm A and H is defined in Eq. (2).
In Example 1 and Example 3, we have seen that the sample complexity of CLUCB is
O(H log(nH/ )) for pure exploration tasks including T OP K, MB and more generally the CPE
tasks with decision classes derived from matroids, i.e., MM ATROID(T ) (including spanning trees).
Hence, our upper and lower bound show that the CLUCB algorithm achieves the optimal sample
complexity within logarithmic factors for these pure exploration tasks. In addition, we remark that
Theorem 2 resolves the conjecture of Bubeck et al. [8] that the lower bounds of sample complexity
of T OP K and MB problems are ? H log(1/ ) , for the cases of Gaussian reward distributions.
On the other hand, for general decision classes with non-constant widths, we see that there is a gap of
2
?
?(width(M)
) between the upper bound Eq. (5) and the lower bound Eq. (6). Notice that we have
width(M) ? n for any decision class M and therefore the gap is relatively benign. Our lower bound
also suggests that the dependency on H of the sample complexity of CLUCB cannot be improved up
to logarithmic factors. Furthermore, we conjecture that the sample complexity lower bound might
inherently depend on the size of exchange sets. In the supplementary material (Appendix C.2), we
6
provide evidences on this conjecture which is a lower bound on the sample complexity of exploration
of the exchange sets.
5
Fixed Budget Algorithm
In this section, we present Combinatorial Successive Accept Reject (CSAR) algorithm, which is a
parameter-free learning algorithm for the CPE problem in the fixed budget setting. Then, we upper
bound the probability of error CSAR in terms of gaps and width(M).
Constrained oracle. The CSAR algorithm requires access to a constrained oracle, which is a function denoted as COracle : Rn ? 2[n] ? 2[n] ! M [ {?} and satisfies
(
arg maxM 2MA,B v(M ) if MA,B 6= ;
COracle(v, A, B) =
(7)
?
if MA,B = ;,
where we define MA,B = {M 2 M | A ? M, B \ M = ;} as the collection of feasible sets
and ? is a null symbol. Hence we see that COracle(v, A, B) returns an optimal set that includes all
elements of A while excluding all elements of B; and if there are no feasible sets, the constrained
oracle COracle(v, A, B) returns the null symbol ?. In the supplementary material (Appendix G),
we show that constrained oracles are equivalent to maximization oracles up to a transformation on
the weight vector. In addition, similar to CLUCB, CSAR does not need any additional knowledge of
M other than accesses to a constrained oracle for M.
Algorithm. The idea of the CSAR algorithm is as follows. The CSAR algorithm divides the budget
of T rounds into n phases. In the end of each phase, CSAR either accepts or rejects a single arm. If
an arm is accepted, then it is included into the final output. Conversely, if an arm is rejected, then it
is excluded from the final output. The arms that are neither accepted nor rejected are sampled for an
equal number of times in the next phase.
Now we describe the procedure of the CSAR algorithm for choosing an arm to accept/reject. Let
At denote the set of accepted arms before phase t and let Bt denote the set of rejected arms before
phase t. We call an arm e to be active if e 62 At [ Bt . In the beginning of phase t, CSAR samples
each active arm for T?t T?t 1 times, where the definition of T?t is given in Algorithm 2. Next,
CSAR calls the constrained oracle to compute an optimal set Mt with respect to the empirical means
w
? t , accepted arms At and rejected arms Bt , i.e., Mt = COracle(w
? t , At , Bt ). It is clear that the
output of COracle(w
? t , At , Bt ) is independent from the input w
?t (e) for any e 2 At [ Bt . Then, for
each active arm e, CSAR estimates the ?empirical gap? of e in the following way. If e 2 Mt , then
? t,e that does not include e, i.e., M
? t,e = COracle(w
CSAR computes an optimal set M
? t , At , Bt [
?
? t,e =
{e}). Conversely, if e 62 Mt , then CSAR computes an optimal Mt,e which includes e, i.e., M
?
COracle(w
? t , At [{e}, Bt ). Then, the empirical gap of e is calculated as w
?t (Mt ) w
?t (Mt,e ). Finally,
CSAR chooses the arm pt which has the largest empirical gap. If pt 2 Mt then pt is accepted,
otherwise pt is rejected. The pseudo-code CSAR is shown in Algorithm 2. We note that CSAR can
be considered as a generalization of the ideas of the two versions of SAR algorithm due to Bubeck
et al. [8], which are designed specifically for the T OP K and MB problems respectively.
5.1
Probability of error
In the following theorem, we bound the probability of error of the CSAR algorithm.
Theorem 3. Given any T > n, any decision class M ? 2[n] and any expected rewards w 2
Rn . Assume that the reward distribution 'e for each arm e 2 [n] has mean w(e) with an R-subGaussian tail. Let (1) , . . . , (n) be a permutation of 1 , . . . , n (defined in Eq. (1)) such that
2
(1) ? . . . . . . (n) . Define H2 , maxi2[n] i (i) . Then, the CSAR algorithm uses at most T
samples and outputs a solution Out 2 M [ {?} such that
?
?
(T n)
2
Pr[Out 6= M? ] ? n exp
,
(8)
?
18R2 log(n)
width(M)2 H2
Pn
?
where log(n)
,
i 1 , M? = arg max
w(M ) and width(M) is defined in Eq. (4).
i=1
M 2M
One can verify that H2 is equivalent to H up to a logarithmic factor: H2 ? H ? log(2n)H2 (see
[1]). Therefore, by setting the probability of error (the RHS of Eq. (8)) to a constant, one can see
2
?
that CSAR requires a budget of T = O(width(M)
H) samples. This is equivalent to the sample
complexity bound of CLUCB up to logarithmic factors. In addition, applying Theorem 3 back to
T OP K and MB, our bound matches the previous fixed budget algorithm due to Bubeck et al. [8].
7
Algorithm 2 CSAR: Combinatorial Successive Accept Reject
Require: Budget: T > 0; Constrained oracle: COracle : Rn ? 2[n] ? 2[n] ! M [ {?}.
P
1
?
1: Define log(n)
, n
i=1 i
?
2: T0
0, A1
;, B1
;
3: for t = 1,l. . . , n do
m
T n
4:
T?t
?
log(n)(n
t+1)
5:
Pull each arm e 2 [n]\(At [ Bt ) for T?t T?t 1 times
6:
Update the empirical means w
? t for each arm e 2 [n]\(At [ Bt )
. set w
?t (e) = 0, 8e 2 At [ Bt
7:
Mt
COracle(w
? t , At , Bt )
8:
if Mt = ? then
9:
fail: set Out
? and return Out
10:
for each e 2 [n]\(At [ Bt ) do
? t,e
11:
if e 2 Mt then M
COracle(w
? t , At , Bt [ {e})
?
12:
else Mt,e
COracle(w
? t , At [ {e}, Bt )
? t,e )
13:
pt
arg maxe2[n]\(At [Bt ) w
?t (Mt ) w
? t (M
. define w
?t (?) = 1; break ties arbitrarily
14:
if pt 2 Mt then
15:
At+1
At [ {pt }, Bt+1
Bt
16:
else
17:
At+1
At , Bt+1
Bt [ {pt }
18: Out
An+1
19: return Out
6
Related Work
The multi-armed bandit problem has been extensively studied in both stochastic and adversarial
settings [22, 3, 2]. We refer readers to [5] for a survey on recent advances. Many work in MABs focus
on minimizing the cumulative regret, which is an objective known to be fundamentally different
from the objective of pure exploration MABs [6]. Among these work, a recent line of research
considers a generalized setting called combinatorial bandits in which a set of arms (satisfying certain
combinatorial constraints) are played on each round [9, 17, 25, 7, 10, 14, 23, 21]. Note that the
objective of these work is to minimize the cumulative regret, which differs from ours.
In the literature of pure exploration MABs, the classical problem of identifying the single best arm
has been well-studied in both fixed confidence and fixed budget settings [24, 11, 6, 1, 13, 15, 16].
A flurry of recent work extend this classical problem to T OP K and MB problems and obtain algorithms with upper bounds [18, 12, 13, 19, 8, 20, 31] and worst-case lower bounds of T OP K [19, 31].
Our framework encompasses these two problems as special cases and covers a much larger class of
combinatorial pure exploration problems, which have not been addressed in current literature. Applying our results back to T OP K and MB, our upper bounds match best available problem-dependent
bounds up to constant factors [13, 19, 8] in both fixed confidence and fixed budget settings; and our
lower bound is the first proven problem-dependent lower bound for these two problems, which are
conjectured earlier by Bubeck et al. [8].
7
Conclusion
In this paper, we proposed a general framework called combinatorial pure exploration (CPE) that
can handle pure exploration tasks for many complex bandit problems with combinatorial constraints,
and have potential applications in various domains. We have shown a number of results for the
framework, including two novel learning algorithms, their related upper bounds and a novel lower
bound. The proposed algorithms support a wide range of decision classes in a unifying way and our
analysis introduced a novel tool called exchange class, which may be of independent interest. Our
upper and lower bounds characterize the complexity of the CPE problem: the sample complexity of
our algorithm is optimal (up to a logarithmic factor) for the decision classes derived from matroids
(including T OP K and MB), while for general decision classes, our upper and lower bounds are
within a relatively benign factor.
Acknowledgments. The work described in this paper was partially supported by the National Grand
Fundamental Research 973 Program of China (No. 2014CB340401 and No. 2014CB340405), the
Research Grants Council of the Hong Kong Special Administrative Region, China (Project No.
CUHK 413212 and CUHK 415113), and Microsoft Research Asia Regional Seed Fund in Big Data
Research (Grant No. FY13-RES-SPONSOR-036).
8
References
[1] J.-Y. Audibert, S. Bubeck, and R. Munos. Best arm identification in multi-armed bandits. In COLT, 2010.
[2] P. Auer, N. Cesa-Bianchi, and P. Fischer. Finite-time analysis of the multiarmed bandit problem. Machine
learning, 47(2-3):235?256, 2002.
[3] P. Auer, N. Cesa-Bianchi, Y. Freund, and R. E. Schapire. The nonstochastic multiarmed bandit problem.
SIAM Journal on Computing, 32(1):48?77, 2002.
[4] C. Berge. Two theorems in graph theory. PNAS, 1957.
[5] S. Bubeck and N. Cesa-Bianchi. Regret analysis of stochastic and nonstochastic multi-armed bandit
problems. Foundations and Trends in Machine Learning, 5:1?122, 2012.
[6] S. Bubeck, R. Munos, and G. Stoltz. Pure exploration in finitely-armed and continuous-armed bandits.
Theoretical Computer Science, 412:1832?1852, 2010.
[7] S. Bubeck, N. Cesa-bianchi, S. M. Kakade, S. Mannor, N. Srebro, and R. C. Williamson. Towards minimax policies for online linear optimization with bandit feedback. In COLT, 2012.
[8] S. Bubeck, T. Wang, and N. Viswanathan. Multiple identifications in multi-armed bandits. In ICML,
pages 258?265, 2013.
[9] N. Cesa-Bianchi and G. Lugosi. Combinatorial bandits. JCSS, 78(5):1404?1422, 2012.
[10] W. Chen, Y. Wang, and Y. Yuan. Combinatorial multi-armed bandit: General framework and applications.
In ICML, pages 151?159, 2013.
[11] E. Even-Dar, S. Mannor, and Y. Mansour. Action elimination and stopping conditions for the multi-armed
bandit and reinforcement learning problems. JMLR, 2006.
[12] V. Gabillon, M. Ghavamzadeh, A. Lazaric, and S. Bubeck. Multi-bandit best arm identification. In NIPS.
2011.
[13] V. Gabillon, M. Ghavamzadeh, and A. Lazaric. Best arm identification: A unified approach to fixed budget
and fixed confidence. In NIPS, 2012.
[14] A. Gopalan, S. Mannor, and Y. Mansour. Thompson sampling for complex online problems. In ICML,
pages 100?108, 2014.
[15] K. Jamieson and R. Nowak. Best-arm identification algorithms for multi-armed bandits in the fixed
confidence setting. In Information Sciences and Systems (CISS), pages 1?6. IEEE, 2014.
[16] K. Jamieson, M. Malloy, R. Nowak, and S. Bubeck. lil?UCB: An optimal exploration algorithm for
multi-armed bandits. COLT, 2014.
[17] S. Kale, L. Reyzin, and R. E. Schapire. Non-stochastic bandit slate problems. In NIPS, 2010.
[18] S. Kalyanakrishnan and P. Stone. Efficient selection of multiple bandit arms: Theory and practice. In
ICML, pages 511?518, 2010.
[19] S. Kalyanakrishnan, A. Tewari, P. Auer, and P. Stone. PAC subset selection in stochastic multi-armed
bandits. In ICML, pages 655?662, 2012.
[20] E. Kaufmann and S. Kalyanakrishnan. Information complexity in bandit subset selection. In COLT, 2013.
[21] B. Kveton, Z. Wen, A. Ashkan, H. Eydgahi, and B. Eriksson. Matroid bandits: Fast combinatorial optimization with learning. In UAI, 2014.
[22] T. L. Lai and H. Robbins. Asymptotically efficient adaptive allocation rules. Advances in applied mathematics, 6(1):4?22, 1985.
[23] T. Lin, B. Abrahao, R. Kleinberg, J. Lui, and W. Chen. Combinatorial partial monitoring game with linear
feedback and its application. In ICML, 2014.
[24] S. Mannor and J. N. Tsitsiklis. The sample complexity of exploration in the multi-armed bandit problem.
The Journal of Machine Learning Research, 5:623?648, 2004.
[25] G. Neu, A. Gy?orgy, and C. Szepesv?ari. The online loop-free stochastic shortest-path problem. In COLT,
pages 231?243, 2010.
[26] J. G. Oxley. Matroid theory. Oxford university press, 2006.
[27] D. Pollard. Asymptopia. Manuscript, Yale University, Dept. of Statist., New Haven, Connecticut, 2000.
[28] O. Rivasplata. Subgaussian random variables: An expository note. 2012.
[29] S. M. Ross. Stochastic processes, volume 2. John Wiley & Sons New York, 1996.
[30] N. Spring, R. Mahajan, and D. Wetherall. Measuring ISP topologies with rocketfuel. ACM SIGCOMM
Computer Communication Review, 32(4):133?145, 2002.
[31] Y. Zhou, X. Chen, and J. Li. Optimal PAC multiple arm identification with applications to crowdsourcing.
In ICML, 2014.
9
| 5433 |@word cpe:24 kong:2 exploitation:1 version:1 kalyanakrishnan:3 series:3 contains:4 ours:1 existing:2 current:2 com:1 nt:1 must:1 john:1 partition:2 benign:3 analytic:1 cis:1 designed:1 update:3 fund:1 beginning:2 oneto:1 characterization:1 mannor:4 cse:1 successive:3 five:1 unbounded:1 become:1 yuan:1 prove:2 introduce:2 hardness:2 indeed:1 expected:11 nor:1 multi:16 resolve:3 armed:15 begin:4 project:1 moreover:1 maximizes:1 null:2 kind:1 minimizes:1 unified:4 finding:5 transformation:2 guarantee:2 pseudo:1 clucb:36 tie:2 returning:1 mmb:2 connecticut:1 grant:2 jamieson:2 before:3 tsinghua:2 analyzing:1 oxford:1 path:13 lugosi:1 might:1 initialization:2 studied:2 china:2 suggests:2 conversely:2 tian:1 range:4 directed:1 unique:1 acknowledgment:1 kveton:1 practice:1 block:1 regret:3 differs:2 procedure:3 axiom:1 empirical:9 reject:5 matching:3 confidence:24 numbered:1 cannot:3 convenience:1 selection:3 operator:2 eriksson:1 applying:6 conventional:1 equivalent:4 lyu1:1 kale:1 attention:1 thompson:1 survey:1 formulate:1 identifying:7 pure:23 rule:1 pull:6 handle:1 notion:3 crowdsourcing:3 sar:1 construction:4 play:6 suppose:5 pt:15 us:5 element:4 trend:1 satisfying:1 continues:1 thinnest:1 observed:1 role:2 jcss:1 wang:2 worst:1 region:1 ensures:1 cycle:2 observes:2 environment:1 complexity:37 reward:28 flurry:1 ghavamzadeh:2 raise:1 solving:2 depend:1 serve:1 bipartite:1 learner:13 basis:3 matchings:8 asymptopia:1 easily:1 isp:1 slate:1 various:1 tx:1 fast:1 describe:2 choosing:1 refined:1 whose:1 quite:1 supplementary:6 larger:1 otherwise:2 fischer:1 transform:2 final:2 online:5 hoc:1 sequence:1 propose:1 mb:19 maximal:1 loop:1 reyzin:1 achieve:1 weic:1 empty:1 i2s:1 object:4 augmenting:3 measured:1 finitely:1 op:20 received:1 eq:8 strong:1 berge:1 resemble:1 convention:1 radius:5 correct:4 stochastic:10 exploration:32 routing:1 material:6 elimination:1 exchange:28 require:3 fix:1 generalization:1 mab:3 adjusted:3 extension:1 exploring:1 hold:1 mm:9 considered:1 ground:1 exp:3 seed:1 lyu:1 mapping:2 achieves:1 adopt:1 smallest:1 combinatorial:32 oxley:1 ross:1 bridge:1 council:1 maxm:6 largest:5 robbins:1 establishes:2 tool:3 clearly:1 gaussian:7 pn:1 zhou:1 derived:4 focus:1 abrahao:1 she:3 rank:1 hk:1 adversarial:1 sense:2 am:2 dependent:10 stopping:3 bt:20 accept:4 her:3 bandit:30 borrowing:1 interested:1 overall:1 among:3 arg:7 aforementioned:1 denoted:1 colt:5 constrained:7 special:4 initialize:2 equal:3 once:2 construct:2 eliminated:1 sampling:1 represents:3 icml:7 nearly:1 t2:1 develops:1 fundamentally:1 haven:1 wen:1 simultaneously:1 national:1 interpolate:1 individual:1 phase:8 microsoft:4 attempt:1 interest:4 investigate:2 introduces:1 predominant:1 light:1 maxi2:1 edge:5 closer:1 worker:2 nowak:2 partial:1 stoltz:1 tree:7 old:1 divide:1 penalizes:1 re:2 theoretical:2 earlier:1 negating:1 cover:4 measuring:1 assignment:1 maximization:9 nr2:1 vertex:2 subset:5 entry:1 uniform:1 delay:1 characterize:1 dependency:1 chooses:2 explores:2 grand:1 fundamental:1 siam:1 michael:1 concrete:2 gabillon:2 squared:1 again:1 satisfied:1 management:1 cesa:5 choose:1 henceforth:2 admit:2 return:8 li:1 account:2 potential:2 singleton:1 gy:1 includes:3 notable:1 audibert:1 ad:2 depends:1 later:4 try:2 break:2 analyze:2 start:1 recover:2 maintains:1 minimize:2 variance:2 kaufmann:1 maximized:2 correspond:6 identify:4 yield:2 identification:10 unifies:2 advertising:2 monitoring:1 footnote:1 ashkan:1 neu:1 definition:4 e2:2 naturally:2 associated:4 recovers:1 sampled:3 stop:3 recall:1 knowledge:4 formalize:1 auer:3 back:4 manuscript:1 asia:3 wei:1 improved:1 formulation:1 done:1 evaluated:1 generality:1 furthermore:4 rejected:5 until:1 hand:2 lack:1 mtop:2 e2m:1 atch:3 building:1 contain:1 concept:2 verify:1 hence:3 excluded:1 symmetric:1 mahajan:1 round:12 during:4 game:10 width:24 encourages:1 hong:2 generalized:1 stone:2 mabs:8 tt:5 tn:1 dedicated:2 radt:10 novel:8 ari:1 mt:30 nh:4 volume:1 tail:4 extend:1 refer:1 multiarmed:3 ai:1 trivially:1 mathematics:1 similarly:2 access:5 etc:1 base:2 own:1 recent:5 conjectured:1 irrelevant:1 belongs:1 route:1 certain:5 arbitrarily:2 seen:3 additional:2 shortest:3 cuhk:3 period:2 multiple:4 full:1 pnas:1 stem:2 match:3 long:1 lin:1 lai:1 sigcomm:1 specializing:1 a1:3 sponsor:1 basic:1 expectation:4 represent:1 addition:6 want:1 szepesv:1 addressed:1 else:3 source:1 regional:1 member:6 call:7 subgaussian:2 revealed:1 easy:1 fit:1 finish:1 matroid:6 nonstochastic:2 topology:1 idea:3 cn:1 tradeoff:2 t0:1 pollard:1 york:1 remark:1 dar:1 action:1 ignored:1 useful:1 generally:1 clear:2 detailed:1 gopalan:1 tewari:1 extensively:1 statist:1 schapire:2 notice:2 disjoint:3 lazaric:2 group:2 key:2 drawn:1 neither:1 graph:7 asymptotically:1 sum:3 inverse:1 uncertainty:1 family:3 reader:1 patch:8 decision:55 prefer:1 appendix:5 bound:57 played:1 display:1 yale:1 oracle:27 nontrivial:1 constraint:5 kleinberg:1 optimality:2 min:1 spring:1 relatively:3 conjecture:5 viswanathan:1 expository:1 belonging:1 son:1 kakade:1 making:1 intuitively:4 pr:4 fail:1 know:1 end:3 available:4 generalizes:2 malloy:1 apply:1 observe:1 eydgahi:1 top:5 denotes:2 include:2 unifying:1 exploit:1 chinese:1 establish:3 build:2 classical:2 objective:8 quantity:2 strategy:1 link:1 separate:1 mail:1 considers:1 trivial:2 spanning:7 besides:2 code:2 modeled:2 minimizing:1 design:2 lil:1 policy:1 unknown:2 bianchi:5 upper:20 finite:2 defining:1 extended:2 communication:2 excluding:1 rn:11 mansour:2 introduced:1 namely:1 pair:2 accepts:1 chen1:1 established:1 nip:3 recurring:1 challenge:5 encompasses:5 program:1 including:11 max:2 belief:1 rocketfuel:1 natural:2 arm:74 representing:1 minimax:1 review:1 literature:6 freund:1 loss:1 permutation:1 interesting:2 allocation:1 acyclic:1 proven:1 srebro:1 h2:5 foundation:1 incurred:1 shouyuan:1 proxy:1 playing:1 supported:2 last:2 free:3 offline:1 tsitsiklis:1 allow:2 pulled:1 wide:4 characterizing:1 munos:2 matroids:10 feedback:2 calculated:1 valid:1 cumulative:2 rich:2 computes:2 author:1 collection:13 refinement:1 reinforcement:1 adaptive:1 active:3 uai:1 b1:1 continuous:1 nature:1 inherently:1 contributes:1 orgy:1 williamson:1 complex:2 domain:4 main:1 rh:1 big:1 gadget:1 site:1 en:1 wiley:1 sub:5 candidate:1 administrative:1 jmlr:1 theorem:11 pac:4 symbol:2 r2:4 evidence:1 intrinsic:1 intractable:1 exists:1 te:1 execution:1 budget:17 demand:1 gap:12 chen:3 intersection:1 logarithmic:8 simply:1 bubeck:13 intern:1 ordered:1 partially:2 corresponds:3 satisfies:4 relies:1 acm:1 ma:4 slot:1 goal:1 formulated:1 king:1 towards:1 feasible:2 included:2 specifically:1 lui:1 called:11 total:1 accepted:5 ucb:1 productivity:1 formally:6 support:4 irwin:1 dept:1 ex:1 |
4,898 | 5,434 | From Stochastic Mixability to Fast Rates
Robert C. Williamson
Research School of Computer Science
Australian National University and NICTA
[email protected]
Nishant A. Mehta
Research School of Computer Science
Australian National University
[email protected]
Abstract
Empirical risk minimization (ERM) is a fundamental learning rule for statistical
learning problems where the data is generated according to some unknown distribution P and returns a hypothesis f chosen from a fixed class F with small loss
? `.
In the parametric setting, depending upon (`, F, P) ERM can have slow (1/ n)
or fast (1/n) rates of convergence of the excess risk as a function of the sample
size n. There exist several results that give sufficient conditions for fast rates in
terms of joint properties of `, F, and P, such as the margin condition and the Bernstein condition. In the non-statistical prediction with expert advice setting, there
is an analogous slow and fast rate phenomenon, and it is entirely characterized in
terms of the mixability of the loss ` (there being no role there for F or P). The
notion of stochastic mixability builds a bridge between these two models of learning, reducing to classical mixability in a special case. The present paper presents
a direct proof of fast rates for ERM in terms of stochastic mixability of (`, F, P),
and in so doing provides new insight into the fast-rates phenomenon. The proof
exploits an old result of Kemperman on the solution to the general moment problem. We also show a partial converse that suggests a characterization of fast rates
for ERM in terms of stochastic mixability is possible.
1
Introduction
Recent years have unveiled central contact points between the areas of statistical and online learning.
These include Abernethy et al.?s [1] unified Bregman-divergence based analysis of online convex
optimization and statistical learning, the online-to-batch conversion of the exponentially weighted
average forecaster (a special case of the aggregating algorithm for mixable losses) which yields the
progressive mixture rule as can be seen e.g. from the work of Audibert [2], and most recently Van
Erven et al.?s [21] injection of the concept of mixability into the statistical learning space in the form
of stochastic mixability. It is this last connection that will be our departure point for this work.
Mixability is a fundamental property of a loss that characterizes when constant regret is possible in
the online learning game of prediction with expert advice [23]. Stochastic mixability is a natural
adaptation of mixability to the statistical learning setting; in fact, in the special case where the function class consists of all possible functions from the input space to the prediction space, stochastic
mixability is equivalent to mixability [21]. Just as Vovk and coworkers (see e.g. [24, 8]) have developed a rich convex geometric understanding of mixability, stochastic mixability can be understood
as a sort of effective convexity.
In this work, we study the O(1/n)-fast rate phenomenon in statistical learning from the perspective
of stochastic mixability. Our motivation is that stochastic mixability might characterize fast rates in
statistical learning. As a first step, Theorem 5 herein establishes via a rather direct argument that
stochastic mixability implies an exact oracle inequality (i.e. with leading constant 1) with a fast rate
for finite function classes, and Theorem 7 extends this result to VC-type classes. This result can be
understood as a new chapter in an evolving narrative that started with Lee et al.?s [13] seminal paper
1
showing fast rates for agnostic learning with squared loss over convex function classes, and that was
continued by Mendelson [18] who showed that fast rates are possible for p-losses (y, y?) 7? |y ? y?|p
over effectively convex function classes by passing through a Bernstein condition (defined in (12)).
We also show that when stochastic mixability does not hold in a certain sense (described in Section 5), then the risk minimizer is not unique in a bad way. This is precisely the situation at the
heart of the works of Mendelson [18] and Mendelson and Williamson [19], which show that having
non-unique minimizers is symptomatic of bad geometry of the learning problem. In such situations,
there are certain targets (i.e. output conditional distributions) close to the original target under which
empirical risk minimization learns (ERM) at a slow rate, where the guilty target depends on the sample size and the target sequence approaches the original target asymptotically. Even the best known
upper bounds have constants that blow up in the case of non-unique minimizers. Thus, whereas
stochastic mixability implies fast rates, a sort of converse is also true, where learning is hard in a
?neighborhood? of statistical learning problems for which stochastic mixability does not hold. In
addition, since a stochastically mixable problem?s function class looks convex from the perspective
of risk minimization, and since when stochastic mixability fails the function class looks non-convex
from the same perspective (it has multiple well-separated minimizers), stochastic mixability characterizes the effective convexity of the learning problem from the perspective of risk minimization.
Much of the recent work in obtaining faster learning rates in agnostic learning has taken place in settings where a Bernstein condition holds, including results based on local Rademacher complexities
[3, 10]. The Bernstein condition appears to have first been used by Bartlett and Mendelson [4] in
their analysis of ERM; this condition is subtly different from the margin condition of Mammen and
Tsybakov [15, 20], which has been used to obtain fast rates for classification. Lecu?e [12] pinpoints
that the difference between the two conditions is that the margin condition applies to the excess loss
relative to the best predictor (not necessarily in the model class) whereas the Bernstein condition
applies to the excess loss relative to the best predictor in the model class. Our approach in this work
is complementary to the approaches of previous works, coming from a different assumption that
forms a bridge to the online learning setting. Yet this assumption is related; the Bernstein condition
implies stochastic mixability under a bounded losses assumption [21]. Further understanding the
connection between the Bernstein condition and stochastic mixability is an ongoing effort.
?
Contributions. The core contribution of this work is to show a new path to the O(1/n)-fast
rate
in statistical learning. We are not aware of previous results that show fast rates from the stochastic
mixability assumption. Secondly, we establish intermediate learning rates that interpolate between
the fast and slow rate under a weaker notion of stochastic mixability. Finally, we show that in a
certain sense stochastic mixability characterizes the effective convexity of the statistical problem.
In the next section we formally define the statistical problem, review stochastic mixability, and
explain our high-level approach toward getting fast rates. This approach involves directly appealing
to the Cram?er-Chernoff method, from which nearly all known concentration inequalities arose in one
way or another. In Section 3, we frame the problem of computing a particular moment of a certain
excess loss random variable as a general moment problem. We sufficiently bound the optimal value
of the moment, which allows for a direct application of the Cram?er-Chernoff method. These results
easily imply a fast rates bound for finite classes that can be extended to parametric (VC-type) classes,
as shown in Section 4. We describe in Section 5 how stochastic mixability characterizes a certain
notion of convexity of the statistical learning problem. In Section 6, we extend the fast rates results to
classes that obey a notion we call weak stochastic mixability. Finally, Section 7 concludes this work
with connections to related topics in statistical learning theory and a discussion of open problems.
2
Stochastic mixability, Cram?er-Chernoff, and ERM
Let (`, F, P) be a statistical learning problem with ` : Y ? R ? R+ a nonnegative loss, F ? RX a
compact function class, and P a probability measure over X ? Y for input space X and output/target
space Y. Let Z be a random variable defined as Z = (X, Y ) ? P. We assume for all f ? F,
`(Y, f (X)) ? V almost surely (a.s.) for some constant V .
A probability measure P operates on functions and loss-composed functions as:
P `(?, f ) = E(X,Y )?P ` Y, f (X) .
P f = E(X,Y )?P f (X)
2
Similarly, an empirical measure Pn associated with an n-sample z, comprising n iid samples
(x1 , y1 ), . . . , (xn , yn ), operates on functions and loss-composed functions as:
n
n
1X
1X
Pn f =
f (xj )
Pn `(?, f ) =
` yj , f (xj ) .
n j=1
n j=1
Let f ? be any function for which P `(?, f ? ) = inf f ?F P `(?, f ). For each f ? F define the excess
risk random variable Zf := ` Y, f (X) ? ` Y, f ? (X) .
We frequently work with the following two subclasses. For any ? > 0, define the subclasses
F? := {f ? F : P Zf ? ?}
F? := {f ? F : P Zf ? ?} .
2.1
Stochastic mixability
For ? > 0, we say that (`, F, P) is ?-stochastically mixable if for all f ? F
log E exp(??Zf ) ? 0.
(1)
If ?-stochastic mixability holds for some ? > 0, then we say that (`, F, P) is stochastically mixable.
Throughout this paper it is assumed that the stochastic mixability condition holds, and we take ? ? to
be the largest ? such that ?-stochastic mixability holds. Condition (1) has a rich history, beginning
from the foundational thesis of Li [14] who studied the special case of ? ? = 1 in density estimation
with log loss from the perspective of information geometry. The connections that Li showed between
this condition and convexity were strengthened by Gr?unwald [6, 7] and Van Erven et al. [21].
2.2
Cram?er-Chernoff
The high-level strategy taken here is to show that with high probability ERM will not select a fixed
hypothesis function f with excess risk above na for some constant a > 0. For each hypothesis, this
guarantee will flow from the Cram?er-Chernoff method [5] by controlling the cumulant generating
function (CGF) of ?Zf in a particular way to yield exponential concentration. This control will be
possible because the ? ? -stochastic mixability condition implies that the CGF of ?Zf takes the value
0 at some ? ? ? ? , a fact later exploited by our key tool Theorem 3.
Let Z be a real-valued random variable. Applying Markov?s inequality to an exponentially transformed random variable yields that, for any ? ? 0 and t ? R
Pr(Z ? t) ? exp(??t + log E exp(?Z));
(2)
the inequality is non-trivial only if t > E Z and ? > 0.
2.3
Analysis of ERM
We consider the ERM estimator f?z := arg minf ?F Pn `(?, f ). That is, given an n-sample z, ERM
selects any f?z ? F minimizing the empirical risk Pn `(?, f ). We say ERM is ?-good when f?z ? F? .
In order to show that ERM is ?-good it is sufficient to show that for all f ? F \ F? we have
P Zf > 0. The goal is to show that with high probability ERM is ?-good, and we will do this by
showing that with high probability uniformly for all f ? F \ F? we have Pn Zf > t for some
slack t > 0 that will come in handy later.
For a real-valued random variable X, recall that the cumulant generating function of X is ? 7?
?X (?) := log E e?X ; we allow ?X (?) to be infinite for some ? > 0.
Theorem 1 (Cram?er-Chernoff Control on ERM). Let a > 0 and select f such that E Zf > 0.
Let t < E Zf . If there exists ? > 0 such that ??Zf (?) ? ? na , then
n
o
Pr Pn `(?, f ) ? Pn `(?, f ? ) + t ? exp(?a + ?t).
Pn
Proof. Let Zf,1 , . . . , Zf,n be iid copies of Zf , and define the sum Sf,n :=
j=1 ?Zf,j . Since
1
(?t) > E n Sf,n , then from (2) we have
X
n
1
1
Pr
Zf,j ? t = Pr
Sf,n ? ?t ? exp (?t + log E exp(?Sf,n ))
n j=1
n
n
= exp(?t) E exp(??Zf ) .
3
Making the replacement ??Zf (?) = log E exp(??Zf ) yields
1
log Pr
Sf,n ? ?t ? ?t + n??Zf (?).
n
By assumption, ??Zf (?) ? ? na , and so Pr{Pn Zf ? t} ? exp(?a + ?t) as desired.
This theorem will be applied by showing that for an excess loss random variable Zf taking values in
[?1, 1], if for some ? > 0 we have E exp(??Zf ) = 1 and if E Zf = na for some constant a (that can
and must depend on n), then ??Zf (?/2) ? ? c?a
n where c > 0 is a universal constant. This is the
nature of the next section. We then extend this result to random variables taking values in [?V, V ].
3
Semi-infinite linear programming and the general moment problem
The key subproblem now is to find, for each excess loss random variable Zf with mean na and
??Zf (?) = 0 (for some ? ? ? ? ), a pair of constants ?0 > 0 and c > 0 for which ??Zf (?0 ) ? ? ca
n.
Theorem 1 would then imply that ERM will prefer f ? over this particular f with high probability for
ca large enough. This subproblem is in fact an instance of the general moment problem, a problem
on which Kemperman [9] has conducted a very nice geometric study. We now describe this problem.
The general moment problem. Let P(A) be the space of probability measures over a measurable
space A = (A, S). For real-valued measurable functions h and (gj )j?[m] on a measurable space
A = (A, S), the general moment problem is
inf
EX?? h(X)
??P(A)
(3)
subject to EX?? gj (X) = yj , j ? {1, . . . , m}.
Let the vector-valued map g : A ? Rm be defined in terms of coordinate functions as (g(x))j =
gj (x), and let the vector y ? Rm be equal to (y1 , . . . , ym ).
Let D? ? Rm+1 be the set
m
X
?
?
m+1
D := d = (d0 , d1 , . . . , dm ) ? R
: h(x) ? d0 +
dj gj (x) for all x ? A .
(4)
j=1
Theorem 3 of [9] states that if y ? int conv g(A), the optimal value of problem (3) equals
m
X
?
?
dj yj : d = (d0 , d1 , . . . , dm ) ? D .
sup d0 +
(5)
j=1
Our instantiation. We choose A = [?1, 1], set m = 2 and define h, (gj )j?{1,2} , and y ? R2 as:
a
h(x) = ?e(?/2)x ,
g1 (x) = x,
g2 (x) = e?x ,
y1 = ? ,
y2 = 1,
n
for any ? > 0, a > 0, and n ? N. This yields the following instantiation of problem (3):
inf
??P([?1,1])
subject to
EX?? ?e(?/2)X
a
n
= 1.
(6a)
EX?? X = ?
(6b)
EX?? e?X
(6c)
Note that equation (5) from the general moment problem now instantiates to
n
o
a
sup d0 ? d1 + d2 : d? = (d0 , d1 , d2 ) ? D? ,
n
with D? equal to the set
n
o
d? = (d0 , d1 , d2 ) ? R3 : ?e(?/2)x ? d0 + d1 x + d2 e?x for all x ? [?1, 1] .
(7)
(8)
Applying Theorem 3 of [9] requires the condition y ? int conv g([?1, 1]). We first characterize
when y ? conv g([?1, 1]) holds and handle the int conv g([?1, 1]) version after Theorem 3.
4
Lemma 2 (Feasible Moments). The point y = ? na , 1 ? conv g([?1, 1]) if and only if
cosh(?) ? 1
a
e? + e?? ? 2
=
?
.
n
e? ? e??
sinh(?)
(9)
Proof. Let W denote the convex hull of g([?1, 1]). We need to see if ? na , 1 ? W . Note that W
is the convex set formed by starting with the graph of x 7? e?x on the domain [?1, 1], including the
line segment connecting this curve?s endpoints (?1, e?? ) to (1, e?x ), and including all of the points
below this line segment but above the aforementioned graph. That is, W is precisely the set
e? + e??
e? ? e??
W := (x, y) ? R2 : e?x ? y ?
+
x, ?x ? [?1, 1] .
2
2
It remains to check that 1 is sandwiched between the lower and upper bounds at x = ? na . Clearly
the lower bound holds. Simple algebra shows that the upper bound is equivalent to condition (9).
Note that if (9) does not hold, then the semi-infinite linear program (6) is infeasible; infeasibility in
turn implies that such an excess loss random variable cannot exist. Thus, we need not worry about
whether (9) holds; it holds for any excess loss random variable satisfying constraints (6b) and (6c).
The following theorem is a key technical result for using stochastic mixability to control the CGF.
The proof is long and can be found in Appendix A.
Theorem 3 (Stochastic Mixability Concentration). Let f be an element of F with Zf taking values in [?1, 1], n ? N, E Zf = na for some a > 0, and ??Zf (?) = 0 for some ? > 0. If
e? + e?? ? 2
a
<
,
n
e? ? e??
E e(?/2)(?Zf ) ? 1 ?
then
(10)
0.18(? ? 1)a
.
n
Note that since log(1 ? x) ? ?x when x < 1, we have ??Zf (?/2) ? ? 0.18(?n? 1)a .
In order to apply Theorem 3, we need (10) to hold, but only (9) is guaranteed to hold. The corner
case is if (9) holds with equality. However, observe that one can always approximate the random
variable X by a perturbed version X 0 which has nearly identical mean a0 ? a and a nearly identical
0
0
? 0 ? ? for which EX 0 ??0 e? X = 1, and yet the inequality in (9) is strict. Later, in the proof
of Theorem 5, for any random variable that required perturbation to satisfy the interior condition
(10), we implicitly apply the analysis to the perturbed version, show that ERM would not pick the
(slightly different) function corresponding to the perturbed version, and use the closeness of the two
functions to show that ERM also would not pick the original function.
We now present a necessary extension for the case of losses with range [0, V ], proved in Appendix A.
Lemma 4 (Bounded Losses). Let g1 (x) = x and y2 = 1 be common settings for the following two
problems. The instantiation of problem (3) with A = [?V, V ], h(x) = ?e(?/2)x , g2 (x) = e?x ,
and y1 = ? na has the same optimal value as the instantiation of problem (3) with A = [?1, 1],
h(x) = ?e(V ?/2)x , g2 (x) = e(V ?)x , and y1 = ? a/V
n .
4
Fast rates
We now show how the above results can be used to obtain an exact oracle inequality with a fast rate.
We first present a result for finite classes and then present a result for VC-type classes (classes with
logarithmic universal metric entropy).
?
Theorem 5 (Finite Classes Exact Oracle Inequality). Let (`, F,
P) be ? -stochastically mixable,
where |F| = N , ` is a nonnegative loss, and supf ?F ` Y, f (X) ? V a.s. for a constant V . Then
for all n ? 1, with probability at least 1 ? ?
n
o
6 max V, ?1? log 1? + log N
?
P `(?, f?z ) ? P `(?, f ) +
.
n
5
(?)
Proof. Let ?n = na for a constant a to be fixed later. For each ? > 0, let F?n ? F?n correspond
to those functions in F?n for which ? is the largest constant such that E exp(??Zf ) = 1. Let
hyper
F?
? F?n correspond to functions f in F?n for which lim??? E exp(??Zf ) < 1. Clearly,
n
S
(?)
hyper
F?n =
??[? ? ,?) F?n ? F?n . The excess loss random variables corresponding to elements
hyper
f ? F?
are ?hyper-concentrated? in the sense that they are infinitely stochastically mixable.
n
However, Lemma 10 in Appendix B shows that for each hyper-concentrated Zf , there exists another
excess loss random variable Zf0 with mean arbitrarily close to that of Zf , with E exp(??Zf0 ) = 1 for
some arbitrarily large but finite ?, and with Zf0 ? Zf with probability 1. The last property implies
that the empirical risk of Zf0 is no greater than that of Zf ; hence for each hyper-concentrated Zf it is
sufficient (from the perspective of ERM) to study a corresponding Zf0 . From now on, we implicitly
S
(?)
make this replacement in F?n itself, so that we now have F?n = ??[?? ,?) F?n .
(?)
Consider an arbitrary a > 0. For some fixed ? ? [? ? , ?) for which |F?n | > 0, consider
(?)
the subclass F?n . Individually for each such function, we will apply Theorem 1 as follows.
From Lemma 4, we have ??Zf (?/2) = ?? V1 Zf (V ?/2). From Theorem 3, the latter is at most
1)(a/V )
? 0.18(V ? ?
= ? (V0.18?a
n
? ? 1)n . Hence, Theorem 1 with t = 0 and the ? from the Theorem taken to be ?/2 implies that the probability of the event Pn `(?, f ) ? Pn `(?, f ? ) is at most
exp ?0.18 V ??? 1 a . Applying the union bound over all of F?n , we conclude that
0.18a
Pr {?f ? F?n : Pn `(?, f ) ? Pn `(?, f ? )} ? N exp ?? ?
.
V ?? ? 1
Since ERM selects hypotheses on their empirical risk, from inversion it holds that with probability at
6 max{V, ?1? }(log ?1 +log N )
.
least 1 ? ? ERM will not select any hypothesis with excess risk at least
n
Before presenting the result for VC-type classes, we require some definitions. For a pseudometric
space (G, d), for any ? > 0, let N (?, G, d) be the ?-covering number of (G, d); that is, N (?, G, d) is
the minimal number of balls of radius ? needed to cover G. We will further constrain the cover (the
set of centers of the balls) to be a subset of G (i.e. to be proper), thus ensuring that the stochastic
mixability assumption transfers to any (proper) cover of F. Note that the ?proper? requirement at
most doubles the constant K below, as shown by Vidyasagar [22, Lemma 2.1].
We now state a localization-based result that allows us to extend the result for finite classes to VCtype classes. Although the localization result can be obtained by combining standard techniques,1
we could not find this particular result in the literature. Below, an ?-net F? of a set F is a subset of
F such that F is contained in the union of the balls of radius ? with centers in F? .
Theorem 6. Let F be a separable function class whose functions have range bounded in [0, V ] and
for which, for a constant K ? 1, for each u ? (0, K] the L2 (P) covering numbers are bounded as
C
K
N (u, F, L2 (P)) ?
.
(11)
u
Suppose F? is a minimal ?-net for F in the L2 (P) norm, with ? = n1 . Denote by ? : F ? F? an
L2 (P)-metric projection from F to F? . Then, provided that ? ? 21 , with probability at most ? can
there exist f ? F such that
s
!
V
1
e
Pn f < Pn (?(f )) ?
1080C log(2Kn) + 90
log
C log(2Kn) + log
.
n
?
?
The proof is presented in Appendix C. We now present the fast rates result for VC-type classes.
The proof (in Appendix C) uses Theorem 6 and the proof of the Theorem 5. Below, we denote the
loss-composed version of a function class F as ` ? F := {`(?, f ) : f ? F}.
1
See e.g. the techniques of Massart and N?ed?elec [16] and equation (3.17) of Koltchinskii [11].
6
Theorem 7 (VC-Type Classes Exact Oracle Inequality). Let (`, F, P) be ? ? -stochastically mixable with ` ? F separable, where, for a constant K ? 1, for each ? ? (0, K] we have
C
N (` ? F, L2 (P), ?) ? K
, and supf ?F ` Y, f (X) ? V a.s. for a constant V ? 1. Then
?
for all n ? 5 and ? ? 12 , with probability at least 1 ? ?
o
n
?
?
C log(Kn) + log 2? ,
8 max V, ?1?
1
?
q
P `(?, fz ) ? P `(?, f ) + max
? 2V 1080C log(2Kn) + 90 log 2 C log(2Kn) + log
n
?
?
?
?
5
2e
?
?
+
1
.
n
Characterizing convexity from the perspective of risk minimization
In the following, when we say (`, F, P) has a unique minimizer we mean that any two minimizers
f1? , f2? of P `(?, f ) over F satisfy ` Y, f1? (X) = ` Y, f2? (X) a.s. We say the excess loss class
{`(?, f ) ? `(?, f ? ) : f ? F} satisfies a (?, B)-Bernstein condition with respect to P for some B > 0
and 0 < ? ? 1 if, for all f ? F:
2
?
P `(?, f ) ? `(?, f ? ) ? B P `(?, f ) ? `(?, f ? )
.
(12)
It already is known that the stochastic mixability condition guarantees that there is a unique minimizer [21]; this is a simple consequence of Jensen?s inequality. This leaves open the question: if
stochastic mixability does not hold, are there necessarily non-unique minimizers? We show that in
a certain sense this is indeed the case, in bad way: the set of minimizers will be a disconnected set.
For any ? > 0, define G? as the class G? := {f ? } ? f ? F : kf ? f ? kL1 (P) ? ? , where in case
there are multiple minimizers in F we arbitrarily select one of them as f ? . Since we assume that F
is compact and G? \ {f ? } is equal to F minus an open set homeomorphic to the unit L1 (P) ball,
G? \ {f ? } is also compact.
Theorem 8 (Non-Unique Minimizers). Suppose there exists some ? > 0 such that G? is not
stochastically mixable. Then there are minimizers
f1? , f2? ? F of P `(?, f ) over F such that it is
?
?
not the case that ` Y, f1 (X) = ` Y, f2 (X) a.s.
Proof. Select ? > 0 as in the theorem and some fixed ? > 0. Since G? is not ?-stochastically
mixable, there exists f? ? G? such that ??Zf? (?) > 0. Note that there exists ? 0 ? (0, ?) with
??Zf? (? 0 ) = 0; if not, lim??0
??Zf (?)???Zf
?
?
?
(0)
> 0 ? ?0?Zf? (0) > 0, so ?0?Zf? (0) = E(?Zf? )
implies that E Zf? < 0, a contradiction! From Lemma 2, E Zf? ?
0
cosh(? 0 )?1
sinh(? 0 ) ;
for ? 0 ? 0 the RHS
0
0
)?1
2 0
1
has upper bound ?2 since the derivative of ?2 ? cosh(?
sinh(? 0 ) is the nonnegative function 2 tanh (? /2)
0
0
)?1
and ?2 ? cosh(?
|?0 =0 = 0. Thus, E Zf? ? 0 as ? ? 0. As G? \ {f ? } is compact, we can take
sinh(? 0 )
a positive decreasing sequence (?j )j approaching 0, corresponding to a sequence (f?j )j ? G? \{f ? }
with limit point g ? ? G? \ {f ? } for which E Zg? = 0, and so there is a risk minimizer in G? \ {f ? }.
The implications of having non-unique risk minimizers. In the case of non-unique risk minimizers, Mendelson [17] showed that for p-losses (y, y?) 7? |y ? y?|p with p ? [2, ?) there is an
n-indexed sequence of probability measures (P(n) )n approaching the true probability measure as
n ? ? such that, for each n, ERM learns at a slow rate under sample size n when the true distribution is P(n) . This behavior is a consequence of the statistical learning problem?s poor geometry:
there are multiple minimizers and the set of minimizers is not even connected. Furthermore, in this
case, the best known fast rate upper bounds (see [18] and [19]) have a multiplicative constant that
approaches ? as the target probability measure approaches a probability measure for which there
are non-unique minimizers. The reason for the poor upper bounds in this case is that the constant B
in the Bernstein condition explodes, and the upper bounds rely upon the Bernstein condition.
6
Weak stochastic mixability
For some ? ? [0, 1], we say (`, F, P) is (?, ?0 )-weakly stochastically mixable if, for every ? > 0, for
all f ? {f ? } ? F? , the inequality log E exp(??? Zf ) ? 0 holds with ?? := ?0 ?1?? . This concept
was introduced by Van Erven et al. [21] without a name.
7
Suppose that some fixed function has excess risk a = ?. Then, roughly, with high probability
ERM does not make a mistake provided that a?a = n1 , i.e. when ? ? ?0 ?1?? = n1 and hence when
? = (?0 n)?1/(2??) . Modifying the proof of the finite classes result (Theorem 5) to consider all
functions in the subclass F?n for ?n = (?0 n)?1/(2??) yields the following corollary of Theorem 5.
Corollary 9. Let (`, F, P) be (?, ?0 )-weakly stochastically
mixable for some ? ? [0, 1], where
|F| = N , ` is a nonnegative loss, and supf ?F ` Y, f (X) ? V a.s. for a constant V . Then for any
n ? ?10 V (1??)/(2??) , with probability at least 1 ? ?
6 log 1? + log N
?
?
P `(?, fz ) ? P `(?, f ) +
.
(?0 n)1/(2??)
It is simple to show a similar result for VC-type classes; the ?-net can still be taken at the resolution
1
?1/(2??)
.
n , but we need only apply the analysis to the subclass of F with excess risk at least (?0 n)
7
Discussion
We have shown that stochastic mixability implies fast rates for VC-type classes, using a direct argument based on the Cram?er-Chernoff method and sufficient control of the optimal value of a certain
instance of the general moment problem. The approach is amenable to localization in that the analysis separately controls the probability of large deviations for individual elements of F. An important
open problem is to extend the results presented here for VC-type classes to results for nonparametric
classes with polynomial metric entropy, and moreover, to achieve rates similar to those obtained for
these classes under the Bernstein condition.
There are still some unanswered questions with regards to the connection between the Bernstein
condition and stochastic mixability. Van Erven et al. [21] showed that for bounded losses the Bernstein condition implies stochastic mixability. Therefore, when starting from a Bernstein condition,
Theorem 5 offers a different path to fast rates. An open problem is to settle the question of whether
the Bernstein condition and stochastic mixability are equivalent. Previous results [21] suggest that
the stochastic mixability does imply a Bernstein condition, but the proof was non-constructive, and
it relied upon a bounded losses assumption. It is well known (and easy to see) that both stochastic
mixability and the Bernstein condition hold only if there is a unique minimizer. Theorem 8 shows in
a certain sense that if stochastic mixability does not hold, then there cannot be a unique minimizer.
Is the same true when the Bernstein condition fails to hold? Regardless of whether stochastic mixability is equivalent to the Bernstein condition, the direct argument presented here and the connection
to classical mixability, which does characterize constant regret in the simpler non-stochastic setting,
motivates further study of stochastic mixability.
Finally, it would be of great interest to discard the bounded losses assumption. Ignoring the dependence of the metric entropy on the maximum possible loss, the upper bound on the loss V enters the
final bound through the difficulty of controlling the minimum value of u? (?1) when ? is large (see
the proof of Theorem 3). From extensive experiments with a grid-approximation linear program,
we have observed that the worst (CGF-wise) random variables for fixed negative mean and fixed
optimal stochastic mixability constant are those which place very little probability mass at ?V and
most of the probability mass at a small positive number that scales with the mean. These random
variables correspond to functions that with low probability beat f ? by a large (loss) margin but with
high probability have slightly higher loss than f ? . It would be useful to understand if this exotic
behavior is a real concern and, if not, find a simple, mild condition on the moments that rules it out.
Acknowledgments
RCW thanks Tim van Erven for the initial discussions around the Cram?er-Chernoff method during
his visit to Canberra in 2013 and for his gracious permission to proceed with the present paper
without him as an author, and both authors thank him for the further enormously helpful spotting
of a serious error in our original proof for fast rates for VC-type classes. This work was supported
by the Australian Research Council (NAM and RCW) and NICTA (RCW). NICTA is funded by the
Australian Government through the Department of Communications and the Australian Research
Council through the ICT Centre of Excellence program.
8
References
[1] Jacob Abernethy, Alekh Agarwal, Peter L. Bartlett, and Alexander Rakhlin. A stochastic view of optimal
regret through minimax duality. In Proceedings of the 22nd Annual Conference on Learning Theory
(COLT 2009), 2009.
[2] Jean-Yves Audibert. Fast learning rates in statistical inference through aggregation. The Annals of Statistics, 37(4):1591?1646, 2009.
[3] Peter L. Bartlett, Olivier Bousquet, and Shahar Mendelson. Local Rademacher complexities. The Annals
of Statistics, 33(4):1497?1537, 2005.
[4] Peter L. Bartlett and Shahar Mendelson. Empirical minimization. Probability Theory and Related Fields,
135(3):311?334, 2006.
[5] St?ephane Boucheron, G?abor Lugosi, and Pascal Massart. Concentration inequalities: A nonasymptotic
theory of independence. Oxford University Press, 2013.
[6] Peter Gr?unwald. Safe learning: bridging the gap between Bayes, MDL and statistical learning theory via
empirical convexity. In Proceedings of the 24th International Conference on Learning Theory (COLT
2011), pages 397?419, 2011.
[7] Peter Gr?unwald. The safe Bayesian. In Proceedings of the 23rd International Conference on Algorithmic
Learning Theory (ALT 2012), pages 169?183. Springer, 2012.
[8] Yuri Kalnishkan and Michael V. Vyugin. The weak aggregating algorithm and weak mixability. In
Proceedings of the 18th Annual Conference on Learning Theory (COLT 2005), pages 188?203. Springer,
2005.
[9] Johannes H.B. Kemperman. The general moment problem, a geometric approach. The Annals of Mathematical Statistics, 39(1):93?122, 1968.
[10] Vladimir Koltchinskii. Local Rademacher complexities and oracle inequalities in risk minimization. The
Annals of Statistics, 34(6):2593?2656, 2006.
[11] Vladimir Koltchinskii. Oracle Inequalities in Empirical Risk Minimization and Sparse Recovery Problems: Ecole dEt?e de Probabilit?es de Saint-Flour XXXVIII-2008, volume 2033. Springer, 2011.
[12] Guillaume Lecu?e. Interplay between concentration, complexity and geometry in learning theory with
applications to high dimensional data analysis. Habilitation a` diriger des recherches, Universit?e ParisEst, 2011.
[13] Wee Sun Lee, Peter L. Bartlett, and Robert C. Williamson. The importance of convexity in learning with
squared loss. IEEE Transactions on Information Theory, 44(5):1974?1980, 1998.
[14] Jonathan Qiang Li. Estimation of mixture models. PhD thesis, Yale University, 1999.
[15] Enno Mammen and Alexandre B. Tsybakov. Smooth discrimination analysis. The Annals of Statistics,
27(6):1808?1829, 1999.
?
[16] Pascal Massart and Elodie
N?ed?elec. Risk bounds for statistical learning. The Annals of Statistics,
34(5):2326?2366, 2006.
[17] Shahar Mendelson. Lower bounds for the empirical minimization algorithm. IEEE Transactions on
Information Theory, 54(8):3797?3803, 2008.
[18] Shahar Mendelson. Obtaining fast error rates in nonconvex situations. Journal of Complexity, 24(3):380?
397, 2008.
[19] Shahar Mendelson and Robert C. Williamson. Agnostic learning nonconvex function classes. In Proceedings of the 15th Annual Conference on Computational Learning Theory (COLT 2002), pages 1?13.
Springer, 2002.
[20] Alexander B. Tsybakov. Optimal aggregation of classifiers in statistical learning. The Annals of Statistics,
32(1):135?166, 2004.
[21] Tim Van Erven, Peter D. Gr?unwald, Mark D. Reid, and Robert C. Williamson. Mixability in statistical
learning. In Advances in Neural Information Processing Systems 25 (NIPS 2012), pages 1700?1708,
2012.
[22] Mathukumalli Vidyasagar. Learning and Generalization with Applications to Neural Networks. Springer,
2002.
[23] Volodya Vovk. A game of prediction with expert advice. Journal of Computer and System Sciences,
56(2):153?173, 1998.
[24] Volodya Vovk. Competitive on-line statistics. International Statistical Review, 69(2):213?248, 2001.
9
| 5434 |@word mild:1 version:5 inversion:1 polynomial:1 norm:1 nd:1 mehta:2 open:5 d2:4 forecaster:1 jacob:1 pick:2 minus:1 moment:13 initial:1 ecole:1 erven:6 yet:2 must:1 discrimination:1 leaf:1 beginning:1 core:1 provides:1 characterization:1 simpler:1 mathematical:1 direct:5 consists:1 excellence:1 indeed:1 roughly:1 behavior:2 frequently:1 decreasing:1 little:1 conv:5 provided:2 bounded:7 moreover:1 exotic:1 agnostic:3 mass:2 developed:1 unified:1 guarantee:2 every:1 subclass:5 universit:1 rm:3 classifier:1 control:5 unit:1 converse:2 yn:1 reid:1 before:1 positive:2 understood:2 aggregating:2 local:3 limit:1 consequence:2 mistake:1 oxford:1 path:2 lugosi:1 might:1 au:2 studied:1 koltchinskii:3 suggests:1 range:2 unique:12 acknowledgment:1 yj:3 union:2 regret:3 handy:1 probabilit:1 area:1 foundational:1 universal:2 evolving:1 empirical:10 projection:1 cram:8 suggest:1 cannot:2 close:2 interior:1 risk:21 applying:3 seminal:1 equivalent:4 measurable:3 map:1 center:2 regardless:1 starting:2 convex:8 resolution:1 recovery:1 contradiction:1 rule:3 insight:1 continued:1 estimator:1 nam:1 his:2 unanswered:1 handle:1 notion:4 coordinate:1 analogous:1 annals:7 target:7 controlling:2 suppose:3 mixable:11 exact:4 programming:1 olivier:1 us:1 hypothesis:5 element:3 satisfying:1 observed:1 role:1 subproblem:2 enters:1 worst:1 connected:1 sun:1 convexity:8 complexity:5 depend:1 weakly:2 segment:2 algebra:1 subtly:1 upon:3 localization:3 f2:4 easily:1 joint:1 chapter:1 separated:1 elec:2 fast:29 effective:3 describe:2 hyper:6 neighborhood:1 abernethy:2 whose:1 jean:1 valued:4 say:6 statistic:8 g1:2 itself:1 final:1 online:5 interplay:1 sequence:4 net:3 coming:1 adaptation:1 combining:1 achieve:1 getting:1 convergence:1 double:1 requirement:1 rademacher:3 generating:2 tim:2 depending:1 school:2 involves:1 implies:10 australian:5 come:1 safe:2 radius:2 modifying:1 stochastic:48 vc:10 hull:1 settle:1 require:1 government:1 f1:4 generalization:1 secondly:1 extension:1 hold:20 sufficiently:1 around:1 exp:17 great:1 algorithmic:1 enno:1 narrative:1 estimation:2 tanh:1 bridge:2 individually:1 largest:2 him:2 council:2 establishes:1 tool:1 weighted:1 minimization:9 clearly:2 always:1 rather:1 arose:1 pn:16 corollary:2 check:1 sense:5 helpful:1 inference:1 minimizers:14 habilitation:1 a0:1 abor:1 transformed:1 comprising:1 selects:2 arg:1 classification:1 aforementioned:1 colt:4 pascal:2 mathukumalli:1 special:4 equal:4 aware:1 field:1 having:2 chernoff:8 identical:2 progressive:1 qiang:1 look:2 nearly:3 minf:1 ephane:1 serious:1 composed:3 wee:1 national:2 interpolate:1 individual:1 divergence:1 geometry:4 replacement:2 n1:3 interest:1 flour:1 mdl:1 mixture:2 unveiled:1 implication:1 amenable:1 bregman:1 partial:1 necessary:1 indexed:1 old:1 desired:1 minimal:2 instance:2 cgf:4 cover:3 deviation:1 subset:2 kl1:1 predictor:2 conducted:1 gr:4 characterize:3 kn:5 elodie:1 perturbed:3 thanks:1 density:1 international:3 fundamental:2 st:1 lee:2 michael:1 connecting:1 ym:1 na:11 squared:2 central:1 thesis:2 choose:1 stochastically:10 corner:1 expert:3 derivative:1 leading:1 return:1 li:3 nonasymptotic:1 de:3 blow:1 int:3 satisfy:2 audibert:2 depends:1 later:4 multiplicative:1 view:1 doing:1 characterizes:4 sup:2 competitive:1 sort:2 relied:1 aggregation:2 bayes:1 contribution:2 formed:1 yves:1 who:2 yield:6 correspond:3 weak:4 bayesian:1 iid:2 rx:1 bob:1 history:1 explain:1 ed:2 definition:1 dm:2 proof:15 associated:1 proved:1 recall:1 lim:2 appears:1 worry:1 alexandre:1 higher:1 symptomatic:1 furthermore:1 just:1 name:1 concept:2 true:4 y2:2 equality:1 hence:3 boucheron:1 recherches:1 game:2 during:1 covering:2 mammen:2 presenting:1 l1:1 wise:1 recently:1 common:1 endpoint:1 exponentially:2 volume:1 extend:4 rd:1 grid:1 similarly:1 centre:1 dj:2 funded:1 alekh:1 gj:5 v0:1 recent:2 showed:4 perspective:7 inf:3 discard:1 certain:8 nonconvex:2 inequality:13 shahar:5 arbitrarily:3 lecu:2 yuri:1 exploited:1 seen:1 minimum:1 greater:1 surely:1 coworkers:1 semi:2 multiple:3 d0:8 smooth:1 technical:1 faster:1 characterized:1 offer:1 long:1 visit:1 ensuring:1 prediction:4 metric:4 agarwal:1 whereas:2 addition:1 separately:1 strict:1 massart:3 subject:2 explodes:1 flow:1 call:1 bernstein:19 intermediate:1 enough:1 easy:1 xj:2 independence:1 approaching:2 det:1 whether:3 bartlett:5 bridging:1 effort:1 peter:7 passing:1 proceed:1 useful:1 johannes:1 nonparametric:1 tsybakov:3 cosh:4 concentrated:3 kalnishkan:1 fz:2 exist:3 key:3 rcw:3 v1:1 asymptotically:1 graph:2 year:1 sum:1 extends:1 place:2 almost:1 throughout:1 prefer:1 appendix:5 entirely:1 bound:15 sinh:4 guaranteed:1 yale:1 oracle:6 nonnegative:4 annual:3 precisely:2 constraint:1 constrain:1 bousquet:1 vyugin:1 argument:3 pseudometric:1 separable:2 injection:1 department:1 according:1 ball:4 disconnected:1 instantiates:1 poor:2 slightly:2 appealing:1 making:1 pr:7 erm:23 heart:1 taken:4 equation:2 remains:1 slack:1 r3:1 turn:1 needed:1 apply:4 obey:1 observe:1 batch:1 permission:1 original:4 include:1 saint:1 exploit:1 homeomorphic:1 build:1 establish:1 classical:2 sandwiched:1 mixability:56 contact:1 already:1 question:3 parametric:2 concentration:5 strategy:1 dependence:1 thank:1 topic:1 guilty:1 trivial:1 toward:1 nicta:3 reason:1 minimizing:1 vladimir:2 robert:4 negative:1 proper:3 motivates:1 unknown:1 conversion:1 upper:8 zf:54 markov:1 finite:7 beat:1 situation:3 extended:1 communication:1 frame:1 y1:5 perturbation:1 arbitrary:1 introduced:1 pair:1 required:1 extensive:1 connection:6 nishant:2 herein:1 nip:1 spotting:1 below:4 departure:1 program:3 including:3 max:4 vidyasagar:2 event:1 natural:1 rely:1 difficulty:1 minimax:1 imply:3 started:1 concludes:1 review:2 geometric:3 understanding:2 nice:1 literature:1 l2:5 kf:1 relative:2 ict:1 loss:35 sufficient:4 supported:1 last:2 copy:1 infeasible:1 weaker:1 allow:1 understand:1 taking:3 characterizing:1 sparse:1 van:6 regard:1 curve:1 xn:1 rich:2 author:2 transaction:2 excess:16 approximate:1 compact:4 implicitly:2 instantiation:4 assumed:1 conclude:1 infeasibility:1 nature:1 transfer:1 ca:2 ignoring:1 obtaining:2 williamson:6 necessarily:2 domain:1 rh:1 motivation:1 complementary:1 x1:1 advice:3 canberra:1 enormously:1 strengthened:1 slow:5 fails:2 pinpoint:1 exponential:1 sf:5 learns:2 theorem:29 bad:3 showing:3 er:8 jensen:1 r2:2 rakhlin:1 alt:1 closeness:1 concern:1 mendelson:10 exists:5 effectively:1 importance:1 phd:1 anu:2 margin:4 gap:1 supf:3 entropy:3 logarithmic:1 infinitely:1 contained:1 g2:3 volodya:2 applies:2 springer:5 minimizer:6 satisfies:1 conditional:1 goal:1 feasible:1 hard:1 infinite:3 reducing:1 vovk:3 operates:2 uniformly:1 lemma:6 duality:1 e:1 unwald:4 zg:1 formally:1 select:5 guillaume:1 mark:1 latter:1 jonathan:1 cumulant:2 alexander:2 ongoing:1 constructive:1 d1:6 phenomenon:3 ex:6 |
4,899 | 5,435 | Beyond Disagreement-based Agnostic Active
Learning
Chicheng Zhang
University of California, San Diego
9500 Gilman Drive, La Jolla, CA 92093
[email protected]
Kamalika Chaudhuri
University of California, San Diego
9500 Gilman Drive, La Jolla, CA 92093
[email protected]
Abstract
We study agnostic active learning, where the goal is to learn a classi?er in a prespeci?ed hypothesis class interactively with as few label queries as possible, while
making no assumptions on the true function generating the labels. The main algorithm for this problem is disagreement-based active learning, which has a high
label requirement. Thus a major challenge is to ?nd an algorithm which achieves
better label complexity, is consistent in an agnostic setting, and applies to general
classi?cation problems.
In this paper, we provide such an algorithm. Our solution is based on two novel
contributions; ?rst, a reduction from consistent active learning to con?dence-rated
prediction with guaranteed error, and second, a novel con?dence-rated predictor.
1
Introduction
In this paper, we study active learning of classi?ers in an agnostic setting, where no assumptions
are made on the true function that generates the labels. The learner has access to a large pool of
unlabelled examples, and can interactively request labels for a small subset of these; the goal is to
learn an accurate classi?er in a pre-speci?ed class with as few label queries as possible. Speci?cally,
we are given a hypothesis class H and a target ?, and our aim is to ?nd a binary classi?er in H
whose error is at most ? more than that of the best classi?er in H, while minimizing the number of
requested labels.
There has been a large body of previous work on active learning; see the surveys by [10, 28] for
overviews. The main challenge in active learning is ensuring consistency in the agnostic setting
while still maintaining low label complexity. In particular, a very natural approach to active learning
is to view it as a generalization of binary search [17, 9, 27]. While this strategy has been extended
to several different noise models [23, 27, 26], it is generally inconsistent in the agnostic case [11].
The primary algorithm for agnostic active learning is called disagreement-based active learning.
The main idea is as follows. A set Vk of possible risk minimizers is maintained with time, and the
label of an example x is queried if there exist two hypotheses h1 and h2 in Vk such that h1 (x) ?=
h2 (x). This algorithm is consistent in the agnostic setting [7, 2, 12, 18, 5, 19, 6, 24]; however, due
to the conservative label query policy, its label requirement is high. A line of work due to [3, 4, 1]
have provided algorithms that achieve better label complexity for linear classi?cation on the uniform
distribution over the unit sphere as well as log-concave distributions; however, their algorithms are
limited to these speci?c cases, and it is unclear how to apply them more generally.
Thus, a major challenge in the agnostic active learning literature has been to ?nd a general active
learning strategy that applies to any hypothesis class and data distribution, is consistent in the agnostic case, and has a better label requirement than disagreement based active learning. This has been
mentioned as an open problem by several works, such as [2, 10, 4].
1
In this paper, we provide such an algorithm. Our solution is based on two key contributions, which
may be of independent interest. The ?rst is a general connection between con?dence-rated predictors and active learning. A con?dence-rated predictor is one that is allowed to abstain from
prediction on occasion, and as a result, can guarantee a target prediction error. Given a con?dencerated predictor with guaranteed error, we show how to to construct an active label query algorithm
consistent in the agnostic setting. Our second key contribution is a novel con?dence-rated predictor
with guaranteed error that applies to any general classi?cation problem. We show that our predictor
is optimal in the realizable case, in the sense that it has the lowest abstention rate out of all predictors
guaranteeing a certain error. Moreover, we show how to extend our predictor to the agnostic setting.
Combining the label query algorithm with our novel con?dence-rated predictor, we get a general
active learning algorithm consistent in the agnostic setting. We provide a characterization of the label
complexity of our algorithm, and show that this is better than the bounds known for disagreementbased active learning in general. Finally, we show that for linear classi?cation with respect to the
uniform distribution and log-concave distributions, our bounds reduce to those of [3, 4].
2
2.1
Algorithm
The Setting
We study active learning for binary classi?cation. Examples belong to an instance space X , and
their labels lie in a label space Y = {?1, 1}; labelled examples are drawn from an underlying data
distribution D on X ? Y. We use DX to denote the marginal on D on X , and DY |X to denote the
conditional distribution on Y |X = x induced by D. Our algorithm has access to examples through
two oracles ? an example oracle U which returns an unlabelled example x ? X drawn from DX and
a labelling oracle O which returns the label y of an input x ? X drawn from DY |X .
Given a hypothesis class H of VC dimension d, the error of any h ? H with respect to a
data distribution ? over X ? Y is de?ned as err? (h) = P(x,y)?? (h(x) ?= y). We de?ne:
h? (?) = argminh?H err? (h), ? ? (?) = err? (h? (?)). For a set S, we abuse notation and use S
to also denote the uniform distribution over the elements of S. We de?ne P? (?) := P(x,y)?? (?),
E? (?) := E(x,y)?? (?).
Given access to examples from a data distribution D through an example oracle U and a labeling
? ? H such that with probability ? 1 ? ?, errD (h)
? ?
oracle O, we aim to provide a classi?er h
? ? (D) + ?, for some target values of ? and ?; this is achieved in an adaptive manner by making
as few queries to the labelling oracle O as possible. When ? ? (D) = 0, we are said to be in the
realizable case; in the more general agnostic case, we make no assumptions on the labels, and thus
? ? (D) can be positive.
Previous approaches to agnostic active learning have frequently used the notion of disagreements.
The disagreement between two hypotheses h1 and h2 with respect to a data distribution ? is
the fraction of examples according to ? to which h1 and h2 assign different labels; formally:
?? (h1 , h2 ) = P(x,y)?? (h1 (x) ?= h2 (x)). Observe that a data distribution ? induces a pseudometric ?? on the elements of H; this is called the disagreement metric. For any r and any h ? H,
de?ne B? (h, r) to be the disagreement ball of radius r around h with respect to the data distribution
?. Formally: B? (h, r) = {h? ? H : ?? (h, h? ) ? r}.
For notational simplicity, we assume that the hypothesis space is ?dense? with repsect to the data
distribution D, in the sense that ?r > 0, suph?BD (h? (D),r) ?D (h, h? (D)) = r. Our analysis will
still apply without the denseness assumption, but will be signi?cantly more messy. Finally, given a
set of hypotheses V ? H, the disagreement region of V is the set of all examples x such that there
exist two hypotheses h1 , h2 ? V for which h1 (x) ?= h2 (x).
This paper establishes a connection between active learning and con?dence-rated predictors with
guaranteed error. A con?dence-rated predictor is a prediction algorithm that is occasionally allowed to abstain from classi?cation. We will consider such predictors in the transductive setting.
Given a set V of candidate hypotheses, an error guarantee ?, and a set U of unlabelled examples,
a con?dence-rated predictor P either assigns a label or abstains from prediction on each unlabelled
2
x ? U . The labels are assigned with the guarantee that the expected disagreement1 between the
label assigned by P and any h ? V is ? ?. Speci?cally,
for all h ? V,
Px?U (h(x) ?= P (x), P (x) ?= 0) ? ?
(1)
This ensures that if some h? ? V is the true risk minimizer, then, the labels predicted by P on U do
not differ very much from those predicted by h? . The performance of a con?dence-rated predictor
which has a guarantee such as in Equation (1) is measured by its coverage, or the probability of
non-abstention Px?U (P (x) ?= 0); higher coverage implies better performance.
2.2
Main Algorithm
Our active learning algorithm proceeds in epochs, where the goal of epoch k is to achieve excess
generalization error ?k = ?2k0 ?k+1 , by querying a fresh batch of labels. The algorithm maintains a
candidate set Vk that is guaranteed to contain the true risk minimizer.
The critical decision at each epoch is how to select a subset of unlabelled examples whose labels
should be queried. We make this decision using a con?dence-rated predictor P . At epoch k, we run
P with candidate hypothesis set V = Vk and error guarantee ? = ?k /64. Whenever P abstains, we
query the label of the example. The number of labels mk queried is adjusted so that it is enough to
achieve excess generalization error ?k+1 .
An outline is described in Algorithm 1; we next discuss each individual component in detail.
Algorithm 1 Active Learning Algorithm: Outline
1: Inputs: Example oracle U , Labelling oracle O, hypothesis class H of VC dimension d,
con?dence-rated predictor P , target excess error ? and target con?dence ?.
2: Set k0 = ?log 1/??. Initialize candidate set V1 = H.
3: for k = 1, 2, ..k0 do
?
4:
Set ?k = ?2k0 ?k+1 , ?k = 2(k0 ?k+1)
2.
5:
Call U to generate a fresh unlabelled sample Uk = {zk,1 , ..., zk,nk } of size nk =
512 2
288
2
192( 512
?k ) (d ln 192( ?k ) + ln ?k ).
6:
Run con?dence-rated predictor P with inpuy V = Vk , U = Uk and error guarantee
? = ?k /64 to get abstention probabilities ?k,1 , . . . , ?k,nk on the examples in U
?kn. k These
?k,i .
probabilities induce a distribution ?k on Uk . Let ?k = Px?Uk (P (x) = 0) = n1k i=1
7:
if in the Realizable Case then
1536?k
k
8:
Let mk = 1536?
+ ln ?48k ). Draw mk i.i.d examples from ?k and query
?k (d ln
?k
O for labels of these examples to get a labelled data set Sk . Update Vk+1 using Sk :
Vk+1 := {h ? Vk : h(x) = y, for all (x, y) ? Sk }.
9:
else
10:
In the non-realizable case, use Algorithm 2 with inputs hypothesis set Vk , distribution
?k
?k , target excess error 8?
, target con?dence ?2k , and the labeling oracle O to get a new
k
hypothesis set Vk+1 .
? ? Vk +1 .
11: return an arbitrary h
0
Candidate Sets. At epoch k, we maintain a set Vk of candidate hypotheses guaranteed to contain
the true risk minimizer h? (D) (w.h.p). In the realizable case, we use a version space as our candidate
set. The version space with respect to a set S of labelled examples is the set of all h ? H such that
h(xi ) = yi for all (xi , yi ) ? S.
Lemma 1. Suppose we run Algorithm 1 in the realizable case with inputs example oracle U , labelling oracle O, hypothesis class H, con?dence-rated predictor P , target excess error ? and target
con?dence ?. Then, with probability 1, h? (D) ? Vk , for all k = 1, 2, . . . , k0 + 1.
In the non-realizable case, the version space is usually empty; we use instead a (1 ? ?)-con?dence
set for the true risk minimizer. Given a set S of n labelled examples, let C(S) ? H be a function of
1
where the expectation is with respect to the random choices made by P
3
S; C(S) is said to be a (1 ? ?)-con?dence set for the true risk minimizer if for all data distributions
? over X ? Y,
PS??n [h? (?) ? C(S)] ? 1 ? ?,
?
Recall that h (?) = argminh?H err? (h). In the non-realizable case, our candidate sets are (1 ? ?)con?dence sets for h? (D), for ? = ?. The precise setting of Vk is explained in Algorithm 2.
Lemma 2. Suppose we run Algorithm 1 in the non-realizable case with inputs example oracle U ,
labelling oracle O, hypothesis class H, con?dence-rated predictor P , target excess error ? and
target con?dence ?. Then with probability 1 ? ?, h? (D) ? Vk , for all k = 1, 2, . . . , k0 + 1.
Label Query. We next discuss our label query procedure ? which examples should we query labels
for, and how many labels should we query at each epoch?
Which Labels to Query? Our goal is to query the labels of the most informative examples. To
choose these examples while still maintaining consistency, we use a con?dence-rated predictor P
with guaranteed error. The inputs to the predictor are our candidate hypothesis set Vk which contains
(w.h.p) the true risk minimizer, a fresh set Uk of unlabelled examples, and an error guarantee ? =
?k /64. For notation simplicity, assume the elements in Uk are distinct. The output is a sequence of
abstention probabilities {?k,1 , ?k,2 , . . . , ?k,nk }, for each example in Uk . It induces a distribution ?k
over Uk , from which we independently draw examples for label queries.
How Many Labels to Query? The goal of epoch k is to achieve excess generalization error ?k .
2
?
To achieve this, passive learning requires O(d/?
k ) labelled examples in the realizable case, and
?
2
?
O(d(? (D) + ?k )/?k ) examples in the agnostic case. A key observation in this paper is that in
order to achieve excess generalization error ?k on D, it suf?ces to achieve a much larger excess
generalization error O(?k /?k ) on the data distribution induced by ?k and DY |X , where ?k is the
fraction of examples on which the con?dence-rated predictor abstains.
1536?k
k
In the realizable case, we achieve this by sampling mk = 1536?
+ ln ?48k ) i.i.d examples
?k (d ln
?k
from ?k , and querying their labels to get a labelled dataset Sk . Observe that as ?k is the abstention
probability of P with guaranteed error ? ?k /64, it is generally smaller than the measure of the
disagreement region of the version space; this key fact results in improved label complexity over
disagreement-based active learning. This sampling procedure has the following property:
Lemma 3. Suppose we run Algorithm 1 in the realizable case with inputs example oracle U , labelling oracle O, hypothesis class H, con?dence-rated predictor P , target excess error ? and target
con?dence ?. Then with probability 1 ? ?, for all k = 1, 2, . . . , k0 + 1, and for all h ? Vk ,
? returned at the end of the algorithm satis?es errD (h)
? ? ?.
errD (h) ? ?k . In particular, the h
The agnostic case has an added complication ? in practice, the value of ? ? is not known ahead of
time. Inspired by [24], we use a doubling procedure(stated in Algorithm 2) which adaptively ?nds
the number mk of labelled examples to be queried and queries them. The following two lemmas
illustrate its properties ? that it is consistent, and that it does not use too many label queries.
Lemma 4. Suppose we run Algorithm 2 with inputs hypothesis set V , example distribution ?,
? Let ?
? be the joint distribution on
labelling oracle O, target excess error ?? and target con?dence ?.
? such that on E,
?
?
? (1)
X ? Y induced by ? and DY |X . Then there exists an event E, P(E) ? 1 ? ?,
Algorithm 2 halts and (2) the set Vj0 has the following properties:
? ?
(2.1) If for h ? H, err?
?/2, then h ? Vj0 .
? (h) ? err?
? (h (?)) ? ?
? ?
?.
(2.2) On the other hand, if h ? Vj0 , then err?
? (h) ? err?
? (h (?)) ? ?
? happens, we say Algorithm 2 succeeds.
When event E
Lemma 5. Suppose we run Algorithm 2 with inputs hypothesis set V , example distribution ?,
? There exists some absolute constant
labelling oracle O, target excess error ?? and target con?dence ?.
? ?
?
). Thus
c1 > 0, such that on the event that Algorithm 2 succeeds, nj0 ? c1 ((d ln 1?? + ln 1?? ) ? (?)+?
??2
? ?
?j0
?
1
1 ? (?)+?
the total number of labels queried is j=1 nj ? 2nj0 ? 2c1 ((d ln ?? + ln ?? ) ??2 ).
2
? hides logarithmic factors
O(?)
4
A naive approach (see Algorithm 4 in the Appendix) which uses an additive VC bound gives a
? ??2 ); Algorithm 2 gives a better sample complexity.
sample complexity of O((d ln(1/?
?) + ln(1/?))?
The following lemma is a consequence of our label query procedure in the non-realizable case.
Lemma 6. Suppose we run Algorithm 1 in the non-realizable case with inputs example oracle U ,
labelling oracle O, hypothesis class H, con?dence-rated predictor P , target excess error ? and
target con?dence ?. Then with probability 1 ? ?, for all k = 1, 2, . . . , k0 + 1, and for all h ? Vk ,
? returned at the end of the algorithm satis?es
errD (h) ? errD (h? (D)) + ?k . In particular, the h
?
? ? errD (h (D)) + ?.
errD (h)
Algorithm 2 An Adaptive Algorithm for Label Query Given Target Excess Error
1: Inputs: Hypothesis set V of VC dimension d, Example distribution ?, Labeling oracle O,
?
target excess error ??, target con?dence ?.
2: for j = 1, 2, . . . do
3:
Draw nj = 2j i.i.d examples from ?; query their labels from O to get a labelled dataset
?
+ 1)).
Sj . Denote ??j := ?/(j(j
?
4:
Train an ERM classi?er hj ? V over Sj .
5:
De?ne the set Vj as follows:
?
?
?
? j ) + ?? + ?(nj , ??j ) + ?(nj , ??j )?S (h, h
?j )
Vj = h ? V : errSj (h) ? errSj (h
j
2
Where ?(n, ?) :=
16
ln 2en
n (2d?
d
6:
if suph?Vj (?(nj , ??j ) +
7:
j0 = j, break
8: return Vj0 .
2.3
+ ln 24
? ).
?
? j )) ?
?(nj , ?j )?S (h, h
j
??
6
then
Con?dence-Rated Predictor
Our active learning algorithm uses a con?dence-rated predictor with guaranteed error to make its
label query decisions. In this section, we provide a novel con?dence-rated predictor with guaranteed
error. This predictor has optimal coverage in the realizable case, and may be of independent interest.
The predictor P receives as input a set V ? H of hypotheses (which is likely to contain the true
risk minimizer), an error guarantee ?, and a set of U of unlabelled examples. We consider a soft
prediction algorithm; so, for each example in U , the predictor P outputs three probabilities that add
up to 1 ? the probability of predicting 1, ?1 and 0. This output is subject to the constraint that the
expected disagreement3 between the ?1 labels assigned by P and those assigned by any h ? V is
at most ?, and the goal is to maximize the coverage, or the expected fraction of non-abstentions.
Our key insight is that this problem can be written as a linear program, which is described in Algorithm 3. There are three variables, ?i , ?i and ?i , for each unlabelled zi ? U ; there are the probabilities with which we predict 1, ?1 and 0 on zi respectively. Constraint (2) ensures that the expected
disagreement between the label predicted and any h ? V is no more than ?, while the LP objective
maximizes the coverage under these constraints. Observe that the LP is always feasible. Although
the LP has in?nitely many constraints, the number of constraints in Equation (2) distinguishable by
Uk is at most (em/d)d , where d is the VC dimension of the hypothesis class H.
The performance of a con?dence-rated predictor is measured by its error and coverage. The error of
a con?dence-rated predictor is the probability with which it predicts the wrong label on an example,
while the coverage is its probability of non-abstention. We can show the following guarantee on the
performance of the predictor in Algorithm 3.
Theorem 1. In the realizable case, if the hypothesis set V is the version space with respect to
a training set, then Px?U (P (x) ?= h? (x), P (x) ?= 0) ? ?. In the non-realizable case, if the
hypothesis set V is an (1 ? ?)-con?dence set for the true risk minimizer h? , then, w.p ? 1 ? ?,
Px?U (P (x) ?= y, P (x) ?= 0) ? Px?U (h? (x) ?= y) + ?.
3
where the expectation is taken over the random choices made by P
5
Algorithm 3 Con?dence-rated Predictor
1: Inputs: hypothesis set V , unlabelled data U = {z1 , . . . , zm }, error bound ?.
2: Solve the linear program:
min
m
?
?i
i=1
subject to:
?i, ?i + ?i + ?i = 1
?
?
?h ? V,
?i +
i:h(zi )=1
i:h(zi )=?1
?i ? ?m
(2)
?i, ?i , ?i , ?i ? 0
3: For each zi ? U , output probabilities for predicting 1, ?1 and 0: ?i , ?i , and ?i .
In the realizable case, we can also show that our con?dence rated predictor has optimal coverage.
Observe that we cannot directly show optimality in the non-realizable case, as the performance
depends on the exact choice of the (1 ? ?)-con?dence set.
Theorem 2. In the realizable case, suppose that the hypothesis set V is the version space with
respect to a training set. If P ? is any con?dence rated predictor with error guarantee ?, and if P is
the predictor in Algorithm 3, then, the coverage of P is at least much as the coverage of P ? .
3
Performance Guarantees
An essential property of any active learning algorithm is consistency ? that it converges to the true
risk minimizer given enough labelled examples. We observe that our algorithm is consistent provided we use any con?dence-rated predictor P with guaranteed error as a subroutine. The consistency of our algorithm is a consequence of Lemmas 3 and 6 and is shown in Theorem 3.
Theorem 3 (Consistency). Suppose we run Algorithm 1 with inputs example oracle U , labelling
oracle O, hypothesis class H, con?dence-rated predictor P , target excess error ? and target
? returned by Algorithm 1 satis?es
con?dence ?. Then with probability 1 ? ?, the classi?er h
?
?
errD (h) ? errD (h (D)) ? ?.
We now establish a label complexity bound for our algorithm; however, this label complexity bound
applies only if we use the predictor described in Algorithm 3 as a subroutine.
For any hypothesis set V , data distribution D, and ?, de?ne ?D (V, ?) to be the minimum abstention probability of a con?dence-rated predictor which guarantees that the disagreement between its
predicted labels and any h ? V under DX is at most ?.
Formally, ?D (V, ?) = min{ED ?(x) : ED [I(h(x) = +1)?(x) + I(h(x) = ?1)?(x)] ?
? for all h ? V, ?(x) + ?(x) + ?(x) ? 1, ?(x), ?(x), ?(x) ? 0}. De?ne ?(r, ?) :=
?D (BD (h? , r), ?). The label complexity of our active learning algorithm can be stated as follows.
Theorem 4 (Label Complexity). Suppose we run Algorithm 1 with inputs example oracle U , labelling oracle O, hypothesis class H, con?dence-rated predictor P of Algorithm 3, target excess
error ? and target con?dence ?. Then there exist constants c3 , c4 > 0 such that with probability
1 ? ?:
(1) In the realizable case, the total number of labels queried by Algorithm 1 is at most:
c3
?log
1
?? ?
k=1
(d ln
?(?k , ?k /256)
?log(1/?)? ? k + 1 ?(?k , ?k /256)
))
+ ln(
?k
?
?k
(2) In the agnostic case, the total number of labels queried by Algorithm 1 is at most:
c4
1
?? ?
?log
k=1
(d ln
?(2? ? (D) + ?k , ?k /256)
?log(1/?)? ? k + 1 ?(2? ? (D) + ?k , ?k /256)
? ? (D)
+ln(
(1+
)
))
?k
?
?k
?k
6
Comparison. The label complexity of disagreement-based active learning is characterized in
terms of the disagreement coef?cient. Given a radius r, the disagreement coef?cent ?(r) is de?ned
as:
P(DIS(BD (h? , r? )))
,
?(r) = sup
r?
r ? ?r
where for any V ? H, DIS(V ) is the disagreement region of V . As P(DIS(BD (h? , r))) =
?
?(r, 0) [13], in our notation, ?(r) = supr? ?r ?(rr?,0) .
In the realizable case, the best known bound for label complexity of disagreement-based active
?
learning is O(?(?)
? ln(1/?) ? (d ln ?(?) + ln ln(1/?))) [20]4 . Our label complexity bound may be
simpli?ed to:
??
?
?
?
?
1
?(?
,
?
/256)
?(?
,
?
/256)
1
k
k
k
k
? ln ?
sup
,
+ ln ln
? d ln
sup
O
? k??log(1/?)?
?k
?k
?
k??log(1/?)?
which is essentially the bound of [20] with ?(?) replaced by supk??log(1/?)? ?(?k ,??kk/256) . As enforcing a lower error guarantee requires more abstention, ?(r, ?) is a decreasing function of ?; as a
result,
?(?k , ?k /256)
? ?(?),
sup
?k
k??log(1/?)?
and our label complexity bound is better.
?
2
?
?
In the agnostic case, [12] provides a label complexity bound of O(?(2?
(D)+?)?(d ? (D)
ln(1/?)+
?2
2
d ln (1/?))) for disagreement-based active-learning. In contrast, by Proposition 1 our label complexity is at most:
?
? ?
??
?
2
?(2?
(D)
+
?
,
?
/256)
(D)
?
k
k
2
?
? d
ln(1/?) + d ln (1/?)
O
sup
2? ? (D) + ?k
?2
k??log(1/?)?
Again, this is essentially the bound of [12] with ?(2? ? (D) + ?) replaced by the smaller quantity
?(2? ? (D) + ?k , ?k /256)
,
2? ? (D) + ?k
k??log(1/?)?
sup
[20] has provided a more re?ned analysis of disagreement-based active learning that gives a label
?
2
?
?
complexity of O(?(?
(D) + ?)( ? (D)
+ ln 1? )(d ln ?(? ? (D) + ?) + ln ln 1? )); observe that their
?2
?
dependence is still on ?(? (D) + ?). We leave a more re?ned label complexity analysis of our
algorithm for future work.
An important sub-case of learning from noisy data is learning under the Tsybakov noise conditions [30]. We defer the discussion into the Appendix.
3.1
Case Study: Linear Classi?cation under the Log-concave Distribution
We now consider learning linear classi?ers with respect to?log-concave data distribution on Rd . In
this case, for any r, the disagreement coef?cient ?(r) ? O( d ln(1/r)) [4]; however, for any ? > 0,
?(r,?)
? O(ln(r/?)) (see Lemma 14 in the Appendix), which is much smaller so long as ?/r is not
r
too small. This leads to the following label complexity bounds.
Corollary 1. Suppose DX is isotropic and log-concave on Rd , and H is the set of homogeneous linear classi?ers on Rd . Then Algorithm 1 with inputs example oracle U , labelling oracle O, hypothesis
class H, con?dence-rated predictor P of Algorithm 3, target excess error ? and target con?dence ?
satis?es the following properties. With probability 1 ? ?:
(1) In the realizable case, there exists some absolute constant c8 > 0 such that the total number of
labels queried is at most c8 ln 1? (d + ln ln 1? + ln 1? ).
4
? notation hides factors logarithmic in 1/?
Here the O(?)
7
(2) In the agnostic case, there exists some absolute constant c9 > 0 such that the total number of la?
2
?
?
?
+ ln 1? ) ln ?+? ? (D) (d ln ?+? ? (D) + ln 1? ) + ln 1? ln ?+? ? (D) ln ln 1? .
bels queried is at most c9 ( ? (D)
?2
(3) If (C0 , ?)-Tsybakov Noise condition holds for D with respect to H, then there exists some
constant c10 > 0 (that depends on C0 , ?) such that the total number of labels queried is at most
2
c10 ? ? ?2 ln 1? (d ln 1? + ln 1? ).
In the realizable case, our bound matches [4]. For disagreement-based
algorithms, the bound is
?
3
2 1
1
?
2
O(d ln ? (ln d + ln ln ? )), which is worse by a factor of O( d ln(1/?)). [4] does not address the
fully agnostic case directly; however, if ? ? (D) is known a-priori, then their algorithm can achieve
roughly the same label complexity as ours.
For the Tsybakov Noise Condition with ? > 1, [3, 4] provides a label complexity bound for
? ?2 ?2 ln2 1 (d + ln ln 1 )) with an algorithm that has a-priori knowledge of C0 and ?. We get
O(?
?
?
a slightly better bound. On the other hand, a disagreement based algorithm [20] gives a label
?
? 32 ln2 1 ? ?2 ?2 (ln d + ln ln 1 )). Again our bound is better by factor of ?( d)
complexity of O(d
?
?
over disagreement-based algorithms. For ? = 1, we can tighten our label complexity to get a
1
1
1
?
O(ln
? (d + ln ln ? + ln ? )) bound, which again matches [4], and is better than the ones provided by
? 32 ln2 1 (ln d + ln ln 1 )) [20].
disagreement-based algorithm ? O(d
?
?
4
Related Work
Active learning has seen a lot of progress over the past two decades, motivated by vast amounts of
unlabelled data and the high cost of annotation [28, 10, 20]. According to [10], the two main threads
of research are exploitation of cluster structure [31, 11], and ef?cient search in hypothesis space,
which is the setting of our work. We are given a hypothesis class H, and the goal is to ?nd an h ? H
that achieves a target excess generalization error, while minimizing the number of label queries.
Three main approaches have been studied in this setting. The ?rst and most natural one is generalized
binary search [17, 8, 9, 27], which was analyzed in the realizable case by [9] and in various limited
noise settings by [23, 27, 26]. While this approach has the advantage of low label complexity, it is
generally inconsistent in the fully agnostic setting [11]. The second approach, disagreement-based
active learning, is consistent in the agnostic PAC model. [7] provides the ?rst disagreement-based
algorithm for the realizable case. [2] provides an agnostic disagreement-based algorithm, which
is analyzed in [18] using the notion of disagreement coef?cient. [12] reduces disagreement-based
active learning to passive learning; [5] and [6] further extend this work to provide practical and ef?cient implementations. [19, 24] give algorithms that are adaptive to the Tsybakov Noise condition.
The third line of work [3, 4, 1], achieves a better label complexity than disagreement-based active
learning for linear classi?ers on the uniform distribution over unit sphere and logconcave distributions. However, a limitation is that their algorithm applies only to these speci?c settings, and it is
not apparent how to apply it generally.
Research on con?dence-rated prediction has been mostly focused on empirical work, with relatively
less theoretical development. Theoretical work on this topic includes KWIK learning [25], conformal prediction [29] and the weighted majority algorithm of [16]. The closest to our work is the recent
learning-theoretic treatment by [13, 14]. [13] addresses con?dence-rated prediction with guaranteed
error in the realizable case, and provides a predictor that abstains in the disagreement region of the
version space. This predictor achieves zero error, and coverage equal to the measure of the agreement region. [14] shows how to extend this algorithm to the non-realizable case and obtain zero
error with respect to the best hypothesis in H. Note that the predictors in [13, 14] generally achieve
less coverage than ours for the same error guarantee; in fact, if we plug them into our Algorithm 1,
then we recover the label complexity bounds of disagreement-based algorithms [12, 19, 24].
A formal connection between disagreement-based active learning in realizable case and perfect
con?dence-rated prediction (with a zero error guarantee) was established by [15]. Our work can
be seen as a step towards bridging these two areas, by demonstrating that active learning can be
further reduced to imperfect con?dence-rated prediction, with potentially higher label savings.
Acknowledgements. We thank NSF under IIS-1162581 for research support. We thank Sanjoy
Dasgupta and Yoav Freund for helpful discussions. CZ would like to thank Liwei Wang for introducing the problem of selective classi?cation to him.
8
References
[1] P. Awasthi, M-F. Balcan, and P. M. Long. The power of localization for ef?ciently learning
linear separators with noise. In STOC, 2014.
[2] M.-F. Balcan, A. Beygelzimer, and J. Langford. Agnostic active learning. J. Comput. Syst.
Sci., 75(1):78?89, 2009.
[3] M.-F. Balcan, A. Z. Broder, and T. Zhang. Margin based active learning. In COLT, 2007.
[4] M.-F. Balcan and P. M. Long. Active and passive learning of linear separators under logconcave distributions. In COLT, 2013.
[5] A. Beygelzimer, S. Dasgupta, and J. Langford. Importance weighted active learning. In ICML,
2009.
[6] A. Beygelzimer, D. Hsu, J. Langford, and T. Zhang. Agnostic active learning without constraints. In NIPS, 2010.
[7] D. A. Cohn, L. E. Atlas, and R. E. Ladner. Improving generalization with active learning.
Machine Learning, 15(2), 1994.
[8] S. Dasgupta. Analysis of a greedy active learning strategy. In NIPS, 2004.
[9] S. Dasgupta. Coarse sample complexity bounds for active learning. In NIPS, 2005.
[10] S. Dasgupta. Two faces of active learning. Theor. Comput. Sci., 412(19), 2011.
[11] S. Dasgupta and D. Hsu. Hierarchical sampling for active learning. In ICML, 2008.
[12] S. Dasgupta, D. Hsu, and C. Monteleoni. A general agnostic active learning algorithm. In
NIPS, 2007.
[13] R. El-Yaniv and Y. Wiener. On the foundations of noise-free selective classi?cation. JMLR,
2010.
[14] R. El-Yaniv and Y. Wiener. Agnostic selective classi?cation. In NIPS, 2011.
[15] R. El-Yaniv and Y. Wiener. Active learning via perfect selective classi?cation. JMLR, 2012.
[16] Y. Freund, Y. Mansour, and R. E. Schapire. Generalization bounds for averaged classi?ers.
The Ann. of Stat., 32, 2004.
[17] Y. Freund, H. S. Seung, E. Shamir, and N. Tishby. Selective sampling using the query by
committee algorithm. Machine Learning, 28(2-3):133?168, 1997.
[18] S. Hanneke. A bound on the label complexity of agnostic active learning. In ICML, 2007.
[19] S. Hanneke. Adaptive rates of convergence in active learning. In COLT, 2009.
[20] S. Hanneke. A statistical theory of active learning. Manuscript, 2013.
[21] S. Hanneke and L. Yang. Surrogate losses in passive and active learning. CoRR, abs/1207.3772,
2012.
[22] D. Hsu. Algorithms for Active Learning. PhD thesis, UC San Diego, 2010.
[23] M. K?aa? ri?ainen. Active learning in the non-realizable case. In ALT, 2006.
[24] V. Koltchinskii. Rademacher complexities and bounding the excess risk in active learning.
JMLR, 2010.
[25] L. Li, M. L. Littman, and T. J. Walsh. Knows what it knows: a framework for self-aware
learning. In ICML, 2008.
[26] M. Naghshvar, T. Javidi, and K. Chaudhuri. Noisy bayesian active learning. In Allerton, 2013.
[27] R. D. Nowak. The geometry of generalized binary search. IEEE Transactions on Information
Theory, 57(12):7893?7906, 2011.
[28] B. Settles. Active learning literature survey. Technical report, University of WisconsinMadison, 2010.
[29] G. Shafer and V. Vovk. A tutorial on conformal prediction. JMLR, 2008.
[30] A. B. Tsybakov. Optimal aggregation of classi?ers in statistical learning. Annals of Statistics,
32:135?166, 2004.
[31] R. Urner, S. Wulff, and S. Ben-David. Plal: Cluster-based active learning. In COLT, 2013.
9
| 5435 |@word exploitation:1 version:7 nd:5 c0:3 open:1 reduction:1 contains:1 ours:2 past:1 err:8 beygelzimer:3 dx:4 bd:4 written:1 additive:1 informative:1 atlas:1 ainen:1 update:1 greedy:1 isotropic:1 characterization:1 provides:5 complication:1 coarse:1 allerton:1 zhang:3 manner:1 expected:4 roughly:1 frequently:1 inspired:1 decreasing:1 provided:4 moreover:1 underlying:1 notation:4 agnostic:29 maximizes:1 lowest:1 what:1 nj:6 guarantee:15 concave:5 wrong:1 uk:9 unit:2 positive:1 consequence:2 nitely:1 abuse:1 koltchinskii:1 studied:1 limited:2 walsh:1 averaged:1 practical:1 practice:1 procedure:4 j0:2 area:1 empirical:1 liwei:1 pre:1 induce:1 get:8 cannot:1 risk:11 independently:1 survey:2 focused:1 simplicity:2 assigns:1 insight:1 notion:2 annals:1 diego:3 target:29 suppose:10 shamir:1 exact:1 homogeneous:1 us:2 hypothesis:36 agreement:1 element:3 gilman:2 predicts:1 wang:1 naghshvar:1 region:5 ensures:2 mentioned:1 complexity:30 messy:1 littman:1 seung:1 localization:1 learner:1 joint:1 k0:9 various:1 train:1 distinct:1 query:24 labeling:3 whose:2 apparent:1 larger:1 solve:1 say:1 statistic:1 transductive:1 noisy:2 sequence:1 rr:1 advantage:1 zm:1 combining:1 chaudhuri:2 achieve:10 rst:4 convergence:1 empty:1 requirement:3 p:1 cluster:2 yaniv:3 generating:1 guaranteeing:1 converges:1 leave:1 perfect:2 ben:1 illustrate:1 rademacher:1 stat:1 measured:2 progress:1 c10:2 coverage:12 c:1 signi:1 predicted:4 implies:1 differ:1 radius:2 vc:5 abstains:4 settle:1 assign:1 generalization:9 proposition:1 theor:1 adjusted:1 hold:1 around:1 predict:1 major:2 achieves:4 label:78 him:1 establishes:1 weighted:2 awasthi:1 always:1 aim:2 hj:1 corollary:1 vk:18 notational:1 contrast:1 realizable:30 sense:2 helpful:1 minimizers:1 el:3 selective:5 subroutine:2 colt:4 priori:2 development:1 initialize:1 uc:1 marginal:1 equal:1 construct:1 saving:1 aware:1 sampling:4 icml:4 future:1 report:1 few:3 individual:1 replaced:2 geometry:1 maintain:1 ab:1 interest:2 satis:4 analyzed:2 accurate:1 nowak:1 supr:1 re:2 theoretical:2 mk:5 instance:1 soft:1 yoav:1 cost:1 introducing:1 subset:2 predictor:46 uniform:4 too:2 tishby:1 kn:1 adaptively:1 broder:1 cantly:1 pool:1 again:3 thesis:1 interactively:2 choose:1 worse:1 return:4 li:1 syst:1 de:8 includes:1 depends:2 view:1 h1:8 break:1 lot:1 sup:6 recover:1 maintains:1 aggregation:1 errd:9 annotation:1 defer:1 chicheng:1 contribution:3 n1k:1 wiener:3 bayesian:1 plal:1 hanneke:4 drive:2 cation:11 coef:4 monteleoni:1 whenever:1 ed:5 urner:1 con:54 hsu:4 dataset:2 treatment:1 recall:1 knowledge:1 manuscript:1 higher:2 wisconsinmadison:1 improved:1 langford:3 hand:2 receives:1 cohn:1 vj0:4 contain:3 true:11 assigned:4 self:1 maintained:1 occasion:1 ln2:3 generalized:2 outline:2 theoretic:1 passive:4 balcan:4 novel:5 abstain:2 ef:3 overview:1 extend:3 belong:1 queried:10 rd:3 consistency:5 access:3 add:1 kwik:1 closest:1 hide:2 recent:1 jolla:2 occasionally:1 certain:1 binary:5 abstention:9 yi:2 seen:2 minimum:1 simpli:1 speci:5 maximize:1 ii:1 reduces:1 technical:1 match:2 unlabelled:11 plug:1 characterized:1 sphere:2 long:3 halt:1 ensuring:1 prediction:12 essentially:2 metric:1 expectation:2 cz:1 achieved:1 c1:3 nj0:2 else:1 induced:3 subject:2 logconcave:2 inconsistent:2 call:1 ciently:1 yang:1 enough:2 zi:5 reduce:1 idea:1 imperfect:1 thread:1 motivated:1 bridging:1 returned:3 generally:6 amount:1 tsybakov:5 induces:2 reduced:1 generate:1 schapire:1 exist:3 nsf:1 tutorial:1 dasgupta:7 key:5 demonstrating:1 drawn:3 ce:1 v1:1 vast:1 fraction:3 run:10 draw:3 decision:3 dy:4 appendix:3 bound:23 guaranteed:12 oracle:26 ahead:1 constraint:6 ri:1 dence:53 generates:1 min:2 optimality:1 c8:2 pseudometric:1 px:6 relatively:1 ned:4 according:2 request:1 ball:1 smaller:3 slightly:1 em:1 lp:3 making:2 happens:1 explained:1 erm:1 taken:1 ln:68 equation:2 discus:2 committee:1 know:2 end:2 conformal:2 apply:3 observe:6 hierarchical:1 disagreement:34 batch:1 disagreementbased:1 cent:1 maintaining:2 cally:2 establish:1 objective:1 added:1 quantity:1 strategy:3 primary:1 dependence:1 surrogate:1 unclear:1 said:2 javidi:1 thank:3 sci:2 majority:1 topic:1 fresh:3 enforcing:1 kk:1 minimizing:2 mostly:1 potentially:1 stoc:1 stated:2 implementation:1 policy:1 ladner:1 observation:1 extended:1 precise:1 mansour:1 ucsd:2 arbitrary:1 david:1 c3:2 connection:3 z1:1 bel:1 c4:2 california:2 established:1 nip:5 address:2 beyond:1 proceeds:1 usually:1 challenge:3 program:2 power:1 critical:1 event:3 natural:2 predicting:2 rated:36 ne:6 naive:1 epoch:7 literature:2 acknowledgement:1 freund:3 fully:2 loss:1 suf:1 suph:2 limitation:1 querying:2 h2:8 foundation:1 consistent:9 free:1 denseness:1 dis:3 formal:1 face:1 absolute:3 dimension:4 made:3 adaptive:4 san:3 tighten:1 transaction:1 excess:20 sj:2 active:58 xi:2 search:4 decade:1 sk:4 learn:2 zk:2 ca:2 improving:1 requested:1 separator:2 vj:3 main:6 dense:1 bounding:1 noise:8 shafer:1 allowed:2 wulff:1 body:1 cient:5 en:1 sub:1 comput:2 candidate:9 lie:1 jmlr:4 third:1 theorem:5 pac:1 er:13 alt:1 exists:5 essential:1 kamalika:2 importance:1 corr:1 phd:1 labelling:12 margin:1 nk:4 logarithmic:2 distinguishable:1 likely:1 doubling:1 supk:1 applies:5 aa:1 minimizer:9 conditional:1 goal:7 ann:1 towards:1 labelled:9 feasible:1 vovk:1 classi:24 lemma:10 conservative:1 called:2 total:6 sanjoy:1 e:4 la:3 succeeds:2 formally:3 select:1 support:1 argminh:2 c9:2 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.