Unnamed: 0
int64 0
7.24k
| id
int64 1
7.28k
| raw_text
stringlengths 9
124k
| vw_text
stringlengths 12
15k
|
---|---|---|---|
4,700 | 5,256 | Parallel Direction Method of Multipliers
Huahua Wang , Arindam Banerjee , Zhi-Quan Luo
University of Minnesota, Twin Cities
{huwang,banerjee}@cs.umn.edu, [email protected]
Abstract
We consider the problem of minimizing block-separable (non-smooth) convex
functions subject to linear constraints. While the Alternating Direction Method of
Multipliers (ADMM) for two-block linear constraints has been intensively studied
both theoretically and empirically, in spite of some preliminary work, effective
generalizations of ADMM to multiple blocks is still unclear. In this paper, we
propose a parallel randomized block coordinate method named Parallel Direction
Method of Multipliers (PDMM) to solve optimization problems with multi-block
linear constraints. At each iteration, PDMM randomly updates some blocks in
parallel, behaving like parallel randomized block coordinate descent. We establish
the global convergence and the iteration complexity for PDMM with constant step
size. We also show that PDMM can do randomized block coordinate descent on
overlapping blocks. Experimental results show that PDMM performs better than
state-of-the-arts methods in two applications, robust principal component analysis
and overlapping group lasso.
1
Introduction
In this paper, we consider the minimization of block-seperable convex functions subject to linear
constraints, with a canonical form:
J
J
X
X
min f (x) =
fj (xj ) , s.t. Ax =
Acj xj = a ,
(1)
{xj ?Xj }
j=1
j=1
where the objective function f (x) is a sum of J block separable P
(nonsmooth) convex functions,
Acj ? Rm?nj is the j-th column block of A ? Rm?n where n = j nj , xj ? Rnj ?1 is the j-th
block coordinate of x, Xj is a local convex constraint of xj and a ? Rm?1 . The canonical form
can be extended to handle linear inequalities by introducing slack variables, i.e., writing Ax ? a as
Ax + z = a, z ? 0.
A variety of machine learning problems can be cast into the linearly-constrained optimization problem (1) [8, 4, 24, 5, 6, 21, 11]. For example, in robust Principal Component Analysis (RPCA) [5],
one attempts to recover a low rank matrix L and a sparse matrix S from an observation matrix M,
i.e., the linear constraint is M = L + S. Further, in the stable version of RPCA [29], an noisy matrix
Z is taken into consideration, and the linear constraint has three blocks, i.e., M = L + S + Z. Problem (1) can also include composite minimization problems which solve a sum of a loss function and
a set of nonsmooth regularization functions. Due to the increasing interest in structural sparsity [1],
composite regularizers have become widely used, e.g., overlapping group lasso [28]. As the blocks
are overlapping in this class of problems, it is difficult to apply block coordinate descent methods
for large scale problems [16, 18] which assume block-separable. By simply splitting blocks and introducing equality constraints, the composite minimization problem can also formulated as (1) [2].
A classical approach to solving (1) is to relax the linear constraints using the (augmented) Lagrangian, i.e.,
?
(2)
L? (x, y) = f (x) + hy, Ax ? ai + kAx ? ak22 ,
2
1
where ? ? 0 is called the penalty parameter. We call x the primal variable and y the dual variable. (2)
usually leads to primal-dual algorithms which update the primal and dual variables alternatively.
While the dual update is simply dual gradient descent, the primal update is to solve a minimization
problem of (2) given y. If ? = 0, the primal update can be solved in a parallel block coordinate
fashion [3, 19], leading to the dual ascent method. While the dual ascent method can achieve massive parallelism, a careful choice of stepsize and some strict conditions are required for convergence,
particularly when f is nonsmooth. To achieve better numerical efficiency and convergence behavior
compared to the dual ascent method, it is favorable to set ? > 0 in the augmented Lagrangian (2)
which we call the method of multipliers. However, (2) is no longer separable and solving entire
augmented Lagrangian (2) exactly is computationally expensive. In [20], randomized block coordinate descent (RBCD) [16, 18] is used to solve (2) exactly, but leading to a double-loop algorithm
along with the dual step. More recent results show (2) can be solved inexactly by just sweeping the
coordinates once using the alternating direction method of multipliers (ADMM) [12, 2]. This paper
attempts to develop a parallel randomized block coordinate variant of ADMM.
When J = 2, ADMM has been widely used to solve the augmented Lagragian (2) in many applications [2]. Encouraged by the success of ADMM with two blocks, ADMM has also been extended to solve the problem with multiple blocks [15, 14, 10, 17, 13, 7]. The variants of ADMM
can be mainly divided into two categories. The first category considers Gauss-Seidel ADMM
(GSADMM) [15, 14], which solves (2) in a cyclic block coordinate manner. In [13], a back substitution step was added so that the convergence of ADMM for multiple blocks can be proved. In
some cases, it has been shown that ADMM might not converge for multiple blocks [7]. In [14], a
block successive upper bound minimization method of multipliers (BSUMM) is proposed to solve
the problem (1). The convergence of BSUMM is established under some fairly strict conditions: (i)
certain local error bounds hold; (ii) the step size is either sufficiently small or decreasing. However,
in general, Gauss-Seidel ADMM with multiple blocks is not well understood and its iteration complexity is largely open. The second category considers Jacobian variants of ADMM [26, 10, 17],
which solves (2) in a parallel block coordinate fashion. In [26, 17], (1) is solved by using two-block
ADMM with splitting variables (sADMM). [10] considers a proximal Jacobian ADMM (PJADMM)
by adding proximal terms. A randomized block coordinate variant of ADMM named RBSUMM
was proposed in [14]. However, RBSUMM can only randomly update one block. Moreover, the
convergence of RBSUMM is established under the same conditions as BSUMM and its iteration
complexity is unknown.
In this paper, we propose a parallel randomized block coordinate method named parallel direction
method of multipliers (PDMM) which randomly picks up any number of blocks to update in parallel,
behaving like randomized block coordinate descent [16, 18]. Like the dual ascent method, PDMM
solves the primal update in a parallel block coordinate fashion even with the augmentation term.
Moreover, PDMM inherits the merits of the method of multipliers and can solve a fairly large class
of problems, including nonsmooth functions. Technically, PDMM has three aspects which make it
distinct from such state-of-the-art methods. First, if block coordinates of the primal x is solved exactly, PDMM uses a backward step on the dual update so that the dual variable makes conservative
progress. Second, the sparsity of A and the number of randomized blocks are taken into consideration to determine the step size of the dual update. Third, PDMM can randomly update arbitrary
number of primal blocks in parallel. Moreover, we show that sADMM and PJADMM are the two extreme cases of PDMM. The connection between sADMM and PJADMM through PDMM provides
better understanding of dual backward step. PDMM can also be used to solve overlapping groups in
a randomized block coordinate fashion. Interestingly, the corresponding problem for RBCD [16, 18]
with overlapping blocks is still an open problem. We establish the global convergence and O(1/T )
iteration complexity of PDMM with constant step size. We evaluate the performance of PDMM in
two applications: robust principal component analysis and overlapping group lasso.
The rest of the paper is organized as follows: We introduce PDMM in Section 2, and establish
convergence results in Section 3. We evaluate the performance of PDMM in Section 4 and conclude
in Section 5. The technical analysis and detailed proofs are provided in the supplement.
Notations: Assume that A ? Rm?n is divided into I ? J blocks. Let Ari ? Rmi ?n be the i-th row
block of A, Acj ? Rm?nj be the j-th column block of A, and Aij ? Rmi ?nj be the ij-th block of
A. Let yi ? Rmi ?1 be the i-th block of y ? Rm?1 . Let N (i) be a set of nonzero blocks Aij in the
2
? i = min{di , K} where
i-th row block Ari and di = |N (i)| be the number of nonzero blocks. Let K
K is the number of blocks randomly chosen by PDMM and T be the number of iterations.
2
Parallel Direction Method of Multipliers
Consider a direct Jacobi version of ADMM which updates all blocks in parallel:
xt+1
= argminxj ?Xj L? (xj , xtk6=j , yt ) ,
j
y
t+1
t
= y + ? ?(Ax
t+1
? a) .
(3)
(4)
where ? is a shrinkage factor for the step size of the dual gradient ascent update. However, empirical
results show that it is almost impossible to make the direct Jacobi updates (3)-(4) to converge even
when ? is extremely small. [15, 10] also noticed that the direct Jacobi updates may not converge.
To address the problem in (3) and (4), we propose a backward step on the dual update. Moreover,
instead of updating all blocks, the blocks xj will be updated in a parallel randomized block coordinate fashion. We call the algorithm Parallel Direction Method of Multipliers (PDMM). PDMM first
randomly select K blocks denoted by set Jt at time t, then executes the following iterates:
t
? t ) + ?jt B?jt (xjt , xtjt ) , jt ? Jt ,
xt+1
jt = argmin L? (xjt , xk6=jt , y
(5)
yit+1 = yit + ?i ?(Ai xt+1 ? ai ) ,
(6)
? it+1
y
(7)
xjt ?Xjt
=
yit+1
? ?i ?(Ai x
t+1
? ai ) ,
where ?i > 0, 0 ? ?i < 1, ?jt ? 0, and B?jt (xjt , xtjt ) is a Bregman divergence. Note xt+1 =
t
(xt+1
Jt , xk?J
/ t ) in (6) and (7). (6) and (7) update all dual blocks. We show that PDMM can also do
? i = min{di , K}. ?i and
randomized dual block coordinate ascent in an extended work [25]. Let K
?i can take the following values:
K
1
, ?i = 1 ?
?i =
.
(8)
?
?
Ki (2J ? K)
Ki
In the xjt -update (5), a Bregman divergence is addded so that exact PDMM and its inexact variants
can be analyzed in an unified framework [23, 11]. In particular, if ?jt = 0, (5) is an exact update. If
?jt > 0, by choosing a suitable Bregman divergence, (5) can be solved by various inexact updates,
often yielding a closed-form for the xjt update (see Section 2.1).
To better understand PDMM, we discuss the following three aspects which play roles in choosing ?i
and ?i : the dual backward step (7), the sparsity of A, and the choice of randomized blocks.
Dual Backward Step: We attribute the failure of the Jacobi updates (3)-(4) to the following observation in (3), which can be rewritten as:
?
xt+1
= argminxj ?Xj fj (xj ) + hyt + ?(Axt ? a), Acj xj i + kAcj (xj ? xtj )k22 .
(9)
j
2
In the primal xj update, the quadratic penalty term implicitly adds full gradient ascent step to the
dual variable, i.e., yt +?(Axt ?a), which we call implicit dual ascent. The implicit dual ascent along
with the explicit dual ascent (4) may lead to too aggressive progress on the dual variable, particularly
when the number of blocks is large. Based on this observation, we introduce an intermediate variable
? t to replace yt in (9) so that the implicit dual ascent in (9) makes conservative progress, e.g.,
y
? t + ?(Axt ? a) = yt + (1 ? ?)?(Axt ? a) , where 0 < ? < 1. y
? t is the result of a ?backward
y
? t = yt ? ??(Axt ? a).
step? on the dual variable, i.e., y
Moreover, one can show that ? and ? have also been implicitly used when using two-block ADMM
with splitting variables (sADMM) to solve (1) [17, 26]. Section 2.2 shows sADMM is a special case
of PDMM. The connection helps in understanding the role of the two parameters ?i , ?i in PDMM.
Interestingly, the step sizes ?i and ?i can be improved by considering the block sparsity of A and
the number of random blocks K to be updated.
Sparsity of A: Assume A is divided into I ? J blocks. While xj can be updated in parallel,
the matrix multiplication Ax in the dual update (4) requires synchronization to gather messages
from
coordinates
jt ? Jt . For updating the i-th block of the dual yi , we need Ai xt+1 =
P all blockt+1
P
t
jt ?Jt Aijt xjt +
k?J
/ t Aik xk which aggregates ?messages? from all xjt . If Aijt is a block of
3
P
zeros, there is no ?message? from xjt to yi . More precisely, Ai xt+1 = jt ?Jt ?N (i) Aijt xt+1
jt +
P
t
k?J
/ t Aik xk where N (i) denotes a set of nonzero blocks in the i-th row block Ai . N (i) can be
considered as the set of neighbors of the i-th dual block yi and di = |N (i)| is the degree of the i-th
dual block yi . If A is sparse, di could be far smaller than J. According to (8), a low di will lead to
bigger step sizes ?i for the dual update and smaller step sizes for the dual backward step (7). Further,
as shown in Section 2.3, when using PDMM with all blocks to solve composite minimization with
overlapping blocks, PDMM can use ?i = 0.5 which is much larger than 1/J in sADMM.
Randomized Blocks: The number of blocks to be randomly chosen also has the effect on ?i , ?i .
1
If randomly choosing one block (K = 1), then ?i = 0, ?i = 2J?1
. The dual backward step (7)
1
1
vanishes. As K increases, ?i increases from 0 to 1 ? di and ?i increases from 2J?1
to d1i . If
updating all blocks (K = J), ?i = d1i , ?i = 1 ? d1i .
PDMM does not necessarily choose any K combination of J blocks. The J blocks can be randomly
partitioned into J/K groups where each group has K blocks. Then PDMM randomly picks some
groups. A simple way is to permutate the J blocks and choose K blocks cyclically.
2.1
Inexact PDMM
If ?jt > 0, there is an extra Bregman divergence term in (5), which can serve two purposes. First,
choosing a suitable Bregman divergence can lead to an efficient solution for (5). Second, if ?jt is
sufficiently large, the dual update can use a large step size (?i = 1) and the backward step (7) can be
removed (?i = 0), leading to the same updates as PJADMM [10] (see Section 2.2).
Given a continuously differentiable and strictly convex function ?jt , its Bregman divergence is
defiend as
B?jt (xjt , xtjt ) = ?jt (xjt ) ? ?jt (xtjt ) ? h??jt (xtjt ), xjt ? xtjt i,
(10)
where ??jt denotes the gradient of ?jt . Rearranging the terms yields
?jt (xjt ) ? B?jt (xjt , xtjt ) = ?jt (xtjt ) + h??jt (xtjt ), xjt ? xtjt i,
(11)
which is exactly the linearization of ?jt (xjt ) at xtjt . Therefore, if solving (5) exactly becomes
difficult due to some problematic terms, we can use the Bregman divergence to linearize these
problematic terms so that (5) can be solved efficiently. More specifically, in (5), we can choose
?jt = ?jt ? ?1j ?jt assuming ?jt is the problematic term. Using the linearity of Bregman divert
gence,
1
B?jt (xjt , xtjt ) = B?jt (xjt , xtjt ) ?
B?jt (xjt , xtjt ) .
(12)
?jt
For instance, if fjt is a logistic function, solving (5) exactly requires an iterative algorithm. Setting
?jt = fjt , ?jt = 21 k? k22 in (12) and plugging into (5) yield
X
?
t
xt+1
yt , Ajt xjt i+ kAjt xjt +
Ak xtk ?ak22 +?jt kxjt ?xtjt k22 ,
jt = argmin h?fjt (xjt ), xjt i+h?
2
xjt ?Xjt
k6=jt
which has a closed-form solution. Similarly, if the quadratic penalty term ?2 kAcjt xjt +
P
?
c
2
c t
2
k6=jt Ak xk ? ak2 is a problematic term, we can set ?jt (xjt ) =
2 kAjt xjt k2 , then
?
c
t
2
t
B?jt (xjt , xjt ) = 2 kAjt (xjt ? xjt )k2 can be used to linearize the quadratic penalty term.
In (12), the nonnegativeness of B?jt implies that B?jt ? ?1j B?jt . This condition can be satisfied
t
as long as ?jt is more convex than ?jt . Technically, we assume that ?jt is ?/?jt -strongly convex
and ?jt has Lipschitz continuous gradient with constant ?, which has been shown in [23].
2.2
Connections to Related Work
Consider the case when all blocks are used in PDMM. There are also two other methods which
update all blocks in parallel. If solving the primal updates exactly, two-block ADMM with splitting
variables (sADMM) is considered in [17, 26]. We show that sADMM is a special case of PDMM
when setting ?i = J1 and ?i = 1 ? J1 (Appendix B in [25]). If the primal updates are solved
inexactly, [10] considers a proximal Jacobian ADMM (PJADMM) by adding proximal terms where
4
the converge rate is improved to o(1/T ) given the sufficiently large proximal terms. We show that
PJADMM [10] is also a special case of PDMM (Appendix C in [25]). sADMM and PJADMM are
two extreme cases of PDMM. The connection between sADMM and PJADMM through PDMM can
provide better understanding of the three methods and the role of dual backward step. If the primal
update is solved exactly which makes sufficient progress, the dual update should take small step, e.g.,
sADMM. On the other hand, if the primal update takes small progress by adding proximal terms,
the dual update can take full gradient step, e.g. PJADMM. While sADMM is a direct derivation of
ADMM, PJADMM introduces more terms and parameters.
In addition to PDMM, RBUSMM [14] can also randomly update one block. The convergence
of RBSUMM requires certain local error bounds to be hold and decreasing step size. Moreover,
the iteration complexity of RBSUMM is still unknown. In contast, PDMM converges at a rate of
O(1/T ) with the constant step size.
2.3
Randomized Overlapping Block Coordinate Descent
Consider the composite minimization problem of a sum of a loss function `(w) and composite
regularizers gj (wj ):
L
X
min `(w) +
gj (wj ) ,
(13)
w
j=1
which considers L overlapping groups wj ? Rb?1 . Let J = L + 1, xJ = w. For 1 ? j ? L,
denote xj = wj , then xj = UTj xJ , where Uj ? Rb?L is the columns of an identity matrix and
extracts the coordinates of xJ . Denote U = [U1 , ? ? ? , UL ] ? Rn?(bL) and A = [IbL , ?UT ] where
bL denotes b ? L. By letting fj (xj ) = gj (wj ) and fJ (xJ ) = `(w), (13) can be written as:
J
X
fj (xj ) s.t. Ax = 0.
(14)
min
x
j=1
b?J
where x = [x1 ; ? ? ? ; xL ; xL+1 ] ? R
. (14) can be solved by PDMM in a randomized block
coordinate fashion. In A, for b rows block, there are only two nonzero blocks, i.e., di = 2. ThereK
fore, ?i = 2(2J?K)
, ?i = 0.5. In particular, if K = J, ?i = ?i = 0.5. In contrast, sADMM uses
?i = 1/J 0.5, ?i = 1 ? 1/J > 0.5 if J is larger.
Remark 1 (a) ADMM [2] can solve (14) where the equality constraint is xj = UTj xJ .
(b) In this setting, Gauss-Seidel ADMM (GSADMM) and BSUMM [14] are the same as ADMM.
BSUMM should converge with constant stepsize ? (not necessarily sufficiently small), although the
theory of BSUMM does not include this special case.
3
Theoretical Results
We establish the convergence results for PDMM under fairly simple assumptions:
Assumption 1
(1) fj : Rnj 7? R ? {+?} are closed, proper, and convex.
(2) A KKT point of the Lagrangian (? = 0 in (2)) of Problem (1) exists.
Assumption 1 is the same as that required by ADMM [2, 22]. Assume that {x?j ? Xj , yi? } satisfies
the KKT conditions of the Lagrangian (? = 0 in (2)), i.e.,
? ATj y? ? ?fj (x?j ) ,
(15)
?
Ax ? a = 0.
(16)
fj0 (xt+1
j )
t+1
?fj (xt+1
j )
= a. Let
During iterations, (16) is satisfied if Ax
?
where ?fj be the
subdifferential of fj . For x?j ? Xj , the optimality conditions for the xj update (5) is
t+1
t+1
t
t
?
hfj0 (xjt+1 )+Acj [yt +(1??)?(Axt?a)+Acj (xt+1
j ?xj )]+?j (??j (xj )???j (xj )), xj ?xj i ? 0 .
When Axt+1 = a, yt+1 = yt . If Acj (xt+1
? xtj ) = 0, then Axt ? a = 0. When ?j ? 0, further
j
t+1
t
assuming B?j (xj , xj ) = 0, (15) will be satisfied. Note x?j ? Xj is always satisfied in (5) in
5
PDMM. Overall, the KKT conditions (15)-(16) are satisfied if the following optimality conditions
are satisfied by the iterates:
Axt+1 = a , Acj (xt+1
? xtj ) = 0 ,
(17)
j
t
B?j (xt+1
(18)
j , xj ) = 0 .
The above optimality conditions are sufficient for the KKT conditions. (17) are the optimality conditions for the exact PDMM. (18) is needed only when ?j > 0.
Let zij = Aij xj ? Rmi ?1 , zri = [zTi1 , ? ? ? , zTiJ ]T ? Rmi J?1 and z = [(zr1 )T , ? ? ? , (zrI )T ]T ?
RJm?1 . Define the residual of optimality conditions (17)-(18) as
I
J
X
?X
? t+1
t 2
r t+1
2
t+1
t
? z kPt +
?i kAi x
? ai k2 +
R(x ) = kz
?j B?j (xt+1
(19)
j , xj ) .
2
2 i=1
j=1
t+1
) ? 0, (17)-(18) will be
where Pt is some positive semi-definite matrix and ?i = JK
? i . If R(x
K
?
?
satisfied and thus PDMM converges to the KKT point {x , y }. Define the current iterate vt =
(xtj , yit ) and h(v? , vt ) as a distance from vt to a KKT point v? = (x?j ? Xj , yi? ):
h(v? , vt ) =
I
J
X
KX 1
?
kyi? ? yit?1 k22 + L?? (xt , yt ) + kz? ? zt k2Q +
?j B?j (x?j , xtj ) , (20)
J i=1 2?i ?
2
j=1
where Q is a positive semi-definite matrix and L?? (xt , yt ) with ?i = K?2(J?K)
+ d1i ? JK
? i is
K
i (2J?K)
I
X
(?i ? ?i )? r t
L?? (xt , yt ) = f (xt ) ? f (x? ) +
hyit , Ari xt ? ai i +
kAi x ? ai k22 .
(21)
2
i=1
The following Lemma shows that h(v? , vt ) ? 0.
Lemma 1 Let vt = (xtj , yit ) be generated by PDMM (5)-(7) and h(v? , vt ) be defined in (20).
K
Setting ?i = 1 ? K?1 and ?i = K? (2J?K)
, we have
i
h(v? , vt ) ?
where ?i =
J?K
? i (2J?K)
K
i
I
?X
J
X
?
?i kAri xt ? ai k22 + kz? ? zt k2Q +
?j B?j (x?j , xtj ) ? 0 .
2 i=1
2
j=1
+
1
di
?
K
?i
JK
(22)
? 0. Moreover, if h(v? , vt ) = 0, then Ari xt = ai , zt = z? and
B?j (x?j , xtj ) = 0. Thus, (15)-(16) are satisfied.
In PDMM, yt+1 depends on xt+1 , which in turn depends on Jt . xt and yt are independent of Jt . xt
depends on the observed realizations of the random variable ?t?1 = {J1 , ? ? ? , Jt?1 } .The following
theorem shows that h(v? , vt ) decreases monotonically and thus establishes the global convergence
of PDMM.
Theorem 1 (Global Convergence) Let vt = (xtj , yit ) be generated by PDMM (5)-(7) and v? =
K
(x?j ? Xj , yi? ) be a KKT point satisfying (15)-(16). Setting ?i = 1 ? K?1 and ?i = K? (2J?K)
, we
i
i
have
0 ? E?t h(v? , vt+1 ) ? E?t?1 h(v? , vt ) , E?t R(xt+1 ) ? 0 .
(23)
The following theorem establishes the iteration complexity of PDMM in an ergodic sense.
?T =
Theorem 2 (Iteration Complexity) Let (xtj , yit ) be generated by PDMM (5)-(7). Let x
PT
1
K
t
? and ?i = K
? i (2J?K) , we have
t=1 x . Setting ?i = 1 ? K
nP i
o
I
1
J
? 2
?? (x1 , y1 ) + ? kz? ? z1 k2 + PJ ?j B? (x? , x1 )
ky
k
+
L
2
j
i
j
j
Q
i=1
j=1
K
2?i ?
2
Ef (?
xT ) ? f (x? ) ?
,
T
2
I
?
0
X
? h(v , v )
r T
2
? ? ai k2 ?
?i kAi x
E
.
T
i=1
where ?i =
K
?i ,
JK
Q is a positive semi-definite matrix, and the expectation is over Jt .
6
4
residual (log)
2
1
0
?1
?2
?3
2
1
0
?1
?2
200
300
400
500
600
700
800
time (s)
?5
0
8.1
8.05
8
7.95
PDMM1
PDMM2
PDMM3
GSADMM
RBSUMM
sADMM
7.85
?4
100
8.15
7.9
?3
?4
?5
0
PDMM1
PDMM2
PDMM3
GSADMM
RBSUMM
sADMM
3
residual (log)
PDMM1
PDMM2
PDMM3
GSADMM
RBSUMM
sADMM
3
objective (log)
4
50
100
150
200
250
iterations
7.8
50
100
150
200
250
300
time (s)
Figure 1: Comparison of the convergence of PDMM with ADMM methods in RPCA.
Table 1: The best results of PDMM with tuning parameters ?i , ?i in RPCA.
time (s) iteration residual(?10?5 ) objective (log)
PDMM1
118.83
40
3.60
8.07
PDMM2
137.46
34
5.51
8.07
PDMM3
147.82
31
6.54
8.07
GSADMM 163.09
28
6.84
8.07
RBSUMM 206.96
141
8.55
8.07
sADMM1
731.51
139
9.73
8.07
Remark 2 PDMM converges at the same rate as ADMM and its variants. In Theorem 2, PDMM
can achieve the fastest convergence by setting J = K = 1, ?i = 1, ?i = 0, i.e., the entire matrix A
is considered as a single block, indicating PDMM reduces to the method of multipliers. In this case,
however, the resulting subproblem may be difficult to solve, as discussed in Section 1. Therefore,
the number of blocks in PDMM depends on the trade-off between the number of subproblems and
how efficiently each subproblem can be solved.
4
Experimental Results
In this section, we evaluate the performance of PDMM in solving robust principal component
analysis (RPCA) and overlapping group lasso [28]. We compared PDMM with ADMM [2] or
GSADMM (no theory guarantee), sADMM [17, 26], and RBSUMM [14]. Note GSADMM includes BSUMM [14]. All experiments are implemented in Matlab and run sequentially. We run
the experiments 10 times and report the average results. The stopping criterion is either when the
residual is smaller than 10?4 or when the number of iterations exceeds 2000.
RPCA: RPCA is used to obtain a low rank and sparse decomposition of a given matrix A corrupted
by noise [5, 17]:
1
(24)
min kX1 k2F + ?2 kX2 k1 + ?3 kX3 k? s.t. A = X1 + X2 + X3 .
2
where A ? Rm?n , X1 is a noise matrix, X2 is a sparse matrix and X3 is a low rank matrix.
A = L + S + V is generated in the same way as [17]1 . In this experiment, m = 1000, n = 5000
and the rank is 100. The number appended to PDMM denotes the number of blocks (K) to be chosen
in PDMM, e.g., PDMM1 randomly updates one block.
Figure 1 compares the convegence results of PDMM with ADMM methods. In PDMM, ? = 1
and ?i , ?i are chosen according to (8), i.e., (?i , ?i ) = {( 51 , 0), ( 14 , 12 ), ( 13 , 13 )} for PDMM1, PDMM2
and PDMM3 respectively. We choose the ?best?results for GSADMM (? = 1) and RBSUMM
11
(? = 1, ? = ? ?t+10
) and sADMM (? = 1). PDMMs perform better than RBSUMM and sADMM.
Note the public available code of sADMM1 does not have dual update, i.e., ?i = 0. sADMM should
be the same as PDMM3 if ?i = 13 . Since ?i = 0, sADMM is the slowest algorithm. Without
tuning the parameters of PDMM, GSADMM converges faster than PDMM. Note PDMM can run
in parallel but GSADMM only runs sequentially. PDMM3 is faster than two randomized version
of PDMM since the costs of extra iterations in PDMM1 and PDMM2 have surpassed the savings
at each iteration. For the two randomized one block coordinate methods, PDMM1 converges faster
than RBSUMM in terms of both the number of iterations and runtime.
The effect of ?i , ?i : We tuned the parameter ?i , ?i in PDMMs. Three randomized methods (RBSUMM, PDMM1 and PDMM2) choose the blocks cyclically instead of randomly. Table 1 compares the ?best?results of PDMM with other ADMM methods. In PDMM, (?i , ?i ) =
1
http://www.stanford.edu/ boyd/papers/prox algs/matrix decomp.html
7
0.5
PA?APG
S?APG
PDMM
ADMM
sADMM
0.3
0.2
0.1
0
0
0
PA?APG
S?APG
PDMM
ADMM
sADMM
0.4
objective
objective
0.4
0.3
0.2
0.1
50
100
time (s)
150
200
0
0
1
21
41
61
81
101
?1
residual (log)
0.5
?2
?3
?4
200
400
600
iteration
800
1000
?5
20
30
40
50
60
70
time (s)
Figure 2: Comparison of convergence of PDMM and other methods in overlapping group Lasso.
{( 12 , 0), ( 13 , 12 ), ( 12 , 21 )}. GSADMM converges with the smallest number of iterations, but PDMMs
can converge faster than GSADMM in terms of runtime. The computation per iteration in
GSADMM is slightly higher than PDMM3 because GSADMM updates the sum X1 + X2 + X3 but
PDMM3 can reuse the sum. Therefore, if the numbers of iterations of the two methods are close,
PDMM3 can be faster than GSADMM. PDMM1 and PDMM2 can be faster than PDMM3. By
simply updating one block, PDMM1 is the fastest algorithm and achieves the lowest residual.
Overlapping Group Lasso: We consider solving the overlapping group lasso problem [28]:
X
1
min
kAw ? bk22 +
dg kwg k2 .
w
g?G
2L?
(25)
where A ? Rm?n , w ? Rn?1 and wg ? Rb?1 is the vector of overlapping group indexed by
g. dg is some positive weight of group g ? G. As shown in Section 2.3, (25) can be rewritten
as the form (14). The data is generated in a same way as [27, 9]: the elements of A are sampled
from normal distribution, b = Ax + with noise sampled from normal distribution, and xj =
(?1)j exp(?(j ? 1)/100). In this experiment, m = 5000, the number of groups is L = 100, and
dg = L1 , ? = L5 in (25). The size of each group is 100 and the overlap is 10. The total number of
blocks in PDMM and sADMM is J = 101. ?i , ?i in PDMM are computed according to (8).
In Figure 2, the first two figures plot the convergence of objective in terms of the number of iterations
and time. PDMM uses all 101 blocks and is the fastest algorithm. ADMM is the same as GSADMM
in this problem, but is slower than PDMM. Since sADMM does not consider the sparsity, it uses
1
1
?i = J+1
, ?i = 1 ? J+1
, leading to slow convergence. The two accelerated methods, PA-APG [27]
and S-APG [9], are slower than PDMM and ADMM.
The effect of K: The third figure shows PDMM with different number of blocks K. Although the
complexity of each iteration is the lowest when K = 1, PDMM takes much more iterations than
other cases and thus takes the longest time. As K increases, PDMM converges faster and faster.
When K = 20, the runtime is already same as using all blocks. When K > 21, PDMM takes less
time to converge than using all blocks. The runtime of PDMM decreases as K increases from 21
to 61. However, the speedup from 61 to 81 is negligable. We tried different set of parameters for
2
+1
(0 ? i ? 5, ? = 0.01, 0.1, 1) or sufficiently small step size, but could not see the
RBSUMM ? ii+t
convergence of the objective within 5000 iterations. Therefore, the results are not included here.
5
Conclusions
We proposed a randomized block coordinate variant of ADMM named Parallel Direction Method of
Multipliers (PDMM) to solve the class of problem of minimizing block-separable convex functions
subject to linear constraints. PDMM considers the sparsity and the number of blocks to be updated
when setting the step size. We show two existing Jacobian ADMM methods are special cases of
PDMM. We also use PDMM to solve overlapping block problems. The global convergence and
the iteration complexity are established with constant step size. Experiments on robust PCA and
overlapping group lasso show that PDMM is faster than existing methods.
Acknowledgment
H. W. and A. B. acknowledge the support of NSF via IIS-1447566, IIS-1422557, CCF-1451986, CNS-1314560,
IIS-0953274, IIS-1029711, IIS-0916750, and NASA grant NNX12AQ39A. H. W. acknowledges the support
of DDF (2013-2014) from the University of Minnesota. A.B. acknowledges support from IBM and Yahoo.
Z.Q. Luo is supported in part by the US AFOSR via grant number FA9550-12-1-0340 and the National Science
Foundation via grant number DMS-1015346.
8
References
[1] F. Bach, R. Jenatton, J. Mairal, and G. Obozinski. Convex Optimization with Sparsity-Inducing Norms.
S. Sra, S. Nowozin, S. J. Wright., editors, Optimization for Machine Learning, MIT Press, 2011.
[2] S. Boyd, E. Chu N. Parikh, B. Peleato, and J. Eckstein. Distributed optimization and statistical learning via
the alternating direction method of multipliers. Foundation and Trends Machine Learning, 3(1):1?122,
2011.
[3] S. Boyd and L. Vandenberghe. Convex Optimization. Cambridge University Press, 2004.
[4] T. Cai, W. Liu, and X. Luo. A constrained `1 minimization approach to sparse precision matrix estimation.
Journal of American Statistical Association, 106:594?607, 2011.
[5] E. J. Candes, X. Li, Y. Ma, and J. Wright. Robust principal component analysis ?. Journal of the ACM,
58:1?37, 2011.
[6] V. Chandrasekaran, P. A. Parrilo, and A. S. Willsky. Latent variable graphical model selection via convex
optimization. Annals of Statistics, 40:1935?1967, 2012.
[7] C. Chen, B. He, Y. Ye, and X. Yuan. The direct extension of ADMM for multi-block convex minimization
problems is not necessarily convergent. Preprint, 2013.
[8] S. Chen, D.L. Donoho, and M.A. Saunders. Atomic decomposition by basis pursuit. SIAM review,
43:129?159, 2001.
[9] X. Chen, Q. Lin, S. Kim, J. G. Carbonell, and E. P. Xing. Smoothing proximal gradient method for
general structured sparse regression. The Annals of Applied Statistics, 6:719752, 2012.
[10] W. Deng, M. Lai, Z. Peng, and W. Yin. Parallel multi-block admm with o(1/k) convergence. ArXiv,
2014.
[11] Q. Fu, H. Wang, and A. Banerjee. Bethe-ADMM for tree decomposition based parallel MAP inference.
In UAI, 2013.
[12] D. Gabay and B. Mercier. A dual algorithm for the solution of nonlinear variational problems via finiteelement approximations. Computers and Mathematics with Applications, 2:17?40, 1976.
[13] B. He, M. Tao, and X. Yuan. Alternating direction method with Gaussian back substitution for separable
convex programming. SIAM Journal of Optimization, pages 313?340, 2012.
[14] M. Hong, T. Chang, X. Wang, M. Razaviyayn, S. Ma, and Z. Luo. A block successive upper bound
minimization method of multipliers for linearly constrained convex optimization. Preprint, 2013.
[15] M. Hong and Z. Luo. On the linear convergence of the alternating direction method of multipliers. ArXiv,
2012.
[16] Y. Nesterov. Efficiency of coordinate descent methods on huge-scale optimization methods. SIAM Journal
on Optimization, 22(2):341362, 2012.
[17] N. Parikh and S. Boyd. Proximal algorithms. Foundations and Trends in Optimization, 1:123?231, 2014.
[18] P. Richtarik and M. Takac. Iteration complexity of randomized block-coordinate descent methods for
minimizing a composite function. Mathematical Programming, 2012.
[19] N. Z. Shor. Minimization Methods for Non-Differentiable Functions. Springer-Verlag, 1985.
[20] R. Tappenden, P. Richtarik, and B. Buke. Separable approximations and decomposition methods for the
augmented lagrangian. Preprint, 2013.
[21] M. J. Wainwright and M. I. Jordan. Graphical models, exponential families, and variational inference.
Foundations and Trends in Machine Learning, 1:1?305, 2008.
[22] H. Wang and A. Banerjee. Online alternating direction method. In ICML, 2012.
[23] H. Wang and A. Banerjee. Bregman alternating direction method of multipliers. In NIPS, 2014.
[24] H. Wang, A. Banerjee, C. Hsieh, P. Ravikumar, and I. Dhillon. Large scale distributed sparse precesion
estimation. In NIPS, 2013.
[25] H. Wang, A. Banerjee, and Z. Luo. Parallel direction method of multipliers. ArXiv, 2014.
[26] X. Wang, M. Hong, S. Ma, and Z. Luo. Solving multiple-block separable convex minimization problems
using two-block alternating direction method of multipliers. Preprint, 2013.
[27] Y. Yu. Better approximation and faster algorithm using the proximal average. In NIPS, 2012.
[28] P. Zhao, G. Rocha, and B. Yu. The composite absolute penalties family for grouped and hierarchical
variable selection. Annals of Statistics, 37:34683497, 2009.
[29] Z. Zhou, X. Li, J. Wright, E. Candes, and Y. Ma. Stable principal component pursuit. In IEEE International Symposium on Information Theory, 2010.
9
| 5256 |@word version:3 norm:1 open:2 tried:1 decomposition:4 hsieh:1 pick:2 cyclic:1 substitution:2 liu:1 zij:1 tuned:1 interestingly:2 existing:2 current:1 luo:7 chu:1 written:1 numerical:1 j1:3 nonnegativeness:1 plot:1 update:39 kxjt:1 xk:4 fa9550:1 provides:1 iterates:2 successive:2 atj:1 along:2 mathematical:1 direct:5 become:1 symposium:1 yuan:2 introduce:2 manner:1 theoretically:1 peng:1 behavior:1 multi:3 decreasing:2 zhi:1 considering:1 increasing:1 becomes:1 provided:1 moreover:7 notation:1 linearity:1 lowest:2 argmin:2 unified:1 nj:4 guarantee:1 acj:8 runtime:4 exactly:8 axt:9 rm:8 k2:6 grant:3 positive:4 understood:1 local:3 ak:2 might:1 studied:1 fastest:3 acknowledgment:1 atomic:1 block:110 definite:3 x3:3 convegence:1 kpt:1 empirical:1 composite:8 boyd:4 spite:1 close:1 selection:2 impossible:1 writing:1 tappenden:1 www:1 map:1 lagrangian:6 yt:14 rnj:2 convex:16 ergodic:1 splitting:4 vandenberghe:1 rocha:1 handle:1 coordinate:27 updated:4 annals:3 pt:2 play:1 massive:1 exact:3 aik:2 programming:2 us:4 pa:3 element:1 trend:3 expensive:1 particularly:2 updating:4 jk:4 satisfying:1 observed:1 role:3 subproblem:2 preprint:4 wang:8 solved:10 wj:5 decrease:2 removed:1 trade:1 vanishes:1 complexity:10 nesterov:1 solving:8 kaw:1 technically:2 serve:1 efficiency:2 basis:1 various:1 derivation:1 distinct:1 effective:1 aggregate:1 choosing:4 saunders:1 seperable:1 widely:2 solve:15 larger:2 kai:3 relax:1 stanford:1 wg:1 statistic:3 noisy:1 online:1 differentiable:2 cai:1 propose:3 loop:1 realization:1 achieve:3 kx1:1 inducing:1 ky:1 convergence:21 double:1 converges:7 help:1 develop:1 linearize:2 ij:1 progress:5 solves:3 implemented:1 c:1 huwang:1 implies:1 direction:15 attribute:1 public:1 generalization:1 preliminary:1 strictly:1 extension:1 hold:2 sufficiently:5 considered:3 wright:3 normal:2 exp:1 achieves:1 smallest:1 purpose:1 favorable:1 estimation:2 rpca:7 grouped:1 city:1 establishes:2 minimization:12 mit:1 always:1 gaussian:1 zhou:1 shrinkage:1 ax:10 inherits:1 longest:1 rank:4 mainly:1 slowest:1 contrast:1 ibl:1 kim:1 sense:1 inference:2 stopping:1 entire:2 tao:1 overall:1 dual:40 html:1 denoted:1 k6:2 yahoo:1 art:2 constrained:3 fairly:3 special:5 ak2:1 smoothing:1 once:1 saving:1 encouraged:1 yu:2 k2f:1 icml:1 nonsmooth:4 np:1 report:1 randomly:13 algs:1 dg:3 divergence:7 national:1 xtj:10 cns:1 attempt:2 interest:1 message:3 huge:1 umn:2 introduces:1 analyzed:1 extreme:2 yielding:1 primal:13 regularizers:2 d1i:4 bregman:9 hyt:1 fj0:1 fu:1 indexed:1 tree:1 theoretical:1 instance:1 column:3 cost:1 introducing:2 too:1 corrupted:1 proximal:9 international:1 randomized:21 siam:3 l5:1 off:1 continuously:1 augmentation:1 satisfied:8 choose:5 american:1 zhao:1 leading:4 li:2 aggressive:1 prox:1 parrilo:1 twin:1 luozq:1 includes:1 ddf:1 depends:4 closed:3 xing:1 recover:1 parallel:24 candes:2 appended:1 sadmm:25 largely:1 efficiently:2 yield:2 richtarik:2 fore:1 executes:1 inexact:3 failure:1 dm:1 proof:1 di:9 jacobi:4 sampled:2 proved:1 intensively:1 ut:1 organized:1 jenatton:1 back:2 nasa:1 higher:1 improved:2 strongly:1 just:1 implicit:3 k2q:2 hand:1 nonlinear:1 banerjee:7 overlapping:17 logistic:1 effect:3 k22:6 ye:1 multiplier:18 ccf:1 regularization:1 equality:2 alternating:8 nonzero:4 dhillon:1 during:1 criterion:1 hong:3 performs:1 l1:1 fj:10 variational:2 consideration:2 arindam:1 ari:4 ef:1 parikh:2 empirically:1 discussed:1 association:1 he:2 bk22:1 cambridge:1 ai:14 tuning:2 mathematics:1 similarly:1 minnesota:2 stable:2 longer:1 behaving:2 gj:3 add:1 recent:1 certain:2 verlag:1 inequality:1 success:1 kx2:1 vt:13 yi:8 deng:1 converge:7 determine:1 argminxj:2 monotonically:1 ii:7 semi:3 multiple:6 full:2 reduces:1 seidel:3 smooth:1 technical:1 exceeds:1 faster:10 bach:1 long:1 lin:1 divided:3 lai:1 ravikumar:1 bigger:1 plugging:1 kax:1 variant:7 xjt:34 regression:1 expectation:1 surpassed:1 arxiv:3 iteration:26 addition:1 subdifferential:1 extra:2 rest:1 ascent:11 strict:2 subject:3 quan:1 jordan:1 call:4 structural:1 intermediate:1 variety:1 xj:43 iterate:1 shor:1 lasso:8 pca:1 reuse:1 ul:1 penalty:5 remark:2 matlab:1 detailed:1 category:3 http:1 canonical:2 problematic:4 nsf:1 per:1 rb:3 group:17 yit:8 kyi:1 pj:1 backward:10 sum:5 run:4 named:4 almost:1 chandrasekaran:1 family:2 appendix:2 bound:4 ki:2 apg:6 convergent:1 quadratic:3 rmi:5 constraint:11 precisely:1 x2:3 hy:1 aspect:2 u1:1 min:7 extremely:1 optimality:5 separable:8 utj:2 xtk:1 speedup:1 structured:1 according:3 combination:1 smaller:3 slightly:1 partitioned:1 taken:2 computationally:1 slack:1 discus:1 turn:1 nnx12aq39a:1 needed:1 merit:1 letting:1 mercier:1 available:1 pursuit:2 rewritten:2 apply:1 kwg:1 hierarchical:1 stepsize:2 slower:2 denotes:4 include:2 graphical:2 k1:1 uj:1 establish:4 negligable:1 classical:1 bl:2 objective:7 noticed:1 added:1 already:1 unclear:1 gradient:7 distance:1 gence:1 carbonell:1 considers:6 willsky:1 assuming:2 code:1 minimizing:3 difficult:3 subproblems:1 proper:1 zt:3 unknown:2 fjt:3 perform:1 upper:2 observation:3 acknowledge:1 descent:9 extended:3 y1:1 rn:2 sweeping:1 arbitrary:1 peleato:1 cast:1 required:2 eckstein:1 connection:4 z1:1 huahua:1 established:3 nip:3 address:1 usually:1 parallelism:1 sparsity:8 including:1 wainwright:1 suitable:2 overlap:1 residual:7 acknowledges:2 extract:1 review:1 understanding:3 multiplication:1 afosr:1 synchronization:1 loss:2 foundation:4 degree:1 gather:1 sufficient:2 editor:1 nowozin:1 ibm:1 row:4 supported:1 aij:3 understand:1 neighbor:1 absolute:1 sparse:7 distributed:2 kari:1 kz:4 rbcd:2 far:1 implicitly:2 global:5 kkt:7 sequentially:2 uai:1 mairal:1 conclude:1 alternatively:1 continuous:1 iterative:1 latent:1 table:2 bethe:1 robust:6 rearranging:1 sra:1 ajt:1 necessarily:3 linearly:2 noise:3 gabay:1 razaviyayn:1 x1:6 augmented:5 fashion:6 slow:1 precision:1 explicit:1 exponential:1 xl:2 third:2 jacobian:4 cyclically:2 theorem:5 xt:29 jt:61 exists:1 adding:3 supplement:1 linearization:1 kx:1 chen:3 ak22:2 yin:1 simply:3 chang:1 springer:1 inexactly:2 satisfies:1 acm:1 ma:4 obozinski:1 identity:1 formulated:1 donoho:1 careful:1 decomp:1 replace:1 admm:39 lipschitz:1 included:1 specifically:1 principal:6 conservative:2 called:1 zr1:1 lemma:2 total:1 experimental:2 gauss:3 takac:1 indicating:1 select:1 support:3 accelerated:1 evaluate:3 |
4,701 | 5,257 | Constant Nullspace Strong Convexity and
Fast Convergence of Proximal Methods under
High-Dimensional Settings
Ian E.H. Yen
Cho-Jui Hsieh
Pradeep Ravikumar
Inderjit Dhillon
Department of Computer Science
University of Texas at Austin
{ianyen,cjhsieh,pradeepr,inderjit}@cs.utexas.edu
Abstract
State of the art statistical estimators for high-dimensional problems take the form
of regularized, and hence non-smooth, convex programs. A key facet of these
statistical estimation problems is that these are typically not strongly convex under a high-dimensional sampling regime when the Hessian matrix becomes rankdeficient. Under vanilla convexity however, proximal optimization methods attain
only a sublinear rate. In this paper, we investigate a novel variant of strong convexity, which we call Constant Nullspace Strong Convexity (CNSC), where we require that the objective function be strongly convex only over a constant subspace.
As we show, the CNSC condition is naturally satisfied by high-dimensional statistical estimators. We then analyze the behavior of proximal methods under this
CNSC condition: we show global linear convergence of Proximal Gradient and local quadratic convergence of Proximal Newton Method, when the regularization
function comprising the statistical estimator is decomposable. We corroborate our
theory via numerical experiments, and show a qualitative difference in the convergence rates of the proximal algorithms when the loss function does satisfy the
CNSC condition.
1 Introduction
There has been a growing interest in high-dimensional statistical problems, where the number of
parameters d is comparable to or even larger than the sample size n, spurred in part by many modern
science and engineering applications. It is now well understood that in order to guarantee statistical
consistency it is key to impose low-dimensional structure, such as sparsity, or low-rank structure,
on the high-dimensional statistical model parameters. A strong line of research has thus developed
classes of regularized M -estimators that leverage such structural constraints, and come with strong
statistical guarantees even under high-dimensional settings [13]. These state of the art regularized
M -estimators typically take the form of convex non-smooth programs.
A facet of computational consequence with these high-dimensional sampling regimes is that these
M -estimation problems, even when convex, are typically not strongly convex. For instance, for
the ?1 -regularized least squares estimator (LASSO), the Hessian is rank deficient when n < d. In
the absence of additional assumptions however, optimization methods to solve general non-smooth
non-strongly convex programs can only achieve a sublinear convergence rate [19, 21]; faster rates
typically require strong convexity [1, 20]. In the past few years, an effort has thus been made to
impose additional assumptions that are stronger than mere convexity, and yet weaker than strong
convexity; and proving faster rates of convergence of optimization methods under these assumptions. Typically these assumptions take the form of a restricted variant of strong convexity, which
incidentally mirror those assumed for statistical guarantees as well, such as the Restricted Isometry
1
Property or Restricted Eigenvalue property. A caveat with these results however is that these statistically motivated assumptions need not hold in general, or require sufficiently large number of samples
to hold with high probability. Moreover, the standard optimization methods have to be modified in
some manner to leverage these assumptions [5, 7, 17]. Another line of research exploits a local error
bound to establish asymptotic linear rate of convergence for a special form of non-strongly convex
functions [16, 8, 6]. However, these do not provide finite-iteration convergence bounds, due to the
potentially large number of iterations spent on early stage.
In this paper, we consider a novel simple condition, which we term Constant Nullspace Strong
Convexity (CNSC). This assumption is motivated not from statistical considerations, but from the
algebraic form of standard M -estimators; indeed as we show, standard M -estimation problems even
under high-dimensional settings naturally satisfy the CNSC condition. Under this CNSC condition,
we then investigate the convergence rates of the class of proximal optimization methods; specifically
the Proximal Gradient method (Prox-GD) [14, 15, 18] and the Proximal Newton method (ProxNewton) [1, 2, 9]. These proximal methods are very amenable to regularized M -estimation problems: they do not treat the M -estimation problem as a black-box convex non-smooth problem, but
instead leverage the composite nature of the objective of the form F (x) = h(x)+f (x), where h(x)
is a possibly non-smooth convex function while f (x) is a convex smooth function with Lipschitzcontinuous gradient. We show that under our CNSC condition, Proximal Gradient achieves global
linear convergence when the non-smooth component is a decomposable norm. We also show that
Proximal Newton, under the CNSC condition, achieves local quadratic convergence as long as the
non-smooth component is Lipschitz-continuous. Note that in the absence of strong convexity, but
under no additional assumptions beyond convexity, the proximal methods can only achieve sublinear convergence as noted earlier. We have thus identified an algebraic facet of the M -estimators
that explains the strong computational performance of standard proximal optimization methods in
practical settings in solving high-dimensional statistical estimation problems.
The paper is organized as follows. In Section 2, we define the CNSC condition and introduce
the Proximal Gradient and Proximal Newton methods. Then we prove global linear convergence
of Prox-GD and local quadratic convergence of Prox-Newton in Section 3 and 4 respectively. In
Section 5, we corroborate our theory via experiments on real high-dimensional data set. We will
leave all the proof of lemmas to the appendix.
2
Preliminaries
We are interested in composite optimization problems of the form
min
x?Rd
F (x) = h(x) + f (x),
(1)
where h(x) is a possibly non-smooth convex function and f (x) is twice differentiable convex function with its Hessian matrix H(x) = ?2 f (x) satisfying
mI ? H(x) ? M I,
?x ? Rd ,
(2)
where for strongly convex f (x) we have m > 0; otherwise, for convex but not strongly convex f (x)
we have m = 0.
2.1
Constant Nullspace Strong Convexity (CNSC)
Before defining our strong convexity variant of Constant Nullspace Strong Convexity (CNSC), we
first provide some intuition by considering the following large class of statistical estimation problems
in high-dimensional machine learning, where f (x) takes the form
f (x) =
n
?
L(aTi x, yi ),
(3)
i=1
where L(u, y) is a non-negative loss function that is convex in its first argument, ai is the observed
feature vector and yi is the observed response of the i-th sample. The Hessian matrix of (3) takes
the form
H(x) = AT D(Ax)A,
(4)
2
where A is a n by d design (data) matrix with Ai,: = aTi and D(Ax) is a diagonal matrix with
??
Dii (x) = L (aTi x, yi ), where the double-derivative in L?? (u, y) is with respect to the first argument. It is easy to see that in high-dimensional problems with d > n, (4) is not positive definite so
that strong convexity would not hold. However, for strictly convex loss function L(?, y), we have
??
L (u, y) > 0 and
(5)
v T H(x)v = 0 iff Av = 0.
As a consequence v T H(x)v > 0 as long as v does not lie in the Nullspace of A; that is, the Hessian
H(x) might satisfy the strong convexity bound in the above restricted sense. We generalize this
concept as follows. We first define the following notation: given a subspace T , we let ?T (?) denote
the orthogonal projection onto T , and let T ? denote the orthogonal subspace to T .
Assumption 1 ( Constant Nullspace Strong Convexity ). A twice-differentiable f (x) satisfies Constant Nullspace Strong Convexity (CNSC) with respect to T (CNSC-T ) iff there is a constant vector
space T s.t. f (x) depends only on z = ?T (x) and its Hessian matrix satisfies
for some m > 0, and ?z ? T ,
v T H(z)v ? m?v?2 , ?v ? T
(6)
H(z)v = 0, ?v ? T ? .
(7)
From the motivating section above, the above condition can be seen to hold for a wide range of loss
functions, such as those arising from linear regression models, as well as generalized linear models
??
(e.g. logistic regression, poisson regression, multinomial regression etc.) 1 . For L (u, y) ? mL >
0, we have m = mL ?min (AT A) > 0 as the constant in (6), where ?min (AT A) is the minimum
positive eigenvalue of AT A.
Then by the assumption, any point x can be decomposed as x = z + y, where z = ?T (x),
y = ?T ? (x), so that the difference between gradient of two points can be written as
? 1
? 1
? 1 , z 2 )?z, (8)
g(x1 ) ? g(x2 ) =
H(s?x + x2 )?xds =
H(s?z + z 2 )?zds = H(z
0
0
?
? 1 , z 2 ) = 1 H(s?z + z 2 )ds is the average Hessian
where ?x = x1 ? x2 , ?z = z 1 ? z 2 , and H(z
0
? 1 , z 2 ) satisfies inequalities (2),
matrix along the path from z 2 to z 1 . It is easy to verify that H(z
(6) and equality (7) for all z 1 , z 2 ? T by just applying inequalities (equality) to each individual
?
Hessian matrix being integrated. Then we have following theorem that shows the uniqueness of z
at optimal.
Theorem 1 (Optimality Condition). For f (x) satisfying CNSC-T ,
? is an optimal solution of (1) iff ?g(?
? for some ?
? ? ?h(?
1. x
x) = ?
x).
? and z
? = ?T (?
2. The optimal ?
x) are unique.
? is an optimal solution iff 0 ? ?h(?
Proof. The first statement is true since x
x) + ?f (?
x). To prove
?1 = z
?1 + y
? 1 and x
?2 = z
?2 + y
? 2 are both optimal. Let ?x = x
?1 ?x
?2
the second statement, suppose x
?1 ? z
? 2 . Since h(x) is convex, ?g(?
and ?z = z
x1 ) ? ?h(?
x1 ) and ?g(?
x2 ) ? ?h(?
x2 ) should satisfy
??g(?
x1 ) + g(?
x2 ), ?x? ? 0.
However, since f (x) satisfies CNSC-T , by (8),
? z1, z
? z1, z
? 2 )?z, ?x? = ??z H(?
? 2 )?z ? ?m??z?22
??g(?
x1 ) + g(?
x2 ), ?x? = ??H(?
? is
for some m > 0. The two inequalities can simultaneously hold only if ??
z = 0. Therefore, z
? z , 0)?
? = ?g(?
unique at optimum, and thus g(?
x) = g(0) + H(?
z and ?
x) are also unique.
In next two sections, we review the Proximal Gradient Method (Prox-GD) and Proximal Newton
Method (Prox-Newton), and introduce some tools that will be used in our analysis.
??
1
Note for many generalized linear models, the second derivative L (u, y) of loss function approaches 0 if |u| ? ?. However, this
could not happen as long as there is a penalty term h(x) which goes to infinity if x diverges, which then serves as a finite constraint bound on
x.
3
2.2
Proximal Gradient Method
The Prox-GD algorithm comprises a gradient descent step
xt+ 12 = xt ?
1
g(xt )
M
followed by a proximal step
xt+1 = proxhM (xt+ 12 ) = arg min h(x) +
x
M
?x ? xt+ 12 ?22 ,
2
(9)
where ? ? ?2 means the Frobinius norm if x is a matrix. For simplicity, we will denote proxhM (.) as
prox(.) in the following discussion when it is clear from the context. In Prox-GD algorithm, it is
assumed that (9) can be computed efficiently, which is true for most of decomposable regularizers.
Here we introduce some properties of proximal operator that can facilitate our analysis.
Lemma 1. Define ?P x = x ? prox(x), the following properties hold for proximal operation (9).
1. M ?P x ? ?h(prox(x)).
2. ?prox(x1 ) ? prox(x2 )?22 ? ?x1 ? x2 ?22 ? ??P x1 ? ?P x2 ?22 .
2.3
Proximal Newton Method
In this section, we introduce the Proximal Newton method, which has been shown to be considerably more efficient than first-order methods in many applications [1], including Sparse Inverse Covariance Estimation [2] and ?1 -regularized Logistic-Regression [9, 10]. Each step of Prox-Newton
solves a local quadratic approximation
1
T
T
x+
t = arg min h(x) + (x ? xt ) Ht (x ? xt ) + g t (x ? xt )
2
x
(10)
to find a search direction x+ ? xt , and then conduct a line search procedure to find t such that
f (xt+1 ) = f (xt + t(x+
t ? xt ))
meets a sufficient decrease condition. Note unlike Prox-GD update (9), in most of cases (10) requires
an iterative procedure to solve. For example if h(x) is ?1 -norm, then a coordinate descent algorithm
is usually employed to solve (10) as an LASSO subproblem [1, 2, 9, 10].
The convergence of Newton-type method comprises two phases [1, 3]. In the first phase, it is possible
that step size t < 1 is chosen, while in the second phase, which occurs when xt is close enough
to optimum, step size t = 1 is always chosen and each step leads to quadratic convergence. In this
paper, we focus on the quadratic convergence phase, while refer readers to [21] for a global analysis
of Prox-Newton without strong convexity assumption. In the quadratic convergence phase, we have
xt+1 = x+
t and the update can be written as
(
)
xt+1 = proxHt xt + ?xnt
, Ht ?xnt
(11)
t
t = ?g t ,
where ?xnt
t is the Newton step when h(x) is absent, and the proximal operator proxH (.) is defined
for any PSD matrix H as
1
proxH (x) = arg min h(v) + ?v ? x?2H .
2
v
(12)
Note while we use ?x?2H to denote xT Hx, we only require H to be PSD instead of PD. Therefore,
?x?H is not a true norm, and (12) might have multiple solutions, where proxH (x) refers to any
one of them. In the following, we show proxH (.) has similar properties as that of prox(.) in
previous section.
Lemma 2. Define ?P x = x ? proxH (x), the following properties hold for the proximal operator:
1. H?P x ? ?h(proxH (x)).
2. ?proxH (x1 ) ? proxH (x2 )?2H ? ?x1 ? x2 ?2H .
4
3
Linear Convergence of Proximal Gradient Method
In this section, we analyze convergence of Proximal Gradient Method for h(x) = ??x?, where ? ? ?
is a decomposable norm defined as follows.
Definition 1 (Decomposable Norm). ? ? ? is a decomposable norm if there are orthogonal subd
J
d
J
spaces
?{Mi }i=1 with R = ?i=1 Mi such that for any point x ? R that can be written as
x = j?E cj aj , where cj > 0 and aj ? Mj , ?aj ?? = 1, we have
?
?x? =
cj , and ??x? = {? | ?Mj (?) = aj , ?j ? E; ??Mj (?)?? ? 1, ?j ?
/ E}, (13)
j?E
where ? ? ?? is the dual norm of ? ? ?.
The above definition includes several well-known examples such as ?1 -norm ?x?1 and group-?1
norm ?X?1,2 . For ?1 -norm, Mj corresponds to vectors with only j-th coordinate not equal to 0,
and E is the set of non-zero coordinates of x. For group-?1 norm, Mj corresponds to vectors with
only j-th group not equal to 0T and E are the set of non-zero groups of X. Under the definition, we
can profile the set of optimal solutions as follows.
Lemma 3 (Optimal Set). Let E? be the active set at optimal and E?+ = {j| ? ?Mj (?
?)?? = ?} be its
? is unique) such that ?Mj (?
augmented set (which is unique since ?
?) = ??
aj , j ? E?+ . The optimal
solutions of (1) form a polyhedral set
{
}
? ,
? and x ? O
X? = x | ?T (x) = z
(14)
}
{
?
? = x|x=
? j , cj ? 0, j ? E?+ is the set of x with ?
? ? ?h(x).
where O
j?E?+ cj a
Given the optimal set is a polyhedron, we can then employ the following lemma to bound the distance of an iterate xt to the optimal set X? .
Lemma 4 (Hoffman?s bound). Consider a polyhedral set S = {x|Ax ? b, Ex = c}. For any point
? ? S such that
x ? Rd , there is a x
[Ax ? b]+
,
? ?2 ? ?(S)
?x ? x
(15)
Ex ? c
2
where ?(S) is a positive constant that depends only on A and E.
The above bound first appears in [11], and was employed in [4] to prove linear convergence of
Feasible Descent method for a class of convex smooth function. A proof of the ?2 -norm version (15)
can be found in [4, lemma 4.3]. By applying (15) to the set X? , the distance of a point x to X? can be
? where the latter can
bounded by infeasible amounts to the two constraints ?T (x) = z and x ? O,
? j ? ? 0, ?j ? E?+ .
be bounded according the following lemma when cj = ?x, a
?2 . . . , a
? |E?+ | ). Suppose ?x? ? R and ?Mj (x) = 0 for j ?
Lemma 5. Let A? = span(?
a1 , a
/ E?+ .
Then
? ?22 ,
?2 ?x ? ?A?(x)?22 ? R2 ?? ? ?
? is as defined in Theorem 1.
where ? ? ?h(x) and ?
Now we are ready to prove the main theorem of this section.
Theorem 2 (Linear Convergence of Prox-GD). Let X? be the set of optimal
( solutions for problem
)
? = ?X? (x) be the solution closest to x. Denote d? = minj ?/ E?+ ? ? ??Mj (?
(1), and x
?)?? > 0.
For the sequence {xt }?
t=0 produced by Proximal Gradient Method, we have:
(a) If xt+1 satisfies the condition that
? j ? < 0,
?j ?
/ E?+ : ?Mj (xt+1 ) ?= 0 or ?j ? E?+ : ?xt+1 , a
we then have:
d2?
? t+1 ?22 ? (1 ? ?)?xt ? x
? t ?22 , ? = 2
?xt+1 ? x
? 0 ?22
M ?x0 ? x
5
(16)
(17)
(b) If xt+1 does not satisfy the condition in (16) but xt does, then
? t+1 ?22 ? (1 ? ?)?xt?1 ? x
? t?1 ?22 , ? =
?xt+1 ? x
d2?
? 0 ?22
M 2 ?x0 ? x
(18)
m
,
M ?(X? )2
(19)
(c) If neither xt+1 , xt satisfy the condition in (16), then
? t+2 ?22 ?
?xt+2 ? x
1
? t ?22 ,
?xt ? x
1+?
?=
where we recall that ?(X? ) is the constant determined by polyhedron X? from Hoffman?s
Bound (15).
? t is an optimal solution, we have x
? t = prox(?
? t,
Proof. Since x
xt ? g(?
xt )/M ). Let ?xt = xt ? x
? = H(z
? t, z
? t ). by Lemma 1, each iterate of Prox-GD
?t = M (xt+ 21 ? xt+1 ) ? ?h(xt+1 ) and H
has
? t ?22 ? ?xt+1 ? x
? t+1 ?22 ? ?xt ? x
? t ?22 ? ?xt+1 ? x
? t ?22
?xt ? x
= ??xt ?22 ? ?prox(xt ? g(xt )/M ) ? prox(?
xt ? g(?
xt )/M )?22
?
??xt ?22
? ?(xt ? g(xt )/M ) ? (?
xt ?
g(?
xt )/M )?22
+ ??t ?
(20)
? ?22 /M 2 .
?
?
Since g(xt ) ? g(?
xt ) = H?x
from (8), we have
2
?
? ?22 /M 2
? t+1 ?22 ? ??xt ?22 ? ??xt ? H?x
? t ?22 ? ?xt+1 ? x
?xt ? x
t /M ?2 + ??t ? ?
(
)
?
? ?22 /M 2
? ?xTt H/M
?xt + ??t ? ?
(21)
? ?22 /M 2 .
? m??z t ?22 /M + ??t ? ?
?
? 2 /M 2 = (H/M
?
?
?
The second inequality holds since 2H/M
?H
)(2I ? H/M
) ? H/M
. The
2
2
? t ? ? ?xt+1 ? x
? t+1 ? ? 0, that is, the distance to the optimal set
inequality tells us ?xt ? x
? t ? is monotonically non-increasing. To get a tighter bound, we consider two cases.
?xt ? x
? j ? < 0 for some j ? E?+ .
Case 1: ?M (xt ) ?= 0 for some j ?
/ E?+ or ?xt , a
j
In this case, suppose there is j ?
/ Et+ with ?Mj (xt ) ?= 0, then 2
? ?22 ? ??Mj (?t ) ? ?Mj (?
?)?? )2 ? d2? .
?)?2? ? (??Mj (?t )?? ? ??Mj (?
??t ? ?
(22)
? j ? < 0 for some j ? E?+ , then we have ?aj , a
? j ? < 0 for ?Mj (?t ) =
On the other hand, if ?xt , a
?aj . Therefore
? j ?) > 2?2 .
? j ?22 = ?2 (2 ? 2?aj , a
? ?22 ? ??Mj (?t ) ? ?Mj (?
?)?22 ? ?2 ?aj ? a
??t ? ?
Either cases we have
? t+1 ?22 ?
? t ?22 ? ?xt+1 ? x
?xt ? x
? ?22
??t ? ?
?
2
M
(
d2?
2
? 0 ?22
M ?x0 ? x
)
? t ?22 .
?xt ? x
(23)
Case 2: Both xt , xt+1 do not fall in Case 1
? defined in
? j ? ? 0, ?j ? E?+ and ?Mj (xt ) = 0, ?j ?
Given ?xt , a
/ E?+ , then x belongs to the set O
2
?
2
Lemma 3 iff ?x ? ?A?(x)?22 = 0. The condition can be also scaled as mM
?(x)?2 = 0,
R2 ?x ? ?A
where R is a bound on ?xt ? holds for ?t, which must exist as long as the regularization parameter
? > 0 in h(x) = ??x?.
By Lemma 4, the distance of point xt to the polyhedral set X? is bounded by its infeasible amount
(
)
?2
2
? t ?22 ? ?(X? )2 ?z t ? z
? ?22 +
?xt ? x
?x
?
?
(x
)?
(24)
?
t
t 2 ,
A
mM R2
2
From our definition of decomposable norm, if a vector v belongs to single subspace Mj , then ?v? = ?v?? = ?v?2 . The reason
is: By the definition, if v ? Mj , then v = cj aj for some cj > 0, aj ? Mj , ?aj ?? = 1, and it has decomposable norm ?v? = cj .
However, we also have ?v?? = ?cj aj ?? = cj ?aj ?? = cj = ?v?. The norm equals to its dual norm only if it is ?2 -norm.
6
where z t = ?T (xt ). Applying (24) to (21) for iteration t + 1, we have
? t+1 ?2 ? ?xt+2 ? x
? t+2 ?2
?xt+1 ? x
? ?2
??
??
?2
m
.
??xt+1 ?2 ? 2 2 ?xt+1 ? ?A?(xt+1 )?22 + t+1 2
2
?
M R
M
M ?(X )
For iteration t, we have
? ?2
m
?? ? ?
? t ?2 ? ?xt+1 ? x
? t+1 ?2 ?
?xt ? x
??z t ?22 + t 2
M
M
. By Lemma 5, adding the two inequalities gives
?
? ?2
??t+1 ? ?
m
m
2
2
??x
?
??z
?
+
+
t+1
t
2
M
M2
M ?(X? )2
m
m
?
??xt+1 ?2 ?
??xt+2 ?2 ,
M ?(X? )2
M ?(X? )2
which yields desired result (18) after arrangement.
? t ?2 ? ?xt+2 ? x
? t+2 ?2 ?
?xt ? x
We note that the descent in the first two cases is actually even stronger than stated above: from the
proofs, that the distance can be seen to reduce by a fixed constant. This is faster than superlinear
convergence since the final solution could then be obtained in a finite number of steps.
4
Quadratic Convergence of Proximal Newton Method
The key idea of the proof is to re-formulate Prox-Newton update (10) as
1
? (z)) + g Tt (z ? z t ) + ?z ? z t ?2Ht
z t+1 = arg min h(z + y
2
z?T
where
? (z) = arg min
y
h(z + y),
y?T ?
(25)
(26)
so that we can focus our convergence analysis on z = ?T (x) as follows.
Lemma 6 (Optimality Condition). For any matrix H satisfying CNSC-T , the update
1
?x = arg min h(x + d) + g(x)T d + ?d?2H
2
d
has
F (x + t?x) ? F (x) ? ?t??z?2H + O(t2 ),
where ?z = ?T (?x). Furthermore, if x is an optimal solution, ?x = 0 satisfies (27).
(27)
(28)
The following lemma then states that, for Prox-Newton, the function suboptimality is bounded by
only distance in the T space.
Lemma 7. Suppose h(x) and f (x) are Lipschitz-continuous with Lipschitz constants Lh and Lf .
In quadratic convergence phase (defined in Theorem 3), Proximal Newton Method has
? ?,
F (xt ) ? F (?
x) ? L?z t ? z
(29)
? = ?T (?
where L = max{Lh , Lf } and z t = ?T (xt ), z
x).
? ? ? ?. Therefore, it suffices
By the above lemma, we have F (xt ) ? F (?
x) ? L? as long as ?z t ? z
? ? to guarantee F (xt ) ? F (?
to show quadratic convergence of ?z t ? z
x) double its precision after
each iteration.
Theorem 3 (Quadratic Convergence of Prox-Newton). For f (x) satisfying CNSC-T with Lipschitzcontinuous second derivative ?2 f (x), the Proximal Newton update (10) has
LH
? ?2 ,
?? ?
?z t ? z
?z t+1 ? z
2m
? = ?T (?
where z
x), z t = ?T (xt ), and LH is the Lipschitz constant for ?2 f (x).
7
? be an optimal solution of (1). By Lemma 6, for any PSD matrix H the update ??
Proof. Let x
x=0
satisfies (27), which means
? = proxHt (?
x
x + ??
xnt ), Ht ??
xnt = ?g(?
x).
(30)
Then by non-expansiveness of proximal operation (Lemma 2), we have
? ?Ht = ?proxHt (xt + ?xnt
?xt+1 ? x
x + ??
xnt )?Ht
t ) ? proxHt (?
? ) + (?xnt
? ?(xt + ?xnt
x + ??
xnt )?Ht = ?(xt ? x
xnt )?Ht
t ) ? (?
t ? ??
(31)
?) +
= ?(z t ? z
?
?
Since for z ? T , ?Ht z?2 ? m?z?Ht , (31) leads to
1
? ?Ht ? ? ?Ht (z t ? z
? ) ? Ht (?z nt
?xt+1 ? x
z nt )?2
t ? ??
m
(32)
LH
1
? ) ? (g t ? g
? )?2 ? ? ?z t ? z
? ?22 ,
= ? ?Ht (z t ? z
m
2 m
? ? T , we have
where last inequality follows from Lipschitz-continuity of ?2 f (x). Since z t+1 , z
?
? ?Ht = ?z t+1 ? z
? ?Ht ? m?z t+1 ? z
? ?2 .
?xt+1 ? x
(33)
Finally, combining (33) with (32),
LH
? ?22 ,
? ?2 ?
?z t ? z
?z t+1 ? z
2m
?
2m
?? < L
where quadratic convergence phase occurs when ?z t ? z
.
H
(?z nt
t
5
??
z nt
t )?Ht .
Numerical Experiments
In this section, we study the convergence behavior of Proximal Gradient method and Proximal Newton method on high-dimensional real data set with and without the CNSC condition.
In particular, two loss functions ? logistic loss L(u, y)=log(1 + exp(?yu)) and ?2 -hinge loss
L(u, y)=max(1 ? yu, 0)2 ? are used in (3) with ?1 -regularization h(x) = ??x?1 , where both
losses are smooth but only logistic loss has strict convexity that implies the CNSC condition. For
Proximal Newton method we employ an randomized coordinate descent algorithm to solve subproblem (10) as in [9]. Figure 5 shows their convergence results of objective value relative to the
optimum on rcv1.1k, subset of a document classification data set with dimension d = 10, 192 and
number of samples n = 1000. From the figure one can clearly observe the linear convergence of
Prox-GD and quadratic convergence of Prox-Newton on problem satisfying CNSC, contrasted to
the qualitatively different behavior on problem without CNSC.
Prox?GD
Prox?Newton
logistic
L2hinge
?2
10
0
logistic
L2hinge
10
?2
10
?4
obj
obj
10
?4
10
?6
10
?6
10
?8
10
?8
10
0.5
1
1.5
iter
2
2.5
3
5
6
x 10
10
15
iter
20
25
30
Figure 1: objective value (relative to optimum) of Proximal Gradient method (left) and Proximal
Newton method (right) with logistic loss and ?2 -hinge loss.
Acknowledgement
This research was supported by NSF grants CCF-1320746 and CCF-1117055. C.-J.H acknowledges
support from an IBM PhD fellowship. P.R. acknowledges the support of ARO via W911NF-12-10390 and NSF via IIS-1149803, IIS-1320894, IIS-1447574, and DMS-1264033.
8
References
[1] J. D. Lee, Y. Sun, and M. A. Saunders. Proximal newton-type methods for minimizing composite functions. In NIPS, 2012.
[2] C.-J. Hsieh, M. A. Sustik, I. S. Dhillon, and P. Ravikumar. Sparse inverse covariance estimation
using quadratic approximation. In NIPS 2011.
[3] S. Boyd and L. Vandenberghe, Convex Optimization, Cambridge Univ. Press, Cambridge, U.K.,
2003.
[4] P.-W. Wang and C.-J. Lin. Iteration Complexity of Feasible Descent Methods for Convex Optimization. Technical report, Department of Computer Science, National Taiwan University,
Taipei, Taiwan, 2013.
[5] A. Agarwal, S. Negahban, and M. Wainwright. Fast Global Convergence Rates of Gradient
Methods for High-Dimensional Statistical Recovery. In NIPS 2010.
[6] K. Hou, Z. Zhou, A. M.-S. So, and Z.-Q. Luo, On the linear convergence of the proximal gradient method for trace norm regularization, in Neural Information Processing Systems (NIPS),
2013.
[7] L. Xiao and T. Zhang, A proximal-gradient homotopy method for the l1-regularized leastsquares problem, in ICML, 2012.
[8] P. Tseng and S. Yun, A coordinate gradient descent method for nonsmooth separable minimization, Math. Prog. B. 117 (2009).
[9] G.-X. Yuan, C.-H. Ho, and C.-J. Lin, An improved GLMNET for l1-regularized logistic regression, Journal of Machine Learning Research, vol. 13, pp. 19992030, 2012
[10] R.-E. Fan, K.-W. Chang, C.-J. Hsieh, X.-R. Wang, and C.-J. Lin, LIBLINEAR: A library for
large linear classification, Journal of Machine Learning Research, vol. 9, pp. 1871-1874, 2008.
[11] Alan J Hoffman. On approximate solutions of systems of linear inequalities. Journal of Research of the National Bureau of Standards, 1952.
[12] Tewari, A, Ravikumar, P, and Dhillon, I S. Greedy Algorithms for Structurally Constrained
High Dimensional Problems. In NIPS, 2011.
[13] S. Negahban, P. Ravikumar, M. J. Wainwright, and B. Yu. A unified framework for highdimensional analysis of M-estimators with decomposable regularizers. In NIPS, 2009.
[14] A. Beck and M. Teboulle. A fast iterative shrinkage-thresholding algorithm for linear inverse
problems. SIAM Journal on Imaging Sciences, 2(1):183?202, 2009.
[15] S. Becker, J. Bobin, and E.J.Candes. Nesta: a fast and accurate first-order method for sparse
recovery. SIAM Journal on Imaging Sciences, 2011.
[16] Z. Q. Luo and P. Tseng. Error bounds and convergence analysis of feasible descent methods: a
general approach. Annals of Operations Research, 46-47:157?178, 1993.
[17] Rahul Garg and Rohit Khandekar. Gradient Descent with Sparsification: an iterative algorithm
for sparse recovery with restricted isometry property. In ICML 2009.
[18] S. Ji and J. Ye. An accelerated gradient method for trace norm minimization. In ICML, 2009.
[19] Y. E. Nesterov, Gradient Methods for Minimizing Composite Objective Function, CORE report, 2007.
[20] Y. Nesterov. Introductory Lectures on Convex Optimization. Kluwer Academic Publishers,
New York, 2004
[21] K. Scheinberg, X. Tang. Practical Inexact Proximal Quasi-Newton Method with Global Complexity Analysis. COR@L Technical Report at Lehigh University. arXiv:1311.6547, 2013.
9
| 5257 |@word version:1 stronger:2 norm:20 d2:4 hsieh:3 covariance:2 liblinear:1 nesta:1 document:1 ati:3 past:1 nt:4 luo:2 yet:1 written:3 must:1 hou:1 numerical:2 happen:1 update:6 greedy:1 core:1 caveat:1 math:1 zhang:1 along:1 qualitative:1 prove:4 yuan:1 introductory:1 polyhedral:3 introduce:4 manner:1 bobin:1 x0:3 indeed:1 behavior:3 growing:1 decomposed:1 considering:1 increasing:1 becomes:1 moreover:1 notation:1 bounded:4 developed:1 unified:1 sparsification:1 guarantee:4 scaled:1 grant:1 before:1 positive:3 engineering:1 local:5 understood:1 treat:1 consequence:2 meet:1 path:1 black:1 might:2 twice:2 garg:1 range:1 statistically:1 practical:2 unique:5 definite:1 lf:2 procedure:2 attain:1 composite:4 projection:1 boyd:1 refers:1 jui:1 get:1 onto:1 close:1 superlinear:1 operator:3 context:1 applying:3 go:1 convex:23 formulate:1 decomposable:9 simplicity:1 recovery:3 m2:1 estimator:9 vandenberghe:1 proving:1 coordinate:5 annals:1 suppose:4 satisfying:5 observed:2 subproblem:2 wang:2 pradeepr:1 sun:1 decrease:1 intuition:1 pd:1 convexity:20 complexity:2 nesterov:2 solving:1 univ:1 fast:4 tell:1 saunders:1 larger:1 solve:4 otherwise:1 final:1 sequence:1 eigenvalue:2 differentiable:2 aro:1 combining:1 iff:5 achieve:2 convergence:36 double:2 optimum:4 diverges:1 incidentally:1 leave:1 spent:1 strong:19 solves:1 c:1 come:1 implies:1 direction:1 dii:1 explains:1 require:4 hx:1 suffices:1 preliminary:1 homotopy:1 tighter:1 leastsquares:1 strictly:1 hold:9 mm:2 sufficiently:1 exp:1 achieves:2 early:1 uniqueness:1 estimation:9 utexas:1 tool:1 hoffman:3 minimization:2 clearly:1 always:1 modified:1 zhou:1 shrinkage:1 ax:4 focus:2 rank:2 polyhedron:2 sense:1 typically:5 integrated:1 quasi:1 comprising:1 interested:1 arg:6 dual:2 classification:2 art:2 special:1 constrained:1 equal:3 sampling:2 yu:3 icml:3 t2:1 report:3 nonsmooth:1 few:1 employ:2 modern:1 simultaneously:1 national:2 individual:1 beck:1 phase:7 psd:3 interest:1 investigate:2 pradeep:1 regularizers:2 amenable:1 accurate:1 lh:6 orthogonal:3 conduct:1 desired:1 re:1 instance:1 earlier:1 facet:3 teboulle:1 corroborate:2 w911nf:1 subset:1 motivating:1 proximal:42 considerably:1 cho:1 gd:10 randomized:1 negahban:2 siam:2 lee:1 satisfied:1 possibly:2 derivative:3 prox:28 includes:1 satisfy:6 depends:2 analyze:2 candes:1 cjhsieh:1 yen:1 square:1 efficiently:1 yield:1 generalize:1 produced:1 mere:1 minj:1 definition:5 inexact:1 pp:2 dm:1 naturally:2 proof:7 mi:3 ianyen:1 recall:1 organized:1 cj:12 actually:1 proxh:8 appears:1 response:1 improved:1 rahul:1 box:1 strongly:7 furthermore:1 just:1 stage:1 d:1 hand:1 continuity:1 logistic:8 aj:14 facilitate:1 ye:1 concept:1 verify:1 true:3 ccf:2 hence:1 regularization:4 equality:2 dhillon:3 noted:1 suboptimality:1 generalized:2 yun:1 xds:1 tt:1 l1:2 cnsc:22 consideration:1 novel:2 multinomial:1 ji:1 kluwer:1 refer:1 cambridge:2 ai:2 rd:3 vanilla:1 consistency:1 etc:1 closest:1 isometry:2 belongs:2 inequality:8 yi:3 seen:2 minimum:1 additional:3 impose:2 employed:2 monotonically:1 ii:3 multiple:1 smooth:11 technical:2 faster:3 academic:1 alan:1 long:5 lin:3 ravikumar:4 a1:1 variant:3 regression:6 poisson:1 arxiv:1 iteration:6 agarwal:1 fellowship:1 publisher:1 unlike:1 strict:1 deficient:1 obj:2 call:1 structural:1 leverage:3 easy:2 enough:1 iterate:2 lasso:2 identified:1 reduce:1 idea:1 texas:1 absent:1 motivated:2 becker:1 effort:1 penalty:1 algebraic:2 hessian:8 york:1 tewari:1 clear:1 amount:2 exist:1 nsf:2 arising:1 vol:2 group:4 key:3 iter:2 neither:1 ht:17 imaging:2 year:1 inverse:3 prog:1 reader:1 appendix:1 comparable:1 bound:11 followed:1 fan:1 quadratic:14 constraint:3 infinity:1 x2:12 argument:2 min:9 optimality:2 span:1 rcv1:1 separable:1 department:2 according:1 restricted:5 scheinberg:1 serf:1 sustik:1 cor:1 operation:3 observe:1 ho:1 bureau:1 spurred:1 hinge:2 newton:26 exploit:1 taipei:1 establish:1 objective:5 arrangement:1 occurs:2 diagonal:1 gradient:21 subspace:4 distance:6 tseng:2 reason:1 khandekar:1 taiwan:2 minimizing:2 potentially:1 statement:2 trace:2 negative:1 stated:1 xnt:11 design:1 av:1 finite:3 descent:9 defining:1 z1:2 nip:6 beyond:1 usually:1 regime:2 sparsity:1 program:3 including:1 max:2 wainwright:2 regularized:8 library:1 ready:1 acknowledges:2 review:1 acknowledgement:1 xtt:1 rohit:1 asymptotic:1 relative:2 loss:12 lecture:1 sublinear:3 sufficient:1 xiao:1 thresholding:1 ibm:1 austin:1 supported:1 last:1 infeasible:2 weaker:1 expansiveness:1 wide:1 fall:1 sparse:4 dimension:1 lipschitzcontinuous:2 made:1 qualitatively:1 approximate:1 ml:2 global:6 active:1 assumed:2 continuous:2 search:2 iterative:3 nature:1 mj:22 main:1 profile:1 subd:1 x1:11 augmented:1 precision:1 structurally:1 comprises:2 lehigh:1 lie:1 nullspace:8 ian:1 tang:1 theorem:7 xt:101 r2:3 adding:1 mirror:1 phd:1 glmnet:1 inderjit:2 chang:1 corresponds:2 satisfies:7 lipschitz:5 absence:2 feasible:3 specifically:1 determined:1 contrasted:1 lemma:19 highdimensional:1 support:2 latter:1 accelerated:1 ex:2 |
4,702 | 5,258 | SAGA: A Fast Incremental Gradient Method With
Support for Non-Strongly Convex Composite
Objectives
Francis Bach
INRIA - Sierra Project-Team
?
Ecole
Normale Sup?erieure, Paris, France
Aaron Defazio
Ambiata ?
Australian National University, Canberra
Simon Lacoste-Julien
INRIA - Sierra Project-Team
?
Ecole
Normale Sup?erieure, Paris, France
Abstract
In this work we introduce a new optimisation method called SAGA in the spirit of
SAG, SDCA, MISO and SVRG, a set of recently proposed incremental gradient
algorithms with fast linear convergence rates. SAGA improves on the theory behind SAG and SVRG, with better theoretical convergence rates, and has support
for composite objectives where a proximal operator is used on the regulariser. Unlike SDCA, SAGA supports non-strongly convex problems directly, and is adaptive to any inherent strong convexity of the problem. We give experimental results
showing the effectiveness of our method.
1
Introduction
Remarkably, recent advances [1, 2] have shown that it is possible to minimise strongly convex
finite sums provably faster in expectation than is possible without the finite sum structure. This is
significant for machine learning problems as a finite sum structure is common in the empirical risk
minimisation setting. The requirement of strong convexity is likewise satisfied in machine learning
problems in the typical case where a quadratic regulariser is used.
In particular, we are interested in minimising functions of the form
n
1X
f (x) =
fi (x),
n i=1
where x ? Rd , each fi is convex and has Lipschitz continuous derivatives with constant L. We will
also consider the case where each fi is strongly convex with constant ?, and the ?composite? (or
proximal) case where an additional regularisation function is added:
F (x) = f (x) + h(x),
where h : R ? R is convex but potentially non-differentiable, and where the proximal operation
of h is easy to compute ? few incremental gradient methods are applicable in this setting [3][4].
d
d
Our contributions are as follows. In Section 2 we describe the SAGA algorithm, a novel incremental
gradient method. In Section 5 we prove theoretical convergence rates for SAGA in the strongly
convex case better than those for SAG [1] and SVRG [5], and a factor of 2 from the SDCA [2]
convergence rates. These rates also hold in the composite setting. Additionally, we show that
?
The first author completed this work while under funding from NICTA. This work was partially supported
by the MSR-Inria Joint Centre and a grant by the European Research Council (SIERRA project 239993).
1
like SAG but unlike SDCA, our method is applicable to non-strongly convex problems without
modification. We establish theoretical convergence rates for this case also. In Section 3 we discuss
the relation between each of the fast incremental gradient methods, showing that each stems from a
very small modification of another.
2
SAGA Algorithm
We start with some known initial vector x0 ? Rd and known derivatives fi0 (?0i ) ? Rd with ?0i = x0
for each i. These derivatives are stored in a table data-structure of length n, or alternatively a n ? d
matrix. For many problems of interest, such as binary classification and least-squares, only a single
floating point value instead of a full gradient vector needs to be stored (see Section 4). SAGA is
inspired both from SAG [1] and SVRG [5] (as we will discuss in Section 3). SAGA uses a step size
of ? and makes the following updates, starting with k = 0:
SAGA Algorithm: Given the value of xk and of each fi0 (?ki ) at the end of iteration k, the updates
for iteration k + 1 is as follows:
1. Pick a j uniformly at random.
2. Take ?k+1
= xk , and store fj0 (?k+1
) in the table. All other entries in the table remain
j
j
k+1
unchanged. The quantity ?j is not explicitly stored.
3. Update x using fj0 (?k+1
), fj0 (?kj ) and the table average:
j
"
w
k+1
k
=x ??
fj0 (?k+1
)
j
?
fj0 (?kj )
#
n
1X 0 k
+
f (? ) ,
n i=1 i i
xk+1 = proxh? wk+1 .
(1)
(2)
The proximal operator we use above is defined as
1
2
h
kx ? yk .
(3)
prox? (y) := argmin h(x) +
2?
x?Rd
In the strongly convex case, when a step size of ? = 1/(2(?n+L)) is chosen, we have the following
convergence rate in the composite and hence also the non-composite case:
2
E
xk ? x?
? 1 ?
?
2(?n + L)
k
0
x ? x?
2 +
n
f (x0 ) ? f 0 (x? ), x0 ? x? ? f (x? ) .
?n + L
We prove this result in Section 5. The requirement of strong convexity can be relaxed from needing
to hold for each fi to just holding on average, but at the expense of a worse geometric rate (1 ?
?
6(?n+L) ), requiring a step size of ? = 1/(3(?n + L)).
In the non-strongly convex case, we have established the convergence rate in terms of the average
Pk
iterate, excluding step 0: x
?k = k1 t=1 xt . Using a step size of ? = 1/(3L) we have
4n 2L
x0 ? x?
2 + f (x0 ) ? f 0 (x? ), x0 ? x? ? f (x? ) .
E F (?
xk ) ? F (x? ) ?
k
n
This result is proved in the supplementary material. Importantly, when this step size ? = 1/(3L) is
used, our algorithm automatically adapts to the level of strong convexity ? > 0 naturally present,
giving a convergence rate of (see the comment at the end of the proof of Theorem 1):
k
2
0
?
1
x ? x?
2 + 2n f (x0 ) ? f 0 (x? ), x0 ? x? ? f (x? ) .
,
E
xk ? x?
? 1 ? min
4n 3L
3L
Although any incremental gradient method can be applied to non-strongly convex problems via the
addition of a small quadratic regularisation, the amount of regularisation is an additional tunable
parameter which our method avoids.
3
Related Work
We explore the relationship between SAGA and the other fast incremental gradient methods in this
section. By using SAGA as a midpoint, we are able to provide a more unified view than is available
in the existing literature. A brief summary of the properties of each method considered in this section
is given in Figure 1. The method from [3], which handles the non-composite setting, is not listed as
its rate is of the slow type and can be up to n times smaller than the one for SAGA or SVRG [5].
2
Strongly Convex (SC)
Convex, Non-SC*
Prox Reg.
Non-smooth
Low Storage Cost
Simple(-ish) Proof
Adaptive to SC
SAGA
3
3
3
7
7
3
3
SAG
3
3
?
7
7
7
3
SDCA
3
7
3[6]
3
7
3
7
SVRG
3
?
3
7
3
3
?
FINITO
3
?
7
7
7
3
?
Figure 1: Basic summary of method properties. Question marks denote unproven, but not experimentally
ruled out cases. (*) Note that any method can be applied to non-strongly convex problems by adding a small
amount of L2 regularisation, this row describes methods that do not require this trick.
SAGA: midpoint between SAG and SVRG/S2GD
In [5], the authors make the observation that the variance of the standard stochastic gradient (SGD)
update direction can only go to zero if decreasing step sizes are used, thus preventing a linear convergence rate unlike for batch gradient descent. They thus propose to use a variance reduction approach
(see [7] and references therein for example) on the SGD update in order to be able to use constant
step sizes and get a linear convergence rate. We present the updates of their method called SVRG
(Stochastic Variance Reduced Gradient) in (6) below, comparing it with the non-composite form
of SAGA rewritten in (5). They also mention that SAG (Stochastic Average Gradient) [1] can be
interpreted as reducing the variance, though they do not provide the specifics. Here, we make this
connection clearer and relate it to SAGA.
We first review a slightly more generalized version of the variance reduction approach (we allow the
updates to be biased). Suppose that we want to use Monte Carlo samples to estimate EX and that
we can compute efficiently EY for another random variable Y that is highly correlated with X. One
variance reduction approach is to use the following estimator ?? as an approximation to EX: ?? :=
?(X ?Y )+EY , for a step size ? ? [0, 1]. We have that E?? is a convex combination of EX and EY :
E?? = ?EX + (1 ? ?)EY . The standard variance reduction approach uses ? = 1 and the estimate
is unbiased E?1 = EX. The variance of ?? is: Var(?? ) = ?2 [Var(X) + Var(Y ) ? 2 Cov(X, Y )],
and so if Cov(X, Y ) is big enough, the variance of ?? is reduced compared to X, giving the method
its name. By varying ? from 0 to 1, we increase the variance of ?? towards its maximum value
(which usually is still smaller than the one for X) while decreasing its bias towards zero.
Both SAGA and SAG can be derived from such a variance reduction viewpoint: here X is the SGD
direction sample fj0 (xk ), whereas Y is a past stored gradient fj0 (?kj ). SAG is obtained by using
? = 1/n (update rewritten in our notation in (4)), whereas SAGA is the unbiased version with ? = 1
(see (5) below). For the same ??s, the variance of the SAG update is 1/n2 times the one of SAGA,
but at the expense of having a non-zero bias. This non-zero bias might explain the complexity of
the convergence proof of SAG and why the theory has not yet been extended to proximal operators.
By using an unbiased update in SAGA, we are able to obtain a simple and tight theory, with better
constants than SAG, as well as theoretical rates for the use of proximal operators.
"
#
n
fj0 (xk ) ? fj0 (?kj )
1X 0 k
k+1
k
(SAG)
x
=x ??
+
f (? ) ,
(4)
n
n i=1 i i
"
#
n
X
1
(SAGA)
xk+1 = xk ? ? fj0 (xk ) ? fj0 (?kj ) +
f 0 (?k ) ,
(5)
n i=1 i i
"
#
n
1X 0
k+1
k
0 k
0
f (?
x) .
(6)
(SVRG)
x
= x ? ? fj (x ) ? fj (?
x) +
n i=1 i
The SVRG update (6) is obtained by using Y = fj0 (?
x) with ? = 1 (and is thus unbiased ? we note
that SAG is the only method that we present in the related work that has a biased update direction).
The vector x
? is not updated every step, but rather the loop over k appears inside an outer loop, where
x
? is updated at the start of each outer iteration. Essentially SAGA is at the midpoint between SVRG
and SAG; it updates the ?j value each time index j is picked, whereas SVRG updates all of ??s as
a batch. The S2GD method [8] has the same update as SVRG, just differing in how the number of
inner loop iterations is chosen. We use SVRG henceforth to refer to both methods.
3
SVRG makes a trade-off between time and space. For the equivalent practical convergence rate it
makes 2x-3x more gradient evaluations, but in doing so it does not need to store a table of gradients,
but a single average gradient. The usage of SAG vs. SVRG is problem dependent. For example for
linear predictors where gradients can be stored as a reduced vector of dimension p ? 1 for p classes,
SAGA is preferred over SVRG both theoretically and in practice. For neural networks, where no
theory is available for either method, the storage of gradients is generally more expensive than the
additional backpropagations, but this is computer architecture dependent.
SVRG also has an additional parameter besides step size that needs to be set, namely the number of
iterations per inner loop (m). This parameter can be set via the theory, or conservatively as m = n,
however doing so does not give anywhere near the best practical performance. Having to tune one
parameter instead of two is a practical advantage for SAGA.
Finito/MISO?
To make the relationship with other prior methods more apparent, we can rewrite the SAGA
algorithm (in P
the non-composite case) in term of an additional intermediate quantity uk , with
n
u0 := x0 + ? i=1 fi0 (x0 ), in addition to the usual xk iterate as described previously:
SAGA: Equivalent reformulation for non-composite case: Given the value of uk and of each
fi0 (?ki ) at the end of iteration k, the updates for iteration k + 1, is as follows:
n
X
1. Calculate xk :
xk = uk ? ?
fi0 (?ki ).
(7)
i=1
2. Update u with uk+1 = uk +
1
k
n (x
3. Pick a j uniformly at random.
? uk ).
4. Take ?k+1
= xk , and store fj0 (?k+1
) in the table replacing fj0 (?kj ). All other entries in
j
j
the table remain unchanged. The quantity ?k+1
is not explicitly stored.
j
Eliminating uk recovers the update (5) for xk . We now describe how the Finito [9] and MISO? [10]
methods are closely related to SAGA. Both Finito and MISO? use updates of the following form,
for a step length ?:
n
X
1X k
xk+1 =
?i ? ?
fi0 (?ki ).
(8)
n i
i=1
The step size used is of the order
P of 1/?n. To simplify the discussion of this algorithm we will
introduce the notation ?? = n1 i ?ki .
SAGA can be interpreted as Finito, but with the quantity ?? replaced with u, which is updated in the
? but in expectation. To see this, consider how ?? changes in expectation:
same way as ?,
1 k
1 k ?k
E ??k+1 = E ??k +
x ? ?kj = ??k +
x ?? .
n
n
The update is identical in expectation to the update for u, uk+1 = uk + n1 (xk ? uk ). There are
three advantages of SAGA over Finito/MISO?. SAGA does not require strong convexity to work,
it has support for proximal operators, and it does not require storing the ?i values. MISO has
proven support for proximal operators only in the case where impractically small step sizes are
used [10]. The big advantage of Finito/MISO? is that when using a per-pass re-permuted access
ordering, empirical speed-ups of up-to a factor of 2x has been observed. This access order can also
be used with the other methods discussed, but with smaller empirical speed-ups. Finito/MISO? is
particularly useful when fi is computationally expensive to compute compared to the extra storage
costs required over the other methods.
SDCA
The Stochastic Dual Coordinate Descent (SDCA) [2] method on the surface appears quite different
from the other methods considered. It works with the convex conjugates of the fi functions. However, in this section we show a novel transformation of SDCA into an equivalent method that only
works with primal quantities, and is closely related to the MISO? method.
4
Consider the following algorithm:
SDCA algorithm in the primal
Step k + 1:
1. Pick an index j uniformly at random.
f
2. Compute ?k+1
= prox?j (z), where ? =
j
Pn
3.
at location j. For i 6= j, the
1
?n and z = ??
Store the gradient fj0 (?k+1
) = ?1 z ? ?k+1
in the table
j
j
0 k+1
0 k
table entries are unchanged (fi (?i ) = fi (?i )).
At completion, return xk = ??
Pn
i
i6=j
fi0 (?ki ).
fi0 (?ki ) .
We claim that this algorithm is equivalent to the version of SDCA where exact block-coordinate
maximisation is used on the dual.1 Firstly, note that while SDCA was originally described for onedimensional outputs (binary classification or regression), it has been expanded to cover the multiclass predictor case [11] (called Prox-SDCA there). In this case, the primal objective has a separate
strongly convex regulariser, and the functions fi are restricted to the form fi (x) := ?i (XiT x), where
Xi is a d?p feature matrix, and ?i is the loss function that takes a p dimensional input, for p classes.
To stay in the same general setting as the other incremental gradient methods, we work directly with
the fi (x) functions rather than the more structured ?i (XiT x). The dual objective to maximise then
becomes
?
?
2
n
n
1 X
X
?
1
D(?) = ??
?i
?
f ? (??i )? ,
2
?n i=1
n i=1 i
where ?i ?s are d-dimensional dual variables. Generalising the exact block-coordinate maximisation
update that SDCA performs to this form, we get the dual update for block j (with xk the current
primal iterate):
(
2 )
?n
k
1
k+1
?
k
k
x +
?j = ?j + argmax ?fj ??j ? ??j ?
??j
(9)
.
2
?n
?aj ?Rd
In the special case where fi (x) = ?i (XiT x), we can see that (9) gives exactly the same update as
Option I of Prox-SDCA in [11, Figure 1], which operates instead on the equivalent p-dimensional
dual variables ?
? i with the relationship that ?i = Xi ?
? i .2 As noted by Shalev-Shwartz & Zhang [11],
the update (9) is actually an instance of the proximal operator of the convex conjugate of fj . Our
primal formulation exploits this fact by using a relation between the proximal operator of a function
and its convex conjugate known as the Moreau decomposition:
?
proxf (v) = v ? proxf (v).
This decomposition allows us to compute the proximal operator of the conjugate via the primal
proximal operator. As this is the only use in the basic SDCA method of the conjugate function,
applying this decomposition allows us to completely eliminate the ?dual? aspect of the algorithm,
yielding the above primal form of SDCA. The dual variables are related to the primal representa0
tives ?i ?s through
P ?i = ?fi (?i ). The KKT conditions ensure that if the ?i values are dual optimal
k
then x = ? i ?i as defined above is primal optimal. The same trick is commonly used to interpret Dijkstra?s set intersection as a primal algorithm instead of a dual block coordinate descent
algorithm [12].
The primal form of SDCA differs from the other incremental gradient methods described in this
section in that it assumes strong convexity is induced by a separate strongly convex regulariser,
rather than each fi being strongly convex. In fact, SDCA can be modified to work without a separate
regulariser, giving a method that is at the midpoint between Finito and SDCA. We detail such a
method in the supplementary material.
1
More precisely, to Option I of Prox-SDCA as described in [11, Figure 1]. We will simply refer to this
method as ?SDCA? in this paper for brevity.
2
This is because fi? (?i ) =
inf
?i? (?
? i ).
?
? i s.t. ?i =Xi ?
?i
5
SDCA variants
The SDCA theory has been expanded to cover a number of other methods of performing the coordinate step [11]. These variants replace the proximal operation in our primal interpretation in the
previous section with an update where ?k+1
is chosen so that: fj0 (?k+1
) = (1??)fj0 (?kj )+?fj0 (xk ),
j
j
P
k+1
1
0 k
does
where xk = ? ?n
i fi (?i ). The variants differ in how ? ? [0, 1] is chosen. Note that ?j
k+1
not actually have to be explicitly known, just the gradient fj0 (?j ), which is the result of the above
interpolation. Variant 5 by Shalev-Shwartz & Zhang [11] does not require operations on the conju?n
gate function, it simply uses ? = L+?n
. The most practical variant performs a line search involving
the convex conjugate to determine ?. As far as we are aware, there is no simple primal equivalent
of this line search. So in cases where we can not compute the proximal operator from the standard
SDCA variant, we can either introduce a tuneable parameter into the algorithm (?), or use a dual
line search, which requires an efficient way to evaluate the convex conjugates of each fi .
4
Implementation
We briefly discuss some implementation concerns:
? For many problems each derivative fi0 is just a simple weighting of the ith data vector.
Logistic regression and least squares have this property. In such cases, instead of storing
the full derivative fi0 for each i, we need only to store the weighting constants. This reduces
the storage requirements to be the same as the SDCA method in practice. A similar trick
can be applied to multi-class classifiers with p classes by storing p ? 1 values for each i.
? Our algorithm assumes that initial gradients are known for each fi at the starting point x0 .
Instead, a heuristic may be used where during the first pass, data-points are introduced oneby-one, in a non-randomized order, with averages computed in terms of those data-points
processed so far. This procedure has been successfully used with SAG [1].
? The SAGA update as stated is slower than necessary when derivatives are sparse. A just-intime updating of u or x may be performed just as is suggested for SAG [1], which ensures
that only sparse updates are done at each iteration.
? We give the form of SAGA for the case where each fi is strongly convex. However in
practice we usually have only convex fi , with strong convexity in f induced by the addition
of a quadratic regulariser. This quadratic regulariser may be split amongst the fi functions
evenly, to satisfy our assumptions. It is perhaps easier to use a variant of SAGA where the
regulariser ?2 ||x||2 is explicit, such as the following modification of Equation (5):
#
"
1X 0 k
k+1
k
0 k
0
k
f (? ) .
x
= (1 ? ??) x ? ? fj (x ) ? fj (?j ) +
n i i i
For sparse implementations instead of scaling xk at each step, a separate scaling constant
? k may be scaled instead, with ? k xk being used in place of xk . This is a standard trick
used with stochastic gradient methods.
For sparse problems with a quadratic regulariser the just-in-time updating can be a little intricate. In
the supplementary material we provide example python code showing a correct implementation that
uses each of the above tricks.
5
Theory
In this section, all expectations are taken with respect to the choice of j at iteration k + 1 and
conditioned on xk and each fi0 (?ki ) unless stated otherwise.
We start with two basic lemmas that just state properties of convex functions, followed by Lemma 1,
which is specific to our algorithm. The proofs of each of these lemmas is in the supplementary
material.
Pn
Lemma 1. Let f (x) = n1 i=1 fi (x). Suppose each fi is ?-strongly convex and has Lipschitz
continuous gradients with constant L. Then for all x and x? :
L??
?
2
hf 0 (x), x? ? xi ?
[f (x? ) ? f (x)] ? kx? ? xk
L
2
6
?
1 X 0 ?
?
2
kfi (x ) ? fi0 (x)k ? hf 0 (x? ), x ? x? i .
2Ln i
L
Lemma 2. We have that for all ?i and x? :
"
#
1X 0
1X
1X 0 ?
0 ? 2
?
?
kfi (?i ) ? fi (x )k ? 2L
fi (?i ) ? f (x ) ?
hfi (x ), ?i ? x i .
n i
n i
n i
Lemma 3. It holds that for any ?ki , x? , xk and ? > 0, with wk+1 as defined in Equation 1:
2
2
2
E
wk+1 ? xk ? ?f 0 (x? )
? ? 2 (1 + ? ?1 )E
fj0 (?kj ) ? fj0 (x? )
+ ? 2 (1 + ?)E
fj0 (xk ) ? fj0 (x? )
2
? ? 2 ?
f 0 (xk ) ? f 0 (x? )
.
Theorem 1. With x? the optimal solution, define the Lyapunov function T as:
2
1X
1 X
0 ? k
T k := T (xk , {?ki }ni=1 ) :=
fi (?ki ) ? f (x? ) ?
fi (x ), ?i ? x? + c
xk ? x?
.
n i
n i
1
1
1
Then with ? = 2(?n+L)
, c = 2?(1???)n
, and ? = ??
, we have the following expected change in the
Lyapunov function between steps of the SAGA algorithm (conditional on T k ):
1
E[T k+1 ] ? (1 ? )T k .
?
Proof. The first three terms in T k+1 are straight-forward to simplify:
#
"
1
1 1X
1X
k+1
k
fi (?i ) = f (x ) + 1 ?
fi (?ki ).
E
n i
n
n n i
"
#
1 X
0 ? k+1
1
1 1 X
0 ? k
?
E ?
fi (x ), ?i ? x
= ? f 0 (x? ), xk ? x? ? 1?
fi (x ), ?i ? x? .
n i
n
n n i
For the change in the last term of T k+1 , we apply the non-expansiveness of the proximal operator3 :
2
2
c
xk+1 ? x?
= c
prox? (wk+1 ) ? prox? (x? ? ?f 0 (x? ))
2
? c
wk+1 ? x? + ?f 0 (x? )
.
We expand the quadratic and apply E[wk+1 ] = xk ? ?f 0 (xk ) to simplify the inner product term:
2
2
cE
wk+1 ? x? + ?f 0 (x? )
= cE
xk ? x? + wk+1 ? xk + ?f 0 (x? )
2
2
= c
xk ? x?
+ 2cE wk+1 ? xk + ?f 0 (x? ), xk ? x? + cE
wk+1 ? xk + ?f 0 (x? )
2
2
= c
xk ? x?
? 2c? f 0 (xk ) ? f 0 (x? ), xk ? x? + cE
wk+1 ? xk + ?f 0 (x? )
2
2
? c
xk ? x?
? 2c? f 0 (xk ), xk ? x? + 2c? f 0 (x? ), xk ? x? ? c? 2 ?
f 0 (xk ) ? f 0 (x? )
2
2
+ 1 + ? ?1 c? 2 E
fj0 (?kj ) ? fj0 (x? )
+ (1 + ?) c? 2 E
fj0 (xk ) ? fj0 (x? )
. (Lemma 3)
The value of ? shall be fixed later. Now we apply Lemma 1 to bound ?2c? f 0 (xk ), xk ? x? and
2
Lemma 2 to bound E
fj0 (?kj ) ? fj0 (x? )
:
2
2
2
c?
E
fj0 (xk ) ? fj0 (x? )
cE
xk+1 ? x?
? (c ? c??)
xk ? x?
+ (1 + ?)c? 2 ?
L
2
0 ? k
2c?(L ? ?)
k
?
?
f (x ) ? f (x ) ? f (x ), x ? x? ? c? 2 ?
f 0 (xk ) ? f 0 (x? )
L
"
#
X
X
1
1
fi (?ki ) ? f (x? ) ?
+ 2 1 + ? ?1 c? 2 L
fi0 (x? ), ?ki ? x? .
n i
n i
3
Note that the first equality below is the only place in the proof where we use the fact that x? is an optimality
point.
7
Function sub-optimality
100
10?4
10?4
10?8
10?8
10?8
10?8
10?12
10?12
10?12
10?12
10?4
5
10
15
20
5
10
15
20
3 ? 10?2
10?1
10?4
5
10
15
20
102
101
5
10
5
10
15
20
15
20
100
100
10?1
10?2
2 ? 10?2
10?2
5
10
15
20
Finito perm
5
10
15
20
10?1
5
Gradient evaluations / n
Finito
SAGA
SVRG
10
15
SAG
20
SDCA
LBFGS
0
4
128
101016
10
Figure 2: From left to right we have the MNIST, COVTYPE, IJCNN1 and MILLIONSONG datasets. Top
row is the L2 regularised case, bottom row the L1 regularised case.
We can now combine the bounds that we have derived for each term in T , and pull out a frac
2
tion ?1 of T k (for any ?
at this point). Together with the inequality ?
f 0 (xk ) ? f 0 (x? )
?
?2? f (xk ) ? f (x? ) ? f 0 (x? ), xk ? x? [13, Thm. 2.1.10], that yields:
h
D
Ei
2c?(L ? ?)
1
1
E[T k+1 ] ? T k ? ? T k +
?
? 2c? 2 ?? f (xk ) ? f (x? ) ? f 0 (x? ), xk ? x?
?
n
L
#
" X
E
1
1
1 XD 0 ?
1
?1
2
k
?
k
?
+ 2(1 + ? )c? L ?
fi (?i ) ? f (x ) ?
fi (x ), ?i ? x
+
?
n
n i
n i
2
2
1
1
+
? ?? c
xk ? x?
+ (1 + ?)? ?
c?E
fj0 (xk ) ? fj0 (x? )
.
(10)
?
L
Note that each of the terms in square brackets are positive, and it can be readily verified that our
1
1
1
, c = 2?(1???)n
, and ? = ??
), together with
assumed values for the constants (? = 2(?n+L)
? = 2?n+L
ensure that each of the quantities in round brackets are non-positive (the constants were
L
determined by setting all the round brackets to zero except the second one ? see [14] for the details).
1
Adaptivity to strong convexity result: Note
when
that
using the ? = 3L step size, the same c as
?
1
1
above can be used with ? = 2 and ? = min 4n , 3L to ensure non-positive terms.
2
Corollary 1. Note that c
xk ? x?
? T k , and therefore by chaining the expectations, plugging
in the constants explicitly and using ?(n ? 0.5) ? ?n to simplify the expression, we get:
2
E
xk ? x?
? 1 ?
?
2(?n + L)
k
0
x ? x?
2 +
n
f (x0 ) ? f 0 (x? ), x0 ? x? ? f (x? ) .
?n + L
Here the expectation is over all choices of index j k up to step k.
6
Experiments
We performed a series of experiments to validate the effectiveness of SAGA. We tested a binary
classifier on MNIST, COVTYPE, IJCNN1 and a least squares predictor on MILLIONSONG. Details
of these datasets can be found in [9]. We used the same code base for each method, just changing the
main update rule. SVRG was tested with the recalibration pass used every n iterations, as suggested
in [8]. Each method had its step size parameter chosen so as to give the fastest convergence.
We tested with a L2 regulariser, which all methods support, and with a L1 regulariser on a subset
of the methods. The results are shown in Figure 2. We can see that Finito (perm) performs the
best on a per epoch equivalent basis, but it can be the most expensive method per step. SVRG is
similarly fast on a per epoch basis, but when considering the number of gradient evaluations per
epoch is double that of the other methods for this problem, it is middle of the pack. SAGA can be
seen to perform similar to the non-permuted Finito case, and to SDCA. Note that SAG is slower
than the other methods at the beginning. To get the optimal results for SAG, an adaptive step size
rule needs to be used rather than the constant step size we used. In general, these tests confirm that
the choice of methods should be done based on their properties as discussed in Section 3, rather than
their convergence rate.
8
1515
20
0
References
[1] Mark Schmidt, Nicolas Le Roux, and Francis Bach. Minimizing finite sums with the stochastic
average gradient. Technical report, INRIA, hal-0086005, 2013.
[2] Shai Shalev-Shwartz and Tong Zhang. Stochastic dual coordinate ascent methods for regularized loss minimization. JMLR, 14:567?599, 2013.
[3] Paul Tseng and Sangwoon Yun. Incrementally updated gradient methods for constrained and
regularized optimization. Journal of Optimization Theory and Applications, 160:832:853,
2014.
[4] Lin Xiao and Tong Zhang. A proximal stochastic gradient method with progressive variance reduction. Technical report, Microsoft Research, Redmond and Rutgers University, Piscataway,
NJ, 2014.
[5] Rie Johnson and Tong Zhang. Accelerating stochastic gradient descent using predictive variance reduction. NIPS, 2013.
[6] Taiji Suzuki. Stochastic dual coordinate ascent with alternating direction method of multipliers.
Proceedings of The 31st International Conference on Machine Learning, 2014.
[7] Evan Greensmith, Peter L. Bartlett, and Jonathan Baxter. Variance reduction techniques for
gradient estimates in reinforcement learning. JMLR, 5:1471?1530, 2004.
[8] Jakub Kone?cn?y and Peter Richt?arik. Semi-stochastic gradient descent methods. ArXiv e-prints,
arXiv:1312.1666, December 2013.
[9] Aaron Defazio, Tiberio Caetano, and Justin Domke. Finito: A faster, permutable incremental
gradient method for big data problems. Proceedings of the 31st International Conference on
Machine Learning, 2014.
[10] Julien Mairal. Incremental majorization-minimization optimization with application to largescale machine learning. Technical report, INRIA Grenoble Rh?one-Alpes / LJK Laboratoire
Jean Kuntzmann, 2014.
[11] Shai Shalev-Shwartz and Tong Zhang. Accelerated proximal stochastic dual coordinate ascent
for regularized loss minimization. Technical report, The Hebrew University, Jerusalem and
Rutgers University, NJ, USA, 2013.
[12] Patrick Combettes and Jean-Christophe Pesquet. Proximal Splitting Methods in Signal Processing. In Fixed-Point Algorithms for Inverse Problems in Science and Engineering. Springer,
2011.
[13] Yu. Nesterov. Introductory Lectures On Convex Programming. Springer, 1998.
[14] Aaron Defazio. New Optimization Methods for Machine Learning. PhD thesis, (draft under
examination) Australian National University, 2014. http://www.aarondefazio.com/pubs.html.
9
| 5258 |@word msr:1 briefly:1 version:3 eliminating:1 middle:1 decomposition:3 pick:3 sgd:3 mention:1 reduction:8 initial:2 series:1 pub:1 ecole:2 past:1 existing:1 current:1 comparing:1 com:1 yet:1 readily:1 update:29 v:1 xk:68 beginning:1 ith:1 draft:1 location:1 firstly:1 zhang:6 prove:2 combine:1 introductory:1 inside:1 introduce:3 theoretically:1 x0:14 expected:1 intricate:1 multi:1 inspired:1 decreasing:2 automatically:1 little:1 frac:1 considering:1 becomes:1 project:3 notation:2 permutable:1 argmin:1 interpreted:2 unified:1 differing:1 transformation:1 nj:2 every:2 xd:1 sag:22 exactly:1 classifier:2 scaled:1 uk:10 grant:1 greensmith:1 positive:3 maximise:1 engineering:1 interpolation:1 inria:5 might:1 therein:1 fastest:1 kfi:2 practical:4 practice:3 block:4 maximisation:2 differs:1 procedure:1 sdca:27 evan:1 empirical:3 composite:10 ups:2 get:4 operator:11 storage:4 risk:1 applying:1 www:1 equivalent:7 go:1 jerusalem:1 starting:2 convex:28 roux:1 alpes:1 splitting:1 estimator:1 rule:2 importantly:1 pull:1 handle:1 coordinate:8 updated:4 suppose:2 exact:2 programming:1 us:4 regularised:2 trick:5 expensive:3 particularly:1 updating:2 taiji:1 observed:1 bottom:1 calculate:1 ensures:1 caetano:1 ordering:1 trade:1 richt:1 yk:1 convexity:8 complexity:1 nesterov:1 tight:1 rewrite:1 predictive:1 completely:1 basis:2 joint:1 fast:5 describe:2 monte:1 sc:3 tuneable:1 shalev:4 apparent:1 quite:1 supplementary:4 heuristic:1 jean:2 otherwise:1 cov:2 advantage:3 differentiable:1 propose:1 product:1 loop:4 adapts:1 fi0:13 validate:1 convergence:14 double:1 requirement:3 incremental:11 sierra:3 ish:1 clearer:1 completion:1 strong:8 australian:2 differ:1 direction:4 lyapunov:2 closely:2 correct:1 stochastic:12 material:4 require:4 tiberio:1 hold:3 considered:2 claim:1 applicable:2 miso:9 council:1 successfully:1 minimization:3 arik:1 modified:1 normale:2 rather:5 pn:3 varying:1 minimisation:1 corollary:1 derived:2 xit:3 dependent:2 eliminate:1 relation:2 expand:1 france:2 interested:1 provably:1 classification:2 dual:14 html:1 constrained:1 special:1 aware:1 having:2 identical:1 progressive:1 yu:1 report:4 simplify:4 inherent:1 few:1 grenoble:1 national:2 floating:1 replaced:1 argmax:1 n1:3 microsoft:1 interest:1 highly:1 evaluation:3 bracket:3 yielding:1 kone:1 behind:1 primal:13 fj0:33 necessary:1 unless:1 ruled:1 re:1 theoretical:4 instance:1 cover:2 cost:2 entry:3 subset:1 predictor:3 johnson:1 stored:6 proximal:18 s2gd:2 st:2 international:2 randomized:1 stay:1 off:1 together:2 thesis:1 satisfied:1 henceforth:1 worse:1 derivative:6 return:1 prox:8 wk:11 satisfy:1 explicitly:4 tion:1 performed:2 view:1 picked:1 later:1 doing:2 francis:2 sup:2 start:3 hf:2 option:2 shai:2 simon:1 majorization:1 contribution:1 square:4 ni:1 variance:15 likewise:1 efficiently:1 yield:1 carlo:1 straight:1 explain:1 recalibration:1 naturally:1 proof:6 recovers:1 proved:1 tunable:1 improves:1 actually:2 proxh:1 appears:2 originally:1 rie:1 formulation:1 done:2 though:1 strongly:16 just:9 anywhere:1 replacing:1 ei:1 incrementally:1 logistic:1 aj:1 perhaps:1 hal:1 usage:1 name:1 usa:1 requiring:1 unbiased:4 multiplier:1 hence:1 equality:1 alternating:1 round:2 during:1 noted:1 chaining:1 proxf:2 generalized:1 yun:1 performs:3 l1:2 fj:6 novel:2 recently:1 fi:34 funding:1 common:1 permuted:2 discussed:2 interpretation:1 onedimensional:1 interpret:1 significant:1 refer:2 rd:5 erieure:2 i6:1 similarly:1 centre:1 had:1 access:2 surface:1 base:1 patrick:1 hfi:1 recent:1 inf:1 store:5 inequality:1 binary:3 christophe:1 seen:1 additional:5 relaxed:1 ey:4 determine:1 signal:1 semi:1 u0:1 full:2 needing:1 reduces:1 stem:1 smooth:1 technical:4 faster:2 bach:2 minimising:1 lin:1 plugging:1 variant:7 basic:3 regression:2 involving:1 optimisation:1 expectation:7 essentially:1 rutgers:2 arxiv:2 iteration:10 addition:3 remarkably:1 want:1 whereas:3 laboratoire:1 biased:2 extra:1 unlike:3 ascent:3 comment:1 induced:2 sangwoon:1 december:1 spirit:1 effectiveness:2 near:1 backpropagations:1 intermediate:1 split:1 easy:1 enough:1 baxter:1 iterate:3 conju:1 architecture:1 pesquet:1 inner:3 cn:1 multiclass:1 minimise:1 expression:1 defazio:3 bartlett:1 accelerating:1 peter:2 generally:1 useful:1 listed:1 tune:1 amount:2 processed:1 reduced:3 http:1 per:6 shall:1 reformulation:1 changing:1 ce:6 verified:1 lacoste:1 sum:4 inverse:1 place:2 scaling:2 ki:14 bound:3 followed:1 quadratic:6 precisely:1 aspect:1 speed:2 min:2 optimality:2 performing:1 expanded:2 structured:1 piscataway:1 combination:1 conjugate:7 remain:2 smaller:3 describes:1 slightly:1 perm:2 modification:3 ijcnn1:2 restricted:1 taken:1 computationally:1 equation:2 ln:1 previously:1 discus:3 end:3 available:2 operation:3 rewritten:2 apply:3 batch:2 schmidt:1 gate:1 slower:2 assumes:2 top:1 ensure:3 completed:1 exploit:1 giving:3 kuntzmann:1 k1:1 establish:1 unchanged:3 objective:4 added:1 quantity:6 question:1 print:1 usual:1 unproven:1 gradient:34 amongst:1 separate:4 outer:2 evenly:1 tseng:1 nicta:1 length:2 besides:1 index:3 relationship:3 code:2 minimizing:1 hebrew:1 potentially:1 holding:1 expense:2 relate:1 stated:2 implementation:4 regulariser:11 perform:1 observation:1 datasets:2 finite:4 descent:5 dijkstra:1 extended:1 excluding:1 team:2 thm:1 introduced:1 namely:1 paris:2 required:1 connection:1 established:1 nip:1 able:3 suggested:2 redmond:1 below:3 usually:2 justin:1 examination:1 regularized:3 largescale:1 tives:1 brief:1 julien:2 kj:11 review:1 geometric:1 literature:1 l2:3 prior:1 python:1 epoch:3 regularisation:4 loss:3 lecture:1 adaptivity:1 proven:1 var:3 xiao:1 viewpoint:1 storing:3 row:3 summary:2 supported:1 last:1 svrg:21 bias:3 allow:1 expansiveness:1 midpoint:4 sparse:4 moreau:1 dimension:1 avoids:1 preventing:1 author:2 conservatively:1 adaptive:3 commonly:1 forward:1 suzuki:1 reinforcement:1 far:2 preferred:1 confirm:1 kkt:1 mairal:1 generalising:1 assumed:1 xi:4 shwartz:4 alternatively:1 continuous:2 search:3 why:1 table:9 additionally:1 pack:1 correlated:1 nicolas:1 european:1 pk:1 main:1 rh:1 big:3 paul:1 n2:1 finito:14 canberra:1 slow:1 tong:4 combettes:1 sub:1 saga:38 explicit:1 jmlr:2 weighting:2 theorem:2 xt:1 specific:2 showing:3 jakub:1 covtype:2 concern:1 mnist:2 adding:1 phd:1 conditioned:1 kx:2 easier:1 intersection:1 simply:2 explore:1 lbfgs:1 partially:1 springer:2 ljk:1 conditional:1 towards:2 lipschitz:2 replace:1 experimentally:1 change:3 typical:1 determined:1 uniformly:3 reducing:1 impractically:1 operates:1 except:1 lemma:9 domke:1 called:3 pas:3 experimental:1 aaron:3 mark:2 support:6 jonathan:1 brevity:1 accelerated:1 evaluate:1 reg:1 tested:3 ex:5 |
4,703 | 5,259 | Time?Data Tradeoffs by Aggressive Smoothing
John J. Bruer1,*
Joel A. Tropp1
Volkan Cevher2
Stephen R. Becker3
1 Dept. of Computing + Mathematical Sciences, California Institute of Technology
2 Laboratory for Information and Inference Systems, EPFL
3 Dept. of Applied Mathematics, University of Colorado at Boulder
* [email protected]
Abstract
This paper proposes a tradeoff between sample complexity and computation time
that applies to statistical estimators based on convex optimization. As the amount of
data increases, we can smooth optimization problems more and more aggressively
to achieve accurate estimates more quickly. This work provides theoretical and
experimental evidence of this tradeoff for a class of regularized linear inverse
problems.
1
Introduction
It once seemed obvious that the running time of an algorithm should increase with the size of the input.
But recent work in machine learning has led us to question this dogma. In particular, Shalev-Shwartz
and Srebro [1] showed that their algorithm for learning a support vector classifier actually becomes
faster when they increase the amount of training data. Other researchers have identified related
tradeoffs [2, 3, 4, 5, 6, 7, 8, 9]. Together, these works support an emerging perspective in statistical
computation that treats data as a computational resource that we can exploit to improve algorithms
for estimation and learning.
In this paper, we consider statistical algorithms based on convex optimization. Our primary contribution is the following proposal:
As the amount of available data increases, we can smooth statistical optimization
problems more and more aggressively. We can solve the smoothed problems
significantly faster without any increase in statistical risk.
Indeed, many statistical estimation procedures balance the modeling error with the complexity of the
model. When we have very little data, complexity regularization is essential to fit an accurate model.
When we have a large amount of data, we can relax the regularization without compromising the
quality of the model. In other words, excess data offers us an opportunity to accelerate the statistical
optimization. We propose to use smoothing methods [10, 11, 12] to implement this tradeoff.
We develop this idea in the context of the regularized linear inverse problem (RLIP) with random
data. Nevertheless, our ideas apply to a wide range of problems. We pursue a more sophisticated
example in a longer version of this work [13].
JJB?s and JAT?s work was supported under ONR award N00014-11-1002, AFOSR award FA9550-09-10643, and a Sloan Research Fellowship. VC?s work was supported in part by the European Commission under
Grant MIRG-268398, ERC Future Proof, SNF 200021-132548, SNF 200021-146750 and SNF CRSII2-147633.
SRB was previously with IBM Research, Yorktown Heights, NY 10598 during the completion of this work.
1
1.1
The regularized linear inverse problem
Let x\ ? Rd be an unknown signal, and let A ? Rm?d be a known measurement matrix. Assume
that we have access to a vector b ? Rm of m linear samples of that signal given by
b := Ax\ .
Given the pair (A, b), we wish to recover the original signal x\ .
We consider the case where A is fat (m < d), so we cannot recover x\ without additional information
about its structure. Let us introduce a proper convex function f : Rd ? R ? {+?} that assigns small
values to highly structured signals. Using the regularizer f , we construct the estimator
D := arg min f (x)
x
subject to Ax = b.
(1)
x
D = x\ , and we refer to this outcome as exact recovery.
We declare the estimator successful when x
While others have studied (1) in the statistical setting, our result is different in character from previous
work. Agarwal, Negahban, and Wainwright [14] showed that gradient methods applied to problems
like (1) converge in fewer iterations due to increasing restricted strong convexity and restricted
smoothness as sample size increases. They did not, however, discuss a time?data tradeoff explicitly,
nor did they recognize that the overall computational cost may rise as the problem sizes grow.
Lai and Yin [15], meanwhile, proposed relaxing the regularizer in (1) based solely on some norm
of the underlying signal. Our relaxation, however, is based on the sample size as well. Our method
results in better performance as sample size increases: a time?data tradeoff.
The RLIP (1) provides a good candidate for studying time?data tradeoffs because recent work in
convex geometry [16] gives a precise characterization of the number of samples needed for exact
recovery. Excess samples allow us to replace the optimization problem (1) with one that we can solve
faster. We do this for sparse vector and low-rank matrix recovery problems in Sections 4 and 5.
2
The geometry of the time?data tradeoff
In this section, we summarize the relevant results that describe the minimum sample size required to
solve the regularized linear inverse problem (1) exactly in a statistical setting.
2.1
The exact recovery condition and statistical dimension
We can state the optimality condition for (1) in a geometric form; cf. [17, Prop. 2.1].
Fact 2.1 (Exact recovery condition). The descent cone of a proper convex function f : Rd ? R?{+?}
at the point x is the convex cone
[(
)
D( f ; x) :=
y ? Rd : f (x + ?y) ? f (x) .
? >0
The regularized linear inverse problem (1) exactly recovers the unknown signal x\ if and only if
D( f ; x\ ) ? null(A) = {0}.
(2)
We illustrate this condition in Figure 1(a).
To determine the number of samples we need to ensure that the exact recovery condition (2) holds,
we must quantify the ?size? of the descent cones of the regularizer f .
Definition 2.2 (Statistical dimension [16, Def. 2.1]). Let C ? Rd be a convex cone. Its statistical
dimension ?(C) is defined as
f
g
?(C) := E k? C (g)k 2 ,
where g ? Rd has independent standard Gaussian entries, and ? C is the projection operator onto C.
When the measurement matrix A is sufficiently random, Amelunxen et al. [16] obtain a precise
characterization of the number m of samples required to achieve exact recovery.
2
x?
x?
nullHAL + x?
nullHAL + x?
?
?
9x : f HxL ? f Ix? M=
9x : f HxL ? f Ix? M=
DI f , x? M + x?
(a)
(b)
Figure 1: The geometric opportunity. Panel (a) illustrates the exact recovery condition (2). Panel (b)
shows a relaxed regularizer f? with larger sublevel sets. The shaded area indicates the difference
between the descent cones of f? and f at x\ . When we have excess samples, Fact 2.3 tells us that
the exact recovery condition holds with high probability, as in panel (a). A suitable relaxtion will
maintain exact recovery, as in panel (b), while allowing us to solve the problem faster.
Fact 2.3 (Exact recovery condition for the random RLIP [16, Thm. II]). Assume that the null space
of the measurement matrix A ? Rm?d in the RLIP (1) is oriented uniformly at random. (In particular,
a matrix with independent standard Gaussian entries has this property.) Then
?
m ? ? D( f ; x\ ) + C? d =? exact recovery holds with probability ? 1 ? ?;
?
m ? ? D( f ; x\ ) ? C? d =? exact recovery holds with probability ? ?,
p
where C? := 8 log(4/?).
In words, the RLIP undergoes a phase transition when the number m of samples equals ?(D( f ; x\ )).
Any additional samples are redundant, so we can try to exploit them to identify x\ more quickly.
2.2
A geometric opportunity
Chandrasekaran and Jordan [6] have identified a time?data tradeoff in the setting of denoising
problems based on Euclidean projection onto a constraint set. They argue that, when they have a large
number of samples, it is possible to enlarge the constraint set without increasing the statistical risk of
the estimator. They propose to use a discrete sequence of relaxations based on algebraic hierarchies.
We have identified a related opportunity for a time?data tradeoff in the RLIP (1). When we have
excess samples, we may replace the regularizer f with a relaxed regularizer f? that is easier to optimize.
In contrast to [6], we propose to use a continuous sequence of relaxations based on smoothing.
Figure 1 illustrates the geometry of our time?data tradeoff. When the number of samples exceeds
?(D( f ; x\ )), Fact 2.3 tells us that the situation shown in Figure 1(a) holds with high probability.
This allows us to enlarge the sublevel sets of the regularizer while still satisfying the exact recovery
condition, as shown in Figure 1(b). A suitable relaxation allows us to solve the problem faster. Our
geometric motivation is similar with [6] although our relaxation method is totally unrelated.
3
A time?data tradeoff via dual-smoothing
This section presents an algorithm that can exploit excess samples to solve the RLIP (1) faster.
3.1
The dual-smoothing procedure
The procedure we use applies Nesterov?s primal-smoothing method from [11] to the dual problem;
see [12]. Given a regularizer f , we introduce a family { f ? : ? > 0} of strongly convex majorants:
?
f ? (x) := f (x) + kxk 2 .
2
3
Algorithm 3.1 Auslender?Teboulle applied to the dual-smoothed RLIP
Input: measurement matrix A, observed vector b
1: z0 ? 0, z?0 ? z0 , ? 0 ? 1
2: for k = 0, 1, 2, . . . do
3:
yk ? (1 ? ? k )zk + ? k z? k
4:
xk ? arg minx f (x) + ?2 kxk 2 ? hyk , Ax ? bi
5:
z? k+1 ? z? k + k A?k 2 ? (b ? Axk )
6:
zk+1 ? (1 ? ? k )zk + ? k z? k+1
7:
? k+1 ? 2/(1 + (1 + 4/? k2 ) 1/2 )
8: end for
In particular, the sublevel sets of f ? grow as ? increases. We then replace f with f ? in the original
RLIP (1) to obtain new estimators of the form
D? := arg min f ? (x) subject to Ax = b.
x
(3)
x
The Lagrangian of the convex optimization problem (3) becomes
?
L ? (x, z) = f (x) + kxk 2 ? hz, Ax ? bi ,
2
where the Lagrange multiplier z is a vector in Rm . This gives a family of dual problems:
maximize g? (z) := min L ? (x, z) subject to z ? Rm .
x
(4)
Since f ? is strongly convex, the Lagrangian L has a unique minimizer xz for each dual point z:
xz := arg min L ? (x, z).
(5)
x
Strong duality holds for (3) and (4) by Slater?s condition [18, Sec. 5.2.3]. Therefore, if we solve the
dual problem (4) to obtain an optimal dual point, (5) returns the unique optimal primal point.
The dual function is differentiable with ?g? (z) = b ? Axz , and the gradient is Lipschitz-continuous
with Lipschitz constant L ? no larger than ??1 kAk 2 ; see [12, 11]. Note that L ? is decreasing in ?,
and so we call ? the smoothing parameter.
3.2
Solving the smoothed dual problem
In order to solve the smoothed dual problem (4), we apply the fast gradient method from Auslender
and Teboulle [19]. We present the pseudocode in Algorithm 3.1.
The computational cost of the algorithm depends on two things: the number of iterations necessary
for convergence and the cost of each iteration. The following result bounds the error of the primal
iterates xk with respect to the true signal x\ . The proof is in the supplemental material.
Proposition 3.1 (Primal convergence of Algorithm 3.1). Assume that the exact recovery condition
holds for the primal problem (3). Algorithm 3.1 applied to the smoothed dual problem (4) converges
?
to an optimal dual point z?
? . Let x ? be the corresponding optimal primal point given by (5). Then
the sequence of primal iterates {xk } satisfies
2 kAk kz?
?k
.
kx\ ? xk k ?
??k
The chosen regularizer affects the cost of Algorithm 3.1, line 4. Fortunately, this step is inexpensive
for many regularizers of interest. Since the matrix?vector product Axk in line 5 dominates the other
vector arithmetic, each iteration requires O(md) arithmetic operations.
3.3
The time?data tradeoff
Proposition 3.1 suggests that increasing the smoothing parameter ? leads to faster convergence of
the primal iterates of the Auslender?Teboulle algorithm. The discussion in Section 2.2 suggests that,
when we have excess samples, we can increase the smoothing parameter while maintaining exact
recovery. Our main technical proposal combines these two observations:
4
Maximal dual-smoothing of the ` 1 norm
Stat. dim. of the dual-smoothed ` 1 descent cones
102
Maximal smoothing parameter (?(m))
Normalized statistical dimension (?/d)
1
0.8
0.6
0.4
?=0
? = 0.1
?=1
? = 10
0.2
0
0
0.2
0.4
0.6
Normalized sparsity (?)
0.8
101
100
10?1
? = 0.01
? = 0.05
? = 0.1
? = 0.2
10?2
0
1
(a)
0.2
0.4
0.6
0.8
Normalized sample size (m/d)
1
(b)
Figure 2: Statistical dimension and maximal smoothing for the dual-smoothed ` 1 norm.
Panel (a) shows upper bounds for the normalized statistical dimension d ?1 D( f ? ; x\ ) of the dualsmoothed sparse vector recovery problem for several choices of ?. Panel (b) shows lower bounds for
the maximal smoothing parameter ?(m) for several choices of the normalized sparsity ? := s/d.
As the number m of measurements in the RLIP (1) increases, we smooth the dual
problem (4) more and more aggressively while maintaining exact recovery. The
Auslender?Teboulle algorithm can solve these increasingly smoothed problems
faster.
In order to balance the inherent tradeoff between smoothing and accuracy, we introduce the maximal
smoothing parameter ?(m). For a sample size m, ?(m) is the largest number satisfying
? D( f ?(m) ; x\ ) ? m.
(6)
Choosing a smoothing parameter ? ? ?(m) ensures that we do not cross the phase transition of
our RLIP. In practice, we need to be less aggressive in order to avoid the ?transition region?. The
following two sections provide examples that use our proposal to achieve a clear time?data tradeoff.
4
Example: Sparse vector recovery
In this section, we apply the method outlined in Section 3 to the sparse vector recovery problem.
4.1
The optimization problem
Assume that x\ is a sparse vector. The ` 1 norm serves as a convex proxy for sparsity, so we choose it
as the regularizer in the RLIP (1). This problem is known as basis pursuit, and it was proposed by
Chen et al. [20]. It has roots in geophysics [21, 22].
We apply the dual-smoothing procedure from Section 3 to obtain the relaxed primal problem, which
is equivalent to the elastic net of Zou and Hastie [23]. The smoothed dual is given by (4).
To determine the exact recovery condition, Fact 2.3, for the dual-smoothed RLIP (3), we must
compute the statistical dimension of the descent cones of f ? . We provide an accurate upper bound.
Proposition 4.1 (Statistical dimension bound for the dual-smoothed ` 1 norm). Let x ? Rd with s
nonzero entries, and define the normalized sparsity ? := s/d. Then
r Z
?
?
?
f
g
?
?
2
1
2
2
2
? D( f ? ; x) ? inf ?
? 1 + ? (1 + ? kxk ` ? ) + (1 ? ?)
(u ? ?) 2 e?u /2 du ?
?
?.
? ?0
d
? ?
?
?
5
150
Cost vs. sample size (` 1 norm)
?1011
? = 0.1
? = ?(m)/4
1
Average cost
Average number of iterations
Iterations vs. sample size (` 1 norm)
100
0.8
? = 0.1
? = ?(m)/4
0.6
50
1
1.5
2
2.5
3
Sample size (m)
(a)
3.5
0.4
4
?104
1
1.5
2
2.5
3
Sample size (m)
3.5
4
?104
(b)
Figure 3: Sparse vector recovery experiment. The average number of iterations (a) and the average
computational cost (b) of 10 random trials of the dual-smoothed sparse vector recovery problem with
ambient dimension d = 40 000 and normalized sparsity ? = 5% for various sample sizes m. The red
curve represents a fixed smoothing parameter ? = 0.1, while the blue curve uses ? = ?(m)/4. The
error bars indicate the minimum and maximum observed values.
The proof is provided in the supplemental material. Figure 2 shows the statistical dimension and
maximal smoothing curves for sparse vectors with ?1 entries. In order to apply this result we only
need estimates of the magnitude and sparsity of the signal.
To apply Algorithm 3.1 to this problem, we must calculate an approximate primal solution xz from a
dual point z (Algorithm 3.1, line 4). This step can be written as
xz ? ?(m) ?1 ? SoftThreshold(AT z, 1),
where [SoftThreshold(x,t)]i = sgn (x i ) ? max {|x i | ? t, 0}. Algorithm 3.1, line 5 dominates the total
cost of each iteration.
4.2
The time?data tradeoff
We can obtain theoretical support for the existence of a time?data tradeoff in the sparse recovery
problem by adapting Proposition 3.1. See the supplemental material for the proof.
Proposition 4.2 (Error bound for dual-smoothed sparse vector recovery). Let x\ ? Rd with s
nonzero entries, m be the sample size, and ?(m) be the maximal smoothing parameter (6). Given a
measurement matrix A ? Rm?d , assume the exact recovery condition (2) holds for the dual-smoothed
sparse vector recovery problem. Then the sequence of primal iterates from Algorithm 3.1 satisfies
f
g1
1
2d 2 ?(A) ? ? (1 + ?(m) kx\ k ` ? ) 2 + (1 ? ?) 2
,
kx\ ? xk k ?
?(m) ? k
where ? := s/d is the normalized sparsity of x\ , and ?(A) is the condition number of the matrix A.
For a fixed number k of iterations, as the number m of samples increases, Proposition 4.2 suggests
that the error decreases like 1/?(m). This observation suggests that we can achieve a time?data
tradeoff by smoothing.
4.3
Numerical experiment
Figure 3 shows the results of a numerical experiment that compares the performance difference
between current numerical practice and our aggressive smoothing approach.
Most practitioners use a fixed smoothing parameter ? that depends on the ambient dimension or
sparsity but not on the sample size. For the constant smoothing case, we choose ? = 0.1 based on the
recommendation in [15]. It is common, however, to see much smaller choices of ? [24, 25].
6
In contrast, our method exploits excess samples by smoothing the dual problem more aggressively.
We set the smoothing parameter ? = ?(m)/4. This heuristic choice is small enough to avoid the phase
transition of the RLIP while large enough to reap performance benefits. Our forthcoming work [13]
addressing the case of noisy samples provides a more principled way to select this parameter.
In the experiment, we fix both the ambient dimension d = 40 000 and the normalized sparsity ? = 5%.
To test each smoothing approach, we generate and solve 10 random sparse vector recovery models for
each value of the sample size m = 12 000, 14 000, 16 000, . . . , 38 000. Each random model comprises
a Gaussian measurement matrix A and a random sparse vector x\ whose nonzero entires are ?1 with
equal probability. We stop Algorithm 3.1 when the relative error kx\ ? xk k / kx\ k is less than 10?3 .
This condition guarantees that both methods maintain the same level of accuracy.
In Figure 3(a), we see that for both choices of ?, the average number of iterations decreases as sample
size increases. When we plot the total computational cost1 in Figure 3(b), we see that the constant
smoothing method cannot overcome the increase in cost per iteration. In fact, in this example, it would
be better to throw away excess data when using constant smoothing. Meanwhile, our aggressive
smoothing method manages to decrease total cost as sample size increases. The maximal speedup
achieved is roughly 2.5?.
We note that if the matrix A were orthonormal, the cost of both smoothing methods would decrease
as sample sizes increase. In particular, the uptick seen at m = 38 000 in Figure 3 would disappear
(but our method would maintain roughly the same relative advantage over constant smoothing).
This suggests that the condition number ?(A) indeed plays an important role in determining the
computational cost. We believe that using a Gaussian matrix A is warranted here as statistical models
often use independent subjects.
Let us emphasize that we use the same algorithm to test both smoothing approaches, so the relative
comparison between them is meaningful. The observed improvement shows that we have indeed
achieved a time?data tradeoff by aggressive smoothing.
5
Example: Low-rank matrix recovery
In this section, we apply the method outlined in Section 3 to the low-rank matrix recovery problem.
5.1
The optimization problem
Assume that X \ ? Rd1 ?d2 is low-rank. Consider a known measurement matrix A ? Rm?d , where
d := d 1 d 2 . We are given linear measurements of the form b = A ? vec(X \ ), where vec returns the
(column) vector obtained by stacking the columns of the input matrix. Fazel [26] proposed using the
Schatten 1-norm k?k S1 , the sum of the matrix?s singular values, as a convex proxy for rank. Therefore,
we follow Recht et al. [27] and select f = k?k S1 as the regularizer in the RLIP (1). The low-rank
matrix recovery problem has roots in control theory [28].
We apply the dual-smoothing procedure to obtain the approximate primal problem and the smoothed
dual problem, replacing the squared Euclidean norm in (3) with the squared Frobenius norm.
As in the sparse vector case, we must compute the statistical dimension of the descent cones of the
strongly convex regularizer f ? . In the case where the matrix X is square, the following is an accurate
upper bound for this quantity. (The non-square case is addressed in the supplemental material.)
Proposition 5.1 (Statistical dimension bound for the dual-smoothed Schatten 1-norm). Let X ?
Rd1 ?d1 have rank r, and define the normalized rank ? := r/d 1 . Then
(
"
1
? D( f ? ; X) ? inf ? + (1 ? ?) ? 1 + ? 2 (1 + ? kX k) 2
2
0?? ?2
d1
#)
p
(1 ? ?)
2
?1
2
2
+
24(1 + ? ) cos (?/2) ? ?(26 + ? ) 4 ? ?
+ o (1) ,
12?
as d 1 ? ? while keeping the normalized rank ? constant.
1We compute total cost as k ? md, where k is the number of iterations taken, and md is the dominant cost of
each iteration.
7
?1011
? = 0.1
? = ?(m)/4
600
Cost vs. sample size (Schatten 1-norm)
3
Average cost
Average number of iterations
Iterations vs. sample size (Schatten 1-norm)
400
? = 0.1
? = ?(m)/4
2
200
1
0
1
1.5
2
2.5
3
Sample size (m)
(a)
3.5
4
1
1.5
?104
2
2.5
3
Sample size (m)
3.5
(b)
4
?104
Figure 4: Low-rank matrix recovery experiment. The average number of iterations (a) and the
average cost (b) of 10 random trials of the dual-smoothed low-rank matrix recovery problem with
ambient dimension d = 200 ? 200 and normalized rank ? = 5% for various sample sizes m. The red
curve represents a fixed smoothing parameter ? = 0.1, while the blue curve uses ? = ?(m)/4. The
error bars indicate the minimum and maximum observed values.
The proof is provided in the supplemental material. The plots of the statistical dimension and maximal
smoothing curves closely resemble those of the ` 1 norm and are in the supplemental material as well.
In this case, Algorithm 3.1, line 4 becomes [12, Sec. 4.3]
Xz ? ?(m) ?1 ? SoftThresholdSingVal(mat(AT z), 1),
where mat is the inverse of the vec operator. Given a matrix X with SVD U ? diag(?) ? V T ,
SoftThresholdSingVal(X,t) = U ? diag (SoftThreshold(?,t)) ? V T .
Algorithm 3.1, line 5 dominates the total cost of each iteration.
5.2
The time?data tradeoff
When we adapt the error bound in Proposition 3.1 to this specific problem, the result is nearly same
as in the ` 1 case (Proposition 4.2). For completeness, we include the full statement of the result in
the supplementary material, along with its proof. Our experience with the sparse vector recovery
problem suggests that a tradeoff should exist for the low-rank matrix recovery problem as well.
5.3
Numerical experiment
Figure 4 shows the results of a substantially similar numerical experiment to the one performed for
sparse vectors. Again, current practice dictates using a smoothing parameter that has no dependence
on the sample size m [29]. In our tests, we choose the constant parameter ? = 0.1 recommended
by [15]. As before, we compare this with our aggressive smoothing method that selects ? = ?(m)/4.
In this case, we use the ambient dimension d = 200 ? 200 and set the normalized rank ? = 5%. We
test each method with 10 random trials of the low-rank matrix recovery problem for each value of the
sample size m = 11 250, 13 750, 16 250, . . . , 38 750. The measurement matrices are again Gaussian,
and the nonzero singular values of the random low-rank matrices X \ are 1. We solve each problem
with Algorithm 3.1, stopping when the relative error in the Frobenius norm is smaller than 10?3 .
In Figure 4, we see that both methods require fewer iterations for convergence as sample size increases.
Our aggressive smoothing method additionally achieves a reduction in total computational cost, while
the constant method does not. The observed speedup from exploiting the additional samples is 5.4?.
The numerical results show that we have indeed identified a time?data tradeoff via smoothing. While
this paper considers only the regularized linear inverse problem, our technique extends to other
settings. Our forthcoming work [13] addresses the case of noisy measurements, provides a connection
to statistical learning problems, and presents additional examples.
8
References
[1] S. Shalev-Shwartz and N. Srebro. SVM optimization: inverse dependence on training set size. In
Proceedings of the 25th Annual International Conference on Machine Learning (ICML 2008), pages
928?935, Helsinki, Finland, 2008. ACM.
[2] L. Bottou and O. Bousquet. The tradeoffs of large scale learning. In Advances in Neural Information
Processing Systems 20, pages 161?168, 2008.
[3] A. A. Amini and M. J. Wainwright. High-dimensional analysis of semidefinite relaxations for sparse
principal components. Ann. Statist., 37(5B):2877?2921, 2009.
[4] A. Agarwal, P. L. Bartlett, and J. C. Duchi. Oracle inequalities for computationally adaptive model
selection. arXiv, 2012, 1208.0129v1.
[5] Q. Berthet and P. Rigollet. Computational Lower Bounds for Sparse PCA. arXiv, 2013, 1304.0828v2.
[6] V. Chandrasekaran and M. I. Jordan. Computational and statistical tradeoffs via convex relaxation. Proc.
Natl. Acad. Sci. USA, 110(13):E1181?E1190, 2013.
[7] A. Daniely, N. Linial, and S. Shalev-Shwartz. More data speeds up training time in learning halfspaces
over sparse vectors. In Advances in Neural Information Processing Systems 26, pages 145?153, 2013.
[8] M. I. Jordan. On statistics, computation and scalability. Bernoulli, 19(4):1378?1390, 2013.
[9] D. Shender and J. Lafferty. Computation-Risk Tradeoffs for Covariance-Thresholded Regression. In
Proceedings of the 30th International Conference on Machine Learning (ICML 2013), pages 756?764,
2013.
[10] A. S. Nemirovsky and D. B. Yudin. Problem complexity and method efficiency in optimization. A
Wiley-Interscience Publication. John Wiley & Sons Inc., New York, 1983.
[11] Y. Nesterov. Smooth minimization of non-smooth functions. Math. Program., 103(1):127?152, 2005.
[12] S. R. Becker, E. J. Cand?s, and M. C. Grant. Templates for convex cone problems with applications to
sparse signal recovery. Math. Program. Comput., 3(3):165?218, 2011.
[13] J. J. Bruer, J. A. Tropp, V. Cevher, and S. R. Becker. Time?Data Tradeoffs by Smoothing. IEEE Journal of
Selected Topics in Signal Processing, submitted, 2014.
[14] A. Agarwal, S. Negahban, and M. J. Wainwright. Fast Global Convergence of Gradient Methods for
High-Dimensional Statistical Recovery. Ann. Statist., 40(5):2452?2482, 2012.
[15] M.-J. Lai and W. Yin. Augmented l(1) and Nuclear-Norm Models with a Globally Linearly Convergent
Algorithm. SIAM J. Imaging Sci., 6(2):1059?1091, 2013.
[16] D. Amelunxen, M. Lotz, M. B. McCoy, and J. A. Tropp. Living on the edge: A geometric theory of phase
transitions in convex optimization. Information and Inference, to appear, 2014.
[17] V. Chandrasekaran, B. Recht, P. A. Parrilo, and A. S. Willsky. The Convex Geometry of Linear Inverse
Problems. Found. Comput. Math., 12(6):805?849, 2012.
[18] S. Boyd and L. Vandenberghe. Convex optimization. Cambridge University Press, Cambridge, 2004.
[19] A. Auslender and M. Teboulle. Interior gradient and proximal methods for convex and conic optimization.
SIAM J. Optim., 16(3):697?725, 2006.
[20] S. S. Chen, D. L. Donoho, and M. A. Saunders. Atomic decomposition by basis pursuit. SIAM J. Sci.
Comput., 20(1):33?61, 1998.
[21] J. F. Claerbout and F. Muir. Robust modeling with erratic data. Geophysics, 38(5):826?844, 1973.
[22] F. Santosa and W. W. Symes. Linear Inversion of Band-Limited Reflection Seismograms. SIAM J. Sci.
Stat. Comput., 7(4):1307?1330, 1986.
[23] H. Zou and T. Hastie. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B Stat.
Methodol., 67:301?320, 2005.
[24] J.-F. Cai, S. Osher, and Z. Shen. Linearized Bregman Iterations for Compressed Sensing. Math. Comp.,
78(267):1515?1536, 2009.
[25] S. Osher, Y. Mao, B. Dong, and W. Yin. Fast linearized Bregman iteration for compressive sensing and
sparse denoising. Commun. Math. Sci., 8(1):93?111, 2010.
[26] M. Fazel. Matrix rank minimization with applications. PhD thesis, Stanford University, 2002.
[27] B. Recht, M. Fazel, and P. A. Parrilo. Guaranteed Minimum-Rank Solutions of Linear Matrix Equations
via Nuclear Norm Minimization. SIAM Rev., 52(3):471?501, 2010.
[28] M. Mesbahi and G. P. Papavassilopoulos. On the rank minimization problem over a positive semidefinite
linear matrix inequality. IEEE Trans. Automat. Control, 42(2):239?243, 1997.
[29] J.-F. Cai, E. J. Cand?s, and Z. Shen. A singular value thresholding algorithm for matrix completion. SIAM
J. Optim., 20(4):1956?1982, 2010.
9
| 5259 |@word trial:3 version:1 inversion:1 norm:17 d2:1 linearized:2 nemirovsky:1 covariance:1 decomposition:1 automat:1 reap:1 reduction:1 current:2 optim:2 must:4 written:1 john:2 numerical:6 plot:2 v:4 fewer:2 selected:1 xk:6 fa9550:1 volkan:1 provides:4 characterization:2 iterates:4 completeness:1 math:5 height:1 mathematical:1 along:1 combine:1 interscience:1 introduce:3 indeed:4 roughly:2 cand:2 nor:1 xz:5 globally:1 decreasing:1 little:1 increasing:3 becomes:3 totally:1 provided:2 underlying:1 unrelated:1 panel:6 null:2 cm:1 pursue:1 emerging:1 substantially:1 compressive:1 supplemental:6 guarantee:1 fat:1 exactly:2 classifier:1 rm:7 k2:1 control:2 ser:1 grant:2 appear:1 before:1 declare:1 positive:1 treat:1 acad:1 solely:1 studied:1 suggests:6 relaxing:1 shaded:1 co:1 limited:1 range:1 bi:2 fazel:3 unique:2 atomic:1 practice:3 implement:1 procedure:5 area:1 snf:3 significantly:1 adapting:1 projection:2 dictate:1 word:2 boyd:1 cannot:2 onto:2 selection:2 operator:2 interior:1 risk:3 context:1 e1181:1 optimize:1 equivalent:1 lagrangian:2 convex:19 shen:2 recovery:38 assigns:1 estimator:5 orthonormal:1 nuclear:2 vandenberghe:1 hierarchy:1 play:1 colorado:1 exact:18 us:2 mirg:1 satisfying:2 slater:1 observed:5 role:1 calculate:1 region:1 ensures:1 decrease:4 yk:1 principled:1 halfspaces:1 convexity:1 complexity:4 nesterov:2 dogma:1 solving:1 linial:1 efficiency:1 basis:2 accelerate:1 various:2 regularizer:12 fast:3 describe:1 tell:2 shalev:3 outcome:1 choosing:1 saunders:1 whose:1 heuristic:1 larger:2 solve:11 supplementary:1 stanford:1 relax:1 compressed:1 statistic:1 g1:1 noisy:2 sequence:4 differentiable:1 advantage:1 net:2 cai:2 propose:3 product:1 maximal:9 relevant:1 achieve:4 frobenius:2 scalability:1 exploiting:1 convergence:5 converges:1 illustrate:1 develop:1 completion:2 stat:4 strong:2 throw:1 soc:1 resemble:1 indicate:2 quantify:1 closely:1 compromising:1 vc:1 sgn:1 material:7 require:1 fix:1 proposition:9 hold:8 sufficiently:1 achieves:1 finland:1 estimation:2 proc:1 largest:1 minimization:4 gaussian:5 avoid:2 mccoy:1 publication:1 hyk:1 ax:5 improvement:1 rank:19 indicates:1 bernoulli:1 contrast:2 amelunxen:2 dim:1 inference:2 stopping:1 epfl:1 entire:1 lotz:1 selects:1 arg:4 overall:1 dual:30 proposes:1 smoothing:42 equal:2 once:1 srb:1 construct:1 enlarge:2 represents:2 icml:2 nearly:1 future:1 others:1 inherent:1 oriented:1 recognize:1 geometry:4 phase:4 maintain:3 interest:1 highly:1 joel:1 semidefinite:2 primal:12 natl:1 regularizers:1 accurate:4 ambient:5 bregman:2 edge:1 necessary:1 experience:1 euclidean:2 theoretical:2 cevher:1 column:2 modeling:2 teboulle:5 cost:19 stacking:1 addressing:1 entry:5 daniely:1 successful:1 commission:1 proximal:1 recht:3 international:2 negahban:2 siam:6 dong:1 together:1 quickly:2 squared:2 again:2 thesis:1 sublevel:3 choose:3 return:2 softthreshold:3 aggressive:7 parrilo:2 sec:2 inc:1 sloan:1 explicitly:1 depends:2 performed:1 try:1 root:2 red:2 recover:2 contribution:1 square:2 accuracy:2 identify:1 hxl:2 manages:1 comp:1 researcher:1 submitted:1 definition:1 inexpensive:1 obvious:1 proof:6 di:1 recovers:1 stop:1 sophisticated:1 actually:1 follow:1 strongly:3 tropp:2 replacing:1 axk:2 undergoes:1 quality:1 believe:1 usa:1 normalized:13 multiplier:1 true:1 regularization:3 aggressively:4 laboratory:1 nonzero:4 during:1 yorktown:1 kak:2 muir:1 duchi:1 reflection:1 common:1 pseudocode:1 rigollet:1 measurement:11 refer:1 cambridge:2 vec:3 smoothness:1 rd:8 outlined:2 mathematics:1 erc:1 e1190:1 access:1 longer:1 dominant:1 recent:2 showed:2 perspective:1 inf:2 commun:1 n00014:1 inequality:2 onr:1 caltech:1 seen:1 minimum:4 additional:4 relaxed:3 fortunately:1 converge:1 determine:2 redundant:1 maximize:1 living:1 signal:10 stephen:1 ii:1 arithmetic:2 full:1 recommended:1 smooth:5 exceeds:1 faster:8 technical:1 adapt:1 offer:1 cross:1 lai:2 award:2 regression:1 arxiv:2 iteration:20 agarwal:3 achieved:2 proposal:3 fellowship:1 addressed:1 grow:2 singular:3 subject:4 hz:1 thing:1 lafferty:1 jordan:3 call:1 practitioner:1 enough:2 affect:1 fit:1 forthcoming:2 hastie:2 identified:4 idea:2 tradeoff:27 pca:1 bartlett:1 becker:2 algebraic:1 york:1 clear:1 amount:4 band:1 statist:2 generate:1 exist:1 seismogram:1 per:1 blue:2 discrete:1 mat:2 nevertheless:1 thresholded:1 v1:1 imaging:1 relaxation:7 cone:9 sum:1 inverse:9 extends:1 family:2 chandrasekaran:3 def:1 bound:10 guaranteed:1 convergent:1 annual:1 oracle:1 constraint:2 mesbahi:1 helsinki:1 bousquet:1 speed:1 min:4 optimality:1 speedup:2 structured:1 smaller:2 increasingly:1 character:1 son:1 rev:1 s1:2 osher:2 restricted:2 boulder:1 taken:1 computationally:1 resource:1 equation:1 previously:1 discus:1 needed:1 end:1 serf:1 studying:1 available:1 operation:1 pursuit:2 apply:8 away:1 v2:1 amini:1 existence:1 original:2 running:1 cf:1 ensure:1 include:1 opportunity:4 maintaining:2 exploit:4 disappear:1 question:1 quantity:1 primary:1 dependence:2 md:3 gradient:5 minx:1 schatten:4 sci:5 topic:1 argue:1 considers:1 willsky:1 balance:2 statement:1 rise:1 jat:1 proper:2 unknown:2 allowing:1 upper:3 observation:2 descent:6 situation:1 precise:2 smoothed:17 thm:1 pair:1 required:2 connection:1 california:1 geophysics:2 trans:1 auslender:5 address:1 bar:2 sparsity:9 summarize:1 program:2 max:1 erratic:1 wainwright:3 suitable:2 regularized:6 methodol:1 improve:1 technology:1 conic:1 geometric:5 determining:1 relative:4 afosr:1 srebro:2 proxy:2 thresholding:1 ibm:1 supported:2 keeping:1 allow:1 institute:1 wide:1 template:1 sparse:21 benefit:1 curve:6 dimension:17 overcome:1 transition:5 yudin:1 seemed:1 crsii2:1 kz:1 berthet:1 adaptive:1 excess:8 approximate:2 emphasize:1 global:1 shwartz:3 continuous:2 additionally:1 zk:3 robust:1 elastic:2 du:1 warranted:1 bottou:1 european:1 meanwhile:2 zou:2 diag:2 did:2 main:1 linearly:1 motivation:1 augmented:1 ny:1 wiley:2 mao:1 comprises:1 wish:1 comput:4 candidate:1 ix:2 z0:2 specific:1 sensing:2 svm:1 evidence:1 dominates:3 essential:1 phd:1 magnitude:1 illustrates:2 kx:6 chen:2 easier:1 rd1:2 led:1 yin:3 lagrange:1 kxk:4 recommendation:1 applies:2 minimizer:1 satisfies:2 acm:1 prop:1 ann:2 donoho:1 replace:3 lipschitz:2 uniformly:1 denoising:2 principal:1 total:6 duality:1 experimental:1 svd:1 meaningful:1 select:2 support:3 dept:2 d1:2 |
4,704 | 526 | Oscillatory Model of Short Term Memory
David Horn
School of Physics and Astronomy
Raymond and Beverly Sackler
Faculty of Exact Sciences
Tel-Aviv University
Tel Aviv 69978, Israel
Marius U sher*
Dept. of Applied Mathematics
and Computer Science
Weizmann Institute of Science
Rehovot 76100, Israel
Abstract
We investigate a model in which excitatory neurons have dynamical thresholds which display both fatigue and potentiation. The fatigue property
leads to oscillatory behavior. It is responsible for the ability of the model
to perform segmentation, i.e., decompose a mixed input into staggered
oscillations of the activities of the cell-assemblies (memories) affected by
it. Potentiation is responsible for sustaining these staggered oscillations
after the input is turned off, i.e. the system serves as a model for short
term memory. It has a limited STM capacity, reminiscent of the magical
number 7 ? 2.
1
Introduction
The limited capacity (7 ? 2) of the short term memory (STM) has been a subject
of major interest in the psychological and physiological literature. It seems quite
natural to assume that the limited capacity is due to the special dynamical nature
of STM. Recently, Crick and Koch (1990) suggested that the working memory
is functionally related to the binding process, and is obtained via synchronized
oscillations of neural populations. The capacity limitation of STM may then result
from the competition between oscillations representing items in STM. In the model
which we investigate this is indeed the case.
?Present address: Division of Biology, 216-76, Caltech, Pasadena CA 91125.
125
126
Horn and Usher
Models of oscillating neural networks can perform various tasks:
1. Phase-locking and synchronization in response to global coherence in the stimuli, such as similarity of orientation and continuity (Kamen et al. 1989; Sompolinsyet al. 1990; Konig & Schillen 1991).
2. Segmentation of incoherent stimuli in low level vision via desynchronization,
using oscillator networks with delayed connections (Schillen & Konig 1991).
3. Segmentation according to semantic content, i.e., separate an input of mixed information into its components which are known memories of the system (Wang
et al. 1990, Horn and Usher 1991). In these models the memories are represented by competing cell a.'3semblies. The input, which affects a subset of these
assemblies, induces staggered oscillations of their activities. This works as long
as the number of memories in the input is small, of the order of 5.
4. Binding, i.e., connecting correctly different attributes of the same object which
appear in the mixed input (Horn et al. 1991). Binding can be interpreted as
matching the phases of oscillations representing attributes of the same object
in two different networks which are coupled in a way which does not assume
any relation between the attributes.
To these we add here the important task of
5. STM, i.e., keeping information about segmentation or binding after the input
is turned off.
In order to qualify as models for STM, the staggered oscillations have to prevail
after the input stimuli disappear. Unfortunately, this does not hold for the models
quot.ed above. Once the input disappears, either the network's activity dies out,
or oscillations of assemblies not included in the original input are turned on. In
other words, the oscillations have no inertia, and thus they do not persist after the
disappearance of the sensory input. Our purpose is to present a model of competing neural assemblies which, upon receiving a mixed input develops oscillations
which prevail after the st.imulus disappears. In order to achieve this, the biological
mechanism of post tetanic potentiation will be used.
2
Dynaillics of Short Ternl Potentiation
It was shown that following a t.etanus of electrophysiological stimulation temporary
modifications in the synaptic strengths, mostly non Hebbian, are observed (Crick
and Koch, 1990; Zucker, 1989). The time scale of these synaptic modifications
ranges between 50 111S to several minutes. A detailed description of the processes
responsible for this mechanism was given by Zucker (1989), exhibiting a rather complex behavior. In the following we will use a simplified version of these mechanisms
involving two processes with different time scales. We assume that following a prolonged activation of a synapse, the synaptic strength exhibits depression on a short
time scale, but recovers and becomes slightly enhanced on a longer time scale. As
illustrated in Fig 1 of Zucker (1989), this captures most of the dynamics of Short
Term Potentiation. The fact that these mechanisms are non Hebbian implies that
all synapses associated with a presynaptic cell are affected, and thus the unit of
change is the presynaptic cell (Crick & Koch 1990).
Oscillatory Model of Shorr Term Memory
Our previous oscillatory neural networks were based on the assumption that, in
addition to the customary properties of the formal neuron, its threshold increases
when the neuron keeps firing, thus exhibiting adaptation or fatigue (Horn & Usher
1989). Motivated by the STP findings we add a new component offacilitaion, which
takes place on a longer time scale than fatigue. We denote the dynamical threshold
by the continuous variable r which is chosen as a sum of two components, I and p,
representing fatigue and potentiation,
(1)
= all - a2p?
r
Their dynamics is governed by the equations
,dl/dt =
m
+ (l/CI
,dp/dt =
1)1
-
m
+ (1/c2 -
l)p
(2)
where m describes the average neuron activity (firing rate) on a time scale which
is large compared to the refractory period. The time constants of the fatigue and
potentiation components, Tj = c,c:' l are chosen so that TI < T2. As a result the
neuron displays fatigue on a short time scale, but recovers and becomes slightly
enhanced (potentiated) on a longer time scale. This is clearly seen in Fig. 1, which
shows the behavior when the activity m of the corresponding neuron is clamped at
1 for some time (due to sensory input) and quenched to zero afterwards.
3
_. -.
,,-
2
\
/
\f
I
\
r
\
1
,
"
.....
"-
.-.- . -. -.-
0
\
'\
'\
'\
-1
-2
""
0
.........
p. . . .
.......
....- ..-
./
.......
.......
./
40
20
60
100
80
time
Figure 1: Behavior of the dynamic threshold r and its fatigue
I and potentiation
p components, when the neuron activity m is clamped as shown. Time scale is
arbitrary. The parameters are
CI
= 1.2 C2 =
1.05
al
= 4 a2 = 1 .
We observe here that the threshold increases during the cell's activation, being
driven to its asymptotic value al c1-I.
After the release of the stimulus the dyCl
namic threshold decreases (i.e . the neuron recovers) and turns negative (signifying
127
128
Horn and Usher
potentiation). The parameters were chosen so that asymptotically the threshold
reaches zero, i.e. no permanent effect is left. In our model we will assume a similar
behavior for the excitatory cell-assemblies which carry the memories in our system.
3
The Model
Our basic model (Horn & Usher 1990) is composed of two kinds of neurons which
are assumed to have excitatory and inhibitory synapses exclusively. Memory patterns are carried by excitatory neurons only. Furthermore, we make the simplifying
assumption that the patterns do not overlap with one another, i.e. the model is
composed of disjoint Hebbian cell-assemblies of excitatory neurons which affect one
another through their interaction with a single assembly of inhibitory neurons.
Let us denote by mS'(t) the fraction of cell-assembly number Il which fires at time t,
and by m I (t) the fraction of active inhibitory neurons. We will refer to mS' as the
activity of the Ilth memory pattern. There are P different memories in the model,
and their activities obey the following differential equations
= -mS' + FT(AmS' -
Bm I - f}S' + is')
dmI/dt == -m I + FT(CM - Dm I - f}I)
(3)
M= LmS'
S'
(4)
dmS' /dt
where
f}S' and f}I are the thresholds of all excitatory and inhibitory neurons correspondingly
and is' represents the input into cell assembly Il. The four parameters ABC and
D are all positive and represent the different couplings between the neurons. This
system is an attractor neural network. In the absence of input and dynamical
thresholds it is a dissipative system which flows into fixed points determined by the
memOrIes.
This system is a generalization of the E-I model of Wilson and Cowan (1972) in
which we have introduced competing memory patterns. The latter make it into an
attractor neural network. Wilson and Cowan have shown that a pair of excitatory
and inhibitory assemblies, when properly connected, will form an oscillator. We
induce oscillations in a different way, keeping the option of having the network
behave either as an attractor neural network or as an oscillating one: we turn the
thresholds of the excitatory neurons into dynamic variables, which are defined by
f}S' = f}t;
+ brS' .
The dynamics of the new variables rS' are chosen to follow equations (1) and (2)
where all elements, r f p and m refer to the same cell-assembly 1-'. To understand
the effects of this change let us first limit ourselves to the fatigue component only,
1 and a2 = 0 in Eq. 1. Imagine a situation in which the system would flow
i.e. a1
into a fixed point mS'
1. rS' will then increase until it reaches the value cI/( C1 -1).
This means that the argument of the FT function in the equation for mS' decreases
by 9 = bCI/(Cl - 1) . If this overcomes the effect of the other terms the amplitude
mS' decreases and the system moves out of the attractor and falls into the basin
of a different center of attraction. This process can continue indefinitely creating
=
=
Oscillatory Model of Short Term Memory
an oscillatory network which moves from one memory to another. Envisage now
turning on a p/lo component leading to an r/lo behavior of the type depicted in Fig.
1. Its effect will evidently be the same as the input i/lo in Eq. (3) during the time
in which it is active. In other words, it will help to reactivate the cell-assembly /-l,
thus carrying the information that this memory was active before. Therefore, its
role in our system is to serve as the inertia component necessary for creating the
effect of STM.
4
Seglnentation and Short Term Memory
In this section we will present. results of numerical investigations of our model. The
parameters used in the following are A = C = D = 1 B = 1.10t; = 0.075 OJ =
-0.55 T = 0.05 b = 0.2 I = 2.5 and the values of ai and Ci of Fig. 1. We let n of the
P memories have a constant input of the form
i/lo
=i
/-l
= 1"
.. , n
=0
i/lo
/-l = n
+ 1,?,?, P.
(5)
An example of the result. of a system with P = 10 and n = 4 is shown in Fig. 2.
1.0 ,...
0.8
-
U1
\l.)
.....
o+J
.....
:>
.....
o+J
CJ
CO
I
,
,.,,11/
" 11? ,\ ,I ""
I
'\'
:
I
I
"
I
. ? '.
I ? .f ??
..
,
'I::
II,?::
0.6 - ,',."
. ",
,. . ., ,I,?:.
0.4 I - / '"t ? I l" r'::
.
.,
tr1
0.2
~
0.0
,
"
;
?
I
t ~
I.
oj
I,!.
I
"
::
"
.'
..
.'
,
1,1 ?
I ~r \
':
"
'
I ~
,\ ,I~I.\.
'.\1J\ ."\.1\t:.).,~~.: \ '.. 1.\,
::
.:
,
A :.
.,11::
I,
-
I
I
I
\<
., I ,'.
"1 1<
. I:
~
1 I
?
I, ::..
,,::
II
"
,.1,::
"~"
,.
'I, "
"
"
,.,,::
I,
'.
I.
\ ..
, ,I I' ,
.
" ' , \,1 I: :
I"
"
,
,,' I:: " :: '\/ I: .
',~~: I~I~"~'
Il'l
~.
"\.;~.'" ?
11,\.t
I, ~
~'" _'
',J\..J ~ '_'l
' ,
?
.A.
---------C>
4
3
2
1
a
a
25
75
50
100
125
time
Figure 2: Result.s of our model for P = 10 memories and n = 4 inputs. The first
frame displays the activities m of the four relevant cell-assemblies, and the second
frame represents their l' values. The arrow indicates the duration of the mixed
input.
129
130
Horn and Usher
Here we display the activities of the cell-assemblies that receive the constant input
and their corresponding average thresholds. While the signal of the mixed input
is on (denoted by an arrow along the time scale) we see how the phenomenon of
segmentation develops. The same staggered oscillation of the four cell-assemblies
which received an input is sustained after the signal is turned off. This indicates that
the system functions as a STM. Note that no synaptic connections were changed
and, once the system will receive a new input its behavior will be revised. However,
as long as it is left alone, it will continue to activate the cell-assemblies affected by
the original input.
We were able to obtain good results only for low n values, n ::; 4. As n is increased
we have difficulties wit.h both segmentation and STM. By modifying slightly the
paradigm we were able to feed 5 different inputs in a STM, as shown in Fig. 3.
This required presenting them at different times, as indicated by the 5 arrows on this
figure. In other words, t.his system does not perform segmentation but it continues
to work as a STM. Note, however, that the order of the different activities is no
longer maintained after the stimuli are turned off.
1.0
.-.-
I
0.8
tZl
tl)
~
.-
>
~
C)
ttl
0.6
.
....
II
, I ? l\
I ,? ?
?
?? ???
,:? ???.
I I
I
0.4
:
\:
0.2
.
:"':
? ;i
,
?:
-;
. /1 .
'1. ,I
I I ,I
., . ,
II
?? ..
.'
1
.'.
I' ::
;; :~
,?
0.0
o
o
25
75
50
100
125
time
Figure 3: Result.s for 5 inputs which are fed in consecutively at the times indicated
by the short arrows. The model functions as STM without segmentation.
Oscillatory Model of Short Term Memory
5
Discussion.
Psychological experiments show that subjects can repeat a sequence of verbal items
in perfect order a.'l long as their number is small (7 ? 2). The items may be numbers
or let.ters but can also be combinations of the latter such as words or recognizable
dates or acronyms. This proves that STM makes use of the encoded material in the
long term memory (Miller 1956). This relation between the two different kinds of
memOt'y lies in the basis of our model. Long term memory is represented by excitatory cell assemblies . Incorporating threshold fatigue into the model, it acquires the
capability of performing temporal segmentation of external input. Adding to the
threshold post tetanic potentiation, the model becomes capable of maintaining the
segmented information in the form of staggered oscillations. This is the property
which we view as responsible for STM.
Both segmentation and STM have very limited capacities. This seems to follow
from t.he oscillatory nature of the system which we use to model these functions.
In contrast with long term memory, whose capacity can be increased endlessly by
adding neurons and synaptic connections, we find here that only a few items can
be st.ored in t.he dynamic fashion of staggered oscillations, irrespective of the size
of the system. Vve regard this result as very significant, in view of the fact that
the same holds for the limited psychological ability of attention and STM. It may
indicate t.hat the oscillatory model contains the key to the understanding of these
psychological findings.
In order to validate the hypothesis that STM is based on oscillatory correlations
between firing rates of neurons, some more experimental neurobiological and psychophysical research is required. While no conclusive results were yet obtained from
recordings of t.he cortical activity in the monkey, some positive support has been obtained in psychophysical experiments. Preliminary results show that an oscillatory
component can be found in the percentage of correct responses in STM matching
experiments (Usher & Sagi 1991).
Our mathematical model is based on many specific assumptions. We believe that
our main results are characteristic of a class of such models which can be obtained
by changing various elements in our system. The main point is that dynamical
storage of information can be achieved through staggered oscillations of memory
activit.ies. Moreover, to sustain them in the absence of an external input, a potentiation capability ha.'l to be present. A model which contains both should be able
to accomodate STM in t.he fashion which we have demonstrated.
A cknowledgenlent s
M. Usher is the recipient of a Dov Biegun post-doctoral fellowship. We wish to
thank S. Popescu for helpful discussions.
References
Crick,F. & Koch,C. 1990. Towards a neurobiological theory of consciousness. Seminars in the Neurosciences 2, 263-275.
131
132
Horn and Usher
Horn,D., Sagi,D. & Usher,M. 1991. Segmentation, binding and illusory conjunctions. Neural Compo 3, 509-524.
Horn,D. & Usher ,M. 1989. Neural networks with dynamical thresholds, Phys. Rev.
A 40, 1036-1044.
Horn,D. & Usher,M. 1990. Excitatory-inhibitory networks with dynamical thresholds, Int. 1. NeuralSyst. 1, 249-257.
Horn,D. & Usher ,M. 1991. Parallel Activation of Memories is an Oscillatory Neural
Network. Neural Compo 3, 31-43.
Kammen,D.M., Holmes,P.J. & Koch C. 1990. Origin of oscillations in visual cortex:
Feedback versus local coupling. In Models of Brain Function, M.Cotterill ed., pp
273-284. Cambridge University Press.
Konig,P. & Schillen,T.B. 1991. Stimulus-dependent assembly formation of oscillatory responses: I. Synchronization, Neural Compo 3, 155-166.
Miller,G. 1956. The magical number seven plus minus two. Psych. Rev., 63,81-97.
Sompolinsky,H., Golomb,D. & Kleinfeld,D. 1990. Global processing of visual stimuli
in a neural network of coupled oscillators. Proc. Natl. Acad. of Sci. USA, 87,
7200-7204.
Schillen,T.B. & Konig,P. 1991. Stimulus-dependent assembly formation of oscillatory responses: I. Synchronization, Neural Compo 3, 155-166.
Wang,D., Buhmann,J. & von der Malsburg,C. 1990. Pattern segmentation in associative memory. Neural Compo 2, 94-106.
Wilson,H .R. & Cowan,J .D. 1972. Excitatory and inhibitory interactions in localized
populations of model neurons. Biophys. 1. 12, 1-24.
Usher,M. & Sa.gi D. 1991, in preparation.
Zucker ,R.S. 1989. Short-term synaptic plasticity. Ann. Rev. Neurosci. 12, 13-31.
PART III
SPEECH
| 526 |@word version:1 faculty:1 seems:2 r:2 simplifying:1 minus:1 carry:1 contains:2 exclusively:1 activation:3 yet:1 reminiscent:1 numerical:1 plasticity:1 alone:1 item:4 short:12 compo:5 indefinitely:1 mathematical:1 along:1 c2:2 differential:1 sustained:1 recognizable:1 indeed:1 behavior:7 brain:1 prolonged:1 endlessly:1 becomes:3 stm:20 moreover:1 golomb:1 israel:2 namic:1 kind:2 interpreted:1 cm:1 ttl:1 monkey:1 psych:1 astronomy:1 finding:2 temporal:1 ti:1 unit:1 appear:1 positive:2 before:1 local:1 sagi:2 limit:1 acad:1 firing:3 plus:1 doctoral:1 dissipative:1 sustaining:1 co:1 limited:5 range:1 weizmann:1 horn:13 responsible:4 matching:2 word:4 quenched:1 induce:1 storage:1 a2p:1 demonstrated:1 center:1 attention:1 duration:1 wit:1 attraction:1 holmes:1 his:1 population:2 enhanced:2 imagine:1 exact:1 hypothesis:1 origin:1 element:2 continues:1 persist:1 observed:1 ft:3 role:1 wang:2 capture:1 connected:1 sompolinsky:1 decrease:3 locking:1 dynamic:6 carrying:1 serve:1 upon:1 division:1 basis:1 various:2 represented:2 activate:1 formation:2 quite:1 encoded:1 whose:1 imulus:1 bci:1 ability:2 gi:1 envisage:1 kammen:1 associative:1 sequence:1 evidently:1 interaction:2 adaptation:1 turned:5 relevant:1 date:1 achieve:1 description:1 validate:1 competition:1 konig:4 oscillating:2 perfect:1 object:2 help:1 coupling:2 school:1 received:1 sa:1 eq:2 implies:1 indicate:1 synchronized:1 exhibiting:2 correct:1 attribute:3 modifying:1 consecutively:1 material:1 potentiation:11 generalization:1 decompose:1 investigation:1 preliminary:1 biological:1 hold:2 koch:5 lm:1 major:1 a2:2 purpose:1 proc:1 clearly:1 rather:1 wilson:3 conjunction:1 release:1 tetanic:2 properly:1 indicates:2 staggered:8 contrast:1 am:1 helpful:1 dependent:2 pasadena:1 relation:2 reactivate:1 stp:1 orientation:1 denoted:1 activit:1 special:1 dmi:1 once:2 having:1 biology:1 represents:2 t2:1 stimulus:8 develops:2 few:1 composed:2 delayed:1 phase:2 ourselves:1 fire:1 attractor:4 interest:1 investigate:2 cotterill:1 tj:1 natl:1 dov:1 capable:1 necessary:1 psychological:4 increased:2 subset:1 st:2 physic:1 receiving:1 off:4 connecting:1 von:1 tr1:1 external:2 creating:2 leading:1 int:1 permanent:1 view:2 option:1 capability:2 parallel:1 il:3 beverly:1 characteristic:1 miller:2 schillen:4 oscillatory:14 synapsis:2 reach:2 phys:1 ed:2 synaptic:6 pp:1 dm:2 associated:1 recovers:3 illusory:1 electrophysiological:1 segmentation:12 cj:1 amplitude:1 feed:1 dt:4 follow:2 response:4 sustain:1 synapse:1 furthermore:1 until:1 correlation:1 working:1 kleinfeld:1 continuity:1 indicated:2 believe:1 aviv:2 usa:1 effect:5 consciousness:1 semantic:1 illustrated:1 during:2 maintained:1 acquires:1 m:6 fatigue:10 presenting:1 recently:1 stimulation:1 refractory:1 he:4 functionally:1 refer:2 significant:1 cambridge:1 ai:1 mathematics:1 zucker:4 similarity:1 longer:4 cortex:1 add:2 driven:1 continue:2 qualify:1 der:1 caltech:1 seen:1 paradigm:1 period:1 signal:2 ii:4 afterwards:1 hebbian:3 segmented:1 long:6 post:3 y:1 a1:1 involving:1 basic:1 vision:1 represent:1 achieved:1 cell:16 c1:2 receive:2 addition:1 fellowship:1 usher:14 subject:2 recording:1 cowan:3 flow:2 iii:1 affect:2 competing:3 br:1 motivated:1 speech:1 depression:1 detailed:1 induces:1 percentage:1 inhibitory:7 neuroscience:1 disjoint:1 correctly:1 rehovot:1 affected:3 key:1 four:3 threshold:15 changing:1 asymptotically:1 fraction:2 sum:1 place:1 oscillation:16 coherence:1 dy:1 sackler:1 display:4 activity:12 strength:2 u1:1 argument:1 performing:1 marius:1 according:1 combination:1 describes:1 slightly:3 rev:3 modification:2 equation:4 turn:2 mechanism:4 fed:1 serf:1 acronym:1 observe:1 obey:1 customary:1 hat:1 magical:2 original:2 recipient:1 assembly:19 maintaining:1 malsburg:1 prof:1 disappear:1 psychophysical:2 move:2 disappearance:1 exhibit:1 dp:1 separate:1 thank:1 sci:1 capacity:6 seven:1 presynaptic:2 unfortunately:1 mostly:1 negative:1 perform:3 potentiated:1 neuron:19 revised:1 behave:1 situation:1 frame:2 arbitrary:1 david:1 introduced:1 pair:1 required:2 connection:3 conclusive:1 temporary:1 address:1 able:3 suggested:1 dynamical:7 pattern:5 oj:2 memory:28 overlap:1 natural:1 difficulty:1 turning:1 buhmann:1 representing:3 disappears:2 irrespective:1 carried:1 popescu:1 incoherent:1 sher:1 coupled:2 raymond:1 literature:1 understanding:1 asymptotic:1 synchronization:3 mixed:6 limitation:1 versus:1 localized:1 basin:1 lo:5 excitatory:11 changed:1 repeat:1 keeping:2 verbal:1 formal:1 understand:1 institute:1 fall:1 correspondingly:1 regard:1 feedback:1 cortical:1 sensory:2 inertia:2 simplified:1 bm:1 neurobiological:2 keep:1 overcomes:1 global:2 active:3 assumed:1 continuous:1 nature:2 ca:1 tel:2 complex:1 cl:1 main:2 neurosci:1 arrow:4 vve:1 fig:6 tl:1 fashion:2 seminar:1 wish:1 lie:1 governed:1 clamped:2 minute:1 specific:1 desynchronization:1 physiological:1 dl:1 incorporating:1 adding:2 prevail:2 ci:4 accomodate:1 biophys:1 depicted:1 visual:2 ters:1 binding:5 abc:1 ann:1 towards:1 oscillator:3 crick:4 content:1 change:2 absence:2 included:1 determined:1 experimental:1 support:1 latter:2 signifying:1 preparation:1 dept:1 phenomenon:1 |
4,705 | 5,260 | Local Linear Convergence of Forward?Backward
under Partial Smoothness
Jingwei Liang and Jalal M. Fadili
GREYC, CNRS-ENSICAEN-Univ. Caen
{Jingwei.Liang,Jalal.Fadili}@greyc.ensicaen.fr
Gabriel Peyr?
CEREMADE, CNRS-Univ. Paris-Dauphine
[email protected]
Abstract
In this paper, we consider the Forward?Backward proximal splitting algorithm to
minimize the sum of two proper closed convex functions, one of which having
a Lipschitz continuous gradient and the other being partly smooth relative to an
active manifold M. We propose a generic framework under which we show that
the Forward?Backward (i) correctly identifies the active manifold M in a finite
number of iterations, and then (ii) enters a local linear convergence regime that we
characterize precisely. This gives a grounded and unified explanation to the typical
behaviour that has been observed numerically for many problems encompassed in
our framework, including the Lasso, the group Lasso, the fused Lasso and the
nuclear norm regularization to name a few. These results may have numerous
applications including in signal/image processing processing, sparse recovery and
machine learning.
1
Introduction
1.1
Problem statement
Convex optimization has become ubiquitous in most quantitative disciplines of science. A common
trend in modern science is the increase in size of datasets, which drives the need for more efficient
optimization methods. Our goal is the generic minimization of composite functions of the form
minn ?(x) = F (x) + J(x) ,
(1.1)
x?R
where
(A.1) J : Rn ? R ? {+?} is a proper, closed and convex function;
(A.2) F is a convex and C 1,1 (Rn ) function whose gradient is ??Lipschitz continuous;
(A.3) Argmin ? 6= ?.
The class of problems (1.1) covers many popular non-smooth convex optimization problems encountered in various fields throughout science and engineering, including signal/image processing,
2
1
machine learning and classification. For instance, taking F = 2?
||y ? A ? || for some A ? Rm?n
and ? > 0, we recover the Lasso problem when J = || ? ||1 , the group Lasso for J = || ? ||1,2 , the
fused Lasso for J = ||D? ? ||1 with D = [DDIF , Id] and DDIF is the finite difference operator,
anti-sparsity regularization when J = || ? ||? , and nuclear norm regularization when J = || ? ||? .
1
The standard (non relaxed) version of Forward?Backward (FB) splitting algorithm [3] for solving
(1.1) updates to a new iterate xk+1 according to
xk+1 = prox?k J xk ? ?k ?F (xk ) ,
(1.2)
starting from any point x0 ? Rn , where 0 < ? ? ?k ? ? < 2/?. Recall that the proximity operator
is defined, for ? > 0, as
prox?J (x) = argminz?Rn
1.2
1
2? ||z
2
? x|| + J(z).
Contributions
In this paper, we present a unified local linear convergence analysis for the FB algorithm to solve
(1.1) when J is in addition partly smooth relative to a manifold M (see Definition 2.1 for details).
The class of partly smooth functions is very large and encompasses all previously discussed examples as special cases. More precisely, we first show that FB has a finite identification property,
meaning that after a finite number of iterations, say K, all iterates obey xk ? M for k ? K. Exploiting this property, we then show that after such a large enough number of iterations, xk converges
locally linearly. We characterize this regime and the rates precisely depending on the structure of the
active manifold M. In general, xk converges locally Q-linearly, and when M is an linear subspace,
the convergence becomes R-linear. Several experimental results on some of the problems discussed
above are provided to support our theoretical findings.
1.3
Related work
Finite support identification and local R-linear convergence of FB to solve the Lasso problem,
though in infinite-dimensional setting, is established in [4] under either a very restrictive injectivity
assumption, or a non-degeneracy assumption which is a specialization of ours (see (3.1)) to the `1
norm. A similar result is proved in [13], for F being a smooth convex and locally C 2 function
and J the `1 norm, under restricted injectivity and non-degeneracy assumptions. The `1 norm is a
partly smooth function and hence covered by our results. [1] proved Q-linear convergence of FB to
solve (1.1) for F satisfying restricted smoothness and strong convexity assumptions, and J being
a so-called convex decomposable regularizer. Again, the latter is a small subclass of partly smooth
functions, and their result is then covered by ours. For example, our framework covers the total
variation (TV) semi-norm and `? -norm regularizers which are not decomposable.
In [15, 16], the authors have shown finite identitification of active manifolds associated to partly
smooth functions for various algorithms, including the (sub)gradient projection method, Newtonlike methods, the proximal point algorithm. Their work extends that of e.g. [28] on identifiable
surfaces from the convex case to a general non-smooth setting. Using these results, [14] considered
the algorithm [25] to solve (1.1) where J is partly smooth, but not necessarily convex and F is
C 2 (Rn ), and proved finite identitification of the active manifold. However, the convergence rate
remains an open problem in all these works.
1.4
Notations
Suppose M ? Rn is a C 2 -manifold around x ? Rn , denote TM (x) the tangent space of M at
x ? Rn . The tangent model subspace is defined as
?
Tx = Lin ?J(x) ,
where Lin(C) is the linear hull of the convex set C ? Rn . For a linear subspace V , we denote PV
the orthogonal projector onto V and for a matrix A ? Rm?n , AV = A ? PV . Define the generalized
sign vector
ex = PTx ?J(x) .
For a convex set C ? Rn , ri(C) denotes its relative interior, i.e. the interior relative to its affine hull.
2
2
Partial smoothness
In addition to (A.1), our central assumption is that J is a partly smooth function. Partial smoothness
of functions is originally defined in [17]. Our definition hereafter specializes it to the case of proper
closed convex functions.
Definition 2.1. Let J be a proper closed convex function such that ?J(x) 6= ?. J is partly smooth
at x relative to a set M containing x if
(1) (Smoothness) M is a C 2 -manifold around x and J restricted to M is C 2 around x.
(2) (Sharpness) The tangent space TM (x) is Tx .
(3) (Continuity) The set?valued mapping ?J is continuous at x relative to M.
In the following, the class of partly smooth functions at x relative to M is denoted as PSx (M).
When M is an affine manifold, then M = x + Tx , and we denote this subclass as PSAx (x + Tx ).
When M is a linear manifold, then M = Tx , and we denote this subclass as PSLx (Tx ).
Capitalizing on the results of [17], it can be shown that under mild transversality assumptions, the
set of continuous convex partly smooth functions is closed under addition and pre-composition by a
linear operator. Moreover, absolutely permutation-invariant convex and partly smooth functions of
the singular values of a real matrix, i.e. spectral functions, are convex and partly smooth spectral
functions of the matrix [10].
It then follows that all the examples discussed in Section 1, including `1 , `1 ?`2 , `? , TV and nuclear
norm regularizers, are partly
smooth. In fact, the nuclear
norm is partly smooth at a matrix x relative
to the manifold M = x0 : rank(x0 ) = rank(x) . The first three regularizers are all part of the
class PSLx (Tx ), see Section 4 and [27] for details.
We now define a subclass of partly smooth functions where the active manifold is actually a subspace
and the generalized sign vector ex is locally constant.
Definition 2.2. J belongs to the class PSSx (Tx ) if and only if J ? PSAx (x+Tx ) or J ? PSLx (Tx )
and ex is constant near x, i.e. there exists a neighbourhood U of x such that ?x0 ? Tx ? U
ex0 = ex .
A typical family of functions that comply with this definition is that of partly polyhedral functions [26, Section 6.5], which includes the `1 and `? norms, and the TV semi-norm.
3
Local linear convergence of the FB method
In this section, we state our main result on finite identification and local linear convergence of FB.
Theorem 3.1. Assume that (A.1)-(A.3) hold. Suppose that the FB scheme is used to create a sequence xk which converges to x? ? Argmin ? such that J ? PSx? (Mx? ), F is C 2 near x? and
??F (x? ) ? ri ?J(x? ) .
(3.1)
Then we have the following holds,
(1) The FB scheme (1.2) has the finite identification property, i.e. there exists K ? 0, such that
for all k ? K, xk ? Mx? .
(2) Suppose moreover that ?? > 0 such that
PT ?2 F (x? )PT ?Id,
where T := Tx? . Then for all k ? K, the following holds.
(3.2)
(i) Q-linear convergence: if 0 < ? ? ?k ? ?? < min 2?? ?2 , 2? ?1 , then given any
1 > ? > ?e,
||xk+1 ? x? || ? ?||xk ? x? ||,
2
where ?e = max q(?), q(?
? ) ? [0, 1[ and q(?) = 1 ? 2?? + ? 2 ? 2 .
3
(ii) R-linear convergence: if J ? PSAx? (x? + T ) or J ? PSLx? (T ), then for 0 < ? ?
?k ? ?? < min 2?? ?2 , 2? ?1 , where ? ? ? is the Lipzchitz constant of PT ?F PT ,
then
||xk+1 ? x? || ? ?k ||xk ? x? ||,
where ?2k = 1 ? 2??k + ? 2 ?k2 ? [0, 1[. Moreover, if
the optimal linear rate can be achieved is
p
?? = 1 ? ?2 /? 2 .
?
?2
? ?? and set ?k ?
?
?2 ,
then
Remark 3.2.
? The non-degeneracy assumption in (3.1) can be viewed as a geometric generalization of strict complementarity of non-linear programming. Building on the arguments
of [16], it turns out that it is almost a necessary condition for finite identification of Mx? .
? Under the non-degeneracy and local strong convexity assumptions (3.1)-(3.2), one can actually show that x? is unique by extending the reasoning in [26].
? For F = G ? A, where G satisfies (A.2), assumption (3.2) and the constant ? can be
restated in terms of local strong convexity of G and restricted injectivity of A on T , i.e.
Ker(A) ? T = {0}.
2
? When F = 21 ||y ? A ? || , not only the minimizer x? is unique, but also the rates in Theorem 3.1 can be refined further as the gradient operator ?F becomes linear.
? Partial smoothness guarantees that xk arrives the active manifold in finite time, hence raising the hope of acceleration using second-order information. For instance, one can think
of turning to geometric methods along the manifold Mx? , where faster convergence rates
can be achieved. This is also the motivation behind the work of e.g. [19].
When J ? PSSx? (T ), it turns out that the restricted convexity assumption (3.2) of Theorem 3.1 can
be removed in some cases, but at the price of less sharp rates.
Theorem 3.3. Assume that (A.1)-(A.3) hold. For x? ? Argmin ?, suppose that
J ? PSSx? (Tx? ),
(3.1) is fulfilled, and there exists a subspace V such that Ker PT ?2 F (x)PT = V for any x ?
B (x? ), > 0. Let the FB scheme be used to create a sequence xk that converges to x? with
0 < ? ? ?k ? ?? < min 2?? ?2 , 2? ?1 , where ? > 0 (see the proof). Then there exists a constant
C > 0 and ? ? [0, 1[ such that for all k large enough
||xk ? x? || ? C?k .
A typical example where this result applies is for F = G ? A with G locally strongly convex, in
which case V = Ker(AT ).
4
Numerical experiments
In this section, we describe some examples to demonstrate the applicability of our results. More
precisely, we consider solving
2
minn 12 ||y ? Ax|| + ?J(x)
(4.1)
x?R
where y ? Rm is the observation, A : Rn ? Rm , ? is the tradeoff parameter, and J is either the
`1 -norm, the `? -norm, the `1 ? `2 -norm, the TV semi-norm or the nuclear norm.
Example 4.1 (`1 -norm). For x ? Rn , the sparsity promoting `1 -norm [8, 23] is
J(x) =
Pn
i=1 |xi |.
It can verified that J is a polyhedral norm, and thus J ? PSSx (Tx ) for the model subspace
M = Tx = u ? Rn : supp(u) ? supp(x) , and ex = sign(x).
The proximity operator of the `1 -norm is given by a simple soft-thresholding.
4
Example 4.2 (`1 ?`2 -norm). The `1 ?`2 -norm is usually used to promote group-structured
S sparsity
[29]. Let the support of x ? Rn be divided into non-overlapping blocks B such that b?B b =
{1, . . . , n}. The `1 ? `2 -norm is given by
J(x) = ||x||B =
P
b?B ||xb ||,
where xb = (xi )i?b ? R|b| . || ? ||B in general is not polyhedral, yet partly smooth relative to the
linear manifold
M = Tx = u ? Rn : suppB (u) ? suppB (x) , and ex = N (xb ) b?B ,
S
where suppB (x) =
b : xb 6= 0 , N (x) = x/||x|| and N (0) = 0. The proximity operator of the
`1 ? `2 norm is given by a simple block soft-thresholding.
Example 4.3 (Total Variation). As stated in the introduction, partial smoothness is preserved under
pre-composition by a linear operator. Let J0 be a closed convex function and D is a linear operator.
Popular examples are the TV semi-norm in which case J0 = || ? ||1 and D? = DDIF is a finite
difference approximation of the derivative [22], or the fused Lasso for D = [DDIF , Id] [24].
If J0 ? PSD? x (M0 ), then it is shown in [17, Theorem 4.2] that under an appropriate transversality
condition, J ? PSx (M) where
M = u ? Rn : D? u ? M0 .
In particular, for the case of the TV semi-norm, we have J ? PSSx (Tx ) with
M = Tx = u ? Rn : supp(D? u) ? I and ex = PTx Dsign(D? x)
where I = supp(D? x). The proximity operator for the 1D TV, though not available in closed form,
can be obtained efficiently using either the taut string algorithm [11] or the graph cuts [7].
Example 4.4 (Nuclear norm). Low-rank is the spectral extension of vector sparsity to matrixvalued data x ? Rn1 ?n2 , i.e. imposing the sparsity on the singular values of x. Let x = U ?x V ? a
reduced singular value decomposition (SVD) of x. The nuclear norm of a x is defined as
J(x) = ||x||? =
Pr
i=1 (?x )i ,
where rank(x) = r. It has been used for instance as SDP convex relaxation for many problems
including in machine learning [2, 12], matrix completion [21, 5] and phase retrieval [6].
It can be shown that the nuclear norm is partly smooth relative to the manifold [18, Example 2]
M = z ? Rn1 ?n2 : rank(z) = r .
The tangent space to M at x and ex are given by
TM (x) = z ? Rn1 ?n2 : z = U L? + M V ? , ?L ? Rn2 ?r , M ? Rn1 ?r , and ex = U V ? .
The proximity operator of the nuclear norm is just soft?thresholding applied to the singular values.
Recovery from random measurements In these examples, the forward observation model is
? ? N (0, ? 2 ),
y = Ax0 + ?,
(4.2)
where A ? Rm?n is generated uniformly at random from the Gaussian ensemble with i.i.d. zeromean and unit variance entries. The tested experimental settings are
(a) `1 -norm m = 48 and n = 128, x0 is 8-sparse;
(b) Total Variation m = 48 and n = 128, (DDIF x0 ) is 8-sparse;
(c) `? -norm m = 123 and n = 128, x0 has 10 saturating entries;
(d) `1 ? `2 -norm m = 48 and n = 128, x0 has 2 non-zero blocks of size 4;
(e) Nuclear norm m = 1425 and n = 2500, x0 ? R50?50 and rank(x0 ) = 5.
5
The number of measurements is chosen sufficiently large, ? small enough and ? of the order of
? so that [27, Theorem 1] applies, yielding that the minimizer of (4.1) is unique and verifies the
non-degeneracy and restricted strong convexity assumptions (3.1)-(3.2).
The convergence profile of ||xk ?x? || are depicted in Figure 1(a)-(e). Only local curves after activity
identification are shown. For `1 , TV and `? , the predicted rate coincides exactly with the observed
one. This is because these regularizers are all partly polyhedral gauges, and the data fidelity is
quadratic, hence making the predictions of Theorem 3.1(ii) exact. For the `1 ? `2 -norm, although
its active manifold is still a subspace, the generalized sign vector ek is not locally constant, which
entails that the the predicted rate of Theorem 3.1(ii) slightly overestimates the observed one. For the
nuclear norm, whose active manifold is not linear. Thus Theorem 3.1(i) applies, and the observed
and predicted rates are again close.
TV deconvolution In this image processing example, y is a degraded image generated according to the same forward model as (4.1), but now A is a convolution with a Gaussian kernel. The
anisotropic TV regularizer is used. The convergence profile is shown in Figure 1(f). Assumptions
(3.1)-(3.2) are checked a posteriori. This together with the fact that the anisotropic TV is polyhedral
justifies that the predicted rate is again exact.
?2
10
theoretical
practical
theoretical
practical
?2
10
0
10
?2
10
?6
10
10
k x k ? x ?k
k x k ? x ?k
?4
k x k ? x ?k
theoretical
practical
10
?4
?6
10
?4
10
?6
10
?8
?8
10
10
?8
10
?10
?10
10
?10
10
380
400
420
440
460
480
500
450
500
550
600
k
650
700
750
10
800
1000
(a) `1 (Lasso)
3000
4000
theoretical
practical
?2
6000
7000
8000
(c) `? -norm
(b) TV semi-norm
2
0
10
5000
k
10
10
theoretical
practical
?2
theoretical
practical
0
10
10
?2
k x k ? x ?k
10
?6
10
k x k ? x ?k
?4
k x k ? x ?k
2000
k
?4
10
?6
10
10
?4
10
?6
10
?8
?8
10
10
?10
?8
10
?10
?10
10
10
10
350
400
450
500
250
300
350
k
(d) `1 ? `2 -norm
400
450
k
(e) Nuclear norm
500
50
100
150
200
250
300
k
(f) TV deconvolution
Figure 1: Observed and predicted local convergence profiles of the FB method (1.2) in terms of
||xk ? x? || for different types of partly smooth functions. (a) `1 -norm; (b) TV semi-norm; (c) `? norm; (d) `1 ? `2 -norm; (e) Nuclear norm; (f) TV deconvolution.
5
Proofs
Lemma 5.1. Suppose that J ? PSx (M). Then for any x0 ? M ? U , where U is a neighbourhood
of x, the projector PM (x0 ) is uniquely valued and C 1 around x, and thus
x0 ? x = PTx (x0 ? x) + o ||x0 ? x|| .
If J ? PSAx (x + Tx ) or J ? PSLx (Tx ), then
x0 ? x = PTx (x0 ? x).
Proof. Partial smoothness implies that M is a C 2 ?manifold around x, then PM (x0 ) is uniquely
valued [20] and moreover C 1 near x [18, Lemma 4]. Thus, continuous differentiability shows
x0 ? x = PM (x0 ) ? PM (x) = DPM (x)(x ? x0 ) + o(||x ? x0 ||).
6
where DPM (x) is the derivative of PM at x. By virtue of [18, Lemma 4] and the sharpness propoerty of J, this derivative is given by
DPM (x) = PTM (x) = PTx ,
The case where M is affine or linear is immediate. This conlcudes the proof.
Proof of Theorem 3.1.
1. Classical convergence results of the FB scheme, e.g. [9], show that xk converges to some x? ?
Argmin ? 6= ? by assumption (A.3). Assumptions (A.1)-(A.2) entail that (3.1) is equivalent
to 0 ? ri ? ?(x? ) . Since F ? C 2 around x? , the smooth perturbation rule of partly smooth
functions [17, Corollary 4.7] ensures that ? ? PSx? (M). By definition of xk+1 , we have
1
?k Gk (xk ) ? Gk (xk+1 ) ? ??(xk+1 ).
where Gk = Id ? ?k ?F . By Baillon-Haddad theorem, Gk is non-expansive, hence
dist 0, ??(xk+1 ) ? ?1k ||Gk (xk ) ? Gk (xk+1 )|| ? ?1k ||xk ? xk+1 ||.
Since lim inf ?k = ? > 0, we obtain dist 0, ??(xk+1 ) ? 0. Owing to assumptions (A.1)(A.2), ? is subdifferentially continuous and thus ?(xk ) ? ?(x? ). Altogether, this shows that
the conditions of [15, Theorem 5.3] are fulfilled, whence the claim follows.
2. Take K > 0 sufficiently large such that for all k ? K, xk ? Mx? and xk ? B (x? ).
(i) Since prox?k J is firmly non-expansive, hence non-expansive, we have
||xk+1 ? x? || = ||prox?k J Gk xk ? prox?k J Gk x? || ? ||Gk xk ? Gk x? ||.
(5.1)
By virtue of Lemma 5.1, we have xk ? x? = PT (xk ? x? ) + o(||xk ? x? ||). This, together
with local C 2 smoothness of F and Lipschitz continuity of ?F entails
hxk ? x? , ?F (xk ) ? ?F (x? )i =
R1
hxk ? x? , ?2 F (x? + t(xk ? x? ))(xk ? x? )idt
R1
2
= 0 hPT (xk ? x? ), ?2 F (x? + t(xk ? x? ))PT (xk ? x? )idt + o ||xk ? x? ||
2
2
? ?||xk ? x? || + o ||xk ? x? || .
(5.2)
0
Since (3.2) holds and ?2 F (x) depends continuously on x, there exists > 0 such that
PT ?2 F (x)PT ?Id, ?x ? B (x? ). Thus, classical development of the right hand side of
(5.1) yields
2
2
||xk+1 ? x? || ? ||Gk xk ? Gk x? || = ||(xk ? x? ) ? ?k (?F (xk ) ? ?F (x? ))||
2
2
2
= ||xk ? x? || ? 2?k hxk ? x? , ?F (xk ) ? ?F (x? )i + ?k2 ||?F (xk ) ? ?F (x? )||
2
2
2
2
? ||xk ? x? || ? 2?k ?||xk ? x? || + ?k2 ? 2 ||xk ? x? || + o ||xk ? x? ||
2
2
= 1 ? 2??k + ? 2 ?k2 ||xk ? x? || + o ||xk ? x? || .
(5.3)
Taking the lim sup in this inequality gives
2
2
lim sup ||xk+1 ? x? || /||xk ? x? || ? q(?k ) = 1 ? 2??k + ? 2 ?k2 .
(5.4)
k?+?
It is clear that for 0 < ? ? ? ? ?? < min 2?? ?2 , 2? ?1 , q(?) ? [0, 1[, and q(?) ? ?e2 =
max q(?), q(?
? ) . Inserting this in (5.4), and using classical arguments yields the result.
(ii) We give the proof for M = T , that for M = x? + T is similar. Since xk and x? belong to
T , from xk+1 = prox?k J (Gk xk ) we have
Gk xk ? xk+1 ? ?k ?J(xk+1 ) ? xk+1 = PT Gk xk ? ?k ?J(xk+1 ) = PT Gk xk ? ?k ek+1 .
Similarily, we have x? = PT Gk x? ? ?k e? . We then arrive at
(xk+1 ? x? ) + ?k (ek+1 ? e? ) = (xk ? x? ) ? ?k PT ?F (PT xk ) ? PT ?F (PT x? ) . (5.5)
7
Moreover, maximal monotonicity of ?k ?J gives
||(xk+1 ? x? ) + ?k (ek+1 ? e? )||
2
2
2
2
= ||xk+1 ? x? || + 2hxk+1 ? x? , ?k (ek+1 ? e? )i + ?k ||ek+1 ? e? || ? ||xk+1 ? x? || .
It is straightforward to see that now, (5.2) becomes
2
hxk ? x? , PT ?F (PT xk ) ? PT ?F (PT x? )i ? ?||xk ? x? || .
Let ? be the Lipschitz constant of PT ?F PT . Obviously ? ? ?. Developing ||PT (Gk xk ?
2
Gk x? )|| similarly to (5.3) we obtain
2
2
2
||xk+1 ? x? || ? 1 ? 2??k + ? 2 ?k2 ||xk ? x? || = ?2k ||xk ? x? || ,
where ?k ? [0, 1[ for 0 < ? ? ?k ? ?? < min 2?/? 2 , 2/? . ?k is minimized at ??2 with the
proposed optimal rate whenever it obeys the given upper-bound.
Proof of Theorem 3.3. Arguing similarly to the proof of Theorem 3.1(ii), and using in addition that
e? = ex? is locally constant, we get
xk+1 ? x? = (xk ? x? ) ? ?k PT ?F (PT xk ) ? PT ?F (PT x? )
= (xk ? x? ) ? ?k
R1
0
PT ?2 F (x? + t(xk ? x? ))PT (xk ? x? )dt,
Denote Ht = PT ?2 F (x? + t(xk ? x? ))PT 0. Using that Ht is self-adjoint, we have
PV xk+1 = PV xk .
Since xk ? x? , it follows that PV xk = PV x? for all k sufficiently large. Observing that xk ? x? =
PV ? (xk ? x? ) for all large k, we get
xk+1 ? x? = xk ? x? ? ?k
R1
0
PV ? Ht PV ? (xk ? x? )dt.
1/2
Observe that V ? ? T . By definition, Bt = Ht PV ? is injective, and therefore, ?? > 0 such that
2
2
||Bt x|| > ?||x|| for all x 6= 0 and t ? [0, 1]. We then have
||xk+1 ? x? ||
2
2
R1
2
0
1
= ||xk ? x? || ? 2?k
= ||xk ? x? || ? 2?k
R
0
2
hxk ? x? , BtT Bt (xk ? x? )idt + ?k2 ||PV ? PT ?F (xk ) ? ?F (x? ) ||
2
2
||Bt (xk ? x? )|| dt + ?k2 ||PV ? PT || ||?F (xk ) ? ?F (x? )||
2
2
2
2
2
2
2
= ||xk ? x? || ? 2?k ?||xk ? x? || + ?k2 ||PT PV ? || ||?F (xk ) ? ?F (x? )||
2
? ||xk ? x? || ? 2?k ?||xk ? x? || + ?k2 ? 2 ||PV ? || ||PV ? (xk ? x? )||
2
2
2
2
2
? ||xk ? x? || ? 2?k ?||xk ? x? || + ?k2 ? 2 ||xk ? x? || = ?2k ||xk ? x? || .
It is easy to see again that ?k ? [0, 1[ whenever 0 < ? ? ?k ? ?? < min 2? ?1 , 2?? ?2 .
References
[1] A. Agarwal, S. Negahban, and M. Wainwright. Fast global convergence of gradient methods for highdimensional statistical recovery. The Annals of Statistics, 40(5):2452?2482, 10 2012.
[2] F. Bach. Consistency of trace norm minimization. The Journal of Machine Learning Research, 9:1019?
1048, 2008.
[3] H. H. Bauschke and P. L. Combettes. Convex analysis and monotone operator theory in Hilbert spaces.
Springer, 2011.
[4] K. Bredies and D. A. Lorenz. Linear convergence of iterative soft-thresholding. Journal of Fourier
Analysis and Applications, 14(5-6):813?837, 2008.
[5] E. J. Cand?s and B. Recht. Exact matrix completion via convex optimization. Foundations of Computational Mathematics, 9(6):717?772, 2009.
8
[6] E. J. Cand?s, T. Strohmer, and V. Voroninski. Phaselift: Exact and stable signal recovery from magnitude measurements via convex programming. Communications on Pure and Applied Mathematics,
66(8):1241?1274, 2013.
[7] A. Chambolle and J. Darbon. A parametric maximum flow approach for discrete total variation regularization. In Image Processing and Analysis with Graphs. CRC Press, 2012.
[8] S. S. Chen, D. L. Donoho, and M. A. Saunders. Atomic decomposition by basis pursuit. SIAM journal
on scientific computing, 20(1):33?61, 1999.
[9] P. L. Combettes and V. R. Wajs. Signal recovery by proximal Forward?Backward splitting. Multiscale
Modeling & Simulation, 4(4):1168?1200, 2005.
[10] A. Daniilidis, D. Drusvyatskiy, and A. S. Lewis. Orthogonal invariance and identifiability. to appear in
SIAM J. Matrix Anal. Appl., 2014.
[11] P. L. Davies and A. Kovac. Local extremes, runs, strings and multiresolution. Ann. Statist., 29:1?65,
2001.
[12] E. Grave, G. Obozinski, and F. Bach. Trace Lasso: a trace norm regularization for correlated designs.
arXiv preprint arXiv:1109.1990, 2011.
[13] E. Hale, W. Yin, and Y. Zhang. Fixed-point continuation for `1 -minimization: Methodology and convergence. SIAM Journal on Optimization, 19(3):1107?1130, 2008.
[14] W. L. Hare. Identifying active manifolds in regularization problems. In H. H. Bauschke, R. S., Burachik,
P. L. Combettes, V. Elser, D. R. Luke, and H. Wolkowicz, editors, Fixed-Point Algorithms for Inverse
Problems in Science and Engineering, volume 49 of Springer Optimization and Its Applications, chapter 13. Springer, 2011.
[15] W. L. Hare and A. S. Lewis. Identifying active constraints via partial smoothness and prox-regularity.
Journal of Convex Analysis, 11(2):251?266, 2004.
[16] W. L. Hare and A. S. Lewis. Identifying active manifolds. Algorithmic Operations Research, 2(2):75?82,
2007.
[17] A. S. Lewis. Active sets, nonsmoothness, and sensitivity. SIAM Journal on Optimization, 13(3):702?725,
2003.
[18] A. S. Lewis and J. Malick. Alternating projections on manifolds. Mathematics of Operations Research,
33(1):216?234, 2008.
[19] S. A. Miller and J. Malick. Newton methods for nonsmooth convex minimization: connections amongLagrangian, Riemannian newton and SQP methods. Mathematical programming, 104(2-3):609?633,
2005.
[20] R. A. Poliquin, R. T. Rockafellar, and L. Thibault. Local differentiability of distance functions. Trans.
Amer. Math. Soc., 352:5231?5249, 2000.
[21] B. Recht, M. Fazel, and P. A. Parrilo. Guaranteed minimum-rank solutions of linear matrix equations via
nuclear norm minimization. SIAM review, 52(3):471?501, 2010.
[22] L. I. Rudin, S. Osher, and E. Fatemi. Nonlinear total variation based noise removal algorithms. Physica
D: Nonlinear Phenomena, 60(1):259?268, 1992.
[23] R. Tibshirani. Regression shrinkage and selection via the Lasso. Journal of the Royal Statistical Society.
Series B. Methodological, 58(1):267?288, 1996.
[24] R. Tibshirani, M. Saunders, S. Rosset, J. Zhu, and K. Knight. Sparsity and smoothness via the fused
Lasso. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 67(1):91?108, 2004.
[25] P. Tseng and S. Yun. A coordinate gradient descent method for nonsmooth separable minimization. Math.
Prog. (Ser. B), 117, 2009.
[26] S. Vaiter, M. Golbabaee, M. J. Fadili, and G. Peyr?. Model selection with low complexity priors. Available at arXiv:1304.6033, 2013.
[27] S. Vaiter, G. Peyr?, and M. J. Fadili.
arXiv:1405.1004, 2014.
Model consistency of partly smooth regularizers.
Available
[28] S. J. Wright. Identifiable surfaces in constrained optimization. SIAM Journal on Control and Optimization, 31(4):1063?1079, 1993.
[29] M. Yuan and Y. Lin. Model selection and estimation in regression with grouped variables. Journal of the
Royal Statistical Society: Series B (Statistical Methodology), 68(1):49?67, 2005.
9
| 5260 |@word mild:1 version:1 norm:48 open:1 simulation:1 decomposition:2 series:3 hereafter:1 ours:2 yet:1 numerical:1 update:1 rudin:1 xk:117 ptm:1 iterates:1 math:2 zhang:1 mathematical:1 along:1 become:1 yuan:1 polyhedral:5 x0:22 cand:2 dist:2 sdp:1 becomes:3 provided:1 notation:1 moreover:5 elser:1 argmin:4 string:2 unified:2 finding:1 wajs:1 guarantee:1 quantitative:1 subclass:4 exactly:1 rm:5 k2:11 ser:1 control:1 unit:1 appear:1 overestimate:1 engineering:2 local:13 matrixvalued:1 id:5 luke:1 appl:1 obeys:1 fazel:1 unique:3 practical:6 arguing:1 atomic:1 block:3 ker:3 j0:3 composite:1 projection:2 jingwei:2 pre:2 davy:1 get:2 onto:1 interior:2 close:1 operator:11 selection:3 equivalent:1 projector:2 straightforward:1 fadili:4 starting:1 convex:24 sharpness:2 restated:1 decomposable:2 splitting:3 recovery:5 pure:1 identifying:3 rule:1 nuclear:14 variation:5 coordinate:1 annals:1 pt:35 suppose:5 exact:4 programming:3 complementarity:1 trend:1 satisfying:1 cut:1 observed:5 preprint:1 enters:1 ensures:1 removed:1 knight:1 convexity:5 complexity:1 solving:2 basis:1 various:2 tx:20 caen:1 regularizer:2 chapter:1 univ:2 fast:1 describe:1 refined:1 saunders:2 whose:2 grave:1 solve:4 valued:3 say:1 statistic:1 think:1 obviously:1 sequence:2 propose:1 maximal:1 fr:2 inserting:1 multiresolution:1 adjoint:1 exploiting:1 convergence:19 regularity:1 sqp:1 extending:1 r1:5 converges:5 depending:1 completion:2 ex:10 strong:4 soc:1 predicted:5 implies:1 owing:1 hull:2 crc:1 behaviour:1 generalization:1 extension:1 physica:1 hold:5 proximity:5 around:6 considered:1 sufficiently:3 wright:1 mapping:1 algorithmic:1 claim:1 m0:2 estimation:1 ex0:1 grouped:1 create:2 gauge:1 minimization:6 hope:1 gaussian:2 pn:1 shrinkage:1 jalal:2 corollary:1 ax:1 methodological:1 rank:7 ceremade:2 expansive:3 whence:1 posteriori:1 cnrs:2 bt:4 voroninski:1 classification:1 fidelity:1 denoted:1 malick:2 development:1 constrained:1 special:1 field:1 having:1 taut:1 promote:1 minimized:1 btt:1 idt:3 nonsmooth:2 few:1 modern:1 phase:1 psd:1 ensicaen:2 arrives:1 extreme:1 yielding:1 hpt:1 behind:1 regularizers:5 strohmer:1 xb:4 partial:7 necessary:1 injective:1 orthogonal:2 phaselift:1 theoretical:7 instance:3 soft:4 modeling:1 cover:2 ax0:1 similarily:1 applicability:1 entry:2 peyr:3 characterize:2 bauschke:2 proximal:3 rosset:1 recht:2 negahban:1 siam:6 sensitivity:1 discipline:1 together:2 fused:4 continuously:1 again:4 central:1 rn1:4 containing:1 ek:6 derivative:3 supp:4 prox:7 parrilo:1 rn2:1 vaiter:2 includes:1 rockafellar:1 depends:1 closed:7 observing:1 sup:2 recover:1 identifiability:1 contribution:1 minimize:1 degraded:1 variance:1 efficiently:1 ensemble:1 yield:2 miller:1 identification:6 daniilidis:1 drive:1 whenever:2 checked:1 definition:7 hare:3 e2:1 associated:1 proof:8 riemannian:1 degeneracy:5 proved:3 wolkowicz:1 popular:2 recall:1 lim:3 ubiquitous:1 hilbert:1 actually:2 originally:1 dt:3 methodology:3 amer:1 though:2 strongly:1 zeromean:1 chambolle:1 just:1 hand:1 multiscale:1 overlapping:1 nonsmoothness:1 nonlinear:2 continuity:2 scientific:1 greyc:2 name:1 building:1 regularization:6 hence:5 bredies:1 alternating:1 self:1 uniquely:2 coincides:1 generalized:3 yun:1 demonstrate:1 reasoning:1 image:5 meaning:1 common:1 volume:1 anisotropic:2 discussed:3 belong:1 numerically:1 measurement:3 composition:2 imposing:1 smoothness:11 consistency:2 pm:5 similarly:2 mathematics:3 stable:1 entail:3 surface:2 belongs:1 inf:1 hxk:6 kovac:1 inequality:1 dauphine:2 injectivity:3 minimum:1 relaxed:1 signal:4 semi:7 ii:6 thibault:1 smooth:25 faster:1 bach:2 lin:3 retrieval:1 divided:1 prediction:1 regression:2 arxiv:4 iteration:3 grounded:1 kernel:1 agarwal:1 achieved:2 preserved:1 addition:4 singular:4 strict:1 dpm:3 flow:1 near:3 enough:3 easy:1 iterate:1 lasso:12 tm:3 tradeoff:1 specialization:1 remark:1 gabriel:2 covered:2 clear:1 locally:7 statist:1 differentiability:2 argminz:1 reduced:1 continuation:1 sign:4 fulfilled:2 correctly:1 darbon:1 tibshirani:2 discrete:1 group:3 verified:1 ht:4 backward:5 ptx:5 graph:2 relaxation:1 monotone:1 sum:1 run:1 inverse:1 extends:1 throughout:1 family:1 almost:1 arrive:1 prog:1 bound:1 guaranteed:1 quadratic:1 encountered:1 identifiable:2 activity:1 burachik:1 precisely:4 constraint:1 ri:3 fourier:1 argument:2 min:6 separable:1 structured:1 tv:15 according:2 developing:1 slightly:1 drusvyatskiy:1 making:1 osher:1 restricted:6 invariant:1 pr:1 equation:1 previously:1 remains:1 turn:2 r50:1 capitalizing:1 available:3 pursuit:1 operation:2 promoting:1 obey:1 observe:1 generic:2 spectral:3 appropriate:1 neighbourhood:2 altogether:1 denotes:1 newton:2 restrictive:1 classical:3 society:3 parametric:1 gradient:6 subspace:7 mx:5 distance:1 manifold:22 tseng:1 fatemi:1 minn:2 psx:5 liang:2 statement:1 gk:19 trace:3 stated:1 anal:1 design:1 proper:4 upper:1 av:1 observation:2 convolution:1 datasets:1 finite:12 descent:1 anti:1 immediate:1 communication:1 peyre:1 rn:17 perturbation:1 sharp:1 paris:1 connection:1 raising:1 established:1 trans:1 usually:1 regime:2 sparsity:6 encompasses:1 haddad:1 including:6 max:2 explanation:1 royal:3 wainwright:1 turning:1 zhu:1 scheme:4 firmly:1 numerous:1 identifies:1 specializes:1 comply:1 geometric:2 review:1 tangent:4 removal:1 prior:1 relative:10 permutation:1 foundation:1 affine:3 thresholding:4 editor:1 side:1 taking:2 sparse:3 curve:1 fb:12 forward:7 author:1 monotonicity:1 global:1 active:13 xi:2 continuous:6 iterative:1 necessarily:1 main:1 linearly:2 motivation:1 noise:1 profile:3 n2:3 verifies:1 encompassed:1 combettes:3 sub:1 pv:15 theorem:14 hale:1 virtue:2 deconvolution:3 exists:5 lorenz:1 magnitude:1 justifies:1 chen:1 depicted:1 yin:1 saturating:1 applies:3 springer:3 minimizer:2 satisfies:1 lewis:5 obozinski:1 goal:1 viewed:1 acceleration:1 donoho:1 ann:1 lipschitz:4 price:1 typical:3 infinite:1 uniformly:1 lemma:4 called:1 total:5 partly:23 experimental:2 svd:1 invariance:1 highdimensional:1 support:3 latter:1 absolutely:1 tested:1 phenomenon:1 correlated:1 |
4,706 | 5,261 | Simple MAP Inference via Low-Rank Relaxations
Roy Frostig?, Sida I. Wang,? Percy Liang, Christopher D. Manning
Computer Science Department, Stanford University, Stanford, CA, 94305
{rf,sidaw,pliang}@cs.stanford.edu, [email protected]
Abstract
We focus on the problem of maximum a posteriori (MAP) inference in Markov
random fields with binary variables and pairwise interactions. For this common
subclass of inference tasks, we consider low-rank relaxations that interpolate between the discrete problem and its full-rank semidefinite relaxation. We develop
new theoretical bounds studying the effect of rank, showing that as the rank grows,
the relaxed objective increases but saturates, and that the fraction in objective value
retained by the rounded discrete solution decreases. In practice, we show two algorithms for optimizing the low-rank objectives which are simple to implement, enjoy
ties to the underlying theory, and outperform existing approaches on benchmark
MAP inference tasks.
1
Introduction
Maximum a posteriori (MAP) inference in Markov random fields (MRFs) is an important problem
with abundant applications in computer vision [1], computational biology [2], natural language
processing [3], and others. To find MAP solutions, stochastic hill-climbing and mean-field inference
are widely used in practice due to their speed and simplicity, but they do not admit any formal
guarantees of optimality. Message passing algorithms based on relaxations of the marginal polytope
[4] can offer guarantees (with respect to the relaxed objective), but require more complex bookkeeping.
In this paper, we study algorithms based on low-rank SDP relaxations which are both remarkably
simple and capable of guaranteeing solution quality.
Our focus is on MAP in a restricted but common class of models, namely those over binary variables
coupled by pairwise interactions. Here, MAP can be cast as optimizing a quadratic function over
the vertices of the n-dimensional hypercube: maxx2{ 1,1}n xT Ax. A standard optimization strategy
is to relax this integer quadratic program (IQP) to a semidefinite program (SDP), and then round
the relaxed solution to a discrete one achieving a constant factor approximation to the IQP optimum
[5, 6, 7]. In practice, the SDP can be solved efficiently using low-rank relaxations [8] of the form
maxX2Rn?k tr(X > AX).
The first part of this paper is a theoretical study of the effect of the rank k on low-rank relaxations of
the IQP. Previous work focused on either using SDPs to solve IQPs [5] or using low-rank relaxations
to solve SDPs [8]. We instead consider the direct link between the low-rank problem and the IQP. We
show that as k increases, the gap between the relaxed low-rank objective and the SDP shrinks, but
vanishes as soon as k rank(A); our bound adapts to the
p problem A and can thereby be considerably
better than the typical data-independent bound of O( n) [9, 10]. We also show that the rounded
objective shrinks in ratio relative to the low-rank objective, but at a steady rate of ?(1/k) on average.
This result relies on the connection we establish between IQP and low-rank relaxations. In the end,
our analysis motivates the use of relatively small values of k, which is advantageous from both a
solution quality and algorithmic efficiency standpoint.
?
Authors contributed equally.
1
The second part of this paper explores the use of very low-rank relaxation and randomized rounding
(R3 ) in practice. We use projected gradient and coordinate-wise ascent for solving the R3 relaxed
problem (Section 4). We note that R3 interfaces with the underlying problem in an extremely simple
way, much like Gibbs sampling and mean-field: only a black box implementation of x 7! Ax is
required. This decoupling permits users to customize their implementation based on the structure
of the weight matrix A: using GPUs for dense A, lists for sparse A, or much faster specialized
algorithms for A that are Gaussian filters [11]. In contrast, belief propagation and marginal polytope
relaxations [2] need to track messages for each edge or higher-order clique, thereby requiring more
memory and a finer-grained interface to the MRF that inhibits flexibility and performance.
Finally, we introduce a comparison framework for algorithms via the x 7! Ax interface, and use it to
compare R3 with annealed Gibbs sampling and mean-field on a range of different MAP inference
tasks (Section 5). We found that R3 often achieves the best-scoring results, and we provide some
intuition for our advantage in Section 4.1.
2
Setup and background
Notation We write Sn for the set of symmetric n ? n real matrices and S k for the unit sphere
{x 2 Rk : kxk2 = 1}. All vectors are columns unless stated otherwise. If X is a matrix, then
Xi 2 R1?k is its i?th row.
This section reviews how MAP inference on binary graphical models with pairwise interactions can
be cast as integer quadratic programs (IQPs) and approximately solved via semidefinite relaxations
and randomized rounding. Let us begin with the definition of an IQP:
Definition 2.1. Let A 2 Sn be a symmetric n ? n matrix. An (indefinite) integer quadratic program
(IQP) is the following optimization problem:
max
def
x2{ 1,1}n
IQP(x) = xT Ax
(1)
Solving (1) is NP-complete in general: the MAX-CUT problem immediately reduces to it [5]. With
an eye towards tractability, consider a first candidate relaxation: maxx2[ 1,1]n xT Ax. This relaxation
is always tight in that the maxima of the relaxed objective and original objective (1) are equal.1
Therefore it is just as hard to solve. Let us then replace each scalar xi 2 [ 1, 1] with a unit vector
Xi 2 Rk and define the following low-rank problem (LRP):
Definition 2.2. Let k 2 {1, . . . , n} and A 2 Sn . Define the low-rank problem LRPk by:
max
X2Rn?k
def
LRPk (X) = tr(X T AX)
subject to kXi k2 = 1, i = 1, . . . , n.
(2)
Note that setting Xi = [xi , 0, . . . , 0] 2 Rk recovers (1). More generally, we have a sequence of
successively looser relaxations as k increases. What we get in return is tractability. The LRPk
objective generally yields a non-convex problem, but if we take k = n, the objective can be rewritten
as tr(X > AX) = tr(AXX > ) = tr(AS), where S is a positive semidefinite matrix with ones on the
diagonal. The result is the classic SDP relaxation, which is convex:
max
S2Sn
def
SDP(S) = tr(AS)
subject to S ? 0, diag(S) = 1
(3)
Although convexity begets easy optimization in a theoretical sense, the number of variables in the
SDP is quadratic in n. Thus for large SDPs, we actually return to the low-rank parameterization (2).
Solving LRPk via simple gradient methods works extremely well in practice and is partially justified
by theoretical analyses in [8, 12].
1
Proof. WLOG, A ? 0 because adding to its diagonal merely adds a constant term to the IQP objective.
The objective is a convex function, as we can factor A = LLT and write xT LLT x = kLT xk22 , so it must be
maximized over its convex polytope domain at a vertex point.
2
To complete the picture, we need to convert the relaxed solutions X 2 Rn?k into integral solutions
x 2 { 1, 1}n of the original IQP (1). This can be done as follows: draw a vector g 2 Rk on the
unit sphere uniformly at random, project each Xi onto g, and take the sign. Formally, we write
x = rrd(X) to mean xi = sign(Xi ? g) for i = 1, . . . , n. This randomized rounding procedure was
pioneered by [5] to give the celebrated 0.878-approximation of MAX-CUT.
3
Understanding the relaxation-rounding tradeoff
The overall IQP strategy is to first relax the integer problem domain, then round back in to it. The
optimal objective increases in relaxation, but decreases in randomized rounding. How do these effects
compound? To guide our choice of relaxation, we analyze the effect that the rank k in (2) has on the
approximation ratio of rounded versus optimal IQP solutions.
More formally, let x? , X ? , and S ? denote global optima of IQP, of LRPk , and of SDP, respectively.
We can decompose the approximation ratio as follows:
1
E[IQP(rrd(X ? ))]
SDP(S ? )
LRPk (X ? )
E[IQP(rrd(X ? ))]
=
?
?
IQP(x? )
IQP(x? )
SDP(S ? )
LRPk (X ? )
|
{z
} | {z }
| {z }
|
{z
}
approximation ratio
constant
1
tightening ratio T (k)
(4)
rounding ratio R(k)
As k increases from 1, the tightening ratio T (k) increases towards 1 and the rounding ratio R(k)
decreases from 1. In this section, we lower bound T and R each in turn, thus lower-bounding the
approximation ratio as a function of k. Specifically, we show that T (k) reaches 1 at small k and that
R(k) falls as ?2 + ?( k1 ).
In practice, one cannot find X ? for general k with guaranteed efficiency (if we could, we would
simply use LRP1 to directly solve the original IQP). However, Section 5 shows empirically that
simple procedures solve LRPk well for even small k.
3.1
The tightening ratio T (k) increases
We now show that, under the assumption of A ? 0, the tightening ratio T (k) plateaus early and
that it approaches this plateau steadily. Hence, provided k is beyond this saturation point, and large
enough so that an LRPk solver is practically capable of providing near-optimal solutions, there is no
advantage in taking k larger.
First, T (k) is steadily bounded below. The following is a result of [13] (that also gives insight into
the theoretical worst-case hardness of optimizing LRPk ):
Theorem 3.1 ([13]). Fix A ? 0 and let S ? be an optimal SDP solution. There is a randomized
? feasible for LRPk such that E ? [LRPk (X)]
?
algorithm that, given S ? , outputs X
(k) ? SDP(S ? ),
X
where
?
?2
? ?
((k + 1)/2)
1
1
def 2
(k) =
=1
+o
(5)
k
(k/2)
2k
k
For example, (1) =
2
?
= 0.6366, (2) = 0.7854, (3) = 0.8488, (4) = 0.8836, (5) = 0.9054.2
? under any probability distribution, so the exisBy optimality of X ? , LRPk (X ? ) EX? [LRPk (X)]
tence of the algorithm in Theorem 3.1 implies that T (k)
(k).
Moreover, T (k) achieves its maximum of 1 at small k, and hence must strictly exceed the (k) lower
bound early on. We can arrive at this fact by bounding the rank of the SDP-optimal solution S ? .
?
This is because S ? factors into S ? = XX T , where X is in Rn?rank S and must be optimal since
?
LRPrank S ? (X) = SDP(S ). Without consideration of A, the following theorem uniformly bounds
this rank at well below n. The theorem was established independently by [9] and [10]:
?
Theorem 3.2 ([9,
p10]). Fix a weight matrix A. There exists an optimal solution S to SDP (3) such
?
that rank S ? 2n.
2
The function (k) generalizes the constant approximation factor 2/? = (1) with regards to the implications of the unique games conjecture: the authors show that no polynomial time algorithm can, in general,
approximate LRPk to a factor greater than (k) assuming P 6= NP and the UGC.
3
1
1.05
R(k)
lower bound
0.9
0.9
0.85
0.85
0.75
0.75
0.7
0.7
0.65
1
2
3
4
5
6
1600
0.8
0.8
R(k) (blue) is close to it 2/(? (k))
lower bound (red) across the small k.
1500
1400
1300
1200
1
2
3
4
5
6
1100
1
2
3
k
k
(a)
SDP
Max
Mean
Mean+Std
Mean?Std
1700
objective
0.95
0.65
1800
?(k)
T(k)=LRPk/SDP
0.95
objective
rounding ratio
1
4
5
6
k
(b) T?(k) (blue), the empirical tightening ra- (c) Rounded objective values vs. k: optimal
tio, clears its lower bound (k) (red) and hits SDP (cyan), best IQP rounding (green), and
its ceiling of 1 at k = 4.
mean IQP rounding ? (black).
100?100
Figure 1: Plots of quantities analyzed in Section 3, under A 2 R
whose entries are sampled
independently from a unit Gaussian. For this instance, the empirical post-rounding objectives are
shown at the right for completeness.
Hence we know already that the tightening ratio T (k) equals 1 by the time k reaches
p
2n.
Taking A into consideration, we can identify a class of problem instances for which T (k) actually
saturates at even smaller k. This result is especially useful when the rank of the weight matrix A is
known, or even under one?s control, while modeling the underlying optimization task:
Theorem 3.3. If A is symmetric, there is an optimal SDP solution S ? such that rank S ? ? rank A.
A complete proof is in Appendix A.1. Because adding to the diagonal of A is equivalent to merely
adding a constant to the objective of all problems considered, Theorem 3.3 can be strengthened:
Corollary 3.4. For any symmetric weight matrix A, there exists an optimal SDP solution S ? such
that rank S ? ? minu2Rn rank(A + diag(u)).
That is, changes to the diagonal of A that reduce its rank may be applied to improve the bound.
In summary,
p T (k) grows at least as fast as (k), from T (k) = 0.6366 at k = 1 to T (k) = 1 at
k = min{ 2n, minu2Rn rank(A + diag(u))}. This is validated empirically in Figure 1b.
3.2
The rounding ratio R(k) decreases
As the dimension k grows for row vectors Xi in the LRPk problem, the rounding procedure incurs a
larger expected drop in objective value. Fortunately, we can bound this drop. Even more fortunately,
the bound grows no faster than (k), exactly the steady lower bound for T (k). We obtain this result
with an argument based on the analysis of [13]:
Theorem 3.5. Fix a weight matrix A ? 0 and any LRPk -feasible X 2 Rn?k . The rounding ratio for
X is bounded below as
?
? ??
E[IQP(rrd(X))]
2
2
1
1
=
1+
+o
(6)
LRPk (X)
? (k)
?
2k
k
Note that X in the theorem need not be optimal ? the bound applies to whatever solution an LRPk
solver might provide. The proof, given in Appendix section A.1, uses Lemma 1 from [13], which is
based on the theory of positive definite functions on spheres [14]. A decrease in R(k) that tracks the
lower bound is observed empirically in Figure 1a.
In summary, considering only the steady bounds (Theorems 3.1 and 3.5), T will always rise opposite
to R at least at the same rate. Then, the added fact that T plateaus early (Theorem 3.2 and Corollary
3.4) means that T in fact rises even faster.
In practice, we would like to take k beyond 1 as we find that the first few relaxations give the optimizer
an increasing advantage in arriving at a good LRPk solution, close to X ? in objective. The rapid rise
of T relative to R just shown then justifies not taking k much larger if at all.
4
4
Pairwise MRFs, optimization, and inference alternatives
Having understood theoretically how IQP relates to low-rank relaxations, we now turn to MAP
inference and empirical evaluation. We will show that the LRPk objective can be optimized via
a simple interface to the underlying MRF. This interface then becomes the basis for (a) a MAP
inference algorithm based on very low-rank relaxations, and (b) a comparison to two other basic
algorithms for MAP: Gibbs sampling and mean-field variational inference.
A binaryPpairwise Markov
random field (MRF) models a function h over x 2 {0, 1}n given by
P
h(x) = i i (xi ) + i<j ?i,j (xi , xj ), where the i and ?i,j are real-valued functions. The MAP
inference problem asks for the variable assignment x? that maximizes the function h. An MRF
being binary-valued and pairwise allows the arbitrary factor tables i and ?i,j to be transformed
with straightforward algebra into weights A 2 Sn for the IQP. For the complete reduction, see
Appendix A.2.
We make Section 3 actionable by defining the randomized relaxation and rounding (R3 ) algorithm for
MAP via low-rank relaxations. The first step of this algorithm involves optimizing LRPk (2) whose
weight matrix encodes the MRF. In practice, MRFs usually have special structure, e.g., edge sparsity,
factor templates, and Gaussian filters [11]. To develop R3 as a general tool, we provide two interfaces
between the solver and MRF representation, both of which allow users to exploit special structure.
Left-multiplication (x 7! Ax) Assume a function F that implements left matrix multiplication by
the MRF matrix A. This suffices to compute the gradient of the relaxed objective: rX LRPk (X) =
2AX. We can optimize the relaxation using projected gradient ascent (PGA): alternate between
taking gradient steps and projecting back onto the feasible set (unit-normalizing the rows Xi if the
norm exceeds 1); see Algorithm 1. A user supplying a left-multiplication routine can parallelize its
implementation on a GPU, use sparse linear algebra, or efficiently implement a dense filter.
Row-product ((i, x) 7! Ai x) If the function F further provides left multiplication by any row of
A, we can optimize LRPk with coordinate-wise ascent (BCA). Fixing all but the i?th row of X gives
a function linear in Xi whose optimum is Ai X normalized to have unit norm.
Left-multiplication is suitable when one expects to parallelize multiplication, or exploit common
dense structure as with filters. Row product is suitable when one already expects to compute Ax
serially. BCA also eliminates the need for the step size scheme in PGA, thus reducing the number of
calls to the left-multiplication interface if this step size is chosen by line search.
X
random initialization in Rk?n
for t
1 to T do
if parallel then
X
?S k (X + 2?t AX) // Parallel update
else
for i
1 to n do
Xi
?S k (hAi , Xi) // Sweep update
for j
1 to M do
x(j)
sign(Xg), where g is a random vector from unit sphere S k (normalized Gaussian)
Output the x(j) for which the objective (x(j) )T Ax(j) is largest.
Algorithm 1: The full randomized relax-and-round (R3 ) procedure, given a weight matrix A;
?S k (?) is row normalization and ?t is the step size in the t?th iteration.
4.1
Comparison to Gibbs sampling and mean-field
The R3 algorithm affords a tidy comparison to two other basic MAP algorithms. First, it is iterative
and maintains a constant amount of state per MRF variable (a length k row vector). Using the
row-product interface, R3 under BCA sequentially sweeps through and updates each variable?s state
(row Xi ) while holding all others fixed. This interface bears a striking resemblance to (annealed)
Gibbs sampling and mean-field iterative updates [4, 15], which are popular due to their simplicity.
Table 1 shows how both can be implemented via the row-product interface.
5
Algorithm
Domain
Gibbs
Mean-field
R3
n
x 2 { 1, 1}
x 2 [ 1, 1]n
X 2 (S k )n
Sweep update
Parallel update
xi ? ?Z (exp(Ai x))
xi
tanh(Ai x)
Xi
?S k (Ai X)
x ? ?Z (exp(Ax))
x
tanh(Ax)
X
?S k (X + 2?t AX)
Table 1: Iterative updates for MAP algorithms that use constant state per MRF variable. ?S k denotes
`2 unit-normalization of rows and ?Z denotes scaling rows so that they sum to 1. The R3 sweep
update is not a gradient step, but rather the analytic maximum for the i?th row fixing the rest.
x1
1 (x1 )
10x1 x2
= x1
x2
2 (x2 )
= x2
"
1 0
1
A=
2 1
1
0
10
1
10
0
#
Figure 2: Consider the two variable MRF on the left (with x1 , x2 2 { 1, 1} for the factor expressions)
and its corresponding matrix A. Note x0 is clamped to 1 as per the reduction (A.2). The optimum is
x = [1, 1, 1]T with a value of xT Ax = 12. If Gibbs or LRP1 is initialized at x = [1, 1, 1]T , then
either one will be unlikely to transition away from its suboptimal objective value of 8 (as flipping
only one of x1 or x2 decreases the objective to 10). Meanwhile, LRP2 succeeds with probability 1
over random initializations. Suppose X = [1, 0; X1 ; X2 ] with X1 = X2 . Then the gradient update
is X1 = ?S 2 (A1 X) = ?S 2 (([1, 0] + 10X1 )/2), which always points towards X1? = X2? = [1, 0]
except in the 0-probability event that X1 = X2 = [ 1, 0] (corresponding to the poor initialization
of [1, 1, 1]T above). The gradient with respect to X1 at points along the unit circle is shown on
the right. The thick arrow represents an X1 ? [ 0.95, 0.3], and the gradient field shows that it will
iteratively update towards the optimum.
Using left-multiplication, R3 updates the state of all variables in parallel. Superficially, both Gibbs
and the iterative mean-field update can be parallelized in this way as well (Table 1), but doing
so incorrectly alters the their convergence properties. Nonetheless, [11] showed that a simple
modification works well in practice for mean-field, so we consider these algorithms for a complete
comparison.3
While Gibbs, mean-field, and R3 are similar in form, they differ in their per-variable state: Gibbs
maintains a number in { 1, 1} whereas R3 stores an entire vector in Rk . We can see by example
that the extra state can help R3 avoid local optima that ensnarls Gibbs. A single coupling edge in a
two-node MRF, described in Figure 2, gives intuition for the advantage of optimizing relaxations
over stochastic hill-climbing.
Another widely-studied family of MAP inference techniques are based on belief propagation or
relaxations of the marginal polytope [4]. For belief propagation, and even for the most basic of the
LP relaxations (relaxing to the local consistency polytope), one needs to store state for every edge in
addition to every variable. This demands a more complex interface to the MRF, introduces substantial
added bookkeeping for dense graphs, and is not amenable to techniques such as the filter of [11].
5
Experiments
We compare the algorithms from Table 1 on three benchmark MRFs and an additional artificial MRF.
We also show the effect of the relaxation k on the benchmarks in Figure 3.
Rounding in practice The theory of Section 3 provides safeguard guarantees by considering the
average-case rounding. In practice, we do far better than average since we take several roundings and
output the best. Similarly, Gibbs? output is taken as the best along its chain.
Budgets Our goal is to see how efficiently each method utilizes the same fixed budget of queries to
the function, so we fix the number queries to the left-multiplication function F of Section 4. A budget
jointly limits the relaxation updates and the number of random roundings taken in R3 . We charge
3
Later, in [16], the authors derive the parallel mean-field update as being that of a concave approximation to
the cross-entropy term in the true mean-field objective.
6
algo.
Gibbs
MF
R3
Gibbs
MF
R3
Gibbs
MF
R3
Gibbs
MF
R3
parallel sweep parallel sweep
low budget
high budget
seg (50)
8.35 (23)
8.36 (23)
8.39 (15)
7.4 (19)
7.4 (26)
7.4 (17)
7.07 (33)
7.03
(9)
7.09 (23)
6.78 (31)
6.75 (12)
6.8 (25)
dataset [name (# of instances)]
dbn (108)
grid40 (8) chain (300)
1.39
(30) 14.5 (7) .473
(37)
1.3
(7) 13.6 (1) .463
(39)
1.42
(71) 13.7 (0) .538 (296)
.826
(3) .843 (0) .124
(3)
1.16
(6) 11.3 (3)
.35
(50)
1.29
(99) 11.3 (5) .418 (282)
1.26
(42) 12.5 (7) .367
(85)
1.16
(4) 11.7 (1)
.33
(39)
1.28
(62) 11.9 (0) .398 (300)
.814
(2) 1.85 (0) .132
(11)
1.1
(2) 10.9 (2) .259
(47)
1.25 (104)
11 (6) .321 (296)
Table 2: Benchmark performance of algorithms in each comparison regime, in which the benchmarks
are held to different computational budgets that cap their access to the left-multiplication routine. The
score shown is an average relative gain in objective over the uniform-random baseline. Parenthesized
is the win count (including ties), and bold text highlights qualitatively notable successes.
seg
4
580
1.3
LRP
Max
Mean
Mean+Std
Mean?Std
1.25
1.2
objective
objective
540
1.4
LRP
Max
Mean
Mean+Std
Mean?Std
520
1.1
500
1
0.95
1
2
3
4
5
6
1.25
1.2
1.05
460
LRP
Max
Mean
Mean+Std
Mean?Std
1.3
1.15
480
grid
x 10
1.35
objective
560
4
dbn
x 10
1.15
1
2
3
4
k
k
5
6
1.1
1
2
3
4
5
6
k
Figure 3: Relaxed and rounded objectives vs. the rank k in an instance of seg, dbn, and grid40. Blue:
max of roundings. Red: value of LRPk . Black: mean of roundings (? ). The relaxation objective
increases with k, suggesting that increasingly good solutions are obtained by increasing k, in spite of
non-convexity (here we are using parallel updates, i.e. using R3 with PGA). The maximum rounding
also improves considerably with k, especially at first when increasing beyond k = 1.
R3 k-fold per use of F when updating, as it queries F with a k-row argument.4 Sweep methods are
charged once per pass through all variables.
We experiment with separate budgets for the sweep and parallel setup, as sweeps typically converge
more quickly. The benchmark is run under separate low and high budget regimes ? the latter more
than double the former to allow for longer-run effects to set in. In Table 2, the sweep algorithms? low
budget is 84 queries; the high budget is 200. The parallel low budget is 180; the high budget is 400.
We set R3 to take 20 roundings under low budgets and 80 under high ones, and the remaining budget
goes towards LRPk updates.
Datasets Each dataset comprises a family of binary pairwise MRFs. The sets seg, dbn, and grid40
are from the PASCAL 2011 Probabilistic Inference Challenge5 ? seg are small segmentation models
(50 instances, average 230 variables, 622 edges), dbn are deep belief networks (108 instances, average
920 variables, 54160 edges), and grid40 are 40x40 grids (8 instances, 1600 variables, 6240 or 6400
edges) whose edge weights outweigh their unaries by an order of magnitude. The chain set comprises
300 randomly generated 20-node chain MRFs with no unary potentials and random unit-Gaussian
edge weights ? it is principally an extension of the coupling two-node example (Figure 2), and serves
as a structural obverse to grid40 in that it lacks cycles entirely. Among these, the dbn set comprises
the largest and most edge-dense instances.
4
5
This conservatively disfavors R3 , as it ignores the possible speedups of treating length-k vectors as a unit.
http://www.cs.huji.ac.il/project/PASCAL/
7
Evaluation To aggregate across instances of a dataset, we measure the average improvement over
a simple baseline that, subject to the budget constraint, draws uniformly random vectors in { 1, 1}n
and selects the highest-scoring among them. Improvement over the baseline is relative: if z is the
solution objective and z 0 is that of the baseline, (z z 0 )/z 0 is recorded for the average. We also
count wins (including ties), the number of times a method obtains the best objective among the
competition. Baseline performance varies with budget so scores are incomparable across sweep and
parallel experiments.
In all experiments,
we use LRP4 , i.e. the width-4 relaxation. The R3 gradient step size scheme is
p
?t = 1/ t. In the parallel setting, mean-field updates are prone to large oscillations, so we smooth
the update with the current point: x
(1 ?)x + ? tanh(Ax). Our experiments set ? = 0.5. Gibbs
is annealed from an initial temperature of 10 down to 0.1. These settings were tuned towards the
benchmarks using a few arbitrary instances from each dataset.
Results are summarized in Table 2. All methods fare well on the seg dataset and find solutions very
near the apparent global optimum. This shows that the rounding scheme of R3 , though elementary,
is nonetheless capable of recovering an actual MAP point. On grid40, R3 is competitive but not
outstanding, and on chain it is a clear winner. Both datasets have edge potentials that dominate
the unaries, but the cycles in the grid help break local frustrations that occur in chain where they
prevents Gibbs from transitioning. On dbn ? the more difficult task grounded in a real model ? R3
outperforms the others by a large margin.
Figure 3 demonstrates that relaxation beyond the quadratic program maxx2[ 1,1]xT Ax (i.e. k = 1) is
crucial, both for optimizing LRPk and for obtaining a good maximum among roundings. Figure 4 in
the appendix visualizes the distribution of rounded objective values across different instances and
relaxations, illustrating that the difficulty of the problem can be apparent in the rounding distribution.
6
Related work and concluding remarks
In this paper, we studied MAP inference problems that can be cast as an integer quadratic program
over hypercube vertices (IQP). Relaxing the IQP to an SDP (3) and rounding back with rrd(?) was
introduced by Goemans and Williamson in the 1990s for MAX-CUT. It was generalized to positive
semidefinite weights shortly thereafter by Nesterov [6].
Separately, in the early 2000s, there was interest in scalably solving SDPs, though not with the
specific goal of solving the IQP. The low-rank reparameterization of an SDP, as in (2), was developed
by [8] and [12]. Recent work has taken this approach to large-scale SDP formulations of clustering,
embedding, matrix completion, and matrix norm optimization for regularization [17, 18]. Upper
bounds on SDP solutions in terms of problem size n, which help justify using a low rank relaxation,
have been known since the 1990s [9, 10].
The natural joint use of these ideas (IQP relaxed to SDP and SDP solved by low-rank relaxation) is
somewhat known. It was applied in a clustering experiment in [19], but no theoretical analysis was
given and no attention paid to rounding directly from a low-rank solution. The benefit of rounding
from low-rank was noticed in coarse MAP experiments in [20], but no theoretical backing was given
and no attention paid to coordinate-wise ascent or budgeted queries to the underlying model.
Other relaxation hierarchies have been studied in the MRF MAP context, namely linear program
(LP) relaxations given by hierarchies of outer bounds on the marginal polytope [21, 2]. They differ
from this paper?s setting in that they maintain state for every MRF clique configuration ? an approach
that extends beyond pairwise MRFs but that scales with the number of factors (unwieldy versus a
large, dense binary pairwise MRF) and requires fine-grained access to the MRF. Sequences of LP and
SDP relaxations form the Sherali-Adams and Lasserre hierarchies, respectively, whose relationship is
discussed in [4] (Section 9). The LRPk hierarchy sits at a lower level: between the IQP (1) and the
first step of the Lasserre hierarchy (the SDP (3)).
From a practical point of view, we have presented an algorithm very similar in form to Gibbs sampling
and mean-field. This provides a down-to-earth perspective on relaxations within the realm of scalable
and simple inference routines. It would be interesting to see if the low-rank relaxation ideas from this
paper can be adapted to other settings (e.g., for marginal inference). Conversely, the rich literature
on the Lasserre hierarchy may offer guidance in extending the low-rank semidefinite approach (e.g.,
beyond the binary pairwise setting).
8
References
[1] S. Geman and D. Geman. Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of
images. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 6:721?741, 1984.
[2] D. Sontag, T. Meltzer, A. Globerson, Y. Weiss, and T. Jaakkola. Tightening LP relaxations for MAP using
message-passing. In Uncertainty in Artificial Intelligence (UAI), pages 503?510, 2008.
[3] A. Rush, D. Sontag, M. Collins, and T. Jaakkola. On dual decomposition and linear programming
relaxations for natural language processing. In Empirical Methods in Natural Language Processing
(EMNLP), pages 1?11, 2010.
[4] M. Wainwright and M. I. Jordan. Graphical models, exponential families, and variational inference.
Foundations and Trends in Machine Learning, 1:1?307, 2008.
[5] M. Goemans and D. Williamson. Improved approximation algorithms for maximum cut and satisfiability
problems using semidefinite programming. Journal of the ACM (JACM), 42(6):1115?1145, 1995.
[6] Y. Nesterov. Semidefinite relaxation and nonconvex quadratic optimization. Optimization methods and
software, 9:141?160, 1998.
[7] N. Alon and A. Naor. Approximating the cut-norm via Grothendieck?s inequality. SIAM Journal on
Computing, 35(4):787?803, 2006.
[8] S. Burer and R. Monteiro. A nonlinear programming algorithm for solving semidefinite programs via
low-rank factorization. Mathematical Programming, 95(2):329?357, 2001.
[9] A. I. Barvinok. Problems of distance geometry and convex properties of quadratic maps. Discrete &
Computational Geometry, 13:189?202, 1995.
[10] G. Pataki. On the rank of extreme matrices in semidefinite programs and the multiplicity of optimal
eigenvalues. Mathematics of Operations Research, 23(2):339?358, 1998.
[11] P. Kr?ahenb?uhl and V. Koltun. Efficient inference in fully connected CRFs with Gaussian edge potentials.
In Advances in Neural Information Processing Systems (NIPS), 2011.
[12] S. Burer and R. Monteiro. Local minima and convergence in low-rank semidefinite programming. Mathematical Programming, 103(3):427?444, 2005.
[13] J. Bri?et, F. M. d. O. Filho, and F. Vallentin. The positive semidefinite Grothendieck problem with rank
constraint. In Automata, Languages and Programming, pages 31?42, 2010.
[14] I. J. Schoenberg. Positive definite functions on spheres. Duke Mathematical Journal, 9:96?108, 1942.
[15] M. I. Jordan, Z. Ghahramani, T. S. Jaakkola, and L. K. Saul. An introduction to variational methods for
graphical models. Machine Learning, 37:183?233, 1999.
[16] P. Kr?ahenb?uhl and V. Koltun. Parameter learning and convergent inference for dense random fields. In
International Conference on Machine Learning (ICML), pages 513?521, 2013.
[17] B. Kulis, A. C. Surendran, and J. C. Platt. Fast low-rank semidefinite programming for embedding and
clustering. In Artificial Intelligence and Statistics (AISTATS), pages 235?242, 2007.
[18] B. Recht and C. R?e. Parallel stochastic gradient algorithms for large-scale matrix completion. Mathematical
Programming Computation, 5:1?26, 2013.
[19] J. Lee, B. Recht, N. Srebro, J. Tropp, and R. Salakhutdinov. Practical large-scale optimization for maxnorm regularization. In Advances in Neural Information Processing Systems (NIPS), pages 1297?1305,
2010.
[20] S. Wang, R. Frostig, P. Liang, and C. Manning. Relaxations for inference in restricted Boltzmann machines.
In International Conference on Learning Representations (ICLR), 2014.
[21] D. Sontag and T. Jaakkola. New outer bounds on the marginal polytope. In Advances in Neural Information
Processing Systems (NIPS), pages 1393?1400, 2008.
9
| 5261 |@word kulis:1 illustrating:1 polynomial:1 advantageous:1 norm:4 scalably:1 decomposition:1 paid:2 asks:1 incurs:1 thereby:2 tr:6 reduction:2 initial:1 celebrated:1 configuration:1 score:2 sherali:1 tuned:1 outperforms:1 existing:1 current:1 must:3 gpu:1 analytic:1 plot:1 drop:2 update:18 treating:1 v:2 intelligence:3 parameterization:1 supplying:1 completeness:1 provides:3 node:3 coarse:1 sits:1 mathematical:4 along:2 direct:1 koltun:2 naor:1 introduce:1 theoretically:1 x0:1 pairwise:9 ra:1 expected:1 unaries:2 hardness:1 rapid:1 sdp:28 salakhutdinov:1 actual:1 solver:3 considering:2 increasing:3 begin:1 project:2 underlying:5 notation:1 bounded:2 moreover:1 maximizes:1 provided:1 what:1 xx:1 developed:1 guarantee:3 every:3 subclass:1 charge:1 concave:1 tie:3 exactly:1 k2:1 hit:1 demonstrates:1 control:1 unit:11 whatever:1 enjoy:1 platt:1 positive:5 understood:1 local:4 limit:1 parallelize:2 approximately:1 pami:1 black:3 might:1 initialization:3 studied:3 conversely:1 relaxing:2 factorization:1 range:1 unique:1 practical:2 globerson:1 practice:11 implement:3 definite:2 procedure:4 empirical:4 spite:1 get:1 onto:2 cannot:1 close:2 context:1 optimize:2 equivalent:1 map:23 charged:1 outweigh:1 www:1 annealed:3 straightforward:1 go:1 attention:2 independently:2 convex:5 focused:1 lrp1:2 automaton:1 simplicity:2 immediately:1 insight:1 dominate:1 reparameterization:1 classic:1 embedding:2 coordinate:3 schoenberg:1 hierarchy:6 suppose:1 user:3 pioneered:1 programming:9 duke:1 us:1 trend:1 roy:1 updating:1 std:8 cut:5 geman:2 observed:1 wang:2 solved:3 worst:1 seg:6 cycle:2 connected:1 decrease:6 highest:1 substantial:1 intuition:2 vanishes:1 convexity:2 nesterov:2 solving:6 tight:1 algebra:2 algo:1 efficiency:2 basis:1 joint:1 fast:2 tidy:1 artificial:3 query:5 aggregate:1 whose:5 apparent:2 stanford:4 widely:2 solve:5 larger:3 relax:3 otherwise:1 valued:2 statistic:1 jointly:1 advantage:4 sequence:2 eigenvalue:1 interaction:3 product:4 disfavor:1 flexibility:1 adapts:1 competition:1 convergence:2 double:1 optimum:7 r1:1 extending:1 guaranteeing:1 adam:1 help:3 coupling:2 develop:2 derive:1 fixing:2 ac:1 completion:2 alon:1 implemented:1 c:2 involves:1 implies:1 recovering:1 differ:2 thick:1 filter:5 stochastic:4 require:1 fix:4 suffices:1 decompose:1 elementary:1 strictly:1 extension:1 practically:1 considered:1 exp:2 algorithmic:1 achieves:2 early:4 optimizer:1 earth:1 tanh:3 largest:2 tool:1 lrp:4 iqp:28 gaussian:6 always:3 rather:1 avoid:1 jaakkola:4 corollary:2 ax:19 focus:2 klt:1 validated:1 improvement:2 rank:48 contrast:1 baseline:5 sense:1 posteriori:2 inference:22 mrfs:7 unary:1 unlikely:1 entire:1 typically:1 transformed:1 selects:1 backing:1 monteiro:2 overall:1 among:4 dual:1 pascal:2 special:2 uhl:2 marginal:6 field:19 equal:2 having:1 once:1 sampling:6 biology:1 represents:1 icml:1 others:3 np:2 few:2 randomly:1 interpolate:1 geometry:2 maintain:1 interest:1 message:3 evaluation:2 introduces:1 analyzed:1 extreme:1 semidefinite:13 held:1 chain:6 implication:1 amenable:1 edge:12 capable:3 integral:1 unless:1 initialized:1 abundant:1 circle:1 rush:1 guidance:1 theoretical:7 instance:11 column:1 modeling:1 restoration:1 assignment:1 tractability:2 vertex:3 entry:1 expects:2 uniform:1 rounding:29 varies:1 kxi:1 considerably:2 recht:2 explores:1 randomized:7 huji:1 siam:1 international:2 probabilistic:1 lee:1 rounded:6 safeguard:1 quickly:1 frustration:1 recorded:1 successively:1 emnlp:1 admit:1 return:2 suggesting:1 potential:3 bold:1 summarized:1 barvinok:1 notable:1 later:1 break:1 view:1 analyze:1 doing:1 red:3 competitive:1 maintains:2 parallel:13 il:1 efficiently:3 maximized:1 yield:1 identify:1 climbing:2 bayesian:1 sdps:4 rx:1 finer:1 visualizes:1 plateau:3 llt:2 reach:2 definition:3 nonetheless:2 steadily:2 proof:3 recovers:1 sampled:1 gain:1 dataset:5 popular:1 realm:1 cap:1 improves:1 satisfiability:1 segmentation:1 routine:3 actually:2 back:3 higher:1 wei:1 improved:1 formulation:1 done:1 shrink:2 box:1 though:2 just:2 tropp:1 christopher:1 nonlinear:1 propagation:3 lack:1 quality:2 resemblance:1 grows:4 name:1 effect:6 requiring:1 normalized:2 true:1 former:1 hence:3 regularization:2 symmetric:4 iteratively:1 round:3 game:1 width:1 customize:1 steady:3 generalized:1 hill:2 complete:5 percy:1 interface:11 temperature:1 image:1 wise:3 consideration:2 variational:3 common:3 bookkeeping:2 specialized:1 empirically:3 winner:1 discussed:1 fare:1 gibbs:20 ai:5 consistency:1 dbn:7 similarly:1 grid:3 mathematics:1 frostig:2 language:4 access:2 longer:1 add:1 showed:1 recent:1 perspective:1 optimizing:6 compound:1 store:2 nonconvex:1 inequality:1 binary:7 success:1 scoring:2 p10:1 minimum:1 greater:1 relaxed:10 fortunately:2 additional:1 somewhat:1 parallelized:1 filho:1 converge:1 sida:1 relates:1 full:2 reduces:1 pataki:1 x2rn:1 exceeds:1 faster:3 smooth:1 burer:2 offer:2 sphere:5 cross:1 post:1 equally:1 a1:1 mrf:17 basic:3 scalable:1 vision:1 iteration:1 normalization:2 grounded:1 ahenb:2 justified:1 background:1 remarkably:1 whereas:1 addition:1 separately:1 fine:1 else:1 standpoint:1 crucial:1 extra:1 eliminates:1 rest:1 ascent:4 subject:3 jordan:2 integer:5 call:1 structural:1 near:2 exceed:1 easy:1 enough:1 meltzer:1 xj:1 opposite:1 suboptimal:1 reduce:1 incomparable:1 idea:2 tradeoff:1 x40:1 expression:1 sontag:3 passing:2 remark:1 deep:1 generally:2 useful:1 clear:2 amount:1 http:1 outperform:1 affords:1 alters:1 sign:3 track:2 per:6 blue:3 discrete:4 write:3 thereafter:1 indefinite:1 achieving:1 budgeted:1 graph:1 relaxation:46 merely:2 fraction:1 convert:1 sum:1 run:2 uncertainty:1 striking:1 arrive:1 family:3 extends:1 looser:1 utilizes:1 oscillation:1 draw:2 appendix:4 scaling:1 entirely:1 pga:3 bound:19 def:4 guaranteed:1 cyan:1 convergent:1 fold:1 quadratic:9 adapted:1 occur:1 constraint:2 x2:11 software:1 encodes:1 speed:1 argument:2 optimality:2 extremely:2 min:1 concluding:1 relatively:1 gpus:1 inhibits:1 conjecture:1 department:1 speedup:1 bri:1 alternate:1 manning:3 poor:1 sidaw:1 across:4 smaller:1 increasingly:1 lp:4 modification:1 projecting:1 restricted:2 multiplicity:1 principally:1 taken:3 ceiling:1 xk22:1 turn:2 r3:29 count:2 know:1 end:1 serf:1 studying:1 generalizes:1 operation:1 rewritten:1 permit:1 away:1 alternative:1 shortly:1 original:3 actionable:1 denotes:2 remaining:1 clustering:3 graphical:3 exploit:2 k1:1 especially:2 establish:1 approximating:1 hypercube:2 ghahramani:1 sweep:11 objective:36 noticed:1 already:2 quantity:1 added:2 flipping:1 strategy:2 diagonal:4 hai:1 gradient:11 win:2 iclr:1 distance:1 link:1 separate:2 outer:2 polytope:7 assuming:1 length:2 retained:1 relationship:1 ratio:15 providing:1 liang:2 setup:2 difficult:1 holding:1 stated:1 tightening:7 rise:3 implementation:3 motivates:1 boltzmann:1 pliang:1 contributed:1 upper:1 markov:3 datasets:2 benchmark:7 incorrectly:1 defining:1 saturates:2 rn:3 arbitrary:2 introduced:1 namely:2 cast:3 required:1 axx:1 connection:1 optimized:1 established:1 nip:3 beyond:6 below:3 usually:1 pattern:1 regime:2 sparsity:1 program:9 saturation:1 rf:1 including:2 memory:1 max:11 belief:4 green:1 wainwright:1 suitable:2 event:1 natural:4 serially:1 difficulty:1 scheme:3 improve:1 eye:1 picture:1 xg:1 coupled:1 grothendieck:2 sn:4 text:1 review:1 understanding:1 bca:3 literature:1 multiplication:10 relative:4 fully:1 bear:1 highlight:1 interesting:1 srebro:1 versus:2 foundation:1 becomes:1 row:16 prone:1 summary:2 soon:1 arriving:1 formal:1 guide:1 allow:2 fall:1 template:1 taking:4 saul:1 sparse:2 benefit:1 regard:1 dimension:1 transition:1 superficially:1 rich:1 conservatively:1 author:3 qualitatively:1 ignores:1 projected:2 far:1 transaction:1 approximate:1 obtains:1 clique:2 global:2 sequentially:1 uai:1 maxnorm:1 xi:19 search:1 iterative:4 table:8 lasserre:3 ca:1 decoupling:1 parenthesized:1 obtaining:1 williamson:2 complex:2 meanwhile:1 domain:3 diag:3 aistats:1 dense:7 arrow:1 bounding:2 x1:14 crfs:1 strengthened:1 wlog:1 comprises:3 exponential:1 candidate:1 kxk2:1 clamped:1 grained:2 rk:6 theorem:11 down:2 transitioning:1 xt:6 specific:1 unwieldy:1 showing:1 list:1 normalizing:1 exists:2 adding:3 kr:2 magnitude:1 tio:1 justifies:1 budget:16 demand:1 margin:1 gap:1 mf:4 entropy:1 simply:1 jacm:1 prevents:1 partially:1 scalar:1 applies:1 relies:1 acm:1 goal:2 towards:6 rrd:5 replace:1 feasible:3 hard:1 change:1 typical:1 specifically:1 uniformly:3 reducing:1 except:1 justify:1 lemma:1 pas:1 goemans:2 succeeds:1 formally:2 latter:1 collins:1 outstanding:1 ex:1 |
4,707 | 5,262 | Inferring synaptic conductances from spike trains
under a biophysically inspired point process model
E. J. Chichilnisky
Department of Neurosurgery
Hansen Experimental Physics Laboratory
Stanford University
[email protected]
Kenneth W. Latimer
The Institute for Neuroscience
The University of Texas at Austin
[email protected]
Fred Rieke
Department of Physiology and Biophysics
Howard Hughes Medical Institute
University of Washington
[email protected]
Jonathan W. Pillow
Princeton Neuroscience Institute
Department of Psychology
Princeton University
[email protected]
Abstract
A popular approach to neural characterization describes neural responses in terms
of a cascade of linear and nonlinear stages: a linear filter to describe stimulus
integration, followed by a nonlinear function to convert the filter output to spike
rate. However, real neurons respond to stimuli in a manner that depends on the
nonlinear integration of excitatory and inhibitory synaptic inputs. Here we introduce a biophysically inspired point process model that explicitly incorporates
stimulus-induced changes in synaptic conductance in a dynamical model of neuronal membrane potential. Our work makes two important contributions. First, on
a theoretical level, it offers a novel interpretation of the popular generalized linear
model (GLM) for neural spike trains. We show that the classic GLM is a special
case of our conductance-based model in which the stimulus linearly modulates excitatory and inhibitory conductances in an equal and opposite ?push-pull? fashion.
Our model can therefore be viewed as a direct extension of the GLM in which we
relax these constraints; the resulting model can exhibit shunting as well as hyperpolarizing inhibition, and time-varying changes in both gain and membrane time
constant. Second, on a practical level, we show that our model provides a tractable
model of spike responses in early sensory neurons that is both more accurate and
more interpretable than the GLM. Most importantly, we show that we can accurately infer intracellular synaptic conductances from extracellularly recorded
spike trains. We validate these estimates using direct intracellular measurements
of excitatory and inhibitory conductances in parasol retinal ganglion cells. The
stimulus-dependence of both excitatory and inhibitory conductances can be well
described by a linear-nonlinear cascade, with the filter driving inhibition exhibiting opposite sign and a slight delay relative to the filter driving excitation. We
show that the model fit to extracellular spike trains can predict excitatory and inhibitory conductances elicited by novel stimuli with nearly the same accuracy as
a model trained directly with intracellular conductances.
1
Introduction
The point process generalized linear model (GLM) has provided a useful and highly tractable tool
for characterizing neural encoding in a variety of sensory, cognitive, and motor brain areas [1?5].
1
inhibitory filter
nonlinearity Poisson
excitatory filter
post-spike filter
stimulus
spikes
Figure 1: Schematic of conductance-based spiking model.
However, there is a substantial gap between descriptive statistical models like the GLM and more
realistic, biophysically interpretable neural models. Cascade-type statistical models describe input
to a neuron in terms of a set of linear (and sometimes nonlinear) filtering steps [6?11]. Real neurons,
on the other hand, receive distinct excitatory and inhibitory synaptic inputs, which drive conductance
changes that alter the nonlinear dynamics governing membrane potential. Previous work has shown
that excitatory and inhibitory conductances in retina and other sensory areas can exhibit substantially
different tuning. [12, 13].
Here we introduce a quasi-biophysical interpretation of the generalized linear model. The resulting
interpretation reveals that the GLM can be viewed in terms of a highly constrained conductancebased model. We expand on this interpretation to construct a more flexible and more plausible
conductance-based spiking model (CBSM), which allows for independent excitatory and inhibitory
synaptic inputs. We show that the CBSM captures neural responses more accurately than the standard GLM, and allows us to accurately infer excitatory and inhibitory synaptic conductances from
stimuli and extracellularly recorded spike trains.
2
A biophysical interpretation of the GLM
The generalized linear model (GLM) describes neural encoding in terms of a cascade of linear,
nonlinear, and probabilistic spiking stages. A quasi-biological interpretation of GLM is known as
?soft threshold? integrate-and-fire [14?17]. This interpretation regards the linear filter output as a
membrane potential, and the nonlinear stage as a ?soft threshold? function that governs how the
probability of spiking increases with membrane potential, specifically:
Vt
rt
yt |rt
= k> xt
= f (Vt )
? Poiss(rt ?t ),
(1)
(2)
(3)
where k is a linear filter mapping the stimulus xt to the membrane potential Vt at time t, a fixed
nonlinear function f maps Vt to the conditional intensity (or spike rate) rt , and spike count yt is a
Poisson random variable in a time bin of infinitesimal width ?t . The log likelihood is
log p(y1:T |x1:T , k) =
T
X
?rt ?t + yt log(rt ?t ) ? log(yt !).
(4)
t=1
The stimulus vector xt can be augmented to include arbitrary covariates of the response such as the
neuron?s own spike history or spikes from other neurons [2, 3]. In such cases, the output does not
form a Poisson process because spiking is history-dependent.
The nonlinearity f is fixed a priori. Therefore, the only parameters are the coefficients of the filter
k. The most common choice is exponential, f (z) = exp(z), corresponding to the canonical ?log?
link function for Poisson GLMs. Prior work [6] has shown that if f grows at least linearly and at
most exponentially, then the log-likelihood is jointly concave in model parameters ?. This ensures
that the log-likelihood has no non-global maxima, and gradient ascent methods are guaranteed to
find the maximum likelihood estimate.
2
3
Interpreting the GLM as a conductance-based model
A more biophysical interpretation of the GLM can be obtained by considering a single-compartment
neuron with linear membrane dynamics and conductance-based input:
dV
dt
= ?gl V + ge (t)(V ? Ee ) ? gi (t)(V ? Ei )
= ?(gl + ge (t) + gi (t))V + ge (t)Ee + gi (t)Ei
= ?gtot (t)V + Is (t),
(5)
where (for simplicity) we have set the leak current reversal potential to zero. The ?total conductance?
at time t is gtot (t) = gl +ge (t)+gi (t) and the ?effective input current? is Is (t) = ge (t)Ee +gi (t)Ei .
Suppose that the stimulus affects the neuron via the synaptic conductances ge and gi . It is then
natural to ask under which conditions, if any, the above model can correspond to a GLM. The
definition of a GLM requires the solution V (t) to be a linear (or affine) function of the stimulus.
This arises if the two following conditions are met:
1. Total conductance gtot is constant. Thus, for some constant c:
ge (t) + gi (t) = c.
(6)
2. The input Is is linear in x. This holds if we set:
ge (xt )
=
ke > xt + be
gi (xt )
=
k i > x t + bi .
(7)
We can satisfy these two conditions by setting ke = ?ki , so that the excitatory and inhibitory
conductances are driven by equal and opposite linear projections of the stimulus. This allows us to
rewrite the membrane equation (eq. 5):
dV
dt
= ?gtot V + (ke > xt + be )Ee + (ki > xt + bi )Ei
= ?gtot V + ktot > xt + btot ,
(8)
where gtot = gl + be + bi is the (constant) total conductance, ktot = ke Ee + ki Ei , and btot =
be Ee + bi Ei . If we take the initial voltage V0 to be btot , the equilibrium voltage in the absence of a
stimulus, then the solution to this differential equation is
Z t
Vt =
e?gtot (t?s) ktot > xs ds + btot
0
=
kleak ? (ktot > xt ) + btot
=
kglm > xt + btot ,
(9)
where kleak ? (ktot > xt ) denotes linear convolution of the exponential decay ?leak? filter kleak (t) =
e?gtot t with the linearly projected stimulus train, and kglm = ktot ? kleak is the ?true? GLM filter
(from eq. 1) that results from temporally convolving the conductance filter with the leak filter. Since
the membrane potential is a linear (affine) function of the stimulus (as in eq. 1), the model is clearly
a GLM.
Thus, to summarize, the GLM can be equated with a synaptic conductance-based dynamical model
in which the GLM filter k results from a common linear filter driving excitatory and inhibitory
synaptic conductances, blurred by convolution with an exponential leak filter determined by the
total conductance.
4
Extending GLM to a nonlinear conductance-based model
From the above, it is easy to see how to create a more realistic conductance-based model of neural
responses. Such a model would allow the stimulus tuning of excitation and inhibition to differ (i.e.,
allow ke 6= ?ki ), and would include a nonlinear relationship between x and the conductances to
3
preclude negative values (e.g., using a rectifying nonlinearity). As with the GLM, we assume that
the only source of stochasticity on the model is in the spiking mechanism: we place no additional
noise on the conductances or the voltage. This simplifying assumption allows us to perform efficient
maximum likelihood inference using standard gradient ascent methods.
We specify the membrane potential of the conductance-based point process model as follows:
dV
= ge (t)(Ee ? V ) + gi (t)(Ei ? V ) + gl (El ? V ),
dt
ge (t) = fe (ke > xt ),
gi (t) = fi (ki > xt ),
(10)
(11)
where fe and fi are nonlinear functions ensuring positivity of the synaptic conductances. In practice,
we evaluate V along a discrete lattice of points (t = 1, 2, 3, . . . T ) of width ?t . Assuming ge and gi
remain constant within each bin, the voltage equation becomes a simple linear differential equation
with the solution
Is (t)
Is (t)
?gtot (t)?t
+
(12)
V (t + 1) = e
V (t) ?
gtot (t)
gtot (t)
V (1) = El
(13)
gtot (t) = ge (t) + gi (t) + gl
(14)
Is (t) = ge (t)Ee + gi (t)Ei + gl El
(15)
The mapping from membrane potential to spiking is similar to that in the standard GLM (eq. 3):
rt
= f (V (t))
(V ? VT )
f (V ) = exp
VS
yt |rt ? Poiss(rt ?t ).
(16)
(17)
(18)
The voltage-to-spike rate nonlinearity f follows the form proposed by Mensi et al. [17], where VT
is a soft spiking threshold and VS determines the steepness of the nonlinearity. To account for
refractory periods or other spike-dependent behaviors, we simply augment the function to include a
GLM-like spike history term:
(V ? VT )
> hist
f (V ) = exp
+h y
(19)
VS
Spiking activity in real neurons influences both the membrane potential and the output nonlinearity.
We could include additional conductance terms that depend on either stimuli or spike history, such as
an after hyper-polarization current; this provides one direction for future work. For spatial stimuli,
the model can include a set of spatially distinct rectified inputs (e.g., as employed in [9]).
To complete the model, we must select a form for the conductance nonlinearities fe and fi . Although
we could attempt to fit these functions (e.g., as in [9, 18]), we fixed them to be the soft-rectifying
function:
fe (?), fi (?) = log(1 + exp(?)).
(20)
Fixing these nonlinearities improved the speed and robustness of maximum likelihood parameter
fitting. Moreover, we examined intracellularly recorded conductances and found that the nonlinear
mapping from linearly projected stimuli to conductance was well described by this function (see
Fig. 4).
The model parameters we estimate are {ke , ki , be , bi , h, gl , El }. We set the remaining model parameters to biologically plausible values: VT = ?70mV, VS = 4mV, Ee = 0mV, and Ei = ?80mV .
To limit the total number of parameters, we fit the linear filters ke and ki using a basis consisting of
12 raised cosine functions, and we used 10 raised cosine functions for the spike history filter [3].
The log-likelihood function for this model is not concave in the model parameters, which increases
the importance to selecting a good initialization point. We initialized the parameters by fitting a
simplified model which had only one conductance. We initialized the leak terms as El = ?70mV
and gl = 200. We assumed a single synaptic conductance with a linear stimulus dependence,
glin (t) = klin > xt (note that this allows for negative conductance values). We initialized this filter
4
filter fits
A
B
?5.27
x 104 fit to
Actual
test data
?0.4
log likelihood
0
?0.8
C
30
L2 error
weight
0.4
estimated filter errors
?5.29
20
?5.31
10
50
100 150 200
time (ms)
0
0
5
10
minutes of training data
?5.33
0
5
10
minutes of training data
Figure 2: Simulation results. (A) Estimates (solid traces) of excitatory (blue) and inhibitory (red) stimulus
filters from 10 minutes of simulated data. (Dashed lines indicate true filters). (B) The L2 norm between
the estimated input filters and the true filters (calculated in the low-dimensional basis) as a function of the
amount of training data. (C) The log-likelihood of the fit CBSM on withheld test data converges to the log
likelihood of the true model.
the GLM fit, and then numerically maximized the likelihood for klin . We then initialized the parameters for the complete model using ke = cklin and ki = ?cklin , where 0 < c ? 1, thereby
exploiting the mapping between the GLM and the CBSM. Although this initialization presumes that
excitation and inhibition have nearly opposite tuning, we found that standard optimization methods successfully converged to the true model parameters even when ke and ki had similar tuning
(simulation results not shown).
5
Results: simulations
To examine the estimation performance, we fit spike train data simulated from a CBSM with known
parameters (see Fig. 2). The simulated data qualitatively mimicked experimental datasets, with input
filters selected to reproduce the stimulus tuning of macaque ON parasol RGCs. The stimulus consisted of a one dimensional white noise signal, binned at a 0.1ms resolution, and filtered with a low
pass filter with a 60Hz cutoff frequency. The simulated cell produced a firing rate of approximately
32spikes/s. We validated our maximum likelihood fitting procedure by examining error in the fitted
parameters, and evaluating the log-likelihood on a held out five-minute test set. With increasing
amounts of training data, the parameter estimates converged to the true parameters, despite the fact
that the model does not have the concavity guarantees of the standard GLM.
To explore the CBSM?s qualitative response properties, we performed simulated experiments using
stimuli with varying statistics (see Fig. 3). We simulated spike responses from a CBSM with
fixed parameters to stimuli with different standard deviations. We then separately fit responses from
each simulation with a standard GLM. The fitted GLM filters exhibit shifts in both peak height
and position for stimuli with different variance. This suggests that the CBSM can exhibit gain
control effects that cannot be captured by a classic GLM with a spike history filter and exponential
nonlinearity.
6
Results: neural data
We fit the CBSM to spike trains recorded from 7 macaque ON parasol RGCs [12]. The spike trains
were obtained by cell attached recordings in response to full-field, white noise stimuli (identical to
the simulations above). Either 30 or 40 trials were recorded from each cell, using 10 unique 6 second
stimuli. After the spike trains were recorded, voltage clamp recordings were used to measure the
excitatory and inhibitory conductances to the same stimuli. We fit the model using the spike trains
for 9 of the stimuli, and the remaining trials were used to test model fit. Thus, the models were
effectively trained using 3 or 4 repeats of 54 seconds of full-field noise stimulus. We compared the
intracellular recordings to the ge and gi estimated from the CBSM (Fig. 5). Additionally, we fit the
measured conductances with the linear-nonlinear cascade model from the CBSM (the terms ge and
5
A
?lters at different contrasts
B
experimental data
0.25x contrast
0.5x
1x
2x
0.03
weight
0.02
0.01
0
?0.01
(Chander & Chichilnisky, 2001)
0
50
100
150
200
time (ms)
Figure 3: Qualitative illustration of model?s capacity to exhibit contrast adaptation (or gain control). (A)
The GLM filters fit to a fixed CBSM simulated at various levels of stimulus variance. (B) Filters fit to two
real retinal ganglion cells at two different levels of contrast (from [19]).
excitatory
measured conductance
50
40
40
data mean
30
30
20
20
10
10
0
0
?10
?30
-15
0
inhibitory
50
15
?10
30
?40
?20
0
20
40
filter output
Figure 4: Measured conductance vs. output of a fitted linear stimulus filter (gray points), for both the
excitatory (left) and inhibitory (right) conductances. The green diamonds correspond to a non-parametric
estimate of the conductance nonlinearity, given by the mean conductance for each bin of filter output. For
both conductances, the function is is well described by a soft-rectifying function (black trace).
gi in eq. 11) with a least-squares fit as an upper bound measure for the best possible conductance
estimate given our model. The CBSM correctly determined the stimulus tuning for excitation and
inhibition for these cells: inhibition is oppositely tuned and slightly delayed from excitation.
For the side-by-side comparison shown in Fig. 5, we introduced a scaling factor in the estimated
conductances in order to compare the conductances estimated from spike trains against recorded
conductances. Real membrane voltage dynamics depend on the capacitance of the membrane, which
we do not include because it introduces an arbitrary scaling factor that cannot be estimated by spike
alone. Therefore, for comparisons we chose a scaling factor for each cell independently. However,
we used a single scaling for the inhibitory and excitatory conductances. Additionally, we often had
2 or 3 repeated trials of the withheld stimulus, and we compared the model prediction to the average
conductance recorded for the stimulus. The CBSM predicted the synaptic conductances with an
average r2 = 0.54 for the excitatory and an r2 = 0.39 for the inhibitory input from spike trains,
compared to an average r2 = 0.72 and r2 = 0.59 for the excitatory and inhibitory conductances respectively from the least-squares fit directly to the conductances (Fig. 6). To summarize, using only
a few minutes of spiking data, the CBSM could account for 71% of the variance of the excitatory
input and 62% of the inhibitory input that can possibly be explained using the LN cascade model of
the conductances (eq. 11).
One challenge we discovered when fitting the model to real spike trains was that one filter, typically
ki , would often become much larger than the other filter. This resulted in one conductance becoming
dominant, which the intracellular recordings indicated was not the case. This was likely due to the
fact that we are data-limited when dealing with intracellular recordings: the spike train recordings
include only 1 minute of unique stimulus. To alleviate this problem, we added a penalty term, ?, to
6
Example Cell 1
estimated filters
0.2
(conductances)
(spikes)
fit to spikes:
ge
0.1
weight
estimated conductances fit to conductance:
0
?0.1
(conductances)
(spikes)
gi
10nS
?0.2
?0.3
fit to conductance:
fit to spikes:
0
50
100 150 200
time (ms)
250ms
Example Cell 2
estimated filters
0.2
ge
0.1
weight
fit to conductance:
fit to spikes:
0
10nS
?0.1
?0.2
fit to conductance:
fit to spikes:
gi
0
50
100 150 200
time (ms)
250ms
Figure 5: Two example ON parasol RGC responses to a full-field noise stimulus fit with the CBSM. The
model parameters were fit to spike train data, and then used to predict excitatory and inhibitory synaptic
currents recorded separately in response to novel stimuli. For comparison, we show predictions of an LN
model fit directly to the conductance data. Left: Linear kernels for the excitatory (blue) and inhibitory
(red) inputs estimated from the conductance-based model (light red, light blue) and estimated by fitting a
linear-nonlinear model directly to the measured conductances (dark red, dark blue). The filters represent a
combination of events that occur in the retinal circuitry in response to a visual stimulus, and are primarily
shaped by the cone transduction process. Right: Conductances predicted by our model on a withheld test
stimulus. Measured conductances (black) are compared to the predictions from the CBSM filters (fit to
spiking data) and an LN model (fit to conductance data).
the log likelihood on the difference of the L2 norms of ke and ki :
2
?(ke , ki ) = ? ||ke ||2 ? ||ki ||2
(21)
This differentiable penalty ensures that the model will not rely too strongly on one filter over the
other, without imposing any prior on the shape of the filters (with ? = 0.05). We note that unlike
the a typical situation with statistical models that contain more abstract parameters, the terms we
wish to regularize can be measured with intracellular recordings. Future work with this model could
include more informative, data-driven priors on ke and ki .
Finally, we fit the CBSM and GLM to a population of nine extracellularly recorded macaque RGCs
in response to a full-field binary noise stimulus [20]. We used a five minute segment for model
fitting, and compared predicted spike rate using a 6s test stimulus for which we had repeated trials.
7
Excitation prediction
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0
Inhibition prediction
1
fit?to?conductance r2
fit?to?conductance r2
1
0.2
0
0.2
0.4
0.6
spike fit r2
0.8
1
0
0
0.2
0.4
0.6
0.8
spike fit r2
1
Figure 6: Summary of the CBSM fits to 7 ON parasol RGCs for which we had both spike train and
conductance recordings. The axes show model?s ability to predict the excitatory (left) and inhibitory (right)
inputs to a new stimulus in terms of r2 . The CBSM fit is compared against predictions of an LN model fit
directly to measured conductances.
A
B
GLM:
CBSM:
50 spks/s
Conductance Model
1
Rate prediction performance
0.9
0.8
on cell
0.7
0.6
0.5
0.4
0.4
250ms
off cell
0.6
GLM
0.8
1
Figure 7: (A) Performance on spike rate (PSTH) prediction. The true rate (black) was estimated using 167
repeat trials. The GLM prediction is in blue and the CBSM is in red. The PSTHs were smoothed with a
Gaussian kernel with a 1ms standard deviation. (B) Spike rate prediction performance for the population
of 9 cells. The red circle indicates cell used in left plot.
The CBSM achieved a 0.08 higher average r2 in PSTH prediction performance compared to the
GLM. All nine cells showed an improved fit with the CBSM.
7
Discussion
The classic GLM is a valuable tool for describing the relationship between stimuli and spike responses. However, the GLM describes this map as a mathematically convenient linear-nonlinear
cascade, which does not take account of the biophysical properties of neural processing. Here we
have shown that the GLM may be interpreted as a biophysically inspired, but highly constrained,
synaptic conductance-based model. We proposed a more realistic model of the conductance, removing the artificial constraints present in the GLM interpretation, which results in a new, more accurate
and more flexible conductance-based point process model for neural responses. Even without the
benefit of a concave log-likelihood, numerical optimization methods provide accurate estimates of
model parameters.
Qualitatively, the CBSM has a stimulus-dependent time constant, which allows it change gain as a
function of stimulus statistics (e.g., contrast), an effect that cannot be captured by a classic GLM. The
model also allows the excitatory and inhibitory conductances to be distinct functions of the sensory
stimulus, as is expected in real neurons. We demonstrate that the CBSM not only achieves improved
performance as a phenomenological model of neural encoding compared to the GLM, the model
accurately estimates the tuning of the excitatory and inhibitory synaptic inputs to RGCs purely from
measured spike times. As we move towards more naturalistic stimulus conditions, we believe that
the conductance-based approach will become a valuable tool for understanding the neural code in
sensory systems.
8
References
[1] K. Harris, J. Csicsvari, H. Hirase, G. Dragoi, and G. Buzsaki. Organization of cell assemblies
in the hippocampus. Nature, 424:552?556, 2003.
[2] W. Truccolo, U. T. Eden, M. R. Fellows, J. P. Donoghue, and E. N. Brown. A point process
framework for relating neural spiking activity to spiking history, neural ensemble and extrinsic
covariate effects. J. Neurophysiol, 93(2):1074?1089, 2005.
[3] J. W. Pillow, J. Shlens, L. Paninski, A. Sher, A. M. Litke, E. J. Chichilnisky, and E. P. Simoncelli. Spatio-temporal correlations and visual signaling in a complete neuronal population.
Nature, 454:995?999, 2008.
[4] S. Gerwinn, J. H. Macke, and M. Bethge. Bayesian inference for generalized linear models for
spiking neurons. Frontiers in Computational Neuroscience, 2010.
[5] I. H. Stevenson, B. M. London, E. R. Oby, N. A. Sachs, J. Reimer, B. Englitz, S. V. David,
S. A. Shamma, T. J. Blanche, K. Mizuseki, A. Zandvakili, N. G. Hatsopoulos, L. E. Miller,
and K. P. Kording. Functional connectivity and tuning curves in populations of simultaneously
recorded neurons. PLoS Comput Biol, 8(11):e1002775, 2012.
[6] L. Paninski. Maximum likelihood estimation of cascade point-process neural encoding models.
Network: Computation in Neural Systems, 15:243?262, 2004.
[7] D. A. Butts, C. Weng, J. Jin, J.M. Alonso, and L. Paninski. Temporal precision in the visual pathway through the interplay of excitation and stimulus-driven suppression. J Neurosci,
31(31):11313?11327, Aug 2011.
[8] B Vintch, A Zaharia, J A Movshon, and E P Simoncelli. Efficient and direct estimation of
a neural subunit model for sensory coding. In Adv. Neural Information Processing Systems
(NIPS*12), volume 25, Cambridge, MA, 2012. MIT Press. To be presented at Neural Information Processing Systems 25, Dec 2012.
[9] J. M. McFarland, Y. Cui, and D. A. Butts. Inferring nonlinear neuronal computation based on
physiologically plausible inputs. PLoS computational biology, 9(7):e1003143, January 2013.
[10] Il M. Park, Evan W. Archer, Nicholas Priebe, and Jonathan W. Pillow. Spectral methods for
neural characterization using generalized quadratic models. In Advances in Neural Information
Processing Systems 26, pages 2454?2462, 2013.
[11] L. Theis, A. M. Chagas, D. Arnstein, C. Schwarz, and M. Bethge. Beyond glms: A generative
mixture modeling approach to neural system identification. PLoS Computational Biology, Nov
2013. in press.
[12] P. K. Trong and F. Rieke. Origin of correlated activity between parasol retinal ganglion cells.
Nature neuroscience, 11(11):1343?51, November 2008.
[13] C. Poo and J. S. Isaacson. Odor representations in olfactory cortex: ?sparse? coding, global
inhibition, and oscillations. Neuron, 62(6):850?61, June 2009.
[14] H. E. Plesser and W. Gerstner. Noise in integrate-and-fire neurons: from stochastic input to
escape rates. Neural Comput, 12(2):367?384, Feb 2000.
[15] W. Gerstner. A framework for spiking neuron models: The spike response model. In F. Moss
and S. Gielen, editors, The Handbook of Biological Physics, volume 4, pages 469?516, 2001.
[16] L. Paninski, J. W. Pillow, and J. Lewi. Statistical models for neural encoding, decoding, and
optimal stimulus design. Progress in brain research, 165:493?507, January 2007.
[17] S. Mensi, R. Naud, and W. Gerstner. From stochastic nonlinear integrate-and-fire to generalized linear models. In NIPS, pages 1377?1385, 2011.
[18] M. B. Ahrens, L. Paninski, and M. Sahani. Inferring input nonlinearities in neural encoding
models. Network: Computation in Neural Systems, 19(1):35?67, January 2008.
[19] D. Chander and E. J. Chichilnisky. Adaptation to Temporal Contrast in Primate and Salamander
Retina. The Journal of Neuroscience, 21(24):9904?16, December 2001.
[20] J. W. Pillow, L. Paninski, V. J. Uzzell, E. P. Simoncelli, and E. J. Chichilnisky. Prediction and
decoding of retinal ganglion cell responses with a probabilistic spiking model. The Journal of
neuroscience, 25(47):11003?11013, November 2005.
9
| 5262 |@word trial:5 norm:2 hippocampus:1 simulation:5 simplifying:1 thereby:1 solid:1 initial:1 selecting:1 tuned:1 current:4 must:1 realistic:3 numerical:1 hyperpolarizing:1 informative:1 shape:1 motor:1 plot:1 interpretable:2 v:5 alone:1 generative:1 selected:1 filtered:1 characterization:2 provides:2 psth:2 five:2 height:1 along:1 direct:3 differential:2 become:2 qualitative:2 fitting:6 pathway:1 olfactory:1 introduce:2 manner:1 expected:1 behavior:1 examine:1 brain:2 inspired:3 actual:1 preclude:1 considering:1 increasing:1 becomes:1 provided:1 moreover:1 interpreted:1 substantially:1 guarantee:1 temporal:3 fellow:1 concave:3 control:2 medical:1 limit:1 despite:1 encoding:6 firing:1 becoming:1 approximately:1 black:3 chose:1 initialization:2 examined:1 suggests:1 limited:1 shamma:1 bi:5 practical:1 unique:2 hughes:1 practice:1 lewi:1 signaling:1 procedure:1 evan:1 area:2 physiology:1 cascade:8 projection:1 glin:1 convenient:1 naturalistic:1 cannot:3 chander:2 influence:1 map:2 yt:5 poo:1 independently:1 ke:14 resolution:1 simplicity:1 importantly:1 shlens:1 pull:1 regularize:1 classic:4 rieke:3 population:4 suppose:1 origin:1 intracellularly:1 capture:1 ensures:2 adv:1 spks:1 plo:3 valuable:2 hatsopoulos:1 substantial:1 leak:5 covariates:1 dynamic:3 trained:2 depend:2 rewrite:1 segment:1 purely:1 basis:2 neurophysiol:1 various:1 train:17 distinct:3 describe:2 effective:1 london:1 artificial:1 oby:1 hyper:1 stanford:2 plausible:3 larger:1 relax:1 ability:1 statistic:2 gi:17 jointly:1 interplay:1 descriptive:1 differentiable:1 biophysical:4 clamp:1 adaptation:2 buzsaki:1 validate:1 exploiting:1 extending:1 converges:1 chaga:1 fixing:1 measured:8 progress:1 aug:1 eq:6 predicted:3 indicate:1 met:1 exhibiting:1 differ:1 direction:1 filter:43 stochastic:2 bin:3 truccolo:1 alleviate:1 biological:2 mathematically:1 extension:1 frontier:1 hold:1 exp:4 equilibrium:1 mapping:4 predict:3 circuitry:1 driving:3 achieves:1 early:1 estimation:3 hansen:1 utexas:1 schwarz:1 create:1 successfully:1 tool:3 neurosurgery:1 clearly:1 mit:1 gaussian:1 ej:1 poi:2 varying:2 voltage:7 validated:1 ax:1 june:1 likelihood:16 indicates:1 salamander:1 contrast:6 litke:1 suppression:1 inference:2 dependent:3 el:5 plesser:1 typically:1 expand:1 quasi:2 reproduce:1 archer:1 flexible:2 augment:1 priori:1 constrained:2 integration:2 trong:1 special:1 spatial:1 raised:2 equal:2 construct:1 field:4 shaped:1 washington:2 reimer:1 btot:6 identical:1 biology:2 park:1 nearly:2 alter:1 future:2 stimulus:52 escape:1 few:1 retina:2 primarily:1 mizuseki:1 simultaneously:1 resulted:1 delayed:1 consisting:1 fire:3 attempt:1 conductance:80 organization:1 highly:3 introduces:1 weng:1 mixture:1 light:2 held:1 accurate:3 initialized:4 circle:1 theoretical:1 fitted:3 soft:5 modeling:1 lattice:1 deviation:2 delay:1 examining:1 too:1 peak:1 probabilistic:2 physic:2 off:1 decoding:2 bethge:2 connectivity:1 recorded:11 latimer:1 possibly:1 positivity:1 cognitive:1 convolving:1 macke:1 presumes:1 account:3 potential:10 nonlinearities:3 stevenson:1 parasol:6 retinal:5 coding:2 coefficient:1 blurred:1 satisfy:1 explicitly:1 mv:5 depends:1 performed:1 extracellularly:3 red:6 elicited:1 rectifying:3 contribution:1 square:2 compartment:1 accuracy:1 il:1 variance:3 maximized:1 correspond:2 ensemble:1 miller:1 biophysically:4 bayesian:1 identification:1 accurately:4 produced:1 drive:1 rectified:1 history:7 converged:2 synaptic:16 definition:1 infinitesimal:1 against:2 frequency:1 gain:4 popular:2 ask:1 mensi:2 higher:1 dt:3 oppositely:1 response:17 specify:1 improved:3 strongly:1 governing:1 stage:3 correlation:1 glms:2 hand:1 d:1 ei:9 nonlinear:18 gtot:12 gray:1 indicated:1 grows:1 believe:1 effect:3 rgcs:5 true:7 consisted:1 contain:1 brown:1 polarization:1 spatially:1 laboratory:1 white:2 width:2 excitation:7 cosine:2 m:9 generalized:7 isaacson:1 complete:3 demonstrate:1 interpreting:1 novel:3 fi:4 common:2 psths:1 functional:1 spiking:16 refractory:1 exponentially:1 attached:1 volume:2 interpretation:9 slight:1 relating:1 numerically:1 measurement:1 cambridge:1 imposing:1 tuning:8 nonlinearity:8 stochasticity:1 had:5 phenomenological:1 cortex:1 inhibition:8 v0:1 feb:1 dominant:1 own:1 showed:1 driven:3 gerwinn:1 binary:1 vt:9 captured:2 additional:2 employed:1 period:1 dashed:1 signal:1 full:4 simoncelli:3 infer:2 offer:1 post:1 shunting:1 biophysics:1 schematic:1 ensuring:1 prediction:12 poisson:4 sometimes:1 kernel:2 represent:1 achieved:1 cell:17 dec:1 receive:1 separately:2 source:1 unlike:1 ascent:2 induced:1 hz:1 recording:8 december:1 incorporates:1 ee:9 easy:1 variety:1 affect:1 fit:39 psychology:1 blanche:1 opposite:4 zandvakili:1 donoghue:1 texas:1 shift:1 penalty:2 movshon:1 nine:2 useful:1 governs:1 amount:2 dark:2 canonical:1 inhibitory:25 ahrens:1 sign:1 neuroscience:6 estimated:12 correctly:1 extrinsic:1 hirase:1 blue:5 discrete:1 steepness:1 threshold:3 eden:1 cutoff:1 kenneth:1 convert:1 cone:1 respond:1 place:1 oscillation:1 scaling:4 ki:14 bound:1 guaranteed:1 followed:1 quadratic:1 activity:3 binned:1 occur:1 constraint:2 speed:1 extracellular:1 department:3 combination:1 cui:1 membrane:14 describes:3 remain:1 slightly:1 biologically:1 primate:1 dv:3 explained:1 glm:41 ln:4 equation:4 describing:1 count:1 mechanism:1 ge:17 tractable:2 reversal:1 spectral:1 nicholas:1 mimicked:1 robustness:1 odor:1 denotes:1 remaining:2 include:8 assembly:1 klin:2 move:1 capacitance:1 added:1 naud:1 spike:47 parametric:1 dependence:2 rt:9 exhibit:5 gradient:2 link:1 simulated:7 capacity:1 alonso:1 dragoi:1 assuming:1 code:1 relationship:2 illustration:1 fe:4 trace:2 negative:2 priebe:1 design:1 perform:1 diamond:1 upper:1 neuron:15 convolution:2 datasets:1 howard:1 withheld:3 jin:1 november:2 january:3 situation:1 subunit:1 y1:1 discovered:1 smoothed:1 arbitrary:2 intensity:1 introduced:1 david:1 chichilnisky:5 csicsvari:1 macaque:3 nip:2 beyond:1 mcfarland:1 dynamical:2 summarize:2 challenge:1 green:1 event:1 natural:1 rely:1 temporally:1 sher:1 moss:1 sahani:1 prior:3 understanding:1 l2:3 theis:1 relative:1 filtering:1 zaharia:1 integrate:3 lters:1 affine:2 editor:1 austin:1 excitatory:25 summary:1 gl:9 repeat:2 side:2 allow:2 institute:3 characterizing:1 sparse:1 benefit:1 regard:1 curve:1 calculated:1 vintch:1 fred:1 pillow:6 evaluating:1 concavity:1 sensory:6 equated:1 qualitatively:2 projected:2 simplified:1 kording:1 nov:1 dealing:1 global:2 butt:2 reveals:1 hist:1 handbook:1 assumed:1 spatio:1 physiologically:1 additionally:2 nature:3 gerstner:3 sachs:1 linearly:4 intracellular:7 neurosci:1 noise:7 repeated:2 x1:1 neuronal:3 augmented:1 fig:6 e1003143:1 fashion:1 transduction:1 n:2 precision:1 inferring:3 position:1 wish:1 exponential:4 comput:2 minute:7 removing:1 xt:15 covariate:1 r2:10 x:1 decay:1 effectively:1 modulates:1 importance:1 push:1 gap:1 simply:1 explore:1 likely:1 ganglion:4 paninski:6 visual:3 gielen:1 determines:1 harris:1 ma:1 conditional:1 viewed:2 towards:1 absence:1 change:4 specifically:1 determined:2 typical:1 total:5 pas:1 experimental:3 select:1 rgc:1 uzzell:1 arises:1 jonathan:2 evaluate:1 princeton:3 biol:1 correlated:1 |
4,708 | 5,263 | Low-dimensional models of neural population activity
in sensory cortical circuits
Evan Archer1,2 , Urs K?oster3 , Jonathan Pillow4 , Jakob H. Macke1,2
1
Max Planck Institute for Biological Cybernetics, T?ubingen
2
Bernstein Center for Computational Neuroscience, T?ubingen
3
Redwood Center for Theoretical Neuroscience, University of California at Berkeley
4
Princeton Neuroscience Institute, Department of Psychology, Princeton University
[email protected], [email protected]
[email protected], [email protected]
Abstract
Neural responses in visual cortex are influenced by visual stimuli and by ongoing spiking activity in local circuits. An important challenge in computational
neuroscience is to develop models that can account for both of these features in
large multi-neuron recordings and to reveal how stimulus representations interact
with and depend on cortical dynamics. Here we introduce a statistical model of
neural population activity that integrates a nonlinear receptive field model with a
latent dynamical model of ongoing cortical activity. This model captures temporal
dynamics and correlations due to shared stimulus drive as well as common noise.
Moreover, because the nonlinear stimulus inputs are mixed by the ongoing dynamics, the model can account for a multiple idiosyncratic receptive field shapes with
a small number of nonlinear inputs to a low-dimensional dynamical model. We
introduce a fast estimation method using online expectation maximization with
Laplace approximations, for which inference scales linearly in both population
size and recording duration. We test this model to multi-channel recordings from
primary visual cortex and show that it accounts for neural tuning properties as
well as cross-neural correlations.
1
Introduction
Neurons in sensory cortices organize into highly-interconnected circuits that share common input,
dynamics, and function. For example, across a cortical column, neurons may share stimulus dependence as a result of sampling the same location of visual space, having similar orientation
preference [1] or receptive fields with shared sub-units [2]. As a result, a substantial fraction of
stimulus-information can be redundant across neurons [3]. Recent advances in electrophysiology
and functional imaging allow us to simultaneously probe the responses of the neurons in a column.
However, the high dimensionality and (relatively) short duration of the resulting data renders analysis a difficult statistical problem.
Recent approaches to modeling neural activity in visual cortex have focused on characterizing the responses of individual neurons by linearly projecting the stimulus on a small feature subspace that optimally drives the cell [4, 5]. Such ?systems-identification? approaches seek to describe the stimulusselectivity of single neurons separately, treating each neuron as an independent computational unit.
Other studies have focused on providing probabilistic models of the dynamics of neural populations,
seeking to elucidate the internal dynamics underlying neural responses [6, 7, 8, 9, 10, 11]. These
approaches, however, typically do not model the effect of the stimulus (or do so using only a linear
stimulus drive). To realize the potential of modern recording technologies and to progress our un1
derstanding of neural population coding, we need methods for extracting both the features that drive
a neural population and the resulting population dynamics [12].
We propose the Quadratic Input Latent Dynamical System (QLDS) model, a statistical model that
combines a low-dimensional representation of population dynamics [9] with a low-dimensional description of stimulus selectivity [13]. A low-dimensional dynamical system governs the population
response, and receives a nonlinear (quadratic) stimulus-dependent input. We model neural spike
responses as Poisson (conditional on the latent state), with exponential firing rate-nonlinearities. As
a result, population dynamics and stimulus drive interact multiplicatively to modulate neural firing. By modeling dynamics and stimulus dependence, our method captures correlations in response
variability while also uncovering stimulus selectivity shared across a population.
linear
filters
linear
dynamics
quadratic
+
population
nonlinear
function noise
spike
response
A
stimulus
...
intrinsic linear
noise update
Figure 1: Schematic illustrating the Quadratic input latent dynamical system model (QLDS).
The sensory stimulus is filtered by multiple units with quadratic stimulus selectivity (only one of
which is shown) which model the feed-forward input into the population. This stimulus-drive provides input into a multi-dimensional linear dynamical system model which models recurrent dynamics and shared noise within the population. Finally, each neuron yi in the population is influenced
by the dynamical system via a linear readout. QLDS therefore models both the stimulus selectivity
as well as the spatio-temporal correlations of the population.
2
2.1
The Quadratic Input Latent Dynamical System (QLDS) model
Model
We summarize the collective dynamics of a population using a linear, low-dimensional dynamical
system with an n-dimensional latent state xt . The evolution of xt is given by
xt = Axt?1 + f? (ht ) + t ,
(1)
where A is the n ? n dynamics matrix and is Gaussian innovation noise with covariance matrix
Q, t ? N (0, Q). Each stimulus ht drives some dimensions of xt via a nonlinear function of the
stimulus, f? , with parameters ?, where the exact form of f (?) will be discussed below. The log
firing rates zt of the population couple to the latent state xt via a loading matrix C,
zt = Cxt + D ? st + d.
(2)
Here, we also include a second external input st , which is used to model the dependence of the
firing rate of each neuron on its own spiking history [14]. We define D ? st to be that vector
PNs
whose k-th element is given by (D ? st )k ? i=1
Dk,i sk,t?i . D therefore models single-neuron
properties that are not explained by shared population dynamics, and captures neural properties such
as burstiness or refractory periods. The vector d represents a constant, private spike rate for each
neuron. The vector xt represents the n-dimensional state of m neurons. Typically n < m, so the
model parameterizes a low-dimensional dynamics for the population.
We assume that, conditional on zt , the observed activity yt of m neurons is Poisson-distributed,
yk,t ? Poisson(exp(zk,t )).
(3)
While the Poisson likelihood provides a realistic probabilistic model for the discrete nature of spiking responses, it makes learning and inference more challenging than it would be for a Gaussian
model. As we discuss in the subsequent section, we rely on computationally-efficient approximations to perform inference under the Poisson observation model for QLDS.
2
2.2
Nonlinear stimulus dependence
Individual neurons in visual cortex respond selectively to only a small subset of stimulus features
[4, 15]. Certain subpopulations of neurons, such as in a cortical column, share substantial receptive
field overlap. We model such a neural subpopulation as sensitive to stimulus variation in a linear
subspace of stimulus space, and seek to characterize this subspace by learning a set of basis vectors,
or receptive fields, wi . In QLDS, a subset of latent states receives a nonlinear stimulus drive, each
of which reflects modulation by a particular receptive field wi . We consider three different forms
of stimulus model: a fully linear model, and two distinct quadratic models. Although it is possible to incorporate more complicated stimulus models within the QLDS framework, the quadratic
models? compact parameterization and analytic elegance make them both flexible and computationally tractable. What?s more, quadratic stimulus models appear in many classical models of neural
computation, e.g. the Adelson-Bergen model for motion-selectivity [16]; quadratic models are also
sometimes used in the classification of simple and complex cells in area V1 [4].
We express our stimulus model by the function f? (ht ), where ? represents the set of parameters describing the stimulus filters wi and mixing parameters ai , bi and ci (in the case of the quadratic models). When fB (ht ) is identically 0 (no stimulus input), the QLDS with Poisson observations reduces
to what has been previously studied as the Poisson Latent Dynamical System (PLDS) [17, 18, 9].
We briefly review three stimulus models we consider, and discuss their computational properties.
Linear: The simplest stimulus model we consider is a linear function of the stimulus,
f (ht ) = Bht ,
(4)
where the rows of B as linear filters, and ? = {B}. This baseline model is identical to [18, 9] and
captures simple cell-like receptive fields since the input to latent states is linear and the observation
process is generalized linear.
Quadratic: Under the linear model, latent dynamics receive linear input from the stimulus along
a single filter dimension, wi . In the quadratic model, we permit the input to each state to be a
quadratic function of wi . We describe the quadratic by including three additional parameters per
latent dimension, so that the stimulus drive takes the form
2
fB,i (ht ) = ai wiT ht + bi wiT ht + ci .
(5)
Here, the parameters ? = {wi , ai , bi , ci : i = 1, . . . , m} include multiple stimulus filters wi and
quadratic parameters (ai , bi , ci ). Equation 5 might result in a stimulus input that has non-zero mean
with respect to the distribution of the stimulus ht , which may be undesirable. Given the covariance
of ht , it is straightforward to constrain the input to be zero-mean by setting ci = ?ai wiT ?wi , where
? is the covariance of ht and we assume the stimulus to have zero mean as well. The quadratic model
enables QLDS to capture phase-invariant responses, like those of complex cells in area V1.
Quadratic with multiplicative interactions: In the above model, there are no interactions between different stimulus filters, which makes it difficult to model suppressive or facilitating interactions between features [4]. Although contributions from different filters combine in the dynamics
of x, any interactions are linear. Our third stimulus model allows for multiplicative interactions
between r < m stimulus filters, with the i-th dimension of the input given by
r
X
f?,i (ht ) =
ai,j wi T ht wjT ht + bi wi T ht + ci .
j=1
Again, we constrain this function to have zero mean by setting ci = ?
2.3
Pr
j=1
ai,j wiT ?wj .
Learning & Inference
We learn all parameters via the expectation-maximization (EM) algorithm. EM proceeds by alternating between expectation (E) and maximization (M) steps, iteratively maximizing a lower-bound
to the log likelihood [19]. In the E-step, one infers the distribution over trajectories xt , given data
and the parameter estimates from the previous iteration. In the M-step, one updates the current parameter estimates by maximizing the expectation of the log likelihood, a lower bound on the log
likelihood. EM is a standard method for fitting latent dynamical models; however, the Poisson
observation model complicates computation and requires the use of approximations.
3
E-step: With Gaussian latent states xt , posterior inference amounts to computing the posterior
means ?t and covariances Qt of the latent states, given data and current parameters. With Poisson observations exact inference becomes intractable, so that approximate inference has to be used
[18, 20, 21, 22]. Here, we apply a global Laplace approximation [20, 9] to efficiently (linearly
in experiment duration T ) approximate the posterior distribution by a Gaussian. We note that each
fB (ht ) in the E-step is deterministic, making posterior inference identical to standard PLDS models.
We found a small number of iterations of Newton?s method sufficient to perform the E-step.
M-step: In the M-step, each parameter is updated using the means ?t and covariances Qt inferred
in the E-step. Given ?t and Qt , the parameters A and Q have closed-form update rules that are
derived in standard texts [23]. For the Poisson likelihood, the M-step requires nonlinear optimization
to update the parameters C, D and d [18, 9]. While for linear stimulus functions f? (ht ) the Mstep has a closed-form solution, for nonlinear stimulus functions we optimize ? numerically. The
objective function for ? given by
T
g(?) = ?
1 X
(?t ? A?t?1 ? f? (ht ))T Q?1 (?t ? A?t?1 ? f? (ht )) + const.,
2 t=2
where ?t = E[xt |yt?1 , ht ]. If ? is represented as a vector concatenating all of its parameters, the
gradient of g(?) takes the form
T
X
?f (ht )
?g(?)
?1
= ?Q
(?t ? A?t?1 ? f? (ht ))
.
??
??
t=2
For the quadratic nonlinearity, the gradients with respect to f (ht ) take the form
h
i
?f (ht )
?f (ht ) T 2
= 2 ai ht T wi + bi ht T ,
= ht wi ,
?wi
?ai
?f (ht )
?f (ht )
= ht T wi ,
= 1.
?bi
?ci
(6)
(7)
(8)
Gradients for the quadratic model with multiplicative interactions take a similar form. When constrained to be 0-mean, the gradient for ci disappears, and is replaced by an additional term in the
gradients for a and wi (arising from the constraint on c).
We found both computation time and quality of fit for QLDS to depend strongly upon the optimization procedure used. For long time series, we split the data into small minibatches. The QLDS E-step
and M-step each naturally parallelize across minibatches. Neurophysiological experiments are often
naturally segmented into separate trials across different stimuli and experimental conditions, making
it possible to select minibatches without boundary effects.
3
Application to simulated data
We illustrate the properties of QLDS using a simulated population recording of 100 neurons, each
responding to a visual stimulus of binary, white spatio-temporal noise of dimensionality 240. We
simulated a recording with T = 50000 samples and a 10-dimensional latent dynamical state. Five of
the latent states received stimulus input from a bank of 5 stimulus filters (see Fig. 2A, top row), and
the remaining latent dimensions only had recurrent dynamics and noise. We aimed to approximate
the properties of real neural populations in early sensory cortex. In particular, we set the dynamics
matrix A by fitting the model to a single neuron recording from V1 [4]. When fitting the model,
we assumed the same dimensionalities (10 latent states, 5 stimulus inputs) as those used to generate
the data. We ran 100 iterations of EM, which?-for the recording length and dimensionality of this
system?took about an hour on a 12?core intel Xeon CPU at 3.5GHz.
The model recovered by EM matched the statistics of the true model well. Linear dynamical system
and quadratic models of stimulus selectivity both commonly have invariances that render a particular
parameterization unidentifiable [4, 15], and QLDS is no exception: the latent state (and its parameters) can be rotated without changing the model?s properties. Hence it is possible only to compare
the subspace recovered by the model, and not the individual filters. In order to visualize subspace
recovery, we computed the best `2 approximation of the 5 ?true? filters in the subspace spanned by
4
Stimulus correlations
?0.2
B
60
80
100
0
20 40 60 80 100
20
40
60
?0.1
80
?0.2
100
true
20 40 60 80 100
60
?0.1
80
?0.2
100
0.2 0.4 0.6 0.8
real
noise vs stimulus
correlations
0.2
0.15
0.1
0.05
0
?0.05
0
0.1
true
0.2
stimulus correlation
0.4
0.3
40
fit
Noise correlations
0
20
?0.5
E
F
0.2
0.1
0
true
0.2
0.1
imaginary
?0.1
40
noise correlation
0
D 0.5 eigenvalues of A
20
probability
0.1
fit
0.2
fit
C
Total correlations
A
true
fit
0.2
0.1
20 40 60 80 100
0
20
40
60
synchronous spikes
Figure 2: Results on simulated data. Low-dimensional subspace recovery from a population of
100 simulated neurons in response to a white noise stimulus. (A) Simulated neurons receive shared
input from 5 spatio-temporal receptive fields (top row). QLDS recovers a subspace capable of
representing the original 5 filters (bottom row). (B) QLDS permits a more compact representation
than the conventional approach of mapping receptive fields for each neuron. For comparison with
the representation in panel A, we here show the spike-triggered averages of the first 60 neurons in the
population. (C) QLDS also models shared variability across neurons, as visualised here by the three
different measures of correlation. Top: Total correlation coefficients between each pair of neurons.
Values below the diagonal are from the simulated data, above the diagonal correspond to correlations
recovered by the model. Center: Stimulus correlations Bottom: Noise correlations. (D) Eigenvalues
of dynamics matrix A (black is ground truth, red is estimated). (E) In this model, stimulus and noise
correlations are dependent on each other, for the parameters chosen in this stimulation, there is a
linear relationship between them. (F) Distribution of population spike counts, i.e. total number of
spikes in each time bin across the population.
MSE (log scale)
1
0
?1
?2
?3
B
reconstruction performance
vs population size
linear
quadratic
quadratic cross
2
MSE (log scale)
A
reconstruction performance
vs experiment length
1
0
?1
?2
?3
?4
?4
200 400 600 800 1000
Population Size (# Cells)
?5
5000
10000
15000
Experiment length (# samples)
Figure 3: Recovery of stimulus subspace as a function of population size (A) and experiment duration (B). Each point represents the best filter reconstruction performance of QLDS over 20 distinct
simulations from the same ?true? model, each initialized randomly and fit using the same number
of EM iterations. Models were fit with each of three distinct stimulus nonlinearities, linear s (blue),
quadratic (green), and quadratic with multiplicative interactions (red). Stimulus input of the ?true?
was a quadratic with multiplicative interactions, and therefore we expect only the multiplicative
model (red) to each low error rates.
? i (see Fig. 2 A bottom row). In QLDS, different neurons share different filters, and
the estimated w
therefore these 5 filters provide a compact description of the stimulus selectivity of the population
[24]. In contrast, for traditional single-neuron analyses [4] ?fully-connected? models such as GLMs
[14] one would estimate the receptive fields of each of the 100 filters in the population, resulting in a
much less compact representation with an order of magnitude more parameters for the stimulus-part
alone (see Fig. 2B).
5
QLDS captures both the stimulus-selectivity of a population and correlations across neurons. In
studies of neural coding, correlations between neurons (Fig. 2C, top) are often divided into stimuluscorrelations and noise-correlations. Stimulus correlations capture correlations explainable by similarity in stimulus dependence (and are calculated by shuffling trials), whereas noise-correlations
capture correlations not explainable by shared stimulus drive (which are calculated by correlating
residuals after subtracting the mean firing rate across multiple presentations of the same stimulus).
The QLDS-model was able to recover both the total, stimulus and noise correlations in our simulation (Fig. 2C), although it was fit only to a single recording without stimulus repeats. Finally, the
model also recovered the eigenvalues of the dynamics (Fig. 2D), the relationship between noise and
stimulus correlations (Fig. 2E) and the distribution of population spike counts (Fig. 2F).
We assume that all stimulus dependence is captured by the subspace parameterized by the filters
of the stimulus model. If this assumption holds, increasing the size of the population increases
statistical power and makes identification of the stimulus selectivity easier rather than harder, in
a manner similar to that of increasing the duration of the experiment. To illustrate this point, we
generated multiple data-sets with larger population sizes, or with longer recording times, and show
that both scenarios lead to improvements in subspace-recovery (see Fig. 3).
4
Applications to Neural Data
Cat V1 with white noise stimulus We evaluate the performance of the QLDS on multi-electrode
recordings from cat primary visual cortex. Data were recorded from anaesthetized cats in response to
a single repeat of a 20 minute long, full-field binary noise movie, presented at 30 frames per second,
and 60 repeats of a 30s long natural movie presented at 150 frames per second. Spiking activity
was binned at the frame rate (33 ms for noise, 6.6 ms for natural movies). For noise, we used the
first 18000 samples for training, and 5000 samples for model validation. For the natural movie, 40
repeats were used for training and 20 for validation. Silicon polytrodes (Neuronexus) were employed
to record multi-unit activity (MUA) from a single cortical column, spanning all cortical layers with
32 channels. Details of the recording procedure are described elsewhere [25]. For our analyses, we
used MUA without further spike-sorting from 22 channels for noise data and 25 channels for natural
movies. We fit a QLDS with 3 stimulus filters, and in each case a 10-dimensional latent state, i.e. 7
of the latent dimensions received no stimulus drive.
Spike trains in this data-set exhibited ?burst-like? events in which multiple units were simultaneously
active (Fig. 4A). The model captured these events by using a dimension of the latent state with
substantial innovation noise, leading substantial variability in population activity across repeated
stimulus presentations. We also calculated pairwise (time-lagged) cross-correlations for each unit
pair, as well as the auto-correlation function for each unit in the data (Fig. 4B, 7 out of 22 neurons
shown, results for other units are qualitatively similar.). We found that samples from the model
(Fig. 4B, red) closely matched the correlations of the data for most units and unit-pairs, indicating
the QLDS provided an accurate representation of the spatio-temporal correlation structure of the
population recording. The instantaneous correlation matrix across all 22 cells was very similar
between the physiological and sampled data (Fig. 4C). We note that receptive fields (Fig. 4F) in this
data did not have spatio-temporal profiles typical of neurons in cat V1 (this was also found when
using conventional analyses such as spike-triggered covariance). Upon inspection, this was likely a
consequence of an LGN afferent also being included in the raw MUA. In our analysis, a 3-feature
model captured stimulus correlations (in held out data) more accurately than 1- and 2- filter models.
However, 10-fold cross validation revealed that 2- and 3- filter models do not improve upon a 1-filter
model in terms of one-step-ahead prediction performance (i.e. trying to predict neural activity on
the next time-step using past observations of population activity and the stimulus).
Macaque V1 with drifting grating stimulus: We wanted to evaluate the ability of the model to
capture the correlation structure (i.e. noise and signal correlations) of a data-set containing multiple
repetitions of each stimulus. To this end, we fit QLDS with a Poisson observation model to the
population activity of 113 V1 neurons from an anaesthetized macaque, as described in [26]. Drifting grating stimuli were presented for 1280ms, followed by a 1280ms blank period, with each of
72 grating orientations repeated 50 times. We fit a QLDS with a 20-dimensional latent state and 15
stimulus filters, where the stimulus was paramterized as a set of phase-shifted sinusoids at the appropriate spatial and temporal frequency (making ht 112-dimensional). We fit the QLDS to 35 repeats,
6
B
data
5
10
?20
15
0
20
0.4
0.4
0.2
0.2
0
0
?20
10 20 30 40 50 60 70 80 90
0
20
0.4
10
15
0.8
noise vs stimulus
correlation
0.6
0.4
0.2
0.4
0.6
0.8
stimulus correlation
0
?20
0
0.4
0.2
0.2
0
0
0
0
0
20
?20
0
20
?20
0
20
?20
0.4
0.4
0.4
0.4
0.2
0.2
0.2
0.2
0
0
0
0
0
20
?20
0
20
?20
0
20
?20
?1
0
?20
0.4
0.2
0.2
0.2
0.2
0.2
0
0
0
0
0
20
?20
0
20
?20
0
20
?20
0
20
0.4
0.2
0
0
20
?20
0.4
0.4
0.4
0.4
0.4
0.4
0.2
0.2
0.2
0.2
0.2
0.2
0
0
0
0
0
0
0
20
?20
0
20
?20
0
20
?20
20
0
20
0.4
0
15
0.2
0
0.4
?20
10
0.4
0.4
20
true
5
20
0.4
0
20
20
0.4
0
20
?20
0
20
?20
0
20
0.4
0.2
0
0
20
?20
0
20
F
eigenvalues of A
feature 1
0.5
feature 2
0
?0.5
20
0.2
?20
E
imaginary
noise correlation
D
10 20 30 40 50 60 70 80 90
time (s)
?0.5
0.2
0
0.4
?20
20
?20
0.2
?20
5
20
15
0.4
0.4
?20
20
0
10
0
20
0
?20
15
0
0.2
0
5
?20
5
0.5
0.4
0.2
10
Total correlations
1
0.2
0
20
Simulated repeats to identical noise stimulus
C
0.4
fit
A
0
0.5
real
1
feature 3
?165ms ?132ms ?99ms ?66ms ?33ms
0ms
Figure 4: QLDS fit to V1 cells with noise stimuli. We fit QLDS to T = 18000 samples of 22
neurons responding to a white noise stimulus, data binned at 33 ms. We used the quadratic with
multiplicative interactions as the stimulus nonlinearity. The QLDS has a 10-dimensional latent state
with 3 stimulus inputs. All results shown here are compared against T = 5000 samples of test-data,
not used to train the model. (A) Top row: Rasters from recordings from 22 cells in cat visual cortex,
where cell index appears on the y axis, and time in seconds on the x. Second and third row: Two independent samples from the QLDS model responding to the same noise stimuli. Note that responses
are highly variable across trials. (B) Auto- and cross-correlations for data (black) and model (red)
cells. For the model, we average across 60 independent samples, thickness of red curves reflects 1
standard deviation from the mean. Panel (i, j) corresponds to cross-correlation between units with
indices i and j, panels along the diagonal show auto-correlations. (C) Total correlations for the true
(lower diagonal) and model (upper diagonal) populations. (D) Noise correlations scattered against
stimulus correlations for the model. As we did not have repeat data for this population, we were not
able to reliably estimate noise correlations, and thereby evaluate the accuracy of this model-based
prediction. (E) Eigenvalues of the dynamics matrix A. (F) Three stimulus filters recovered by
QLDS. We selected the 3-filter QLDS by inspection, having observed that fitting with larger number
of stimulus filters did not improve the fit. We note that although two of the filters appear similar,
that they drive separate latent dimensions with distinct mixing weights ai , bi and ci .
and held out 15 for validation. The QLDS accurately captured the stimulus and noise correlations of
the full population (Fig. 5A). Further, a QLDS with 15 shared receptive fields captured simple and
complex cell behavior of all 113 cells, as well as response variation across orientation (Fig. 5B).
5
Discussion
We presented QLDS, a statistical model for neural population recordings from sensory cortex that
combines low-dimensional, quadratic stimulus dependence with a linear dynamical system model.
The stimulus model can capture simple and complex cell responses, while the linear dynamics capture temporal dynamics of the population and shared variability between neurons. We applied QLDS
to population recordings from primary visual cortex (V1). The cortical microcircuit in V1 consists of
highly-interconnected cells that share receptive field properties such as orientation preference [27],
with a well-studied laminar organization [1]. Layer IV cells have simple cell receptive field properties, sending excitatory connections to complex cells in the deep and superficial layers. Quadratic
7
A
Stimulus correlations
Noise correlations
0.1
60
0.05
0
?0.5
?1
B
spike rate
Cell 49
0.6
0 degrees
80
100
data
20 40 60 80 100
cell index
45 degrees
90 degrees
0
?0.05
?0.1
data
20 40 60 80 100
cell index
135 degrees
180 degrees
225 degrees
500 1000 1500
time (ms)
500 1000 1500
time (ms)
500 1000 1500
time (ms)
0.4
0.2
0
stimulus off
Cell 50 0.6
spike rate
model
40
0.5
model
20
1
0.4
0.2
0
500 1000 1500
time (ms)
500 1000 1500
time (ms)
500 1000 1500
time (ms)
Figure 5: QLDS fit to 113 V1 cells across 35 repeats of each of 72 grating orientations. (A)
Comparison of total correlations in the data and generated from the model, (B) For two cells (cells
49 and 50, using the index scheme from A) and 6 orientations (0, 45, 90, 135, 180, and 225 degrees),
we show the posterior mean prediction performance (red traces) in in comparison to the average
across 15 held-out trials (black traces). In each block, we show predicted and actual spike rate
(y-axis) over time binned at 10 ms (x-axis). Stimulus offset is denoted by a vertical blue line.
stimulus models such as the classical ?energy model? [16] of complex cells reflect this structure.
The motivation of QLDS is to provide a statistical description of receptive fields in the different
cortical layers, and to parsimoniously capture both stimulus dependence and correlations across an
entire population.
Another prominent neural population model is the GLM (Generalized Linear Model, e.g. [14]; or
the ?common input model?, [28]), which includes a separate receptive field for each neuron, as
well as spike coupling terms between neurons. While the GLM is a successful model of a population?s statistical response properties, its fully?connected parameterization scales quadratically with
population size. Furthermore, the GLM supposes direct couplings between pairs of neurons, while
monosynaptic couplings are statistically unlikely for recordings from a small number of neurons
embedded in a large network.
In QLDS, latent dynamics mediate both stimulus and noise correlations. This reflects the structure
of the cortex, where recurrent connectivity gives rise to both stimulus-dependent and independent
correlations. Without modeling a separate receptive field for each neuron, the model complexity of
QLDS grows only linearly in population size, rather than quadratically as in fully-connected models
such as the GLM [14]. Conceptually, our modeling approach treats the entire recorded population
as a single ?computational unit?, and aims to characterize its joint feature-selectivity and dynamics.
Neurophysiology and neural coding are progressing toward recording and analyzing datasets of ever
larger scale. Population-level parameterizations, such as QLDS, provide a scalable strategy for
representing and analyzing the collective computational properties of neural populations.
Acknowledgements
We are thankful to Arnulf Graf and the co-authors of [26] for sharing the data used in Fig. 5, and to
Memming Park for comments on the manuscript. JHM and EA were funded by the German Federal
Ministry of Education and Research (BMBF; FKZ: 01GQ1002, Bernstein Center T?ubingen) and the
Max Planck Society, and UK by National Eye Institute grant #EY019965. Collaboration between
EA, JP and JHM initiated at the ?MCN? Course at the Marine Biological Laboratory, Woods Hole.
8
References
[1] D. Hubel and T. Wiesel, ?Receptive fields, binocular interaction and functional architecture in the cat?s
visual cortex,? J Physiol, pp. 106?154, 1962.
[2] S. L. Smith and M. H?ausser, ?Parallel processing of visual space by neighboring neurons in mouse visual
cortex,? Nature Neurosci, vol. 13, no. 9, pp. 1144?9, 2010.
[3] D. S. Reich, F. Mechler, and J. D. Victor, ?Independent and redundant information in nearby cortical
neurons,? Science, vol. 294, pp. 2566?2568, 2001.
[4] N. C. Rust, O. Schwartz, J. A. Movshon, and E. P. Simoncelli, ?Spatiotemporal elements of macaque v1
receptive fields,? Neuron, vol. 46, no. 6, pp. 945?56, 2005.
[5] T. O. Sharpee, ?Computational identification of receptive fields,? Annu Rev Neurosci, vol. 36, pp. 103?20,
2013.
[6] M. M. Churchland, B. M. Yu, M. Sahani, and K. V. Shenoy, ?Techniques for extracting single-trial activity
patterns from large-scale neural recordings,? vol. 17, no. 5, pp. 609?618, 2007.
[7] B. M. Yu, J. P. Cunningham, G. Santhanam, S. I. Ryu, K. V. Shenoy, and M. Sahani, ?Gaussian-process
factor analysis for low-dimensional single-trial analysis of neural population activity,? vol. 102, no. 1,
pp. 614?635, 2009.
[8] W. Truccolo, L. R. Hochberg, and J. P. Donoghue, ?Collective dynamics in human and monkey sensorimotor cortex: predicting single neuron spikes,? Nat Neurosci, vol. 13, no. 1, pp. 105?111, 2010.
[9] J. H. Macke, L. B?using, J. P. Cunningham, B. M. Yu, K. V. Shenoy, and M. Sahani., ?Empirical models
of spiking in neural populations,? in Adv in Neural Info Proc Sys, vol. 24, 2012.
[10] D. Pfau, E. A. Pnevmatikakis, and L. Paninski, ?Robust learning of low-dimensional dynamics from large
neural ensembles,? in Adv in Neural Info Proc Sys, pp. 2391?2399, 2013.
[11] V. Mante, D. Sussillo, K. V. Shenoy, and W. T. Newsome, ?Context-dependent computation by recurrent
dynamics in prefrontal cortex,? Nature, vol. 503, pp. 78?84, Nov. 2013.
[12] A. Fairhall, ?The receptive field is dead. long live the receptive field?,? Curr Opin Neurobiol, vol. 25,
pp. ix?xii, 2014.
[13] I. M. Park, E. W. Archer, N. Priebe, and J. W. Pillow, ?Spectral methods for neural characterization using
generalized quadratic models,? in Adv in Neural Info Proc Sys 26, pp. 2454?2462, 2013.
[14] J. W. Pillow, J. Shlens, L. Paninski, A. Sher, A. M. Litke, E. J. Chichilnisky, and E. P. Simoncelli, ?Spatiotemporal correlations and visual signalling in a complete neuronal population,? Nature, vol. 454, no. 7207,
pp. 995?999, 2008.
[15] J. W. Pillow and E. P. Simoncelli, ?Dimensionality reduction in neural models: an information-theoretic
generalization of spike-triggered average and covariance analysis,? J Vis, vol. 6, no. 4, pp. 414?28, 2006.
[16] E. H. Adelson and J. R. Bergen, ?Spatiotemporal energy models for the perception of motion,? J Opt Soc
Am A, vol. 2, no. 2, pp. 284?99, 1985.
[17] A. C. Smith and E. N. Brown, ?Estimating a state-space model from point process observations,? vol. 15,
no. 5, pp. 965?991, 2003.
[18] J. E. Kulkarni and L. Paninski, ?Common-input models for multiple neural spike-train data,? Network,
vol. 18, no. 4, pp. 375?407, 2007.
[19] A. P. Dempster, N. M. Laird, and D. B. Rubin, ?Maximum likelihood from incomplete data via the em
algorithm,? J R Stat Soc Ser B, vol. 39, no. 1, pp. 1?38, 1977.
[20] L. Paninski, Y. Ahmadian, D. Ferreira, S. Koyama, K. Rahnama Rad, M. Vidne, J. Vogelstein, and W. Wu,
?A new look at state-space models for neural data,? vol. 29, pp. 107?126, 2010.
[21] A. Z. Mangion, K. Yuan, V. Kadirkamanathan, M. Niranjan, and G. Sanguinetti, ?Online variational inference for state-space models with point-process observations,? Neural Comput, vol. 23, no. 8, pp. 1967?
1999, 2011.
[22] M. Emtiyaz Khan, A. Aravkin, M. Friedlander, and M. Seeger, ?Fast dual variational inference for nonconjugate latent gaussian models,? in Proceedings of ICML, 2013.
[23] Z. Ghahramani and G. E. Hinton, ?Parameter estimation for linear dynamical systems,? Univ. Toronto
Tech Report, vol. 6, no. CRG-TR-96-2, 1996.
[24] J. H. Macke, G. Zeck, and M. Bethge, ?Receptive fields without spike-triggering,? in Adv in Neural Info
Proc Sys, vol. 20, pp. 969?976, 2008.
[25] U. K?oster, J. Sohl-Dickstein, C. M. Gray, and B. A. Olshausen, ?Modeling higher-order correlations
within cortical microcolumns,? PLoS Comput Bio, vol. 10, no. 7, p. e1003684, 2014.
[26] A. B. Graf, A. Kohn, M. Jazayeri, and J. A. Movshon, ?Decoding the activity of neuronal populations in
macaque primary visual cortex,? Nature neuroscience, vol. 14, no. 2, pp. 239?245, 2011.
[27] V. Mountcastle, ?Modality and topographic properties of single neurons of cat?s somatic sensory cortex,?
J Neurophysiol, 1957.
[28] M. Vidne, Y. Ahmadian, J. Shlens, J. Pillow, J. Kulkarni, A. Litke, E. Chichilnisky, E. Simoncelli, and
L. Paninski, ?Modeling the impact of common noise inputs on the network activity of retinal ganglion
cells,? J Comput Neurosci, 2011.
9
| 5263 |@word neurophysiology:1 trial:6 illustrating:1 briefly:1 private:1 wiesel:1 loading:1 seek:2 simulation:2 covariance:7 thereby:1 tr:1 harder:1 reduction:1 series:1 past:1 imaginary:2 current:2 com:1 recovered:5 blank:1 realize:1 physiol:1 realistic:1 subsequent:1 shape:1 analytic:1 enables:1 mstep:1 treating:1 wanted:1 update:4 opin:1 v:4 alone:1 selected:1 parameterization:3 signalling:1 inspection:2 sys:4 marine:1 smith:2 short:1 core:1 record:1 filtered:1 provides:2 parameterizations:1 characterization:1 location:1 preference:2 toronto:1 five:1 along:2 burst:1 direct:1 yuan:1 consists:1 combine:3 fitting:4 manner:1 introduce:2 pairwise:1 behavior:1 mpg:2 multi:5 cpu:1 actual:1 increasing:2 becomes:1 provided:1 monosynaptic:1 moreover:1 underlying:1 circuit:3 matched:2 panel:3 estimating:1 what:2 neurobiol:1 monkey:1 temporal:8 berkeley:1 axt:1 ferreira:1 uk:1 schwartz:1 unit:12 grant:1 ser:1 appear:2 planck:2 organize:1 shenoy:4 zeck:1 bio:1 local:1 treat:1 consequence:1 analyzing:2 initiated:1 parallelize:1 firing:5 modulation:1 might:1 black:3 studied:2 challenging:1 co:1 bi:8 statistically:1 block:1 procedure:2 evan:2 area:2 empirical:1 subpopulation:2 rahnama:1 undesirable:1 context:1 live:1 optimize:1 conventional:2 deterministic:1 center:4 yt:2 maximizing:2 straightforward:1 duration:5 focused:2 wit:4 recovery:4 rule:1 spanned:1 shlens:2 population:57 variation:2 laplace:2 updated:1 elucidate:1 exact:2 element:2 observed:2 bottom:3 capture:12 readout:1 wj:1 connected:3 adv:4 plo:1 burstiness:1 substantial:4 yk:1 ran:1 visualised:1 complexity:1 dempster:1 dynamic:30 depend:2 churchland:1 upon:3 basis:1 neurophysiol:1 joint:1 represented:1 cat:7 train:3 univ:1 distinct:4 fast:2 describe:2 ahmadian:2 macke1:1 whose:1 larger:3 ability:1 statistic:1 topographic:1 laird:1 online:2 triggered:3 eigenvalue:5 un1:1 took:1 propose:1 reconstruction:3 interconnected:2 interaction:10 subtracting:1 neighboring:1 mixing:2 description:3 electrode:1 rotated:1 thankful:1 illustrate:2 develop:1 recurrent:4 stat:1 coupling:3 sussillo:1 qt:3 received:2 progress:1 grating:4 soc:2 predicted:1 aravkin:1 closely:1 filter:26 human:1 mangion:1 bin:1 education:1 truccolo:1 generalization:1 mua:3 opt:1 biological:2 crg:1 hold:1 ground:1 exp:1 mapping:1 predict:1 visualize:1 early:1 estimation:2 proc:4 integrates:1 sensitive:1 pnevmatikakis:1 repetition:1 reflects:3 federal:1 gaussian:6 aim:1 rather:2 derived:1 improvement:1 likelihood:6 tech:1 contrast:1 seeger:1 litke:2 baseline:1 progressing:1 am:1 inference:10 dependent:4 bergen:2 typically:2 entire:2 unlikely:1 cunningham:2 archer:2 lgn:1 uncovering:1 classification:1 orientation:6 flexible:1 denoted:1 dual:1 constrained:1 spatial:1 field:24 having:2 sampling:1 identical:3 represents:4 park:2 adelson:2 yu:3 look:1 icml:1 report:1 stimulus:106 modern:1 randomly:1 simultaneously:2 national:1 individual:3 parsimoniously:1 replaced:1 phase:2 curr:1 organization:1 highly:3 held:3 accurate:1 capable:1 iv:1 incomplete:1 initialized:1 theoretical:1 jazayeri:1 complicates:1 column:4 modeling:6 xeon:1 newsome:1 maximization:3 deviation:1 subset:2 successful:1 optimally:1 characterize:2 thickness:1 supposes:1 spatiotemporal:3 st:4 probabilistic:2 off:1 decoding:1 bethge:1 mouse:1 connectivity:1 again:1 reflect:1 recorded:2 containing:1 prefrontal:1 external:1 dead:1 macke:2 leading:1 account:3 potential:1 nonlinearities:2 de:2 retinal:1 coding:3 includes:1 coefficient:1 afferent:1 vi:1 multiplicative:7 bht:1 closed:2 red:7 recover:1 complicated:1 parallel:1 memming:1 cxt:1 contribution:1 accuracy:1 efficiently:1 ensemble:1 correspond:1 emtiyaz:1 conceptually:1 identification:3 raw:1 accurately:2 trajectory:1 drive:12 cybernetics:1 history:1 influenced:2 sharing:1 against:2 raster:1 energy:2 sensorimotor:1 frequency:1 pp:22 elegance:1 naturally:2 recovers:1 couple:1 sampled:1 dimensionality:5 infers:1 ea:2 appears:1 feed:1 manuscript:1 higher:1 nonconjugate:1 response:16 unidentifiable:1 microcircuit:1 strongly:1 furthermore:1 binocular:1 correlation:53 glms:1 receives:2 nonlinear:10 quality:1 reveal:1 gray:1 grows:1 olshausen:1 effect:2 brown:1 true:10 evolution:1 hence:1 sinusoid:1 alternating:1 iteratively:1 laboratory:1 white:4 m:18 generalized:3 trying:1 prominent:1 complete:1 theoretic:1 motion:2 variational:2 instantaneous:1 common:5 functional:2 spiking:5 stimulation:1 rust:1 refractory:1 jp:1 discussed:1 numerically:1 silicon:1 ai:10 shuffling:1 tuning:1 nonlinearity:2 had:1 funded:1 reich:1 cortex:17 similarity:1 longer:1 posterior:5 own:1 recent:2 ausser:1 pns:1 selectivity:10 certain:1 scenario:1 ubingen:3 binary:2 yi:1 victor:1 captured:5 ministry:1 additional:2 employed:1 redundant:2 period:2 signal:1 vogelstein:1 multiple:8 full:2 simoncelli:4 reduces:1 segmented:1 cross:6 long:4 divided:1 niranjan:1 schematic:1 prediction:3 scalable:1 impact:1 expectation:4 poisson:11 iteration:4 sometimes:1 cell:26 receive:2 whereas:1 separately:1 suppressive:1 modality:1 exhibited:1 comment:1 recording:19 extracting:2 bernstein:2 split:1 identically:1 revealed:1 neuronexus:1 fit:17 psychology:1 architecture:1 fkz:1 triggering:1 parameterizes:1 donoghue:1 synchronous:1 kohn:1 explainable:2 movshon:2 render:2 deep:1 governs:1 aimed:1 amount:1 simplest:1 generate:1 mcn:1 shifted:1 neuroscience:5 arising:1 per:3 estimated:2 blue:2 xii:1 discrete:1 vol:22 dickstein:1 express:1 santhanam:1 changing:1 ht:32 v1:12 imaging:1 fraction:1 wood:1 parameterized:1 respond:1 wu:1 hochberg:1 bound:2 layer:4 followed:1 fold:1 quadratic:30 laminar:1 mante:1 activity:16 fairhall:1 binned:3 ahead:1 constraint:1 constrain:2 nearby:1 relatively:1 department:1 mechler:1 across:17 em:7 ur:2 wi:15 rev:1 making:3 projecting:1 explained:1 invariant:1 pr:1 glm:4 computationally:2 equation:1 previously:1 discus:2 describing:1 count:2 german:1 tractable:1 end:1 sending:1 plds:2 permit:2 probe:1 apply:1 appropriate:1 spectral:1 drifting:2 original:1 vidne:2 responding:3 top:5 include:2 remaining:1 newton:1 const:1 ghahramani:1 polytrodes:1 classical:2 society:1 anaesthetized:2 seeking:1 objective:1 spike:19 kadirkamanathan:1 receptive:23 primary:4 dependence:8 strategy:1 diagonal:5 traditional:1 gradient:5 subspace:11 separate:4 simulated:8 koyama:1 tuebingen:2 spanning:1 toward:1 arnulf:1 length:3 index:5 relationship:2 multiplicatively:1 providing:1 innovation:2 difficult:2 idiosyncratic:1 info:4 trace:2 priebe:1 rise:1 lagged:1 reliably:1 collective:3 zt:3 perform:2 upper:1 vertical:1 neuron:43 observation:9 datasets:1 hinton:1 variability:4 ever:1 frame:3 redwood:1 jakob:2 somatic:1 inferred:1 pair:4 chichilnisky:2 khan:1 connection:1 rad:1 pfau:1 california:1 quadratically:2 ryu:1 hour:1 macaque:4 able:2 proceeds:1 dynamical:15 below:2 pattern:1 perception:1 challenge:1 summarize:1 max:2 including:1 green:1 power:1 overlap:1 event:2 natural:4 rely:1 predicting:1 residual:1 representing:2 scheme:1 improve:2 movie:5 technology:1 eye:1 disappears:1 axis:3 auto:3 sher:1 sahani:3 text:1 review:1 oster:1 acknowledgement:1 mountcastle:1 friedlander:1 graf:2 embedded:1 fully:4 expect:1 mixed:1 validation:4 degree:7 gq1002:1 sufficient:1 rubin:1 bank:1 share:5 collaboration:1 row:7 elsewhere:1 excitatory:1 course:1 repeat:8 allow:1 institute:3 characterizing:1 distributed:1 ghz:1 boundary:1 dimension:8 cortical:11 calculated:3 pillow:5 curve:1 fb:3 sensory:6 forward:1 commonly:1 qualitatively:1 author:1 approximate:3 compact:4 nov:1 global:1 correlating:1 active:1 hubel:1 assumed:1 spatio:5 sanguinetti:1 latent:28 sk:1 jhm:2 channel:4 zk:1 robust:1 nature:5 learn:1 superficial:1 interact:2 mse:2 complex:6 did:3 linearly:4 neurosci:4 motivation:1 noise:36 profile:1 mediate:1 repeated:2 facilitating:1 neuronal:2 fig:17 intel:1 scattered:1 bmbf:1 sub:1 exponential:1 concatenating:1 comput:3 third:2 ix:1 minute:1 annu:1 xt:9 offset:1 dk:1 physiological:1 intrinsic:1 intractable:1 sohl:1 ci:10 magnitude:1 nat:1 hole:1 sorting:1 easier:1 electrophysiology:1 paninski:5 likely:1 ganglion:1 neurophysiological:1 visual:15 corresponds:1 truth:1 minibatches:3 conditional:2 modulate:1 presentation:2 wjt:1 shared:10 included:1 typical:1 total:7 invariance:1 experimental:1 sharpee:1 exception:1 selectively:1 select:1 indicating:1 internal:1 jonathan:1 kulkarni:2 ongoing:3 incorporate:1 evaluate:3 princeton:3 |
4,709 | 5,264 | Inferring sparse representations of continuous signals
with continuous orthogonal matching pursuit
Jacob L. Yates
Department of Neuroscience
The University of Texas at Austin
[email protected]
Karin C. Knudson
Department of Mathematics
The University of Texas at Austin
[email protected]
Alexander C. Huk
Center for Perceptual Systems
Departments of Psychology & Neuroscience
The University of Texas at Austin
[email protected]
Jonathan W. Pillow
Princeton Neuroscience Institute and
Department of Psychology
Princeton University
[email protected]
Abstract
Many signals, such as spike trains recorded in multi-channel electrophysiological
recordings, may be represented as the sparse sum of translated and scaled copies
of waveforms whose timing and amplitudes are of interest. From the aggregate
signal, one may seek to estimate the identities, amplitudes, and translations of the
waveforms that compose the signal. Here we present a fast method for recovering these identities, amplitudes, and translations. The method involves greedily
selecting component waveforms and then refining estimates of their amplitudes
and translations, moving iteratively between these steps in a process analogous
to the well-known Orthogonal Matching Pursuit (OMP) algorithm [11]. Our approach for modeling translations borrows from Continuous Basis Pursuit (CBP)
[4], which we extend in several ways: by selecting a subspace that optimally captures translated copies of the waveforms, replacing the convex optimization problem with a greedy approach, and moving to the Fourier domain to more precisely
estimate time shifts. We test the resulting method, which we call Continuous Orthogonal Matching Pursuit (COMP), on simulated and neural data, where it shows
gains over CBP in both speed and accuracy.
1
Introduction
It is often the case that an observed signal is a linear combination of some other target signals that
one wishes to resolve from each other and from background noise. For example, the voltage trace
from an electrode (or array of electrodes) used to measure neural activity in vivo may be recording
from a population of neurons, each of which produces many instances of its own stereotyped action
potential waveform. One would like to decompose an analog voltage trace into a list of the timings
and amplitudes of action potentials (spikes) for each neuron.
Motivated in part by the spike-sorting problem, we consider the case where we are given a signal
that is the sum of known waveforms whose timing and amplitude we seek to recover. Specifically,
we suppose our signal can be modeled as:
y(t) =
Nf J
X
X
an,j fn (t ? ?n,j ),
n=1 j=1
1
(1)
where the waveforms fn are known, and we seek to estimate positive amplitudes an,j and event
times ?n,j . Signals of this form have been studied extensively [12, 9, 4, 3].
This a difficult problem in part because of the nonlinear dependence of y on ? . Moreover, in most
applications we do not have access to y(t) for arbitrary t, but rather have a vector of sampled (noisy)
measurements on a grid of discrete time points. One way to simplify the problem is to discretize ? ,
considering only a finite set of possible time shift ?n,j ? {?, 2?..., N? ?} and approximating the
signal as
y?
Nf J
X
X
an,j fn (t ? in,j ?), in,j ? 1, ..., N?
(2)
n=1 j=1
Once discretized in this way, the problem is one of sparse recovery: we seek to represent the
observed signal with a sparse linear combination of elements of a finite dictionary {fn,j (t) :=
fn (t ? j?), n ? 1, ..., Nf , j ? 1, ..., N? }. Framing the problem as sparse recovery, one can
bring tools from compressed sensing to bear. However, the discretization introduces several new
difficulties. First, we can only approximate the translation ? by values on a discrete grid. Secondly,
choosing small ? allows us to more closely approximate ? , but demands more computation, and
such finely spaced dictionary elements yield a highly coherent dictionary, while sparse recovery
algorithms generally have guarantees for low-coherence dictionaries.
A previously introduced algorithm that uses techniques of sparse recovery and returns accurate and
continuous valued estimates of a and ? is Continuous Basis Pursuit (CBP) [4], which we describe
below. CBP proceeds (roughly speaking) by augmenting the discrete dictionary fn,j (t) with other
carefully chosen basis elements, and then solving a convex optimization problem inspired by basis
pursuit denoising. We extend ideas introduced in CBP to present a new method for recovering
the desired time shifts ? and amplitudes a that leverage the speed and tractability of solving the
discretized problem while still ultimately producing continuous valued estimates of ? , and partially
circumventing the problem of too much coherence.
Basis pursuit denoising and other convex optimization or `1 -minimization based methods have been
effective in the realm of sparse recovery and compressed sensing. However, greedy methods have
also been used with great success. Our approach begins with the augmented bases used in CBP,
but adds basis vectors greedily, drawing on the well known Orthogonal Matching Pursuit algorithm
[11]. In the regimes considered, our greedy approach is faster and more accurate than CBP.
Broadly speaking, our approach has three parts. First, we augment the discretized basis in one of
several ways. We draw on [4] for two of these choices, but also present another choice of basis that
is in some sense optimal. Second, we greedily select candidate time bins of size ? in which we
suspect an event has occurred. Finally, we move from this rough, discrete-valued estimate of timing
? to continuous-valued estimates of ? and a. We iterate the second and third steps, greedily adding
candidate time bins and updating our estimates of ? and a until a stopping criterion is reached.
The structure of the paper is as follows. In Section 2 we describe the method of Continuous Basis
Pursuit (CBP), which our method builds upon. In Section 3 we develop our method, which we call
Continuous Orthogonal Matching Pursuit (COMP). In Section 4 we present the performance of our
method on simulated and neural data.
2
Continuous basis pursuit
Continuous Basis Pursuit (CBP) [4, 3, 5] is a method for recovering the time shifts and amplitudes
of waveforms present in a signal of the form (1). A key element of CBP is augmenting or replacing
the set {fn,j (t)} with certain additional dictionary elements that are chosen to smoothly interpolate
the one dimensional manifold traced out by fn,j (t ? ? ) as ? varies in (??/2, ?/2).
The benefit of a dictionary that is expanded in this way is twofold. First, it increases the ability
of the dictionary to represent shifted copies of the waveform fn (t ? ? ) without introducing as
much correlation as would be introduced by simply using a finer discretization (decreasing ?),
which is an advantage because dictionaries with smaller coherence are generally better suited for
sparse recovery techniques. Second, one can move from recovered coefficients in this augmented
dictionary to estimates an,j and continuous-valued estimates of ?n,j .
2
In general, there are three ingredients for CBP: basis elements, an interpolator with corresponding
mapping function ?, and a convex constraint set, C. There are K basis elements {gn,j,k (t) =
gn,k (t ? j?)}k=K
k=1 , for each waveform and width-? time bin, which together can be used to linearly
interpolate fn,j (t ? ? ), |? | < ?/2. The function ? maps from amplitude a and time shift ? to KPK (k)
(1)
(K)
tuples of coefficients ?(a, ? ) = (cn,j , ..., cn,j ), so afn,j (t ? ? ) ? k=1 cn,j gn,j,k (t). The convex
constraint set C is for K-tuples of coefficients of {gn,j,k }k=K
k=1 and corresponds to the requirement
that a > 0 and |? | < ?/2. If the constraint region corresponding to these requirements is not convex
(e.g. in the polar basis discussed below), its convex relaxation is used.
As a concrete example, let us first consider (as discussed in [4]) the dictionary augmented with
0
shifted copies of each waveform?s derivative : {fn,j
(t) := fn0 (t ? j?)}. Assuming fn is sufficiently
0
smooth, we have from the Taylor expansion that for small ? , afn,j (t ? ? ) ? afn,j (t) ? a? fn,j
(t). If
0
we recover a representation of y as c1 fn,j (t)+c2 fn,j (t), then we can estimate the amplitude a of the
waveform present in y as c1 , the time shift ? as ?c2 /c1 . Hence, we estimate y ? c1 fn,j (t+c2 /c1 ) =
c1 fn (t ? j? + c2 /c1 ). Note that the estimate of the time shift ? varies continuously with c1 , c2 .
In contrast, using shifted copies of the waveforms only as a basis would not allow for a time shift
j=N?
.
estimate off of the grid {j?}j=1
Once a suitable dictionary is chosen, one must still recover coefficients (i.e. c1 , c2 above). Motivated
by the assumed sparsity of the signal (i.e. y is the sum of relatively few shifted copies of waveforms,
so the coefficients of most dictionary elements will be zero), CBP draws on the basis pursuit denoising, which has been effective in the compressive sensing setting and elsewhere [10],[1]. Specifically,
CBP (with a Taylor basis) recovers coefficients using:
2
Nf
Nf
X
X
? (1)
(1)
(1)
(2)
(1)
0 (2)
argminc
(Fn cn + Fn cn ) ? y
+
?
c
? n, i (3)
c
n
s.t. cn,i ? 0 , |cn,i | ?
2 i,n
1
n=1
n=1
2
0
Here we denote by F the matrix with columns {fn,j (t)} and F0 the matrix with columns {fn,j
(t)}.
The `1 penalty encourages sparsity, pushing most of the estimated amplitudes to zero, with higher
(1)
? encouraging greater sparsity. Then, for each (n, j) such that cn,j 6= 0, one estimates that there is
(1)
(2)
(1)
a waveform in the shape of fn with amplitude a
? = cn,j and time shift j? ? ?? = j? ? cn,j /cn,j
present in the signal. The inequality constraints in the optimization problem ensure first that we only
recover positive amplitudes a
?, and second that estimates ?? satisfy |?
? | < ?/2. Requiring ?? to fall
in this range keeps the estimated ? in the time bin represented by fn,j and also in the regime where
they Taylor approximation to fn,j (t?? ) is accurate. Note that (3) is a convex optimization problem.
Better results in [4] are obtained for a second order Taylor interpolation and the best results come
from a polar interpolator, which represents each manifold of time-shifted waveforms fn,j (t ?
? ), |? | ? ?/2 as an arc of the circle that is uniquely defined to pass through fn,j (t), fn,j (t ? ?/2),
and fn,j (t+?/2). Letting the radius of the arc be r, and its angle be 2? one represents points on this
2?
arc by linear combinations of functions w, u, v: f (t ? ? ) ? w(t) + r cos( 2?
? ?)u(t) + r sin( ? ?)v(t).
The Taylor and polar bases consist of shifted copies of elements chosen in order to linearly interpolate the curve in function space defined by fn (t ? ? ) as ? varies from ??/2 to ?/2. Let Gn,k be
the matrix whose columns are gn,j,k (t) for j ? 1, ..., N? . With choices of basis elements, interpolator, and corresponding convex constraint set C in place, one proceeds to estimate coefficients in
the chosen basis by solving:
2
Nf K
Nf
X
X
X
(1)
(K)
(k)
argminc
y ?
Gn,k cn
+ ?k
c(1)
n k1 subject to (cn,j , ..., cn,j ) ? C ?(n, j)
n=1 k=1
n=1
(4)
2
(1)
(K)
One then maps back from each nonzero K-tuple of recovered coefficients cn,j , ..., cn,j to corresponding a
?n,j , ??n,j that represent the amplitude and timing of the nth waveform present in
the jth time bin. This can be done by inverting ?, if possible, or estimating (?
an,j , ??n,j ) =
(1)
(K) 2
argmina,? k?(a, ? ) ? (cn,j , ..., cn,j )k2 .
3
Table 1: Basis choices (see also [4], Table 1.)
Interpolator
Basis Vectors
?(a, ? )
2
C
Taylor
(K=3)
0
(t)},
{fn,j (t)}, {fn,j
00
{fn,j (t)}
(a, ?a?, a ?2 )
c(1) , c(3) > 0, |c(2) | < c(1) ?
2,
2
|c(3) | < c(1) ?8
Polar
{wn,j }, {un,j },
{vn,j }
(a, ar cos( 2?
? ?),
ar sin( 2?
?))
?
p
c(1) ? 0, (c(2) )2 + (c(3) )2 ? rc(1)
rc(1) cos(?) ? c(2) ? rc(1)
SVD
{u1n,j }...{uK
n,j }.
(See Section 3.1)
(See Section 3.1)
3
Continuous Orthogonal Matching Pursuit
We now present our method for recovery, which makes use of the idea of augmented bases presented
above, but differs from CBP in several important ways. First, we introduce a different choice of basis
that we find enables more accurate estimates. Second, we make use of a greedy method that iterates
between choosing basis vectors and estimating time shifts and amplitudes, rather than proceeding
via a single convex optimization problem as CBP does. Lastly, we introduce an alternative to the
step of mapping back from recovered coefficients via ? that notably improves the accuracy of the
recovered time estimates.
Greedy methods such as Orthogonal Matching Pursuit (OMP) [11], Subspace Pursuit [2], and Compressive Sampling Matching Pursuit (CoSaMP) [8] have proven to be fast and effective in the realm
of compressed sensing. Since the number of iterations of these greedy methods tend to go as the
sparsity (when the algorithms succeed), they tend to be extremely fast when for very sparse signals. Moreover, our the greedy method eliminates the need to choose a regularization constant ?,
a choice that can vastly alter the effectiveness of CBP. (We still need to choose K and ?.) Our
method is most closely analogous to OMP, but recovers continuous time estimates, so we call it
Continuous Orthogonal Matching Pursuit (COMP). However, the steps below could be adapted in a
straightforward way to create analogs of other greedy methods.
3.1
Choice of finite basis
We build upon [4], choosing as our basis N? shifted copies of a set of K basis vectors for each
waveform in such away that these K basis vectors can effectively linearly interpolate fn (t ? ? )
for |? | < ?/2. In our method, as in Continuous Basis Pursuit, these basis vectors allow us to
represent continuous time shifts instead of discrete time shifts, and expand the descriptive power of
our dictionary without introducing undue amounts of coherence. While previous work introduced
Taylor and polar bases, we obtain the best recovery from a different basis, which we describe now.
The basis comes from a singular value decomposition of a matrix whose columns correspond to
discrete points on the curve in function space traced out by fn,j (t ? ? ) as we vary ? for |? | < ?/2.
Within one time bin of size ?, consider discretizing further into N? = ?/? time bins of size ? ?.
Let F? be the matrix with columns that are these (slightly) shifted copies of the waveform, so that
the ith column of F? is fn,j (t ? i? + ?/2) for a discrete vector of time points t. Each column of
this matrix is a discrete point on the curve traced out by fn,j (t ? ? ) as ? varies.
In choosing a basis, we seek the best choice of K vectors to use to linearly interpolate this curve. We
might instead seek to solve the related problem of finding the best K vectors to represent these finely
spaced points on the curve, in which case a clear choice for these K vectors is the first K left singular
vectors of F? . This choice is optimal in the sense that the singular value decomposition yields the
best rank-K approximation to a matrix. If F? = U?VT is the singular value decomposition, and
PK
uk , vk are the columns of U and V respectively, then kF? ? k=1 uk ?k,k (vk )T k ? kF ? Ak for
any rank-K matrix A and any unitarily invariant norm k ? k.
4
In order to use this SVD basis with CBP or COMP, one must specify a convex constraint set for the
PK
coefficients of this basis. Since afn,j (t ? i?) = k=1 auk ?k,k vik a reasonable and simply enforced
constraint set would be to assume that the recovered coefficients c(k) corresponding to each basis
vector uk , when divided by c(1) to account for scaling, be between mini ?k,k vik and maxi ?k,k vik . A
PK
simple way to recover a and ? would to choose ? = i? and a, i to minimize k=1 (c(k) ?a?k,k vik )2 .
In figure 3.1, we compare the error between shifted copies of a sample waveform f (t ? ? ) for
|? | < 0.5 and the best (least-squares) approximation of that waveform as a linear combination of
K = 3 vectors from the Taylor, polar, and SVD bases. The structure of the error as a function of the
time shift ? reflects the structure of these bases. The Taylor approximation is chosen to be exactly
accurate at ? = 0 while the polar basis is chosen to be precisely accurate at ? = 0, ?/2, ??/2. The
SVD basis gives the lowest mean error across time shifts.
Original
Waveform
0
0.5
Taylor
2
5
0
t
5
1
Polar
0
0
0
2
1
0.2
5
0
t
5
5
0
t
SVD
0.2
5
0.08
l2 error
f(t)
0.5
Approximation Error
Basis Vectors
0.06
Taylor
Polar
SVD
0.04
0.02
5
0
t
5
0.5
0
time shift
0.5
Figure 1: Using sample waveform f (t) ? t exp(?t2 ) (left panel), we compare the error introduced
by approximating f (t ? ? ) for varying ? with a linear combination of K = 3 basis vectors, from the
Taylor, polar or SVD bases. Basis vectors are shown in the middle three panels, and error in the far
right panel. The SVD basis introduces the least error on average over the shift ? . The average errors
for the Taylor, polar, and SVD bases are 0.026, 0.027, and 0.014 respectively.
3.2
Greedy recovery
Taylor: 0.027
Polar: 0.027
recoverSVD:
the time
bins
0.014
Having chosen our basis, we then greedily
in which an occurrence of each
waveform appears to be present. We would like to build up a set of pairs (n, j) corresponding to
an instance of the nth waveform in the j th time bin. (In our third step, we will refine the estimate
within the chosen bins.)
Our greedy method is motivated by Orthogonal Matching Pursuit (OMP), which is used to recover a
sparse solution x from measurements y = Ax. In OMP [11], one greedily adds a single dictionary
element to an estimated support set S at each iteration, and then projects orthogonally to adjust the
coefficients of all chosen dictionary elements. After initializing with S = ?, x = 0, one iterates the
following until a stopping criterion is met:
r = y ? Ax
j = argmaxj {|haj , ri| s.t. j ? {1, ...J}\S}
S = S ? {j}
x = argminz {||y ? Az||2 s.t. zi = 0 ? i ?
/ S}
If we knew the sparsity of the signal, we could use that as our stopping condition. Normally we do
not know the sparsity a priori; we stop when changes in the residual become sufficiently small.
We adjust this method to choose at each step not a single additional element but rather a set of
K associated basis vectors. S is again initialized to be empty, but at each step we add a timebin/waveform pair (n, j), which is associated with K basis vectors. In this way, we are adding K
vectors at each step, instead of one as in OMP. We greedily add the next index (n, j) according to:
(
)
k
X
(k) (k)
2
c
(n, j) = argminm,i min{k
cm,i gm,i ? rk2 s.t. cm,i ? C} , (m, i) ? S
(5)
cm,i
i=1
5
(k)
Here {gm,i } are the chosen basis vectors (Taylor, polar, or SVD), and C is the corresponding constraint set, as in Section 2.
In comparison with the greedy step in OMP, choosing j as in (5) is more costly, because we need
to perform a constrained optimization over a K dimensional space for each n, j. Fortunately, it is
not necessary to repeat the optimization for each of the Nf ? N? possible indices each time we add
an index. Assuming waves are localized in time, we need only update the results of the constrained
optimization locally. When we update the residual r by subtracting the newly identified waveform
n in the j th bin, the residual only changes in the bins at or near the j th bin, so we need only update
Pk
(k) (k)
the quantity mincn,j0 {k i=1 cn,j 0 gn,j 0 ? rk22 s.t. cn,j 0 ? C } for j 0 neighboring j.
3.3
Estimating time shifts
Having greedily added a new waveform/timebin index pair (n, j), we next define our update step,
which will correspond to the orthogonal projection in OMP. We present two alternatives, one of
which most closely mirrors the corresponding step in OMP, the other of which works within the
Fourier domain to obtain more accurate recovery.
To most closely follow the steps of OMP, at each iteration after updating S we update coefficients c
according to:
2
X X
K
(k) (k)
cn,j gn,j ? y
(6)
argminc
subject to cn,j ? C ? (n, j) ? S
(n,j)?S k=1
2
We alternate between the greedily updating S via (5), and updating c as in (6), at each iteration
P
PK (k) (k)
finding the new residual r = (n,j)?S k=1 cn,j gn,j ?y ) until the `2 stopping criterion is reached.
Then, one maps back from {cn,j }(n,j)?S to {a(n,j) , ?(n,j) }(n,j)?S as described in Section 2.
Alternatively we may replace the orthogonal projection step with a more accurate recovery of spike
timings that involves working in the Fourier domain. We use the property of the Fourier transform
with respect to translation that: (f (t ? ? ))? = e2?i? f?. This allows us to estimate a, ? directly via:
X
argmina,? k(
an,j e2?i??n,j f?n,j (?)) ? y?(?)k2 subject to |?n,j | < ?/2 ? (n, j) ? S
(7)
n,j?S
This is a nonlinear and non-convex constrained optimization problem. However, it can be solved reasonably quickly using, for example, trust region methods. The search space is dramatically reduced
because ? has only |S| entries, each constrained to be small in absolute value. By searching directly
for a, ? as in (7) we sacrifice convexity, but with the benefit of eliminating from this step error of
interpolation introduced as we map back from c to a, ? using ??1 or a least squares estimation.
It is easy and often helpful to add inequality constraints to a as well, for example requiring a to be
in some interval around 1, and we do impose this in our spike-sorting simulations and analysis in
Section 4. Such a requirement effectively imposes a uniform prior on a over the chosen interval. It
would be an interesting future project to explore imposing other priors on a.
4
Results
We test COMP and CBP for each choice of basis on simulated and neural data. Here, COMP denotes
the greedy method that includes direct estimation of a and ? during the update set as in (7). The
convex optimization for CBP is implemented using the cvx package for MATLAB [7], [6].
4.1
Simulated data
We simulate a signal y as the sum of time-shifted copies of two sample waveforms f1 (t) ?
4
2
t exp(?t2 ) and f2 (t) ? e?t /16 ? e?t (Figure 2a). There are s1 = s2 = 5 shifted copies of
f1 and f2 , respectively. The time shifts are independently generated for each of the two waveforms
using a Poisson process (truncated after 5 spikes), and independent Gaussian noise of variance ? 2 is
6
0
0.5
0
5
(b)
0
0.5
5
t
5
0
t
5
1
0.5
1
0
20
40 t 60
80
100
waveform 1
0.5
1
0.5
0
waveform 2
0
0.5
1
0
20
40
t 60
100
20
40
60
80
100
True
CBP SVD
1
2.5
CBP Taylor
CBP Polar
CBP SVD
COMP Taylor
COMP Polar
COMP SVD
2
1.5
1
0.5
0.5
0
0
0
20
40
60
80
100
COMP-SVD
1.5
(f )
0
.05
.1
Noise ( )
.2
.4
0.5
0.4
1
0.5
0
20
40
t
60
80
100
1.5
True
COMP SVD
1
0.3
0.2
0.1
0.5
0
80
0
1.5
0
0.5
(e)
1.5
(d)
0
1
CBP-SVD
(c)
(Misses + False Positives)/s
waveform 2
0.5
Average Hit Error
waveform 1
waveform 1
0.5
waveform 2
(a)
0
20
40
60
80
100
0
0
.05
.1
.2
.4
Noise ( )
Figure 2: (a) Waveforms present in the signal. (b) A noiseless (top) and noisy (bottom) signal with
? = .2. (c) Recovery using CBP. (d) Recovery using COMP (with a, ? updated as in (7)). (e) For
each recovery method over different values of the standard deviation of the noise ?, misses plus false
positives, divided by the total number of events present, s = s1 + s2 . (f) Average distance between
the true and estimated spike for each hit.
added at each time point. Figures 2b,c show an example noise-free signal (? = 0), and noisy signal
(? = .2) on which each recovery method will be run.
We run CBP with the Taylor and polar bases, but also with our SVD basis, and COMP with all three
bases. Since COMP here imposes a lower bound on a, we also impose a thresholding step after
recovery with CBP, discarding any recovered waveforms with amplitude less than .3. We find the
thresholding generally improved the performance of the CBP algorithm by pruning false positives.
Throughout, we use K = 3, since the polar basis requires 3 basis vectors per bin.
We categorize hits, false positive and misses based on whether a time shift estimate is within a
threshold of = 1 of the true value. The ?average hit error? of Figure 2h, 3b is the average distance
between the true and estimated event time for each estimate that is categorized as a hit. Results are
averaged over 20 trials.
We compare CBP and COMP over different parameter regimes, varying the noise (?) and the bin
size (?). Figures 2g and 3a show misses plus false positives for each method, normalized by the total
number of events present. Figures 2f and 3b show average distance between the true and estimated
spike for each estimate categorized as a hit. The best performance by both measures across nearly
all parameter regimes considered is achieved by COMP using the SVD basis. COMP is more robust
to noise (Figure 2g), and also to increases in bin width ?. Since both algorithms are faster for
higher ?, robustness with respect to ? is an advantage. We also note a significant increase in CBP?s
robustness to noise when we implement it with our SVD basis rather than with the Taylor or polar
basis (Figure 2e).
A significant advantage of COMP over CBP is its speed. In Figure 3c we compare the speed of
COMP (solid) and CBP (dashed) algorithms for each basis. COMP yields vast gains in speed. The
comparison is especially dramatic for small ?, where results are most accurate across methods.
4.2
Neural data
We now present recovery of spike times and identities from neural data. Recordings were made
using glass-coated tungsten electrodes in the lateral intraparietal sulcus (LIP) of a macaque monkey
performing a motion discrimination task. In addition to demonstrating the applicability of COMP
to sorting spikes in neural data, this section also shows the resistance of COMP to a certain kind of
error that recovery via CBP can systematically commit, and which is relevant to neural data.
7
0.8
0.7
1.5
1
0.5
(b)
500
0.5
0.4
0.3
0.2
1
1.5
Bin Width ( )
2
0
0.5
2.5
CBP Taylor
CBP Polar
CBP SVD
COMP Taylor
COMP Polar
COMP SVD
300
200
100
0.1
0
0.5
(c)
400
0.6
Computing Time
(a)
Average Hit Error
(Misses + False Positives)/s
2
1
1.5
Bin Width ( )
2
2.5
0
0.5
1
1.5
Bin Width ( )
2
2.5
Figure 3: (a) Misses plus false positives, divided by the total number of events present, s = s1 + s2
over different values of bin width ?. (b) Average distance between the true and estimated spike for
each hit for each recovery method. (c) Run time for COMP (solid) and CBP (dashed) for each basis.
0.3
0.4
0.5
0
(c)
COMP-SVD
CBP-SVD
0
1
0.5
0
0
10
20
30
40
50
60
70
80
90
100
1.5
1
0.5
0.5
1
1.5
2
time (ms)
0
0.5
0
20
1
0
0.5
0.1
0
40
60
time (ms)
80
100
0.1
0
10
20
30
40
50
60
time (ms)
1
Voltage Trace
0.5
0.5
1.5
Neuron 2
0.2
(Misses + False Positives)/s
0
Recovered Spikes
1.5
Neuron 1
Neuron 2
2
0.1
0.1
(b)
Waveforms
Neuron 1
(a)
1.5
Bin Width ( )
2
70
80
90
100
0.2
70
70.5
71
71.5
time (ms)
72
72.5
73
2.5
Figure 4: (a) Two neural waveforms; each is close to as scaled copy of the other (b) Recovery of
spikes via COMP (magenta) and CBP (cyan) using the SVD basis. CBP tends to recover smallamplitude instances of waveform one where COMP recovers large amplitude instances of waveform
two (c) Top: recovered traces. Lower panel: zooming in on an area of disagreement between COMP
and CBP. The large-ampltude copy of waveform two more closely matches the trace
In the data, the waveform of one neuron resembles a scaled copy of another (Figure 4a).The similarity causes problems for CBP or any other `1 minimization based method that penalizes large
amplitudes. When the second waveform is present with an amplitude of one, CBP is likely to incorrectly add a low-amplitude copy of the first waveform (to reduce the amplitude penalty), instead of
correctly choosing the larger copy of the second waveform; the amplitude penalty for choosing the
correct waveform can outweigh the higher `2 error caused by including the incorrect waveform.
This misassignment is exactly what we observe (Figure 4b). We see that CBP tends to report smallamplitude copies of waveform one where COMP reports large-amplitude copies of waveform two.
Although we lack ground truth, the closer match of the recovered signal to data (Figure 4c) indicates
that the waveform identities and amplitudes identified via COMP better explain the observed signal.
5
Discussion
We have presented a new greedy method called Continuous Orthogonal Matching Pursuit (COMP)
for identifying the timings and amplitudes for waveforms from a signal that has the form of a (noisy)
sum of shifted and scaled copies of several known waveforms. We draw upon the method of Continuous Basis Pursuit, and extend it in several ways. We leverage the success of Orthogonal Matching
Pursuit in the realm of sparse recovery, use a different basis derived from a singular value decomposition, and also introduce a move to the Fourier domain to fine-tune the recovered time shifts.
Our SVD basis can also be used with CBP and in our simulations it increased performance of CBP
as compared to previously used bases. In our simulations COMP obtains increased accuracy as
well as greatly increased speed over CBP across nearly all regimes tested. Our results suggest that
greedy methods of the type introduced here may be quite promising for, among other applications,
spike-sorting during the processing of neural data.
Acknowledgments
This work was supported by the McKnight Foundation (JP), NSF CAREER Award IIS-1150186
(JP), and grants from the NIH (NEI grant EY017366 and NIMH grant MH099611 to AH & JP).
8
References
[1] Scott Shaobing Chen, David L Donoho, and Michael A Saunders. Atomic decomposition by
basis pursuit. SIAM journal on scientific computing, 20(1):33?61, 1998.
[2] Wei Dai and Olgica Milenkovic. Subspace pursuit for compressive sensing signal reconstruction. Information Theory, IEEE Transactions on, 55(5):2230?2249, 2009.
[3] Chaitanya Ekanadham, Daniel Tranchina, and Eero P Simoncelli. A blind deconvolution
method for neural spike identification. In Proceedings of the 25th Annual Conference on Neural Information Processing Systems (NIPS11), volume 23, 2011.
[4] Chaitanya Ekanadham, Daniel Tranchina, and Eero P Simoncelli. Recovery of sparse
translation-invariant signals with continuous basis pursuit. Signal Processing, IEEE Transactions on, 59(10):4735?4744, 2011.
[5] D. Ekanadham, C.vand Tranchina and E. P. Simoncelli. A unified framework and method for
automatic neural spike identification. Journal of Neuroscience Methods, 222:47?55, 2014.
[6] M. Grant and S. Boyd. Graph implementations for nonsmooth convex programs. In V. Blondel,
S. Boyd, and H. Kimura, editors, Recent Advances in Learning and Control, Lecture Notes in
Control and Information Sciences, pages 95?110. Springer-Verlag Limited, 2008. http:
//stanford.edu/?boyd/graph_dcp.html.
[7] CVX Research Inc. CVX: Matlab software for disciplined convex programming, version 2.0.
http://cvxr.com/cvx, August 2012.
[8] Deanna Needell and Joel A Tropp. Cosamp: Iterative signal recovery from incomplete and
inaccurate samples. Applied and Computational Harmonic Analysis, 26(3):301?321, 2009.
[9] Jonathan W Pillow, Jonathon Shlens, EJ Chichilnisky, and Eero P Simoncelli. A model-based
spike sorting algorithm for removing correlation artifacts in multi-neuron recordings. PloS
one, 8(5):e62123, 2013.
[10] Robert Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal
Statistical Society. Series B (Methodological), pages 267?288, 1996.
[11] Joel A Tropp and Anna C Gilbert. Signal recovery from random measurements via orthogonal
matching pursuit. Information Theory, IEEE Transactions on, 53(12):4655?4666, 2007.
[12] Martin Vetterli, Pina Marziliano, and Thierry Blu. Sampling signals with finite rate of innovation. Signal Processing, IEEE Transactions on, 50(6):1417?1428, 2002.
9
| 5264 |@word trial:1 milenkovic:1 version:1 middle:1 eliminating:1 norm:1 blu:1 seek:6 simulation:3 jacob:1 decomposition:5 dramatic:1 solid:2 series:1 selecting:2 daniel:2 recovered:10 discretization:2 com:1 must:2 fn:36 shape:1 enables:1 update:6 discrimination:1 greedy:14 afn:4 ith:1 iterates:2 math:1 cbp:47 rc:3 c2:6 direct:1 become:1 incorrect:1 compose:1 introduce:3 blondel:1 sacrifice:1 notably:1 roughly:1 multi:2 discretized:3 inspired:1 decreasing:1 resolve:1 encouraging:1 considering:1 begin:1 estimating:3 moreover:2 project:2 panel:4 lowest:1 what:1 cm:3 kind:1 monkey:1 compressive:3 unified:1 finding:2 kimura:1 guarantee:1 nf:8 exactly:2 scaled:4 k2:2 uk:4 hit:8 normally:1 grant:4 control:2 producing:1 positive:10 timing:7 tends:2 ak:1 interpolation:2 might:1 plus:3 studied:1 resembles:1 argminc:3 co:3 limited:1 tungsten:1 range:1 averaged:1 acknowledgment:1 atomic:1 implement:1 differs:1 j0:1 area:1 matching:13 projection:2 boyd:3 suggest:1 rk22:1 close:1 selection:1 argminm:1 outweigh:1 map:4 gilbert:1 center:1 go:1 straightforward:1 independently:1 convex:15 recovery:24 identifying:1 needell:1 array:1 shlens:1 population:1 searching:1 analogous:2 updated:1 target:1 suppose:1 gm:2 programming:1 us:1 element:13 updating:4 tranchina:3 observed:3 bottom:1 initializing:1 capture:1 solved:1 region:2 plo:1 convexity:1 nimh:1 ultimately:1 solving:3 upon:3 f2:2 basis:60 translated:2 represented:2 train:1 fast:3 describe:3 effective:3 aggregate:1 choosing:7 saunders:1 whose:4 quite:1 larger:1 valued:5 solve:1 stanford:1 drawing:1 compressed:3 ability:1 commit:1 transform:1 noisy:4 advantage:3 descriptive:1 reconstruction:1 subtracting:1 neighboring:1 relevant:1 az:1 electrode:3 requirement:3 cosamp:2 empty:1 produce:1 develop:1 augmenting:2 thierry:1 recovering:3 implemented:1 involves:2 come:2 karin:1 met:1 waveform:54 radius:1 closely:5 correct:1 jonathon:1 bin:21 f1:2 decompose:1 secondly:1 sufficiently:2 considered:2 around:1 ground:1 exp:2 great:1 mapping:2 dictionary:16 vary:1 polar:20 estimation:2 utexas:3 mh099611:1 create:1 tool:1 reflects:1 minimization:2 rough:1 gaussian:1 rather:4 ej:1 shrinkage:1 varying:2 voltage:3 ax:2 derived:1 refining:1 vk:2 methodological:1 rank:2 indicates:1 greatly:1 contrast:1 greedily:9 sense:2 glass:1 helpful:1 stopping:4 inaccurate:1 expand:1 among:1 html:1 augment:1 priori:1 undue:1 constrained:4 once:2 having:2 sampling:2 represents:2 nearly:2 alter:1 future:1 report:2 auk:1 t2:2 simplify:1 few:1 nonsmooth:1 interpolate:5 interest:1 highly:1 joel:2 adjust:2 introduces:2 accurate:9 tuple:1 closer:1 necessary:1 orthogonal:14 incomplete:1 taylor:21 initialized:1 desired:1 circle:1 penalizes:1 chaitanya:2 instance:4 column:8 modeling:1 increased:3 gn:10 ar:2 tractability:1 introducing:2 deviation:1 entry:1 applicability:1 ekanadham:3 uniform:1 too:1 optimally:1 varies:4 siam:1 off:1 michael:1 together:1 continuously:1 concrete:1 quickly:1 vastly:1 again:1 recorded:1 choose:4 derivative:1 return:1 account:1 potential:2 includes:1 coefficient:13 inc:1 satisfy:1 caused:1 blind:1 reached:2 haj:1 recover:7 wave:1 vivo:1 minimize:1 square:2 fn0:1 accuracy:3 variance:1 spaced:2 yield:3 correspond:2 identification:2 graph_dcp:1 comp:34 finer:1 ah:1 explain:1 kpk:1 e2:2 associated:2 recovers:3 gain:2 sampled:1 stop:1 newly:1 realm:3 improves:1 electrophysiological:1 vetterli:1 amplitude:26 carefully:1 back:4 appears:1 higher:3 follow:1 specify:1 improved:1 wei:1 disciplined:1 done:1 lastly:1 until:3 correlation:2 working:1 tropp:2 replacing:2 trust:1 nonlinear:2 lack:1 artifact:1 scientific:1 unitarily:1 requiring:2 rk2:1 true:7 normalized:1 hence:1 regularization:1 iteratively:1 nonzero:1 sin:2 during:2 width:7 encourages:1 uniquely:1 criterion:3 m:4 motion:1 bring:1 harmonic:1 nih:1 jp:3 volume:1 extend:3 analog:2 occurred:1 discussed:2 measurement:3 significant:2 imposing:1 automatic:1 grid:3 mathematics:1 moving:2 access:1 f0:1 similarity:1 base:11 add:7 argmina:2 own:1 recent:1 certain:2 verlag:1 inequality:2 discretizing:1 success:2 vt:1 additional:2 greater:1 fortunately:1 omp:10 impose:2 dai:1 signal:31 dashed:2 ii:1 simoncelli:4 smooth:1 faster:2 match:2 divided:3 award:1 regression:1 noiseless:1 poisson:1 iteration:4 represent:5 achieved:1 c1:9 background:1 addition:1 fine:1 interval:2 singular:5 eliminates:1 finely:2 recording:4 suspect:1 subject:3 tend:2 effectiveness:1 call:3 near:1 leverage:2 easy:1 wn:1 iterate:1 psychology:2 zi:1 identified:2 lasso:1 reduce:1 idea:2 cn:24 vik:4 texas:3 shift:20 whether:1 motivated:3 penalty:3 resistance:1 speaking:2 cause:1 action:2 matlab:2 dramatically:1 generally:3 clear:1 tune:1 amount:1 extensively:1 locally:1 argminz:1 reduced:1 http:2 nsf:1 shifted:12 neuroscience:4 estimated:7 per:1 intraparietal:1 correctly:1 tibshirani:1 broadly:1 discrete:8 yates:1 key:1 threshold:1 demonstrating:1 traced:3 interpolator:4 sulcus:1 vast:1 graph:1 circumventing:1 relaxation:1 sum:5 enforced:1 run:3 angle:1 package:1 place:1 throughout:1 reasonable:1 vn:1 cvx:4 draw:3 coherence:4 scaling:1 bound:1 cyan:1 refine:1 annual:1 activity:1 adapted:1 precisely:2 constraint:9 ri:1 software:1 shaobing:1 fourier:5 speed:6 simulate:1 extremely:1 min:1 performing:1 expanded:1 relatively:1 martin:1 department:4 according:2 alternate:1 combination:5 mcknight:1 smaller:1 slightly:1 across:4 s1:3 invariant:2 previously:2 argmaxj:1 know:1 letting:1 pursuit:27 observe:1 away:1 disagreement:1 occurrence:1 alternative:2 robustness:2 original:1 denotes:1 top:2 ensure:1 pushing:1 k1:1 build:3 especially:1 approximating:2 society:1 move:3 added:2 quantity:1 spike:17 costly:1 dependence:1 subspace:3 distance:4 zooming:1 simulated:4 lateral:1 manifold:2 assuming:2 modeled:1 index:4 mini:1 innovation:1 difficult:1 robert:1 trace:5 implementation:1 perform:1 discretize:1 neuron:8 arc:3 finite:4 truncated:1 incorrectly:1 arbitrary:1 nei:1 august:1 introduced:7 inverting:1 pair:3 david:1 chichilnisky:1 coherent:1 framing:1 macaque:1 deanna:1 proceeds:2 below:3 scott:1 regime:5 sparsity:6 program:1 including:1 royal:1 power:1 event:6 suitable:1 difficulty:1 residual:4 nth:2 orthogonally:1 prior:2 l2:1 kf:2 lecture:1 bear:1 interesting:1 proven:1 borrows:1 ingredient:1 localized:1 foundation:1 imposes:2 olgica:1 thresholding:2 editor:1 systematically:1 translation:7 austin:3 elsewhere:1 repeat:1 supported:1 copy:20 free:1 jth:1 allow:2 institute:1 fall:1 absolute:1 sparse:13 benefit:2 curve:5 pillow:3 made:1 far:1 transaction:4 approximate:2 pruning:1 obtains:1 keep:1 assumed:1 eero:3 tuples:2 knew:1 alternatively:1 coated:1 continuous:21 un:1 search:1 iterative:1 table:2 lip:1 promising:1 channel:1 reasonably:1 robust:1 career:1 huk:2 expansion:1 domain:4 anna:1 pk:5 stereotyped:1 linearly:4 s2:3 noise:9 cvxr:1 categorized:2 augmented:4 inferring:1 wish:1 candidate:2 perceptual:1 third:2 magenta:1 removing:1 discarding:1 sensing:5 list:1 maxi:1 deconvolution:1 consist:1 false:8 adding:2 effectively:2 mirror:1 demand:1 sorting:5 chen:1 suited:1 smoothly:1 simply:2 knudson:1 explore:1 likely:1 partially:1 springer:1 corresponds:1 truth:1 succeed:1 identity:4 donoho:1 twofold:1 replace:1 change:2 specifically:2 denoising:3 miss:7 total:3 called:1 pas:1 svd:25 select:1 support:1 jonathan:2 alexander:1 categorize:1 princeton:3 tested:1 |
4,710 | 5,265 | Fast Sampling-Based Inference in Balanced Neuronal
Networks
Guillaume Hennequin1
[email protected]
Laurence Aitchison2
[email protected]
M?at?e Lengyel1
[email protected]
1
2
Computational & Biological Learning Lab, Dept. of Engineering, University of Cambridge, UK
Gatsby Computational Neuroscience Unit, University College London, UK
Abstract
Multiple lines of evidence support the notion that the brain performs probabilistic
inference in multiple cognitive domains, including perception and decision making. There is also evidence that probabilistic inference may be implemented in the
brain through the (quasi-)stochastic activity of neural circuits, producing samples
from the appropriate posterior distributions, effectively implementing a Markov
chain Monte Carlo algorithm. However, time becomes a fundamental bottleneck
in such sampling-based probabilistic representations: the quality of inferences depends on how fast the neural circuit generates new, uncorrelated samples from
its stationary distribution (the posterior). We explore this bottleneck in a simple, linear-Gaussian latent variable model, in which posterior sampling can be
achieved by stochastic neural networks with linear dynamics. The well-known
Langevin sampling (LS) recipe, so far the only sampling algorithm for continuous variables of which a neural implementation has been suggested, naturally fits
into this dynamical framework. However, we first show analytically and through
simulations that the symmetry of the synaptic weight matrix implied by LS yields
critically slow mixing when the posterior is high-dimensional. Next, using methods from control theory, we construct and inspect networks that are optimally fast,
and hence orders of magnitude faster than LS, while being far more biologically
plausible. In these networks, strong ? but transient ? selective amplification of
external noise generates the spatially correlated activity fluctuations prescribed by
the posterior. Intriguingly, although a detailed balance of excitation and inhibition
is dynamically maintained, detailed balance of Markov chain steps in the resulting
sampler is violated, consistent with recent findings on how statistical irreversibility can overcome the speed limitation of random walks in other domains.
1
Introduction
The high speed of human sensory perception [1] is puzzling given its inherent computational complexity: sensory inputs are noisy and ambiguous, and therefore do not uniquely determine the state
of the environment for the observer, which makes perception akin to a statistical inference problem.
Thus, the brain must represent and compute with complex and often high-dimensional probability
distributions over relevant environmental variables. Most state-of-the-art machine learning techniques for large scale inference trade inference accuracy for computing speed (e.g. [2]). The brain,
on the contrary, seems to enjoy both simultaneously [3].
Some probabilistic computations can be made easier through an appropriate choice of representation for the probability distributions of interest. Sampling-based representations used in Monte Carlo
1
techniques, for example, make computing moments of the distribution or its marginals straightforward. Indeed, recent behavioural and neurophysiological evidence suggests that the brain uses such
sampling-based representations by neural circuit dynamics implementing a Markov chain Monte
Carlo (MCMC) algorithm such that their trajectories in state space produce sequential samples from
the appropriate posterior distribution [4, 5, 6].
However, for sampling-based representations, speed becomes a key bottleneck: computations involving the posterior distribution become accurate only after enough samples have been collected,
and one has no choice but to wait for those samples to be delivered by the circuit dynamics. For
sampling to be of any practical use, the interval that separates the generation of two independent
samples must be short relative to the desired behavioral timescale. Single neurons can integrate
their inputs on a timescale ?m ? 10 ? 50 ms, whereas we must often make decisions in less than
a second: this leaves just enough time to use (i.e. read out) a few tens of samples. What kinds of
neural circuit dynamics are capable of producing uncorrelated samples at ?100 Hz remains unclear.
Here, we introduce a simple yet non-trivial generative model and seek plausible neuronal network
dynamics for fast sampling from the corresponding posterior distribution. While some standard
machine learning techniques such as Langevin or Gibbs sampling do suggest ?neural network?type solutions to sampling, not only are the corresponding architectures implausible in fundamental
ways (e.g. they violate Dale?s law), but we show here that they lead to unacceptably slow mixing
in high dimensions. Although the issue of sampling speed in general is well appreciated in the
context of machine learning, there have been no systematic approaches to tackle it owing to a large
part to the fact that sampling speed can only be evaluated empirically in most cases. In contrast,
the simplicity of our generative model allowed us to draw an analytical picture of the problem
which in turn suggested a systematic approach for solving it. Specifically, we used methods from
robust control to discover the fastest neural-like sampler for our generative model, and to study its
structure. We find that it corresponds to greatly non-symmetric synaptic interactions (leading to
statistical irreversibility), and mathematically nonnormal1 circuit dynamics [7, 8] in close analogy
with the dynamical regime in which the cortex has been suggested to operate [9].
2
Linear networks perform sampling under a linear Gaussian model
We focus on a linear Gaussian latent variable model which generates observations h ? RM as
weighted sums of N features A ? (a1 ; . . . ; aN ) ? RM ?N with jointly Gaussian coefficients r ?
RN , plus independent additive noise terms (Fig. 1, left). More formally:
p(r) = N (r; 0, C)
and
p(h|r) = N h; Ar, ?h2 I
(1)
where I denotes the identity matrix. The posterior distribution is multivariate Gaussian, p(r|h) =
N (r; ?(h), ?), with
?1
? = C?1 + A> A/?h2
and
?(h) = ?A> h/?h2 .
(2)
where we made explicit the fact that under this simple model, only the mean, ?(h), but not the
covariance of the posterior, ?, depends on the input, h.
We are interested in neural circuit dynamics for sampling from p(r|h), whereby the data (observation) h is given as a constant feedforward input to a population of recurrently connected neurons,
each of which encodes one of the latent variables and also receives inputs from an external, private
source of noise ? (Fig. 1, right). Our goal is to devise a network such that the activity fluctuations
r(t) in the recurrent layer have a stationary distribution that matches the posterior, for any h.
Specifically, we consider linear recurrent stochastic dynamics of the form:
r
dt
2
dr =
[?r(t) + Wr(t) + Fh] + ??
d?(t)
?m
?m
(3)
where ?m = 20 ms is the single-unit ?membrane? time constant, and d? is a Wiener process of unit
variance, which is scaled by a scalar noise intensity ?? . The activity ri (t) could represent either the
1
?Nonnormal? should not be confused with ?non-Gaussian?: a matrix M is nonnormal iff MM> 6=
M> M.
2
Posterior sampling:
noise ?
Linear Gaussian latent variable model:
latent
variables
r
P(r) = N (r; 0, C)
observations
h
P(h|r) = N h; Ar, ?h2 I
network r(t)
input h(t)
Figure 1: Sampling under a
linear Gaussian latent variable model using neuronal
Left:
W network dynamics.
schematics of the generative
F
model. Right: schematics of
the recognition model. See text
for details.
membrane potential of neuron i, or the deviation of its momentary firing rate from a baseline. The
matrices F and W contain the feedforward and recurrent connection weights, respectively.
r
?1
The stationary distribution
of r is rindeed Gaussian
with a mean ? (h) = (I ? W) Fh and a cor
r >
variance matrix ? ? (r(t) ? ? )(r(t) ? ? ) t . For the following, we will use the dependence
of ?r on W (and ?? ) given implicitly by the following Lyapunov equation [10]:
(W ? I)?r + ?r (W ? I)> = ?2??2 I
(4)
Note that in the absence of recurrent connectivity (W = 0), the variance of every ri (t) would be
exactly ??2 . Note also that, just as required (see above), only the mean, ?r (h), but not the covariance,
?r , depends on the input, h.
In order for the dynamics of Eq. 3 to sample from the correct posteriors, we must choose F, W and
?? such that ?r (h) = ?(h) for any h, and ?r = ?. One possible solution (which, importantly, is
not unique, as we show later) is
2
F = (?? /?h ) A>
W = WL ? I ? ??2 ??1
and
(5)
with arbitrary ?? > 0.
In the following, we will be interested in the likelihood matrix A only insofar as it affects the
posterior covariance matrix ?, which turns out to be the main determinant of sampling speed. We
will therefore directly choose some covariance matrix ?, and set h = 0 without loss of generality.
3
Langevin sampling is very slow
Langevin sampling (LS) is a common sampling technique [2, 11, 12], and in fact the only one that
has been proposed to be neurally implemented for continuous variables [6, 13]. According to LS, a
stochastic dynamical system performs ?noisy gradient ascent of the log posterior?:
dr =
1 ?
log p(r|h) dt + d?
2 ?r
(6)
where d? is a unitary Wiener process. When r|h is Gaussian, Eq. 6 reduces to Eq. 3 for ?? = 1 and
the choice of F and W given in Eq. 5 ? hence the notation WL above. Note that WL is symmetric.
As we show now, this choice of weight matrix leads to critically slow mixing (i.e. very long autocorrelation time scales in r(t)) when N is large. In a linear network, the average autocorrelation
length is dominated by the decay time constant ?max of the slowest eigenmode, i.e. the eigenvector
of (W ? I) associated with the eigenvalue ?W?I
max which, of all the eigenvalues of (W ? I), has the
largest real part (which must still be negative, to ensure stability). The contribution
of the slowest
W?I
eigenmode to the sample autocorrelation
time
is
?
=
??
/Re
?
,
so
sampling
becomes
max
m
max
very slow when Re ?W?I
approaches
0.
This
is,
in
fact,
what
happens
with
LS
as
N
?
?. Inmax
deed, we could derive the following generic lower bound (details can be found in our Supplementary
Information, SI):
?(?? /?0 )2
WL ?I
(7)
?max
? p
1 + N ?r2
which is shown as dashed lines in Fig. 2. Thus, LS becomes infinitely slow in the large N limit
1
when pairwise correlations do not vanish in that limit (or at least not as fast as N ? 2 in their std.).
Slowing becomes even worse when ? is drawn from the inverse Wishart distribution with ? degrees
of freedom and scale matrix ? ?2 I (Fig. 2). We choose ? = N ?1+b?r?2 c and ? ?2 = ?02 (? ?N ?1)
3
1000
?max /?m
simulation (inverse Wishart)
theory (inverse Wishart)
lower bound (general)
0
L ?I
?W
max
slowing
100
factor
-0.2
-0.4
-0.6
10
?r = 0.10
?r = 0.20
-0.8
(? N (0, ?r ))
-1
1
1
10 100 1000
network size N
1
10 100 1000
network size N
-1 -0.5 0 0.5 1
pairwise corr.
Figure 2: Langevin sampling (LS) is slow in high-dimension. Random covariance matrices ? of
size N are drawn from an inverse Wishart distribution with parameters chosen such that the average
diagonal element (variance) is ?02 = 1 and the distribution of pairwise correlations has zero mean
and variance ?r2 (right). Sampling from N (0, ?) using a stochastic neural network (cf. Fig. 1) with
W = WL (LS, symmetric solution) becomes increasingly slow as N grows, as indicated by the
relative decay time constant ?max /?m of the slowest eigenmode of (WL ? I) (left), which is also
the negative inverse of its largest eigenvalue (middle). Dots indicate the numerical evaluation of the
corresponding quantities, and errorbars (barely noticeable) denote standard deviation across several
random realizations of ?. Dashed lines correspond to the generic bound in Eq. 7. Solid lines are
obtained from random matrix theory under the asssumption that ? is drawn from an inverse Wishart
distribution (Eq. 8). Parameters: ?? = ?0 = 1.
such that the expected value of a diagonal element (variance) in ? is ?02 , and the distribution of
pairwise correlations is centered on zero with variance ?r2 . The asymptotic behavior of the largest
eigenvalue of ??1 (the square of the smallest singular value of a random ? ? N rectangular matrix)
is known from random matrix theory (e.g. [14]), and we have for large N :
q
2
?
(?? /?0 )2
1
?2
L ?I
?W
?
?
(8)
N
?
1
+
b?
c
?
N
?
?O
r
max
N
b?r?2 c ? 2
This scaling behavior is shown in Fig. 2 (solid lines). In fact, we can also show (cf. SI) that LS is
(locally) the slowest possible choice (see Sec. 4 below for a precise definition of ?slowest?, and SI
for details).
Note that both Eqs. 7-8 are inversely proportional to the ratio (?0 /?? ), which tells us how much
the recurrent interactions must amplify the external noise in order to produce samples from the
right stationary activity distribution. The more amplification is required (?0 ?? ), the slower the
dynamics of LS. Conversely, one
? could potentially make Langevin sampling faster by increasing ?? ,
but ?? would need to scale as N to annihilate the critical slowing problem. This ? in itself ? is
unrealistic; moreover, it would also require the resulting connectivity matrix to have a large negative
diagonal (O(?N )) ? ie. the intrinsic neuronal time constant ?m to scale as O(1/N ) ?, which is
perhaps even more unrealistic.2
Note also that LS can be sped up by appropriate ?preconditioning? (e.g. [15, 16]), for example using
the inverse Hessian of the log-posterior. In our case, a simple calculation shows that this corresponds
to removing all recurrent connections, and pushing the posterior covariance matrix to the external
noise sources, which is only postponing the problem to some other brain network.
Finally, LS is fundamentally implausible as a neuronal implementation: it imposes symmetric synaptic interactions, which is simply not possible in the brain due to the existence of distinct classes of
excitatory and inhibitory neurons (Dale?s principle). In the following section, we show that networks
can be constructed that overcome all the above limitations of LS in a principled way.
4
General solution and quantification of sampling speed
While Langevin dynamics (Eq. 6) provide a general recipe for sampling from any given posterior
density, they unduly constrain the recurrent interactions to be symmetric ? at least in the Gaussian
2
From a pure machine learning perspective, increasing ?? is not an option either: the increasing stiffness of
Eq. 6 would either require the use of a very small integration step, or would lead to arbitrarily small acceptance
ratios in the context of Metropolis-Hastings proposals.
4
Langevin
optimal
optimal E/I
random S (? = 0.2; 0.4; 0.8; 1.6)
Newton (unconnected net)
Gibbs (update time ?m )
0 2 4 6 8 10
time lag k (units of ?m )
B
1
0.1
S ? N (0, ? 2 )
C
weight RMS
kK(k ?m )kF
kK(0)kF
1
0.8
0.6
0.4
0.2
0
?slow
A
0.01
0.01 0.1
1
?
10
S ? N (0, ? 2 )
10
1
0.1
0.01
0.01 0.1
1
10
?
Figure 3: How fast is the fastest sampler? (A) Scalar measure of the statistical dependency between any two samples collected k?m seconds apart (cf. main text), for Langevin sampling (black),
Gibbs sampling (blue, assuming a full update sweep is done every ?m ), a series of networks (brown
to red) with connectivities given by Eq. 9 where the elements of the skew-symmetric matrix S were
drawn iid. from N (0, ? 2 ) for different values of ? (see also panel B), the unconstrained optimized
network (yellow), and the optimized E/I network (green). For reference, the dashed gray line shows
the behavior of a network in which there are no recurrent interactions, and the posterior covariance
is encoded in the covariance of the input noise, which in fact corresponds to Langevin sampling
with inverse Hessian (?Newton?-like) preconditioning [16]. (B) Total slowing cost ?slow (S) when
Si<j ? N (0, ? 2 ), for increasing values of ?. The Langevin and the two optimized networks are
shown as horizontal lines for comparison. (C) Same as in (B), showing the root mean square (RMS)
value of the synaptic weights. Parameter values: N = 200, NI = 100, ?? = 1, ?m = 20 ms.
case. To see why this is a drastic restriction, let us observe that any connectivity matrix of the form
W(S) = I + ???2 I + S ??1
(9)
where S is an arbitrary skew-symmetric matrix (S> = ?S), solves Eq. 4, and therefore induces
the correct stationary distribution N (?, ?) under the linear stochastic dynamics of Eq. 3. Note that
Langevin sampling corresponds to S = 0 (cf. Eq. 5). In general, though, there are O(N 2 ) degrees of
freedom in the skew-symmetric matrix S, which could perhaps be exploited to increase the mixing
rate. In Sec. 5, we will show that indeed a large gain in sampling speed can be obtained through an
appropriate choice of S. For now, let us quantify slowness.
Let
matrix that contains all the posterior variances, and K(S, ? ) ?
? ? diag (?) be the diagonal
(r(t + ? ) ? ?)(r(t) ? ?)> t be the matrix of lagged covariances among neurons under the sta1
1
tionary distribution of the dynamics (so that ?? 2 K(S, ? )?? 2 is the autocorrelation matrix of the
network). Note that K(S, 0) = ? is the posterior covariance matrix, and that for fixed ?, ??2 and
?m , K(S, ? ) depends only on the lag ? and on the matrix of recurrent weights W, which itself
depends only on the skew-symmetric matrix S of free parameters. We then define a ?total slowing
cost?
Z ?
2
1
? 21
? 21
?slow (S) =
K(S,
?
)?
(10)
?
d?
2?m N 2 0
F
which penalizes the magnitude of the temporal (normalized) autocorrelations and pairwise cross2
correlations in the sequence of samples generated by the circuit dynamics. Here kMkF ?
P
2
trace(MM> ) = ij Mij
is the squared Frobenius norm of M.
Using the above measure of slowness, we revisit the mixing behavior of LS on a toy covariance
matrix ? drawn from the same inverse Wishart distribution mentioned above with parameters N =
200, ?02 = 2 and ?r = 0.2. We further regularize ? by adding the identity matrix to it, which
does not change anything in terms of the scaling law of Eq. 8 but ensures that the diagonal of WL
remains
bounded as N grows
large. We will use the same ? in the rest of the paper. Figure 3A
shows
??1/2 K(S, ? )??1/2
F as a function of the time lag ? : as predicted in Sec. 3, mixing is
indeed an order of magnitude slower for LS (S = 0, solid black line) than the single-neuron time
constant ?m (grey dashed line). Note that ?slow (Eq. 10, Fig. 3B) is proportional to the area under
the squared curve shown in Fig. 3A. Sample activity traces for this network, implementing LS, can
be found in Fig. 4B (top).
Using the same measure of slowness, we also inspected the speed of Gibbs sampling, another widely
used sampling technique (e.g. [17]) inspiring neural network dynamics for sampling from distributions over binary variables [18, 19, 20]. Gibbs sampling defines a Markov chain that operates in
5
discrete time, and also uses a symmetric weight matrix. In order to compare its mixing speed with
that of our continuous stochastic dynamics, we assume that a full update step (in which all neurons
have been updated once) takes time ?m . We estimated the integrand of the slowing cost (Eq. 10)
numerically using 30?000 samples generated by the Gibbs chain (Fig. 3A, blue). Gibbs sampling is
comparable to LS here: samples are still correlated on a timescale of order ? 50 ?m .
Finally, one may wonder how a random choice of S would perform in terms of decorrelation speed.
We drew random skew-symmetric S matrices from the Gaussian ensemble, Si<j ? N (0, ? 2 ), and
computed the slowing cost (Fig. 3, red). As the magnitude ? of S increases, sampling becomes
faster and faster until the dynamics is about as fast as the single-neuron time constant ?m . However,
the synaptic weights also grow with ? (Fig. 3C), and we show in Sec. 5 that an even faster sampler
exists that has comparatively weaker synapses. It is also interesting to note that the slope of ?slow at
? = 0 is zero, suggesting that LS is in fact maximally slow (we prove this formally in the SI).
5
What is the fastest sampler?
We now show that the skew-symmetric matrix S can be optimized for sampling speed, by directly
minimizing the slowing cost ?slow (S) (Eq. 10), subject to an L2 -norm penalty. We thus seek to
minimize:
?L2
2
L(S) ? ?slow (S) +
kW(S)kF .
(11)
2N 2
The key to performing this minimization is to use classical Ornstein-Uhlenbeck theory (e.g. [10]) to
bring our slowness cost under a form mathematically analogous to a different optimization problem
that has arisen recently in the field of robust control [21]. We can then use analytical results obtained
there concerning the gradient of ?slow , and obtain the overall gradient:
?L(S)
?S
=
?L
1 ?1
(? PQ)> ? (??1 PQ) + 22 S??2 + ??2 S
2
N
N
(12)
where matrices P and Q are obtained by solving two dual Lyapunov equations. All details can be
found in our SI.
We initialized S with random, weak and uncorrelated elements (cf. the end of Sec. 4, with ? = 0.01),
and ran the L-BFGS optimization algorithm using the gradient of Eq. 12 to minimize L(S) (with
?L2 = 0.1). The resulting, optimal sampler is an order of magnitude faster than either Langevin or
Gibbs sampling: samples are decorrelated on a timescale that is even faster than the single-neuron
time constant ?m (Fig. 3A, orange). We also found that fast solutions (with correlation length ? ?m )
can be found irrespective of the size N of the state space (not shown), meaning that the relative
speed-up between the optimal solution and LS grows with N (cf. Fig. 2).
The optimal Sopt induces a weight matrix Wopt given by Eq. 9 and shown in Fig. 4A (middle).
Notably, Wopt is no longer symmetric, and its elements are much larger than in the Langevin
symmetric solution WL with the same stationary covariance, albeit orders of magnitude smaller
than in random networks of comparable decorrelation speed (Fig. 3C).
It is illuminating to visualize activity trajectories in the plane defined by the topmost and bottommost
eigenvectors of ?, i.e. the first and last principal components (PCs) of the network activity (Fig. 4C).
The distribution of interest is broad along some dimensions, and narrow along others. In order to
sample efficiently, large steps ought to be taken along directions in which the distribution is broad,
and small steps along directions in which the distribution is narrow. This is exactly what our optimal
sampler does, whereas LS takes small steps along both broad and narrow directions (Fig. 4C).
6
Balanced E/I networks for fast sampling
We can further constrain our network to obey Dale?s law, i.e. the separation of neurons into separate
excitatory (E) and inhibitory (I) groups. The main difficulty in building such networks is that picking
an arbitrary skew-symmetric matrix S in Eq. 9 will not yield the column sign structure of an E/I
network in general. Therefore, we no longer have a parametric form for the solution matrix manifold
on which to find the fastest network. However, by extending the methods of Sec. 5, described in
6
optimized net.
neuron #
postsynaptic
-1
-1
presynaptic
neuron #
postsynaptic
-0.5
0
-4
ri (t)
-20
0
20
-3
0
3
-1
0
1
-20
0
20
-3
0
3
-1
0
1
8
4
20
0
-4
1
4
0
-4
500 ms
-8
100ms
40
0.5
4
0
-4
D
8
4
optimized E/I net.
0
ri (t)
20
1
0
4
0
-4
dist. of
increments
(1 ms steps)
-8
100ms
40
1
0
0
-4
1
-0.1
8
4
20
trajectories in
state space
(1 ms steps)
last PC
neuron #
postsynaptic
0
ri (t)
40
0.1
C
sample activity traces
Langevin
last PC
B
weight matrices
last PC
A
-8
100ms
-20 0 20
first PC
-3
0
3 -1 0 1
step along
E/I corr.
{first|last} PC
Figure 4: Fast sampling with optimized networks. (A) Synaptic weight matrices for the Langevin
network (top), the fastest sampler (middle) and the fastest sampler that obeys Dale?s law (bottom).
Note that the synaptic weights in both optimized networks are an order of magnitude larger than in
the symmetric Langevin solution. The first two networks are of size N = 200, while the optimized
E/I network has size N + NI = 300. (B) 500 ms of spontaneous network activity (h = 0) in each of
the three networks, for all of which the stationary distribution of r (restricted here to the first 40 neurons) is the same multivariate Gaussian. (C) Left: activity trajectories (the same 500 ms as shown
in (B)) in the plane defined by the topmost and bottommost eigenvectors of the posterior covariance matrix ? (corresponding to the first and last principal components of the activity fluctuations
r(t)). For the E/I network, the projection is restricted to the excitatory neurons. Right: distribution of increments along both axes, measured in 1 ms time steps. Langevin sampling takes steps of
comparable size along all directions, while the optimized networks take much larger steps along the
directions of large variance prescribed by the posterior. (D) Distributions of correlations between
the time courses of total excitatory and inhibitory input in individual neurons.
detail in our SI, we can still formulate the problem as one of unconstrained optimization, and obtain
the fastest, balanced E/I sampler.
We consider the posterior to be encoded in the activity of the N = 200 excitatory neurons, and add
NI = 100 inhibitory neurons which we regard as auxiliary variables, in the spirit of Hamiltonian
Monte Carlo methods [11]. Consequently, the E-I and I-I covariances are free parameters, while
the E-E covariance is given by the target posterior. For additional biological realism, we also forbid
self-connections as they can be interpreted as a modification of the intrinsic membrane time constant
of the single neurons, which in principle cannot be arbitrarily learned.
The speed optimization yields the connectivity matrix shown in Fig. 4A (bottom). Results for this
network are presented in a similar format as before, in the same figures. Sampling is almost as fast
as in the best (regularized) unconstrained network (compare yellow and green in Fig. 3), indicating
that Dale?s law ? unlike the symmetry constraint implicitly present in Langevin sampling ? is not
fundamentally detrimental to mixing speed. Moreover, the network operates in a regime of excitation/inhibition balance, whereby the total E and I input time courses are correlated in single cells
(Fig. 4D, bottom). This is true also in the unconstrained optimal sampler. In contrast, E and I inputs
are strongly anti-correlated in LS.
7
7
Discussion
We have studied sampling for Bayesian inference in neural circuits, and observed that a linear
stochastic network is able to sample from the posterior under a linear Gaussian latent variable model.
Hidden variables are directly encoded in the activity of single neurons, and their joint activity undergoes moment-to-moment fluctuations that visit each portion of the state space at a frequency
given by the target posterior density. To achieve this, external noise sources fed into the network are
amplified by the recurrent circuitry, but preferentially amplified along the state-space directions of
large posterior variance. Although, for the very simple linear Gaussian model we considered here,
a purely feed-forward architecture would also trivially be able to provide independent samples (ie.
provide samples that are decorrelated at the time scale of ?m ), the network required to achieve this
is deeply biologically implausible (see SI).
We have shown that the choice of a symmetric weight matrix ? equivalent to LS, a popular machine learning technique [2, 11, 12] that has been suggested to underlie neuronal network dynamics
sampling continuous variables [6, 13] ? is most unfortunate. We presented an analytical argument
predicting dramatic slowing in high-dimensional latent spaces, supported by numerical simulations.
Even in moderately large networks, samples were correlated on timescales much longer than the
single-neuron decay time constant.
We have also shown that when the above symmetry constraint is relaxed, a family of other solutions
opens up that can potentially lead to much faster sampling. We chose to explore this possibility
from a normative viewpoint, optimizing the network connectivity directly for sampling speed. The
fastest sampler turned out to be highly asymmetric and typically an order of magnitude faster than
Langevin sampling. Notably, we also found that constraining each neuron to be either excitatory
or inhibitory does not impair performance while giving a far more biologically plausible sampler.
Dale?s law could even provide a natural safeguard against reaching slow symmetric solutions such
as Langevin sampling, which we saw was the worst-case scenario (cf. also SI).
It is worth noting that Wopt is strongly nonnormal.3 Deviation from normality has important consequences for the dynamics of our networks: it makes the network sensitive to perturbations along
some directions in state space. Such perturbations are rapidly amplified into large, transient excursions along other, relevant directions. This phenomenon has been shown to explain some key
features of spontaneous activity in primary visual cortex [9] and primary motor cortex [22].
Several aspects would need to be addressed before our proposal can crystalize into a more thorough
understanding of the neural implementation of the sampling hypothesis. First, can local synaptic
plasticity rules perform the optimization that we have approached from an algorithmic viewpoint?
Second, what is the origin of the noise that we have hypothesized to come from external sources?
Third, what kind of nonlinearity must be added in order to allow sampling from non-Gaussian distributions, whose shapes may have non-trivial dependencies on the observations? Also, does the main
insight reached here ? namely that fast samplers are to be found among nonsymmetric, nonnormal
networks ? carry over to the nonlinear case? As a proof of principle, in preliminary simulations, we
have shown that speed optimization in a linearized version of a nonlinear network (with a tanh gain
function) does yield fast sampling in the nonlinear regime, even when fluctuations are strong enough
to trigger the nonlinearity and make the resulting sampled distribution non-Gaussian (details in SI).
Finally, we have also shown (see SI) that the Langevin solution is the only network that satisfies the
detailed balance condition [23] in our model class; reversibility is violated in all other stochastic networks we have presented here (random, optimal, optimal E/I). The fact that these networks are faster
samplers is in line with recent machine learning studies on how non-reversible Markov chains can
mix faster than their reversible counterparts [24]. The construction of such Monte-Carlo algorithms
has proven challenging [25, 26, 27], suggesting that the brain ? if it does indeed use sampling-based
representations ? might have something yet to teach us about machine learning.
Acknowledgements This work was supported by the Wellcome Trust (GH, ML), the Swiss National Science Foundation (GH) and the Gatsby Charitable Foundation (LA). Our code will be made
freely available from GH?s personal webpage.
Indeed, the sum of the squared moduli of its eigenvalues accounts for only 25% of kWopt k2F [7]. For a
P
normal matrix W (such as the Langevin solution, WL ), i |?i |2 = kWk2F , i.e. this ratio is 100%.
3
8
References
[1] S. Thorpe, D. Fize, and C. Marlot. Speed of processing in the human visual system. Nature, 381:520?
522, 1996.
[2] D. MacKay. Information theory, inference, and learning algorithms. Cambridge University Press, 2003.
[3] D. Knill and A. Pouget. The Bayesian brain: the role of uncertainty in neural coding and computation.
Trends in Neurosciences, 27:712?719, 2004.
[4] J. Fiser, P. Berkes, G. Orb?an, and M. Lengyel. Statistically optimal perception and learning: from behavior
to neural representations. Trends in Cognitive Sciences, 14:119?130, 2010.
[5] P. Berkes, G. Orb?an, M. Lengyel, and J. Fiser. Spontaneous cortical activity reveals hallmarks of an
optimal internal model of the environment. Science, 331:83?87, 2011.
[6] R. Moreno-Bote, D. C. Knill, and A. Pouget. Bayesian sampling in visual perception. Proceedings of the
National Academy of Sciences, 108:12491?12496, 2011.
[7] L. N. Trefethen and M. Embree. Spectra and pseudospectra: the behavior of nonnormal matrices and
operators. Princeton University Press, 2005.
[8] G. Hennequin, T. P. Vogels, and W. Gerstner. Non-normal amplification in random balanced neuronal
networks. Physical Review E, 86:011909, 2012.
[9] B. K. Murphy and K. D. Miller. Balanced amplification: A new mechanism of selective amplification of
neural activity patterns. Neuron, 61:635?648, 2009.
[10] C. W. Gardiner. Handbook of stochastic methods: for physics, chemistry, and the natural sciences. Berlin:
Springer, 1985.
[11] R. Neal. MCMC using Hamiltonian dynamics. Handbook of MCMC, pages 113?162, 2011.
[12] M. Welling and Y. W. Teh. Bayesian learning via stochastic gradient Langevin dynamics. In Proceedings
of the International Conference on Machine Learning, 2011.
[13] A. Grabska-Barwinska, J. Beck, A. Pouget, and P. Latham. Demixing odors - fast inference in olfaction.
In C.J.C. Burges, L. Bottou, M. Welling, Z. Ghahramani, and K.Q. Weinberger, editors, Advances in
Neural Information Processing Systems 26, pages 1968?1976. Curran Associates, Inc., 2013.
[14] Mark Rudelson and Roman Vershynin. Smallest singular value of a random rectangular matrix. Communications on Pure and Applied Mathematics, 62:1707?1739, 2009.
[15] M. Girolami and B. Calderhead. Riemann manifold Langevin and Hamiltonian Monte Carlo methods.
Journal of the Royal Statistical Society: Series B (Statistical Methodology), 73:123?214, 2011.
[16] J. Martin, L. C. Wilcox, C. Burstedde, and O. Ghattas. A stochastic Newton MCMC method for largescale statistical inverse problems with application to seismic inversion. SIAM Journal on Scientific Computing, 34:A1460?A1487, 2012.
[17] M. Mezard and A. Montanari. Information, physics, and computation. Oxford University Press, 2009.
[18] G. E. Hinton and T. J. Sejnowski. Learning and relearning in Boltzmann machines. In D E Rumelhart, J L
McClelland, and the PDP Research Group, editors, Parallel Distributed Processing: Explorations in the
Microstructure of Cognition, volume 1: Foundations, chapter 7, pages 282?317. MIT Press, Cambridge,
MA, 1986.
[19] G E Hinton, P Dayan, B J Frey, and R M Neal. The ?wake-sleep? algorithm for unsupervised neural
networks. Science, 268(5214):1158?1161, 1995.
[20] L. Buesing, J. Bill, B. Nessler, and W. Maass. Neural dynamics as sampling: a model for stochastic
computation in recurrent networks of spiking neurons. PLoS Computational Biology, 7:1?22, 2011.
[21] J. Vanbiervliet, B. Vandereycken, W. Michiels, S. Vandewalle, and M. Diehl. The smoothed spectral
abscissa for robust stability optimization. SIAM Journal on Optimization, 20:156?171, 2009.
[22] G. Hennequin, T. P. Vogels, and W. Gerstner. Optimal control of transient dynamics in balanced networks
supports generation of complex movements. Neuron, 82, 2014.
[23] W. Keith Hastings. Monte Carlo sampling methods using Markov chains and their applications.
Biometrika, 57:97?109, 1970.
[24] Akihisa Ichiki and Masayuki Ohzeki. Violation of detailed balance accelerates relaxation. Physical
Review E, 88:020101, 2013.
[25] Y. Sun, J. Schmidhuber, and F. J. Gomez. Improving the asymptotic performance of Markov Chain
Monte-Carlo by inserting vortices. In J.D. Lafferty, C.K.I. Williams, J. Shawe-Taylor, R.S. Zemel, and
A. Culotta, editors, Advances in Neural Information Processing Systems 23, pages 2235?2243. 2010.
[26] K. S. Turitsyn, M. Chertkov, and M. Vucelja. Irreversible Monte Carlo algorithms for efficient sampling.
Physica D: Nonlinear Phenomena, 240:410?414, 2011.
[27] Joris Bierkens. Non-reversible Metropolis-Hastings. arXiv:1401.8087 [math], 2014.
9
| 5265 |@word private:1 middle:3 version:1 determinant:1 laurence:2 seems:1 norm:2 inversion:1 open:1 grey:1 simulation:4 seek:2 linearized:1 eng:1 covariance:15 dramatic:1 solid:3 carry:1 moment:3 series:2 contains:1 si:12 yet:2 must:7 additive:1 numerical:2 plasticity:1 shape:1 motor:1 moreno:1 update:3 stationary:7 generative:4 leaf:1 unacceptably:1 slowing:9 plane:2 hamiltonian:3 short:1 realism:1 math:1 along:12 constructed:1 become:1 prove:1 behavioral:1 autocorrelation:4 introduce:1 pairwise:5 notably:2 expected:1 indeed:5 behavior:6 abscissa:1 dist:1 brain:9 riemann:1 increasing:4 becomes:7 confused:1 discover:1 notation:1 moreover:2 circuit:9 nonnormal:5 panel:1 bounded:1 what:6 grabska:1 kind:2 interpreted:1 eigenvector:1 finding:1 ought:1 temporal:1 thorough:1 every:2 tackle:1 exactly:2 biometrika:1 rm:2 scaled:1 uk:5 control:4 unit:4 underlie:1 enjoy:1 producing:2 before:2 engineering:1 local:1 frey:1 limit:2 consequence:1 irreversible:1 vortex:1 oxford:1 fluctuation:5 firing:1 black:2 plus:1 chose:1 might:1 studied:1 dynamically:1 suggests:1 conversely:1 challenging:1 fastest:8 statistically:1 obeys:1 practical:1 unique:1 swiss:1 area:1 deed:1 projection:1 wait:1 suggest:1 amplify:1 close:1 cannot:1 operator:1 context:2 nessler:1 restriction:1 equivalent:1 bill:1 straightforward:1 williams:1 l:23 rectangular:2 formulate:1 simplicity:1 pure:2 pouget:3 rule:1 insight:1 importantly:1 hennequin:2 regularize:1 population:1 stability:2 notion:1 increment:2 analogous:1 updated:1 spontaneous:3 inspected:1 target:2 trigger:1 construction:1 us:2 curran:1 hypothesis:1 origin:1 associate:1 element:5 trend:2 recognition:1 rumelhart:1 std:1 asymmetric:1 bottom:3 observed:1 role:1 worst:1 culotta:1 ensures:1 connected:1 sun:1 plo:1 trade:1 movement:1 ran:1 balanced:6 principled:1 environment:2 mentioned:1 complexity:1 topmost:2 deeply:1 moderately:1 cam:2 dynamic:24 personal:1 solving:2 bottommost:2 purely:1 calderhead:1 preconditioning:2 kmkf:1 joint:1 chapter:1 distinct:1 fast:14 london:1 monte:9 sejnowski:1 approached:1 tell:1 zemel:1 whose:1 lag:3 supplementary:1 plausible:3 encoded:3 widely:1 larger:3 trefethen:1 timescale:4 jointly:1 noisy:2 itself:2 delivered:1 sequence:1 eigenvalue:5 analytical:3 net:3 ucl:1 interaction:5 inserting:1 relevant:2 turned:1 realization:1 rapidly:1 mixing:8 iff:1 achieve:2 amplified:3 academy:1 amplification:5 frobenius:1 recipe:2 webpage:1 extending:1 produce:2 derive:1 recurrent:11 ac:3 measured:1 ij:1 noticeable:1 keith:1 eq:20 solves:1 strong:2 implemented:2 predicted:1 auxiliary:1 indicate:1 come:1 quantify:1 lyapunov:2 orb:2 girolami:1 direction:8 correct:2 owing:1 stochastic:13 centered:1 human:2 exploration:1 transient:3 implementing:3 require:2 microstructure:1 preliminary:1 biological:2 mathematically:2 physica:1 mm:2 considered:1 normal:2 algorithmic:1 cognition:1 visualize:1 circuitry:1 smallest:2 fh:2 tanh:1 saw:1 sensitive:1 largest:3 wl:9 weighted:1 minimization:1 mit:1 gaussian:17 reaching:1 ax:1 focus:1 likelihood:1 slowest:5 greatly:1 contrast:2 baseline:1 inference:10 dayan:1 typically:1 hidden:1 quasi:1 selective:2 interested:2 issue:1 among:2 dual:1 overall:1 art:1 integration:1 orange:1 mackay:1 field:1 construct:1 once:1 intriguingly:1 reversibility:1 sampling:60 biology:1 kw:1 broad:3 unsupervised:1 k2f:1 others:1 fundamentally:2 roman:1 inherent:1 thorpe:1 few:1 simultaneously:1 national:2 individual:1 murphy:1 beck:1 olfaction:1 freedom:2 interest:2 acceptance:1 possibility:1 highly:1 marlot:1 evaluation:1 vandereycken:1 violation:1 pc:6 chain:8 accurate:1 capable:1 taylor:1 walk:1 desired:1 re:2 penalizes:1 initialized:1 masayuki:1 column:1 ar:2 cost:6 deviation:3 wonder:1 vandewalle:1 optimally:1 dependency:2 vershynin:1 density:2 fundamental:2 international:1 forbid:1 ie:2 siam:2 probabilistic:4 systematic:2 physic:2 picking:1 safeguard:1 connectivity:6 squared:3 choose:3 wishart:6 dr:2 worse:1 cognitive:2 external:6 leading:1 toy:1 suggesting:2 potential:1 account:1 bfgs:1 chemistry:1 sec:6 coding:1 coefficient:1 inc:1 depends:5 ornstein:1 later:1 root:1 observer:1 lab:1 red:2 portion:1 reached:1 option:1 parallel:1 slope:1 contribution:1 minimize:2 square:2 ni:3 accuracy:1 wiener:2 variance:10 efficiently:1 ensemble:1 yield:4 correspond:1 miller:1 yellow:2 weak:1 bayesian:4 buesing:1 critically:2 iid:1 carlo:9 trajectory:4 worth:1 lengyel:3 explain:1 implausible:3 synapsis:1 decorrelated:2 synaptic:8 eigenmode:3 definition:1 against:1 frequency:1 naturally:1 associated:1 proof:1 gain:2 sampled:1 popular:1 feed:1 dt:2 methodology:1 maximally:1 evaluated:1 done:1 though:1 generality:1 strongly:2 just:2 fiser:2 correlation:6 until:1 receives:1 hastings:3 horizontal:1 trust:1 nonlinear:4 reversible:3 autocorrelations:1 defines:1 undergoes:1 quality:1 indicated:1 perhaps:2 gray:1 grows:3 vogels:2 scientific:1 building:1 modulus:1 hypothesized:1 contain:1 brown:1 normalized:1 true:1 counterpart:1 analytically:1 hence:2 spatially:1 read:1 symmetric:18 maass:1 neal:2 self:1 uniquely:1 maintained:1 ambiguous:1 excitation:2 whereby:2 anything:1 m:12 bote:1 latham:1 performs:2 bring:1 gh:3 meaning:1 hallmark:1 recently:1 common:1 sped:1 spiking:1 empirically:1 sopt:1 physical:2 volume:1 nonsymmetric:1 marginals:1 numerically:1 cambridge:3 lengyel1:1 gibbs:8 unconstrained:4 trivially:1 mathematics:1 nonlinearity:2 shawe:1 dot:1 pq:2 cortex:3 longer:3 inhibition:2 add:1 berkes:2 something:1 posterior:28 multivariate:2 recent:3 asssumption:1 perspective:1 optimizing:1 apart:1 scenario:1 schmidhuber:1 slowness:4 binary:1 arbitrarily:2 devise:1 exploited:1 additional:1 relaxed:1 freely:1 determine:1 dashed:4 multiple:2 violate:1 neurally:1 reduces:1 full:2 mix:1 barwinska:1 faster:11 match:1 calculation:1 long:1 concerning:1 visit:1 a1:1 schematic:2 involving:1 arxiv:1 represent:2 uhlenbeck:1 arisen:1 achieved:1 cell:1 proposal:2 whereas:2 interval:1 addressed:1 wake:1 singular:2 source:4 grow:1 operate:1 rest:1 unlike:1 ascent:1 hz:1 subject:1 contrary:1 lafferty:1 spirit:1 unitary:1 noting:1 feedforward:2 constraining:1 enough:3 insofar:1 affect:1 fit:1 architecture:2 bottleneck:3 rms:2 akin:1 penalty:1 hessian:2 detailed:4 eigenvectors:2 ten:1 locally:1 induces:2 inspiring:1 mcclelland:1 inhibitory:5 revisit:1 sign:1 neuroscience:2 estimated:1 wr:1 blue:2 discrete:1 group:2 key:3 drawn:5 fize:1 relaxation:1 sum:2 inverse:10 uncertainty:1 almost:1 family:1 excursion:1 separation:1 draw:1 decision:2 scaling:2 comparable:3 accelerates:1 layer:1 bound:3 wopt:3 gomez:1 sleep:1 activity:18 gardiner:1 constraint:2 constrain:2 ri:5 encodes:1 dominated:1 generates:3 integrand:1 speed:20 argument:1 prescribed:2 aspect:1 performing:1 format:1 martin:1 according:1 membrane:3 across:1 smaller:1 increasingly:1 postsynaptic:3 metropolis:2 making:1 biologically:3 happens:1 modification:1 restricted:2 taken:1 wellcome:1 behavioural:1 equation:2 remains:2 turn:2 skew:7 mechanism:1 fed:1 drastic:1 cor:1 end:1 available:1 stiffness:1 observe:1 obey:1 appropriate:5 generic:2 spectral:1 odor:1 weinberger:1 slower:2 existence:1 denotes:1 top:2 ensure:1 cf:7 unfortunate:1 rudelson:1 newton:3 pushing:1 joris:1 giving:1 ghahramani:1 classical:1 comparatively:1 society:1 implied:1 sweep:1 added:1 quantity:1 parametric:1 primary:2 dependence:1 diagonal:5 unclear:1 gradient:5 detrimental:1 separate:2 berlin:1 irreversibility:2 manifold:2 presynaptic:1 collected:2 trivial:2 barely:1 assuming:1 length:2 code:1 kk:2 ratio:3 balance:5 minimizing:1 preferentially:1 postponing:1 potentially:2 teach:1 trace:3 negative:3 lagged:1 implementation:3 boltzmann:1 perform:3 teh:1 seismic:1 inspect:1 neuron:25 observation:4 markov:7 anti:1 langevin:25 unconnected:1 communication:1 precise:1 hinton:2 pdp:1 rn:1 perturbation:2 smoothed:1 arbitrary:3 intensity:1 namely:1 required:3 connection:3 optimized:10 errorbars:1 learned:1 unduly:1 narrow:3 able:2 suggested:4 impair:1 dynamical:3 perception:5 below:1 pattern:1 regime:3 including:1 max:9 green:2 royal:1 unrealistic:2 critical:1 decorrelation:2 difficulty:1 quantification:1 regularized:1 predicting:1 natural:2 largescale:1 normality:1 inversely:1 picture:1 irrespective:1 text:2 review:2 understanding:1 l2:3 acknowledgement:1 kf:3 relative:3 law:6 asymptotic:2 loss:1 generation:2 limitation:2 proportional:2 interesting:1 analogy:1 proven:1 h2:4 integrate:1 illuminating:1 degree:2 foundation:3 consistent:1 imposes:1 principle:3 viewpoint:2 charitable:1 editor:3 uncorrelated:3 excitatory:6 course:2 supported:2 last:6 free:2 appreciated:1 weaker:1 allow:1 burges:1 distributed:1 regard:1 overcome:2 dimension:3 curve:1 cortical:1 sensory:2 dale:6 made:3 forward:1 far:3 welling:2 implicitly:2 annihilate:1 ml:1 reveals:1 handbook:2 spectrum:1 continuous:4 latent:8 why:1 nature:1 robust:3 diehl:1 symmetry:3 improving:1 gerstner:2 complex:2 bottou:1 domain:2 diag:1 main:4 timescales:1 montanari:1 kwk2f:1 noise:10 knill:2 allowed:1 neuronal:7 fig:21 gatsby:3 slow:18 mezard:1 momentary:1 explicit:1 vanish:1 third:1 chertkov:1 removing:1 showing:1 normative:1 recurrently:1 r2:3 decay:3 evidence:3 demixing:1 intrinsic:2 exists:1 albeit:1 sequential:1 effectively:1 corr:2 adding:1 drew:1 magnitude:8 tionary:1 relearning:1 easier:1 simply:1 explore:2 infinitely:1 neurophysiological:1 visual:3 scalar:2 springer:1 mij:1 corresponds:4 environmental:1 satisfies:1 ma:1 identity:2 goal:1 consequently:1 absence:1 change:1 specifically:2 operates:2 sampler:15 principal:2 total:4 la:1 indicating:1 formally:2 guillaume:1 college:1 puzzling:1 support:2 internal:1 mark:1 violated:2 phenomenon:2 dept:1 mcmc:4 princeton:1 wilcox:1 correlated:5 |
4,711 | 5,266 | Information-based learning by agents in unbounded
state spaces
Shariq A. Mobin, James A. Arnemann, Friedrich T. Sommer
Redwood Center for Theoretical Neuroscience
University of California, Berkeley
Berkeley, CA 94720
[email protected], [email protected], [email protected]
Abstract
The idea that animals might use information-driven planning to explore an unknown environment and build an internal model of it has been proposed for quite
some time. Recent work has demonstrated that agents using this principle can efficiently learn models of probabilistic environments with discrete, bounded state
spaces. However, animals and robots are commonly confronted with unbounded
environments. To address this more challenging situation, we study informationbased learning strategies of agents in unbounded state spaces using non-parametric
Bayesian models. Specifically, we demonstrate that the Chinese Restaurant Process (CRP) model is able to solve this problem and that an Empirical Bayes version is able to efficiently explore bounded and unbounded worlds by relying on
little prior information.
1
Introduction
Learning in animals involves the active gathering of sensor data, presumably selecting those sensor
inputs that are most useful for learning a model of the world. Thus, a theoretical framework for
the learning in agents, where learning itself is the primary objective, would be essential for making
testable predictions for neuroscience and psychology [9, 7], and it would also impact applications
such as optimal experimental design and building autonomous robots [3].
It has been proposed that information theory-based objective functions, such as those based on the
comparison of learned probability distributions, could guide exploratory behavior in animals and artificial agents [13, 18]. Although reinforcement learning theory has largely advanced in describing
action planning in fully or partially observable worlds with a fixed reward function, e.g., [17], the
study of planning with internally defined and gradually decreasing reward functions has been rather
slow. A few recent studies [20, 11, 12] developed remarkably efficient action policies for learning
an internal model of an unknown fully observable world that are driven by maximizing an objective of predicted information gain. Although using somewhat different definitions of information
gain, the key insights of these studies are that optimization has to be non-greedy, with a longer time
horizon, and that gain in information also translates to efficient reward gathering. However, these
models are still quite limited and cannot be applied to agents in more realistic environments. They
only work in observable, discrete and bounded state spaces. Here, we relax one of these restrictions and present a model for unbounded, observable discrete state spaces. Using methods from
non-parametric Bayesian statistics, specifically the Chinese Restaurant Process (CRP), the resulting
agent can efficiently learn the structure of an unknown, unbounded state space. To our knowledge
this is the first use of CRPs to address this problem, however, CRPs have been introduced earlier to
reinforcement learning for other purposes, such as state clustering [2].
1
2
2.1
Model
Mathematical framework for embodied active learning
In this study we follow [12] and use Controlled Markov Chains (CMC) to describe how an agent
can interact with its environment in closed, embodied, action-perception loops. A CMC is a Markov
Chain with an additional control variable to allow for switching between different transition distributions in each state, e.g. [6]. Put differently, it is a Markov Decision Process (MDP) without the
reward function. A CMC is described by a 3-tuple (S , A , ?) where S denotes a finite set of states,
A is a finite set of actions the agent can take, and ? is a 3-dimensional CMC kernel describing the
transition probabilities between states for each action
?sas0 = ps0 |s,a = P (st+1 = s0 |st = s, at = a)
(1)
Like in [12] we consider the exploration task of the agent to be the formation of an accurate estimate,
b of the true CMC kernel, ?, that describes its world.
or internal model ?,
2.2
Modeling the transition in unbounded state spaces
Let t be the current number of observations of states S and Kt be the number of different states
discovered so far. The observed counts are denoted by Ct := {#1 , ..., #Kt }.
Species sampling models have been proposed as generalizations of the Dirichlet process [14], which
are interesting for non-parametric Bayesian inference in unbounded state spaces. A species sampling
sequence (SSS) describes the distribution of the next observation St+1 . It is defined by
St+1 |S1 , , St ?
Kt
X
pi (Ct )?S? + pKt+1 (Ct )
(2)
i=1
with ?S? a degenerate probability measure, see [10] for details. In order to define a valid SSS,
the sequence (p1 , p2 , ...) must sum to one and be an Exchangeable Partition Probability Function
(EPPF). The exchangeability condition requires that the probabilities depend only on the counts Ct ,
not on the order of how the agent sampled the transitions.
Here we consider one of most common EPPF models in the literature, the Chinese Restaurant Process (CRP) or Polya urn process [1]. According to the CRP model, the probability of observing a
state is
#i
pi (Ct ) =
for i = 1, ..., Kt
(3)
t+?
?
(4)
p? (Ct ) ? pKt+1 (Ct ) =
t+?
where (3) describes revisiting a state and (4) describes the undiscovered probability mass (UPM),
i.e., the probability of discovering a new state, which is then labeled Kt+1 . In the following, the set
of undiscovered states will be denoted by ?. Using this formalism, the agent must define a separate
CRP for each state action pair s, a. The internal model is then described by
b sas0 = ps0 |s,a (Ct ),
?
(5)
b sas0 is suppressed for the sake of notational ease.
updated according to (3, 4). The t index in ?
Our simplest agent uses a CRP (3, 4) with fixed ?. Further, we will investigate an Empirical Bayes
CRP, referred to as EB-CRP, in which the parameter ? is learned and adjusted from observations
online using a maximum likelihood estimate (MLE). This is similar to the approach of [22] but we
follow a more straightforward path and derive a MLE of ? using the EPPF of the CRP and employing
an approximation of the harmonic series.
The likelihood of observing a given number of state counts is described by the EPPF of the CRP [8]
Kt
Y
?Kt
?(Ct ; ?) = Qt?1
(#i ? 1)!
i=0 (? + i) i=1
2
(6)
Maximizing the log likelihood
t?1
d
Kt X 1
ln(?(Ct ; ?)) =
?
=0
d?
?
?+i
i=0
(7)
yields
Kt
(8)
1 ,
1
? 12t
ln(t) + ? + 2t
2
where (8) uses a closed form approximation of the harmonic series in (7) with Euler?s Mascheroni
constant ?. In our EB-CRP agent, the parameter ? is updated after each observation according to
(8).
?(t) ?
2.3
Information-theoretic assessment of learning
Assessing or guiding the progress of the agent in the exploration process can be done by comparing
probability distributions. For example, the learning progress should increase the similarity between
b of the agent and the true model, ?. A popular measure for comparing disthe internal model, ?,
tributions of the same dimensions is the KL Divergence, DKL . However, in our case, with the size
b models of
of the underlying state space unknown and states being discovered successively in ?,
different sizes have to be compared.
To address this, we apply the following padding procedure to the smaller model with fewer discovb has n undiscovered state transitions
ered states and transitions (Figure 1). If the smaller model, ?,
from a known origin state, one splits the UPM uniformly into n equal probabilities (Figure 1a). The
resulting padded model is given by
?
b sa?
?
b sas0 = 0
?
[Figure 1a]
? (|S?sa |?|S?b sa |) , ?
P
b
?sas0 = 1/|S? |,
(9)
s?
/ S?
[Figure 1b]
b
sa
?
?b
b
?sas0 ,
?sas0 > 0
where |S?sa | is the number of known states reachable from state s by taking action a in ?. Further,
b one adds such states and a uniform transition kernel to
if there are undiscovered origin states in ?,
potential target states (Figure 1b).
Figure 1: Illustration of the padding procedure for adding unknown states and state transitions
b of an unbounded environment in order to compare it with a
in a smaller, less informed model, ?,
larger, better informed model, ?. (a) If transitions to target states are missing, we uniformly split the
UPM into equal transition probabilities to the missing target states, which are in fact the unknown
elements of the set ?. (b) If a state is not discovered yet, we paste this state in with a uniform
transition distribution to all target states reachable in the larger model, ?.
With this type of padding procedure we can define a distance between two unequally sized models,
!
X
0
?
sas
P
b sa? ) := DKL (?sa? ||?
b sa? ) :=
?sas0 log2
DKLP (?sa? ||?
,
(10)
bP 0
?
sas
s0 ?S?
sa
and use it to extend previous information measures for assessing and guiding explorative learning
[12] to unbounded state spaces. First, we define Missing Information,
X
b :=
b sa? ),
IM (?||?)
DKLP (?sa? ||?
(11)
s?S ,a?A
3
a quantity an external observer can use for assessing the deficiency of the internal model of the agent
with respect to the true model. Second, we define Information Gain,
0
b ? IM (?||?
b s,a?s ),
IG (s, a, s0 ) := IM (?||?)
(12)
a quantity measuring the improvement between two models, in this case, between the current internal
b and an improved one, ?
b s,a?s0 , which represents an updated model after
model of the agent, ?,
0
observing a new state transition from s to s under action a.
2.4
Predicted information gain
Predicted information gain (PIG) as used in [12] is the expected information gain for a given state
action pair. To extend the previous formula in [12] to compute this expectation in the non-parametric
setting, we again make use of the padding procedure described in the last section
P IG(s, a)
:= Es0 ,?|Ct [IG (s, a, s0 )]
=
b sa? DKLP (?
b s,a??
b sa? ) +
?
||?
sa?
s,a?s0 b
b sas0 DKL (?
b sa?
?
||?sa? )
X
(13)
s0 ?S?
b sa
Here, DKLP handles the case where the agent, during its planning, hypothetically discovers a new
target state, ? ? ?, from the state action pair, s, a. There is one small difference in calculating the
s,a?? . Thus
DKLP from the previous section, which is that in equation (9) S?sa is replaced by S?
b sa
the RHS of (13) can be computed internally by the agent for action planning as it does not contain
the true model, ?.
2.5
Value Iteration
When states of low information gain separate the agent from states of high information gain in the
environment, greedy maximization of PIG performs poorly. Thus, like in [12], we employ value
iteration using the Bellman equations [4]. We begin at a distant time point (? = 0) assigning initial
values to PIG. Then, we propogate backward in time calculating the expected reward.
Q0 (s, a)
Q? ?1 (s, a)
:= P IG(s, a)
(14)
h
b sa? V? (?) +
:= P IG(s, a) + ? ?
X
i
b sas0 V? (s0 )
?
(15)
s0 ?S?
b sa
V? (s)
:= max Q? (s, a)
(16)
a
With the discount factor, ?, set to 0.95, one can define how actions are chosen by all our PIG agents
aP IG := argmax Q?10 (s, a)
(17)
a
3
Experimental Results
Here we describe simulation experiments with our two models, CRP-PIG and EB-CRP-PIG, and
compare them with published approaches. The models are tested in environments defined in the
literature and also in an unbounded world.
First the agents were tested in a bounded maze environment taken from [12] (Figure 2). The state
space in the maze consists of the |S | = 36 rooms. There are |A | = 4 actions that correspond to noisy
translations in the four cardinal directions, drawn from a Dirichlet distribution. To make the task of
learning harder, 30 transporters are distributed amongst the walls which lead to an absorbing state
(state 29 marked by concentric rings in Figure 2). Absorbing states, such as at the bottom of gravity
wells, are common in real world environments and pose serious challenges for many exploration
algorithms [12].
We compare the learning strategies proposed here, CRP-PIG and EB-CRP-PIG, with the following
strategies:
4
Random action: A negative control, representing the minimally directed action policy that
any directed action policy should beat.
Least Taken Action (LTA): A well known explorative strategy that simply takes the action it
has taken least often in the current state [16].
Counter-Based Exploration (CB): Another explorative strategy from the literature that attempts to induce a uniform sampling across states [21].
DP-PIG: The strategy of [12] which applies the same objective function as described here,
but is given the size of the state space and is therefore at an advantage. This agent uses a
Dirichlet process (DP) with ? set to 0.20, which was found empirically to be optimal for the
maze environment.
Unembodied: An agent which can choose any action from any state at each time step (hence
unembodied) and can therefore attain the highest PIG possible at every sampling step. This
strategy represents a positive control.
Figure 2: Bounded Maze environment. Two transition
distributions, ?sa? , are depicted, one for (s=13, a=?left?)
and one for (s=9, a=?up?). Dark versus light gray arrows represent high versus low probabilities. For (s=13,
a=?left?), the agent moves with highest probability left
into a transporter (blue line), leading it to the absorbing
state 29 (blue concentric rings). With smaller probabilities the agent moves up, down or is reflected back to its
current state by the wall to the right. The second transition distribution is displayed similarly.
Figure 3 depicts the missing information (11) in the bounded maze for the various learning strategies over 3000 sampling steps averaged over 200 runs. All PIG-based embodied strategies exhibit
a faster decrease of missing information with sampling, however, still significantly slower than the
unembodied control. In this finite environment the DP-PIG agent with the correct Dirichlet prior
(experimentally optimized ?-parameter) has an advantage over the CRP based agents and reduced
the missing information more quickly. However, the new strategies for unbounded state space still
outperform the competitor agents from the literature by far. Interestingly, EB-CRP-PIG with continuously adjusted ? can reduce missing information significantly faster than CRP-PIG with fixed,
experimentally optimized ? = 0.25.
Figure 3: Missing Information vs. Time for EB-CRP-PIG and several other strategies in the bounded
maze environment.
To directly assess how efficient learning translates to the ability to harvest reward, we consider the 5state ?Chain? problem [19], shown in Figure 4, a popular benchmark problem. In this environment,
agents have two actions available, a and b, which cause transitions between the five states. At each
time step the agent ?slips? and performs the opposite action with probability pslip = 0.2. The agent
receives a reward of 2 for taking action b in any state and a reward of 0 for taking action a in
5
Figure 4: Chain Environment.
every state but the last, in which it receives a reward of 10. The optimal policy is to always choose
action a to reach the highest reward at the end of the chain, it is used as a positive control for this
experiment. We follow the protocol in previous publications and report the cumulative reward in
1000 steps, averaged over 500 runs. Our agent EB-CRP-PIG-R executes the EB-CRP-PIG strategy
for S steps, then computes the best reward policy given its internal model and executes it for the
remaining 1000-S steps. We found S=120 to be roughly optimal for our agent and display the
results of the experiment in Table 1, taking the results of the competitor algorithms directly from
the corresponding papers. The competitor algorithms define their own balance between exploitation
and exploration, leading to different results.
Method
RAM-RMAX [5]
BOSS [2]
exploit [15]
Bayesian DP [19]
EB-CRP-PIG-R
Optimal
Reward
2810
3003
3078
3158 ? 31
3182 ? 25
3658 ? 14
Table 1: Cumulative reward for 1000 steps in the chain environment.
The EB-CRP-PIG-R agent is able to perform the best and significantly outperforms many of the
other strategies. This result is remarkable because the EB-CRP-PIG-R agent has no prior knowledge
of the state space size, unlike all the competitor models. We also note that our algorithm is extremely
efficient computationally, it must approximate the optimal policy only once and then simply execute
it. In comparison, the exploit strategy [15] must compute the approximation at each time step.
Further, we interpret our competitive edge over BOSS to reflect a more efficient exploration strategy.
Specifically, BOSS uses LTA for exploration and Figure 3 indicates that the learning performance
of LTA is far worse than the performance of the PIG-based models.
Figure 5: Missing Information vs. Time for EB-CRP-PIG and CRP-PIG in the unbounded maze
environment.
Finally, we consider an unbounded maze environment with |S | being infinite and with multiple
absorbing states. Figure 5 shows the decrease of missing information (11) for the two CRP based
strategies. Interestingly, like in the bounded maze the Empirical Bayes version reduces the missing
information more rapidly than a CRP which has a fixed, but experimentally optimized, parameter
value. What is important about this result is that EB-CRP-PIG is not only better but it requires no
prior parameter tuning since ? is adjusted intrinsicially. Figure 6 shows how an EB-CRP-PIG and
an LTA agent explore the environment over 6000 steps. The missing information for each state is
6
Figure 6: Unbounded Maze environment. Exploration is depicted for two different agents (a) EBCRP-PIG and (b) LTA, after 2000, 4000, and 6000 exploration steps respectively. Initially all states
are white (not depicted), which represent unexplored states. Transporters (blue lines) move the agent
to the closest gravity well (small blue concentric rings). The current position of the agent is indicated
by the purple arrow.
color coded, light yellow representing high missing information, and red representing low missing
information, less than 1 bit. Note that the EB-CRP-PIG agent explores a much bigger area than the
LTA agent.
The two agents are also tested in a reward task in the unbounded environment for assessing whether
the exploration of EB-CRP-PIG leads to efficient reward acquisition. Specifically, we assign a reward to each state equal to the Euclidian distances from the starting state. Like for the Chain problem
before, we create two agents EB-CRP-PIG-R and LTA-R which each run for 1000 total steps, exploring for S=750 steps (defined previously) and then calculating their best reward policy and executing
it for the remaining 250 steps. The agents are repositioned to the start state after S steps and the
best reward policy is calculated. The simulation results are shown in Table 2. Clearly, the increased
coverage of the EB-CRP-PIG agent also results in higher reward acquisition.
Method
EB-CRP-PIG-R
LTA-R
Reward
1053
812
Table 2: Cumulative reward after 1000 steps in the unbounded maze environment.
7
4
Discussion
To be able to learn environments whose number of states is unknown or even unbounded is crucial
for applications in biology, as well as in robotics. Here we presented a principled information-based
strategy for an agent to learn a model of an unknown, unbounded environment. Specifically, the
proposed model uses the Chinese Restaurant Process (CRP) and a version of predicted information
gain (PIG) [12], adjusted for being able to accommodate comparisons of models with different
numbers of states.
We evaluated our model in three different environments in order to assess its performance. In the
bounded maze environment the new algorithm performed quite similarly to DP-PIG despite being at
a disadvantage in terms of prior knowledge. This result suggests that agents exploring environments
of unknown size can still develop accurate models of it quite rapidly. Since the new model is based
on the CRP, calculating the posterior and sampling from it is easily tractable.
The experiments in a simple bounded reward task, the Chain environment, were equally encouraging. Although the agent was unaware of the size of its environment, it was able to learn the states
and their transition probabilities quickly and retrieved a cumulative reward that was competitive with
published results. Some of the competitor strategies (exploit [15]) required to recompute the best
reward policy for each step. In contrast, EB-CRP-PIG computed the best policy only once, yet, was
able to outperform the exploit [15] strategy.
In the unbounded maze environment, EB-CRP-PIG was able to outperform CRP-PIG even though
it required no prior parameter tuning. In addition, it covered much more ground during exploration
than LTA, one of the few existing competitor models able to function in unbounded environments.
Specifically, the EB-CRP-PIG model evenly explored a large number of environmental states. In
contrast, LTA, exhaustively explored a much smaller area limited by two nearby absorbing states.
Two caveats need to be mentioned. First, although the computational complexity of the CRP is low,
the complexity of the value iteration algorithm scales linearly with the number of states discovered.
Thus, tractability of value iteration is an issue in EB-CRP-PIG. A possible remedy to this problem
would be to only calculate value iteration for states that are reachable from the current state in the
calculated time horizon. Second, the described padding procedure implicitly sets a balance between
seeking to discover new state transitions versus sampling from known ones. For different goals or
environments this balance may not be optimal, a future investigation of alternatives for comparing
models of different sizes would be very interesting.
All told, the proposed novel models overcome a major limitation of information-based learning
methods, the assumption of a bounded state space of known size. Since the new models are based
on the CRP, sampling is quite tractable. Interestingly, by applying Empirical Bayes for continuously
updating the parameter of the CRP, we are able to build agents that can explore bounded or unbounded environments with very little prior information. For describing learning in animals, models
that easily adapt to diverse environments could be crucial. Of course, other restrictictions in these
models still need to be addressed, in particular, the limitation to discrete and fully observable state
spaces. For example, the need to act in continuous state spaces is obviously crucial for animals and
robots. Further, recent literature [7] supports that information-based learning in partially observable
state spaces, like POMDPs [17], will be important to address applications in neuroscience.
5
Acknowledgements
JAA was funded by NSF grant IIS-1111765. FTS was supported by the Director, Office of Science, Office of Advanced Scientific Computing Research, Applied Mathematics program of the
U.S. Department of Energy under Contract No. DE-AC02-05CH11231. The authors thank Bruno
Olshausen, Tamara Broderick, and the members of the Redwood Center for Theoretical Neuroscience for their valuable input.
8
References
?
? e de Probabilit?es de Saint-Flour XIII1983,
[1] David Aldous. Exchangeability and related topics. Ecole
d?Et?
pages 1?198, 1985.
[2] John Asmuth, Lihong Li, Michael L Littman, Ali Nouri, and David Wingate. A bayesian sampling
approach to exploration in reinforcement learning. In Proceedings of the Twenty-Fifth Conference on
Uncertainty in Artificial Intelligence, pages 19?26. AUAI Press, 2009.
[3] Nihat Ay, Nils Bertschinger, Ralf Der, Frank G?uttler, and Eckehard Olbrich. Predictive information and
explorative behavior of autonomous robots. The European Physical Journal B, 63(3):329?339, 2008.
[4] Richard Bellman. E. 1957. dynamic programming. Princeton UniversityPress. BellmanDynamic programming1957, 1957.
[5] Ronen I Brafman and Moshe Tennenholtz. R-max-a general polynomial time algorithm for near-optimal
reinforcement learning. The Journal of Machine Learning Research, 3:213?231, 2003.
[6] Hugo Gimbert. Pure stationary optimal strategies in markov decision processes. In STACS 2007, pages
200?211. Springer, 2007.
[7] Jacqueline Gottlieb, Pierre-Yves Oudeyer, Manuel Lopes, and Adrien Baranes. Information-seeking,
curiosity, and attention: computational and neural mechanisms. Trends in cognitive sciences, 17(11):585?
593, 2013.
[8] Hemant Ishwaran and Lancelot F James. Generalized weighted chinese restaurant processes for species
sampling mixture models. Statistica Sinica, 13(4):1211?1236, 2003.
[9] Laurent Itti and Pierre Baldi. Bayesian surprise attracts human attention. Advances in neural information
processing systems, 18:547, 2006.
[10] Jaeyong Lee, Fernando A Quintana, Peter M?uller, and Lorenzo Trippa. Defining predictive probability
functions for species sampling models. Statistical science: a review journal of the Institute of Mathematical Statistics, 28(2):209, 2013.
[11] Daniel Y Little and Friedrich T Sommer. Learning in embodied action-perception loops through exploration. arXiv preprint arXiv:1112.1125, 2011.
[12] Daniel Y Little and Friedrich T Sommer. Learning and exploration in action-perception loops. Frontiers
in neural circuits, 7, 2013.
[13] David JC MacKay. Information-based objective functions for active data selection. Neural computation,
4(4):590?604, 1992.
[14] Jim Pitman. Exchangeable and partially exchangeable random partitions. Probability theory and related
fields, 102(2):145?158, 1995.
[15] Pascal Poupart, Nikos Vlassis, Jesse Hoey, and Kevin Regan. An analytic solution to discrete bayesian
reinforcement learning. In Proceedings of the 23rd international conference on Machine learning, pages
697?704. ACM, 2006.
[16] Mitsuo Sato, Kenichi Abe, and Hiroshi Takeda. Learning control of finite markov chains with an explicit trade-off between estimation and control. Systems, Man and Cybernetics, IEEE Transactions on,
18(5):677?684, 1988.
[17] Adhiraj Somani, Nan Ye, David Hsu, and Wee Sun Lee. Despot: Online pomdp planning with regularization. In Advances in Neural Information Processing Systems, pages 1772?1780, 2013.
[18] Jan Storck, Sepp Hochreiter, and J?urgen Schmidhuber. Reinforcement driven information acquisition
in non-deterministic environments. In Proceedings of the International Conference on Artificial Neural
Networks, Paris, volume 2, pages 159?164. Citeseer, 1995.
[19] Malcolm Strens. A bayesian framework for reinforcement learning. In ICML, pages 943?950, 2000.
[20] Yi Sun, Faustino Gomez, and J?urgen Schmidhuber. Planning to be surprised: Optimal bayesian exploration in dynamic environments. In Artificial General Intelligence, pages 41?51. Springer, 2011.
[21] Sebastian B Thrun. Efficient exploration in reinforcement learning. 1992.
[22] Jian Zhang, Zoubin Ghahramani, and Yiming Yang. A probabilistic model for online document clustering
with application to novelty detection. In NIPS, volume 4, pages 1617?1624, 2004.
9
| 5266 |@word nihat:1 exploitation:1 version:3 polynomial:1 simulation:2 citeseer:1 euclidian:1 accommodate:1 jacqueline:1 harder:1 initial:1 series:2 selecting:1 daniel:2 ecole:1 document:1 undiscovered:4 interestingly:3 outperforms:1 existing:1 current:6 comparing:3 manuel:1 yet:2 assigning:1 must:4 john:1 explorative:4 realistic:1 partition:2 distant:1 analytic:1 v:2 stationary:1 greedy:2 discovering:1 fewer:1 intelligence:2 ps0:2 recompute:1 caveat:1 zhang:1 five:1 unbounded:22 mathematical:2 director:1 surprised:1 consists:1 baldi:1 ch11231:1 expected:2 roughly:1 p1:1 planning:7 behavior:2 bellman:2 relying:1 decreasing:1 little:4 es0:1 encouraging:1 begin:1 discover:1 bounded:12 underlying:1 circuit:1 mass:1 what:1 rmax:1 developed:1 informed:2 berkeley:5 every:2 unexplored:1 act:1 auai:1 gravity:2 control:7 exchangeable:3 internally:2 grant:1 positive:2 before:1 switching:1 despite:1 hemant:1 laurent:1 path:1 ap:1 might:1 eb:23 minimally:1 suggests:1 challenging:1 ease:1 limited:2 averaged:2 directed:2 ered:1 procedure:5 lancelot:1 jan:1 probabilit:1 area:2 empirical:4 attain:1 significantly:3 induce:1 zoubin:1 cannot:1 selection:1 put:1 applying:1 restriction:1 deterministic:1 demonstrated:1 center:2 maximizing:2 missing:14 straightforward:1 attention:2 starting:1 jesse:1 upm:3 pomdp:1 sepp:1 mascheroni:1 pure:1 insight:1 ralf:1 eppf:4 handle:1 exploratory:1 autonomous:2 updated:3 target:5 programming:1 us:5 slip:1 origin:2 element:1 trend:1 updating:1 tributions:1 stacs:1 labeled:1 observed:1 bottom:1 ft:1 preprint:1 wingate:1 informationbased:1 calculate:1 revisiting:1 sun:2 counter:1 highest:3 decrease:2 valuable:1 trade:1 principled:1 mentioned:1 environment:37 complexity:2 broderick:1 reward:25 littman:1 exhaustively:1 dynamic:2 depend:1 ali:1 predictive:2 unequally:1 easily:2 baranes:1 differently:1 various:1 describe:2 hiroshi:1 artificial:4 formation:1 kevin:1 quite:5 whose:1 larger:2 solve:1 tested:3 relax:1 ability:1 statistic:2 itself:1 noisy:1 online:3 obviously:1 confronted:1 sequence:2 advantage:2 loop:3 rapidly:2 degenerate:1 poorly:1 takeda:1 assessing:4 ring:3 executing:1 yiming:1 derive:1 develop:1 pose:1 qt:1 polya:1 progress:2 sa:25 p2:1 coverage:1 predicted:4 involves:1 cmc:5 direction:1 correct:1 exploration:16 human:1 assign:1 generalization:1 wall:2 investigation:1 olbrich:1 im:3 adjusted:4 exploring:2 frontier:1 somani:1 ground:1 presumably:1 cb:1 propogate:1 major:1 purpose:1 estimation:1 faustino:1 create:1 weighted:1 uller:1 clearly:1 sensor:2 always:1 rather:1 exchangeability:2 publication:1 office:2 notational:1 improvement:1 likelihood:3 indicates:1 contrast:2 bos:3 storck:1 inference:1 initially:1 issue:1 pascal:1 denoted:2 adrien:1 animal:6 mackay:1 urgen:2 equal:3 once:2 field:1 sampling:12 biology:1 represents:2 icml:1 future:1 report:1 cardinal:1 few:2 employ:1 serious:1 richard:1 wee:1 divergence:1 replaced:1 argmax:1 attempt:1 detection:1 mitsuo:1 investigate:1 flour:1 mixture:1 light:2 chain:9 accurate:2 kt:9 tuple:1 edge:1 quintana:1 theoretical:3 increased:1 formalism:1 earlier:1 modeling:1 disadvantage:1 measuring:1 maximization:1 tractability:1 euler:1 uniform:3 despot:1 st:5 explores:1 international:2 lee:2 probabilistic:2 told:1 contract:1 off:1 michael:1 quickly:2 continuously:2 again:1 reflect:1 successively:1 choose:2 worse:1 external:1 cognitive:1 leading:2 itti:1 li:1 potential:1 de:3 jc:1 performed:1 observer:1 closed:2 observing:3 red:1 competitive:2 bayes:4 start:1 ass:2 purple:1 yves:1 largely:1 efficiently:3 yield:1 correspond:1 ronen:1 yellow:1 bayesian:9 pomdps:1 cybernetics:1 published:2 executes:2 reach:1 sebastian:1 definition:1 competitor:6 energy:1 acquisition:3 tamara:1 james:2 gain:10 sampled:1 hsu:1 popular:2 knowledge:3 color:1 back:1 higher:1 asmuth:1 follow:3 reflected:1 improved:1 done:1 execute:1 evaluated:1 though:1 crp:47 receives:2 assessment:1 gray:1 indicated:1 scientific:1 mdp:1 olshausen:1 building:1 ye:1 contain:1 true:4 remedy:1 hence:1 regularization:1 q0:1 white:1 during:2 strens:1 generalized:1 ay:1 theoretic:1 demonstrate:1 performs:2 nouri:1 harmonic:2 discovers:1 novel:1 common:2 absorbing:5 empirically:1 physical:1 hugo:1 volume:2 extend:2 interpret:1 tuning:2 rd:1 mathematics:1 similarly:2 bruno:1 reachable:3 funded:1 lihong:1 robot:4 longer:1 similarity:1 add:1 closest:1 own:1 recent:3 posterior:1 retrieved:1 aldous:1 driven:3 schmidhuber:2 der:1 yi:1 additional:1 somewhat:1 nikos:1 novelty:1 fernando:1 ii:1 multiple:1 reduces:1 faster:2 adapt:1 mle:2 equally:1 dkl:3 coded:1 controlled:1 impact:1 prediction:1 bigger:1 adhiraj:1 expectation:1 arxiv:2 iteration:5 kernel:3 represent:2 robotics:1 hochreiter:1 addition:1 remarkably:1 addressed:1 jian:1 crucial:3 unlike:1 member:1 near:1 yang:1 split:2 transporter:3 restaurant:5 psychology:1 attracts:1 opposite:1 reduce:1 idea:1 ac02:1 translates:2 whether:1 padding:5 harvest:1 peter:1 cause:1 action:26 useful:1 covered:1 discount:1 dark:1 simplest:1 reduced:1 outperform:3 nsf:1 neuroscience:4 blue:4 diverse:1 discrete:5 paste:1 key:1 four:1 drawn:1 backward:1 ram:1 padded:1 sum:1 sas0:10 run:3 uncertainty:1 lope:1 decision:2 bit:1 ct:11 nan:1 gomez:1 display:1 sato:1 deficiency:1 bp:1 sake:1 nearby:1 extremely:1 urn:1 department:1 arnemann:2 according:3 kenichi:1 describes:4 smaller:5 across:1 suppressed:1 making:1 s1:1 gradually:1 hoey:1 gathering:2 taken:3 ln:2 equation:2 computationally:1 previously:1 describing:3 count:3 lta:10 mechanism:1 tractable:2 end:1 available:1 apply:1 ishwaran:1 pierre:2 alternative:1 slower:1 denotes:1 clustering:2 sommer:3 dirichlet:4 remaining:2 saint:1 log2:1 calculating:4 exploit:4 testable:1 ghahramani:1 build:2 chinese:5 seeking:2 objective:5 move:3 quantity:2 moshe:1 strategy:20 parametric:4 primary:1 exhibit:1 amongst:1 dp:5 distance:2 separate:2 thank:1 thrun:1 evenly:1 topic:1 poupart:1 index:1 illustration:1 balance:3 sinica:1 frank:1 negative:1 design:1 policy:10 unknown:9 perform:1 twenty:1 observation:4 markov:5 benchmark:1 finite:4 displayed:1 beat:1 situation:1 defining:1 vlassis:1 jim:1 discovered:4 redwood:2 abe:1 concentric:3 introduced:1 david:4 pair:3 required:2 kl:1 paris:1 friedrich:3 optimized:3 california:1 learned:2 nip:1 address:4 able:10 tennenholtz:1 curiosity:1 perception:3 pig:38 challenge:1 program:1 max:2 advanced:2 representing:3 lorenzo:1 ss:2 embodied:4 prior:7 literature:5 acknowledgement:1 review:1 fully:3 interesting:2 limitation:2 regan:1 versus:3 remarkable:1 agent:51 s0:9 principle:1 pi:2 translation:1 course:1 supported:1 last:2 brafman:1 guide:1 allow:1 institute:1 taking:4 fifth:1 pitman:1 distributed:1 overcome:1 dimension:1 calculated:2 world:7 transition:17 valid:1 maze:13 cumulative:4 computes:1 commonly:1 reinforcement:8 unaware:1 ig:6 author:1 far:3 employing:1 transaction:1 approximate:1 observable:6 implicitly:1 active:3 repositioned:1 continuous:1 table:4 pkt:2 learn:5 ca:1 interact:1 european:1 protocol:1 statistica:1 linearly:1 rh:1 arrow:2 referred:1 depicts:1 slow:1 position:1 guiding:2 explicit:1 formula:1 down:1 explored:2 essential:1 adding:1 bertschinger:1 horizon:2 surprise:1 depicted:3 simply:2 explore:4 partially:3 applies:1 springer:2 environmental:1 acm:1 sized:1 marked:1 goal:1 room:1 man:1 experimentally:3 specifically:6 infinite:1 uniformly:2 gottlieb:1 total:1 specie:4 nil:1 oudeyer:1 experimental:2 e:1 hypothetically:1 internal:8 support:1 princeton:1 malcolm:1 |
4,712 | 5,267 | On the Computational Efficiency of Training Neural
Networks
Roi Livni
The Hebrew University
[email protected]
Shai Shalev-Shwartz
The Hebrew University
[email protected]
Ohad Shamir
Weizmann Institute of Science
[email protected]
Abstract
It is well-known that neural networks are computationally hard to train. On the
other hand, in practice, modern day neural networks are trained efficiently using SGD and a variety of tricks that include different activation functions (e.g.
ReLU), over-specification (i.e., train networks which are larger than needed), and
regularization. In this paper we revisit the computational complexity of training
neural networks from a modern perspective. We provide both positive and negative results, some of them yield new provably efficient and practical algorithms
for training certain types of neural networks.
1
Introduction
One of the most significant recent developments in machine learning has been the resurgence of
?deep learning?, usually in the form of artificial neural networks. A combination of algorithmic
advancements, as well as increasing computational power and data size, has led to a breakthrough
in the effectiveness of neural networks, and they have been used to obtain very impressive practical
performance on a variety of domains (a few recent examples include [17, 16, 24, 10, 7]).
A neural network can be described by a (directed acyclic) graph, where each vertex in the graph corresponds to a neuron and each edge is associated with a weight. Each neuron calculates a weighted
sum of the outputs of neurons which are connected to it (and possibly adds a bias term). It then
passes the resulting number through an activation function ? : R ? R and outputs the resulting
number. We focus on feed-forward neural networks, where the neurons are arranged in layers, in
which the output of each layer forms the input of the next layer. Intuitively, the input goes through
several transformations, with higher-level concepts derived from lower-level ones. The depth of the
network is the number of layers and the size of the network is the total number of neurons.
From the perspective of statistical learning theory, by specifying a neural network architecture (i.e.
the underlying graph and the activation function) we obtain a hypothesis class, namely, the set of all
prediction rules obtained by using the same network architecture while changing the weights of the
network. Learning the class involves finding a specific set of weights, based on training examples,
which yields a predictor that has good performance on future examples. When studying a hypothesis
class we are usually concerned with three questions:
1. Sample complexity: how many examples are required to learn the class.
2. Expressiveness: what type of functions can be expressed by predictors in the class.
3. Training time: how much computation time is required to learn the class.
For simplicity, let us first consider neural networks with a threshold activation function (i.e. ?(z) =
1 if z > 0 and 0 otherwise), over the boolean input space, {0, 1}d , and with a single output in
{0, 1}. The sample complexity of such neural networks is well understood [3]. It is known that the
VC dimension grows linearly with the number of edges (up to log factors). It is also easy to see that
no matter what the activation function is, as long as we represent each weight of the network using
1
a constant number of bits, the VC dimension is bounded by a constant times the number of edges.
This implies that empirical risk minimization - or finding weights with small average loss over the
training data - can be an effective learning strategy from a statistical point of view.
As to the expressiveness of such networks, it is easy to see that neural networks of depth 2 and
sufficient size can express all functions from {0, 1}d to {0, 1}. However, it is also possible to show
that for this to happen, the size of the network must be exponential in d (e.g. [19, Chapter 20]).
Which functions can we express using a network of polynomial size? The theorem below shows that
all boolean functions that can be calculated in time O(T (d)), can also be expressed by a network of
depth O(T (d)) and size O(T (d)2 ).
Theorem 1. Let T : N ? N and for every d, let Fd be the set of functions that can be implemented
by a Turing machine using at most T (d) operations. Then there exist constants b, c ? R+ such that
for every d, there is a network architecture of depth c T (d) + b, size of (c T (d) + b)2 , and threshold
activation function, such that the resulting hypotesis class contains Fd .
The proof of the theorem follows directly from the relation between the time complexity of programs
and their circuit complexity (see, e.g., [22]), and the fact that we can simulate the standard boolean
gates using a fixed number of neurons.
We see that from the statistical perspective, neural networks form an excellent hypothesis class; On
one hand, for every runtime T (d), by using depth of O(T (d)) we contain all predictors that can be
run in time at most T (d). On the other hand, the sample complexity of the resulting class depends
polynomially on T (d).
The main caveat of neural networks is the training time. Existing theoretical results are mostly
negative, showing that successfully learning with these networks is computationally hard in the worst
case. For example, neural networks of depth 2 contain the class of intersection of halfspaces (where
the number of halfspaces is the number of neurons in the hidden layer). By reduction to k-coloring,
it has been shown that finding the weights that best fit the training set is NP-hard ([9]). [6] has
shown that even finding weights that result in close-to-minimal empirical error is computationally
infeasible. These hardness results focus on proper learning, where the goal is to find a nearly-optimal
predictor with a fixed network architecture A. However, if our goal is to find a good predictor, there
is no reason to limit ourselves to predictors with one particular architecture. Instead, we can try,
for example, to find a network with a different architecture A0 , which is almost as good as the
best network with architecture A. This is an example of the powerful concept of improper learning,
which has often proved useful in circumventing computational hardness results. Unfortunately, there
are hardness results showing that even with improper learning, and even if the data is generated
exactly from a small, depth-2 neural network, there are no efficient algorithms which can find a
predictor that performs well on test data. In particular, [15] and [12] have shown this in the case of
learning intersections of halfspaces, using cryptographic and average case complexity assumptions.
On a related note, [4] recently showed positive results on learning from data generated by a neural
network of a certain architecture and randomly connected weights. However, the assumptions used
are strong and unlikely to hold in practice.
Despite this theoretical pessimism, in practice, modern-day neural networks are trained successfully
in many learning problems. There are several tricks that enable successful training:
? Changing the activation function: The threshold activation function, ?(a) = 1a>0 , has zero
derivative almost everywhere. Therefore, we cannot apply gradient-based methods with this activation function. To circumvent this problem, we can consider other activation functions. Most
1
widely known is a sigmoidal activation, e.g. ?(a) = 1+e
a , which forms a smooth approximation of the threshold function. Another recent popular activation function is the rectified linear
unit (ReLU) function, ?(a) = max{0, a}. Note that subtracting a shifted ReLU from a ReLU
yields an approximation of the threshold function, so by doubling the number of neurons we can
approximate a network with threshold activation by a network with ReLU activation.
? Over-specification: It was empirically observed that it is easier to train networks which are larger
than needed. Indeed, we empirically demonstrate this phenomenon in Sec. 5.
? Regularization: It was empirically observed that regularizing the weights of the network speeds
up the convergence (e.g. [16]).
2
The goal of this paper is to revisit and re-raise the question of neural network?s computational efficiency, from a modern perspective. This is a challenging topic, and we do not pretend to give any
definite answers. However, we provide several results, both positive and negative. Most of them are
new, although a few appeared in the literature in other contexts. Our contributions are as follows:
? We make a simple observation that for sufficiently over-specified networks, global optima are
ubiquitous and in general computationally easy to find. Although this holds only for extremely
large networks which will overfit, it can be seen as an indication that the computational hardness of learning does decrease with the amount of over-specification. This is also demonstrated
empirically in Sec. 5.
? Motivated by the idea of changing the activation function, we consider the quadratic activation
function, ?(a) = a2 . Networks with the quadratic activation compute polynomial functions of
the input in Rd , hence we call them polynomial networks. Our main findings for such networks
are as follows:
? Networks with quadratic activation are as expressive as networks with threshold activation.
? Constant depth networks with quadratic activation can be learned in polynomial time.
? Sigmoidal networks of depth 2, and with `1 regularization, can be approximated by polynomial
networks of depth O(log log(1/)). It follows that sigmoidal networks with `1 regularization
can be learned in polynomial time as well.
? The aforementioned positive results are interesting theoretically, but lead to impractical algorithms. We provide a practical, provably correct, algorithm for training depth-2 polynomial
networks. While such networks can also be learned using a linearization trick, our algorithm is
more efficient and returns networks whose size does not depend on the data dimension. Our algorithm follows a forward greedy selection procedure, where each step of the greedy selection
procedure builds a new neuron by solving an eigenvalue problem.
? We generalize the above algorithm to depth-3, in which each forward greedy step involves an
efficient approximate solution to a tensor approximation problem. The algorithm can learn a
rich sub-class of depth-3 polynomial networks.
? We describe some experimental evidence, showing that our practical algorithm is competitive
with state-of-the-art neural network training methods for depth-2 networks.
2
Sufficiently Over-Specified Networks Are Easy to Train
We begin by considering the idea of over-specification, and make an observation that for sufficiently
over-specified networks, the optimization problem associated with training them is generally quite
easy to solve, and that global optima are in a sense ubiquitous. As an interesting contrast, note that
for very small networks (such as a single neuron with a non-convex activation function), the associated optimization problem is generally hard, and can exhibit exponentially many local (non-global)
minima [5]. We emphasize that our observation only holds for extremely large networks, which will
overfit in any reasonable scenario, but it does point to a possible spectrum where computational cost
decreases with the amount of over-specification.
To present the result, let X ? Rd,m be a matrix of m training examples in Rd . We can think of the
network as composed of two mappings. The first maps X into a matrix Z ? Rn,m , where n is the
number of neurons whose outputs are connected to the output layer. The second mapping is a linear
mapping Z 7? W Z, where W ? Ro,n , that maps Z to the o neurons in the output layer. Finally,
there is a loss function ` : Ro,m ? R, which we?ll assume to be convex, that assesses the quality of
the prediction on the entire data (and will of course depend on the m labels). Let V denote all the
weights that affect the mapping from X to Z, and denote by f (V ) the function that maps V to Z.
The optimization problem associated with learning the network is therefore minW,V `(W f (V )).
The function `(W f (V )) is generally non-convex, and may have local minima. However, if n ? m,
then it is reasonable to assume that Rank(f (V )) = m with large probability (under some random
choice of V ), due to the non-linear nature of the function computed by neural networks1 . In that
case, we can simply fix V and solve minW `(W f (V )), which is computationally tractable as ` is
1
For example, consider the function computed by the first layer, X 7? ?(Vd X), where ? is a sigmoid
function. Since ? is non-linear, the columns of ?(Vd X) will not be linearly dependent in general.
3
assumed to be convex. Since f (V ) has full rank, the solution of this problem corresponds to a global
optima of `, and hence to a global optima of the original optimization problem. Thus, for sufficiently
large networks, finding global optima is generally easy, and they are in a sense ubiquitous.
3
The Hardness of Learning Neural Networks
We now review several known hardness results and apply them to our learning setting. For simplicity, throughout most of this section we focus on the PAC model in the binary classification case, over
the Boolean cube, in the realizable case, and with a fixed target accuracy.2
Fix some , ? ? (0, 1). For every dimension d, let the input space be Xd = {0, 1}d and let H be a
hypothesis class of functions from Xd to {?1}. We often omit the subscript d when it is clear from
context. A learning algorithm A has access to an oracle that samples x according to an unknown
distribution D over X and returns (x, f ? (x)), where f ? is some unknown target hypothesis in H.
The objective of the algorithm is to return a classifier f : X ? {?1}, such that with probability of
at least 1 ? ?,
Px?D [f (x) 6= f ? (x)] ? .
We say that A is efficient if it runs in time poly(d) and the function it returns can also be evaluated
on a new instance in time poly(d). If there is such A, we say that H is efficiently learnable.
In the context of neural networks, every network architecture defines a hypothesis class, Nt,n,? ,
that contains all target functions f that can be implemented using a neural network with t layers, n
neurons (excluding input neurons), and an activation function ?. The immediate question is which
Nt,n,? are efficiently learnable. We will first address this question for the threshold activation function, ?0,1 (z) = 1 if z > 0 and 0 otherwise.
Observing that depth-2 networks with the threshold activation function can implement intersections
of halfspaces, we will rely on the following hardness results, due to [15].
Theorem 2 (Theorem 1.2 in [15]). Let X = {?1}d , let
H a = x ? ?0,1 w> x ? b ? 1/2 : b ? N, w ? Nd , |b| + kwk1 ? poly(d) ,
and let Hka = {x ? h1 (x) ? h2 (x) ? . . . ? hk (x) : ?i, hi ? H a }, where k = d? for some constant
? > 0. Then under a certain cryptographic assumption, Hka is not efficiently learnable.
Under a different complexity assumption, [12] showed a similar result even for k = ?(1).
As mentioned before, neural networks of depth ? 2 and with the ?0,1 activation function can
express intersections of halfspaces: For example, the first layer consists of k neurons computing the
P k halfspaces, and the second layer computes their conjunction by the mapping x 7?
?0,1 ( i xi ? k + 1/2). Trivially, if some class H is not efficiently learnable, then any class containing it is also not efficiently learnable. We thus obtain the following corollary:
Corollary 1. For every t ? 2, n = ?(1), the class Nt,n,?0,1 is not efficiently learnable (under the
complexity assumption given in [12]).
What happens when we change the activation function? In particular, two widely used activation
functions for neural networks are the sigmoidal activation function, ?sig (z) = 1/(1 + exp(?z)),
and the rectified linear unit (ReLU) activation function, ?relu (z) = max{z, 0}.
As a first observation, note that for |z| 1 we have that ?sig (z) ? ?0,1 (z). Our data domain is
the discrete Boolean cube, hence if we allow the weights of the network to be arbitrarily large, then
Nt,n,?0,1 ? Nt,n,?sig . Similarly, the function ?relu (z)??relu (z?1) equals ?0,1 (z) for every |z| ? 1.
As a result, without restricting the weights, we can simulate each threshold activated neuron by two
ReLU activated neurons, which implies that Nt,n,?0,1 ? Nt,2n,?relu . Hence, Corollary 1 applies to
both sigmoidal networks and ReLU networks as well, as long as we do not regularize the weights of
the network.
2
While we focus on the realizable case (i.e., there exists f ? ? H that provides perfect predictions), with a
fixed accuracy () and confidence (?), since we are dealing with hardness results, the results trivially apply to
the agnostic case and to learning with arbitrarily small accuracy and confidence parameters.
4
What happens when we do regularize the weights? Let Nt,n,?,L be all target functions that can be
implemented using a neural network of depth t, size n, activation function ?, and when we restrict
the input weights of each neuron to be kwk1 + |b| ? L.
One may argue that in many real world distributions, the difference between the two classes, Nt,n,?,L
and Nt,n,?0,1 is small. Roughly speaking, when the distribution density is low around the decision
boundary of neurons (similarly to separation with margin assumptions), then sigmoidal neurons will
be able to effectively simulate threshold activated neurons.
In practice, the sigmoid and ReLU activation functions are advantageous over the threshold activation function, since they can be trained using gradient based methods. Can these empirical successes
be turned into formal guarantees? Unfortunately, a closer examination of Thm. 2 demonstrates that
if L = ?(d) then learning N2,n,?sig ,L and N2,n,?relu ,L is still hard. Formally, to apply these networks to binary classification, we follow a standard definition of learning with a margin assumption:
We assume that the learner receives examples of the form (x, sign(f ? (x))) where f ? is a real-valued
function that comes from the hypothesis class, and we further assume that |f ? (x)| ? 1. Even under
this margin assumption, we have the following:
Corollary 2. For every t ? 2, n = ?(1), L = ?(d), the classes Nt,n,?sig ,L and Nt,n,?relu ,L are not
efficiently learnable (under the complexity assumption given in [12]).
A proof is provided in the appendix. What happens when L is much smaller? Later on in the paper
we will show positive results for L being a constant and the depth being fixed. These results will be
obtained using polynomial networks, which we study in the next section.
4
Polynomial Networks
In the previous section we have shown several strong negative results for learning neural networks
with the threshold, sigmoidal, and ReLU activation functions. One way to circumvent these hardness
results is by considering another activation function. Maybe the simplest non-linear function is
the squared function, ?2 (x) = x2 . We call networks that use this activation function polynomial
networks, since they compute polynomial functions of their inputs. As in the previous section, we
denote by Nt,n,?2 ,L the class of functions that can be implemented using a neural network of depth
t, size n, squared activation function, and a bound L on the `1 norm of the input weights of each
neuron. Whenever we do not specify L we refer to polynomial networks with unbounded weights.
Below we study the expressiveness and computational complexity of polynomial networks. We
note that algorithms for efficiently learning (real-valued) sparse or low-degree polynomials has been
studied in several previous works (e.g. [13, 14, 8, 2, 1]). However, these rely on strong distributional
assumptions, such as the data instances having a uniform or log-concave distribution, while we are
interested in a distribution-free setting.
4.1
Expressiveness
We first show that, similarly to networks with threshold activation, polynomial networks of polynomial size can express all functions that can be implemented efficiently using a Turing machine.
Theorem 3 (Polynomial networks can express Turing Machines). Let Fd and T be as in Thm. 1.
Then there exist constants b, c ? R+ such that for every d, the class Nt,n,?2 ,L , with t =
c T (d) log(T (d)) + b, n = t2 , and L = b, contains Fd .
The proof of the theorem relies on the result of [18] and is given in the appendix.
Another relevant expressiveness result, which we will use later, shows that polynomial networks can
approximate networks with sigmoidal activation functions:
?
Theorem 4. Fix 0 < < 1, L ? 3 and t ? N. There are Bt ? O(log(tL
+ L log 1 )) and
1
?
Bn ? O(tL + L log ) such that for every f ? Nt,n,?sig ,L there is a function g ? NtBt ,nBn ,?2 , such
that supkxk? <1 kf (x) ? g(x)k? ? .
The proof relies on an approximation of the sigmoid function based on Chebyshev polynomials, as
was done in [21], and is given in the appendix.
5
4.2
Training Time
We now turn to the computational complexity of learning polynomial networks. We first show that
it is hard to learn polynomial networks of depth ?(log(d)). Indeed, by combining Thm. 4 and
Corollary 2 we obtain the following:
Corollary 3. The class Nt,n,?2 , where t = ?(log(d)) and n = ?(d), is not efficiently learnable.
On the flip side, constant-depth polynomial networks can be learned in polynomial time, using a
simple linearization trick. Specifically, the class of polynomial networks of constant depth t is
contained in the class of multivariate polynomials of total degree at most s = 2t . This class can
be represented as a ds -dimensional linear space, where each vector is the coefficient vector of some
such polynomial. Therefore, the class of polynomial networks of depth t can be learned in time
t
poly(d2 ), by mapping each instance vector x ? Rd to all of its monomials, and learning a linear
predictor on top of this representation (which can be done efficiently in the realizable case, or when
a convex loss function is used). In particular, if t is a constant then so is 2t and therefore polynomial
networks of constant depth are efficiently learnable. Another way to learn this class is using support
vector machines with polynomial kernels.
An interesting application of this observation is that depth-2 sigmoidal networks are efficiently learnable with sufficient regularization, as formalized in the result below. This contrasts with corollary 2,
which provides a hardness result without regularization.
Theorem 5. The class N2,n,?sig ,L can be learned, to accuracy , in time poly(T ) where T =
2
(1/) ? O(d4L ln(11L +1) ).
The idea of the proof is as follows. Suppose that we obtain data from some f ? N2,n,?sig ,L . Based
on Thm. 4, there is g ? N2Bt ,nBn ,?2 that approximates f to some fixed accuracy 0 = 0.5, where Bt
and Bn are as defined in Thm. 4 for t = 2. Now we can learn N2Bt ,nBn ,?2 by considering the class
of all polynomials of total degree 22Bt , and applying the linearization technique discussed above.
Since f is assumed to separate the data with margin 1 (i.e. y = sign(f ? (x)),|f ? (x)| ? 1|), then g
separates the data with margin 0.5, which is enough for establishing accuracy in sample and time
that depends polynomially on 1/.
4.3
Learning 2-layer and 3-layer Polynomial Networks
While interesting theoretically, the above results are not very practical, since the time and sample
complexity grow very fast with the depth of the network.3 In this section we describe practical,
provably correct, algorithms for the special case of depth-2 and depth-3 polynomial networks, with
some additional constraints. Although such networks can be learned in polynomial time via explicit
linearization (as described in section 4.2), the runtime and resulting network size scales quadratically
(for depth-2) or cubically (for depth-3) with the data dimension d. In contrast, our algorithms and
guarantees have a much milder dependence on d.
We first consider 2 layer polynomial networks, of the following form:
(
)
k
X
>
> 2
P2,k = x 7? b + w0 x +
?i (wi x) : ?i ? 1, |?i | ? 1, kwi k2 = 1 .
i=1
This networks corresponds to one hidden layer containing r neurons with the squared activation
function, where we restrict the input weights of all neurons in the network to have bounded `2 norm,
and where we also allow a direct linear dependency between the input layer and the output layer.
We?ll describe an efficient algorithm for learning this class, which is based on the GECO algorithm
for convex optimization with low-rank constraints [20].
3
If one uses SVM with polynomial kernels, the time and sample complexity may be small under margin
assumptions in a feature space corresponding to a given kernel. Note, however, that large margin in that space
is very different than the assumption we make here, namely, that there is a network with a small number of
hidden neurons that works well on the data.
6
The goal of the algorithm is to find f that minimizes the objective
m
R(f ) =
1 X
`(f (xi ), yi ),
m i=1
(1)
where ` : R ? R ? R is a loss function. We?ll assume that ` is ?-smooth and convex.
The basic idea of the algorithm is to gradually add hidden neurons to the hidden layer, in a greedy
manner, so as to decrease the loss function over the data. To do so, define V = {x 7? (w> x)2 :
kwk2 = 1} the set of functions that can be implemented by hidden neurons. Then every f ? P2,r
is an affine function plus a weighted sum of functions from V. The algorithm starts with f being
the minimizer of R over all affine functions. Then at each greedy step, we search for g ? V that
minimizes a first order approximation of R(f + ?g):
m
R(f + ?g) ? R(f ) + ?
1 X 0
` (f (xi ), yi )g(xi ) ,
m i=1
(2)
where `0 is the derivative of ` w.r.t. its first argument. Observe that for every g ? V there is some w
> 2
>
>
with kwk2 = 1 for which g(x) = (w
the right-hand side of Eq. (2) can
Pm x) 0 = w xx w.>Hence,
> 1
be rewritten as R(f ) + ? w m i=1 ` (f (xi ), yi )xi xi w . The vector w that minimizes this
Pm 0
1
>
expression (for positive ?) is the leading eigenvector of the matrix m
i=1 ` (f (xi ), yi )xi xi .
We add this vector as a hidden neuron to the network.4 Finally, we minimize R w.r.t. the weights
from the hidden layer to the output layer (namely, w.r.t. the weights ?i ).
The following theorem, which follows directly from Theorem 1 of [20], provides convergence guarantee for GECO. Observe that the theorem gives guarantee for learning P2,k if we allow to output
an over-specified network.
Theorem 6. Fix some > 0. Assume that the loss function is convex and ?-smooth. Then if
2
the GECO Algorithm is run for r > 2?k
iterations, it outputs a network f ? N2,r,?2 for which
?
R(f ) ? minf ? ?P2,k R(f ) + .
We next consider a hypothesis class consisting of third degree polynomials, which is a subset of
3-layer polynomial networks (see Lemma 1nin the appendix) . The hidden neurons
o will be functions
Qi
from the class: V = ?3i=1 Vi where Vi = x 7? j=1 (wj> x) : ?j, kwj k2 = 1 . The hypothesis
n
o
Pk
class we consider is P3,k = x 7? i=1 ?i gi (x) : ?i, |?i | ? 1, gi ? V .
The basic idea of the algorithm is the same as for 2-layer networks. However, while in the 2-layer
case we could implement efficiently each greedy step by solving an eigenvalue problem, we now
face the following tensor approximation problem at each greedy step:
m
max
g?V3
m
1 X 0
1 X 0
` (f (xi ), yi )g(xi ) =
max
` (f (xi ), yi )(w> xi )(u> xi )(v> xi ) .
m i=1
kwk=1,kuk=1,kvk=1 m
i=1
While this is in general a hard optimization problem, we can approximate it ? and luckily, an approximate greedy step suffices for success of the greedy procedure. This procedure is given in Figure 1,
and is again based on an approximate eigenvector computation. A guarantee for the quality of approximation is given in the appendix, and this leads to the following theorem, whose proof is given
in the appendix.
Theorem 7. Fix some ?, > 0. Assume that the loss function is convex and ?-smooth. Then if the
4d?k2
GECO Algorithm is run for r > (1??
)2 iterations, where each iteration relies on the approximation
procedure given in Fig. 1, then with probability (1??)r , it outputs a network f ? N3,5r,?2 for which
R(f ) ? minf ? ?P3,k R(f ? ) + .
4
It is also possible to find an approximate solution to the eigenvalue problem and still retain the performance
guarantees (see [20]). Since an approximate eigenvalue can be found in time O(d) using the power method, we
obtain the runtime of GECO depends linearly on d.
7
d
m
Input: {xi }m
i=1 ? R ? ? R , ? ,?
?
approximate
solution to
Output: A 1??
d
X
max
F (w, u, v) =
?i (w> xi )(u> xi )(v> xi )
kwk,kuk,kvk=1
i
Pick randomly w1 , . . . , ws iid according to N (0, Id ).
For t = 1, . . . , 2d log 1?
wt
wt ? kw
tk
P
Let A = i ?i (wt> xi )xi x>
i and set ut , vt s.t:
>
T r(u>
Av
)
?
(1
?
?
)
max
t
kuk,kvk=1 T r(u Av).
t
Return w, u, v the maximizers of maxi?s F (wi , ui , ui ).
Figure 1: Approximate tensor maximization.
5
Experiments
To demonstrate the practicality of GECO to train neural networks for real world problems, we considered a pedestrian detection problem as follows. We collected 200k training examples of image
patches of size 88x40 pixels containing either pedestrians (positive examples) or hard negative examples (containing images that were classified as pedestrians by applying a simple linear classifier in
a sliding window manner). See a few examples of images above. We used half of the examples as a
training set and the other half as a test set. We calculated HoG
features ([11]) from the images5 . We then trained, using GECO,
0.1
SGD ReLU
a depth-2 polynomial network on the resulting features. We
SGD Squared
GECO
9 ? 10
used 40 neurons in the hidden layer. For comparison we trained
the same network architecture (i.e. 40 hidden neurons with a
8 ? 10
squared activation function) by SGD. We also trained a similar
7 ? 10
network (40 hidden neurons again) with the ReLU activation
6 ? 10
function. For the SGD implementation we tried the following
tricks to speed up the convergence: heuristics for initialization
5 ? 10
of the weights, learning rate rules, mini-batches, Nesterov?s mo0
0.2
0.4
0.6
0.8
1
iterations
mentum (as explained in [23]), and dropout. The test errors of
?10
SGD as a function of the number of iterations are depicted on
4
1
2
the top plot of the Figure on the side. We also mark the perfor4
mance of GECO as a straight line (since it doesn?t involve SGD
3
8
iterations). As can be seen, the error of GECO is slightly bet2
ter than SGD. It should be also noted that we had to perform a
very large number of SGD iterations to obtain a good solution,
1
while the runtime of GECO was much faster. This indicates that
GECO may be a valid alternative approach to SGD for training
0
0
0.2
0.4
0.6
0.8
1
depth-2 networks. It is also apparent that the squared activation
#iterations
?10
function is slightly better than the ReLU function for this task.
?2
Error
?2
?2
?2
?2
MSE
5
5
The second plot of the side figure demonstrates the benefit of
over-specification for SGD. We generated random examples in R150 and passed them through a
random depth-2 network that contains 60 hidden neurons with the ReLU activation function. We
then tried to fit a new network to this data with over-specification factors of 1, 2, 4, 8 (e.g., overspecification factor of 4 means that we used 60 ? 4 = 240 hidden neurons). As can be clearly seen,
SGD converges much faster when we over-specify the network.
Acknowledgements: This research is supported by Intel (ICRI-CI). OS was also supported by
an ISF grant (No. 425/13), and a Marie-Curie Career Integration Grant. SSS and RL were also
supported by the MOS center of Knowledge for AI and ML (No. 3-9243). RL is a recipient of the
Google Europe Fellowship in Learning Theory, and this research is supported in part by this Google
Fellowship. We thank Itay Safran for spotting a mistake in a previous version of Sec. 2 and to James
Martens for helpful discussions.
5
Using the Matlab implementation provided in http://www.mathworks.com/matlabcentral/
fileexchange/33863-histograms-of-oriented-gradients.
8
References
[1] A. Andoni, R. Panigrahy, G. Valiant, and L. Zhang. Learning polynomials with neural networks. In ICML, 2014.
[2] A. Andoni, R. Panigrahy, G. Valiant, and L. Zhang. Learning sparse polynomial functions. In
SODA, 2014.
[3] M. Anthony and P. Bartlett. Neural Network Learning - Theoretical Foundations. Cambridge
University Press, 2002.
[4] S. Arora, A. Bhaskara, R. Ge, and T. Ma. Provable bounds for learning some deep representations. arXiv preprint arXiv:1310.6343, 2013.
[5] P. Auer, M. Herbster, and M. Warmuth. Exponentially many local minima for single neurons.
In NIPS, 1996.
[6] P. L. Bartlett and S. Ben-David. Hardness results for neural network approximation problems.
Theor. Comput. Sci., 284(1):53?66, 2002.
[7] Y. Bengio, A. Courville, and P. Vincent. Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35:1798?1828,
2013.
[8] E. Blais, R. O?Donnell, and K. Wimmer. Polynomial regression under arbitrary product distributions. Machine Learning, 80(2-3):273?294, 2010.
[9] A. Blum and R. Rivest. Training a 3-node neural network is np-complete. Neural Networks,
5(1):117?127, 1992.
[10] G. Dahl, T. Sainath, and G. Hinton. Improving deep neural networks for lvcsr using rectified
linear units and dropout. In ICASSP, 2013.
[11] N. Dalal and B. Triggs. Histograms of oriented gradients for human detection. In CVPR, 2005.
[12] A. Daniely, N. Linial, and S. Shalev-Shwartz. From average case complexity to improper
learning complexity. In FOCS, 2014.
[13] A. Kalai, A. Klivans, Y. Mansour, and R. Servedio. Agnostically learning halfspaces. SIAM J.
Comput., 37(6):1777?1805, 2008.
[14] A. Kalai, A. Samorodnitsky, and S.-H. Teng. Learning and smoothed analysis. In FOCS, 2009.
[15] A. Klivans and A. Sherstov. Cryptographic hardness for learning intersections of halfspaces.
In FOCS, 2006.
[16] A. Krizhevsky, I. Sutskever, and G. Hinton. Imagenet classification with deep convolutional
neural networks. In NIPS, 2012.
[17] Q. V. Le, M.-A. Ranzato, R. Monga, M. Devin, G. Corrado, K. Chen, J. Dean, and A. Y. Ng.
Building high-level features using large scale unsupervised learning. In ICML, 2012.
[18] N. Pippenger and M. Fischer. Relations among complexity measures. Journal of the ACM
(JACM), 26(2):361?381, 1979.
[19] S. Shalev-Shwartz and S. Ben-David. Understanding Machine Learning: From Theory to
Algorithms. Cambridge University Press, 2014.
[20] S. Shalev-Shwartz, A. Gonen, and O. Shamir. Large-scale convex minimization with a lowrank constraint. In ICML, 2011.
[21] S. Shalev-Shwartz, O. Shamir, and K. Sridharan. Learning kernel-based halfspaces with the
0-1 loss. SIAM Journal on Computing, 40(6):1623?1646, 2011.
[22] M. Sipser. Introduction to the Theory of Computation. Thomson Course Technology, 2006.
[23] I. Sutskever, J. Martens, G. Dahl, and G. Hinton. On the importance of initialization and
momentum in deep learning. In ICML, 2013.
[24] M. Zeiler and R. Fergus. Visualizing and understanding convolutional neural networks. arXiv
preprint arXiv:1311.2901, 2013.
9
| 5267 |@word version:1 dalal:1 polynomial:42 advantageous:1 norm:2 nd:1 triggs:1 d2:1 tried:2 bn:2 pick:1 sgd:12 reduction:1 contains:4 existing:1 com:1 nt:16 activation:43 must:1 devin:1 happen:1 plot:2 greedy:9 half:2 advancement:1 intelligence:1 warmuth:1 caveat:1 provides:3 node:1 sigmoidal:9 zhang:2 unbounded:1 direct:1 focs:3 consists:1 manner:2 theoretically:2 indeed:2 hardness:12 roughly:1 window:1 considering:3 increasing:1 begin:1 provided:2 underlying:1 bounded:2 circuit:1 agnostic:1 xx:1 rivest:1 what:5 minimizes:3 eigenvector:2 finding:6 transformation:1 impractical:1 guarantee:6 every:12 concave:1 xd:2 runtime:4 exactly:1 ro:2 classifier:2 demonstrates:2 k2:3 sherstov:1 unit:3 grant:2 omit:1 positive:7 before:1 understood:1 local:3 limit:1 supkxk:1 mistake:1 despite:1 id:1 establishing:1 subscript:1 plus:1 initialization:2 studied:1 specifying:1 challenging:1 weizmann:2 directed:1 practical:6 practice:4 definite:1 implement:2 procedure:5 networks1:1 empirical:3 confidence:2 cannot:1 close:1 selection:2 risk:1 context:3 applying:2 www:1 map:3 demonstrated:1 center:1 marten:2 dean:1 go:1 sainath:1 convex:10 simplicity:2 formalized:1 rule:2 regularize:2 shamir:4 target:4 suppose:1 itay:1 us:1 hypothesis:9 sig:8 trick:5 approximated:1 distributional:1 observed:2 preprint:2 worst:1 wj:1 connected:3 improper:3 ranzato:1 decrease:3 halfspaces:9 mentioned:1 complexity:17 ui:2 nesterov:1 mentum:1 trained:6 raise:1 depend:2 solving:2 linial:1 efficiency:2 learner:1 icassp:1 chapter:1 represented:1 train:5 fast:1 effective:1 describe:3 artificial:1 shalev:5 whose:3 quite:1 larger:2 widely:2 solve:2 say:2 valued:2 otherwise:2 heuristic:1 cvpr:1 gi:2 fischer:1 think:1 indication:1 eigenvalue:4 subtracting:1 product:1 turned:1 relevant:1 combining:1 sutskever:2 convergence:3 optimum:5 nin:1 perfect:1 converges:1 ben:2 tk:1 ac:3 lowrank:1 eq:1 p2:4 strong:3 implemented:6 c:1 involves:2 implies:2 come:1 correct:2 vc:2 luckily:1 human:1 enable:1 fix:5 suffices:1 safran:1 theor:1 hold:3 sufficiently:4 around:1 considered:1 roi:2 exp:1 algorithmic:1 mapping:6 mo:1 a2:1 label:1 successfully:2 weighted:2 minimization:2 clearly:1 kalai:2 conjunction:1 corollary:7 derived:1 focus:4 rank:3 indicates:1 hk:1 contrast:3 sense:2 realizable:3 helpful:1 milder:1 dependent:1 cubically:1 unlikely:1 entire:1 a0:1 bt:3 hidden:14 relation:2 w:1 interested:1 provably:3 pixel:1 aforementioned:1 classification:3 among:1 development:1 art:1 breakthrough:1 special:1 integration:1 cube:2 equal:1 having:1 ng:1 kw:1 unsupervised:1 icml:4 nearly:1 minf:2 future:1 np:2 t2:1 few:3 modern:4 randomly:2 oriented:2 composed:1 ourselves:1 consisting:1 detection:2 fd:4 kvk:3 activated:3 edge:3 closer:1 ohad:2 minw:2 re:1 theoretical:3 minimal:1 instance:3 column:1 boolean:5 maximization:1 cost:1 vertex:1 subset:1 monomials:1 daniely:1 uniform:1 predictor:8 krizhevsky:1 successful:1 dependency:1 answer:1 density:1 herbster:1 siam:2 huji:2 retain:1 donnell:1 pessimism:1 w1:1 squared:6 again:2 containing:4 possibly:1 derivative:2 leading:1 return:5 sec:3 coefficient:1 matter:1 pedestrian:3 depends:3 vi:2 sipser:1 later:2 view:1 try:1 h1:1 observing:1 kwk:2 competitive:1 start:1 shai:1 curie:1 contribution:1 ass:1 il:3 minimize:1 accuracy:6 convolutional:2 efficiently:15 yield:3 generalize:1 vincent:1 iid:1 rectified:3 straight:1 classified:1 whenever:1 definition:1 servedio:1 james:1 associated:4 proof:6 proved:1 popular:1 knowledge:1 ut:1 ubiquitous:3 auer:1 coloring:1 feed:1 higher:1 day:2 follow:1 specify:2 arranged:1 evaluated:1 done:2 overfit:2 hand:4 receives:1 d:1 apparent:1 expressive:1 o:1 google:2 defines:1 quality:2 icri:1 grows:1 building:1 concept:2 contain:2 regularization:6 hence:5 visualizing:1 ll:3 noted:1 complete:1 demonstrate:2 thomson:1 performs:1 hka:2 image:3 recently:1 sigmoid:3 empirically:4 rl:2 exponentially:2 discussed:1 approximates:1 isf:1 kwk2:2 significant:1 refer:1 cambridge:2 ai:1 rd:4 images5:1 trivially:2 pm:2 similarly:3 had:1 specification:7 access:1 impressive:1 europe:1 add:3 multivariate:1 recent:3 showed:2 perspective:5 scenario:1 certain:3 binary:2 arbitrarily:2 kwk1:2 success:2 vt:1 yi:6 seen:3 minimum:3 additional:1 v3:1 corrado:1 sliding:1 full:1 smooth:4 faster:2 long:2 calculates:1 prediction:3 qi:1 basic:2 regression:1 arxiv:4 iteration:8 represent:1 kernel:4 histogram:2 monga:1 fellowship:2 grow:1 pass:1 kwi:1 sridharan:1 effectiveness:1 call:2 ter:1 bengio:1 easy:6 concerned:1 enough:1 variety:2 affect:1 fit:2 relu:20 architecture:10 restrict:2 agnostically:1 idea:5 chebyshev:1 x40:1 motivated:1 expression:1 bartlett:2 passed:1 lvcsr:1 speaking:1 matlab:1 deep:5 useful:1 generally:4 clear:1 involve:1 maybe:1 amount:2 simplest:1 http:1 exist:2 revisit:2 shifted:1 sign:2 discrete:1 express:5 threshold:14 blum:1 changing:3 marie:1 kuk:3 dahl:2 graph:3 circumventing:1 sum:2 run:4 turing:3 everywhere:1 powerful:1 soda:1 almost:2 reasonable:2 throughout:1 separation:1 p3:2 patch:1 decision:1 appendix:6 nbn:3 bit:1 dropout:2 layer:24 hi:1 bound:2 courville:1 quadratic:4 oracle:1 constraint:3 x2:1 n3:1 simulate:3 speed:2 extremely:2 argument:1 klivans:2 px:1 according:2 combination:1 smaller:1 slightly:2 wi:2 happens:3 intuitively:1 gradually:1 explained:1 computationally:5 ln:1 turn:1 mathworks:1 needed:2 flip:1 ge:1 tractable:1 studying:1 operation:1 rewritten:1 mance:1 apply:4 observe:2 batch:1 alternative:1 gate:1 original:1 recipient:1 top:2 include:2 zeiler:1 pretend:1 practicality:1 build:1 tensor:3 objective:2 geco:12 question:4 strategy:1 dependence:1 shais:1 exhibit:1 gradient:4 separate:2 thank:1 sci:1 vd:2 w0:1 topic:1 mail:1 argue:1 collected:1 reason:1 provable:1 panigrahy:2 mini:1 hebrew:2 mostly:1 unfortunately:2 hog:1 negative:5 resurgence:1 implementation:2 proper:1 cryptographic:3 unknown:2 perform:1 av:2 neuron:35 observation:5 immediate:1 hinton:3 excluding:1 blais:1 rn:1 mansour:1 smoothed:1 arbitrary:1 thm:5 expressiveness:5 david:2 namely:3 required:2 specified:4 imagenet:1 learned:7 quadratically:1 wimmer:1 nip:2 address:1 able:1 spotting:1 usually:2 below:3 pattern:1 appeared:1 gonen:1 program:1 max:6 power:2 rely:2 circumvent:2 examination:1 technology:1 arora:1 ss:1 review:2 literature:1 acknowledgement:1 understanding:2 kf:1 loss:8 interesting:4 acyclic:1 h2:1 foundation:1 degree:4 affine:2 sufficient:2 course:2 supported:4 free:1 infeasible:1 bias:1 allow:3 formal:1 side:4 institute:1 face:1 livni:2 sparse:2 benefit:1 boundary:1 depth:33 dimension:5 calculated:2 world:2 rich:1 computes:1 doesn:1 forward:3 valid:1 polynomially:2 transaction:1 approximate:10 emphasize:1 dealing:1 ml:1 global:6 assumed:2 xi:22 shwartz:5 fergus:1 spectrum:1 search:1 learn:6 nature:1 career:1 improving:1 mse:1 excellent:1 poly:5 anthony:1 domain:2 pk:1 main:2 linearly:3 n2:5 fig:1 intel:1 tl:2 sub:1 momentum:1 explicit:1 exponential:1 comput:2 third:1 bhaskara:1 theorem:15 specific:1 showing:3 pac:1 learnable:10 maxi:1 svm:1 evidence:1 maximizers:1 exists:1 restricting:1 andoni:2 effectively:1 valiant:2 ci:1 importance:1 linearization:4 margin:7 chen:1 easier:1 intersection:5 led:1 depicted:1 simply:1 jacm:1 expressed:2 contained:1 doubling:1 kwj:1 applies:1 corresponds:3 minimizer:1 relies:3 acm:1 ma:1 goal:4 pippenger:1 samorodnitsky:1 hard:8 change:1 specifically:1 wt:3 lemma:1 total:3 teng:1 experimental:1 formally:1 support:1 mark:1 regularizing:1 phenomenon:1 |
4,713 | 5,268 | Attentional Neural Network: Feature Selection Using
Cognitive Feedback
Qian Wang
Department of Biomedical Engineering
Tsinghua University
Beijing, China 100084
[email protected]
Jiaxing Zhang
Microsoft Research Asia
5 Danning Road, Haidian District
Beijing, China 100080
[email protected]
Sen Song ?
Department of Biomedical Engineering
Tsinghua University
Beijing, China 100084
[email protected]
Zheng Zhang * ?
Department of Computer Science
NYU Shanghai
1555 Century Ave, Pudong
Shanghai, China 200122
[email protected]
Abstract
Attentional Neural Network is a new framework that integrates top-down cognitive bias and bottom-up feature extraction in one coherent architecture. The
top-down influence is especially effective when dealing with high noise or difficult segmentation problems. Our system is modular and extensible. It is also
easy to train and cheap to run, and yet can accommodate complex behaviors. We
obtain classification accuracy better than or competitive with state of art results
on the MNIST variation dataset, and successfully disentangle overlaid digits with
high success rates. We view such a general purpose framework as an essential
foundation for a larger system emulating the cognitive abilities of the whole brain.
1
Introduction
How our visual system achieves robust performance against corruptions is a mystery. Although its
performance may degrade, it is capable of performing denoising and segmentation tasks with different levels of difficulties using the same underlying architecture. Consider the first two examples
in Figure 1. Digits overlaid over random images are harder to recognize than those over random
noise, since pixels in the background images are structured and highly correlated. It is even more
challenging if two digits are overlaid altogether, in a benchmark that we call MNIST-2. Yet, with
different levels of efforts (and error rates), we are able to recognize these digits for all three cases.
Figure 1: Handwriting digits with different corruptions. From left to right: random background
noise, random background images, and MNIST-2
?
?
These authors supervised the project jointly and are co-corresponding authors.
Work partially done while at Microsoft Resarch Asia
1
Another interesting property of the human visual system is that recognition is fast for low noise
level but takes longer for cluttered scenes. Testers perform well on recognition tasks even when
the exposure duration is short enough to allow only one feed-forward pass [18], while finding the
target in cluttered scenes requires more time[4]. These evidences suggest that our visual system is
simultaneously optimized for the common, and over-engineered for the worst. One hypothesis is
that, when challenged with high noise, top-down ?explanations? propagate downwards via feedback
connections, and modulate lower level features in an iterative refinement process[19].
Inspired by these intuitions, we propose a framework called attentional neural network (aNN). aNN
is composed of a collection of simple modules. The denoising module performs multiplicative
feature selection controlled by a top-down cognitive bias, and returns a modified input. The classification module receives inputs from the denoising module and generates assignments. If necessary,
multiple proposals can be evaluated and compared to pick the final winner. Although the modules
are simple, their combined behaviors can be complex, and new algorithms can be plugged in to
rewire the behavior, e.g., a fast pathway for low noise, and an iterative mode for complex problems
such as MNIST-2. We have validated the performance of aNN on the MNIST variation dataset.
We obtained accuracy better than or competitive to state of art. In the challenging benchmark of
MNIST-2, we are able to predict one digit or both digits correctly more than 95% and 44% of the
time, respectively. aNN is easy to train and cheap to run. All the modules are trained with known
techniques (e.g. sparse RBM and back propagation), and inference takes much fewer rounds of
iterations than existing proposals based on generative models.
2
Model
aNN deals with two related issues: 1) constructing a segmentation module under the influence of
cognitive bias and 2) its application to the challenging task of classifying highly corrupted data. We
describe them in turn, and will conclude with a brief description of training methodologies.
2.1
Segmentation with cognitive bias
?
M
?
?
feedback
?
?
?
M
?>?
?
?
C
? = ?(? ? ?)
?
(b)
?? = ? ? ?
? = ?(? ? ?)
?
??
?
?
M
? = ?(?? ? ?? )
?
?
(a)
?>?
?
?
C
(c)
Figure 2: Segmentation module with cognitive bias (a) and classification based on that (b,c).
As illustrated in Figure 2(a), the objective of the segmentation module M is to segment out an object
y belonging to one of N classes in the noisy input image x. Unlike in the traditional deonising
models such as autoencoders, M is given a cognitive bias vector b ? {0, 1}N , whose ith element
indicates a prior belief on the existence of objects belonging to the i-th class in the noisy image.
During the bottom up pass, input image x is mapped into a feature vector h = ?(W ? x), where
W is the feature weight matrix and ? represents element-wise nonlinear Sigmoid function. During
the top-down pass, b generates a gating vector g = ?(U ? b) with the feedback weights U . g selects
and de-selects the features by modifying hidden activation hg = h g, where means pair-wised
multiplication. Reconstruction occurs from hg by z = ?(W 0 ? hg ). In general, bias b can be a
probability distribution indicating a mixture of several guesses, but in this paper we only use two
simpler scenarios: a binary vector to indicate whether there is a particular object with its associated
weights UG , or a group bias bG with equal values for all objects, which indicates the presence of
some object in general.
2
2.2
Classification
A simple strategy would be to feed the segmented input y into a classifier C. However, this suffers
from the loss of details during M ?s reconstruction and is prone to hallucinations, i.e. y transforming
to a wrong digit when given a wrong bias. We opted to use the reconstruction y to gate the raw
image x with a threshold to produce gated image z = (y > ) x for classification (Figure 2b). To
segment complex images, we explored an iterative design that is reminiscent of a recurrent network
(Figure 2c). At time step t, the input to the segmentation module M is zt = (yt?1 > ) x, and the
result yt is used for the next iteration. Consulting the raw input x each time prevents hallucination.
Alternatively, we could feed the intermediate representation hg to the classifier and such a strategy
gives reasonable performance (see section 3.2 group bias subsection), but in general this suffers
from loss of modularity.
For iterative classification, we can give the system an initial cognitive bias, and the system produces
a series of guesses b and classification results given by C. If the guess b is confirmed by the output
of C, then we consider b as a candidate for the final classification result. A wrong bias b will lead the
network to transform x to a different class, but the segmented images with the correct bias is often
still better than transformed images under the wrong bias. In the simplest version, we can give initial
bs over all classes and compare the fitness of the candidates. Such fitness metrics can be the number
of iterations it takes C to confirm the guess, the confidence of the confirmation , or a combination
of many related factors. For simplicity, we use the entropy of outputs of C, but more sophisticated
extensions are possible (see section 3.2 making it scalable subsection).
2.3
Training the model
We used a shallow network of RBM for the generative model, and autoencoders gave qualitatively
similar results. The parameters to be learned include the feature weights W and the feedback
weights U . The multiplicative nature of feature selection makes learning both W and U simultaneously problematic, and we overcame this problem with a two-step procedure: firstly, W is trained
with noisy data in a standalone RBM (i.e. with the feedback disabled); next, we fix W and learn
U with the noisy data as input but with clean data as target, using the standard back propagation
procedure. This forces U to learn to select relevant features and de-select distractors. We find it
helpful to use different noise levels in these two stages. In the results presented below, training W
and U uses half and full noise intensity, respectively. In practice, this simple strategy is surprisingly
effective (see Section 3). We found it important to use sparsity constraint when learning W to
produce local features. Global features (e.g. templates) tend to be activated by noises and data
alike, and tend to be de-selected by the feedback weights. We speculate that feature locality might
be especially important when compositionality and segmentation is considered. Jointly training the
features and the classifier is a tantalizing idea but proves to be difficult in practice as the procedure is
iterative and the feedback weights need to be handled. But attempts could be made in this direction
in the future to fine-tune performance for a particular task. Another hyper-parameter is the threshold
. We assume that there is a global minimum, and used binary search on a small validation set. 1
3
Results and Analysis
We used the MNIST variation dataset and MNIST-2 to evaluate the effectiveness of our framework.
MNIST-2 is composed by overlaying two randomly chosen clean MNIST digits. Unless otherwise
stated, we used an off-the-shelf classifier: a 3-layer perceptron with 256 hidden nodes, trained on
clean MNIST data with a 1.6% error rate. In the following sections, we will discuss bias-induced
feature selection, its application in denosing, segmentation and finally classification.
3.1
Effectiveness of feedback
If feature selection is sensitive to the cognitive bias b, then a given b should leads to the activation
of the corresponding relevant features. In Figure 3(a), we sorted the hidden units by the associated
1
The training and testing code can be found in https://github.com/qianwangthu/feedback-nips2014-wq.git
3
input
sum
no bias
b=0
group bias
b=1
b=2
correct bias
b=8
wrong bias
(a) Top features
(b) Reconstruction
sum
activated
group bias
b=1
b=2
(c) feature selection
Figure 3: The effectiveness of bias-controlled feature selection. (a) top features selected by different
cognitive bias (0, 1, 2, 8) and their accumulation; (b) denoising without bias, with group bias, correct
bias and wrong bias (b = 1); (c) how bias selects and de-selects features, the second and the third
rows correspond to the correct and wrong bias, respectively.
weights in U for a given bias from the set {0, 1, 2, 8}, and inspected their associated feature weights
in W. The top features, when superimposed, successfully compose a crude version of the target digit.
Since b controls feature selection, it can lead to effective segmentation (shown in Figure 3(b)))
By comparing the reconstruction results in the second row without bias, with those in the third
and fouth rows (with group bias and correct bias respectively), it is clear that segmentation quality
progressively improves. On the other hand, a wrong bias (fifth row) will try to select features to
its favor in two ways: selecting features shared with the correct bias, and hallucinating incorrect
features by segmenting from the background noises. Figure 3(c) goes further to reveal how feature
selection works. The first row shows features for one noisy input, sorted by their activity levels
without the bias. Next three rows show their deactivtion by the cognitive biases. The last column
shows a reconstructed image using the selected features in this figure. It is clear how a wrong bias
fails to produce a reasonable reconstructed image.
guess
guess
1?
1?
5
5
5
5
7
5
2?
1
1
1
1
1
1
2
2
2
2
2
2
3
3
3
3
3
3
4
4
4
4
4
4
5
5
5
5
5
5
7
9
9
9
9
9
2?
2
2
2
2
2
2
3?
3?
2
2
2
2
2
2
4?
4?
2
7
7
7
7
7
7?
5?
7
7
7
7
7
7
9?
9?
2
7
7
7
7
7
(a)
(b)
Figure 4: Recurrent segmentation examples in six iterations. In each iteration, the classification
result is shown under the reconstructed image, along with the confidence (red bar, the longer the
higher confidence).
As described in Section 2, segmentation might take multiple iterations, and each iteration produces a
reconstruction that can be processed by an off-the-shelf classifier. Figure 4 shows two cases, with as4
sociated predictions generated by the 3-layer MLP. In the first example (Figure 4(a)), two cognitive
biase guesses 2 and 7 are confirmed by the network, and the correct guess 2 has a greater confidence.
The second example (Figure 4(b)) illustratess that, under high intensity background, transformations
can happen and a hallucinated digit can be ?built? from a patch of high intensity region since they can
indiscriminately activate features. Such transformations constitute false-positives (i.e. confirming a
wrong guess) and pose challenges to classification. More complicated strategies such as local contrast normalization can be used in the future to deal with such cases. This phenomenon is not at all
uncommon in everyday life experiences: when truth is flooded with high noises, all interpretations
are possible, and each one picks evidence in its favor while ignoring others.
As described in Section 2, we used an entropy confidence metric to select the winner from candidates. The MLP classifier C produces a predicted score for the likelihood of each class, and we take
the total confidence as the entropy of the prediction distribution, normalized by its class average
obtained under clean data. This confidence metric, as well as the associated classification result, are
displayed under each reconstruction. The first example shows that confidence under the right guess
(i.e. 2) is higher. On the other hand, the second example shows that, with high noise, confidences
of many guesses are equally poor. Furthermore, more iterations often lead to higher confidence,
regardless of whether the guess is correct or not. This self-fulfilling process locks predictions to
their given biases, instead of differentiating them, which is also a familiar scenario.
3.2
Classification
0.2
Table 1: Classification performance
mnist-background-noise
mnist-background-image
err rate
0.15
back-rand
back-image
RBM
11.39
15.42
0.05
imRBM
10.46
16.35
0
discRBM
10.29
15.56
DBN-3
6.73
16.31
CAE-2
10.90
15.50
PGBM
6.08
12.25
sDBN
4.48
14.34
aNN - ?rand
3.22
22.30
aNN - ?image
6.09
15.33
0.1
(a)
0
0.2
0.4
0.6
background level
0.8
1
0.25
err rate
0.2
false negative
false positive
0.15
(b)
0.1
0.05
0
1
2
3
iteration
4
5
Figure 5: (a) error vs. background level. (b) error vs. iteration number.
To compare with previous results, we used the standard training/testing split (12K/50K) of the
MNIST variation set, and results are shown in the Table 1. We ran one-iteration denoising, and
then picked the winner by comparing normalized entropies among the candidates, i.e. those with
biases matching the prediction of the 3-layer MLP classifier. We trained two parameter sets separately in random-noise background (?rand ) and image background dataset(?image ). To test transfer
abilities, we also applied ?image to random-noise background data and ?rand to image background
data. On MNIST-back-rand and MNIST-back-image dataset, ?noise achieves 3.22% and 22.3% err
rate respectively, while ?image achieves 6.09% and 15.33%.
Figure 5(a) shows how the performance deteriorates with increasing noise level. In these experiments, random noise and random images are modulated by scaling down their pixel intensity linearly. Intuitively, at low noise the performance should approach the default accuracy of the classifier
C and is indeed the case.
The effect of iterations: We have chosen to run only one iteration because under high noise, each
guess will insist on picking features to its favor and some hallucination can still occur. With more
iterations, false positive rates will rise and false negative rates will decrease, as confidence scores for
5
both the right and the wrong guesses will keep on improving. This is shown in Figure-5(b). As such,
more iterations do not necessarily lead to better performance. In the current model, the predicted
class from the previous step is not feed into the next step, and more sophisticated strategies with
such an extension might produce better results in the future.
The power of group bias: For this benchmark, good performance mostly depends on the quality of
segmentation. Therefore, a simpler approach is to denoise with coarse-grained group bias, followed
by classification. For ?image , we attached a SVM to the hidden units with bG turned on, and obtained
a 16.2% error rate. However, if we trained a SVM with 60K samples, the error rate drops to 12.1%.
This confirms that supervised learning can achieve better performance with more training data.
Making it scalable. So far, we enumerate over all the guesses. This is clearly not scalable if number
of classes is large. One sensible solution is to first denoise with a group bias bG , and pick top-K
candidates from the prediction distribution, and then iterate among them.
Finally, we emphasize that the above results are obtained with only one up-down pass. This is in
stark contrast to other generative model based systems. For example, in PGBM [15], each inference
takes 25 rounds.
3.3
MNIST-2 problem
Compared to corruption by background noises, MNIST-2 is a much more challenging task, even for
a human observer. It is a problem of segmentation, not denoising. In fact, such segmentation requires
semantic understanding of the object. Knowing which features are task-irrelevant is not sufficient,
we need to discover and utilize per-class features. Any denoising architectures only removing taskirrelevant features will fail on such a task without additional mechanisms. In aNN, each bias has its
own associated features and explicitly call these features out in the reconstruction phase (modulated
by input activations). Meanwhile, its framework permits multiple predictions so it can accommodate
such problems.
ground truth
ground truth
guess
ground truth
guess
2?
guess
2?
2
2
2
2
2
2
6?
1?
2
2
2
2
2
2
7?
6
6
6
6
6
6
1?
7
7
7
7
7
7
0?
2
2
2
2
2
2
4?
6
4
4
4
4
2
4
2
2
2
2
2
2
2
2
2
2
1
1
1
1
5
5
5
5
5
5
8
8
8
2
2
2
3
3
3
3
3
3
8
4
4
4
4
4
3?
1
1
1
1
1
1
4?
2
1
2?
1?
8?
2
5?
4?
7
7
4
(a)
(b)
4
4
4
(c)
Figure 6: Sample results on MNIST-2. In each example, each column is one iteration. The first two
rows are runs with two ground truth digits, others are with wrong biases.
For the MNIST-2 task, we used the same off-the-shelf 3-layer classifier to validate a guess. In the
first two examples in Figure 6, the pair of digits in the ground truth is correctly identified. Supplying
either digit as the bias successfully segments its features, resulting in imperfect reconstructions that
are nonetheless confident enough to win over competing proposals. One would expect that the
random nature of MNIST-2 would create much more challenging (and interesting) cases that either
defy or confuse any segmentation attempts. This is indeed true. The last example is an overlay of
the digit 1 and 5 that look like a perfect 8. Each of the 5 biases successfully segment out their target
?digit?, and sometimes creatively. It is satisfying to see that a human observor would make similar
misjudgements in those cases.
6
ground truth image
result
ground truth image
result
ground truth image
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
(a)
(b)
result
(c)
Figure 7: Sample results on MNIST-2 when adding background noises. (a) (b) (c) are examples
three groups of results, when both digits, one digit, or none are predicted, respectively.
Out of the 5000 MNIST-2 pairs, there are 95.46% and 44.62% cases where at least one digit or
both digits get correctly predicted, respectively. Given the challenging nature of the benchmark, we
are surprised by this performance. Contrary to random background dataset, in this problem, more
iterations conclusively lead to better performance. The above accuracy is obtained with 5 iterations,
and the accuracy for matching both digits will drop to 36.28% if only 1 iteration is used. Even more
interestingly, this performance is resilient against background noise (Figure 7), the accuracy only
drops slightly (93.72% and 41.66%). The top-down biases allowed us to achieve segmentaion and
denoising at the same time.
4
4.1
Discussion and Related Work
Architecture
Feedforward multilayer neural networks have achieved good performance in many classification
tasks in the past few years, notably achieving the best performance in the ImageNet competition
in vision([21] [7]). However, they typically give a fixed outcome for each input image, therefore
cannot naturally model the influence of cognitive biases and are difficult to incorporate into a larger
cognitive framework. The current frontier of vision research is to go beyond object recognition
towards image understanding [16]. Inspired by neuroscience research, we believe that an unified
module which integrates feedback predictions and interpretations with information from the world
is an important step towards this goal.
Generative models have been a popular approach([5, 13]). They are typically based on a probabilistic framework such as Boltzmann Machines and can be stacked into a deep architecture. They have
advantages over discriminative models in dealing with object occlusion. In addition, prior knowledge can be easily incorporated in generative models in the forms of latent variables. However,
despite the mathematical beauty of a probabilistic framework, this class of models currently suffer
from the difficulty of generative learning and have been mostly successful in learning small patches
of natural images and objects [17, 22, 13]. In addition, inferring the hidden variables from images
is a difficult process and many iterations are typically needed for the model to converge[13, 15]. A
recent trend is to first train a DBN or DBM model then turn the model into a discriminative network
for classification. This allows for fast recognition but the discriminative network loses the generative
ability and cannot combine top-down and bottom-up information.
We sought a simple architecture that can flexibly navigate between discriminative and generative
frameworks. This should ideally allow for one-pass quick recognition for images with easy and
well-segmented objects, but naturally allow for iteration and influence by cognitive-bias when the
need for segmentation arises in corrupted or occluded image settings.
4.2
Models of Attention
In the field of computational modeling of attention, many models have been proposed to model the
saliency map and used to predict where attention will be deployed and provide fits to eye-tracking
data[1]. We are instead more interested in how attentional signals propagating back from higher lev7
els in the visual hierarchy can be merged with bottom up information. Volitional top-down control
could update, bias or disambiguate the bottom-up information based on high-level tasks, contextual cues or behavior goals. Computational models incorporating this principle has so far mostly
focused on spatial attention [12, 1]. For example, in a pedestrian detection task, it was shown that
visual search can be sped up if the search is limited to spatial locations of high prior or posterior
probabilities [3]. However, human attention abilities go beyond simple highlighting based on location. For example, the ability to segment and disentangle object based on high level expectations
as in the MNIST-2 dataset represents an interesting case. Here, we demonstrate that top-down attention can also be used to segment out relevant parts in a cluttered and entangled scene guided
by top-down interpretation, demonstrating that attentional bias can be successfully deployed on a
far-more fine-grained level than previous realized.
We have chosen the image-denoising and image-segmentation tasks as our test cases. In the context
of image-denoising, feedforward neural networks have been shown to have good performance [6,
20, 11]. However, their work has not included a feedback component and has no generative ability.
Several Boltzmann machine based architectures have been proposed[9, 8]. In PGBM, gates on input
images are trained to partition such pixel as belonging to objects or backgrounds, which are modeled
by two RBMs separately [15]. The gates and the RBM components make up a high-order RBM.
However, such a high-order RBM is difficult to train and needs costly iterations during inference.
sDBN [17] used a RBM to model the distribution of the hidden layer, and then denoises the hidden
layer by Gibbs sampling over the hidden units affected by noise. Besides the complexity of Gibbs
sampling, the process of iteratively finding which units are affected by noise is also complicated and
costly, as there is a process of Gibbs sampling for each unit. When there are multiple digits appearing
in the image as in the case of MNSIT-2, the hidden layer denoising step leads to uncertain results,
and the best outcome is an arbitrary choice of one of the mixed digits. a DBM based architecture has
also been proposed for modeling attention, but the complexity of learning and inference also makes
it difficult to apply in practice [10]. All those works also lack the ability of controlled generation
and input reconstruction under the direction of a top-down bias.
In our work, top-down biases influence the processing of feedforward information at two levels. The
inputs are gated at the raw image stage by top-down reconstructions. We propose that this might be
equivalent to the powerful gating influence of the thalamus in the brain [1, 15]. If the influence of
input image is shut off at this stage, then the system can engage in hallucination and might get into a
state akin to dreams, as when the thalamic gates are closed. Top-down biases also affect processing
at a higher stage of high-level features. We think this might be equivalent to the processing level of
V4 in the visual hierarchy. At this level, top-down biases mostly suppresses task-irrelevant features
and we have modeled the interactions as multiplicative in accordance with results from neuroscience
research [1, 2].
4.3
Philosophical Points
The issue of whether top-down connections and iterative processing are useful for object recognition has been a point of hot contention. Early work inspired by Hopfield network and the tradition
of probabilistic models based on Gibbs sampling argue for the usefulness of feedback and iteration
[14],[13], but results from neuroscience research and recent success by purely feedforward networks
argue against it [18],[7]. In our work, we find that feedforward processing is sufficient for good performance on clean digits. Feedback connections play an essential role for digit denoising. However,
one pass with a simple cognitive bias towards digits seems to suffice and iteration seems only to confirm the initial bias and does not improve performance. We hypothesize that this ?see what you want
to see? is a side-effect of our ability to denoise a cluttered scene, as the deep hierarchy possesses the
ability to decompose objects into many shareable parts. In the more complex case of MNIST-2, performance does increase with iteration. This suggests that top-down connections and iteration might
be particularly important for good performance in the case of cluttered scenes. The architecture we
proposed can naturally accommodate all these task requirements simultaneously with essentially no
further fine-tuning. We view such a general purpose framework as an essential foundation for a
larger system emulating the cognitive abilities of the whole brain.
8
References
[1] F. Baluch and L. Itti. Mechanisms of top-down attention. Trends in neurosciences, 34(4):210?
224, 2011.
[2] T. C
? ukur, S. Nishimoto, A. G. Huth, and J. L. Gallant. Attention during natural vision warps
semantic representation across the human brain. Nature neuroscience, 16(6):763?770, 2013.
[3] K. A. Ehinger, B. Hidalgo-Sotelo, A. Torralba, and A. Oliva. Modelling search for people in
900 scenes: A combined source model of eye guidance. Visual cognition, 17(6-7):945?978,
2009.
[4] J. M. Henderson, M. Chanceaux, and T. J. Smith. The influence of clutter on real-world scene
search: Evidence from search efficiency and eye movements. Journal of Vision, 9(1):32, 2009.
[5] G. E. Hinton, S. Osindero, and Y.-W. Teh. A fast learning algorithm for deep belief nets.
Neural computation, 18(7):1527?1554, 2006.
[6] V. Jain and H. S. Seung. Natural image denoising with convolutional networks. In NIPS,
volume 8, pages 769?776, 2008.
[7] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional
neural networks. In NIPS, volume 1, page 4, 2012.
[8] H. Larochelle and Y. Bengio. Classification using discriminative restricted boltzmann machines. In Proceedings of the 25th international conference on Machine learning, pages 536?
543. ACM, 2008.
[9] V. Nair and G. E. Hinton. Implicit mixtures of restricted boltzmann machines. In NIPS,
volume 21, pages 1145?1152, 2008.
[10] D. P. Reichert, P. Series, and A. J. Storkey. A hierarchical generative model of recurrent objectbased attention in the visual cortex. In Artificial Neural Networks and Machine Learning?
ICANN 2011, pages 18?25. Springer, 2011.
[11] S. Rifai, P. Vincent, X. Muller, X. Glorot, and Y. Bengio. Contractive auto-encoders: Explicit
invariance during feature extraction. In Proceedings of the 28th International Conference on
Machine Learning (ICML-11), pages 833?840, 2011.
[12] A. L. Rothenstein and J. K. Tsotsos. Attention links sensing to recognition. Image and Vision
Computing, 26(1):114?126, 2008.
[13] R. Salakhutdinov and G. E. Hinton. Deep boltzmann machines. In International Conference
on Artificial Intelligence and Statistics, pages 448?455, 2009.
[14] F. Schwenker, F. T. Sommer, and G. Palm. Iterative retrieval of sparsely coded associative
memory patterns. Neural Networks, 9(3):445?455, 1996.
[15] K. Sohn, G. Zhou, C. Lee, and H. Lee. Learning and selecting features jointly with pointwise gated {B} oltzmann machines. In Proceedings of The 30th International Conference on
Machine Learning, pages 217?225, 2013.
[16] C. Tan, J. Z. Leibo, and T. Poggio. Throwing down the visual intelligence gauntlet. In Machine
Learning for Computer Vision, pages 1?15. Springer, 2013.
[17] Y. Tang and C. Eliasmith. Deep networks for robust visual recognition. In Proceedings of the
27th International Conference on Machine Learning (ICML-10), pages 1055?1062, 2010.
[18] S. Thorpe, D. Fize, C. Marlot, et al. Speed of processing in the human visual system. nature,
381(6582):520?522, 1996.
[19] S. Ullman. Sequence seeking and counter streams: a computational model for bidirectional
information flow in the visual cortex. Cerebral cortex, 5(1):1?11, 1995.
[20] P. Vincent, H. Larochelle, Y. Bengio, and P.-A. Manzagol. Extracting and composing robust
features with denoising autoencoders. In Proceedings of the 25th international conference on
Machine learning, pages 1096?1103. ACM, 2008.
[21] M. D. Zeiler and R. Fergus. Visualizing and understanding convolutional neural networks.
arXiv preprint arXiv:1311.2901, 2013.
[22] D. Zoran and Y. Weiss. From learning models of natural image patches to whole image restoration. In Computer Vision (ICCV), 2011 IEEE International Conference on, pages 479?486.
IEEE, 2011.
9
| 5268 |@word version:2 seems:2 confirms:1 propagate:1 git:1 pick:3 harder:1 accommodate:3 initial:3 series:2 score:2 selecting:2 interestingly:1 past:1 existing:1 err:3 current:2 com:4 comparing:2 segmentaion:1 contextual:1 activation:3 gmail:2 yet:2 reminiscent:1 happen:1 partition:1 confirming:1 cheap:2 hypothesize:1 drop:3 progressively:1 standalone:1 v:2 update:1 generative:10 fewer:1 guess:19 half:1 selected:3 cue:1 shut:1 intelligence:2 ith:1 smith:1 short:1 supplying:1 coarse:1 consulting:1 node:1 location:2 district:1 simpler:2 zhang:2 firstly:1 mathematical:1 along:1 surprised:1 incorrect:1 pathway:1 compose:1 combine:1 notably:1 indeed:2 behavior:4 brain:4 inspired:3 insist:1 salakhutdinov:1 increasing:1 project:1 discover:1 underlying:1 suffice:1 what:1 suppresses:1 unified:1 finding:2 transformation:2 classifier:9 wrong:12 control:2 unit:5 overlaying:1 segmenting:1 positive:3 engineering:2 local:2 accordance:1 tsinghua:2 despite:1 might:7 china:4 suggests:1 challenging:6 co:1 limited:1 contractive:1 testing:2 practice:3 digit:26 procedure:3 matching:2 confidence:11 road:1 suggest:1 get:2 cannot:2 selection:9 context:1 influence:8 accumulation:1 equivalent:2 map:1 quick:1 yt:2 nishimoto:1 exposure:1 go:3 regardless:1 cluttered:5 duration:1 flexibly:1 attention:11 focused:1 simplicity:1 qian:1 century:1 variation:4 target:4 inspected:1 hierarchy:3 engage:1 play:1 tan:1 us:1 hypothesis:1 element:2 trend:2 recognition:8 satisfying:1 particularly:1 storkey:1 sparsely:1 bottom:5 role:1 module:11 preprint:1 wang:1 worst:1 region:1 decrease:1 movement:1 counter:1 ran:1 intuition:1 transforming:1 complexity:2 ideally:1 seung:1 occluded:1 trained:6 zoran:1 segment:6 purely:1 efficiency:1 cae:1 easily:1 hopfield:1 train:4 stacked:1 jain:1 fast:4 effective:3 describe:1 activate:1 artificial:2 hyper:1 outcome:2 sociated:1 whose:1 modular:1 larger:3 otherwise:1 ability:10 favor:3 statistic:1 think:1 jointly:3 noisy:5 transform:1 final:2 associative:1 advantage:1 sequence:1 net:1 sen:2 propose:2 reconstruction:11 interaction:1 relevant:3 turned:1 achieve:2 description:1 validate:1 everyday:1 competition:1 sutskever:1 requirement:1 produce:7 perfect:1 object:14 recurrent:3 propagating:1 pose:1 predicted:4 indicate:1 larochelle:2 direction:2 tester:1 guided:1 merged:1 correct:8 modifying:1 human:6 engineered:1 eliasmith:1 resilient:1 fix:1 decompose:1 extension:2 frontier:1 considered:1 ground:8 overlaid:3 cognition:1 predict:2 dbm:2 achieves:3 sought:1 early:1 torralba:1 purpose:2 integrates:2 currently:1 sensitive:1 create:1 successfully:5 clearly:1 modified:1 zhou:1 shelf:3 beauty:1 validated:1 modelling:1 indicates:2 superimposed:1 likelihood:1 opted:1 ave:1 contrast:2 tradition:1 helpful:1 inference:4 el:1 typically:3 hidden:9 transformed:1 selects:4 interested:1 pixel:3 issue:2 classification:19 among:2 art:2 spatial:2 equal:1 field:1 extraction:2 sampling:4 zz:1 represents:2 look:1 icml:2 future:3 others:2 few:1 thorpe:1 randomly:1 composed:2 simultaneously:3 recognize:2 fitness:2 familiar:1 phase:1 occlusion:1 microsoft:3 attempt:2 detection:1 mlp:3 highly:2 zheng:1 marlot:1 hallucination:4 henderson:1 uncommon:1 mixture:2 activated:2 hg:4 capable:1 necessary:1 experience:1 poggio:1 unless:1 creatively:1 plugged:1 guidance:1 uncertain:1 column:2 modeling:2 extensible:1 challenged:1 assignment:1 restoration:1 usefulness:1 krizhevsky:1 successful:1 osindero:1 encoders:1 corrupted:2 combined:2 confident:1 international:7 probabilistic:3 off:4 v4:1 lee:2 picking:1 misjudgements:1 cognitive:18 denoises:1 itti:1 return:1 stark:1 ullman:1 de:4 speculate:1 pedestrian:1 explicitly:1 bg:3 stream:1 depends:1 multiplicative:3 view:2 try:1 closed:1 picked:1 thu:1 observer:1 red:1 competitive:2 thalamic:1 complicated:2 objectbased:1 accuracy:6 convolutional:3 correspond:1 saliency:1 raw:3 vincent:2 none:1 confirmed:2 corruption:3 suffers:2 against:3 nonetheless:1 rbms:1 naturally:3 associated:5 rbm:8 handwriting:1 dataset:7 popular:1 subsection:2 distractors:1 improves:1 knowledge:1 segmentation:19 sophisticated:2 back:7 feed:4 bidirectional:1 higher:5 supervised:2 asia:2 methodology:1 wei:1 rand:5 done:1 evaluated:1 furthermore:1 biomedical:2 stage:4 implicit:1 autoencoders:3 hand:2 receives:1 nonlinear:1 propagation:2 lack:1 mode:1 quality:2 reveal:1 disabled:1 believe:1 effect:2 normalized:2 true:1 iteratively:1 semantic:2 illustrated:1 deal:2 round:2 visualizing:1 during:6 self:1 demonstrate:1 performs:1 image:45 wise:1 contention:1 common:1 sigmoid:1 ug:1 sped:1 shanghai:2 winner:3 attached:1 volume:3 cerebral:1 interpretation:3 gibbs:4 tuning:1 dbn:2 longer:2 cortex:3 disentangle:2 posterior:1 own:1 recent:2 irrelevant:2 scenario:2 binary:2 success:2 life:1 muller:1 minimum:1 greater:1 additional:1 converge:1 signal:1 multiple:4 full:1 thalamus:1 segmented:3 retrieval:1 equally:1 coded:1 controlled:3 prediction:7 scalable:3 oliva:1 multilayer:1 vision:7 metric:3 expectation:1 essentially:1 arxiv:2 iteration:26 normalization:1 sometimes:1 achieved:1 proposal:3 background:18 addition:2 fine:3 separately:2 want:1 entangled:1 source:1 unlike:1 posse:1 induced:1 tend:2 contrary:1 flow:1 effectiveness:3 call:2 extracting:1 presence:1 intermediate:1 split:1 easy:3 enough:2 feedforward:5 iterate:1 affect:1 fit:1 gave:1 bengio:3 architecture:9 identified:1 competing:1 imperfect:1 idea:1 rifai:1 knowing:1 whether:3 six:1 handled:1 hallucinating:1 effort:1 akin:1 song:2 suffer:1 as4:1 constitute:1 deep:6 enumerate:1 useful:1 clear:2 tune:1 clutter:1 processed:1 sohn:1 simplest:1 http:1 overlay:1 problematic:1 deteriorates:1 neuroscience:5 correctly:3 per:1 affected:2 group:10 threshold:2 demonstrating:1 achieving:1 clean:5 leibo:1 utilize:1 sotelo:1 fize:1 volitional:1 tsotsos:1 sum:2 beijing:3 denosing:1 run:4 mystery:1 year:1 powerful:1 you:1 reasonable:2 patch:3 scaling:1 layer:7 overcame:1 followed:1 activity:1 occur:1 constraint:1 throwing:1 scene:7 generates:2 speed:1 performing:1 department:3 structured:1 palm:1 combination:1 poor:1 belonging:3 across:1 slightly:1 shallow:1 b:1 making:2 alike:1 intuitively:1 restricted:2 fulfilling:1 iccv:1 turn:2 discus:1 fail:1 mechanism:2 needed:1 permit:1 apply:1 hierarchical:1 appearing:1 altogether:1 gate:4 existence:1 reichert:1 top:22 include:1 sommer:1 zeiler:1 lock:1 especially:2 prof:1 seeking:1 objective:1 realized:1 occurs:1 strategy:5 costly:2 traditional:1 win:1 attentional:5 mapped:1 link:1 sensible:1 degrade:1 argue:2 dream:1 code:1 besides:1 modeled:2 pointwise:1 manzagol:1 difficult:6 mostly:4 stated:1 negative:2 rise:1 huth:1 design:1 zt:1 boltzmann:5 perform:1 gated:3 gallant:1 teh:1 benchmark:4 displayed:1 emulating:2 incorporated:1 hinton:4 arbitrary:1 intensity:4 compositionality:1 pair:3 optimized:1 connection:4 hallucinated:1 imagenet:2 philosophical:1 coherent:1 learned:1 nip:3 able:2 bar:1 beyond:2 below:1 pattern:1 nips2014:1 sparsity:1 challenge:1 built:1 memory:1 explanation:1 belief:2 power:1 hot:1 difficulty:2 force:1 natural:4 improve:1 github:1 brief:1 eye:3 auto:1 prior:3 understanding:3 multiplication:1 loss:2 expect:1 mixed:1 interesting:3 generation:1 validation:1 foundation:2 sufficient:2 principle:1 classifying:1 row:7 prone:1 surprisingly:1 last:2 bias:58 allow:3 side:1 perceptron:1 warp:1 template:1 differentiating:1 fifth:1 sparse:1 feedback:14 default:1 world:2 author:2 forward:1 refinement:1 collection:1 qualitatively:1 made:1 far:3 reconstructed:3 emphasize:1 conclusively:1 keep:1 dealing:2 confirm:2 global:2 conclude:1 discriminative:5 fergus:1 alternatively:1 search:6 iterative:7 latent:1 shareable:1 modularity:1 table:2 disambiguate:1 nature:5 learn:2 robust:3 composing:1 confirmation:1 ignoring:1 transfer:1 defy:1 improving:1 complex:5 necessarily:1 constructing:1 meanwhile:1 icann:1 linearly:1 whole:3 noise:25 denoise:3 allowed:1 downwards:1 deployed:2 ehinger:1 fails:1 inferring:1 explicit:1 candidate:5 crude:1 third:2 grained:2 tang:1 down:21 removing:1 navigate:1 gating:2 sensing:1 nyu:2 explored:1 svm:2 evidence:3 glorot:1 essential:3 incorporating:1 mnist:25 indiscriminately:1 false:5 adding:1 confuse:1 locality:1 entropy:4 tantalizing:1 visual:12 prevents:1 highlighting:1 tracking:1 partially:1 chanceaux:1 springer:2 truth:9 loses:1 acm:2 nair:1 modulate:1 sorted:2 goal:2 ann:8 towards:3 shared:1 included:1 denoising:14 called:1 total:1 pas:6 invariance:1 indicating:1 select:4 wq:1 people:1 modulated:2 arises:1 incorporate:1 evaluate:1 phenomenon:1 correlated:1 |
4,714 | 5,269 | Expectation Backpropagation: Parameter-Free
Training of Multilayer Neural Networks with
Continuous or Discrete Weights
Daniel Soudry1 , Itay Hubara2 , Ron Meir2
(1) Department of Statistics, Columbia University
(2) Department of Electrical Engineering, Technion, Israel Institute of Technology
[email protected],[email protected],[email protected]
Abstract
Multilayer Neural Networks (MNNs) are commonly trained using gradient
descent-based methods, such as BackPropagation (BP). Inference in probabilistic
graphical models is often done using variational Bayes methods, such as Expectation Propagation (EP). We show how an EP based approach can also be used
to train deterministic MNNs. Specifically, we approximate the posterior of the
weights given the data using a ?mean-field? factorized distribution, in an online
setting. Using online EP and the central limit theorem we find an analytical approximation to the Bayes update of this posterior, as well as the resulting Bayes
estimates of the weights and outputs.
Despite a different origin, the resulting algorithm, Expectation BackPropagation
(EBP), is very similar to BP in form and efficiency. However, it has several additional advantages: (1) Training is parameter-free, given initial conditions (prior)
and the MNN architecture. This is useful for large-scale problems, where parameter tuning is a major challenge. (2) The weights can be restricted to have discrete
values. This is especially useful for implementing trained MNNs in precision limited hardware chips, thus improving their speed and energy efficiency by several
orders of magnitude.
We test the EBP algorithm numerically in eight binary text classification tasks.
In all tasks, EBP outperforms: (1) standard BP with the optimal constant learning
rate (2) previously reported state of the art. Interestingly, EBP-trained MNNs with
binary weights usually perform better than MNNs with continuous (real) weights
- if we average the MNN output using the inferred posterior.
1
Introduction
Recently, Multilayer1 Neural Networks (MNNs) with deep architecture have achieved state-of-theart performance in various supervised learning tasks [11, 14, 8]. Such networks are often massive
and require large computational and energetic resources. A dense, fast and energetically efficient
hardware implementation of trained MNNs could be built if the weights were restricted to discrete
values. For example, with binary weights, the chip in [13] can perform 1012 operations per second
with 1mW power efficiency. Such performances will enable the integration of massive MNNs into
small and low-power electronic devices.
Traditionally, MNNs are trained by minimizing some error function using BackPropagation (BP) or
related gradient descent methods [15]. However, such an approach cannot be directly applied if the
weights are restricted to binary values. Moreover, crude discretization of the weights is usually quite
1
i.e., having more than a single layer of adjustable weights.
1
destructive [20]. Other methods have been suggested in the 90?s (e.g., [23, 3, 18]), but it is not clear
whether these approaches are scalable.
The most efficient methods developed for training Single-layer2 Neural Networks (SNN) with binary
weights use approximate Bayesian inference, either implicitly [6, 1] or explicitly [24, 22]. In theory,
given a prior, the Bayes estimate of the weights can be found from their posterior given the data.
However, storing or updating the full posterior is usually intractable. To circumvent this problem,
these previous works used a factorized ?mean-field? form the posterior of the weights given the data.
As explained in [22], this was done using a special case of the widely applicable Expectation Propagation (EP) algorithm [19] - with an additional approximation that the fan-in of all neurons is large,
so their inputs are approximately Gaussian. Thus, given an error function, one can analytically
obtain the Bayes estimate of the weights or the outputs, using the factorized approximation of the
posterior. However, to the best of our knowledge, it is still unknown whether such an approach could
be generalized to MNNs, which are more relevant for practical applications.
In this work we derive such generalization, using similar approximations (section 3). The end result
is the Expectation BackPropagation (EBP, section 4) algorithm for online training of MNNs where
the weight values can be either continuous (i.e., real numbers) or discrete (e.g., ?1 binary). Notably,
the training is parameter-free (with no learning rate), and insensitive to the magnitude of the input.
This algorithm is very similar to BP. Like BP, it is very efficient in each update, having a linear
computational complexity in the number of weights.
We test the EBP algorithm (section 5) on various supervised learning tasks: eight high dimensional
tasks of classifying text into one of two semantic classes, and one low dimensional medical discrimination task. Using MNNs with two or three weight layers, EBP outperforms both standard BP, as
well as the previously reported state of the art for these tasks [7]. Interestingly, the best performance
of EBP is usually achieved using the Bayes estimate of the output of MNNs with binary weights.
This estimate can be calculated analytically, or by averaging the output of several such MNNs, with
weights sampled from the inferred posterior.
2
Preliminaries
General Notation A non-capital boldfaced letter x denotes a column vector with components xi ,
a boldfaced capital letter X denotes a matrix with components Xij . Also, if indexed, the components of xl are denoted xi,l and those of Xl are denoted Xij,l . We denote by P (x) the probability distribution (in the discrete case)
or density (in the? continuous case) of a random variable X,
?
P (x|y) = P (x, y) /P (y),hxi = xP (x) dx, hx|yi = xP (x|y) dx, Cov (x, y) = hxyi?hxi hyi
and Var (x) = Cov (x, x). Integration is exchanged with summation in the discrete case. For any
condition A, we make use of I {A}, the indicator function (i.e., I {A} = 1 if A holds, and zero
otherwise), and ?ij = I {i = j}, Kronecker?s delta function. If x ? N (?, ?) then it is Gaussian
with mean ? and covariance matrix ?, and we denote
? x its density by N (x|?, ?). Furthermore, we
use the cumulative distribution function ? (x) = ?? N (u|0, 1) du.
Model We consider a general feedforward Multilayer Neural Network (MNN) with connections
between adjacent layers (Fig. 2.1). For analytical simplicity, we focus here on deterministic binary
(?1) neurons. However, the framework can be straightforwardly extended to other types of neurons
(deterministic or stochastic). The MNN has L layers, where Vl is the width of the l-th layer, and
L
W = {Wl }l=1 is the collection of Vl ? Vl?1 synaptic weight matrices which connect neuronal
L
L?1
layers sequentially. The outputs of the layers are {vl }l=0 , where v0 is the input layer, {vl }l=1 are
the hidden layers and vL is the output layer. In each layer,
vl = sign (Wl vl?1 )
(2.1)
where each sign ?activation function? (a neuronal layer) operates component-wise (i.e., ?i :
(sign (x))i = sign (xi )). The output of the network is therefore
vL = g (v0 , W) = sign (WL sign (WL?1 sign (? ? ? W1 v0 ))) .
2
i.e., having only a single layer of adjustable weights.
2
(2.2)
We assume that the weights are constrained to some set
S, with the specific restrictions on each weight denoted
by Sij,l , so Wij,l ? Sij,l and W ? S. If Sij,l = {0},
then we say that Wij,l is ?disconnected?. For simplicity, we assume that in each layer the ?fan-in? Kl =
|{j|Sij,l 6= {0}}| is constant for all i. Biases can be optionally included in the standard way, by adding a constant output v0,l = 1 to each layer.
Task We examine a supervised classification learning Figure 2.1: Our MNN model (Eq. 2.2).
task, in Bayesian framework. We are given a fixed set of
N
sequentially labeled data pairs DN = x(n) , y(n) n=1
(so D0 = ?), where each x(n) ? RV0 is a data point, and
V
each y(n) is a label taken from a binary set Y ? {?1, +1} L . For brevity, we will sometimes
suppress the sample index n, where it is clear from the context. As common for supervised learning
with MNNs, we assume that for all n the relation x(n) ? y(n) can be represented by a MNN with
known architecture (the ?hypothesis class?), and unknown weights W ? S. This is a reasonable
assumption since a MNN can approximate any deterministic function, given that it has sufficient
number of neurons
[12] (if L ? 2). Specifically, there exists some W ? ? S, so that y(n) =
(n)
?
f x , W (see Eq. 2.2). Our goals are: (1) estimate the most probable W ? for this MNN, (2)
estimate the most probable y given some (possibly unseen) x.
3
Theory
In this section we explain how a specific learning algorithm for MNNs (described in section 4) arises
from approximate (mean-field) Bayesian inference, used in this context (described in section 2).
3.1
Online Bayesian learning in MNNs
We approach this task within a Bayesian framework, where we assume some prior distribution on the
weights - P (W|D0 ). Our aim is to find P (W|DN ), the posterior probability for the configuration
of the weights W, given the data. With this posterior, one can select the most probable weight
configuration - the Maximum A Posteriori (MAP) weight estimate
W ? = argmaxW?S P (W|DN ) ,
(3.1)
?
minimizing the expected zero-one loss over the weights (I {W 6= W}). This weight estimate can
be implemented in a single MNN, which can provide an estimate of the label y for (possibly unseen)
data points x through y =g (x, W ? ). Alternatively, one can aim to minimize the expected loss over
the output - as more commonly done in the MNN literature. For example, if the aim is to reduce
classification error then one should use the MAP output estimate
X
y? = argmaxy?Y
I {g (x, W) = y} P (W|DN ) ,
(3.2)
W
?
which minimizes the zero-one loss (I {y 6= g (x, W)}) over the outputs. The resulting estimator
does not generally have the form of a MNN (i.e., y =g (x, W) with W ? S), but can be approximated by averaging the output over many such MNNs with W values sampled from the posterior.
Note that averaging the output of several MNNs is a common method to improve performance.
We aim to find the posterior P (W|DN ) in an online setting, where samples arrive sequentially.
After the n-th sample is received, the posterior is updated according to Bayes rule:
P (W|Dn ) ? P y(n) |x(n) , W P (W|Dn?1 ) ,
(3.3)
for n = 1, . . . , N . Note that the MNN is deterministic, so the likelihood (per data point) has the
following simple form3
n
o
P y(n) |x(n) , W = I g x(n) , W = y(n) .
(3.4)
3
MNN with stochastic activation functions will have a ?smoothed out? version of this.
3
Therefore, the Bayes update in Eq. 3.3 simply makes sure that P (W|Dn ) = 0 in any ?illegal? configuration (i.e., any W 0 such that g x(k) , W 0 6= y(k) ) for some 1 ? k ? n. In other words, the
posterior is equal to the prior, restricted to the ?legal? weight domain, and re-normalized appropriately. Unfortunately, this update is generally intractable for large networks, mainly because we need
to store and update an exponential number of values for P (W|Dn ). Therefore, some approximation
is required.
3.2
Approximation 1: mean-field
In order to reduce computational complexity, instead of storing P (W|Dn ), we will store its factorized (?mean-field?) approximation P? (W|Dn ), for which
Y
P? (W|Dn ) =
P? (Wij,l |Dn ) ,
(3.5)
i,j,l
where each factor must be normalized. Notably, it is easy to find the MAP estimate of the weights
(Eq. 3.1) under this factorized approximation ?i, j, l
?
Wij,l
= argmaxWij,l ?Sij,l P? (Wij,l |DN ) .
(3.6)
The factors P? (Wij,l |Dn ) can be found using a standard variational approach [5, 24]. For each n,
we first perform the Bayes update in Eq. 3.3 with P? (W|Dn?1 ) instead of P (W|Dn?1 ). Then, we
project the resulting posterior onto the family of distributions factorized as in Eq. 3.5, by minimizing the reverse Kullback-Leibler divergence (similarly to EP [19, 22]). A straightforward calculation
shows that the optimal factor is just a marginal of the posterior (appendix A, available in the supplementary material). Performing this marginalization on the Bayes update and re-arranging terms, we
obtain a Bayes-like update to the marginals ?i, j, l
P? (Wij,l |Dn ) ? P? y(n) |x(n) , Wij,l , Dn?1 P? (Wij,l |Dn?1 ) ,
(3.7)
where
P? y(n) |x(n) , Wij,l , Dn?1 =
X
P y(n) |x(n) , W 0
0 =W
W 0 :Wij,l
ij,l
Y
0
P? Wkr,m
|Dn?1 (3.8)
{k,r,m}6={i,j,l}
is the marginal likelihood. Thus we can directly update the factor P? (Wij,l |Dn ) in a single step.
However, the last equation is still problematic, since it contains a generally intractable summation
over an exponential number of values, and therefore requires simplification. For simplicity, from
now on we replace any P? with P , in a slight abuse of notation (keeping in mind that the distributions
are approximated).
3.3
Simplifying the marginal likelihood
In order to be able to use the
update rule in Eq. 3.7, we must first calculate the marginal likelihood
P y(n) |x(n) , Wij,l , Dn?1 using Eq. 3.8. For brevity, we suppress the index n and the dependence
on Dn?1 and x, obtaining
Y
X
0
P (y|W 0 )
P Wkr,m
,
(3.9)
P (y|Wij,l ) =
0 =W
W 0 :Wij,l
ij,l
{k,r,m}6={i,j,l}
where we recall that P (y|W 0 ) is simply an indicator function (Eq. 3.4). Since, by assumption,
P (y|W 0 ) arises from a feed-forward MNN with input v0 = x and output vL = y, we can perform
the summations in Eq. 3.9 in a more convenient way - layer by layer. To do this, we define
?
? ?
?
Vm?1
Vm
m?1
?
?VY
XY
X
0
0
?I vk,m
? (3.10)
P (vm |vm?1 ) =
vr,m?1 Wkr,m
>0
P Wkr,m
?
?
0
Wm k=1
r=1
r=1
and P (vl |vl?1 , Wij,l ), which is defined identically to P (vl |vl?1 ), except that the summation is
0
performed over all configurations in which Wij,l is fixed (i.e., Wl0 : Wij,l
= Wij,l ) and we set
4
P (Wij,l ) = 1. Now we can write recursively P (v1 ) = P (v1 |v0 = x)
X
?m ? {2, .., l ? 1} : P (vm ) =
P (vm |vm?1 ) P (vm?1 )
(3.11)
vm?1
P (vl |Wij,l ) =
X
P (vl |vl?1 , Wij,l ) P (vl?1 )
(3.12)
vl?1
?m ? {l + 1, l + 2, .., L} : P (vm |Wij,l ) =
X
P (vm |vm?1 ) P (vm?1 |Wij,l )
(3.13)
vm?1
Thus we obtain the result of Eq. 3.9, through P (y|Wij,l ) = P (vL = y|Wij,l ). However, this
computation is still generally intractable, since all of the above summations (Eqs. 3.10-3.13) are still
over an exponential number of values. Therefore, we need to make one additional approximation.
3.4
Approximation 2: large fan-in
Next we simplify the above summations (Eqs. 3.10-3.13) assuming that the neuronal fan-in is
?large?. We keep in mind that i, j and l are the specific indices of the fixed weight Wij,l . All
the other weights beside Wij,l can be treated as independent random variables, due to the mean field
approximation (Eq. 3.5). Therefore, in the limit of a infinite neuronal fan-in (?m : Km ? ?) we
can use the Central Limit Theorem (CLT) and say that the normalized input to each neuronal layer,
is distributed according to a Gaussian distribution
p
?m : um = Wm vm?1 / Km ? N (?m , ?m ) .
(3.14)
Since Km is actually finite, this would be only an approximation - though a quite common and
effective one (e.g., [22]). Using the approximation in Eq. 3.14 together with vm = sign (um ) (Eq.
2.1) we can calculate (appendix B) the distribution of um and vm sequentially for all the layers
m ? {1, . . . , L}, for any given value of v0 and Wij,l . These effectively simplify the summations in
3.10-3.13 using Gaussian integrals (appendix B).
At the end of this ?forward pass? we will be able to find P (y|Wij,l ) = P (vL = y|Wij,l ) , ?i, j, l.
This takes a polynomial number of steps (appendix B.3), instead of a direct calculation through
Eqs. 3.11-3.13, which is exponentially hard. Using P (y|Wij,l ) and Eq. 3.7 we can now update the
distribution of P (Wij,l ). This immediately gives the Bayes estimate of the weights (Eq. 3.6) and
outputs (Eq. 3.2).
As we note in appendix B.3, the computational complexity of the forward pass is significantly lower
if ?m is diagonal. This is true exactly only in special cases. For example, this is true if all hidden
neurons have a fan-out of one - such as in a 2-layer network with a single output. However, in order
to reduce the computational complexity in cases that ?m is not diagonal, we will approximate the
distribution of um with its factorized (?mean-field?) version. Recall that the optimal factor is the
marginal of the distribution (appendix A). Therefore, we can now find P (y|Wij,l ) easily (appendix
2
B.1), as all the off-diagonal components in ?m are zero, so ?kk0 ,m = ?k,m
?kk0 .
A direct calculation of P (vL = y|Wij,l ) for every i, j, l would be computationally wasteful, since
we will repeat similar calculations many times. In order to improve the algorithm?s efficiency,
we again exploit the fact that Kl is large. We approximate ln P (vL = y|Wij,l ) using a Taylor
?1/2
expansion of Wij,l around its mean, hWij,l i, to first order in Kl
. The first order terms in this
expansion can be calculated using backward propagation of derivative terms
?k,m = ? ln P (vL = y) /??k,m ,
(3.15)
similarly to the BP algorithm (appendix C). Thus, after a forward pass for m = 1, . . . , L, and a
backward pass for l = L, . . . , 1, we obtain P (vL = y|Wij,l ) for all Wij,l and update P (Wij,l ).
4
The Expectation Backpropagation Algorithm
Using our results we can efficiently update the posterior distribution P (Wij,l |Dn ) for all the weights
with O (|W|) operations, according to Eqs. 3.7. Next, we summarize the resulting general algorithm
- the Expectation BackPropgation (EBP) algorithm. In appendix D, we exemplify how to apply the
5
algorithm in the special cases of MNNs with binary, ternary or real (continuous) weights. Similarly
to the original BP algorithm (see review in [16]), given input x and desired output y, first we perform
a forward pass to calculate the mean output hvl i for each layer. Then we perform a backward pass
to update P (Wij,l |Dn ) for all the weights.
Forward pass In this pass we perform the forward calculation of probabilities, as in Eq. 3.11.
Recall that hWkr,m i is the mean of the posterior distribution P (Wkr,m |Dn ). We first initialize the
MNN input hvk,0 i = xk for all k and calculate recursively the following quantities for m = 1, . . . , L
and all k
?k,m = ?
2
?k,m
=
Vm?1
1 X
hWkr,m i hvr,m?1 i ; hvk,m i = 2? (?k,m /?k,m ) ? 1 .
Km r=1
Vm?1
1 X
2
2
2
2
Wkr,m ?m,1 hvr,m?1 i ? 1 + 1 ? hWkr,m i hvr,m?1 i ,
Km r=1
(4.1)
(4.2)
where ?m and ? 2m are, respectively, the mean and variance of um , the input of layer m (Eq. 3.14),
and hvm i is the resulting mean of the output of layer m.
Backward pass In this pass we perform the Bayes update of the posterior (Eq. 3.7) using a Taylor
expansion. Recall Eq. 3.15. We first initialize4
2
N 0|?i,L , ?i,L
?i,L = yi
.
(4.3)
? (yi ?i,L /?i,L )
for all i. Then, for l = L, . . . , 1 and ?i, j we calculate
V
?i,l?1
=
m
X
2
2
? N 0|?i,l?1 , ?i,l?1
hWji,l i ?j,l .
Kl
j=1
(4.4)
ln P (Wij,l |Dn )
=
1
ln P (Wij,l |Dn?1 ) + ? Wij,l ?i,l hvj,l?1 i + C ,
Kl
(4.5)
where C is some unimportant constant (which does not depend on Wij,l ).
Output Using the posterior distribution, the optimal configuration can be immediately found
through the MAP weights estimate (Eq. 3.6) ?i, j, l
?
Wij,l
= argmaxWij,l ?Sij,l ln P (Wij,l |Dn ) .
(4.6)
The output of a MNN implementing these weights would be g (x, W ? ) (see Eq. 2.2). We define this
to be the ?deterministic? EBP output (EBP-D).
Additionally, the MAP output (Eq. 3.2) can be calculated directly
"
#
X 1 + hvk,L i yk
?
y = argmaxy?Y ln P (vL = y) = argmaxy?Y
ln
1 ? hvk,L i
(4.7)
k
using hvk,L i from Eq. 4.1, or as an ensemble average over the outputs of all possible MNN with the
weights Wij,l being sampled from the estimated posterior P (Wij,l |Dn ). We define the output in Eq.
4.7 to be the Probabilistic EBP output (EBP-P). Note that in the case of a single output Y = {?1, 1},
so this output simplifies to y = sign (hvk,L i).
4
Due to numerical inaccuracy, calculating ?i,L using Eq. 4.3 can generate nonsensical values (??, NaN)
if |?i,L /?i,L | becomes to large. If this happens, we use instead the asymptotic form in that limit
?i,L = ?
?i,L
?
I {yi ?i,L < 0}
2
?i,L
KL
6
5
Numerical Experiments
We use several high dimensional text datasets to assess the performance of the EBP algorithm in
a supervised binary classification task. The datasets (taken from [7]) contain eight binary tasks
from four datasets: ?Amazon (sentiment)?, ?20 Newsgroups?, ?Reuters? and ?Spam or Ham?. Data
specification (N =#examples and M =#features) and results (for each algorithm) are described in
Table 1. More details on the data including data extraction and labeling can be found in [7].
We test the performance of EBP on MNNs with a 2-layer architecture of M ? 120 ? 1, and
bias weights. We examine two special cases: (1) MNNs with real weights (2) MNNs with binary
weights (and real bias). Recall the motivation for the latter (section 1) is that they can be efficiently
implemented in hardware (real bias has negligible costs). Recall also that for each type of MNN, the
algorithm gives two outputs - EBP-D (deterministic) and EBP-P (probabilistic), as explained near
Eqs. 4.6-4.7.
To evaluate our results we compare EBP to: (1) the AROW algorithm, which reports state-of-the-art
results on the tested datasets [7] (2) the traditional Backpropagation (BP) algorithm, used to train an
M ? 120 ? 1 MNN with real weights. In the latter case, we used both Cross Entropy (CE) and
Mean Square Error (MSE) as loss functions. On each dataset we report the results of BP with the
loss function which achieved the minimal error. We use a simple parameter scan for both AROW
(regularization parameter) and the traditional BP (learning rate parameter). Only the results with
the optimal parameters (i.e., achieving best results) are reported in Table 1. The optimal parameters
found were never at the edges of the scanned field. Lastly, to demonstrate the destructive effect of
naive quantization, we also report the performance of the BP-trained MNNs, after all the weights
(except the bias) were clipped using a sign function.
During training the datasets were repeatedly presented in three epochs (in all algorithms, additional
epochs did not reduce test error). On each epoch the examples were shuffled at random order for BP
and EBP (AROW determines its own order). The test results are calculated after each epoch using
8-fold cross-validation, similarly to [7]. Empirically, EBP running time is similar to BP with real
weights, and twice slower with binary weights. For additional implementation details, see appendix
E.1. The code is available on the author?s website.
The minimal values achieved over all three epochs are summarized in Table 1. As can be seen, in all
datasets EBP-P performs better then AROW, which performs better then BP. Also, EBP-P usually
perfroms better with binary weights. In appendix E.2 we show that this ranking remains true even if
the fan-in is small (in contrast to our assumptions), or if a deeper 3-layer architecture is used.
Dataset
#Examples
#Features
Real EBP-D
Real EBP-P
Binary EBP-D
Binary EBP-P
AROW
BP
Clipped BP
Reuters news I6
Reuters news I8
Spam or ham d0
Spam or ham d1
20News group comp vs HW
20News group elec vs med
Amazon Book reviews
Amazon DVD reviews
2000
2000
2500
2500
1943
1971
3880
3880
11463
12167
26580
27523
29409
38699
221972
238739
14.5%
15.65%
1.28%
1.0%
5.06%
3.36%
2.14%
2.06%
11.35%
15.25%
1.11%
0.96%
4.96%
3.15%
2.09%
2.14%
21.7%
23.15%
7.93%
3.85%
7.54%
6.0%
2.45%
5.72%
9.95%
16.4%
0.76%
0.96%
4.44%
2.08%
2.01%
2.27%
11.72%
15.27%
1.12%
1.4%
5.79%
2.74%
2.24%
2.63%
13.3%
18.2%
1.32%
1.36%
7.02%
3.96%
2.96%
2.94%
26.15%
26.4%
7.97%
7.33%
13.07%
14.23%
3.81%
5.15%
Table 1: Data specification, and test errors (with 8-fold cross-validation). Best results are boldfaced.
6
Discussion
Motivated by the recent success of MNNs, we developed the Expectation BackPropagation algorithm (EBP - see section 4) for approximate Bayesian inference of the synaptic weights of a MNN.
Given a supervised classification task with labeled training data and a prior over the weights, this
deterministic online algorithm can be used to train deterministic MNNs (Eq. 2.2) without the need
to tune learning parameters (e.g., learning rate). Furthermore, each synaptic weight can be restricted
to some set - which can be either finite (e.g., binary numbers) or infinite (e.g., real numbers). This
opens the possibility of implementing trained MNNs in power-efficient hardware devices requiring
limited parameter precision.
7
This algorithm is essentially an analytic approximation to the intractable Bayes calculation of the
posterior distribution of the weights after the arrival of a new data point. To simplify the intractable
Bayes update rule we use several approximations. First, we approximate the posterior using a product of its marginals - a ?mean field? approximation. Second, we assume the neuronal layers have a
large fan-in, so we can approximate them as Gaussian. After these two approximations each Bayes
update can be tractably calculated in polynomial time in the size of the MNN. However, in order to
further improve computational complexity (to O (|W|) in each step, like BP), we make two additional approximations. First, we use the large fan-in to perform a first order expansion. Second, we
optionally5 perform a second ?mean field? approximation - to the distribution of the neuronal inputs.
Finally, after we obtain the approximated posterior using the algorithm, the Bayes estimates of the
most probable weights and the outputs are found analytically.
Previous approaches to obtain these Bayes estimates were too limited for our purposes. The Monte
Carlo approach [21] achieves state-of-the-art performance for small MNNs [26], but does not scale
well [25]. The Laplace approximation [17] and variational Bayes [10, 2, 9] based methods require
real-value weights, tuning of the learning rate parameter, and stochastic neurons (to ?smooth? the
likelihood). Previous EP [24, 22] and message passing [6, 1] (a special case of EP[5]) based methods
were derived only for SNNs.
In contrast, the EBP allows parameter free and scalable training of various types of MNNs (deterministic or stochastic) with discrete (e.g., binary) or continuous weights. In appendix F, we see that
for continuous weights EBP is almost identical to standard BP with a specific choice of activation
function s (x) = 2? (x) ? 1, CE loss and learning rate ? = 1. The only difference is that the input
is normalized by its standard deviation (Eq. 4.1, right), which depends on the weights and inputs
(Eq. 4.2). This re-scaling makes the learning algorithm invariant to the amplitude changes in the
neuronal input. This results from the same invariance of the sign activation functions. Note that in
standard BP algorithm the performance is directly affected by the amplitude of the input, so it is a
recommended practice to re-scale it in pre-processing [16].
We numerically evaluated the algorithm on binary classification tasks using MNNs with two or three
synaptic layers. In all data sets and MNNs EBP performs better than standard BP with the optimal
constant learning rate, and even achieves state-of-the-art results in comparison to [7]. Surprisingly,
EBP usually performs best when it is used to train binary MNNs. As suggested by a reviewer, this
could be related to the type of problems examined here. In text classification tasks have large sparse
input spaces (bag of words), and presence/absence of features (words) is more important than their
real values (frequencies). Therefore, (distributions over) binary weights and a threshold activation
function may work well.
In order to get such a good performance in binary MNNs, one must average over the output the
inferred (approximate) posterior of the weights. The EBP-P output of the algorithm calculates this
average analytically. In hardware this output could be realizable by averaging the output of several
binary MNNs, by sampling weights from P (Wij,l |Dn ). This can be done efficiently (appendix G).
Our numerical testing mainly focused on high-dimensional text classification tasks, where shallow architectures seem to work quite well. In other domains, such as vision [14] and speech [8],
deep architectures achieve state-of-the-art performance. Such deep MNNs usually require considerable fine-tuning and additional ?tricks? such as unsupervised pre-training [8], weight sharing [14]
or momentum6 . Integrating such methods into EBP and using it to train deep MNNs is a promising direction for future work. Another important generalization of the algorithm, which is rather
straightforward, is to use activation functions other than sign (?). This is particularly important for
the last layer - where a linear activation function would be useful for regression tasks, and joint
activation functions7 would be useful for multi-class tasks[4].
Acknowledgments The authors are grateful to C. Baldassi, A. Braunstein and R. Zecchina for
helpful discussions and to A. Hallak, T. Knafo and U. S?mb?l for reviewing parts of this manuscript.
The research was partially funded by the Technion V.P.R. fund, by the Intel Collaborative Research
Institute for Computational Intelligence (ICRI-CI), and by the Gruss Lipper Charitable Foundation.
5
This approximation is not required if all neurons in the MNN have a fan-out of one.
Which departs from the online framework considered here, since it requires two samples in each update.
7
i.e., activation functions for which (f (x))i 6= f (xi ), such as softmax or argmax.
6
8
References
[1] C Baldassi, A Braunstein, N Brunel, and R Zecchina. Efficient supervised learning in networks with
binary synapses. PNAS, 104(26):11079?84, 2007.
[2] D Barber and C M Bishop. Ensemble learning for multi-layer networks. In Advances in Neural Information Processing Systems, pages 395?401, 1998.
[3] R Battiti and G Tecchiolli. Training neural nets with the reactive tabu search. IEEE transactions on neural
networks, 6(5):1185?200, 1995.
[4] C M Bishop. Neural networks for pattern recognition. 1995.
[5] C M Bishop. Pattern recognition and machine learning. Springer, Singapore, 2006.
[6] A Braunstein and R Zecchina. Learning by message passing in networks of discrete synapses. Physical
review letters, 96(3), 2006.
[7] K Crammer, A Kulesza, and M Dredze. Adaptive regularization of weight vectors. Machine Learning,
91(2):155?187, March 2013.
[8] G E Dahl, D Yu, L Deng, and A Acero. Context-Dependent Pre-Trained Deep Neural Networks for
Large-Vocabulary Speech Recognition. Audio, Speech, and Language Processing, 20(1):30?42, 2012.
[9] A Graves. Practical variational inference for neural networks. Advances in Neural Information Processing
Systems, pages 1?9, 2011.
[10] G E Hinton and D Van Camp. Keeping the neural networks simple by minimizing the description length
of the weights. In COLT ?93, 1993.
[11] G E Hinton, L Deng, D Yu, G E Dahl, A R Mohamed, N Jaitly, A Senior, V Vanhoucke, P Nguyen, T N
Sainath, and B Kingsbury. Deep neural networks for acoustic modeling in speech recognition: The shared
views of four research groups. Signal Processing Magazine, IEEE, 29(6):82?97, 2012.
[12] K Hornik. Approximation capabilities of multilayer feedforward networks.
4(1989):251?257, 1991.
Neural networks,
[13] R Karakiewicz, R Genov, and G Cauwenberghs. 1.1 TMACS/mW Fine-Grained Stochastic Resonant
Charge-Recycling Array Processor. IEEE Sensors Journal, 12(4):785?792, 2012.
[14] A Krizhevsky, I Sutskever, and G E Hinton. Imagenet classification with deep convolutional neural
networks. In NIPS, 2012.
[15] Y LeCun and L Bottou. Gradient-based learning applied to document recognition. Proceedings of the
IEEE, 86(11):2278?2324, 1998.
[16] Y LeCun, L Bottou, G B Orr, and K R M?ller. Efficient Backprop. In G Montavon, G B Orr, and K-R
M?ller, editors, Neural networks: Tricks of the Trade. Springer, Heidelberg, 2nd edition, 2012.
[17] D J C MacKay. A practical Bayesian framework for backpropagation networks. Neural computation,
472(1):448?472, 1992.
[18] E Mayoraz and F Aviolat. Constructive training methods for feedforward neural networks with binary
weights. International journal of neural systems, 7(2):149?66, 1996.
[19] T P Minka. Expectation Propagation for Approximate Bayesian Inference. NIPS, pages 362?369, 2001.
[20] P Moerland and E Fiesler. Neural Network Adaptations to Hardware Implementations. In Handbook of
neural computation. Oxford University Press, New York, 1997.
[21] R M Neal. Bayesian learning for neural networks. PhD thesis, University of Toronto, 1995.
[22] F Ribeiro and M Opper. Expectation propagation with factorizing distributions: a Gaussian approximation
and performance results for simple models. Neural computation, 23(4):1047?69, April 2011.
[23] D Saad and E Marom. Training Feed Forward Nets with Binary Weights Via a Modified CHIR Algorithm.
Complex Systems, 4:573?586, 1990.
[24] S A Solla and O Winther. Optimal perceptron learning: an online Bayesian approach. In On-Line Learning
in Neural Networks. Cambridge University Press, Cambridge, 1998.
[25] N Srivastava and G E Hinton. Dropout: A Simple Way to Prevent Neural Networks from Overfitting.
Journal of Machine Learning, 15:1929?1958, 2014.
[26] H Y Xiong, Y Barash, and B J Frey. Bayesian prediction of tissue-regulated splicing using RNA sequence
and cellular context. Bioinformatics (Oxford, England), 27(18):2554?62, October 2011.
9
| 5269 |@word version:2 polynomial:2 nd:1 nonsensical:1 open:1 km:5 covariance:1 simplifying:1 recursively:2 initial:1 configuration:5 contains:1 daniel:2 document:1 interestingly:2 outperforms:2 com:2 discretization:1 activation:9 gmail:2 dx:2 must:3 numerical:3 analytic:1 update:18 fund:1 discrimination:1 v:2 intelligence:1 device:2 website:1 xk:1 ron:1 toronto:1 kingsbury:1 dn:33 direct:2 boldfaced:3 notably:2 expected:2 examine:2 multi:2 snn:1 becomes:1 project:1 moreover:1 notation:2 factorized:7 israel:1 kk0:2 minimizes:1 developed:2 barash:1 hallak:1 zecchina:3 every:1 charge:1 exactly:1 um:5 medical:1 negligible:1 engineering:1 frey:1 limit:4 soudry:1 despite:1 oxford:2 approximately:1 abuse:1 twice:1 hwij:1 examined:1 limited:3 practical:3 acknowledgment:1 lecun:2 testing:1 ternary:1 practice:1 backpropagation:9 itayhubara:1 braunstein:3 significantly:1 illegal:1 convenient:1 word:3 pre:3 integrating:1 get:1 cannot:1 onto:1 acero:1 context:4 baldassi:2 restriction:1 deterministic:10 map:5 reviewer:1 straightforward:2 sainath:1 focused:1 simplicity:3 amazon:3 immediately:2 estimator:1 rule:3 array:1 tabu:1 traditionally:1 arranging:1 updated:1 laplace:1 itay:1 massive:2 magazine:1 hypothesis:1 origin:1 jaitly:1 trick:2 lipper:1 approximated:3 particularly:1 updating:1 recognition:5 labeled:2 ep:7 electrical:1 calculate:5 news:4 chir:1 solla:1 trade:1 yk:1 ham:3 complexity:5 trained:8 depend:1 grateful:1 reviewing:1 efficiency:4 easily:1 joint:1 chip:2 various:3 represented:1 train:5 elec:1 fast:1 effective:1 wkr:6 monte:1 labeling:1 quite:3 widely:1 supplementary:1 say:2 tested:1 otherwise:1 statistic:1 cov:2 unseen:2 online:8 advantage:1 sequence:1 analytical:2 net:2 product:1 mb:1 adaptation:1 relevant:1 achieve:1 description:1 sutskever:1 derive:1 ac:1 ij:3 received:1 eq:35 implemented:2 direction:1 stochastic:5 enable:1 material:1 implementing:3 backprop:1 require:3 hx:1 generalization:2 preliminary:1 probable:4 summation:7 hold:1 around:1 considered:1 major:1 achieves:2 purpose:1 applicable:1 bag:1 label:2 wl:4 sensor:1 gaussian:6 rna:1 aim:4 modified:1 rather:1 derived:1 focus:1 vk:1 likelihood:5 mainly:2 contrast:2 realizable:1 camp:1 posteriori:1 inference:6 helpful:1 dependent:1 vl:26 hidden:2 relation:1 wij:51 layer2:1 classification:9 colt:1 denoted:3 art:6 integration:2 special:5 constrained:1 marginal:5 field:10 equal:1 initialize:1 having:3 extraction:1 never:1 sampling:1 identical:1 yu:2 unsupervised:1 theart:1 future:1 report:3 simplify:3 softmax:1 divergence:1 hvj:1 argmax:1 message:2 possibility:1 argmaxy:3 integral:1 edge:1 xy:1 indexed:1 taylor:2 exchanged:1 re:4 desired:1 hvm:1 minimal:2 column:1 modeling:1 cost:1 deviation:1 technion:3 krizhevsky:1 too:1 reported:3 straightforwardly:1 connect:1 density:2 grus:1 international:1 winther:1 probabilistic:3 vm:19 off:1 fiesler:1 together:1 w1:1 again:1 central:2 thesis:1 possibly:2 book:1 derivative:1 orr:2 summarized:1 explicitly:1 ranking:1 depends:1 performed:1 view:1 cauwenberghs:1 wm:2 bayes:19 capability:1 collaborative:1 minimize:1 il:1 ass:1 square:1 convolutional:1 variance:1 efficiently:3 ensemble:2 bayesian:11 moerland:1 carlo:1 comp:1 processor:1 tissue:1 explain:1 synapsis:2 sharing:1 synaptic:4 energy:1 frequency:1 destructive:2 mohamed:1 minka:1 sampled:3 dataset:2 recall:6 knowledge:1 exemplify:1 amplitude:2 actually:1 manuscript:1 feed:2 marom:1 supervised:7 april:1 done:4 though:1 evaluated:1 furthermore:2 just:1 lastly:1 propagation:5 ebp:33 icri:1 dredze:1 effect:1 normalized:4 true:3 contain:1 requiring:1 analytically:4 regularization:2 shuffled:1 leibler:1 semantic:1 neal:1 adjacent:1 during:1 width:1 generalized:1 demonstrate:1 performs:4 hwji:1 variational:4 wise:1 recently:1 common:3 empirically:1 physical:1 insensitive:1 exponentially:1 slight:1 numerically:2 marginals:2 cambridge:2 tuning:3 similarly:4 backpropgation:1 i6:1 language:1 funded:1 hxi:2 specification:2 v0:7 posterior:25 own:1 recent:1 reverse:1 store:2 binary:27 success:1 battiti:1 yi:4 seen:1 additional:7 deng:2 ller:2 hyi:1 clt:1 recommended:1 signal:1 full:1 pnas:1 d0:3 smooth:1 england:1 calculation:6 cross:3 hvk:6 calculates:1 prediction:1 scalable:2 regression:1 multilayer:4 essentially:1 expectation:10 vision:1 sometimes:1 achieved:4 fine:2 appropriately:1 saad:1 sure:1 med:1 seem:1 ee:1 mw:2 near:1 presence:1 feedforward:3 easy:1 identically:1 newsgroups:1 marginalization:1 architecture:7 reduce:4 simplifies:1 whether:2 motivated:1 energetically:1 energetic:1 sentiment:1 speech:4 passing:2 york:1 repeatedly:1 deep:7 useful:4 generally:4 clear:2 unimportant:1 tune:1 hardware:6 generate:1 xij:2 problematic:1 vy:1 singapore:1 sign:12 delta:1 estimated:1 rmeir:1 per:2 discrete:8 write:1 affected:1 group:3 four:2 threshold:1 achieving:1 capital:2 wasteful:1 prevent:1 ce:2 dahl:2 backward:4 v1:2 letter:3 arrive:1 family:1 reasonable:1 clipped:2 electronic:1 almost:1 resonant:1 splicing:1 appendix:13 scaling:1 dropout:1 layer:29 nan:1 simplification:1 fan:10 fold:2 scanned:1 kronecker:1 bp:22 hvl:1 dvd:1 speed:1 performing:1 department:2 according:3 march:1 disconnected:1 shallow:1 happens:1 argmaxw:1 explained:2 restricted:5 sij:6 invariant:1 taken:2 legal:1 resource:1 equation:1 previously:2 computationally:1 ln:7 remains:1 mind:2 end:2 available:2 operation:2 eight:3 apply:1 xiong:1 slower:1 original:1 denotes:2 running:1 graphical:1 recycling:1 calculating:1 exploit:1 especially:1 quantity:1 dependence:1 diagonal:3 traditional:2 gradient:3 regulated:1 barber:1 cellular:1 assuming:1 code:1 length:1 index:3 minimizing:4 optionally:1 unfortunately:1 october:1 suppress:2 implementation:3 adjustable:2 perform:10 unknown:2 wl0:1 neuron:7 datasets:6 finite:2 descent:2 extended:1 hinton:4 smoothed:1 inferred:3 pair:1 required:2 kl:6 connection:1 imagenet:1 acoustic:1 inaccuracy:1 tractably:1 nip:2 able:2 suggested:2 usually:7 pattern:2 kulesza:1 challenge:1 summarize:1 built:1 including:1 power:3 treated:1 circumvent:1 indicator:2 improve:3 technology:1 naive:1 columbia:1 text:5 prior:5 literature:1 review:4 epoch:5 genov:1 asymptotic:1 graf:1 beside:1 loss:6 var:1 validation:2 foundation:1 vanhoucke:1 sufficient:1 xp:2 editor:1 charitable:1 i8:1 storing:2 classifying:1 repeat:1 last:2 free:4 keeping:2 surprisingly:1 bias:5 senior:1 deeper:1 perceptron:1 institute:2 sparse:1 distributed:1 van:1 calculated:5 vocabulary:1 opper:1 cumulative:1 forward:8 commonly:2 collection:1 author:2 adaptive:1 spam:3 nguyen:1 ribeiro:1 transaction:1 approximate:11 implicitly:1 kullback:1 keep:1 sequentially:4 overfitting:1 handbook:1 xi:4 alternatively:1 factorizing:1 continuous:7 search:1 table:4 additionally:1 promising:1 obtaining:1 hornik:1 mnn:22 improving:1 heidelberg:1 du:1 expansion:4 mse:1 bottou:2 complex:1 domain:2 did:1 dense:1 reuters:3 motivation:1 arrival:1 edition:1 neuronal:8 fig:1 intel:1 vr:1 precision:2 exponential:3 xl:2 crude:1 montavon:1 hw:1 grained:1 theorem:2 departs:1 specific:4 bishop:3 intractable:6 exists:1 quantization:1 adding:1 rv0:1 effectively:1 ci:1 phd:1 magnitude:2 entropy:1 simply:2 snns:1 arow:5 partially:1 brunel:1 springer:2 determines:1 mackay:1 goal:1 shared:1 replace:1 absence:1 considerable:1 hard:1 change:1 included:1 specifically:2 except:2 operates:1 infinite:2 averaging:4 pas:10 invariance:1 select:1 latter:2 arises:2 scan:1 brevity:2 reactive:1 crammer:1 bioinformatics:1 constructive:1 evaluate:1 audio:1 d1:1 srivastava:1 |
4,715 | 527 | Improved Hidden Markov Model
Speech Recognition Using
Radial Basis Function Networks
Elliot Singer and Richard P. Lippmann
Lincoln Laboratory, MIT
Lexington, MA 02173-9108, USA
Abstract
A high performance speaker-independent isolated-word hybrid speech recognizer was developed which combines Hidden Markov Models (HMMs)
and Radial Basis Function (RBF) neural networks. In recognition experiments using a speaker-independent E-set database, the hybrid recognizer had an error rate of 11.5% compared to 15.7% for the robust
unimodal Gaussian HMM recognizer upon which the hybrid system was
based. These results and additional experiments demonstrate that RBF
networks can be successfully incorporated in hybrid recognizers and suggest that they may be capable of good performance with fewer parameters
than required by Gaussian mixture classifiers. A global parameter optimization method designed to minimize the overall word error rather than
the frame recognition error failed to reduce the error rate.
1
HMM/RBF HYBRID RECOGNIZER
A hybrid isolated-word speech recognizer was developed which combines neural
network and Hidden Markov Model (HMM) approaches. The hybrid approach is
an attempt to capitalize on the superior static pattern classification performance of
neural network classifiers [6] while preserving the temporal alignment properties of
HMM Viterbi decoding. Our approach is unique when compared to other studies
[2, 5] in that we use Radial Basis Function (RBF) rather than multilayer sigmoidal
networks. RBF networks were chosen because their static pattern classification
performance is comparable to that of other networks and they can be trained rapidly
using a one-pass matrix inversion technique [8] .
The hybrid HMM/RBF isolated-word recognizer is shown in Figure 1. For each
159
160
Singer and Lippmann
BEST WORD MATCH
WORD
MODELS
UNKNOWN
WORD
BACKGROUND
NOISE MODEL
_
Figure 1: Block diagram of the hybrid recognizer for a two word vocabulary.
pattern presented at the input layer, the RBF network produces nodal outputs
which are estimates of Bayesian probabilities [9]. The RBF network consists of an
input layer, a hidden layer composed of Gaussian basis functions, and an output
layer. Connections from the input layer to the hidden layer are fixed at unity
while those from the hidden layer to the output layer are trained by minimizing
the overall mean-square error between actual and desired output values. Each
RBF output node has a corresponding state in a set of HMM word models which
represent the words in the vocabulary. HMM word models are left-to-right with
no skip states and have a one-state background noise model at either end. The
background noise models are identical for all words. In the simplified diagram of
Figure 1, the vocabulary consists of 2 E-set words and the HMMs contain 3 states
per word model. The number of RBF output nodes (classes) is thus equal to the
total number of HMM non-background states plus one to account for background
noise. In recognition, Viterbi decoders use the nodal outputs of the RBF network
as observation probabilities to produce word likelihood scores. Since the outputs of
the RBF network can take on any value, they were initially hard limited to 0.0 and
1.0. The transition probabilities estimated as part of HMM training are retained.
The final response of the recognizer corresponds to that word model which produces
the highest Viterbi likelihood. Note that the structure of the HMM/RBF hybrid
recognizer is identical to that of a tied-mixture HMM recognizer. For a discussion
and comparison of the two recognizers, see [10].
Training of the hybrid recognizer begins with the preliminary step of training an
HMM isolated-word recognizer. The robust HMM recognizer used provides good
recognition performance on many standard difficult isolated-word speech databases
[7]. It uses continuous density, unimodal diagonal-covariance Gaussian classifiers
for each word state. Variances of all states are equal to the grand variance averaged
over all words and states. The trained HMM recognizer is used to force an alignment
of every training token and assign a label to each frame. Labels correspond to both
states of HMM word models and output nodes of the RBF network.
The Gaussian centers in the RBF hidden layer are obtained by performing k-means
Improved Hidden Markov Model Speech Recognition Using Radial Basis Function Networks
clustering on speech frames and separate clustering on noise frames, where speech
and noise frames are distinguished on the basis of the initial Viterbi alignment. The
RBF weights from the hidden layer to the output layer are computed by presenting
input frames to the RBF network and setting the desired network outputs to 1.0
for the output node corresponding to the frame label and 0.0 for all other nodes.
The RBF hidden node outputs and their correlations are accumulated across all
training tokens and are used to estimate weights to the RBF output nodes using a
fast one-pass algorithm [8]. Unlike the performance of the system reported in [5],
additional training iterations using the hybrid recognizer to label frames did not
improve performance.
2
DATABASE
All experiments were performed using a large, speaker-independent E-set (9 word)
database derived from the ISOLET Spoken Letter Database [4]. The training set
consisted of 1,080 tokens (120 tokens per word) spoken by 60 female and 60 male
speakers for a total of 61,466 frames. The test set consisted of 540 tokens (60
tokens per word) spoken by a different set of 30 female and 30 male speakers for
a total of 30,406 frames . Speech was sampled at 16 kHz and had an average SNR
of 31.5 dB. Input vectors were based on a mel-cepstrum analysis of the speech
waveform as described in [7]. The input analysis window was 20ms wide and was
advanced at 10ms intervals. Input vectors were created by adjoining the first 12
non-energy cepstral coefficients, the first 13 first-difference cepstral coefficients, and
the first 13 second-difference cepstral coefficients. Since the hybrid was based on
an 8 state-per-word robust HMM recognizer, the RBF network contained a total of
73 output nodes (72 speech nodes and 1 background node). The error rate of the 8
state-per-word robust HMM recognizer on the speaker-independent E-set task was
15.7%.
3
MODIFICATIONS TO THE HYBRID RECOGNIZER
The performance of the baseline HMM/RBF hybrid recognizer described in Section 1 is quite poor. We found it necessary to select the recognizer structure carefully
and utilize intermediate outputs properly to achieve a higher level of performance.
A full description of these modifications is presented in [10]. Briefly, they include
normalizing the hidden node outputs to sum to 1.0, normalizing the RBF outputs
by the corresponding a priori class probabilities as estimated from the initial Viterbi
alignment, expanding the RBF network into three individually trained subnetworks
corresponding to the ceptrum, first difference cepstrum, and second difference cepstrum data streams, setting a lower limit of 10- 5 on the values produced at the RBF
output nodes, adjusting a global scaling factor applied to the variances of the RBF
centers, and setting the number of centers to 33,33, and 65 for the first, second, and
third subnets, respectively. The structure of the final hybrid recognizer is shown in
Figure 2. This recognizer has an error rate of 11.5% (binomial standard deviation
= ?1.4) on the E-set test data compared to 15.7% (?1.6) for the 8 state-per-word
unimodal Gaussian HMM recognizer, and 9.6% (?1.3) for a considerably more complex tied-mixture HMM recognizer [10]. The final hybrid system contained a total
of 131 Gaussians and 9,563 weights. On a SUN SPARCstation 2, training time for
161
162
Singer and Lippmann
the final hybrid recognizer was about 1 hour and testing time was about 10 minutes.
BEST WORD MATCH
Figure 2: Block diagram of multiple sub net hybrid recognizer.
4
GLOBAL OPTIMIZATION
In the hybrid recognizer described above, discriminative training is performed at
the frame level. A preliminary segmentation by the HMM recognizer assigns each
speech frame to a specific RBF output node or, equivalently, an HMM word state.
The RBF network weights are then computed to minimize the squared error between the network output and the desired output over all input frames. The goal of
the recognizer, however, is to classify words. To meet this goal, discriminant training should be performed on word-level rather than frame-level outputs. Recently,
several investigators have described techniques that optimize parameters based on
word-level discriminant criteria [1, 3]. These techniques seek to maximize a mutual
information type of criterion:
Lc
C logy,
=
where Lc. is the likelihood score of the word model corresponding to the correct
Lw Lw is the sum of the word likelihood scores for all models. By
result and L
computing oC/oO, the gradient of C with respect to parameter 0, we can optimize
any parameter in the hybrid recognizer using the update equation
=
where 0 is the new value of parameter 0, () is the previous value, and TJ is a gain
term proportional to the learning rate. Following [1], we refer to the word-level
optimization technique as "global optimization."
Improved Hidden Markov Model Speech Recognition Using Radial Basis Function Networks
To apply global optimization to the HMM/RBF hybrid recognizer, we derived the
formulas for the gradient of C with respect to
the weight connecting RBF center
i to RBF output node j in subnet k; Pj, the RBF output normalization factor for
RBF output node j in subnet k; and mfl' the Ith element of the mean of center i of
subnet k. For each token of length T frames, these are given by
wt '
8C
J:lwk
U
ij
= (be; L- Pw )
w
T
<I>~
"'"' frjt{3jt
kIt'
t=1
L..J
St
and
likelihood score for word model w,
Lw / Lw Lw is the normalized word likelihood,
{ I if RBF output node j is a member of the correct word model
o otherwise,
forward partial probability of HMM state j at time t,
backward partial probability of HMM state j at time t,
unnormalized output of RBF node j of subnet k at time t,
normalized output of ith Gaussian center of sub net k at time t,
~
~,. <I>~t
I
=1
,
Ith element of the input vector for subnet k at time t,
global scaling factor for the variances of sub net k,
[th component of the standard deviation of the ith Gaussian center
of subnet k,
number of RBF output nodes in sub net k.
In implementing global optimization, the frame-level training procedure described
earlier serves to initialize system parameters and hill climbing methods are used to
reestimate parameters iteratively. Thus, weights are initialized to the values derived
using the one-pass matrix inversion procedure, RBF output normalization factors
are initialized to the class priors, and Gaussian means are initialized to the k-means
clustering values. Note that while the priors sum to one, no such constraint was
placed on the RBF output normalization factors during global optimization.
It is worth noting that since the RBF network outputs in the hybrid recognizer
are a posteriori probabilities normalized by a priori class probabilities, their values
may exceed 1. The accumulation of these quantities in the Viterbi decoders often
leads to values of (Xjt{3jt and Lw in the range of 10 80 or greater. Numerical problems
with the implementation of the global optimization equations were avoided by using
log arithmetic for intermediate operations and working with the quantity {3jt! Lw
throughout. Values of 7J which produced reasonable results were generally in the
range of 10- 10 to 10- 6
163
164
Singer and Lippmann
The results of using the global optimization technique to estimate the RBF weights
are shown in Figure 3. Figure 3( a) shows the recognition performance on the training and test sets versus the number of training iterations and Figure 3(b) tracks
the value of the criterion C = Lei L on the training and test set under the same
conditions. It is apparent that the method succeeds in iteratively increasing the
value of the criterion and in significantly lowering the error rate on the training
data. Unfortunately, this behavior does not extend to improved performance on
the test data. This suggests that global optimization is overfitting the hybrid word
models to the training data. Results using global optimization to estimate RBF
output normalization factors and the Gaussian means produced similar results.
20
%ERROR
,.---~----r-----,r-----"
TEST
C = log (Lc I L)
0
-0.2
TRAIN
-0.4
10
-0.6
TEST
-0.8
o o~--~---~--~~--~
5
10
15
20
NUMBER OF ITERATIONS
-1
0
5
10
15
20
NUMBER OF ITERATIONS
Figure 3: (a) Error rates for training and test data. (b) Criterion C for training
and test data.
5
ACCURACY OF BAYES PROBABILITY
ESTIMATION
Three methods were used to determine how well RBF outputs estimate Bayes probabilities. First, since network outputs must sum to one if they are probabilities, we
computed the RMS error between the sum of the RBF outputs and unity for all
frames of the test data. The average RMS error was low (10- 4 or less for each
subnet). Second, the average output of each RBF node was computed because this
should equal the a priori probability of the class associated with the node [9]. This
condition was true for each subnet with an average RMS error on the order of 10- 5 .
For the final method, we partitioned the outputs into 100 equal size bins between
0.0 and 1.0. For each input pattern, we used the output values to select the appropriate bins and incremented the corresponding bin counts by one. In addition, we
incremented the correct-class bin count for the one bin which corresponded to the
class of the input pattern. For example, data indicated that for the 61,466 frames
of training tokens, nodal outputs of the cepstra subnet in the range 0.095-0.105 occurred 29,698 times and were correct classifications (regardless of class) 3,067 times.
If the outputs of the network were true Bayesian probabilities, we would expect the
Improved Hidden Markov Model Speech Recognition Using Radial Basis Function Networks
relative frequency of correct labeling to be close to 0.1. Similarly, relative frequencies measured in other intervals would also be expected to be close to the value of
the corresponding center of the interval. Thus, a plot of the relative frequencies for
each bin versus the bin centers should show the measured values lying close to the
diagonal.
The measured relative frequency data for the cepstra subnet and ?2u bounds for
the binomial standard deviations of the relative frequencies are shown in Figure 4.
Outputs below 0.0 and above 1.0 are fixed at 0.0 and 1.0, respectively. Although
the relative frequencies tend to be clustered around the diagonal, many values lie
test
outside the bounds. Furthermore, goodness-of-fit measurements using the
indicate that fits fail at significance levels well below .01. We conclude that although
the system provides good recognition accuracy, better performance may be obtained
with improved estimation of Bayesian probabilities.
x:
1r-------------------------------------------------------------~?
.J
~0.9
:3 0.8
t;
wO.7
IX
IX
0 0 .6
o
~0.5
gO.4
IX
~0.3
1= 0.2
:5
~O.1
0.1
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
RBF NETWORK OUTPUT (All Nodes)
1
Figure 4: Relative frequency of correct class labeling and ?2u bounds for the binomial standard deviation, cepstra subnet.
6
SUMMARY AND CONCLUSIONS
This paper describes a hybrid isolated-word speech recognizer which successfully
integrates Radial Basis Function neural networks and Hidden Markov Models. The
hybrid's performance is better than that of a tied-mixture recognizer of comparable
complexity and near that of a tied-mi..xture recognizer of considerably greater complexity. The structure of the RBF networks and the processing of network outputs
had to be carefully selected to provide this level of performance. A global optimization technique designed to maximize a word discrimination criterion did not
succeed in improving performance further. Statistical tests indicated that the accuracy of the Bayesian probability estimation performed by the RBF networks could
165
166
Singer and Lippmann
be improved. We conclude that RBF networks can be used to provide good performance and short training times in hybrid recognizers and that these systems may
require fewer parameters than Gaussian-mixture-based recognizers at comparable
performance levels.
Acknowledgements
This work was sponsored by the Defense Advanced Research Projects Agency. The
views expressed are those of the authors and do not reflect the official policy or
position of the U.S. Government.
References
[1] Yoshua Bengio, Renato De Mori, Giovanni Flammia, and Ralf Kompe. Global
optimization of a neural network - Hidden Markov model hybrid. Technical
Report TR-SOCS-90.22, MgGill University School of Computer Science, Montreal, Qc., Canada, December 1990.
[2] Herve Bourlard and Nelson Morgan. A continuous speech recognition system
embedding MLP into HMM. In D. Touretzky, editor, Advances in Neural
Information Processing 2, pages 186-193. Morgan Kaufmann, San Mateo, CA,
1990.
[3] John S. Bridle. Alpha-nets: A recurrent neural network architecture with a
hidden Markov model interpretation. Speech Communication, 9:83-92, 1990.
[4] Ron Cole, Yeshwant Muthusamy, and Mark Fanty. The Isolet spoken letter
database. Technical Report CSE 90-004, Oregon Graduate Institute of Science
and Technology, Beverton, OR, March 1990.
[5] Michael Franzini, Kai-Fu Lee, and Alex Waibel. Connectionist viterbi training: A new hybrid method for continuous speech recognition. In Proceedings
of IEEE International Conference on Acoustics Speech and Signal Processing.
IEEE, April 1990.
[6] Richard P. Lippmann. Pattern classification using neural networks. IEEE
Communications Magazine, 27(11):47-54, November 1989.
[7] Richard P. Lippmann and Ed A. Martin. Mqlti-style training for robust
isolated-word speech recognition. In Proceedings International Conference on
Acoustics Speech and Signal Processing, pages 705-708. IEEE, April 1987.
[8] Kenney Ng and Richard P. Lippmann. A comparative study of the practical characteristics of neural network and conventional pattern classifiers. In
R. P. Lippmann, J. Moody, and D. S. Touretzky, editors, Advances in Neural
Information Processing 3. Morgan Kaufmann, San Mateo, CA, 1991.
[9] Mike D. Richard and Richard P. Lippmann. Neural network classifiers estimate
Bayesian a posteriori probabilities. Neural Computation, In Press.
[10] Elliot Singer and Richard P. Lippmann. A speech recognizer using radial basis function neural networks in an HMM framework. In Proceedings of the
International Conference on Acoustics, Speech, and Signal Processing. IEEE,
1992.
| 527 |@word briefly:1 inversion:2 pw:1 lwk:1 seek:1 covariance:1 tr:1 initial:2 score:4 must:1 john:1 numerical:1 designed:2 plot:1 update:1 sponsored:1 discrimination:1 fewer:2 selected:1 ith:4 short:1 provides:2 node:21 ron:1 cse:1 sigmoidal:1 nodal:3 consists:2 combine:2 kenney:1 expected:1 behavior:1 actual:1 window:1 increasing:1 begin:1 project:1 developed:2 sparcstation:1 spoken:4 lexington:1 temporal:1 every:1 classifier:5 limit:1 meet:1 plus:1 mateo:2 suggests:1 hmms:2 limited:1 range:3 graduate:1 averaged:1 unique:1 practical:1 testing:1 block:2 procedure:2 significantly:1 word:42 radial:8 suggest:1 close:3 optimize:2 accumulation:1 conventional:1 center:9 go:1 regardless:1 qc:1 assigns:1 isolet:2 ralf:1 embedding:1 magazine:1 us:1 element:2 recognition:13 database:6 mike:1 sun:1 highest:1 incremented:2 agency:1 complexity:2 trained:4 upon:1 basis:10 train:1 fast:1 corresponded:1 labeling:2 outside:1 quite:1 apparent:1 kai:1 otherwise:1 final:5 net:5 fanty:1 rapidly:1 achieve:1 lincoln:1 description:1 produce:3 comparative:1 oo:1 recurrent:1 montreal:1 subnets:1 measured:3 ij:1 school:1 skip:1 indicate:1 waveform:1 correct:6 implementing:1 bin:7 subnet:11 require:1 government:1 assign:1 clustered:1 preliminary:2 lying:1 around:1 viterbi:7 recognizer:36 estimation:3 integrates:1 label:4 cole:1 individually:1 successfully:2 mit:1 gaussian:11 rather:3 derived:3 properly:1 likelihood:6 baseline:1 posteriori:2 accumulated:1 initially:1 hidden:16 overall:2 classification:4 priori:3 initialize:1 mutual:1 equal:4 ng:1 identical:2 capitalize:1 yoshua:1 report:2 connectionist:1 richard:7 kompe:1 composed:1 attempt:1 mlp:1 alignment:4 male:2 mixture:5 adjoining:1 tj:1 fu:1 capable:1 partial:2 necessary:1 herve:1 initialized:3 desired:3 isolated:7 classify:1 earlier:1 goodness:1 deviation:4 snr:1 reported:1 considerably:2 st:1 density:1 grand:1 international:3 lee:1 decoding:1 michael:1 connecting:1 moody:1 squared:1 reflect:1 logy:1 style:1 account:1 de:1 coefficient:3 oregon:1 stream:1 performed:4 view:1 bayes:2 minimize:2 square:1 accuracy:3 variance:4 kaufmann:2 characteristic:1 correspond:1 climbing:1 bayesian:5 produced:3 worth:1 reestimate:1 touretzky:2 ed:1 energy:1 frequency:7 associated:1 mi:1 static:2 bridle:1 sampled:1 gain:1 adjusting:1 segmentation:1 carefully:2 higher:1 response:1 improved:7 april:2 cepstrum:3 furthermore:1 correlation:1 working:1 indicated:2 lei:1 usa:1 contain:1 consisted:2 normalized:3 true:2 laboratory:1 iteratively:2 elliot:2 during:1 speaker:6 mel:1 oc:1 unnormalized:1 m:2 criterion:6 presenting:1 hill:1 demonstrate:1 yeshwant:1 recently:1 superior:1 khz:1 extend:1 occurred:1 interpretation:1 refer:1 measurement:1 similarly:1 had:3 recognizers:4 female:2 preserving:1 morgan:3 additional:2 greater:2 kit:1 determine:1 maximize:2 signal:3 arithmetic:1 full:1 unimodal:3 multiple:1 technical:2 match:2 xjt:1 multilayer:1 iteration:4 represent:1 normalization:4 background:6 addition:1 interval:3 diagram:3 flammia:1 unlike:1 tend:1 db:1 member:1 december:1 near:1 noting:1 intermediate:2 bengio:1 exceed:1 muthusamy:1 fit:2 architecture:1 reduce:1 rms:3 defense:1 wo:1 speech:22 generally:1 estimated:2 per:6 track:1 pj:1 utilize:1 backward:1 lowering:1 sum:5 letter:2 throughout:1 reasonable:1 scaling:2 comparable:3 layer:11 bound:3 renato:1 constraint:1 alex:1 cepstra:3 performing:1 martin:1 waibel:1 march:1 poor:1 across:1 describes:1 unity:2 partitioned:1 modification:2 mori:1 equation:2 count:2 fail:1 singer:6 end:1 subnetworks:1 serf:1 gaussians:1 operation:1 apply:1 appropriate:1 distinguished:1 binomial:3 clustering:3 include:1 franzini:1 quantity:2 diagonal:3 gradient:2 separate:1 hmm:27 decoder:2 nelson:1 discriminant:2 length:1 retained:1 minimizing:1 equivalently:1 difficult:1 unfortunately:1 implementation:1 policy:1 unknown:1 observation:1 markov:9 november:1 incorporated:1 communication:2 frame:18 canada:1 required:1 connection:1 acoustic:3 hour:1 below:2 pattern:7 hybrid:29 force:1 bourlard:1 advanced:2 improve:1 technology:1 created:1 prior:2 acknowledgement:1 relative:7 expect:1 proportional:1 versus:2 editor:2 summary:1 token:8 placed:1 institute:1 wide:1 cepstral:3 vocabulary:3 transition:1 giovanni:1 forward:1 author:1 san:2 simplified:1 avoided:1 alpha:1 lippmann:11 global:14 overfitting:1 conclude:2 discriminative:1 continuous:3 robust:5 expanding:1 ca:2 improving:1 complex:1 official:1 did:2 significance:1 noise:6 lc:3 sub:4 position:1 lie:1 tied:4 lw:7 third:1 ix:3 minute:1 formula:1 specific:1 jt:3 normalizing:2 failed:1 expressed:1 contained:2 corresponds:1 ma:1 succeed:1 goal:2 rbf:47 hard:1 wt:1 total:5 pas:3 succeeds:1 select:2 mark:1 investigator:1 |
4,716 | 5,270 | An Autoencoder Approach to Learning
Bilingual Word Representations
Sarath Chandar A P1 ? , Stanislas Lauly2 ? , Hugo Larochelle2 , Mitesh M Khapra3 ,
Balaraman Ravindran1 , Vikas Raykar3 , Amrita Saha3
1
Indian Institute of Technology Madras, 2 Universit?e de Sherbrooke, 3 IBM Research India
[email protected], {stanislas.lauly,hugo.larochelle}@usherbrooke.ca,
{mikhapra,viraykar,amrsaha4}@in.ibm.com, [email protected]
?
Both authors contributed equally
Abstract
Cross-language learning allows one to use training data from one language to
build models for a different language. Many approaches to bilingual learning require that we have word-level alignment of sentences from parallel corpora. In this
work we explore the use of autoencoder-based methods for cross-language learning of vectorial word representations that are coherent between two languages,
while not relying on word-level alignments. We show that by simply learning to
reconstruct the bag-of-words representations of aligned sentences, within and between languages, we can in fact learn high-quality representations and do without
word alignments. We empirically investigate the success of our approach on the
problem of cross-language text classification, where a classifier trained on a given
language (e.g., English) must learn to generalize to a different language (e.g., German). In experiments on 3 language pairs, we show that our approach achieves
state-of-the-art performance, outperforming a method exploiting word alignments
and a strong machine translation baseline.
1
Introduction
The accuracy of Natural Language Processing (NLP) tools for a given language depend heavily on
the availability of annotated resources in that language. For example, high quality POS taggers
[1], parsers [2], sentiment analyzers [3] are readily available for English. However, this is not the
case for many other languages such as Hindi, Marathi, Bodo, Farsi, and Urdu, for which annotated
data is scarce. This situation was acceptable in the past when only a few languages dominated the
digital content available online and elsewhere. However, the ever increasing number of languages
on the web today has made it important to accurately process natural language data in such resourcedeprived languages also. An obvious solution to this problem is to improve the annotated inventory
of these languages, but the cost, time and effort required act as a natural deterrent to this.
Another option is to exploit the unlabeled data available in a language. In this context, vectorial text
representations have proven useful for multiple NLP tasks [4, 5]. It has been shown that meaningful representations, capturing syntactic and semantic similarity, can be learned from unlabeled data.
While the majority of previous work on vectorial text representations has concentrated on the monolingual case, there has also been considerable interest in learning word and document representations
that are aligned across languages [6, 7, 8, 9, 10, 11, 12]. Such aligned representations allow the use
of resources from a resource-fortunate language to develop NLP capabilities in a resource-deprived
language.
One approach to cross-lingual exploitation of resources is to project parameters learned from the
annotated data of one language to another language [13, 14, 15, 16, 17]. These approaches rely on a
1
bilingual resource such as a Machine Translation (MT) system. Recent attempts at learning common
bilingual representations [9, 10, 11] aim to eliminate the need of such an MT system. A common
property of these approaches is that a word-level alignment of translated sentences is leveraged to
derive a regularization term relating word embeddings across languages. Such methods not only
eliminate the need for an MT system but also outperform MT based projection approaches.
In this paper, we experiment with methods that learn bilingual word representations without word-toword alignments of bilingual corpora during training. Unlike previous approaches, we only require
aligned sentences and do not rely on word-level alignments (e.g., extracted using GIZA++, as is
usual), simplifying the learning procedure. To do so, we propose and investigate bilingual autoencoder models, that learn hidden encoder representations of paired bag-of-words sentences that are
not only informative of the original bag-of-words but also predictive of the other language. Word
representations can then easily be extracted from the encoder and used in the context of a supervised NLP task. Specifically, we demonstrate the quality of these representations for the task of
cross-language document classification, where a labeled data set can be available in one language,
but not in another one. As we?ll see, our approach is able to reach state-of-the-art performance,
outperforming a method exploiting word alignments and a strong machine translation baseline.
2
Autoencoder for Bags-of-Words
Let x be the bag-of-words representation of a sentence. Specifically, each xi is a word index from
a fixed vocabulary of V words. As this is a bag-of-words, the order of the words within x does not
correspond to the word order in the original sentence. We wish to learn a D-dimensional vectorial
representation of our words from a training set of sentence bags-of-words {x(t) }Tt=1 .
We propose to achieve this by using an autoencoder model that encodes an input bag-of-words x with
a sum of the representations (embeddings) of the words present in x, followed by a non-linearity.
Specifically, let matrix W be the D ? V matrix whose columns are the vector representations for
each word. The encoder?s computation will involve summing over the columns of W for each
word in the bag-of-word. We will denote this encoder function ?(x). Then, using a decoder, the
autoencoder will be trained to optimize a loss function that measures how predictive of the original
bag-of-words is the encoder representation ?(x) .
There are different variations we can consider in the design of the encoder/decoder and the choice of
loss function. One must be careful however, as certain choices can be inappropriate for training on
word observations, which are intrinsically sparse and high-dimensional. In this paper, we explore
and compare two different approaches, described in the next two sub-sections.
2.1
Binary bag-of-words reconstruction training with merged bags-of-words
In the first approach, we start from the conventional autoencoder architecture, which minimizes a
cross-entropy loss that compares a binary vector observation with a decoder reconstruction. We thus
convert the bag-of-words x into a fixed-size but sparse binary vector v(x), which is such that v(x)xi
is 1 if word xi is present in x and otherwise 0.
From this representation, we obtain an encoder representation by multiplying v(x) with the word
representation matrix W
a(x) = c + Wv(x), ?(x) = h(a(x))
(1)
where h(?) is an element-wise non-linearity such as the sigmoid or hyperbolic tangent, and c is a
D-dimensional bias vector. Encoding thus involves summing the word representations of the words
present at least once in the bag-of-words.
To produce a reconstruction, we parametrize the decoder using the following non-linear form:
b (x) = sigm(V?(x) + b)
v
(2)
where V = WT , b is the bias vector of the reconstruction layer and sigm(a) = 1/(1 + exp(?a)) is
the sigmoid non-linearity.
2
Then, the reconstruction is compared to the original binary bag-of-words as follows:
`(v(x)) = ?
V
X
v(x)i log(b
v (x)i ) + (1 ? v(x)i ) log(1 ? vb(x)i ) .
(3)
i=1
Training proceeds by optimizing the sum of reconstruction cross-entropies across the training set,
e.g., using stochastic or mini-batch gradient descent.
Note that, since the binary bags-of-words are very high-dimensional (the dimensionality corresponds
to the size of the vocabulary, which is typically large), the above training procedure which aims at
reconstructing the complete binary bag-of-word, will be slow. Since we will later be training on
millions of sentences, training on each individual sentence bag-of-words will be expensive.
Thus, we propose a simple trick, which exploits the bag-of-words structure of the input. Assuming
we are performing mini-batch training (where a mini-batch contains a list of the bags-of-words of
adjacent sentences), we simply propose to merge the bags-of-words of the mini-batch into a single
bag-of-words and perform an update based on that merged bag-of-words. The resulting effect is that
each update is as efficient as in stochastic gradient descent, but the number of updates per training
epoch is divided by the mini-batch size . As we?ll see in the experimental section, this trick produces
good word representations, while sufficiently reducing training time. We note that, additionally, we
could have used the stochastic approach proposed by Dauphin et al. [18] for reconstructing binary
bag-of-words representations of documents, to further improve the efficiency of training. They use
importance sampling to avoid reconstructing the whole V -dimensional input vector.
2.2
Tree-based decoder training
The previous autoencoder architecture worked with a binary vectorial representation of the input
bag-of-words. In the second autoencoder architecture we investigate, we consider an architecture
that instead works with the bag (unordered list) representation more directly.
First, the encoder representation will now involve a sum of the representation of all words, reflecting
the relative frequency of each word:
a(x) = c +
|x|
X
W?,xi , ?(x) = h (a(x)) .
(4)
i=1
Moreover, decoder training will assume that, from the decoder?s output, we can obtain a probability
distribution p(b
x|?(x)) over any word x
b observed at the reconstruction output layer. Then, we can
treat the input bag-of-words as a |x|-trials multinomial sample from that distribution and use as the
reconstruction loss its negative log-likelihood:
`(x) =
V
X
? log p(b
x = xi |?(x)) .
(5)
i=1
We now must ensure that the decoder can compute p(b
x = xi |?(x)) efficiently from ?(x). Specifically, we?d like to avoid a procedure scaling linearly with the vocabulary size V , since V will be very
large in practice. This precludes any procedure that would compute the numerator of p(b
x = w|?(x))
for each possible word w separately and normalize it so it sums to one.
We instead opt for an approach borrowed from the work on neural network language models [19, 20].
Specifically, we use a probabilistic tree decomposition of p(b
x = xi |?(x)). Let?s assume each word
has been placed at the leaf of a binary tree. We can then treat the sampling of a word as a stochastic
path from the root of the tree to one of the leaves.
We denote as l(x) the sequence of internal nodes in the path from the root to a given word x, with
l(x)1 always corresponding to the root. We will denote as ?(x) the vector of associated left/right
branching choices on that path, where ?(x)k = 0 means the path branches left at internal node l(x)k
and otherwise branches right if ?(x)k = 1. Then, the probability p(b
x = x|?(x)) of reconstructing
a certain word x observed in the bag-of-words is computed as
|?(?
x)|
p(b
x|?(x)) =
Y
k=1
3
p(?(b
x)k |?(x))
(6)
where p(?(b
x)k |?(x)) is output by the decoder. By using a full binary tree of words, the number of
different decoder outputs required to compute p(b
x|?(x)) will be logarithmic in the vocabulary size
V . Since there are |x| words in the bag-of-words, at most O(|x| log V ) outputs are required from
the decoder. This is of course a worst case scenario, since words will share internal nodes between
their paths, for which the decoder output can be computed just once. As for organizing words into a
tree, as in Larochelle and Lauly [21] we used a random assignment of words to the leaves of the full
binary tree, which we have found to work well in practice.
Finally, we need to choose a parametrized form for the decoder. We choose the following form:
p(?(b
x)k = 1|?(x)) = sigm(bl(?xi )k + Vl(?xi )k ,? ?(x))
(7)
where b is a (V -1)-dimensional bias vector and V is a (V ?1)?D matrix. Each left/right branching
probability is thus modeled with a logistic regression model applied on the encoder representation
of the input bag-of-words ?(x).
3
Bilingual autoencoders
Let?s now assume that for each sentence bag-of-words x in some source language X , we have an
associated bag-of-words y for this sentence translated in some target language Y by a human expert.
Assuming we have a training set of such (x, y) pairs, we?d like to use it to learn representations in
both languages that are aligned, such that pairs of translated words have similar representations.
To achieve this, we propose to augment the regular autoencoder proposed in Section 2 so that, from
the sentence representation in a given language, a reconstruction can be attempted of the original
sentence in the other language. Specifically, we now define language specific word representation
matrices Wx and Wy , corresponding to the languages of the words in x and y respectively. Let
V X and V Y also be the number of words in the vocabulary of both languages, which can be different. The word representations however are of the same size D in both languages. For the binary
reconstruction autoencoder, the bag-of-words representations extracted by the encoder become
?(x) = h c + WX v(x) , ?(y) = h c + WY v(y)
and are similarly extended for the tree-based autoencoder. Notice that we share the bias c before the
non-linearity across encoders, to encourage the encoders in both languages to produce representations on the same scale.
From the sentence in either languages, we want to be able to perform a reconstruction of the original
sentence in both the languages. In particular, given a representation in any language, we?d like a
decoder that can perform a reconstruction in language X and another decoder that can reconstruct in
language Y. Again, we use decoders of the form proposed in either Section 2.1 or 2.2 (see Figure 1),
but let the decoders of each language have their own parameters (bX , VX ) and (bY , VY ).
This encoder/decoder decomposition structure allows us to learn a mapping within each language
and across the languages. Specifically, for a given pair (x, y), we can train the model to (1) construct
y from x (loss `(x, y)), (2) construct x from y (loss `(y, x)), (3) reconstruct x from itself (loss
`(x)) and (4) reconstruct y from itself (loss `(y)). We follow this approach in our experiments and
optimize the sum of the corresponding 4 losses during training.
3.1
Joint reconstruction and cross-lingual correlation
We also considered incorporating two additional terms to the loss function, in an attempt to favour
even more meaningful bilingual representations:
`(x, y) + `(y, x) + `(x) + `(y) + ?`([x, y], [x, y]) ? ? ? cor(a(x), a(y))
(8)
The term `([x, y], [x, y]) is simply a joint reconstruction term, where both languages are simultanouesly presented as input and reconstructed. The second term cor(a(x), a(y)) encourages correlation between the representation of each language. It is the sum of the scalar correlations between
each pair a(x)k , a(y)k , across all dimensions k of the vectors a(x), a(y)1 . To obtain a stochastic
estimate of the correlation, during training, small mini-batches are used.
1
While we could have applied the correlation term on ?(x), ?(y) directly, applying it to the pre-activation
function vectors was found to be more numerically stable.
4
Figure 1: Left: Bilingual autoencoder based on the binary reconstruction error. Right: Tree-based
bilingual autoencoder. In this example, they both reconstruct the bag-of-words for the English sentence ?the dog barked? from its French translation ?le chien a japp?e?.
3.2
Document representations
Once we learn the language specific word representation matrices Wx and Wy as described above,
we can use them to construct document representations, by using their columns as word vector
representations. Given a document d written in language Z ? {X , Y} and containing m words,
zP
1 , z2 , . . . , zm , we represent it as the tf-idf weighted sum of its words? representations ?(d) =
m
Z
i=1 tf-idf(zi ) ? W.,zi . We use the document representations thus obtained to train our document
classifiers, in the cross-lingual document classification task described in Section 5.
4
Related Work
Recent work that has considered the problem of learning bilingual representations of words usually
has relied on word-level alignments. Klementiev et al. [9] propose to train simultaneously two neural
network languages models, along with a regularization term that encourages pairs of frequently
aligned words to have similar word embeddings. Thus, the use of this regularization term requires
to first obtain word-level alignments from parallel corpora. Zou et al. [10] use a similar approach,
with a different form for the regularizer and neural network language models as in [5]. In our work,
we specifically investigate whether a method that does not rely on word-level alignments can learn
comparably useful multilingual embeddings in the context of document classification.
Looking more generally at neural networks that learn multilingual representations of words or
phrases, we mention the work of Gao et al. [22] which showed that a useful linear mapping between
separately trained monolingual skip-gram language models could be learned. They too however
rely on the specification of pairs of words in the two languages to align. Mikolov et al. [11] also propose a method for training a neural network to learn useful representations of phrases, in the context
of a phrase-based translation model. In this case, phrase-level alignments (usually extracted from
word-level alignments) are required. Recently, Hermann and Blunsom [23], [24] proposed neural
network architectures and a margin-based training objective that, as in this work, does not rely on
word alignments. We will briefly discuss this work in the experiments section.
5
Experiments
The techniques proposed in this paper enable us to learn bilingual embeddings which capture crosslanguage similarity between words. We propose to evaluate the quality of these embeddings by using
them for the task of cross-language document classification. We followed closely the setup used by
Klementiev et al. [9] and compare with their method, for which word representations are publicly
available2 . The set up is as follows. A labeled data set of documents in some language X is available
to train a classifier, however we are interested in classifying documents in a different language Y
at test time. To achieve this, we leverage some bilingual corpora, which is not labeled with any
2
http://people.mmci.uni-saarland.de/?aklement/data/distrib/
5
document-level categories. This bilingual corpora is used to learn document representations that are
coherent between languages X and Y. The hope is thus that we can successfully apply the classifier
trained on document representations for language X directly to the document representations for
language Y. Following this setup, we performed experiments on 3 data sets of language pairs:
English/German (EN/DE), English/French (EN/FR) and English/Spanish (EN/ES).
5.1
Data
For learning the bilingual embeddings, we used sections of the Europarl corpus [25] which contains
roughly 2 million parallel sentences. We considered 3 language pairs. We used the same preprocessing as used by Klementiev et al. [9]. We tokenized the sentences using NLTK [26], removed
punctuations and lowercased all words. We did not remove stopwords.
As for the labeled document classification data sets, they were extracted from sections of the Reuters
RCV1/RCV2 corpora, again for the 3 pairs considered in our experiments. Following Klementiev
et al. [9], we consider only documents which were assigned exactly one of the 4 top level categories
in the topic hierarchy (CCAT, ECAT, GCAT and MCAT). These documents are also pre-processed
using a similar procedure as that used for the Europarl corpus. We used the same vocabularies as
those used by Klementiev et al. [9] (varying in size between 35, 000 and 50, 000).
For each pair of languages, our overall procedure for cross-language classification can be summarized as follows:
Train representation: Train bilingual word representations Wx and Wy on sentence pairs extracted from Europarl for languages X and Y. Optionally, we also use the monolingual documents
from RCV1/RCV2 to reinforce the monolingual embeddings (this choice is cross-validated). These
non-parallel documents can be used through the losses `(x) and `(y) (i.e. by reconstructing x from x
or y from y). Note that Klementiev et al. [9] also used this data when training word representations.
Train classifier: Train document classifier on the Reuters training set for language X , where documents are represented using the word representations Wx (see Section 3.2). As in Klementiev et al.
[9] we used an averaged perceptron trained for 10 epochs, for all the experiments.
Test-time classification: Use the classifier trained in the previous step on the Reuters test set for
language Y, using the word representations Wy to represent the documents.
We trained the following autoencoders3 : BAE-cr which uses reconstruction error based decoder
training (see Section 2.1) and BAE-tr which uses tree-based decoder training (see Section 2.2).
Models were trained for up to 20 epochs using the same data as described earlier. BAE-cr used
mini-batch (of size 20) stochastic gradient descent, while BAE-tr used regular stochastic gradient.
All results are for word embeddings of size D = 40, as in Klementiev et al. [9]. Further, to speed
up the training for BAE-cr we merged each 5 adjacent sentence pairs into a single training instance,
as described in Section 2.1. For all language pairs, the joint reconstruction ? was fixed to 1 and
the cross-lingual correlation factor ? to 4 for BAE-cr. For BAE-tr, none of these additional terms
were found to be particularly beneficial, so we set their weights to 0 for all tasks. The other hyperparameters were tuned to each task using a training/validation set split of 80% and 20% and using
the performance on the validation set of an averaged perceptron trained on the smaller training set
portion (notice that this corresponds to a monolingual classification experiment, since the general
assumption is that no labeled data is available in the test set language).
5.2
Comparison of the performance of different models
We now present the cross language classification results obtained by using the embeddings produced
by our two autoencoders. We also compare our models with the following approaches:
Klementiev et al.: This model uses word embeddings learned by a multitask neural network language model with a regularization term that encourages pairs of frequently aligned words to have
similar word embeddings. From these embeddings, document representations are computed as described in Section 3.2.
3
Our word representations and code are available at http://www.sarathchandar.in/crl.html
6
Table 1: Cross-lingual classification accuracy for 3 language pairs, with 1000 labeled examples.
BAE-tr
BAE-cr
Klementiev et al.
MT
Majority Class
EN ? DE
81.8
91.8
77.6
68.1
46.8
DE ? EN
60.1
74.2
71.1
67.4
46.8
EN ? FR
70.4
84.6
74.5
76.3
22.5
FR ? EN
61.8
74.2
61.9
71.1
25.0
EN ? ES
59.4
49.0
31.3
52.0
15.3
ES ? EN
60.4
64.4
63.0
58.4
22.2
Table 2: Example English words along with the closest words both in English (EN) and German
(DE), using the Euclidean distance between the embeddings learned by BAE-cr.
Word
Lang Nearest neighbors
Word
Lang Nearest neighbors
EN
january, march, october
EN
oil, supply, supplies, gas
january
oil
DE
januar, m?arz, oktober
DE
o? l, boden, befindet, ger?at
EN
president, i, mr, presidents
EN
microsoft, cds, insider
president
microsoft
DE
pr?asident, pr?asidentin
DE
microsoft, cds, warner
EN
said, told, say, believe
EN
market, markets, single
said
market
DE
gesagt, sagte, sehr, heute
DE
markt, marktes, m?arkte
MT: Here, test documents are translated to the language of the training documents using a standard
phrase-based MT system, MOSES4 which was trained using default parameters and a 5-gram language model on the Europarl corpus (same as the one used for inducing our bilingual embeddings).
Majority Class: Test documents are simply assigned the most frequent class in the training set.
For the EN/DE language pairs, we directly report the results from Klementiev et al. [9]. For the other
pairs (not reported in Klementiev et al. [9]), we used the embeddings available online and performed
the classification experiment ourselves. Similarly, we generated the MT baseline ourselves.
Table 1 summarizes the results. They were obtained using 1000 RCV training examples. We report
results in both directions, i.e. language X to Y and vice versa. The best performing method is always
either BAE-cr or BAE-tr, with BAE-cr having the best performance overall. In particular, BAE-cr
often outperforms the approach of Klementiev et al. [9] by a large margin.
We also mention the recent work of Hermann and Blunsom [23], who proposed two neural network
architectures for learning word and document representations using sentence-aligned data only. Instead of an autoencoder paradigm, they propose a margin-based objective that aims to make the
representation of aligned sentences closer than non-aligned sentences. While their trained embeddings are not publicly available, they report results for the EN/DE classification experiments, with
representations of the same size as here (D = 40) and trained on 500K EN/DE sentence pairs. Their
best model reaches accuracies of 83.7% and 71.4% respectively for the EN ? DE and DE ? EN
tasks. One clear advantage of our model is that unlike their model, it can use additional monolingual data. Indeed, when we train BAE-cr with 500k EN/DE sentence pairs, plus monolingual RCV
documents (which come at no additional cost), we get accuracies of 87.9% (EN ? DE) and 76.7%
(DE ? EN), still improving on their best model. If we do not use the monolingual data, BAE-cr?s
performance is worse but still competitive at 86.1% for EN ? DE and 68.8% for DE ? EN.
We also evaluate the effect of varying the amount of supervised training data for training the classifier. For brevity, we report only the results for the EN/DE pair, which are summarized in Figure 2.
We observe that BAE-cr clearly outperforms the other models at almost all data sizes. More importantly, it performs remarkably well at very low data sizes (100), suggesting it learns very meaningful
embeddings, though the method can still benefit from more labeled data (as in the DE ? EN case).
Table 2 also illustrates the properties captured within and across languages, for the EN/DE pair5 .
For a few English words, the words with closest word representations (in Euclidean distance) are
shown, for both English and German. We observe that words that form a translation pair are close,
but also that close words within a language are syntactically/semantically similar as well.
4
5
http://www.statmt.org/moses/
See also the supplementary material for a t-SNE visualization of the word representations.
7
Figure 2: Cross-lingual classification accuracy results, from EN ? DE (left), and DE ? EN (right).
The excellent performance of BAE-cr suggests that merging several sentences into single bags-ofwords can still yield good word embeddings. In other words, not only we do not need to rely
on word-level alignments, but exact sentence-level alignment is also not essential to reach good
performances. We experimented with the merging of 5, 25 and 50 adjacent sentences (see the
supplementary material). Generally speaking, these experiments also confirm that even coarser
merges can sometimes not be detrimental. However, for certain language pairs, there can be an
important decrease in performance. On the other hand, when comparing the performance of BAE-tr
with the use of 5-sentences merges, no substantial impact is observed.
6
Conclusion and Future Work
We presented evidence that meaningful bilingual word representations could be learned without
relying on word-level alignments or using fairly coarse sentence-level alignments. In particular, we
showed that even though our model does not use word level alignments, it is able to reach state-ofthe-art performance, even compared to a method that exploits word-level alignments. In addition, it
also outperforms a strong machine translation baseline.
For future work, we would like to investigate extensions of our bag-of-words bilingual autoencoder
to bags-of-n-grams, where the model would also have to learn representations for short phrases.
Such a model should be particularly useful in the context of a machine translation system. We
would also like to explore the possibility of converting our bilingual model to a multilingual model
which can learn common representations for multiple languages given different amounts of parallel
data between these languages.
Acknowledgement
We would like to thank Alexander Klementiev and Ivan Titov for providing the code for the classifier
and data indices. This work was supported in part by Google.
References
[1] Kristina Toutanova, Dan Klein, Christopher D. Manning, and Yoram Singer. Feature-rich part-of-speech
tagging with a cyclic dependency network. In Proceedings of the 2003 Conference of the North American
Chapter of the Association for Computational Linguistics, NAACL ?03, pages 173?180, 2003.
[2] Richard Socher, John Bauer, Christopher D. Manning, and Ng Andrew Y. Parsing with compositional
vector grammars. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 455?465, Sofia, Bulgaria, August 2013.
[3] Bing Liu. Sentiment Analysis and Opinion Mining. Synthesis Lectures on Human Language Technologies.
Morgan & Claypool Publishers, 2012.
8
[4] Joseph Turian, Lev Ratinov, and Yoshua Bengio. Word representations: A simple and general method for
semi-supervised learning. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics (ACL2010), pages 384?394, 2010.
[5] Ronan Collobert, Jason Weston, L?eon Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa.
Natural Language Processing (Almost) from Scratch. Journal of Machine Learning Research, 12, 2011.
[6] Susan T Dumais, Todd A Letsche, Michael L Littman, and Thomas K Landauer. Automatic crosslanguage retrieval using latent semantic indexing. AAAI spring symposium on cross-language text and
speech retrieval, 15:21, 1997.
[7] John C. Platt, Kristina Toutanova, and Wen-tau Yih. Translingual document representations from discriminative projections. In Proceedings of the 2010 Conference on Empirical Methods in Natural Language
Processing, EMNLP ?10, pages 251?261, Stroudsburg, PA, USA, 2010.
[8] Wen-tau Yih, Kristina Toutanova, John C. Platt, and Christopher Meek. Learning discriminative projections for text similarity measures. In Proceedings of the Fifteenth Conference on Computational Natural
Language Learning, CoNLL ?11, pages 247?256, Stroudsburg, PA, USA, 2011.
[9] Alexandre Klementiev, Ivan Titov, and Binod Bhattarai. Inducing Crosslingual Distributed Representations of Words. In Proceedings of the International Conference on Computational Linguistics, 2012.
[10] Will Y. Zou, Richard Socher, Daniel Cer, and Christopher D. Manning. Bilingual Word Embeddings for
Phrase-Based Machine Translation. In Empirical Methods in Natural Language Processing, 2013.
[11] Tomas Mikolov, Quoc Le, and Ilya Sutskever. Exploiting Similarities among Languages for Machine
Translation. Technical report, arXiv, 2013.
[12] Manaal Faruqui and Chris Dyer. Improving vector space word representations using multilingual correlation. In Proceedings of the 14th Conference of the European Chapter of the Association for Computational
Linguistics, pages 462?471, Gothenburg, Sweden, April 2014.
[13] David Yarowsky and Grace Ngai. Inducing multilingual pos taggers and np bracketers via robust projection across aligned corpora. In Proceedings of the second meeting of the North American Chapter of the
Association for Computational Linguistics on Language technologies, pages 1?8, Pennsylvania, 2001.
[14] Dipanjan Das and Slav Petrov. Unsupervised part-of-speech tagging with bilingual graph-based projections. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics:
Human Language Technologies, pages 600?609, Portland, Oregon, USA, June 2011.
[15] Rada Mihalcea, Carmen Banea, and Janyce Wiebe. Learning multilingual subjective language via crosslingual projections. In Proceedings of the 45th Annual Meeting of the Association of Computational
Linguistics, pages 976?983, Prague, Czech Republic, June 2007.
[16] Xiaojun Wan. Co-training for cross-lingual sentiment classification. In Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural
Language Processing of the AFNLP, pages 235?243, Suntec, Singapore, August 2009.
[17] Sebastian Pad?o and Mirella Lapata. Cross-lingual annotation projection for semantic roles. Journal of
Artificial Intelligence Research (JAIR), 36:307?340, 2009.
[18] Yann Dauphin, Xavier Glorot, and Yoshua Bengio. Large-Scale Learning of Embeddings with Reconstruction Sampling. In Proceedings of the 28th International Conference on Machine Learning (ICML
2011), pages 945?952. Omnipress, 2011.
[19] Frederic Morin and Yoshua Bengio. Hierarchical Probabilistic Neural Network Language Model. In
Proceedings of the 10th International Workshop on Artificial Intelligence and Statistics (AISTATS 2005),
pages 246?252. Society for Artificial Intelligence and Statistics, 2005.
[20] Andriy Mnih and Geoffrey E Hinton. A Scalable Hierarchical Distributed Language Model. In Advances
in Neural Information Processing Systems 21 (NIPS 2008), pages 1081?1088, 2009.
[21] Hugo Larochelle and Stanislas Lauly. A Neural Autoregressive Topic Model. In Advances in Neural
Information Processing Systems 25 (NIPS 25), 2012.
[22] Jianfeng Gao, Xiaodong He, Wen-tau Yih, and Li Deng. Learning continuous phrase representations for
translation modeling. In Proceedings of the 52nd Annual Meeting of the Association for Computational
Linguistics (Volume 1: Long Papers), pages 699?709, Baltimore, Maryland, June 2014.
[23] Karl Moritz Hermann and Phil Blunsom. Multilingual models for compositional distributed semantics.
In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, ACL 2014,
June 22-27, 2014, Baltimore, MD, USA, Volume 1: Long Papers, pages 58?68, 2014.
[24] Karl Moritz Hermann and Phil Blunsom. Multilingual Distributed Representations without Word Alignment. In Proceedings of International Conference on Learning Representations (ICLR), 2014.
[25] Philipp Koehn. Europarl: A parallel corpus for statistical machine translation. In MT Summit, 2005.
[26] Edward Loper Bird Steven and Ewan Klein. Natural Language Processing with Python. OReilly Media
Inc., 2009.
9
| 5270 |@word multitask:1 trial:1 exploitation:1 briefly:1 nd:2 simplifying:1 decomposition:2 pavel:1 mention:2 tr:6 yih:3 cyclic:1 contains:2 liu:1 daniel:1 tuned:1 document:32 past:1 outperforms:3 subjective:1 com:2 z2:1 comparing:1 activation:1 gmail:1 lang:2 must:3 readily:1 lauly:3 written:1 john:3 parsing:1 ronan:1 informative:1 wx:5 remove:1 update:3 kristina:3 intelligence:3 leaf:3 short:1 coarse:1 cse:1 node:3 philipp:1 org:1 tagger:2 stopwords:1 saarland:1 along:2 become:1 supply:2 symposium:1 dan:1 xiaojun:1 tagging:2 indeed:1 market:3 kuksa:1 p1:1 frequently:2 warner:1 roughly:1 relying:2 inappropriate:1 increasing:1 project:1 linearity:4 moreover:1 medium:1 rcv:2 minimizes:1 act:1 exactly:1 universit:1 klementiev:15 classifier:9 platt:2 yarowsky:1 before:1 bodo:1 treat:2 todd:1 encoding:1 iitm:1 lev:1 path:5 merge:1 blunsom:4 plus:1 acl:2 mirella:1 bird:1 suggests:1 co:1 averaged:2 practice:2 procedure:6 mihalcea:1 empirical:2 hyperbolic:1 projection:7 word:129 pre:2 regular:2 morin:1 get:1 unlabeled:2 close:2 giza:1 faruqui:1 context:5 applying:1 rcv2:2 optimize:2 conventional:1 www:2 phil:2 tomas:1 sarath:1 importantly:1 variation:1 president:3 target:1 today:1 heavily:1 parser:1 hierarchy:1 exact:1 us:3 trick:2 element:1 pa:2 expensive:1 particularly:2 summit:1 coarser:1 labeled:7 observed:3 role:1 steven:1 capture:1 worst:1 susan:1 decrease:1 removed:1 substantial:1 littman:1 sherbrooke:1 trained:12 depend:1 predictive:2 efficiency:1 translated:4 po:2 easily:1 joint:5 represented:1 chapter:3 regularizer:1 sigm:3 train:9 artificial:3 jianfeng:1 insider:1 whose:1 supplementary:2 koehn:1 say:1 reconstruct:5 otherwise:2 encoder:11 precludes:1 grammar:1 statistic:2 syntactic:1 itself:2 online:2 afnlp:1 sequence:1 advantage:1 manaal:1 propose:9 reconstruction:18 zm:1 fr:3 frequent:1 aligned:11 organizing:1 achieve:3 inducing:3 normalize:1 exploiting:3 sutskever:1 zp:1 produce:3 stroudsburg:2 derive:1 develop:1 ac:1 andrew:1 nearest:2 borrowed:1 strong:3 edward:1 involves:1 skip:1 larochelle:3 come:1 direction:1 hermann:4 merged:3 annotated:4 closely:1 stochastic:7 human:3 vx:1 enable:1 opinion:1 material:2 require:2 opt:1 extension:1 sufficiently:1 considered:4 claypool:1 exp:1 mapping:2 achieves:1 bag:37 vice:1 tf:2 successfully:1 tool:1 weighted:1 hope:1 clearly:1 always:2 aim:3 avoid:2 cr:13 varying:2 validated:1 loper:1 june:4 portland:1 likelihood:1 balaraman:1 baseline:4 vl:1 eliminate:2 typically:1 pad:1 hidden:1 interested:1 semantics:1 statmt:1 overall:2 classification:15 html:1 dauphin:2 augment:1 among:1 available2:1 art:3 fairly:1 once:3 construct:3 having:1 ng:1 sampling:3 koray:1 cer:1 unsupervised:1 icml:1 stanislas:3 future:2 report:5 yoshua:3 np:1 richard:2 few:2 wen:3 simultaneously:1 individual:1 ourselves:2 microsoft:3 attempt:2 interest:1 investigate:5 possibility:1 mining:1 mnih:1 alignment:21 punctuation:1 encourage:1 closer:1 bulgaria:1 sweden:1 mcat:1 tree:10 euclidean:2 instance:1 column:3 earlier:1 modeling:1 assignment:1 phrase:8 cost:2 republic:1 too:1 reported:1 encoders:2 dependency:1 dumais:1 st:1 international:5 probabilistic:2 told:1 michael:2 synthesis:1 ilya:1 again:2 aaai:1 containing:1 wan:1 leveraged:1 choose:2 emnlp:1 worse:1 expert:1 american:2 bx:1 li:1 suggesting:1 de:27 lapata:1 unordered:1 summarized:2 chandar:1 availability:1 north:2 oregon:1 inc:1 collobert:1 later:1 root:3 performed:2 jason:1 bhattarai:1 portion:1 start:1 relied:1 option:1 parallel:6 capability:1 competitive:1 annotation:1 publicly:2 accuracy:5 who:1 efficiently:1 correspond:1 yield:1 ofthe:1 generalize:1 kavukcuoglu:1 accurately:1 comparably:1 ccat:1 none:1 produced:1 multiplying:1 reach:4 sebastian:1 petrov:1 frequency:1 obvious:1 associated:2 intrinsically:1 dimensionality:1 oreilly:1 reflecting:1 gcat:1 alexandre:1 jair:1 supervised:3 follow:1 april:1 though:2 just:1 autoencoders:2 correlation:7 hand:1 web:1 christopher:4 google:1 french:2 logistic:1 quality:4 believe:1 oil:2 effect:2 naacl:1 usa:4 xiaodong:1 marathi:1 regularization:4 assigned:2 xavier:1 moritz:2 semantic:3 adjacent:3 ll:2 during:3 numerator:1 branching:2 encourages:3 spanish:1 tt:1 demonstrate:1 complete:1 performs:1 syntactically:1 omnipress:1 wise:1 recently:1 common:3 sigmoid:2 multinomial:1 mt:9 hugo:3 empirically:1 ngai:1 volume:3 million:2 association:9 crosslingual:2 he:1 relating:1 numerically:1 versa:1 automatic:1 similarly:2 analyzer:1 language:97 stable:1 specification:1 similarity:4 align:1 closest:2 own:1 recent:3 showed:2 optimizing:1 scenario:1 certain:3 outperforming:2 success:1 usherbrooke:1 binary:13 wv:1 meeting:8 captured:1 morgan:1 additional:4 mr:1 deng:1 urdu:1 converting:1 paradigm:1 semi:1 branch:2 multiple:2 full:2 karlen:1 technical:1 cross:19 long:3 retrieval:2 divided:1 equally:1 paired:1 impact:1 scalable:1 regression:1 mmci:1 fifteenth:1 arxiv:1 represent:2 sometimes:1 addition:1 want:1 separately:2 remarkably:1 baltimore:2 source:1 publisher:1 unlike:2 prague:1 leverage:1 split:1 embeddings:21 japp:1 bengio:3 ecat:1 ivan:2 zi:2 architecture:6 pennsylvania:1 andriy:1 favour:1 whether:1 effort:1 sentiment:3 speech:3 speaking:1 compositional:2 useful:5 generally:2 clear:1 involve:2 amount:2 concentrated:1 processed:1 category:2 http:3 outperform:1 vy:1 singapore:1 notice:2 moses:1 per:1 klein:2 ravi:1 bae:19 graph:1 sum:7 convert:1 ratinov:1 almost:2 yann:1 acceptable:1 scaling:1 vb:1 summarizes:1 wiebe:1 conll:1 capturing:1 layer:2 followed:2 meek:1 annual:7 vectorial:5 worked:1 idf:2 letsche:1 encodes:1 dominated:1 speed:1 carmen:1 spring:1 performing:2 mikolov:2 rcv1:2 slav:1 march:1 manning:3 lingual:8 across:8 beneficial:1 reconstructing:5 smaller:1 joseph:1 quoc:1 deprived:1 pr:2 indexing:1 translingual:1 resource:6 visualization:1 bing:1 discus:1 german:4 singer:1 dyer:1 cor:2 available:9 parametrize:1 apply:1 observe:2 titov:2 hierarchical:2 batch:7 vikas:1 original:6 thomas:1 top:1 nlp:4 ensure:1 linguistics:10 madras:1 exploit:3 yoram:1 eon:1 build:1 society:1 bl:1 objective:2 ofwords:1 ewan:1 usual:1 grace:1 md:1 said:2 gradient:4 detrimental:1 iclr:1 distance:2 thank:1 reinforce:1 maryland:1 majority:3 decoder:20 parametrized:1 chris:1 topic:2 assuming:2 tokenized:1 code:2 index:2 modeled:1 mini:7 providing:1 optionally:1 setup:2 october:1 sne:1 negative:1 design:1 contributed:1 perform:3 observation:2 descent:3 mitesh:1 january:2 gas:1 situation:1 extended:1 ever:1 looking:1 hinton:1 august:2 david:1 pair:23 required:4 dog:1 sentence:32 coherent:2 learned:6 merges:2 czech:1 nip:2 able:3 proceeds:1 wy:5 usually:2 tau:3 natural:9 rely:6 scarce:1 hindi:1 improve:2 technology:4 autoencoder:16 text:5 epoch:3 acknowledgement:1 tangent:1 python:1 relative:1 loss:11 lecture:1 monolingual:8 ger:1 proven:1 geoffrey:1 digital:1 validation:2 classifying:1 share:2 cd:2 ibm:2 karl:2 translation:12 elsewhere:1 course:1 placed:1 supported:1 english:10 bias:4 allow:1 perceptron:2 india:1 institute:1 neighbor:2 sparse:2 benefit:1 bauer:1 lowercased:1 dimension:1 vocabulary:6 gram:3 default:1 rich:1 distributed:4 dipanjan:1 author:1 made:1 autoregressive:1 preprocessing:1 reconstructed:1 uni:1 chien:1 multilingual:8 confirm:1 corpus:11 summing:2 xi:9 discriminative:2 landauer:1 continuous:1 latent:1 table:4 additionally:1 learn:15 robust:1 ca:1 boden:1 improving:2 inventory:1 excellent:1 bottou:1 zou:2 european:1 da:1 did:1 aistats:1 linearly:1 whole:1 reuters:3 bilingual:23 hyperparameters:1 sofia:1 turian:1 en:31 slow:1 sub:1 wish:1 fortunate:1 suntec:1 learns:1 nltk:1 specific:2 list:2 experimented:1 frederic:1 evidence:1 glorot:1 incorporating:1 essential:1 toutanova:3 merging:2 socher:2 importance:1 workshop:1 illustrates:1 margin:3 entropy:2 logarithmic:1 simply:4 explore:3 gao:2 scalar:1 corresponds:2 extracted:6 weston:1 careful:1 toword:1 crl:1 content:1 considerable:1 markt:1 specifically:8 reducing:1 semantically:1 wt:1 experimental:1 e:3 attempted:1 meaningful:4 internal:3 people:1 distrib:1 brevity:1 alexander:1 indian:1 evaluate:2 scratch:1 europarl:5 |
4,717 | 5,271 | Pre-training of Recurrent Neural Networks via
Linear Autoencoders
Luca Pasa, Alessandro Sperduti
Department of Mathematics
University of Padova, Italy
{pasa,sperduti}@math.unipd.it
Abstract
We propose a pre-training technique for recurrent neural networks based on linear
autoencoder networks for sequences, i.e. linear dynamical systems modelling the
target sequences. We start by giving a closed form solution for the definition of
the optimal weights of a linear autoencoder given a training set of sequences. This
solution, however, is computationally very demanding, so we suggest a procedure
to get an approximate solution for a given number of hidden units. The weights
obtained for the linear autoencoder are then used as initial weights for the inputto-hidden connections of a recurrent neural network, which is then trained on the
desired task. Using four well known datasets of sequences of polyphonic music,
we show that the proposed pre-training approach is highly effective, since it allows
to largely improve the state of the art results on all the considered datasets.
1
Introduction
Recurrent Neural Networks (RNN) constitute a powerful computational tool for sequences modelling and prediction [1]. However, training a RNN is not an easy task, mainly because of the well
known vanishing gradient problem which makes difficult to learn long-term dependencies [2]. Although alternative architectures, e.g. LSTM networks [3], and more efficient training procedures,
such as Hessian Free Optimization [4], have been proposed to circumvent this problem, reliable and
effective training of RNNs is still an open problem.
The vanishing gradient problem is also an obstacle to Deep Learning, e.g., [5, 6, 7]. In that context,
there is a growing evidence that effective learning should be based on relevant and robust internal
representations developed in autonomy by the learning system. This is usually achieved in vectorial
spaces by exploiting nonlinear autoencoder networks to learn rich internal representations of input
data which are then used as input to shallow neural classifiers or predictors (see, for example, [8]).
The importance to start gradient-based learning from a good initial point in the parameter space has
also been pointed out in [9]. Relationship between autoencoder networks and Principal Component
Analysis (PCA) [10] is well known since late ?80s, especially in the case of linear hidden units [11,
12]. More recently, linear autoencoder networks for structured data have been studied in [13, 14, 15],
where an exact closed-form solution for the weights is given in the case of a number of hidden units
equal to the rank of the full data matrix.
In this paper, we borrow the conceptual framework presented in [13, 16] to devise an effective pretraining approach, based on linear autoencoder networks for sequences, to get a good starting point
into the weight space of a RNN, which can then be successfully trained even in presence of longterm dependencies. Specifically, we revise the theoretical approach presented in [13] by: i) giving
a simpler and direct solution to the problem of devising an exact closed-form solution (full rank
case) for the weights of a linear autoencoder network for sequences, highlighting the relationship
between the proposed solution and PCA of the input data; ii) introducing a new formulation of
1
the autoencoder learning problem able to return an optimal solution also in the case of a number
of hidden units which is less than the rank of the full data matrix; iii) proposing a procedure for
approximate learning of the autoencoder network weights under the scenario of very large sequence
datasets. More importantly, we show how to use the linear autoencoder network solution to derive a
good initial point into a RNN weight space, and how the proposed approach is able to return quite
impressive results when applied to prediction tasks involving long sequences of polyphonic music.
2
Linear Autoencoder Networks for Sequences
In [11, 12] it is shown that principal directions of a set of vectors xi ? Rk are related to solutions
obtained by training linear autoencoder networks
oi = Woutput Whidden xi , i = 1, . . . , n,
(1)
where Whidden ? Rp?k , Woutput ? Rk?p , p k, and the network is trained so to get oi = xi , ?i.
When considering a temporal sequence x1 , x2 , . . . , xt , . . . of input vectors, where t is a discrete time
index, a linear autoencoder can be defined by considering the coupled linear dynamical systems
xt
= Cyt
(3)
yt = Axt + Byt?1 (2)
yt?1
It should be noticed that eqs. (2) and (3) extend the linear transformation defined in eq. (1) by
introducing a memory term involving matrix B ? Rp?p . In fact, yt?1 is inserted in the right part
of equation (2) to keep track of the input history through time: this is done exploiting a state space
representation. Eq. (3) represents the decoding part of the autoencoder: when a state yt is multiplied
by C, the observed input xt at time t and state at time t ? 1, i.e. yt?1 , are generated. Decoding
can then continue from yt?1 . This formulation has been proposed, for example, in [17] where an
iterative procedure to learn weight matrices A and B, based on Oja?s rule, is presented. No proof
of convergence for the proposed procedure is however given. More recently, an exact closed-form
solution for the weights has been given in the case of a number of hidden units equal to the rank of
the full data matrix (full rank case) [13, 16]. In this section, we revise this result. In addition, we
give an exact solution also for the case in which the number of hidden units is strictly less than the
rank of the full data matrix.
The basic idea of [13, 16] is to look for directions of high variance into the state space of the
dynamical linear system (2). Let start by considering a single sequence x1 , x2 , . . . , xt , . . . , xn and
the state vectors of the corresponding induced state sequence collected as rows of a matrix Y =
T
[y1 , y2 , y3 , ? ? ? , yn ] . By using the initial condition y0 = 0 (the null vector), and the dynamical
linear system (2), we can rewrite the Y matrix as
? T
?? T
?
A
x1 0
0
0
??? 0
? xT xT
?
? AT BT
0
0
??? 0 ?
2
1
? T
? ? T 2T
?
T
T
?
?
? x3 x2
?
x
0
?
?
?
0
A
B
1
Y=?
??
?
? ..
?
..
..
..
..
.. ? ?
..
? .
?
?
?
.
.
.
.
.
.
T
T
T
T
T
T
T
n?1
xn xn?1 xn?2 ? ? ? x2 x1
A B
|
{z
}|
{z
}
?
?
n?s
where, given s = kn, ? ? R
is a data matrix collecting all the (inverted) input subsequences
(including the whole sequence) as rows, and ? is the parameter matrix of the dynamical system.
Now, we are interested in using a state space of dimension p n, i.e. yt ? Rp , such that as
much information as contained in ? is preserved. We start by factorizing ? using SVD, obtaining
? = V?UT where V ? Rn?n is an unitary matrix, ? ? Rn?s is a rectangular diagonal matrix
with nonnegative real numbers on the diagonal with ?1,1 ? ?2,2 ? ? ? ? ? ?n,n (the singular values),
and UT ? Rs?n is a unitary matrix.
It is important to notice that columns of UT which correspond to nonzero singular values, apart
some mathematical technicalities, basically correspond to the principal directions of data, i.e. PCA.
If the rank of ? is p, then only the first p elements of the diagonal of ? are not null, and the
T
above decomposition can be reduced to ? = V(p) ?(p) U(p) where V(p) ? Rn?p , ?(p) ? Rp?p ,
2
T
T
and U(p) ? Rp?n . Now we can observe that U(p) U(p) = I (where I is the identity matrix of
dimension p), since by definition the columns of U(p) are orthogonal, and by imposing ? = U(p) ,
we can derive ?optimal? matrices A ? Rp?k and B ? Rp?p for our dynamical system, which will
T
have corresponding state space matrix Y(p) = ?? = ?U(p) = V(p) ?(p) U(p) U(p) = V(p) ?(p) .
(p)
Thus, if we represent U(p) as composed of n submatrices Ui , each of size k ? p, the problem
reduces to find matrices A and B such that
? T
? ? (p) ?
A
U1
(p) ?
? AT BT
? ?
U2 ?
? T 2T
? ?
(p) ?
? A B
? ?
(p)
U3 ?
(4)
?=?
?=?
?=U .
?
? ?
..
.
?
?
? ?
?
.. ?
.
T
(p)
AT Bn?1
Un
The reason to impose ? = U(p) is to get a state space where the coordinates are uncorrelated so
to diagonalise the empirical sample covariance matrix of the states. Please, note that in this way
each state (i.e., row of the Y matrix) corresponds to a row of the data matrix ?, i.e. the unrolled
(sub)sequence read up to a given time t. If the rows of ? were vectors, this would correspond to
compute PCA, keeping only the fist p principal directions.
In the following, we demonstrate that there exists a solution to the above equation. We start
by observing that ? owns a special structure, i.e. given? = [?1 ?2 ? ? ? ?n ], where
?i ?
0
0
1?1
1?(n?1)
Rn?k , then for i = 1, . . . , n ? 1, ?i+1 = Rn ?i =
?i , and
I(n?1)?(n?1) 0(n?1)?1
Rn ?n = 0, i.e. the null matrix of size n ? k. Moreover, by singular value decomposition, we
(p) T
T
have ?i = V(p) ?(p) Ui , for i = 1, . . . , n. Using the fact that V(p) V(p) = I, and
(p)
(p)
combining the above equations, we get Ui+t = Ui Qt , for i = 1, . . . , n ? 1, and t =
?1
T
(p)
(p) (p)
1, . . . , n ? i, where Q = ?(p) V(p) RT
?
. Moreover, we have that Un Q = 0 since
nV
?1
(p) (p) (p) T T (p) (p) ?1
(p)
= (Rn ?n )T V(p) ?(p) . Thus, eq. (4) is satisfied by
Rn V ?
Un Q = Un ? V
| {z }
=0
(p) T
U1
T
A =
and B = Q . It is interesting to note that the original data ? can be recovered by
T
T
computing Y(p) U(p) = V(p) ?(p) U(p) = ?, which can be achieved by running the system
xt
AT
yt
=
yt?1
BT
AT
starting from yn , i.e.
is the matrix C defined in eq. (3).
BT
Finally, it is important to remark that the above construction works not only for a single sequence,
but also for a set of sequences of different length. For example, let consider the two sequences
(xa1 , xa2 , xa3 ) and (xb 1 , xb 2 ). Then, we have
? aT
?
"
#
x1
0
0
bT
x
0
1
? and ?b =
?a = ? xa2 T xa1 T 0
T
bT
x
xb1
aT
aT
aT
2
x3
x2
x1
?a
R4
, and R =
.
which can be collected together to obtain ? =
?b 02?1
R2 02?1
As a final remark, it should be stressed that the above construction only works if p is equal to the
rank of ?. In the next section, we treat the case in which p < rank(?).
2.1
Optimal solution for low dimensional autoencoders
T
? i = V(p) L(p) U(p) 6= ?i , and
When p < rank(?) the solution given above breaks down because ?
i
? i+1 6= Rn ?
? i . So the question is whether the proposed solutions for A and B still
consequently ?
hold the best reconstruction error when p < rank(?).
3
In this paper, we answer in negative terms to this question by resorting to a new formulation of our
(p)
problem where we introduce slack-like matrices Ei ? Rk?p , i = 1, . . . , n + 1 collecting the
reconstruction errors, which need to be minimised:
n+1
X
min
(p)
Q?Rp?p ,Ei
?
subject to :
?
?
?
?
?
?
?
(p)
i=1
(p)
U1 + E 1
(p)
(p)
U2 + E 2
(p)
(p)
U3 + E 3
..
.
(p)
(p)
kEi k2F
(p)
Un + E n
(p)
(p)
?
?
?
?
?
?
?
?Q = ?
?
?
? (p)
?
? Un + En(p)
?
(p)
En+1
?
?
?
?
?
?
?
?
?
U2 + E 2
(p)
(p)
U3 + E 3
..
.
(5)
Notice that the problem above is convex both in the objective function and in the constraints; thus
(p)
it only has global optimal solutions E?i and Q? , from which we can derive AT = U1 + E?1 and
T
?
T
(p)
(p)
B = Q . Specifically, when p = rank(?), Rs,k U is in the span of U and the optimal
T
(p)
solution is given by E?i = 0k?p ?i, and Q? = U(p) RT
, i.e. the solution we have already
s,k U
described. If p < rank(?), the optimal solution cannot have ?i, E?i = 0k?p . However, it is not
difficult to devise an iterative procedure to reach the minimum. Since in the experimental section we
do not exploit the solution to this problem for reasons that we will explain later, here we just sketch
(p)
such procedure. It helps to observe that, given a fixed Q, the optimal solution for Ei is given by
(p)
(p)
(p)
(p)
(p)
(p)
(p)
(p)
(p)
+
2
3
? ,E
? ,...,E
?
[E
1
2
n+1 ] = [U1 Q ? U2 , U1 Q ? U3 , U1 Q ? U4 , . . .] MQ
?
?
?Q ?Q2 ?Q3 ? ? ?
0
0
??? ?
? I
? 0
+
I
0
??? ?
?
?.
where MQ is the pseudo inverse of MQ = ?
0
I
??? ?
? 0
?
..
..
..
..
.
.
.
.
h
i
T
T
T
T T
? (p) = E
? (p) , E
? (p) , E
? (p) , ? ? ? , E
? n(p)
In general, E
can be decomposed into a component in the
1
2
3
?
?
span of U(p) and a component E(p) orthogonal to it. Notice that E(p) cannot be reduced, while
? (p) = U(p) + E(p) ? and taking
(part of) the other component can be absorbed into Q by defining U
h
i
T
T
T T
? = (U
? (p) )+ U
? (p) , U
? (p) , ? ? ? , U
? (p)T , E(p)
Q
.
n
2
3
n+1
? the new optimal values for E(p) are obtained and the process iterated till convergence.
Given Q,
i
3
Pre-training of Recurrent Neural Networks
Here we define our pre-training procedure for recurrent neural networks with one hidden layer of p
units, and O output units:
ot = ?(Woutput h(xt )) ? RO , h(xt ) = ?(Winput xt + Whidden h(xt?1 )) ? Rp
(6)
T
where Woutput ? RO?p , Whidden ? Rp?k , for a vector z ? Rm , ?(z) = [?(z1 ), . . . , ?(zm )] ,
?zi
.
and here we consider the symmetric sigmoid function ?(zi ) = 1?e
1+e?zi
The idea is to exploit the hidden state representation obtained by eqs. (2) as initial hidden state representation for the RNN described by eqs. (6). This is implemented by initialising the weight matrices
Winput and Whidden of (6) by using the matrices that jointly solve eqs. (2) and eqs. (3), i.e. A and
B (since C is function of A and B). Specifically, we initialize Winput with A, and Whidden with
B. Moreover, the use of symmetrical sigmoidal functions, which do give a very good approximation
of the identity function around the origin, allows a good transferring of the linear dynamics inside
4
RNN. For what concerns Woutput , we initialise it by using the best possible solution, i.e. the pseudoinverse of H times the target matrix T, which does minimise the output squared error. Learning
is then used to introduce nonlinear components that allow to improve the performance of the model.
More formally, let consider a prediction task where for each sequence sq ? (xq1 , xq2 , . . . , xqlq )
of length lq in the training set, a sequence tq of target vectors is defined, i.e. a training sequence is given by hsq , tq i ? h(xq1 , tq1 ), (xq2 , tq2 ), . . . , (xqlq , tqlq )i, where tqi ? RO . Given a trainPN
ing set with N sequences, let define the target matrix T ? RL?O , where L =
q=1 lq , as
1 1
1
2
?
N T
T = t1 , t2 , . . . , tl1 , t1 , . . . , tlN . The input matrix ? will have size L ? k. Let p be the desired number of hidden units for the recurrent neural network (RNN). Then the pre-training procedure can be defined as follows: i) compute the linear autoencoder for ? using p? principal direc?
?
?
tions, obtaining the optimal matrices A? ? Rp ?k and B? ? Rp ?p ; i) set Winput = A? and
?
Whidden = B ; iii) run the RNN over the training sequences, collecting the hidden activities vec?
tors (computed using symmetrical sigmoidal functions) over time as rows of matrix H ? RL?p ;
+
+
iv) set Woutput = H T, where H is the (left) pseudoinverse of H.
3.1
Computing an approximate solution for large datasets
In real world scenarios the application of our approach may turn difficult because of the size of
the data matrix. In fact, stable computation of principal directions is usually obtained by SVD decomposition of the data matrix ?, that in typical application domains involves a number of rows
and columns which is easily of the order of hundreds of thousands. Unfortunately, the computational complexity of SVD decomposition is basically cubic in the smallest of the matrix dimensions.
Memory consumption is also an important issue. Algorithms for approximate computation of SVD
have been suggested (e.g., [18]), however, since for our purposes we just need matrices V and ?
with a predefined number of columns (i.e. p), here we present an ad-hoc algorithm for approximate
computation of these matrices. Our solution is based on the following four main ideas: i) divide ?
in slices of k (i.e., size of input at time t) columns, so to exploit SVD decomposition at each slice
separately; ii) compute approximate V and ? matrices, with p columns, incrementally via truncated
SVD of temporary matrices obtained by concatenating the current approximation of V? with a new
slice; iii) compute the SVD decomposition of a temporary matrix via either its kernel or covariance
matrix, depending on the smallest between the number of rows and the number of columns of the
temporary matrix; iv) exploit QR decomposition to compute SVD decomposition.
Algorithm 1 shows in pseudo-code the main steps of our procedure. It maintains a temporary matrix
T which is used to collect incrementally an approximation of the principal subspace of dimension p
of ?. Initially (line 4) T is set equal to the last slices of ?, in a number sufficient to get a number
of columns larger than p (line 2). Matrices V and ? from the p-truncated SVD decomposition of
T are computed (line 5) via the K E C O procedure, described in Algorithm 2, and used to define a
new T matrix by concatenation with the last unused slice of ?. When all slices are processed, the
current V and ? matrices are returned. The K E C O procedure, described in Algorithm 2 , reduces
the computational burden by computing the p-truncated SVD decomposition of the input matrix
M via its kernel matrix (lines 3-4) if the number of rows of M is no larger than the number of
columns, otherwise the covariance matrix is used (lines 6-8). In both cases, the p-truncated SVD
decomposition is implemented via QR decomposition by the INDIRECT SVD procedure described in
Algorithm 3. This allows to reduce computation time when large matrices must be processed [19].
1
Finally, matrices V and S 2 (both kernel and covariance matrices have squared singular values of
M) are returned.
We use the strategy to process slices of ? in reverse order since, moving versus columns with larger
indices, the rank as well as the norm of slices become smaller and smaller, thus giving less and less
contribution to the principal subspace of dimension p. This should reduce the approximation error
cumulated by dropping the components from p + 1 to p + k during computation [20]. As a final
remark, we stress that since we compute an approximate solution for the principal directions of ?,
it makes no much sense to solve the problem given in eq. (5): learning will quickly compensate
for the approximations and/or sub-optimality of A and B obtained by matrices V and ? returned
by Algorithm 1. Thus, these are the matrices we have used for the experiments described in next
section.
5
Algorithm 1 Approximated V and ? with p components
1: function SVF OR B IG DATA(?, k, p)
2:
nStart = dp/ke
. Number of starting slices
3:
nSlice = (?.columns/k) ? nStart
. Number of remaining slices
4:
T = ?[:, k ? nSlice : ?.columns]
5:
V, ? =K E C O(T, p)
. Computation of V and ? for starting slices
6:
for i in REVERSED(range(nSlice)) do
. Computation of V and ? for remaining slices
7:
T = [?[:, i ? k:(i + 1) ? k], V?]
8:
V, ? =K E C O(T, p)
9:
end for
10:
return V, ?
11: end function
Algorithm 2 Kernel vs covariance computation Algorithm 3 Truncated SVD by QR
1: function K E C O(M, p)
1: function INDIRECT SVD(M, p)
2:
if M.rows <= ?.columns then
2:
Q, R =QR(M)
3:
K = MMT
3:
Vr , S, UT =SVD(R)
T
4:
V, Ssqr , U =INDIRECT SVD(K, p) 4:
V = QVr
5:
else
5:
S = S[1 : p, 1 : p]
6:
C = MT M
6:
V = V[1 : p, :]
7:
V, Ssqr , UT =INDIRECT SVD(C, p) 7:
UT = UT [:, 1 : p]
1
?
8:
return V, S, UT
8:
V = MUT Ssqr2
9: end function
9:
end if
1
2
10:
return V, Ssqr
11: end function
4
Experiments
In order to evaluate our pre-training approach, we decided to use the four polyphonic music sequences datasets used in [21] for assessing the prediction abilities of the RNN-RBM model. The
prediction task consists in predicting the notes played at time t given the sequence of notes played
till time t ? 1. The RNN-RBM model achieves state-of-the-art in such demanding prediction task.
As performance measure we adopted the accuracy measure used in [21] and described in [22]. Each
dataset is split in training set, validation set, and test set. Statistics on the datasets, including largest
sequence length, are given in columns 2-4 of Table 1. Each sequence in the dataset represents a song
having a maximum polyphony of 15 notes (average 3.9); each time step input spans the whole range
of piano from A0 to C8 and it is represented by using 88 binary values (i.e. k = 88).
Our pre-training approach (PreT-RNN) has been assessed by using a different number of hidden
units (i.e., p is set in turn to 50, 100, 150, 200, 250) and 5000 epochs of RNN training1 using the
Theano-based stochastic gradient descent software available at [23].
Random initialisation (Rnd) has also been used for networks with the same number of hidden units.
Specifically, for networks with 50 hidden units, we have evaluated the performance of 6 different
random initialisations. Finally, in order to verify that the nonlinearity introduced by the RNN is
actually useful to solve the prediction task, we have also evaluated the performance of a network
with linear units (250 hidden units) initialised with our pre-training procedure (PreT-Lin250).
To give an idea of the time performance of pre-training with respect to the training of a RNN, in
column 5 of Table 1 we have reported the time in seconds needed to compute pre-training matrices
c
c
(Pre-) (on Intel
Xeon
CPU E5-2670 @2.60GHz with 128 GB) and to perform training of a
RNN with p = 50 for 5000 epochs (on GPU NVidia K20). Please, note that for larger values of p,
the increase in computation time of pre-training is smaller than the increment in computation time
needed for training a RNN.
1
Due to early overfitting, for the Muse dataset we used 1000 epochs.
6
Dataset
Nottingham
Piano-midi.de
MuseData
JSB Chorales
Set
Training
(39165 ? 56408)
Test
Validation
Training
(70672 ? 387640)
Test
Validation
Training
(248479 ? 214192)
Test
Validation
Training
(27674 ? 22792)
Test
Validation
# Samples
195
Max length
641
170
173
87
1495
1229
4405
25
12
524
2305
1740
2434
25
135
229
2305
2523
259
77
76
320
289
(Pre-)Training Time
seconds
(226) 5837
p = 50
5000 epochs
seconds
(2971) 4147
p = 50
5000 epochs
seconds
(7338) 4190
p = 50
5000 epochs
seconds
(79) 6411
p = 50
5000 epochs
Model
RNN (w. HF)
RNN-RBM
PreT-RNN
PreT-Lin250
RNN (w. HF)
RNN-RBM
PreT-RNN
PreT-Lin250
RNN (w. HF)
RNN-RBM
PreT-RNN
PreT-Lin250
RNN (w. HF)
RNN-RBM
PreT-RNN
PreT-Lin250
ACC% [21]
62.93 (66.64)
75.40
75.23 (p = 250)
73.19
19.33 (23.34)
28.92
37.74 (p = 250)
16.87
23.25 (30.49)
34.02
57.57 (p = 200)
3.56
28.46 (29.41)
33.12
65.67 (p = 250)
38.32
Table 1: Datasets statistics including data matrix size for the training set (columns 2-4), computational times in seconds to perform pre-training and training for 5000 epochs with p = 50 (column
5), and accuracy results for state-of-the-art models [21] vs our pre-training approach (columns 6-7).
The acronym (w. HF) is used to identify an RNN trained by Hessian Free Optimization [4].
Training and test curves for all the models described above are reported in Figure 1. It is evident that
random initialisation does not allow the RNN to improve its performance in a reasonable amount of
epochs. Specifically, for random initialisation with p = 50 (Rnd 50), we have reported the average
and range of variation over the 6 different trails: different initial points do not change substantially
the performance of RNN. Increasing the number of hidden units allows the RNN to slightly increase
its performance. Using pre-training, on the other hand, allows the RNN to start training from a quite
favourable point, as demonstrated by an early sharp improvement of performances. Moreover, the
more hidden units are used, the more the improvement in performance is obtained, till overfitting is
observed. In particular, early overfitting occurs for the Muse dataset. It can be noticed that the linear
model (Linear) reaches performances which are in some cases better than RNN without pre-training.
However, it is important to notice that while it achieves good results on the training set (e.g. JSB and
Piano-midi), the corresponding performance on the test set is poor, showing a clear evidence of overfitting. Finally, in column 7 of Table 1, we have reported the accuracy obtained after validation on
the number of hidden units and number of epochs for our approaches (PreT-RNN and PreT-Lin250)
versus the results reported in [21] for RNN (also using Hessian Free Optimization) and RNN-RBM.
In any case, the use of pre-training largely improves the performances over standard RNN (with
or without Hessian Free Optimization). Moreover, with the exception of the Nottingham dataset,
the proposed approach outperforms the state-of-the-art results achieved by RNN-RBM. Large improvements are observed for the Muse and JSB datasets. Performance for the Nottingham dataset
is basically equivalent to the one obtained by RNN-RBM. For this dataset, also the linear model
with pre-training achieves quite good results, which seems to suggest that the prediction task for
this dataset is much easier than for the other datasets. The linear model outperforms RNN without
pre-training on Nottingham and JSB datasets, but shows problems with the Muse dataset.
5
Conclusions
We have proposed a pre-training technique for RNN based on linear autoencoders for sequences.
For this kind of autoencoders it is possible to give a closed form solution for the definition of the
?optimal? weights, which however, entails the computation of the SVD decomposition of the full
data matrix. For large data matrices exact SVD decomposition cannot be achieved, so we proposed
a computationally efficient procedure to get an approximation that turned to be effective for our
goals. Experimental results for a prediction task on datasets of sequences of polyphonic music
show the usefulness of the proposed pre-training approach, since it allows to largely improve the
state of the art results on all the considered datasets by using simple stochastic gradient descend for
learning. Even if the results are very encouraging the method needs to be assessed on data from
other application domains. Moreover, it is interesting to understand whether the analysis performed
in [24] on linear deep networks for vectors can be extended to recurrent architectures for sequences
and, in particular, to our method.
7
0.4
0.2
0.1
0
-0.1
Rnd 50 (6 trials)
Linear 250
Rnd 100
0
200
600
PreT 200
PreT 250
800
Nottingham
Test Set
1000
0.8
Epoch
0.7
0.7
0.6
0.6
0.5
0.5
Accuracy
Accuracy
PreT 50
PreT 150
PreT 100
400
Nottingham Training
Set
0.8
0.4
0.3
0.4
0.3
0.2
0.2
0.1
0.1
0
0
0
500 1000 1500 2000 2500 3000 3500 4000 4500 5000
Epoch
0
500 1000 1500 2000 2500 3000 3500 4000 4500 5000
Epoch
Piano-Midi.de Training Set
Piano-Midi.de Test Set
0.55
0.4
0.5
0.35
Accuracy
0.45
0.4
0.3
0.35
0.25
0.3
0.25
0.2
0.15
0.2
0.15
0.1
0.1
0.05
0.05
0
0
0
500 1000 1500 2000 2500 3000 3500 4000 4500 5000
Epoch
0
500 1000 1500 2000 2500 3000 3500 4000 4500 5000
Epoch
Muse Dataset Training Set
Muse Dataset Test Set
0.7
0.6
0.6
0.5
0.5
0.4
Accuracy
Accuracy
0.4
0.3
0.3
0.2
0.2
0.1
0.1
0
0
0
200
400
600
800
1000
0
Epoch
200
400
600
800
1000
Epoch
JSB Chorales Training Set
JSB Chorales Test Set
0.8
0.7
0.7
0.6
0.6
0.5
0.5
Accuracy
-0.3
Rnd 150
Rnd 200
Rnd 250
Accuracy
-0.2
Accuracy
Accuracy
0.3
0.4
0.3
0.4
0.3
0.2
0.2
0.1
0.1
0
0
0
500 1000 1500 2000 2500 3000 3500 4000 4500 5000
Epoch
0
500 1000 1500 2000 2500 3000 3500 4000 4500 5000
Epoch
Figure 1: Training (left column) and test (right column) curves for the assessed approaches on the
four datasets. Curves are sampled at each epoch till epoch 100, and at steps of 100 epochs afterwards.
8
References
[1] S. C. Kremer. Field Guide to Dynamical Recurrent Networks. Wiley-IEEE Press, 2001.
[2] Y. Bengio, P. Simard, and P. Frasconi. Learning long-term dependencies with gradient descent
is difficult. IEEE Transactions on Neural Networks, 5(2):157?166, 1994.
[3] S. Hochreiter and J. Schmidhuber. Lstm can solve hard long time lag problems. In NIPS, pages
473?479, 1996.
[4] J. Martens and I. Sutskever. Learning recurrent neural networks with hessian-free optimization.
In ICML, pages 1033?1040, 2011.
[5] G. E. Hinton and R. R. Salakhutdinov. Reducing the dimensionality of data with neural networks. Science, 313(5786):504?507, July 2006.
[6] G. E. Hinton, S. Osindero, and Y. W. Teh. A fast learning algorithm for deep belief nets. Neural
Computation, 18(7):1527?1554, 2006.
[7] P. di Lena, K. Nagata, and P. Baldi. Deep architectures for protein contact map prediction.
Bioinformatics, 28(19):2449?2457, 2012.
[8] Y. Bengio. Learning deep architectures for ai. Foundations and Trends in Machine Learning,
2(1):1?127, 2009.
[9] I. Sutskever, J. Martens, G. E. Dahl, and G. E. Hinton. On the importance of initialization and
momentum in deep learning. In ICML (3), pages 1139?1147, 2013.
[10] I.T. Jolliffe. Principal Component Analysis. Springer-Verlag New York, Inc., 2002.
[11] H. Bourlard and Y. Kamp. Auto-association by multilayer perceptrons and singular value
decomposition. Biological Cybernetics, 59(4-5):291?294, 1988.
[12] P. Baldi and K. Hornik. Neural networks and principal component analysis: Learning from
examples without local minima. Neural Networks, 2(1):53?58, 1989.
[13] A. Sperduti. Exact solutions for recursive principal components analysis of sequences and
trees. In ICANN (1), pages 349?356, 2006.
[14] A. Micheli and A. Sperduti. Recursive principal component analysis of graphs. In ICANN (2),
pages 826?835, 2007.
[15] A. Sperduti. Efficient computation of recursive principal component analysis for structured
input. In ECML, pages 335?346, 2007.
[16] A. Sperduti. Linear autoencoder networks for structured data. In NeSy?13:Ninth International
Workshop onNeural-Symbolic Learning and Reasoning, 2013.
[17] T. Voegtlin. Recursive principal components analysis. Neural Netw., 18(8):1051?1063, 2005.
[18] G. Martinsson et al. Randomized methods for computing the singular value decomposition
(svd) of very large matrices. In Works. on Alg. for Modern Mass. Data Sets, Palo Alto, 2010.
[19] E. Rabani and S. Toledo. Out-of-core svd and qr decompositions. In PPSC, 2001.
[20] Z. Zhang and H. Zha. Structure and perturbation analysis of truncated svds for columnpartitioned matrices. SIAM J. on Mat. Anal. and Appl., 22(4):1245?1262, 2001.
[21] N. Boulanger-Lewandowski, Y. Bengio, and P. Vincent. Modeling temporal dependencies in
high-dimensional sequences: Application to polyphonic music generation and transcription.
In ICML, 2012.
[22] M. Bay, A. F. Ehmann, and J. S. Downie. Evaluation of multiple-f0 estimation and tracking
systems. ISMIR, pages 315?320, 2009.
[23] https://github.com/gwtaylor/theano-rnn.
[24] A. M. Saxe, J. L. McClelland, and S. Ganguli. Exact solutions to the nonlinear dynamics of
learning in deep linear neural networks. arXiv preprint arXiv:1312.6120, 2013.
9
| 5271 |@word trial:1 longterm:1 norm:1 seems:1 open:1 r:2 bn:1 decomposition:17 covariance:5 initial:6 initialisation:4 outperforms:2 recovered:1 current:2 com:1 must:1 gpu:1 xb1:1 polyphonic:5 v:2 devising:1 pasa:2 vanishing:2 core:1 math:1 sigmoidal:2 simpler:1 zhang:1 mathematical:1 direct:1 become:1 consists:1 inside:1 baldi:2 introduce:2 growing:1 lena:1 salakhutdinov:1 decomposed:1 cpu:1 encouraging:1 considering:3 increasing:1 moreover:6 alto:1 mass:1 null:3 what:1 kind:1 substantially:1 q2:1 developed:1 proposing:1 transformation:1 temporal:2 pseudo:2 y3:1 collecting:3 axt:1 ro:3 classifier:1 rm:1 unit:17 yn:2 t1:2 local:1 treat:1 rnns:1 initialization:1 studied:1 r4:1 collect:1 appl:1 range:3 decided:1 recursive:4 x3:2 sq:1 procedure:15 empirical:1 rnn:43 submatrices:1 pre:23 protein:1 suggest:2 get:7 cannot:3 symbolic:1 context:1 equivalent:1 map:1 demonstrated:1 yt:9 marten:2 xa1:2 starting:4 convex:1 rectangular:1 ke:1 rule:1 importantly:1 borrow:1 k20:1 lewandowski:1 mq:3 initialise:1 coordinate:1 increment:1 variation:1 target:4 construction:2 exact:7 trail:1 unipd:1 origin:1 element:1 trend:1 approximated:1 jsb:6 u4:1 observed:3 inserted:1 preprint:1 descend:1 thousand:1 svds:1 alessandro:1 ui:4 complexity:1 dynamic:2 trained:4 rewrite:1 easily:1 indirect:4 represented:1 fast:1 effective:5 quite:3 lag:1 larger:4 solve:4 otherwise:1 ability:1 statistic:2 jointly:1 final:2 hoc:1 sequence:32 net:1 propose:1 reconstruction:2 zm:1 relevant:1 combining:1 turned:1 till:4 qr:5 exploiting:2 convergence:2 sutskever:2 assessing:1 downie:1 help:1 derive:3 recurrent:10 tions:1 depending:1 qt:1 eq:10 implemented:2 involves:1 direction:6 stochastic:2 saxe:1 tln:1 biological:1 strictly:1 voegtlin:1 hold:1 around:1 considered:2 u3:4 tor:1 achieves:3 smallest:2 tq1:1 early:3 purpose:1 estimation:1 palo:1 largest:1 successfully:1 tool:1 cyt:1 q3:1 improvement:3 modelling:2 rank:14 mainly:1 sense:1 ganguli:1 bt:6 transferring:1 a0:1 initially:1 hidden:19 interested:1 issue:1 art:5 special:1 initialize:1 equal:4 field:1 having:1 frasconi:1 represents:2 look:1 k2f:1 icml:3 tl1:1 t2:1 modern:1 oja:1 composed:1 mut:1 tq:2 highly:1 evaluation:1 xb:2 predefined:1 orthogonal:2 tree:1 iv:2 divide:1 sperduti:6 desired:2 theoretical:1 column:21 xeon:1 obstacle:1 modeling:1 introducing:2 predictor:1 hundred:1 usefulness:1 osindero:1 reported:5 dependency:4 kn:1 answer:1 lstm:2 international:1 randomized:1 siam:1 decoding:2 minimised:1 together:1 quickly:1 squared:2 satisfied:1 pret:17 simard:1 return:5 de:3 inc:1 ad:1 later:1 break:1 performed:1 closed:5 observing:1 start:6 hf:5 maintains:1 nagata:1 zha:1 contribution:1 oi:2 accuracy:12 variance:1 largely:3 correspond:3 identify:1 kamp:1 vincent:1 iterated:1 basically:3 cybernetics:1 history:1 acc:1 explain:1 reach:2 definition:3 initialised:1 proof:1 rbm:9 di:1 sampled:1 dataset:12 revise:2 ut:8 improves:1 dimensionality:1 woutput:6 actually:1 formulation:3 done:1 evaluated:2 just:2 nottingham:6 autoencoders:4 sketch:1 hand:1 ei:3 nonlinear:3 incrementally:2 xa2:2 verify:1 y2:1 read:1 symmetric:1 nonzero:1 during:1 please:2 stress:1 evident:1 demonstrate:1 reasoning:1 recently:2 sigmoid:1 mt:1 rl:2 extend:1 association:1 martinsson:1 imposing:1 vec:1 ai:1 resorting:1 mathematics:1 pointed:1 nonlinearity:1 moving:1 stable:1 entail:1 impressive:1 f0:1 italy:1 apart:1 reverse:1 scenario:2 schmidhuber:1 nvidia:1 verlag:1 binary:1 continue:1 devise:2 inverted:1 minimum:2 impose:1 july:1 ii:2 fist:1 full:7 afterwards:1 multiple:1 reduces:2 ing:1 long:4 compensate:1 luca:1 prediction:10 involving:2 basic:1 multilayer:1 arxiv:2 represent:1 kernel:4 achieved:4 hochreiter:1 preserved:1 addition:1 separately:1 else:1 singular:6 ot:1 nv:1 induced:1 subject:1 unitary:2 presence:1 unused:1 svf:1 iii:3 easy:1 split:1 bengio:3 zi:3 architecture:4 reduce:2 idea:4 minimise:1 whether:2 pca:4 gb:1 song:1 returned:3 hessian:5 york:1 constitute:1 pretraining:1 remark:3 deep:7 useful:1 clear:1 amount:1 processed:2 mcclelland:1 reduced:2 http:1 notice:4 track:1 discrete:1 dropping:1 mat:1 four:4 dahl:1 graph:1 run:1 inverse:1 powerful:1 reasonable:1 ismir:1 initialising:1 layer:1 played:2 nonnegative:1 activity:1 vectorial:1 constraint:1 x2:5 software:1 u1:7 rabani:1 min:1 span:3 optimality:1 c8:1 department:1 structured:3 poor:1 smaller:3 slightly:1 y0:1 shallow:1 theano:2 computationally:2 equation:3 slack:1 turn:2 jolliffe:1 needed:2 xq1:2 end:5 acronym:1 adopted:1 available:1 multiplied:1 observe:2 alternative:1 rp:12 original:1 running:1 remaining:2 muse:6 music:5 exploit:4 giving:3 especially:1 boulanger:1 contact:1 objective:1 noticed:2 question:2 already:1 occurs:1 strategy:1 rt:2 diagonal:3 gradient:6 dp:1 subspace:2 reversed:1 micheli:1 concatenation:1 consumption:1 collected:2 reason:2 padova:1 length:4 code:1 index:2 relationship:2 unrolled:1 difficult:4 unfortunately:1 negative:1 anal:1 perform:2 teh:1 datasets:13 musedata:1 descent:2 ecml:1 truncated:6 defining:1 extended:1 hinton:3 y1:1 rn:9 perturbation:1 ninth:1 sharp:1 introduced:1 connection:1 z1:1 temporary:4 toledo:1 nip:1 able:2 suggested:1 dynamical:7 usually:2 reliable:1 memory:2 including:3 max:1 belief:1 demanding:2 circumvent:1 predicting:1 bourlard:1 chorale:3 improve:4 github:1 autoencoder:17 coupled:1 auto:1 epoch:22 piano:5 interesting:2 generation:1 versus:2 validation:6 foundation:1 sufficient:1 uncorrelated:1 autonomy:1 row:10 last:2 free:5 keeping:1 kremer:1 guide:1 allow:2 understand:1 taking:1 ghz:1 slice:12 curve:3 dimension:5 xn:4 world:1 rich:1 ig:1 kei:1 transaction:1 approximate:7 netw:1 midi:4 transcription:1 keep:1 technicality:1 global:1 pseudoinverse:2 overfitting:4 conceptual:1 owns:1 symmetrical:2 xi:3 factorizing:1 subsequence:1 iterative:2 un:6 bay:1 table:4 learn:3 polyphony:1 robust:1 obtaining:2 hornik:1 e5:1 alg:1 domain:2 icann:2 main:2 mmt:1 whole:2 x1:6 intel:1 en:2 byt:1 cubic:1 vr:1 wiley:1 sub:2 momentum:1 gwtaylor:1 lq:2 concatenating:1 late:1 rk:3 down:1 xa3:1 xt:11 showing:1 favourable:1 r2:1 evidence:2 concern:1 exists:1 burden:1 workshop:1 cumulated:1 importance:2 direc:1 easier:1 absorbed:1 highlighting:1 contained:1 tracking:1 u2:4 rnd:7 springer:1 corresponds:1 identity:2 goal:1 consequently:1 change:1 hard:1 specifically:5 typical:1 reducing:1 principal:15 svd:21 experimental:2 perceptrons:1 exception:1 formally:1 internal:2 stressed:1 assessed:3 bioinformatics:1 evaluate:1 |
4,718 | 5,272 | Using Convolutional Neural Networks
to Recognize Rhythm Stimuli
from Electroencephalography Recordings
Sebastian Stober, Daniel J. Cameron and Jessica A. Grahn
Brain and Mind Institute, Department of Psychology, Western University
London, Ontario, Canada, N6A 5B7
{sstober,dcamer25,jgrahn}@uwo.ca
Abstract
Electroencephalography (EEG) recordings of rhythm perception might contain enough
information to distinguish different rhythm types/genres or even identify the rhythms
themselves. We apply convolutional neural networks (CNNs) to analyze and classify
EEG data recorded within a rhythm perception study in Kigali, Rwanda which comprises
12 East African and 12 Western rhythmic stimuli ? each presented in a loop for 32
seconds to 13 participants. We investigate the impact of the data representation and
the pre-processing steps for this classification tasks and compare different network
structures. Using CNNs, we are able to recognize individual rhythms from the EEG
with a mean classification accuracy of 24.4% (chance level 4.17%) over all subjects by
looking at less than three seconds from a single channel. Aggregating predictions for
multiple channels, a mean accuracy of up to 50% can be achieved for individual subjects.
1 Introduction
Musical rhythm occurs in all human societies and is related to many phenomena, such as the perception of a
regular emphasis (i.e., beat), and the impulse to move one?s body. It is a universal human phenomenon, but
differs between human cultures. The influence of culture on the processing of rhythm in the brain as well
as the brain mechanisms underlying musical rhythm are still not fully understood. In order to study these,
we recruited participants in East Africa and North America to test their ability to perceive and produce
rhythms derived from East African and Western music. Besides several behavioral tasks, which have
already been discussed in [1], the East African participants also underwent electroencephalography (EEG)
recording while listening to East African and Western musical rhythms thus enabling us to study the neural
mechanisms underlying rhythm perception.
Using two popular deep learning techniques ? stacked denoising autoencoders (SDAs) [2] and convolutional
neural networks (CNNs) [3] ? we already obtained encouraging early results for distinguishing East
African and Western stimuli in a binary classification task based on the recorded EEG [4]. In this paper, we
address the much harder classification problem of recognizing the 24 individual rhythms. In the following,
we will review related work in Section 2, describe the data acquisition and pre-processing in Section 3,
present our experimental findings in Section 4, and discuss further steps in Section 5.
2 Related work
How the brain responses to auditory rhythms has already been investigated in several studies using EEG
and magnoencephalography (MEG): Oscillatory neural activity in the gamma (20-60 Hz) frequency band
is sensitive to accented tones in a rhythmic sequence and anticipates isochronous tones [5]. Oscillations
in the beta (20-30 Hz) band increase in anticipation of strong tones in a non-isochronous sequence [6, 7, 8].
1
Another approach has measured the magnitude of steady state evoked potentials (SSEPs) (reflecting neural
oscillations entrained to the stimulus) while listening to rhythmic sequences [9, 10]. Here, enhancement
of SSEPs was found for frequencies related to the metrical structure of the rhythm (e.g., the frequency
of the beat). In contrast to these studies investigating the oscillatory activity in the brain, other studies
have used EEG to investigate event-related potentials (ERPs) in responses to tones occurring in rhythmic
sequences. This approach has been used to show distinct sensitivity to perturbations of the rhythmic pattern
vs. the metrical structure in rhythmic sequences [11], and to suggest that similar responses persist even
when attention is diverted away from the rhythmic stimulus [12]. Further, Will and Berg [13] observed
a significant increase in brain wave synchronization after periodic auditory stimulation with drum sounds
and clicks with repetition rates of 1?8Hz. Vlek et al. [14] already showed that imagined auditory accents
can be recognized from EEG. They asked ten subjects to listen to and later imagine three simple metric
patterns of two, three and four beats on top of a steady metronome click. Using logistic regression to
classify accented versus unaccented beats, they obtained an average single-trial accuracy of 70% for
perception and 61% for imagery. These results are very encouraging to further investigate the possibilities
for retrieving information about the perceived rhythm from EEG recordings.
Very recently, the potential of deep learning techniques for neuroimaging has been demonstrated for
functional and structural magnetic resonance imaging (MRI) data [15]. However, applications of deep
learning techniques within neuroscience and specifically for processing EEG recordings have been very
limited so far. Wulsin et al. [16] used deep belief nets (DBNs) to detect anomalies related to epilepsy
in EEG recordings of 11 subjects by classifying individual ?channel-seconds?, i.e., one-second chunks
from a single EEG channel without further information from other channels or about prior values. Their
classifier was first pre-trained layer by layer as an autoencoder on unlabelled data, followed by a supervised
fine-tuning with backpropagation on a much smaller labeled data set. They found that working on raw,
unprocessed data (sampled at 256Hz) led to a classification accuracy comparable to hand-crafted features.
Langkvist et al. [17] similarly employed DBNs combined with hidden Markov models (HMMs) to classify
different sleep stages. Their data for 25 subjects comprised EEG as well as recordings of eye movements
and skeletal muscle activity. Again, the data was segmented into one-second chunks. Here, a DBN on
raw data showed a classification accuracy close to one using 28 selected features.
3 Data acquisition & pre-processing
3.1 Stimuli
The African rhythm stimuli were derived from recordings of traditional East African music [18]. The
author (DC) composed the Western rhythmic stimuli. Rhythms were presented as sequences of sine
tones that were 100ms in duration with intensity ramped up/down over the first/final 50ms and a pitch
of either 375 or 500 Hz. All rhythms had a temporal structure of 12 equal units, in which each unit could
contain a sound or not. For each rhythmic stimulus, two individual rhythmic sequences were overlaid
whereby one sequence was played at the high pitch and the other at the low pitch. There were two groups
of three individual rhythmic sequences for each cultural type of rhythm as shown in Table 1. With three
combinations within each group and two possible pitch assignments, this resulted in six rhythmic stimuli
for each group, 12 per rhythm type and 24 in total.1 Finally, rhythmic stimuli could be played back at
one of two tempi, having a minimum inter-onset interval of either 180 or 240ms.
Furthermore, we also formed groups based on how these stimuli were created. These allowed a more coarse
classification with fewer classes. Ignoring the pitch assignments and thus considering the pairs [a,b] and [b,a]
as equivalent, 12 groups were formed. At the next level, the stimuli derived from the same of the four groups
of three sequences were grouped resulting in four groups of six stimuli. Finally, distinguishing East African
from Western stimuli resulted in the binary classification problem that we addressed in our earlier work.
3.2 Study description
Sixteen East African participants were recruited in Kigali, Rwanda (3 female, mean age: 23 years, mean
musical training: 3.4 years, mean dance training: 2.5 years). The participants first completed three
behavioral tasks: a rhythm discrimination task, a rhythm reproduction task, and a beat tapping task.
Afterward, thirteen subjects also participated in the EEG portion of the study. All participants were over
1
The 24 rhythm stimuli are available at http://dx.doi.org/10.6084/m9.figshare.1213903
2
Table 1: Rhythmic sequences in groups of three that pairings were based on. All ?x?s denote onsets. Larger,
bold ?X?s denote the beginning of a 12 unit cycle (downbeat).
Western Rhythms
1Xxxx xx
2X
x xx
3X xx xx
xx
Xxxx xx
x xX
x xx
xxxxX xx xx
East African Rhythms
1X xxxxx
2X x x x
3X x
x
xx
x x
xxxx
4X xx xx
x xX xx xx
x x
5Xxxx
xx x
Xxxx
xx x
6X xx xx xxxxX xx xx xxxx
xxxxX xxxxx
x xX
x x
x
X x x x
xxxx
x
4X xxx xxx xxX xxx xxx xx
5X xx xx xx xX xx xx xx x
6X xx xx x x X xx xx x x
the age of 18, had normal hearing, and had spent the majority of their lives in East Africa. They all gave
informed consent prior to participating and were compensated for their participation, as per approval by
the ethics boards at the Centre Hospitalier Universitaire de Kigali and the University of Western Ontario.
The participants were instructed to sit with eyes closed and without moving for the duration of the EEG
recording, and to maintain their attention on the auditory stimuli. All rhythms were repeated for 32 seconds,
presented in counterbalanced blocks (all East African rhythms then all Western rhythms, or vice versa),
and with randomized order within blocks. 12 rhythms of each type were presented ? all at the same tempo,
and each rhythm was preceded by 4 seconds of silence. EEG was recorded via a portable Grass EEG
system using 14 channels at a sampling rate of 400Hz and impedances were kept below 10k?.
3.3 Data pre-processing
EEG recordings are usually very noisy. They contain artifacts caused by muscle activity such as eye blinking
as well as possible drifts in the impedance of the individual electrodes over the course of a recording. Furthermore, the recording equipment is very sensitive and easily picks up interferences from the surroundings.
For instance, in this experiment, the power supply dominated the frequency band around 50Hz. All these
issues have led to the common practice to invest a lot of effort into pre-processing EEG data, often even manually rejecting single frames or channels. In contrast to this, we decided to put only little manual work into
cleaning the data and just removed obviously bad channels, thus leaving the main work to the deep learning
techniques. After bad channel removal, 12 channels remained for subjects 1?5 and 13 for subjects 6?13.
We followed the common practice in machine learning to partition the data into training, validation (or
model selection) and test sets. To this end, we split each 32s-long trial recording into three non-overlapping
pieces. The first T seconds after an optional offset were used for the validation set. The rationale behind
this was that we expected that the participants would need a few seconds in the beginning of each trial to
get used to the new rhythm. Thus, the data would be less suited for training but might still be good enough
to estimate the model accuracy on unseen data. The main part of each recording was used for training
and the remaining T seconds for testing. The time length T was tempo-dependent and corresponded to
the length of a single bar in the stimuli. Naturally, one would prefer segments that are as long as the 2-bar
stimuli. However, this would have reduced the amount of data left for training significantly and since only
the East African rhythm sequences 2 and 3 had differences between the first and second bar (cf. Table 1),
we only used 1 bar. With the optional offset, the data sets were aligned to start at the same position within
a bar.2 The specific values for the two tempi are listed in Table 2. Furthermore, we decided to process and
classify each EEG channel individually. Combining all 12 or 13 EEG channels in the analysis might allow
to detect spatial patterns and most likely lead to an increase of the classification performance. However,
this would increase the model complexity (number of parameters) by a factor of more than ten while at the
same time reducing the number of training and test examples by the same factor. Under these conditions,
the amount of data would not be sufficient to effectively train the CNN and lead to severe overfitting.
The data was finally converted into the input format required by the CNN to be learned.3 If the network
just took the raw EEG data, each waveform was normalized to a maximum amplitude of 1 and then split
into equally sized frames of length T matching the size of the network?s input layer. No windowing
2
With offset, the validation and test set would correspond to the same section of the stimuli for the fast tempo
whereas for the fast tempo, it would differ by 1 bar because of the odd number of bars in between.
3
Most of the processing was implemented through the librosa library available at https://github.com/
bmcfee/librosa/.
3
Table 2: Differences between slow and fast stimuli.
tempo participants beat length bar length T
bars optional offset training segment length
fast
slow
1?3, 7?9
4?6, 10?13
180ms
240ms
2160ms
2880ms
14.815
11.111
1760ms
320ms
27680ms - offset
26240ms - offset
function was applied and the hop size (controlling the overlap of consecutive windows) was either 24,
which corresponded to 60ms at the sampling rate of 400Hz, or the equivalent of T in samples. If the
network was designed to process the frequency spectrum, the processing involved:
1. computing the short-time Fourier transform (STFT) with given window length of 96 samples
and a hop size of 24 (This resulted in a new frequency spectrum vector every 60ms.),
2. computing the log amplitude,
3. scaling linearly to a maximum of 1 (per sequence),
4. (optionally) cutting of all frequency bins above the number requested by the network,
5. splitting the data into frames of length T (matching the network?s input dimensionality) with
a given hop size of 1 (60ms) or the equivalent of T .
Hops of 60ms were chosen as this equals to one fourth or one third of the beat length in the slow and
fast rhythms respectively. With this choice, we hoped to be able to pick up beat-related effects but also
to have a window size big enough for a sufficient frequency resolution in the spectrum. Including the
zero-frequency band, this resulted in 49 frequency bins up to 200Hz with a resolution of 4.17Hz. Using
the log amplitude in combination with the normalization had turned out to be the best approach in our
previous experiments trying to distinguish East African from Western stimuli [4].
4 Experiments
CNNs, as for instance described in [3], have a variety of structural parameters which need to be chosen
carefully. In general, CNNs are artificial neural networks (ANNs) with one or more convolutional layers.
In such layers, linear convolution operations are applied for local segments of the input followed by a nonlinear transformation and a pooling operation over neighboring segments. If the EEG data is represented as
waveform, the input has only one dimension (width) which corresponds to the time. If it is represented as
frequency spectrum, it has a second dimension (height) which corresponds to the frequency. The kernel for
each convolution operation is described by a weight matrix of a certain shape. Here, we only considered
the kernel width as free parameter and kept the height maximal. Multiple kernels can be applied in parallel
within the same layer whereby each corresponds to a different output channel of the layer. The stride parameter controls how much the kernels should advance on the input data between successive applications. Here,
we fixed this parameter at 1 resulting in a maximal overlap of consecutive input segments. Finally, the pooling parameter controls how many values of neighboring segments are aggregated using the max operation.
Like in our previous work, we used a DLSVM output layer as proposed in [19].4 This special kind of
output layer for classification uses the hinge loss as cost function and replaces the commonly applied
softmax. The convolutional layers applied the rectifier non-linearity f(x) = max(0,x) which does not
saturate like sigmoid functions and thus facilitates faster learning as proposed in [20]. The input length
in the time dimension was adapted to match the bar length T . All models were trained for 50 epochs using
stochastic gradient descent (SGD) (on mini-batches of size 100) with exponential decay of the learning
rate after each epoch and momentum. The best model was selected based on the accuracy on the validation
set. Furthermore, we applied dropout regularization [21]. In total, this resulted in four learning parameters
with value ranges derived from earlier experiments:
? the initial learning rate (between 0.001 and 0.01),
? the exponential learning rate decay per epoch (between 1.0 and 1.1),
? the initial momentum (between 0.0 and 0.5), and
? and the final momentum in the last epoch (between 0.0 and 0.99)
and three structural parameters for each convolutional layer
? the kernel width (between 1 and the input width for the layer),
? the number of channels (between 1 and 30), and
4
We used the experimental implementation for pylearn2 provided by Kyle Kastner at https:
//github.com/kastnerkyle/pylearn2/blob/svm_layer/pylearn2/models/mlp.py
4
? the pooling width (between 1 and 10).
In our previous work, we successfully applied CNNs with two convolutional layers to classify the perceived
rhythms into types (East African vs. Western) as well as to identify individual rhythms in a pilot experiment
[4]. However, we were only able to test a small number of manually tuned structural configurations,
leaving a considerable potential for further improvement. Here, we took a systematic approach for finding
good structural and learning parameters for the CNNs. To this end, we applied a Bayesian optimization
technique for hyper-parameter selection in machine learning algorithms, which has recently been described
by Snoek et al. [22] and has been implemented in Spearmint library.5 The basic idea is to treat the learning
algorithm?s generalization performance as a sample from a Gaussian process and select the next parameter
configuration to test based on the expected improvement. The authors showed that this way, the number
of experiment runs to minimize a given objective can be significantly reduced while surpassing the
performance of parameters chosen by human experts. We implemented6 our experiments using Theano
[23] and pylearn2 [24]. The computations were run on a dedicated 12-core workstation with two Nvidia
graphics cards ? a Tesla C2075 and a Quadro 2000.
We followed the common practice to optimize the performance on the validation set. Because the 24
classes we would like to predict were perfectly balanced, we chose the accuracy, i.e., the percentage of
correctly classified instances, as primary evaluation measure.7 Furthermore, ranking the 24 classes by their
corresponding network output values, we also computed the precision at rank 3 (prec.@3) and the mean
reciprocal rank (MRR) ? two commonly used information retrieval measures. The former corresponds to accuracy considering the top three classes in the ranking instead of just the first one. The latter is computed as:
|D|
1 X 1
MRR=
|D| i=1 ranki
(1)
where D is the set of test instances and ranki is the rank of the correct class for instance i. The value
range is (0,1] where the best value, 1, is obtained if the correct class is always ranked first.
4.1 Impact of pre-processing (subject 4)
At first, we analyzed the impact of the pre-processing on the performance of a model with a single
convolutional layer. For this, we only considered the recordings from subject 4 who were easiest to classify
in our earlier experiments. The exponential learning rate decay was fixed at 1.08 leaving three structural
and three learning parameters for the Bayesian optimization. Results are shown in Figure 1 (left).
Generally, CNNs using the frequency spectrum representation were faster. A possible reason could be
that the graphics cards performed better using two-dimensional kernels instead of long one-dimensional
ones. Furthermore, the search for good parameters was much harder for the waveform representation
because the value range for the kernel width was much wider ([1,1152] instead of [1,45]). Thus, the search
took much longer. For instance, using the large hop size, an accuracy of more than 20% was only achieved
after 208 runs for CNNs using waveform input with offset and after 47 runs without offset. Comparable
values were already obtained after 1 and 2 runs respectively for the CNNs with frequency spectrum input
and the values shown in Figure 1 (left) were obtained after 45 and 105 runs respectively. Consequently,
the frequency spectrum appeared to be the clearly preferable choice for the input representation.
With the small hop size of 60ms, a lot more training instances were generated because of the high overlap.
This slowed down learning by a factor of more than 10. Hence, fewer configurations could be tested within
the same time. Overall, the large hop size corresponding to 1 bar was favorable because of the significant
speed-up without an impact on accuracy. By using the offset in combination with the hop size of 1 bar,
all instances for training, validation and testing were aligned to the same position within a bar. This could
explain the increase in accuracy for this parameter combination together with the spectrum representation.
In combination with the waveform input, the inverse effect was observed. However, as it was generally
harder to find good solutions in this setting, it could be that testing more configurations eventually would
lead to the same result as for the spectrum.
5
https://github.com/JasperSnoek/spearmint
The code to run the experiments is available as supplementary material at http://dx.doi.org/10.6084/
m9.figshare.1213903
7
As the Bayesian optimization aims to minimize an objective, we let our learner report the misclassification rate
instead which is one minus the accuracy.
6
5
SVM
60ms
(60 runs)
offset
waveform
no 33.3% 233.7s
yes 34.8% 119.5s
33.7% 22.3s
33.0% 16.4s
1 bar
no 33.0% 12.7s
(300 runs) yes 24.7% 5.3s
33.3% 0.4s
35.8% 0.3s
60ms
no
yes
1 bar
no
yes
40
freq. spectrum
35
accuracy (%)
CNN
hop size
training did not finish
within 48 hours
11.1%
12.2%
30
25
20
15
10
5
22.2%
24.3%
0
0
10
20
30
40
50
number of frequency bins (4.167 Hz per bin)
Figure 1: Impact of pre-processing. Left: Classification accuracy and average epoch processing time for
different combinations of the pre-processing parameters. CNN structural and learning parameters were
obtained through Bayesian optimization for 300 runs for hop size 1 bar and 60 runs for hop size 60ms.
Processing times for CNNs were measured separately as single process using the Tesla C2075 graphics
card and averaged over 50 epochs. For comparison, SVM classification accuracies were obtained using
LIBSVM with polynomial kernel (degree 1?5). (Only the best values are shown.) Right: Impact of the
optional frequency bin cutoff on the accuracy.
For a comparison, we also trained support vector machine (SVM) classifiers using LIBSVM [25] on the
same pre-processed data. Here, training did not finish within 48 hours for the small hop size because of the
amount of training data. For waveform data, a polynomial kernel with degree 2 worked best, whereas for the
frequency spectrum, it was a polynomial kernel with degree 4. All values were significantly (more than 10%
absolute) below those obtained with a CNN. This shows using CNNs leads to a substantial improvement.
Next, we analyzed the impact of the optional frequency bin cutoff. To this end, we used the best
pre-processing parameter combination from the above comparison. This time, we fixed the momentum
parameters to an initial value of 0.5 and a final value of 0.99 as these clearly dominated within the best
configurations found so far. Instead, we did not fix the exponential learning rate decay. This resulted in
5 parameters to be optimized. We sampled the number of frequency bins from the range of [1,49] with
higher density for lower values and let the Bayesian optimization run 300 experiments for each value.
Results are shown in Figure 1 (right). A very significant accuracy increase can be observed between
12 and 15 bins which corresponds to a frequency band of 45.8?62.5 Hz in the high gamma range which
has been associated with beat perception, e.g., in [5]. The accuracy increase between 28 and 36 bins
(116?145 Hz) is hard to explain as EEG frequency ranges beyond 100 Hz have barely been studied so
far. Here, a further investigation of the learned patterns (reflected in the CNN kernels) could lead to more
insight. This analysis is still subject of ongoing research. The effect on the processing time was negligible.
Based on these findings, we chose the following pre-processing parameters for the remaining experiments:
The EEG data was represented as frequency spectrum using 49 bins. Input frames were obtained with a
hop size corresponding to the length of 1 bar, T , and with a offset to align all instances to the same position
within a bar.
4.2 One vs. two convolutional layers (all subjects)
Having determined the optimal pre-processing parameters for subject 4 and CNNs with a single
convolutional layer, we also used these settings to train individual models with one and two convolutional
layers for all subjects. This time, we allowed 500 runs of the Bayesian optimization to find the best
parameters in each setting. Additionally, we considered three groups of subjects. The ?fast? and ?slow?
group contained all subjects with the respective stimulus tempo (cf. Table 2) whereas the ?all? group
contained all 13 subjects. For the groups, we stopped the Bayesian optimization after 100 runs as there was
no more improvement and the processing time was much longer due to the bigger size of the combined
data sets. Results are shown in Table 3. Apart from the performance values for classifying individual
instances that correspond a segment from an EEG channel, we also aggregated all predictions from the
12 or 13 different channels of the same trial into one prediction by a simple majority vote. The obtained
accuracies are listed in Table 3 (right). Additionally, we computed the accuracies for the more coarse
variants of the classification problem with fewer classes (cf. Section 3.1).
6
Table 3: Structural parameters and performance values of the best CNNs with one or two convolutional
layers after Bayesian parameter optimization for each subject (500 runs) and the three subject groups (100
runs). Layer structure is written as [kernel shape] / pooling width x number of channels. (A more detailed
table can be found in the supplementary material.)
network structure
subject input 1st layer
2nd layer
1
2
3
4
5
6
7
8
9
10
11
12
13
19.1%
27.1%
21.9%
36.1%
18.1%
29.5%
23.1%
24.0%
21.8%
26.6%
26.6%
32.1%
20.2%
36.1%
46.5%
38.2%
63.5%
34.7%
48.1%
43.9%
44.2%
33.7%
51.0%
55.1%
60.9%
37.2%
0.34
0.42
0.36
0.55
0.33
0.45
0.40
0.41
0.36
0.44
0.45
0.51
0.36
25.0%
37.5%
20.8%
50.0%
16.7%
37.5%
33.3%
41.7%
25.0%
33.3%
33.3%
29.2%
25.0%
29.2%
37.5%
25.0%
62.5%
25.0%
41.7%
45.8%
41.7%
29.2%
33.3%
37.5%
33.3%
29.2%
58.3%
50.0%
45.8%
75.0%
41.7%
54.2%
54.2%
58.3%
58.3%
45.8%
41.7%
54.2%
50.0%
79.2%
87.5%
66.7%
83.3%
70.8%
75.0%
66.7%
91.7%
91.7%
66.7%
75.0%
83.3%
70.8%
mean (1 convolutional layer)
mean (2 convolutional layers)
24.4%
24.4%
46.4%
44.2%
0.41
0.40
30.8%
29.5%
36.5%
34.0%
51.6%
52.2%
74.7%
77.2%
fast
9.7%
9.5%
9.9%
9.1%
7.3%
7.2%
22.1%
21.6%
22.9%
24.3%
19.0%
18.4%
0.23
0.23
0.24
0.24
0.21
0.20
10.4%
11.8%
10.7%
10.1%
7.7%
8.7%
16.7%
19.4%
13.7%
13.1%
12.2%
12.5%
35.4%
38.9%
32.7%
31.5%
29.2%
31.4%
66.7%
67.4%
56.5%
58.9%
57.1%
55.4%
slow
all
33x49
33x49
33x49
45x49
45x49
45x49
33x49
33x49
33x49
45x49
45x49
45x49
45x49
33x49
33x49
45x49
45x49
33x49
33x49
[5x49]/3x16
[10x49]/1x22
[17x49]/1x30
[35x49]/1x30
[40x49]/2x30
[26x49]/5x30
[15x49]/1x13
[5x49]/2x21
[13x49]/2x21
[7x49]/1x30
[27x49]/1x30
[5x49]/5x30
[18x49]/10x21
[16x1]/5x12
channel mean (24 classes)
aggregated trial accuracy
accuracy prec.@3 MRR 24 classes 12 classes 4 classes 2 classes
[1x1]/10x30
[2x1]/2x24
[6x1]/4x30
[5x1]/10x30
[1x1]/6x30
[8x49]/1x22
[1x49]/1x30 [17x1]/1x30
[31x49]/1x30
[1x49]/10x23 [12x1]/5x27
[1x49]/1x30
[3x49]/9x22 [5x1]/5x18
As expected, models learned for groups of participants did not perform very well. Furthermore, the
classification accuracy varied a lot between subjects with the best accuracy (36.1% for subject 4) twice
as high as the worst (18.1% for subject 5). This was most likely due to strong individual differences in
the rhythm perception. But it might at least have been partly caused by the varying quality of the EEG
recordings. For instance, the signal was much noisier than usual for subject 5. For most subjects, the
aggregation per trial significantly increased the classification accuracy. Only in cases where the accuracy
for individual channels was low, such as for subject 5, the aggregation did not yield an improvement.
Overall, the performance of the simpler models with a single convolutional layer was on par with the
more complex ones ? and often even better. One possible reason for this could be that the models with
two convolutional layers had twice as many structural parameters and thus it was potentially harder to
find good configurations. Furthermore, with more weights to be learned and thus more degrees of freedom
to adapt, they were more prone to overfitting on this rather small data set. Figure 2 (left) visualizes the
confusion between the different rhythms for subject 4 where the best overall accuracy was achieved.8
Remarkably, only few of the East African rhythms were misclassified as Western (upper right quadrant)
and vice versa (lower left). For the East African music, confusion was mostly amongst neighbors (i.e.,
similar rhythms; upper left quadrant) ? especially rhythms based on sequences 2 and 3 that were the only
ones that cannot be captured correctly in a window of 1 bar ? whereas for the Western rhythms, there
were patterns indicating a strong perceived similarity between rhythm sequences 1 and 4. The accuracies
obtained for the classification tasks with fewer classes (cf. Table 3, right) paint a similar picture indicating
strong stimulus similarity as the main reason for confusion. In the mean confusion matrix, this effect is far
less pronounced. However, it can be observed in most of the confusion matrices for the individual subjects.
The results reported here still need to be taken with a grain of salt. Because of the study design, there is
only one trial session (of 32 seconds) per stimulus for each subject. Thus, there is the chance that the neural
networks learned to identify the individual trials and not the stimuli based on artifacts in the recordings that
only occurred sporadically throughout the experiment. Or there could have been brain processes unrelated
8
The respective confusion matrices for the models with two convolutional layers look very similar. They can be
found in the supplementary material together with the matrices for the other participants.
7
13
12
11
10
9
True label
8
7
6
5
4
3
2
1
0
[5, 4, 'a']
[6, 4, 'a']
[6, 5, 'a']
[2, 1, 'a']
[3, 2, 'a']
[3, 1, 'a']
[5, 6, 'a']
[2, 3, 'a']
[1, 2, 'a']
[1, 3, 'a']
[4, 5, 'a']
[4, 6, 'a']
[6, 4, 'w']
[2, 1, 'w']
[5, 4, 'w']
[2, 3, 'w']
[5, 6, 'w']
[3, 1, 'w']
[3, 2, 'w']
[4, 6, 'w']
[1, 3, 'w']
[1, 2, 'w']
[4, 5, 'w']
[6, 5, 'w']
subject 4 (labels in trial order)
Predicted label
13
12
11
10
9
8
7
6
5
4
3
2
1
[5, 4, 'a']
[6, 4, 'a']
[6, 5, 'a']
[2, 1, 'a']
[3, 2, 'a']
[3, 1, 'a']
[5, 6, 'a']
[2, 3, 'a']
[1, 2, 'a']
[1, 3, 'a']
[4, 5, 'a']
[4, 6, 'a']
[6, 4, 'w']
[2, 1, 'w']
[5, 4, 'w']
[2, 3, 'w']
[5, 6, 'w']
[3, 1, 'w']
[3, 2, 'w']
[4, 6, 'w']
[1, 3, 'w']
[1, 2, 'w']
[4, 5, 'w']
[6, 5, 'w']
[1, 2, 'a']
[1, 3, 'a']
[2, 1, 'a']
[2, 3, 'a']
[3, 1, 'a']
[3, 2, 'a']
[4, 5, 'a']
[4, 6, 'a']
[5, 4, 'a']
[5, 6, 'a']
[6, 4, 'a']
[6, 5, 'a']
[1, 2, 'w']
[1, 3, 'w']
[2, 1, 'w']
[2, 3, 'w']
[3, 1, 'w']
[3, 2, 'w']
[4, 5, 'w']
[4, 6, 'w']
[5, 4, 'w']
[5, 6, 'w']
[6, 4, 'w']
[6, 5, 'w']
True label
subject 4
[1, 2, 'a']
[1, 3, 'a']
[2, 1, 'a']
[2, 3, 'a']
[3, 1, 'a']
[3, 2, 'a']
[4, 5, 'a']
[4, 6, 'a']
[5, 4, 'a']
[5, 6, 'a']
[6, 4, 'a']
[6, 5, 'a']
[1, 2, 'w']
[1, 3, 'w']
[2, 1, 'w']
[2, 3, 'w']
[3, 1, 'w']
[3, 2, 'w']
[4, 5, 'w']
[4, 6, 'w']
[5, 4, 'w']
[5, 6, 'w']
[6, 4, 'w']
[6, 5, 'w']
0
Predicted label
Figure 2: Confusion matrices for the CNN with a single convolutional layer for subject 4. Labels contain
the ids of the high-pitched and low-pitch rhythm sequence (c.f. Table 1) and the rhythm type (?a? for
African, ?w? for Western). Left: Labels arranged such that most similar rhythms are close together. Right:
Labels in the order of the trials for this subject. More plots are provided in the supplementary material.
to rhythm perception that were only present during some of the trials. Re-arranging the labels within the confusion matrix such that they correspond to the order of the stimuli presentation (Figure 2, right) shows some
confusion between successive trials (blocks along the diagonal) which supports this hypothesis. Repeating
the experiment with multiple trials per stimulus for each subject should give more insights into this matter.
5 Conclusions
Distinguishing the rhythm stimuli used in this study is not easy as a listener. They are all presented in
the same tempo and comprise two 12/8 bars. Consequently, none of the participants scored more than
83% in the behavioral rhythm discrimination test. Considering this and the rather sub-par data quality of
the EEG recordings, the accuracies obtained for some of the participants are remarkable. They demonstrate
that perceived rhythms may be identified from EEG recorded during their auditory presentation using
convolutional neural networks that look only at a short segment of the signal from a single EEG channel
(corresponding to the length of a single bar of a two-bar stimulus).
We hope that our finding will encourage the application of deep learning techniques for EEG analysis
and stimulate more research in this direction. As a next step, we want to analyze the learned models
as they might provide some insight into the important underlying patterns within the EEG signals and
their corresponding neural processes. However, this is largely still an open problem. (As a first attempt,
visualizations of the kernel weight matrices and of input patterns producing the highest activations can
be found in the supplementary material.) We are also looking to correlate the classification performance
values with the subjects? scores in the behavioral part of the study.
The study is currently being repeated with North America participants and we are curious to see whether
we can replicate our findings. In particular, we hope to further improve the classification accuracy through
higher data quality of the new EEG recordings. Furthermore, we want to conduct a behavioral study to
obtain information about the perceived similarity between the stimuli. Finally, encouraged by our results,
we want to extend our focus by also considering more complex and richer stimuli such as audio recordings
of rhythms with realistic instrumentation instead of artificial sine tones.
Acknowledgments
This work was supported by a fellowship within the Postdoc-Program of the German Academic Exchange
Service (DAAD), by the Natural Sciences and Engineering Research Council of Canada (NSERC), through
the Western International Research Award R4911A07, and by an AUCC Students for Development Award.
8
References
[1] D.J. Cameron, J. Bentley, and J.A. Grahn. Cross-cultural influences on rhythm processing: Reproduction,
discrimination, and beat tapping. Frontiers in Human Neuroscience, to appear.
[2] P. Vincent, H. Larochelle, I. Lajoie, Y. Bengio, and P.-A. Manzagol. Stacked denoising autoencoders: Learning
useful representations in a deep network with a local denoising criterion. Journal of Machine Learning Research,
11:3371?3408, 2010.
[3] A. Krizhevsky, I. Sutskever, and G.E. Hinton. Imagenet classification with deep convolutional neural networks.
In Neural Information Processing Systems (NIPS?12), pages 1097?1105, 2012.
[4] S. Stober, D.J. Cameron, and J.A. Grahn. Classifying EEG recordings of rhythm perception. In 15th International
Society for Music Information Retrieval Conference (ISMIR?14), pages 649?654, 2014.
[5] J.S. Snyder and E.W. Large. Gamma-band activity reflects the metric structure of rhythmic tone sequences.
Cognitive Brain Research, 24(1):117?126, 2005.
[6] J.R. Iversen, B.H. Repp, and A.D. Patel. Top-down control of rhythm perception modulates early auditory
responses. Annals of the New York Academy of Sciences, 1169(1):58?73, 2009.
[7] T. Fujioka, L.J. Trainor, E.W. Large, and B. Ross. Beta and gamma rhythms in human auditory cortex during
musical beat processing. Annals of the New York Academy of Sciences, 1169(1):89?92, 2009.
[8] T. Fujioka, L.J. Trainor, E.W. Large, and B. Ross. Internalized timing of isochronous sounds is represented
in neuromagnetic beta oscillations. The Journal of Neuroscience, 32(5):1791?1802, 2012.
[9] S. Nozaradan, I. Peretz, M. Missal, and A. Mouraux. Tagging the neuronal entrainment to beat and meter. The
Journal of Neuroscience, 31(28):10234?10240, 2011.
[10] S. Nozaradan, I. Peretz, and A. Mouraux. Selective neuronal entrainment to the beat and meter embedded in
a musical rhythm. The Journal of Neuroscience, 32(49):17572?17581, 2012.
[11] E. Geiser, E. Ziegler, L. Jancke, and M. Meyer. Early electrophysiological correlates of meter and rhythm
processing in music perception. Cortex, 45(1):93?102, 2009.
[12] O. Ladinig, H. Honing, G. H?aden, and I. Winkler. Probing attentive and preattentive emergent meter in adult
listeners without extensive music training. Music Perception, 26(4):377?386, 2009.
[13] U. Will and E. Berg. Brain wave synchronization and entrainment to periodic acoustic stimuli. Neuroscience
Letters, 424(1):55?60, 2007.
[14] R.J. Vlek, R.S. Schaefer, C.C.A.M. Gielen, J.D.R. Farquhar, and P. Desain. Shared mechanisms in perception
and imagery of auditory accents. Clinical neurophysiology : official journal of the International Federation
of Clinical Neurophysiology, 122(8):1526?1532, 2011.
[15] S.M. Plis, D.R. Hjelm, R.Salakhutdinov, and V.D. Calhoun. Deep learning for neuroimaging: a validation study.
arXiv preprint arXiv:1312.5847, 2013.
[16] D.F. Wulsin, J.R. Gupta, R. Mani, J.A. Blanco, and B. Litt. Modeling electroencephalography waveforms with
semi-supervised deep belief nets: fast classification and anomaly measurement. Journal of Neural Engineering,
8(3):036015, 2011.
[17] M. L?angkvist, L. Karlsson, and M. Loutfi. Sleep stage classification using unsupervised feature learning.
Advances in Artificial Neural Systems, 2012:5:5?5:5, 2012.
[18] G.F. Barz. Music in East Africa: experiencing music, expressing culture. Oxford University Press, 2004.
[19] Y. Tang. Deep Learning using Linear Support Vector Machines. arXiv preprint arXiv:1306.0239, 2013.
[20] X. Glorot, A. Bordes, and Y. Bengio. Deep sparse rectifier networks. In NIPS 2010 Workshop on Deep Learning
and Unsupervised Feature Learning, 2010.
[21] G.E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, and R.R. Salakhutdinov. Improving neural networks
by preventing co-adaptation of feature detectors. arXiv preprint arXiv:1207.0580, 2012.
[22] J. Snoek, H. Larochelle, and R.P. Adams. Practical bayesian optimization of machine learning algorithms. In
F. Pereira, C.J.C. Burges, L. Bottou, and K.Q. Weinberger, editors, Neural Information Processing Systems
(NIPS?12), pages 2951?2959. Curran Associates, Inc., 2012.
[23] J. Bergstra, O. Breuleux, F. Bastien, P. Lamblin, R. Pascanu, G. Desjardins, J. Turian, D. Warde-Farley, and
Y. Bengio. Theano: a CPU and GPU math expression compiler. In Proceedings of the Python for Scientific
Computing Conference (SciPy), 2010.
[24] I.J. Goodfellow, D. Warde-Farley, P. Lamblin, V. Dumoulin, M. Mirza, R. Pascanu, J. Bergstra, F. Bastien, and
Y. Bengio. Pylearn2: a machine learning research library. arXiv preprint arXiv:1308.4214, 2013.
[25] Chih-Chung Chang and Chih-Jen Lin. LIBSVM: A Library for Support Vector Machines. ACM Transactions
on Intelligent Systems and Technology, 2(3):27:1?27:27, 2011.
9
| 5272 |@word neurophysiology:2 trial:13 cnn:7 mri:1 polynomial:3 replicate:1 nd:1 open:1 pick:2 sgd:1 minus:1 harder:4 initial:3 configuration:6 score:1 daniel:1 tuned:1 africa:3 com:3 activation:1 dx:2 written:1 gpu:1 grain:1 realistic:1 partition:1 shape:2 designed:1 plot:1 v:3 discrimination:3 grass:1 selected:2 fewer:4 tone:7 beginning:2 reciprocal:1 short:2 core:1 coarse:2 pascanu:2 math:1 successive:2 org:2 simpler:1 height:2 along:1 beta:3 supply:1 retrieving:1 pairing:1 ramped:1 behavioral:5 inter:1 tagging:1 snoek:2 expected:3 themselves:1 quadro:1 brain:9 approval:1 salakhutdinov:2 encouraging:2 little:1 window:4 electroencephalography:4 considering:4 cpu:1 provided:2 xx:35 underlying:3 cultural:2 linearity:1 unrelated:1 easiest:1 kind:1 informed:1 finding:5 transformation:1 temporal:1 every:1 preferable:1 classifier:2 control:3 unit:3 appear:1 producing:1 negligible:1 understood:1 aggregating:1 local:2 treat:1 service:1 engineering:2 timing:1 id:1 oxford:1 tapping:2 erps:1 might:5 chose:2 emphasis:1 twice:2 studied:1 evoked:1 co:1 hmms:1 limited:1 range:6 averaged:1 decided:2 acknowledgment:1 practical:1 testing:3 practice:3 block:3 differs:1 backpropagation:1 universal:1 significantly:4 matching:2 pre:14 regular:1 quadrant:2 anticipation:1 suggest:1 get:1 cannot:1 close:2 selection:2 put:1 influence:2 py:1 optimize:1 equivalent:3 demonstrated:1 compensated:1 attention:2 duration:2 resolution:2 ranki:2 splitting:1 perceive:1 scipy:1 insight:3 lamblin:2 arranging:1 annals:2 imagine:1 dbns:2 controlling:1 experiencing:1 anomaly:2 cleaning:1 distinguishing:3 us:1 hypothesis:1 curran:1 goodfellow:1 associate:1 persist:1 labeled:1 observed:4 preprint:4 worst:1 cycle:1 movement:1 removed:1 highest:1 balanced:1 substantial:1 complexity:1 asked:1 neuromagnetic:1 warde:2 trained:3 segment:8 learner:1 easily:1 emergent:1 represented:4 america:2 genre:1 listener:2 stacked:2 train:2 distinct:1 fast:8 describe:1 london:1 accented:2 doi:2 universitaire:1 corresponded:2 artificial:3 hyper:1 schaefer:1 richer:1 larger:1 supplementary:5 federation:1 calhoun:1 ability:1 winkler:1 unseen:1 transform:1 noisy:1 final:3 obviously:1 sequence:17 blob:1 net:2 took:3 maximal:2 adaptation:1 neighboring:2 aligned:2 loop:1 stober:2 combining:1 consent:1 turned:1 honing:1 ontario:2 academy:2 description:1 participating:1 pronounced:1 blanco:1 invest:1 sutskever:2 enhancement:1 electrode:1 spearmint:2 produce:1 adam:1 spent:1 wider:1 measured:2 odd:1 strong:4 implemented:2 predicted:2 larochelle:2 differ:1 direction:1 waveform:8 correct:2 cnns:14 stochastic:1 human:6 material:5 bin:10 exchange:1 fix:1 generalization:1 investigation:1 frontier:1 around:1 considered:3 normal:1 overlaid:1 predict:1 desjardins:1 early:3 consecutive:2 perceived:5 favorable:1 label:9 currently:1 ross:2 sensitive:2 individually:1 grouped:1 council:1 repetition:1 vice:2 successfully:1 ziegler:1 reflects:1 hope:2 clearly:2 gaussian:1 always:1 aim:1 rather:2 x27:1 varying:1 derived:4 focus:1 improvement:5 rank:3 contrast:2 equipment:1 plis:1 detect:2 dependent:1 hidden:1 misclassified:1 selective:1 issue:1 classification:21 overall:3 development:1 resonance:1 spatial:1 special:1 softmax:1 equal:2 comprise:1 having:2 sampling:2 manually:2 hop:13 encouraged:1 look:2 unsupervised:2 kastner:1 report:1 stimulus:33 mirza:1 intelligent:1 few:2 aden:1 surroundings:1 composed:1 x49:38 recognize:2 resulted:6 individual:14 gamma:4 maintain:1 attempt:1 jessica:1 freedom:1 mlp:1 investigate:3 possibility:1 karlsson:1 evaluation:1 severe:1 analyzed:2 farley:2 behind:1 x22:3 metrical:2 encourage:1 isochronous:3 culture:3 respective:2 conduct:1 re:1 stopped:1 instance:11 classify:6 earlier:3 increased:1 modeling:1 assignment:2 cost:1 hearing:1 comprised:1 recognizing:1 krizhevsky:2 graphic:3 reported:1 periodic:2 anticipates:1 combined:2 chunk:2 st:1 density:1 international:3 sensitivity:1 randomized:1 systematic:1 together:3 imagery:2 again:1 recorded:4 cognitive:1 expert:1 chung:1 potential:4 m9:2 de:1 converted:1 stride:1 bergstra:2 bold:1 student:1 north:2 metronome:1 matter:1 inc:1 caused:2 ranking:2 onset:2 piece:1 later:1 sine:2 lot:3 closed:1 performed:1 analyze:2 dumoulin:1 portion:1 wave:2 start:1 participant:14 parallel:1 aggregation:2 sporadically:1 compiler:1 litt:1 minimize:2 formed:2 accuracy:31 convolutional:20 musical:6 who:1 largely:1 correspond:3 identify:3 yield:1 yes:4 daad:1 raw:3 bayesian:9 vincent:1 rejecting:1 none:1 visualizes:1 african:17 classified:1 anns:1 oscillatory:2 explain:2 detector:1 sebastian:1 manual:1 attentive:1 acquisition:2 frequency:23 involved:1 naturally:1 associated:1 workstation:1 pitched:1 sampled:2 auditory:8 pilot:1 popular:1 x13:1 listen:1 blinking:1 dimensionality:1 ethic:1 electrophysiological:1 amplitude:3 carefully:1 reflecting:1 back:1 higher:2 supervised:2 xxx:5 reflected:1 response:4 arranged:1 furthermore:9 just:3 stage:2 autoencoders:2 xxxx:7 working:1 hand:1 nonlinear:1 overlapping:1 western:16 accent:2 logistic:1 artifact:2 diverted:1 impulse:1 quality:3 stimulate:1 scientific:1 bentley:1 effect:4 contain:4 normalized:1 true:2 former:1 regularization:1 hence:1 mani:1 freq:1 during:3 width:7 whereby:2 rhythm:57 steady:2 m:19 criterion:1 trying:1 demonstrate:1 confusion:9 dedicated:1 kyle:1 recently:2 common:3 sigmoid:1 stimulation:1 functional:1 preceded:1 salt:1 imagined:1 discussed:1 occurred:1 extend:1 epilepsy:1 surpassing:1 significant:3 measurement:1 expressing:1 versa:2 tuning:1 dbn:1 stft:1 similarly:1 session:1 centre:1 had:6 moving:1 longer:2 similarity:3 cortex:2 align:1 showed:3 female:1 mrr:3 apart:1 instrumentation:1 certain:1 nvidia:1 binary:2 life:1 muscle:2 captured:1 minimum:1 sdas:1 employed:1 recognized:1 aggregated:3 signal:3 semi:1 multiple:3 sound:3 windowing:1 segmented:1 unlabelled:1 faster:2 match:1 adapt:1 long:3 retrieval:2 academic:1 cross:1 clinical:2 lin:1 cameron:3 equally:1 bigger:1 award:2 impact:7 prediction:3 pitch:6 regression:1 basic:1 variant:1 metric:2 arxiv:8 normalization:1 kernel:13 achieved:3 whereas:4 remarkably:1 fine:1 participated:1 interval:1 addressed:1 separately:1 want:3 leaving:3 fellowship:1 breuleux:1 recording:21 subject:35 recruited:2 hz:14 entrained:1 pooling:4 facilitates:1 structural:9 curious:1 split:2 enough:3 easy:1 bengio:4 variety:1 b7:1 finish:2 psychology:1 gave:1 counterbalanced:1 perfectly:1 click:2 identified:1 drum:1 idea:1 listening:2 unprocessed:1 whether:1 six:2 expression:1 effort:1 york:2 deep:13 generally:2 useful:1 detailed:1 listed:2 amount:3 repeating:1 band:6 ten:2 processed:1 reduced:2 http:5 percentage:1 neuroscience:6 per:8 correctly:2 skeletal:1 snyder:1 group:14 four:4 cutoff:2 libsvm:3 kept:2 imaging:1 year:3 run:16 inverse:1 letter:1 fourth:1 throughout:1 chih:2 ismir:1 oscillation:3 prefer:1 scaling:1 comparable:2 dropout:1 layer:27 followed:4 distinguish:2 played:2 sleep:2 replaces:1 activity:5 adapted:1 worked:1 dominated:2 fourier:1 speed:1 format:1 x12:1 department:1 combination:7 smaller:1 slowed:1 theano:2 interference:1 taken:1 visualization:1 discus:1 eventually:1 mechanism:3 german:1 mind:1 end:3 available:3 operation:4 x23:1 apply:1 away:1 prec:2 magnetic:1 tempo:9 batch:1 weinberger:1 top:3 remaining:2 cf:4 completed:1 x21:3 hinge:1 iversen:1 music:9 especially:1 society:2 move:1 objective:2 already:5 paint:1 occurs:1 primary:1 usual:1 traditional:1 diagonal:1 desain:1 gradient:1 amongst:1 card:3 majority:2 lajoie:1 portable:1 reason:3 barely:1 meg:1 besides:1 length:13 code:1 mini:1 manzagol:1 optionally:1 neuroimaging:2 mostly:1 thirteen:1 potentially:1 farquhar:1 implementation:1 design:1 perform:1 upper:2 convolution:2 markov:1 enabling:1 descent:1 beat:13 optional:5 hinton:2 looking:2 uwo:1 dc:1 perturbation:1 frame:4 varied:1 canada:2 intensity:1 drift:1 pair:1 required:1 extensive:1 optimized:1 imagenet:1 acoustic:1 learned:6 hour:2 pylearn2:5 nip:3 address:1 able:3 bar:22 beyond:1 below:2 perception:13 pattern:7 usually:1 adult:1 appeared:1 program:1 including:1 max:2 belief:2 power:1 event:1 overlap:3 ranked:1 misclassification:1 natural:1 participation:1 improve:1 github:3 technology:1 eye:3 library:4 picture:1 created:1 dance:1 autoencoder:1 review:1 prior:2 epoch:6 removal:1 meter:4 python:1 synchronization:2 fully:1 loss:1 par:2 rationale:1 embedded:1 afterward:1 versus:1 sixteen:1 remarkable:1 age:2 validation:7 degree:4 sufficient:2 x18:1 editor:1 classifying:3 bordes:1 prone:1 course:1 supported:1 last:1 free:1 silence:1 allow:1 burges:1 institute:1 neighbor:1 underwent:1 rhythmic:15 absolute:1 sparse:1 dimension:3 preventing:1 author:2 instructed:1 commonly:2 far:4 correlate:2 transaction:1 patel:1 cutting:1 overfitting:2 investigating:1 spectrum:12 search:2 table:12 impedance:2 additionally:2 channel:20 ca:1 ignoring:1 eeg:34 improving:1 requested:1 investigated:1 complex:2 postdoc:1 bottou:1 official:1 did:5 main:3 linearly:1 big:1 scored:1 turian:1 allowed:2 repeated:2 tesla:2 body:1 x1:9 crafted:1 neuronal:2 board:1 x16:1 slow:5 probing:1 precision:1 sub:1 position:3 comprises:1 momentum:4 meyer:1 exponential:4 x24:1 pereira:1 third:1 tang:1 down:3 remained:1 saturate:1 bad:2 specific:1 rectifier:2 bastien:2 jen:1 offset:11 decay:4 svm:3 gupta:1 reproduction:2 sit:1 glorot:1 workshop:1 effectively:1 modulates:1 magnitude:1 hoped:1 occurring:1 x30:15 suited:1 led:2 gielen:1 likely:2 contained:2 nserc:1 chang:1 corresponds:5 chance:2 acm:1 sized:1 presentation:2 consequently:2 rwanda:2 shared:1 hjelm:1 considerable:1 hard:1 specifically:1 determined:1 reducing:1 entrainment:3 denoising:3 total:2 partly:1 experimental:2 vote:1 east:18 preattentive:1 indicating:2 select:1 berg:2 support:4 latter:1 noisier:1 ongoing:1 audio:1 tested:1 phenomenon:2 srivastava:1 |
4,719 | 5,273 | Neurons as Monte Carlo Samplers: Bayesian
Inference and Learning in Spiking Networks
Rajesh P.N. Rao
University of Washington
[email protected]
Yanping Huang
University of Washington
[email protected]
Abstract
We propose a spiking network model capable of performing both approximate
inference and learning for any hidden Markov model. The lower layer sensory
neurons detect noisy measurements of hidden world states. The higher layer neurons with recurrent connections infer a posterior distribution over world states
from spike trains generated by sensory neurons. We show how such a neuronal
network with synaptic plasticity can implement a form of Bayesian inference similar to Monte Carlo methods such as particle filtering. Each spike in the population
of inference neurons represents a sample of a particular hidden world state. The
spiking activity across the neural population approximates the posterior distribution of hidden state. The model provides a functional explanation for the Poissonlike noise commonly observed in cortical responses. Uncertainties in spike times
provide the necessary variability for sampling during inference. Unlike previous
models, the hidden world state is not observed by the sensory neurons, and the
temporal dynamics of the hidden state is unknown. We demonstrate how such
networks can sequentially learn hidden Markov models using a spike-timing dependent Hebbian learning rule and achieve power-law convergence rates.
1
Introduction
Humans are able to routinely estimate unknown world states from ambiguous and noisy stimuli,
and anticipate upcoming events by learning the temporal dynamics of relevant states of the world
from incomplete knowledge of the environment. For example, when facing an approaching tennis
ball, a player must not only estimate the current position of the ball, but also predict its trajectory
by inferring the ball?s velocity and acceleration before deciding on the next stroke. Tasks such as
these can be modeled using a hidden Markov model (HMM), where the relevant states of the world
are latent variables X related to sensory observations Z via a likelihood model (determined by the
emission probabilities). The latent states themselves evolve over time in a Markovian manner, the
dynamics being governed by a transition probabilities. In these tasks, the optimal way of combining such noisy sensory information is to use Bayesian inference, where the level of uncertainty for
each possible state is represented as a probability distribution [1]. Behavioral and neuropsychophysical experiments [2, 3, 4] have suggested that the brain may indeed maintain such a representation
and employ Bayesian inference and learning in a great variety of tasks in perception, sensori-motor
integration, and sensory adaptation. However, it remains an open question how the brain can sequentially infer the hidden state and learn the dynamics of the environment from the noisy sensory
observations.
Several models have been proposed based on populations of neurons to represent probability distribution [5, 6, 7, 8]. These models typically assume a static world state X. To get around this
limitation, firing-rate models [9, 10] have been proposed to used responses in populations of neurons to represent the time-varying posterior distributions of arbitrary hidden Markov models with
discrete states. For the continuous state space, similar models based on line attractor networks [11]
1
have been introduced for implementing the Kalman filter, which assumes all distributions are Gaussian and the dynamics is linear. Bobrowski et al. [12] proposed a spiking network model that can
compute the optimal posterior distribution in continuous time. The limitation of these models is that
model parameters (the emission and transition probabilities) are assumed to be known a priori. Deneve [13, 14] proposed a model for inference and learning based on the dynamics of a single neuron.
However, the maximum number of world state in her model is limited to two.
In this paper, we explore a neural implementation of HMMs in networks of spiking neurons that
perform approximate Bayesian inference similar to the Monte Carlo method of particle filtering [15].
We show how the time-varying posterior distribution P (Xt |Z1:t ) can be directly represented by
mean spike counts in sub-populations of neurons. Each model neuron in the neuron population
behaves as a coincidence detector, and each spike is viewed as a Monte Carlo sample of a particular
world state. At each time step, the probability of a spike in one neuron is shown to approximate
the posterior probability of the preferred state encoded by the neuron. Nearby neurons within the
same sub-population (analogous to a cortical column) encode the same preferred state. The model
thus provides a concrete neural implementation of sampling ideas previously suggested in [16, 17,
18, 19, 20]. In addition, we demonstrate how a spike-timing based Hebbian learning rule in our
network can implement an online version of the Expectation-Maximization(EM) algorithm to learn
the emission and transition matrices of HMMs.
2
Review of Hidden Markov Models
For clarity of notation, we briefly review the equations behind a discrete-time ?grid-based? Bayesian
filter for a hidden Markov model. Let the hidden state be {Xk ? X, k ? N} with dynamics
Xk+1 | (Xk = x0 ) ? f (x|x0 ), where f (x|x0 ) is the transition probability density, X is a discrete
state space of Xk , N is the set of time steps, and ??? denotes distributed according to. We focus
on estimating Xk by constructing its posterior distribution, based only on noisy measurements or
observations {Zk } ? Z where Z can be discrete or continuous. {Zk } are conditional independent
given {Xk } and are governed by the emission probabilities Zk | (Xk = x) ? g(z|x).
i
The posterior probability P (Xk = i|Z1:k ) = ?k|k
may be updated in two stages: a prediction stage
(Eq 1) and a measurement update (or correction) stage (Eq 2):
PX
j
i
P (Xk+1 = i | Z1:k ) =
?k+1|k
= j=1 ?k|k
f (xi |xj ),
(1)
i
P (Xk+1 = i | Z1:k+1 ) = ?k+1|k+1
=
i
?k+1|k
g(Zk+1 |xi )
PX
.
j
j
?
j=1 k+1|k g(Zk+1 |x )
(2)
This process is repeated for each time step. These two recursive equations above are the foundation
for any exact or approximate solution to Bayesian filtering, including well-known examples such as
Kalman filtering when the original continuous state space has been discretized into X bins.
3
Neural Network Model
We now describe the two-layer spiking neural network model we use (depicted in the central panel of
Figure 1(a)). The noisy observation Zk is not directly observed by the network, but sensed through
an array of Z sensory neurons, The lower layer consists of an array of sensory neurons, each of
which will be activated at time k if the observation Zk is in the receptive field. The higher layer
consists of an array of inference neurons, whose activities can be defined as:
s(k) = sgn(a(k) ? b(k))
(3)
where s(k) describes the binary response of an inference neuron at time k, the sign function
sgn(x) = 1 only when x > 0. a(k) represents the sum of neuron?s recurrent inputs, which is
determined by the recurrent weight matrix W among the inference neurons and the population responses sk?1 from the previous time step. b(k) represents the sum of feedforward inputs, which is
determined by the feed-forward weight matrix M as well as the activities in sensory neurons.
Note that Equation 3 defines the output of an abstract inference neuron which acts as a coincidence
detector and fires if and only if both recurrent and sensory inputs are received. In the supplementary
materials, we show that this abstract model neuron can be implemented using the standard leakyintegrate-and-fire (LIF) neurons used to model cortical neurons.
2
(a)
(b)
Figure 1: a. Spiking network model for sequential Monte Carlo Bayesian inference. b. Graphical
representation of spike distribution propagation
3.1
Neural Representation of Probability Distributions
Similar to the idea of grid-based filtering, we first divide the inference neurons into X subpopulations. s = {sil , i = 1, . . . X , l = 1, . . . , L}. We have sil (k) = 1 if there is a spike in the
l-th neuron of the i-th sub-population at time step k. Each sub-population of L neurons share the
same preferred world state, there being X such sub-populations representing each of X preferred
states. One can, for example, view such a neuronal sub-population as a cortical column, within
which neurons encode similar features [21].
Figure 1(a) illustrates how our neural network encodes a simple hidden Markov model with X =
Z = 1, . . . , 100. Xk = 50 is a static state and P (Zk |Xk ) is normally distributed. The network
utilizes 10,000 neurons for the Monte Carlo approximation, with each state preferred by a subpopulation of 100 neurons. At time k, the network observe Zk and the corresponding sensory
neuron whose receptive field contains Zk is activated and sends inputs to the inference neurons.
Combining with recurrent inputs from the previous time step, the responses in the inference neurons
are updated at each time step. As shown in the raster plot of Figure 1(a), the spikes across the entire
inference layer population form a Monte-Carlo approximation to the current posterior distribution:
nik|k :=
L
X
i
sil (k) ? ?k|k
(4)
l=1
where nik|k is the number of spiking neurons in the ith sub-population at time k, which can also be
PX
regarded as the instantaneous firing rate for sub-population i. Nk = i=1 nik|k is the total spike
count in the inference layer population. The set {nik|k } represents the un-normalized conditional
probabilities of Xk , so that P? (Xk = i|Z1:k ) = ? i = ni /Nk .
k|k
3.2
k|k
Bayesian Inference with Stochastic Synaptic Transmission
In this section, we assume the network is given the model parameters in a HMM and there is no
learning in connection weights in the network. To implement the prediction Eq 1 in a spiking
network, we initialize the recurrent connections between the inference neurons as the transition
probabilities: Wij = f (xj |xi )/CW , where CW is a scaling constant. We will discuss how our
network learns the HMM parameters from random initial synaptic weights in section 4.
We define the recurrent weight Wij to be the synaptic release probability between the i-th neuron
sub-population and the j-th neuron sub-population in the inference layer. Each neuron that spikes
at time step k will randomly evoke, with probability Wij , one recurrent excitatory post-synaptic
potential (EPSP) at time step k + 1, after some network delay. We define the number of recurrent
EPSPs received by neuron l in the j-th sub-population as ajl . Thus, ajl is the sum of Nk independent
(but not identically distributed) Bernoulli trials:
ajl (k + 1)
=
X X
L
X
i=1
il0 sil0 (k),
l0 =1
3
?l = 1 . . . L.
(5)
where P (il = 1) = Wij and P (il = 0) = 1 ? Wij . The sum ajl follows the so-called ?Poisson
binomial? distribution [22] and in the limit approaches the Poisson distribution:
X
Nk j
P (ajl (k + 1) ? 1) '
Wij nik|k =
(6)
?
CW k+1|k
i
The detailed analysis of the distribution of ail and the proof of equation 6 are provided in the supplementary materials.
The definition of model neuron in Eq 3 indicates that recurrent inputs alone are not strong enough
to make the inference neurons fire ? these inputs leave the neurons partially activated. We can
view these partially activated neurons as the proposed samples drawn from the prediction density
P (Xk+1 |Xk ). Let njk+1|k be the number of proposed samples in j-th sub-population, we have
E[njk+1|k |{nik|k }] = L
X
X
Wij nik|k = L
i=1
Nk j
? Var[njk+1|k |{nik|k }]
?
CW k+1|k
(7)
Thus, the prediction probability in equation 1 is represented by the expected number of neurons that
receive recurrent inputs.
When a new observation Zk+1 is received, the network will correct the prediction distribution based
on the current observation. Similar to rejection sampling used in sequential Monte Carlo algorithms [15], these proposed samples are accepted with a probability proportional to the observation
likelihood P (Zk+1 |Xk+1 ). We assume for simplicity that receptive fields of sensory neurons do not
overlap with each other (in the supplementary materials, we discuss the more general overlapping
case). Again we define the feedforward weight Mij to be the synaptic release probability between
sensory neuron i and inference neurons in the j-th sub-population. A spiking sensory neuron i
causes an EPSP in a neuron in the j-th sub-population with probability Mij , which is initialized
proportional to the likelihood:
P (bil (k + 1) ? 1) = g(Zk+1 |xi )/CM
i
(8)
j
where CM is a scaling constant such that Mij = g(Zk+1 = z | x )/CM .
Finally, an inference neuron fires a spike at time k + 1 if and only if it receives both recurrent and
sensory inputs. The corresponding firing probability is then the product of the probabilities of the
two inputs:P (sil (k + 1) = 1) = P (ail (k + 1) ? 1)P (bil (k + 1) ? 1)
PL
Let nik+1|k+1 = l=1 sil (k + 1) be the number of spikes in i-th sub-population at time k + 1, we
have
Nk
i
E[nik+1|k+1 |{nik|k }] = L
P (Zk+1 |Z1:k )?k+1|k+1
(9)
CW CM
Nk
i
g(Zk+1 |xi )?k+1|k
(10)
Var[nik+1|k+1 |{nik|k }] ' L
CW CM
Equation 9 ensures that the expected spike distribution at time k + 1 is a Monte Carlo approximation
to the updated posterior probability P (Xk+1 |Z1:k+1 ). It also determines how many neurons are
activated at time k + 1. To keep the number of spikes at different time steps relatively constant, the
scaling constant CM , CW and the number of neurons L could be of the same order of magnitude:
for example, CW = L = 10 ? N1 and CM (k + 1) = 10 ? Nk /N1 , resulting in a form of divisive
inhibition [23]. If the overall neural activity is weak at time k, then the global inhibition regulating
M is decreased to allow more spikes at time k + 1. Moreover, approximations in equations 6 and
N2
10 become exact when C 2k ? 0.
W
3.3
Filtering Examples
Figure 1(b) illustrates how the model network implements Bayesian inference with spike samples.
The top three rows of circles in the left panel in Figure 1(b) represent the neural activities in the
inference neurons, approximating respectively the prior, prediction, and posterior distributions in
the right panel. At time k, spikes (shown as filled circles) in the posterior population represent the
4
(a)
(b)
(c)
Figure 2: Filtering results for uni-modal (a) and bi-modal posterior distributions ((b) and (c) - see
text for details).
distribution P (Xk |Z1:k ). With recurrent weights W ? f (Xk+1 |Xk ), spiking neurons send EPSPs
to their neighbors and make them partially activated (shown as half-filled circles in the second row).
The distribution of partially activated neurons is a Monte-Carlo approximation to the prediction
distribution P (Xk+1 |Z1:k ). When a new observation Zk+1 arrives, the sensory neuron (filled circles
the bottom row) whose receptive field contains Zk+1 is activated, and sends feedforward EPSPs to
the inference neurons using synaptic weights M = g(Z|X). The inference neurons at time k +1 fire
only if they receive both recurrent and feedforward inputs. With the firing probability proportional to
the product of prediction probability P (Xk+1 |Z1:k ) and observation likelihood g(Zk+1 |Xk+1 ), the
spike distribution at time k + 1 (filled circles in the third row) again represents the updated posterior
P (Xk+1 |Z1:k+1 ).
We further tested the filtering results of the proposed neural network with two other example HMMs.
The first example is the classic stochastic volatility model, where X = Z = R. The transition model
of the hidden volatility variable f (Xk+1 |Xk ) = N (0.91Xk , 1.0), and the emission model of the
observed price given volatility is g(Zk |Xk ) = N (0, 0.25 exp(Xk )). The posterior distribution of
this model is uni-modal. In simulation we divided X into 100 bins, and initial spikes N1 = 1000.
We plotted the expected volatility with estimated standard deviation from the population posterior
distribution in Figure 2(a). We found that the neural network does indeed produce a reasonable
estimate of volatility and plausible confidence interval. The second example tests the network?s
ability to approximate bi-modal posterior distributions by comparing the time varying population
posterior distribution with the true one using heat maps (Figures 2(b) and 2(c)). The vertical axis
represents the hidden state and the horizontal axis represents time steps. The magnitude of the
probability is represented by the color. In this example, X = {1, . . . , 8} and there are 20 time steps.
3.4
Convergence Results and Poisson Variability
In this section, we discuss some convergence results for Bayesian filtering using the proposed spiking network and show our population estimator of the posterior probability is a consistent one. Let
ni
P?ki = Nk|k
be the population estimator of the true posterior probability P (Xk = i|Z1:k ) at time k.
k
Suppose the true distribution is known only at initial time k = 1: P? i = ? i . We would like to
1
1|1
investigate how the mean and variance of P?ki vary over time. We derived the updating equations for
mean and variance (see supplementary materials) and found two implications. First, the variance of
neural response is roughly proportional to the mean. Thus, rather than representing noise, Poisson
variability in the model occurs as a natural consequence of sampling and sparse coding. Second, the
variance Var[P?kj ] ? 1/N1 . Therefore Var[P?kj ] ? 0 as N1 ? ?, showing that P?kj is a consistent
j
estimator of ?k|k
. We tested the above two predictions using numerical experiments on arbitrary
HMMs, where we choose X = {1, 2, . . . 20}, Zk ? N (Xk ,P
5), the transition matrix f (xj |xi ) first
uniformly drawn from [0, 1], and then normalized to ensure j f (xj |xi ) = 1.
In Figures 3(a-c), each data point represents Var[P?kj ] along the vertical axis and E[P?kj ] ? E 2 [P?kj ]
along the horizontal axis, calculated over 100 trials with the same random transition matrix f , and
k = 1, . . . 10, j = 1, . . . 20. The solid lines represent a least squares power law fit to the data:
Var[P?kj ] = CV ? (E[P?kj ] ? E 2 [P?kj ])CE . For 100 different random transition matrices f , the means
5
?2
?4
10
10
y = 0.00355627 * x1.13
?3
10
y = 0.028804 * x1.2863
?5
10
k
Var[pjk]
?6
Var[pj ]
Var[pjk]
?4
10
?5
10
?5
10
10
?7
10
?6
?8
10
10
?7
10
?7
10 ?5
10
y = 0.000303182 * x1.037
?3
10
?9
?3
10
E[pj ] ? E2[pj ]
k
?1
10
0
10
?5
10
k
?3
10
E[pj ] ? E2[pj ]
k
?1
10
0
10
10 ?5
10
k
?4
10
?3
?2
10
10
E[pj ] ? E2[pj ]
k
(a)
(b)
(c)
(d)
(e)
(f)
?1
10
0
10
k
Figure 3: Variance versus Mean of estimator for different initial spike counts
of the exponential term CE were 1.2863, 1.13, and 1.037, with standard deviations 0.13, 0.08, and
0.03 respectively, for N1 = 100 and X = 4, 20, and 100. The mean of CE continues to approach
1 when X is increased, as shown in figure 3(d). Since Var[P?kj ] ? (E[P?kj ] ? E 2 [P?kj ]) implies
Var[njk|k ] ? E[njk|k ] (see supplementary material for derivation), these results verify the Poisson
variability prediction of our neural network.
The term CV represents the scaling constant for the variance. Figure 3(e) shows that the mean of
CV over 100 different transition matrices f (over 100 different trials with the same f ) is inversely
proportional to initial spike count N1 , with power law fit CV = 1.77N1?0.9245 . This indicates that
the variance of P?kj converges to 0 if N1 ? ?. The bias between estimated and true posterior
probability can be calculated as:
bias(f) =
X
K
1 X X ?i
i
(E[Pk ] ? ?k|k
)2
XK
i=1 k=1
The relationship between the mean of the bias (over 100 different f ) versus initial count N1 is shown
in figure 3(f). We also have an inverse proportionality between bias and N1 . Therefore, as the figure
j
shows, for arbitrary f , the estimator P?kj is a consistent estimator of ?k|k
.
4
On-line parameter learning
In the previous section, we assumed that the model parameters, i.e., the transition probabilities
f (Xk+1 |Xk ) and the emission probabilities g(Zk |Xk ), are known. In this section, we describe how
these parameters ? = {f, g} can be learned from noisy observations {Zk }. Traditional methods
to estimate model parameters are based on the Expectation-Maximization (EM) algorithm, which
maximizes the (log) likelihood of the unknown parameters log P? (Z1:k ) given a set of observations
collected previously. However, such an ?off-line? approach is biologically implausible because (1)
it requires animals to store all of the observations before learning, and (2) evolutionary pressures
dictate that animals update their belief over ? sequentially any time a new measurement becomes
available.
We therefore propose an on-line estimation method where observations are used for updating parameters as they become available and then discarded. We would like to find the parameters ? that
Pk
maximize the log likelihood: log P? (Z1:k ) = t=1 log P? (Zt |Zt?1 ). Our approach is based on recursively calculating the sufficient statistics of ? using stochastic approximation algorithms and the
6
Figure 4: Performance of the Hebbian Learning Rules.
Monte Carlo method, and employs an online EM algorithm obtained by approximating the expected
sufficient statistic T?(?k ) using the stochastic approximation (or Robbins-Monoro) procedure. Based
on the detailed derivations described in the supplementary materials, we obtain a Hebbian learning
rule for updating the synaptic weights based on the pre-synaptic and post-synaptic activities:
k
Mij
=
?k
Wijk
=
?k
njk|k
Nk
j
nk|k
n
? i (k)
k?1
+ (1 ? ?k
?P i
) ? Mij
? (k)
Nk
in
nik?1|k?1
Nk?1
?
njk|k
Nk
+ (1 ? ?k
nik?1|k?1
Nk?1
when njk|k > 0,
) ? Wijk?1
(11)
when nik?1|k?1 > 0, (12)
where n
? i (k) is the number of pre-synaptic spikes in the i-th sub-population of sensory neurons at
time k, ?k is the learning rate.
Learning both emission and transition probability matrices at the same time using the online EM
algorithm with stochastic approximation is in general very difficult because there are many local
minima in the likelihood function. To verify the correctness of our learning algorithms individually,
we first divide the learning process into two phases. The first phase involves learning the emission
probability g when the hidden world state is stationary, i.e., Wij = fij = ?ij . This corresponds to
learning the observation model of static objects at the center of gaze before learning the dynamics
f of objects. After an observation model g is learned, we relax the stationarity constraint, and allow
the spiking network to update the recurrent weights W to learn the arbitrary transition probability f .
Figure 4 illustrates the performance of learning rules (11) and (12) for a discrete HMM with X = 4
and Z = 12. X and Z values are spaced equally apart: X ? {1, . . . , 4} and Z ? { 23 , 1, 43 , . . . , 4 31 }.
The transition probability matrix f then involves 4?4 = 16 parameters and the emission probability
matrix g involves 12 ? 4 = 48 parameters.
In Figure 4(a), we examine the performance of learning rule (11) for the feedforward weights
M k , with fixed transition matrix. The true emission probability matrix has the
form g.j
=?
2
N (xj , ?Z
). The solid blue curve shows the mean square error (Frobenius norm)
M k ? g
F =
qP
k
k
2
ij (Mij ? gij ) between the learned feedforward weights M and the true emission probability matrix g over trials with different g,. The dotted lines show ? 1 standard deviation for MSE
based on 10 different trials. ?Z varied from trial to trial and was drawn uniformly between 0.2
and 0.4, representing different levels of observation noises. The initial spike distribution was uni0
form ni0|0 = nj0|0 , ?i, j = 1 . . . , X and the initial estimate Mi,j
= Z1 . The learning rate was set
to ?k = k1 , although a small constant learning rate such as ?k = 10?5 also gives rise to similar
learning results. A notable feature in Figure 4(a) is that the average MSE exhibits a fast powerlaw decrease. The red solid line in Figure 4(a) represents the power-law fit to the average MSE:
M SE(k) ? k ?1.1 . Furthermore, the standard deviation of MSE approaches zero as k grows large.
7
Figure 4(a) thus shows the asymptotic convergence of equation (11) irrespective of the ?Z of the
true emission matrix g.
We next examined the performance of learning rule 12 for the recurrent weights W k , given the
learned emission probability matrix g (the true transition probabilities f are unknown to the network). The initial estimator Wij0 = X1 . Similarly, Performance was evaluated by calculating the
qP
(W k ? fij )2 between the learned recurrent weight W k
mean square error
W k ? f
=
F
ij
ij
and the true f . Different randomly chosen transition matrices f were tested. When ?Z = 0.04, the
observation noise is 0.04
1/3 = 12% of the separation between two observed states. Hidden state identification in this case is relatively easy. The red solid line in figure 4(b) represents the power-law fit to
the average MSE: M SE(k) ? k ?0.36 . Similar convergence results can still be obtained for higher
?Z , e.g., ?Z = 0.4 (figure 4(c)). In this case, hidden state identification is much more difficult as
the observation noise is now 1.2 times the separation between two observed states. This difficulty
is reflected in a slower asymptotic convergence rate, with a power-law fit M SE(k) ? k ?0.21 , as
indicated by the red solid line in figure 4(c).
Finally, we show the results for learning both emission and transition matrices simultaneously in
figure 4(d,e). In this experiment, the true emission and transition matrices are deterministic, the
weight matrices are initialized as the sum of the true one and a uniformly random one: Wij0 ?
0
fij + and Mij
? gij + where is a uniform distributed noise between 0 and 1/NX . Although
the asymptotic convergence rate for this case is much slower, it still exhibits desired power-law
convergences in both M SEW (k) ? k ?0.02 and M SEM (k) ? k ?0.08 over 100 trials starting with
different initial weight matrices.
5
Discussion
Our model suggests that, contrary to the commonly held view, variability in spiking does not reflect ?noise? in the nervous system but captures the animal?s uncertainty about the outside world.
This suggestion is similar to some previous models [17, 19, 20], including models linking firing rate
variability to probabilistic representations [16, 8] but differs in the emphasis on spike-based representations, time-varying inputs, and learning. In our model, a probability distribution over a finite
sample space is represented by spike counts in neural sub-populations. Treating spikes as random
samples requires that neurons in a pool of identical cells fire independently. This hypothesis is supported by a recent experimental findings [21] that nearby neurons with similar orientation tuning and
common inputs show little or no correlation in activity. Our model offers a functional explanation
for the existence of such decorrelated neuronal activity in the cortex.
Unlike many previous models of cortical computation, our model treats synaptic transmission between neurons as a stochastic process rather than a deterministic event. This acknowledges the
inherent stochastic nature of neurotransmitter release and binding. Synapses between neurons usually have only a small number of vesicles available and a limited number of post-synaptic receptors
near the release sites. Recent physiological studies [24] have shown that only 3 NMDA receptors
open on average per release during synaptic transmission. These observations lend support to the
view espoused by the model that synapses should be treated as probabilistic computational units
rather than as simple scalar parameters as assumed in traditional neural network models.
The model for learning we have proposed builds on prior work on online learning [25, 26]. The
online algorithm used in our model for estimating HMM parameters involves three levels of approximation. The first level involves performing a stochastic approximation to estimate the expected
complete-data sufficient statistics over the joint distribution of all hidden states and observations.
Cappe and Moulines [26] showed that under some mild conditions, such an approximation produces
a consistent, asymptotically efficient estimator of the true parameters. The second approximation
comes from the use of filtered rather than smoothed posterior distributions. Although the convergence reported in the methods section is encouraging, a rigorous proof of convergence remains to
be shown. The asymptotic convergence rate using only the filtered distribution is about one third
the convergence rate obtained for the algorithms in [25] and [26], where the smoothed distribution
is used. The third approximation results from Monte-Carlo sampling of the posterior distribution.
As discussed in the methods section, the Monte Carlo approximation converges in the limit of large
numbers of particles (spikes).
8
References
[1] R.S. Zemel, Q.J.M. Huys, R. Natarajan, and P. Dayan. Probabilistic computation in spiking populations.
Advances in Neural Information Processing Systems, 17:1609?1616, 2005.
[2] D. Knill and W. Richards. Perception as Bayesian inference. Cambridage University Press, 1996.
[3] K. Kording and D. Wolpert. Bayesian integration in sensorimotor learning. Nature, 427:244?247, 2004.
[4] K. Doya, S. Ishii, A. Pouget, and R. P. N. Rao. Bayesian Brain: Probabilistic Approaches to Neural
Coding. Cambridge, MA: MIT Press, 2007.
[5] K. Zhang, I. Ginzburg, B.L. McNaughton, and T.J.Sejnowski. Interpreting neuronal population activity
by reconstruction: A unified framework with application to hippocampal place cells. Journal of Neuroscience, 16(22), 1998.
[6] R. S. Zemel and P. Dayan. Distributional population codes and multiple motion models. Advances in
neural information procession system, 11, 1999.
[7] S. Wu, D. Chen, M. Niranjan, and S.I. Amari. Sequential Bayesian decoding within a population of
neurons. Neural Computation, 15, 2003.
[8] W.J. Ma, J.M. Beck, P.E. Latham, and A. Pouget. Bayesian inference with probabilistic population codes.
Nature Neuroscience, 9(11):1432?1438, 2006.
[9] R.P.N. Rao. Bayesian computation in recurrent neural circuits. Neural Computation, 16(1):1?38, 2004.
[10] J.M. Beck and A. Pouget. Exact inferences in a neural implementation of a hidden Markov model. Neural
Computation, 19(5):1344?1361, 2007.
[11] R.C. Wilson and L.H. Finkel. A neural implmentation of the kalman filter. Advances in Neural Information Processing Systems, 22:2062?2070, 2009.
[12] O. Bobrowski, R. Meir, and Y. Eldar. Bayesian filtering in spiking neural networks: noise adaptation and
multisensory integration. Neural Computation, 21(5):1277?1320, 2009.
[13] S. Deneve. Bayesian spiking neurons i: Inference. Neural Computation, 20:91?117, 2008.
[14] S. Deneve. Bayesian spiking neurons ii: Learning. Neural Computation, 20:118?145, 2008.
[15] A. Doucet, N. de Freitas, and N. Gordon. Sequential Monte Carlo methods in practice. Springer-Verlag,
2001.
[16] P.O. Hoyer, A. Hyrinen, and A.H. Arinen. Interpreting neural response variability as Monte Carlo sampling of the posterior. Advances in Neural Information Processing Systems 15, 2002.
[17] M G Paulin. Evolution of the cerebellum as a neuronal machine for Bayesian state estimation. J. Neural
Eng., 2:S219?S234, 2005.
[18] N.D. Daw and A.C. Courville. The pigeon as particle lter. Advances in Neural Information Processing
Systems, 19, 2007.
[19] L. Buesing, J. Bill, B. Nessler, and W. Maass. Neural dynamics as sampling: A model for stochastic
computation in recurrent networks of spiking neurons. PLoS Comput Biol, 7(11), 2011.
[20] P. Berkes, G. Orban, M. Lengye, and J. Fisher. Spontaneous cortical activity reveals hallmarks of an
optimal internal model of the environment. Science, 331(6013), 2011.
[21] A. S. Ecker, P. Berens, G.A. Kelirls, M. Bethge, N. K. Logothetis, and A. S. Tolias. Decorrelated neuronal
firing in cortical microcircuits. Science, 327(5965):584?587, 2010.
[22] Jr. Hodges, J. L. and Lucien Le Cam. The Poisson approximation to the Poisson binomial distribution.
The Annals of Mathematical Statistics, 31(3):737?740, 1960.
[23] Frances S. Chance and L. F. Abbott. Divisive inhibition in recurrent networks. Network, 11:119?129,
2000.
[24] E.A. Nimchinsky, R. Yasuda, T.G. Oertner, and K. Svoboda. The number of glutamate receptors opened
by synaptic stimulation in single hippocampal spines. J Neurosci, 24:2054?2064, 2004.
[25] G. Mongillo and S. Deneve. Online learning with hidden Markov models. Neural Computation, 20:1706?
1716, 2008.
[26] O. Cappe and E. Moulines. Online EM algorithm for latent data models, 2009.
9
| 5273 |@word mild:1 trial:8 version:1 briefly:1 norm:1 open:2 proportionality:1 simulation:1 sensed:1 eng:1 pressure:1 solid:5 recursively:1 initial:10 contains:2 njk:8 freitas:1 current:3 comparing:1 must:1 numerical:1 plasticity:1 motor:1 plot:1 treating:1 update:3 alone:1 half:1 stationary:1 nervous:1 xk:36 ith:1 paulin:1 filtered:2 poissonlike:1 provides:2 zhang:1 mathematical:1 along:2 become:2 consists:2 behavioral:1 manner:1 x0:3 expected:5 indeed:2 spine:1 themselves:1 examine:1 roughly:1 brain:3 discretized:1 moulines:2 little:1 encouraging:1 becomes:1 provided:1 estimating:2 notation:1 moreover:1 panel:3 maximizes:1 circuit:1 cm:7 ail:2 unified:1 finding:1 temporal:2 act:1 normally:1 unit:1 before:3 timing:2 local:1 treat:1 limit:2 consequence:1 receptor:3 firing:6 emphasis:1 examined:1 suggests:1 hmms:4 limited:2 bi:2 huys:1 recursive:1 practice:1 implement:4 differs:1 procedure:1 dictate:1 confidence:1 pre:2 subpopulation:2 get:1 nessler:1 bill:1 map:1 deterministic:2 center:1 ecker:1 send:1 starting:1 independently:1 simplicity:1 powerlaw:1 pouget:3 rule:7 estimator:8 array:3 regarded:1 population:35 classic:1 mcnaughton:1 analogous:1 updated:4 annals:1 spontaneous:1 suppose:1 logothetis:1 svoboda:1 exact:3 hypothesis:1 velocity:1 natarajan:1 updating:3 continues:1 richards:1 distributional:1 observed:6 bottom:1 coincidence:2 capture:1 ensures:1 ni0:1 plo:1 decrease:1 environment:3 cam:1 dynamic:9 vesicle:1 joint:1 routinely:1 represented:5 neurotransmitter:1 derivation:2 train:1 heat:1 fast:1 describe:2 monte:15 sejnowski:1 zemel:2 outside:1 whose:3 encoded:1 supplementary:6 plausible:1 relax:1 amari:1 ability:1 statistic:4 noisy:7 online:7 propose:2 reconstruction:1 product:2 epsp:2 adaptation:2 relevant:2 combining:2 achieve:1 frobenius:1 sew:1 convergence:12 transmission:3 produce:2 leave:1 converges:2 object:2 volatility:5 recurrent:20 ij:4 received:3 eq:4 strong:1 epsps:3 implemented:1 c:2 involves:5 implies:1 come:1 fij:3 correct:1 filter:3 stochastic:9 opened:1 human:1 sgn:2 material:6 implementing:1 bin:2 arinen:1 pjk:2 anticipate:1 pl:1 correction:1 around:1 deciding:1 great:1 exp:1 predict:1 vary:1 estimation:2 lucien:1 robbins:1 individually:1 correctness:1 mit:1 gaussian:1 rather:4 finkel:1 varying:4 wilson:1 encode:2 release:5 emission:15 focus:1 l0:1 derived:1 bernoulli:1 likelihood:7 indicates:2 rigorous:1 ishii:1 detect:1 inference:33 dependent:1 dayan:2 typically:1 entire:1 hidden:22 her:1 wij:8 france:1 overall:1 among:1 orientation:1 eldar:1 priori:1 animal:3 integration:3 lif:1 initialize:1 field:4 oertner:1 washington:2 sampling:7 identical:1 represents:11 stimulus:1 gordon:1 inherent:1 employ:2 bil:2 randomly:2 simultaneously:1 beck:2 phase:2 fire:6 attractor:1 maintain:1 n1:11 stationarity:1 regulating:1 investigate:1 wijk:2 arrives:1 behind:1 activated:8 held:1 implication:1 rajesh:1 capable:1 necessary:1 filled:4 incomplete:1 divide:2 initialized:2 circle:5 plotted:1 desired:1 increased:1 column:2 rao:4 markovian:1 maximization:2 deviation:4 uniform:1 delay:1 reported:1 density:2 probabilistic:5 off:1 decoding:1 pool:1 gaze:1 bethge:1 concrete:1 again:2 hodges:1 central:1 espoused:1 reflect:1 huang:1 choose:1 yasuda:1 potential:1 de:1 coding:2 notable:1 view:4 red:3 mongillo:1 il:2 ni:2 square:3 variance:7 spaced:1 sensori:1 weak:1 bayesian:21 identification:2 buesing:1 carlo:15 trajectory:1 stroke:1 detector:2 implausible:1 synapsis:2 decorrelated:2 synaptic:15 definition:1 raster:1 sensorimotor:1 e2:3 proof:2 mi:1 static:3 knowledge:1 color:1 nmda:1 cappe:2 feed:1 higher:3 reflected:1 response:7 modal:4 evaluated:1 microcircuit:1 furthermore:1 stage:3 correlation:1 receives:1 horizontal:2 overlapping:1 propagation:1 defines:1 indicated:1 grows:1 normalized:2 true:12 verify:2 procession:1 evolution:1 maass:1 cerebellum:1 during:2 ambiguous:1 hippocampal:2 complete:1 demonstrate:2 latham:1 motion:1 interpreting:2 hallmark:1 instantaneous:1 common:1 behaves:1 functional:2 spiking:19 qp:2 stimulation:1 linking:1 discussed:1 approximates:1 measurement:4 cambridge:1 cv:4 tuning:1 grid:2 similarly:1 particle:4 tennis:1 cortex:1 inhibition:3 berkes:1 posterior:24 recent:2 showed:1 apart:1 ajl:5 store:1 verlag:1 binary:1 minimum:1 maximize:1 ii:1 multiple:1 infer:2 hebbian:4 offer:1 divided:1 post:3 equally:1 niranjan:1 prediction:10 expectation:2 poisson:7 wij0:2 represent:5 cell:2 receive:2 addition:1 nj0:1 decreased:1 interval:1 sends:2 unlike:2 contrary:1 near:1 feedforward:6 identically:1 enough:1 easy:1 variety:1 xj:5 fit:5 approaching:1 idea:2 cause:1 detailed:2 se:3 meir:1 dotted:1 sign:1 estimated:2 neuroscience:2 per:1 blue:1 discrete:5 drawn:3 clarity:1 pj:7 ce:3 abbott:1 lter:1 uw:2 deneve:4 asymptotically:1 sum:5 inverse:1 uncertainty:3 place:1 reasonable:1 wu:1 doya:1 utilizes:1 separation:2 scaling:4 layer:8 ki:2 courville:1 activity:10 constraint:1 encodes:1 nearby:2 orban:1 performing:2 px:3 relatively:2 according:1 ball:3 jr:1 across:2 describes:1 em:5 biologically:1 ginzburg:1 equation:9 remains:2 previously:2 discus:3 count:6 available:3 observe:1 slower:2 existence:1 original:1 assumes:1 denotes:1 binomial:2 top:1 ensure:1 graphical:1 calculating:2 k1:1 build:1 approximating:2 upcoming:1 question:1 spike:30 occurs:1 receptive:4 traditional:2 evolutionary:1 exhibit:2 hoyer:1 cw:8 hmm:5 nx:1 collected:1 kalman:3 code:2 modeled:1 relationship:1 difficult:2 yanping:1 rise:1 implementation:3 zt:2 unknown:4 perform:1 vertical:2 neuron:70 observation:21 markov:9 discarded:1 finite:1 variability:7 varied:1 smoothed:2 arbitrary:4 introduced:1 connection:3 z1:15 learned:5 daw:1 able:1 suggested:2 usually:1 perception:2 including:2 explanation:2 belief:1 lend:1 power:7 event:2 overlap:1 natural:1 difficulty:1 treated:1 glutamate:1 representing:3 inversely:1 axis:4 irrespective:1 acknowledges:1 kj:14 text:1 review:2 prior:2 evolve:1 asymptotic:4 law:7 suggestion:1 limitation:2 filtering:10 proportional:5 facing:1 var:11 sil:5 versus:2 foundation:1 sufficient:3 consistent:4 share:1 row:4 excitatory:1 supported:1 bias:4 allow:2 neighbor:1 sparse:1 distributed:4 curve:1 calculated:2 cortical:7 world:13 transition:19 sensory:18 forward:1 commonly:2 kording:1 approximate:5 uni:2 preferred:5 evoke:1 keep:1 global:1 sequentially:3 doucet:1 reveals:1 assumed:3 tolias:1 xi:7 continuous:4 latent:3 un:1 sk:1 learn:4 zk:23 nature:3 sem:1 mse:5 berens:1 constructing:1 pk:2 neurosci:1 noise:8 n2:1 knill:1 repeated:1 x1:4 neuronal:6 site:1 sub:17 position:1 inferring:1 exponential:1 comput:1 governed:2 third:3 learns:1 xt:1 showing:1 physiological:1 sequential:4 nik:16 magnitude:2 illustrates:3 nk:15 chen:1 rejection:1 wolpert:1 depicted:1 pigeon:1 explore:1 partially:4 il0:1 scalar:1 binding:1 springer:1 mij:7 corresponds:1 determines:1 chance:1 ma:2 conditional:2 viewed:1 acceleration:1 price:1 fisher:1 determined:3 uniformly:3 sampler:1 bobrowski:2 total:1 called:1 accepted:1 divisive:2 gij:2 player:1 experimental:1 multisensory:1 internal:1 support:1 tested:3 biol:1 |
4,720 | 5,274 | A framework for studying synaptic plasticity
with neural spike train data
Scott W. Linderman
Harvard University
Cambridge, MA 02138
Christopher H. Stock
Harvard College
Cambridge, MA 02138
Ryan P. Adams
Harvard University
Cambridge, MA 02138
[email protected]
[email protected]
[email protected]
Abstract
Learning and memory in the brain are implemented by complex, time-varying
changes in neural circuitry. The computational rules according to which synaptic
weights change over time are the subject of much research, and are not precisely
understood. Until recently, limitations in experimental methods have made it challenging to test hypotheses about synaptic plasticity on a large scale. However, as
such data become available and these barriers are lifted, it becomes necessary
to develop analysis techniques to validate plasticity models. Here, we present
a highly extensible framework for modeling arbitrary synaptic plasticity rules
on spike train data in populations of interconnected neurons. We treat synaptic weights as a (potentially nonlinear) dynamical system embedded in a fullyBayesian generalized linear model (GLM). In addition, we provide an algorithm
for inferring synaptic weight trajectories alongside the parameters of the GLM and
of the learning rules. Using this method, we perform model comparison of two
proposed variants of the well-known spike-timing-dependent plasticity (STDP)
rule, where nonlinear effects play a substantial role. On synthetic data generated
from the biophysical simulator NEURON, we show that we can recover the weight
trajectories, the pattern of connectivity, and the underlying learning rules.
1
Introduction
Synaptic plasticity is believed to be the fundamental building block of learning and memory in the
brain. Its study is of crucial importance to understanding the activity and function of neural circuits.
With innovations in neural recording technology providing access to the simultaneous activity of
increasingly large populations of neurons, statistical models are promising tools for formulating and
testing hypotheses about the dynamics of synaptic connectivity. Advances in optical techniques [1,
2], for example, have made it possible to simultaneously record from and stimulate large populations
of synaptically connected neurons. Armed with statistical tools capable of inferring time-varying
synaptic connectivity, neuroscientists could test competing models of synaptic plasticity, discover
new learning rules at the monosynaptic and network level, investigate the effects of disease on
synaptic plasticity, and potentially design stimuli to modify neural networks.
Despite the popularity of GLMs for spike data, relatively little work has attempted to model the
time-varying nature of neural interactions. Here we model interaction weights as a dynamical system
governed by parametric synaptic plasticity rules. To perform inference in this model, we use particle
Markov Chain Monte Carlo (pMCMC) [3], a recently developed inference technique for complex
time series. We use this new modeling framework to examine the problem of using recorded data to
distinguish between proposed variants of spike-timing-dependent plasticity (STDP) learning rules.
1
time
Figure 1: A simple network of four sparsely connected neurons whose synaptic weights are changing over time.
Here, the neurons have inhibitory self connections to mimic refractory effects, and are connected via a chain of
excitatory synapses, as indicated by the nonzero entries A1?2 , A2?3 , and A3?4 . The corresponding weights
of these synapses are strengthening over time (darker entries in W ), leading to larger impulse responses in the
firing rates and a greater number of induced post-synaptic spikes (black dots), as shown below.
2
Related Work
The GLM is a probabilistic model that considers spike trains to be realizations from a point process
with conditional rate ?(t) [4, 5]. From a biophysical perspective, we interpret this rate as a nonlinear
function of the cell?s membrane potential. When the membrane potential exceeds the spiking threshold potential of the cell, ?(t) rises to reflect the rate of the cell?s spiking, and when the membrane
potential decreases below the spiking threshold, ?(t) decays to zero. The membrane potential is
modeled as the sum of three terms: a linear function of the stimulus, I(t), for example a low-pass
filtered input current, the sum of excitatory and inhibitory PSPs induced by presynaptic neurons, and
n
a constant background rate. In a network of N neurons, let Sn = {sn,m }M
m=1 ? [0, T ] be the set of
observed spike times for neuron n, where T is the duration of the recording and Mn is the number
of spikes. The conditional firing rate of a neuron n can be written,
?
?
Z t
N M
n0
X
X
hn0 ?n (t ? sn0 ,m ) ? I[sn0 ,m < t]? , (1)
?n (t) = g ?bn +
kn (t ? ? ) ? I(? ) d? +
0
n0 =1 m=1
where bn is the background rate, the second term is a convolution of the (potentially vector-valued)
stimulus with a linear stimulus filter, kn (?t), and the third is a linear summation of impulse responses, hn0 ?n (?t), which preceding spikes on neuron n0 induce on the membrane potential of
neuron n. Finally, the rectifying nonlinearity g : R ? R+ converts this linear function of stimulus
and spike history into a nonnegative rate. While the spiking threshold potential is not explicitly
modeled in this framework, it is implicitly inferred in the amplitude of the impulse responses.
From this semi-biophysical perspective it is clear that one shortcoming of the standard GLM is that it
does not account for time-varying connectivity, despite decades of research showing that changes in
synaptic weight occur over a variety of time scales and are the basis of many fundamental cognitive
processes. This absence is due, in part, to the fact that this direct biophysical interpretation is not
warranted in most traditional experimental regimes, e.g., in multi-electrode array (MEA) recordings
where electrodes are relatively far apart. However, as high resolution optical recordings grow in
popularity, this assumption must be revisited; this is a central motivation for the present model.
There have been a few efforts to incorporate dynamics into the GLM. Stevenson and Koerding [6]
extended the GLM to take inter-spike intervals as a covariates and formulated a generalized bilinear
model for weights. Eldawlatly et al. [7] modeled the time-varying parameters of a GLM using a
dynamic Bayesian network (DBN). However, neither of these approaches accommodate the breadth
of synaptic plasticity rules present in the literature. For example, parametric STDP models with hard
2
bounds on the synaptic weight are not congruent with the convex optimization techniques used by
[6], nor are they naturally expressed in a DBN. Here we model time-varying synaptic weights as a
potentially nonlinear dynamical system and perform inference using particle MCMC.
Nonstationary, or time-varying, models of synaptic weights have also been studied outside the context of GLMs. For example, Petreska et al. [8] applied hidden switching linear dynamical systems models to neural recordings. This approach has many merits, especially in traditional MEA
recordings where synaptic connections are less likely and nonlinear dynamics are not necessarily
warranted. Outside the realm of computational neuroscience and spike train analysis, there exist a
number of dynamic statistical models, such as West et al. [9], which explored dynamic generalized
linear models. However, the types of models we are interested in for studying synaptic plasticity
are characterized by domain-specific transition models and sparsity structure, and until recently, the
tools for effectively performing inference in these models have been limited.
3
A Sparse Time-Varying Generalized Linear Model
In order to capture the time-varying nature of synaptic weights, we extend the standard GLM by first
factoring the impulse responses in the firing rate of Equation 1 into a product of three terms:
hn0 ?n (?t, t) ? An0 ?n Wn0 ?n (t) rn0 ?n (?t).
(2)
Here, An0 ?n ? {0, 1} is a binary random variable indicating the presence of a direct synapse
from neuron n0 to neuron n, Wn0 ?n (t) : [0, T ] ? R is a non stationary synaptic ?weight? trajectory
R ?associated with the synapse, and rn0 ?n (?t) is a nonnegative, normalized impulse response,
i.e. 0 rn0 ?n (? )d? = 1. Requiring rn0 ?n (?t) to be normalized gives meaning to the synaptic
weights: otherwise W would only be defined up to a scaling factor. For simplicity, we assume r(?t)
does not change over time, that is, only the amplitude and not the duration of the PSPs are timevarying. This restriction could be adapted in future work.
As is often done in GLMs, we model the normalized impulse responses as a linear combination of
basis functions. In order to enforce the normalization of r(?), however, we use a convex combination
of normalized, nonnegative basis functions. That is,
rn0 ?n (?t) ?
B
X
(n0 ?n)
?b
rb (?t),
b=1
R?
PB
(n0 ?n)
where 0 rb (? ) d? = 1, ?b and b=1 ?b
= 1, ?n, n0 . The same approach is used to model
the stimulus filters, kn (?t), but without the normalization and non-negativity constraints.
The binary random variables An0 ?n , which can be collected into an N ? N binary matrix A,
model the connectivity of the synaptic network. Similarly, the collection of weight trajectories {{Wn0 ?n (t)}}n0 ,n , which we will collectively refer to as W (t), model the time-varying synaptic weights. This factorization is often called a spike-and-slab prior [10], and it allows us to separate
our prior beliefs about the structure of the synaptic network from those about the evolution of synaptic weights. For example, in the most general case we might leverage a variety of random network
models [11] as prior distributions for A, but here we limit ourselves to the simplest network model,
the Erd?os-Renyi model. Under this model, each An0 ?n is an independent identically distributed
Bernoulli random variable with sparsity parameter ?.
Figure 1 illustrates how the adjacency matrix and the time-varying weights are integrated into the
GLM. Here, a four-neuron network is connected via a chain of excitatory synapses, and the synapses
strengthen over time due to an STDP rule. This is evidenced by the increasing amplitude of the
impulse responses in the firing rates. With larger synaptic weights comes an increased probability
of postsynaptic spikes, shown as black dots in the figure. In order to model the dynamics of the
time-varying synaptic weights, we turn to a rich literature on synaptic plasticity and learning rules.
3.1
Learning rules for time-varying synaptic weights
Decades of research on synapses and learning rules have yielded a plethora of models for the evolution of synaptic weights [12]. In most cases, this evolution can be written as a dynamical system,
dW (t)
= ` (W (t), {sn,m : sn,m < t} ) + (W (t), t),
dt
3
where ` is a potentially nonlinear learning rule that determines how synaptic weights change as a
function of previous spiking. This framework encompasses rate-based rules such as the Oja rule
[13] and timing-based rules such as STDP and its variants. The additive noise, (W (t), t), need not
be Gaussian, and many models require truncated noise distributions.
Following biological intuition, many common learning rules factor into a product of simpler functions.
For example, STDP (defined below) updates each synapse independently such that dWn0 ?n (t)/dt only depends on Wn0 ?n (t) and the presynaptic spike history Sn<t = {sn,m : sn,m < t}. Biologically speaking, this means that plasticity is local to the
synapse. More sophisticated rules allow dependencies among the columns of W . For example, the
incoming weights to neuron n may depend upon one another through normalization, as in the Oja
rule [13], which scales synapse strength according to the total strength of incoming synapses.
Extensive research in the last fifteen years has identified the relative spike timing between the preand postsynaptic neurons as a key component of synaptic plasticity, among other factors such as
mean firing rate and dendritic depolarization [14]. STDP is therefore one of the most prominent
learning rules in the literature today, with a number of proposed variants based on cell type and
biological plausibility. In the experiments to follow, we will make use of two of these proposed variants. First, consider the canonical STDP rule with a ?double-exponential? function parameterized
by ?? , ?+ , A? , and A+ [15], in which the effect of a given pair of pre-synaptic and post-synaptic
spikes on a weight may be written:
` (Wn0 ?n (t), Sn0 , Sn ) = I[t ? Sn ] `+ (Sn0 ; A+ , ?+ ) ? I[t ? Sn0 ] `? (Sn ; A? , ?? ),
(3)
X
X
`+ (Sn0 ; A+ , ?+ ) =
A+ e(t?sn0 ,m )/?+
`? (Sn ; A? , ?? ) =
A? e(t?sn,m )/?? .
sn0 ,m ?Sn0 <t
sn,m ?Sn<t
This rule states that weight changes only occur at the time of pre- or post-synaptic spikes, and that
the magnitude of the change is a nonlinear function of interspike intervals.
A slightly more complicated model known as the multiplicative STDP rule extends this by bounding
the weights above and below by Wmax and Wmin , respectively [16]. Then, the magnitude of the
weight update is scaled by the distance from the threshold:
` (Wn0 ?n (t), Sn0 , Sn ) = I[t ? Sn ] `?+ (Sn0 ; A+ , ?+ ) (Wmax ? Wn0 ?n (t)),
? I[t ? Sn0 ] `?? (Sn ; A? , ?? ) (Wn0 ?n (t) ? Wmin ).
(4)
Here, by setting `?? = min(`? , 1), we enforce that the synaptic weights always fall
within [Wmin , Wmax ]. With this rule, it often makes sense to set Wmin to zero.
Similarly, we can construct an additive, bounded model which is identical to the standard additive
STDP model except that weights are thresholded at a minimum and maximum value. In this model,
the weight never exceeds its set lower and upper bounds, but unlike the multiplicative STDP rule,
the proposed weight update is independent of the current weight except at the boundaries. Likewise,
whereas with the canonical STDP model it is sensible to use Gaussian noise for (t) in the bounded
multiplicative model we use truncated Gaussian noise to respect the hard upper and lower bounds
on the weights. Note that this noise is dependent upon the current weight, Wn0 ?n (t).
The nonlinear nature of this rule, which arises from the multiplicative interactions among the parameters, ?` = {A+ , ?+ , A? , ?? , Wmax , Wmax }, combined with the potentially non-Gaussian noise
models, pose substantial challenges for inference. However, the computational cost of these detailed
models is counterbalanced by dramatic expansions in the flexibility of the model and the incorporation of a priori knowledge of synaptic plasticity. These learning models can be interpreted as strong
regularizers of models that would otherwise be highly underdetermined, as there are N 2 weight trajectories and only N spike trains. In the next section we will leverage powerful new techniques for
Bayesian inference in order to capitalize on these expressive models of synaptic plasticity.
4
Inference via particle MCMC
The traditional approach to inference in the standard GLM is penalized maximum likelihood estimation. The log likelihood of a single conditional Poisson process is well known to be,
Z T
Mn
X
N
L ?n (t); {Sn }n=1 , I(t) = ?
?n (t) dt +
log (?n (sn,m )) ,
(5)
0
4
m=1
and the log likelihood of a population of non-interacting spike trains is simply the sum of
each of the log likelihoods for each neuron. The likelihood depends upon the parameters ?GLM = {bn , kn , {hn0 ?n (?t)}N
n0 =1 } through the definition of the rate function given in Equation 1. For some link functions g, the log likelihood is a concave function of ?GLM , and the MLE can
be found using efficient optimization techniques. Certain dynamical models, namely linear Gaussian latent state space models, also support efficient inference via point process filtering techniques
[17].
Due to the potentially nonlinear and non-Gaussian nature of STDP, these existing techniques are
not applicable here. Instead we use particle MCMC [3], a powerful technique for inference in time
series. Particle MCMC samples the posterior distribution over weight trajectories, W (t), the adjacency matrix A, and the model parameters ?GLM and ?` , given the observed spike trains, by combining particle filtering with MCMC. We represent the conditional distribution over weight trajectories
with a set of discrete particles. Let the instantaneous weights at (discretized) time t be represented
(p)
N ?N
by a set of P particles, {W t }P
and are assigned normalized parp=1 . The particles live in R
PP
1
ticle weights , ?p , which approximate the true distribution via Pr(W t ) ? p=1 ?p ?W (p) (W t ).
t
Particle filtering is a method of inferring a distribution over weight trajectories by iteratively propagating forward in time and reweighting according to how well the new samples explain the data. For
(p)
each particle W t at time t, we propagate forward one time step using the learning rule to obtain
(p)
a particle W t+1 . Then, using Equation 5, we evaluate the log likelihood of the spikes that occurred
in the window [t, t + 1) and update the weights. Since some of these particles may have very low
weights, after each step we resample the particles. After the T -th time step we are left with a set of
(p)
(p)
weight trajectories {(W 0 , . . . , W T )}P
p=1 , each associated with a particle weight ?p .
Particle filtering only yields a distribution over weight trajectories, and implicitly assumes that the
other parameters have been specified. Particle MCMC provides a broader inference algorithm for
both weights and other parameters. The idea is to interleave conditional particle filtering steps
that sample the weight trajectory given the current model parameters and the previously sampled
weights, with traditional Gibbs updates to sample the model parameters given the current weight
trajectory. This combination leaves the stationary distribution of the Markov chain invariant and
allows joint inference over weights and parameters. Gibbs updates for the remaining model parameters, including those of the learning rule, are described in the supplementary material.
Collapsed sampling of A and W (t) In addition to sampling of weight trajectories and model
parameters, particle MCMC approximates the marginal likelihood of entries in the adjacency matrix, A, integrating out the corresponding weight trajectory. We have, up to a constant,
Pr(An0 ?n | S, ?` , ?GLM , A?n0 ?n , W ?n0 ?n (t))
Z TZ ?
=
p(An0 ?n , Wn0 ?n (t) | S, ?` , ?GLM , A?n0 ?n , W ?n0 ?n (t)) dWn0 ?n (t) dt
0
??
"
#
T
P
Y
1 X (p)
?
?
?
Pr(An0 ?n ),
P p=1 t
t=1
where ?n0 ? n indicates all entries except for n0 ? n, and the particle weights are obtained by
running a particle filter for each assignment of An0 ?n . This allows us to jointly sample An?n0
and Wn?n0 (t) by first sampling An?n0 and then Wn?n0 (t) given An?n0 . By marginalizing out the
weight trajectory, our algorithm is able to explore the space of adjacency matrices more efficiently.
We capitalize on a number of other opportunities for computational efficiency as well. For example, if the learning rule factors into independent updates for each Wn0 ?n (t), then we can update
each synapse?s weight trajectory separately and reduce the particles to one-dimensional objects. In
our implementation, we also make use of a pMCMC variant with ancestor sampling [18] that significantly improves convergence. Any distribution may be used to propagate the particles forward;
using the learning rule is simply the easiest to implement and understand. We have omitted a number
of details in this description; for a thorough overview of particle MCMC, the reader should consult
[3, 18].
1
Note that the particle weights are not the same as the synaptic weights.
5
Figure 2: We fit time-varying weight trajectories to spike trains simulated from a GLM with two neurons
undergoing no plasticity (top row), an additive, unbounded STDP rule (middle), and a multiplicative, saturating
STDP rule (bottom row). We fit the first 50 seconds with four different models: MAP for an L1-regularized
GLM, and fully-Bayesian inference for a static, additive STDP, and multiplicative STDP learning rules. In all
cases, the correct models yield the highest predictive log likelihood on the final 10 seconds of the dataset.
5
Evaluation
We evaluated our technique with two types of synthetic data. First, we generated data from our
model, with known ground-truth. Second, we used the well-known simulator NEURON to simulate
driven, interconnected populations of neurons undergoing synaptic plasticity. For comparison, we
show how the sparse, time-varying GLM compares to a standard GLM with a group LASSO prior
on the impulse response coefficients for which we can perform efficient MAP estimation.
5.1
GLM-based simulations
As a proof of concept, we study a single synapse undergoing a variety of synaptic plasticity rules
and generating spikes according to a GLM. The neurons also have inhibitory self-connections to
mimic refractory effects. We tested three synaptic plasticity mechanisms: a static synapse (i.e., no
plasticity), the unbounded, additive STDP rule given by Equation 3, and the bounded, multiplicative
STDP rule given by Equation 4. For each learning rule, we simulated 60 seconds of spiking activity
at 1kHz temporal resolution, updating the synaptic weights every 1s. The baseline firing rates were
normally distributed with mean 20Hz and standard deviation of 5Hz. Correlations in the spike timing
led to changes in the synaptic weight trajectories that we could detect with our inference algorithm.
Figure 2 shows the true and inferred weight trajectories, the inferred learning rules, and the predictive
log likelihood on ten seconds of held out data for each of the three ground truth learning rules. When
the underlying weights are static (top row), MAP estimation and static learning rules do an excellent
6
mV
9
Figure 3: Evaluation of synapse detection on a 60 second spike train from a network of 10 neurons undergoing
synaptic plasticity with a saturating, additive STDP rule, simulated with NEURON. The sparse, time-varying
GLM with an additive rule outperforms the fully-Bayesian model with static weights, MAP estimation with L1
regularization, and simple thresholding of the cross-correlation matrix.
job of detecting the true weight whereas the two time-varying models must compensate by either
setting the learning rule as close to zero as possible, as the additive STDP does, or setting the
threshold such that the weight trajectory is nearly constant, as the multiplicative model does. Note
that the scales of the additive and multiplicative learning rules are not directly comparable since the
weight updates in the multiplicative case are modulated by how close the weight is to the threshold.
When the underlying weights vary (middle and bottom rows), the static models must compromise
with an intermediate weight. Though the STDP models are both able to capture the qualitative
trends, the correct model yields a better fit and better predictive power in both cases.
In terms of computational cost, our approach is clearly more expensive than alternative approaches
based on MAP estimation or MLE. We developed a parallel implementation of our algorithm to
capitalize on conditional independencies across neurons, i.e. for the additive and multiplicative
STDP rules we can sample the weights W ??n independently of the weights W ??n0 . On the two
neuron examples we achieve upward of 2 iterations per second (sampling all variables in the model),
and we run our model for 1000 iterations. Convergence of the Markov chain is assessed by analyzing
the log posterior of the samples, and typically stabilizes after a few hundred iterations. As we scale
to networks of ten neurons, our running time quickly increases to roughly 20 seconds per iteration,
which is mostly dominated by slice sampling the learning rule parameters. In order to evaluate the
conditional probability of a learning rule parameter, we need to sample the weight trajectories for
each synapse. Though these running times are nontrivial, they are not prohibitive for networks that
are realistically obtainable for optical study of synaptic plasticity.
5.2
Biophysical simulations
Using the biophysical simulator NEURON, we performed two experiments. First, we considered a
network of 10 sparsely interconnected neurons (28 excitatory synapses) undergoing synaptic plasticity according to an additive STDP rule. Each neuron was driven independently by a hidden
population of 13 excitatory neurons and 5 inhibitory neurons connected to the visible neuron with
probability 0.8 and fixed synaptic weights averaging 3.0 mV. The visible synapses were initialized
close to 6.0 mV and allowed to vary between 0.0 and 10.5 mV. The synaptic delay was fixed at
1.0 ms for all synapses. This yielded a mean firing rate of 10 Hz among visible neurons. Synaptic weights were recorded every 1.0 ms. These parameters were chosen to demonstrate interesting
variations in synaptic strength, and as we transition to biological applications it will be necessary to
evaluate the sensitivity of the model to these parameters and the appropriate regimes for the circuits
under study.
We began by investigating whether the model is able to accurately identify synapses from spikes, or
whether it is confounded by spurious correlations. Figure 3 shows that our approach identifies the
28 excitatory synapses in our network, as measured by ROC curve (Add. STDP AUC=0.99), and
outperforms static models and cross-correlation. In the sparse, time-varying GLM, the probability
of an edge is measured by the mean of A under the posterior, whereas in the standard GLM with
MAP estimation, the likelihood of an edge is measured by area under the impulse response.
7
mV
12
Figure 4: Analogously to Figure 2, a sparse, time-varying GLM can capture the weight trajectories and learning
rules from spike trains simulated by NEURON. Here an excitatory synapse undergoes additive STDP with a
hard upper bound on the excitatory postsynaptic current. The weight trajectory inferred by our model with the
same parametric form of the learning rule matches almost exactly, whereas the static models must compromise
in order to capture early and late stages of the data, and the multiplicative weight exhibits qualitatively different
trajectories. Nevertheless, in terms of predictive log likelihood, we do not have enough information to correctly
determine the underlying learning rule. Potential solutions are discussed in the main text.
Looking into the synapses that are detected by the time-varying model and missed by the static
model, we find an interesting pattern. The improved performance comes from synapses that decay
in strength over the recording period. Three examples of these synaptic weight trajectories are shown
in the right panel of Figure 3. The time-varying model assigns over 90% probability to each of the
three synapses, whereas the static model infers less than a 40% probability for each synapse.
Finally, we investigated our model?s ability to distinguish various learning rules by looking at a
single synapse, analogous to the experiment performed on data from the GLM. Figure 4 shows
the results of a weight trajectory for a synapse under additive STDP with a strict threshold on the
excitatory postsynaptic current. The time-varying GLM with an additive model captures the same
trajectory, as shown in the left panel. The GLM weights have been linearly rescaled to align with the
true weights, which are measured in millivolts. Furthermore, the inferred additive STDP learning
rule, in particular the time constants and relative amplitudes, perfectly match the true learning rule.
These results demonstrate that a sparse, time-varying GLM is capable of discovering synaptic weight
trajectories, but in terms of predictive likelihood, we still have insufficient evidence to distinguish
additive and multiplicative STDP rules. By the end of the training period, the weights have saturated
at a level that almost surely induces postsynaptic spikes. At this point, we cannot distinguish two
learning rules which have both reached saturation. This motivates further studies that leverage this
probabilistic model in an optimal experimental design framework, similar to recent work by Shababo
et al. [19], in order to conclusively test hypotheses about synaptic plasticity.
6
Discussion
Motivated by the advent of optical tools for interrogating networks of synaptically connected neurons, which make it possible to study synaptic plasticity in novel ways, we have extended the GLM
to model a sparse, time-varying synaptic network, and introduced a fully-Bayesian inference algorithm built upon particle MCMC. Our initial results suggest that it is possible to infer weight
trajectories for a variety of biologically plausible learning rules.
A number of interesting questions remain as we look to apply these methods to biological recordings. We have assumed access to precise spike times, though extracting spike times from optical
recordings poses inferential challenges of its own. Solutions like those of Vogelstein et al. [20]
could be incorporated into our probabilistic model. Computationally, particle MCMC could be replaced with stochastic EM to achieve improved performance [18], and optimal experimental design
could aid in the exploration of stimuli to distinguish between learning rules. Beyond these direct extensions, this work opens up potential to infer latent state spaces with potentially nonlinear dynamics
and non-Gaussian noise, and to infer learning rules at the synaptic or even the network level.
Acknowledgments This work was partially funded by DARPA YFA N66001-12-1-4219 and NSF IIS1421780. S.W.L. was supported by an NDSEG fellowship and by the NSF Center for Brains, Minds, and
Machines.
8
References
[1] Adam M Packer, Darcy S Peterka, Jan J Hirtz, Rohit Prakash, Karl Deisseroth, and Rafael Yuste. Twophoton optogenetics of dendritic spines and neural circuits. Nature methods, 9(12):1202?1205, 2012.
[2] Daniel R Hochbaum, Yongxin Zhao, Samouil L Farhi, Nathan Klapoetke, Christopher A Werley, Vikrant
Kapoor, Peng Zou, Joel M Kralj, Dougal Maclaurin, Niklas Smedemark-Margulies, et al. All-optical
electrophysiology in mammalian neurons using engineered microbial rhodopsins. Nature methods, 2014.
[3] Christophe Andrieu, Arnaud Doucet, and Roman Holenstein. Particle Markov chain Monte Carlo methods. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 72(3):269?342, 2010.
[4] Liam Paninski. Maximum likelihood estimation of cascade point-process neural encoding models. Network: Computation in Neural Systems, 15(4):243?262, January 2004.
[5] Wilson Truccolo, Uri T. Eden, Matthew R. Fellows, John P. Donoghue, and Emery N. Brown. A point
process framework for relating neural spiking activity to spiking history, neural ensemble, and extrinsic
covariate effects. Journal of Neurophysiology, 93(2):1074?1089, 2005.
[6] Ian Stevenson and Konrad Koerding. Inferring spike-timing-dependent plasticity from spike train data. In
Advances in Neural Information Processing Systems, pages 2582?2590, 2011.
[7] Seif Eldawlatly, Yang Zhou, Rong Jin, and Karim G Oweiss. On the use of dynamic Bayesian networks
in reconstructing functional neuronal networks from spike train ensembles. Neural Computation, 22(1):
158?189, 2010.
[8] Biljana Petreska, Byron Yu, John P Cunningham, Gopal Santhanam, Stephen I Ryu, Krishna V Shenoy,
and Maneesh Sahani. Dynamical segmentation of single trials from population neural data. In Neural
Information Processing Systems, pages 756?764, 2011.
[9] Mike West, P Jeff Harrison, and Helio S Migon. Dynamic generalized linear models and Bayesian forecasting. Journal of the American Statistical Association, 80(389):73?83, 1985.
[10] T. J. Mitchell and J. J. Beauchamp. Bayesian Variable Selection in Linear Regression. Journal of the
American Statistical Association, 83(404):1023?-1032, 1988.
[11] James Robert Lloyd, Peter Orbanz, Zoubin Ghahramani, and Daniel M Roy. Random function priors
for exchangeable arrays with applications to graphs and relational data. Advances in Neural Information
Processing Systems, 2012.
[12] Natalia Caporale and Yang Dan. Spike timing-dependent plasticity: a Hebbian learning rule. Annual
Review of Neuroscience, 31:25?46, 2008.
[13] Erkki Oja. Simplified neuron model as a principal component analyzer. Journal of Mathematical Biology,
15(3):267?273, 1982.
[14] Daniel E Feldman. The spike-timing dependence of plasticity. Neuron, 75(4):556?71, August 2012.
[15] S Song, K D Miller, and L F Abbott. Competitive Hebbian learning through spike-timing-dependent
synaptic plasticitye. Nature Neuroscience, 3(9):919?26, September 2000. ISSN 1097-6256.
[16] Abigail Morrison, Markus Diesmann, and Wulfram Gerstner. Phenomenological models of synaptic
plasticity based on spike timing. Biological cybernetics, 98(6):459?478, 2008.
[17] Anne C Smith and Emery N Brown. Estimating a state-space model from point process observations.
Neural Computation, 15(5):965?91, May 2003.
[18] Fredrik Lindsten, Michael I Jordan, and Thomas B Sch?on. Ancestor sampling for particle Gibbs. In
Advances in Neural Information Processing Systems, pages 2600?2608, 2012.
[19] Ben Shababo, Brooks Paige, Ari Pakman, and Liam Paninski. Bayesian inference and online experimental
design for mapping neural microcircuits. In Advances in Neural Information Processing Systems, pages
1304?1312, 2013.
[20] Joshua T Vogelstein, Brendon O Watson, Adam M Packer, Rafael Yuste, Bruno Jedynak, and Liam Paninski. Spike inference from calcium imaging using sequential Monte Carlo methods. Biophysical journal,
97(2):636?655, 2009.
9
| 5274 |@word neurophysiology:1 trial:1 middle:2 interleave:1 open:1 simulation:2 propagate:2 bn:3 dramatic:1 fifteen:1 accommodate:1 deisseroth:1 initial:1 series:3 daniel:3 hirtz:1 outperforms:2 existing:1 current:7 anne:1 written:3 must:4 john:2 additive:17 visible:3 plasticity:31 interspike:1 update:9 n0:21 stationary:2 leaf:1 prohibitive:1 discovering:1 shababo:2 smith:1 record:1 filtered:1 provides:1 detecting:1 revisited:1 beauchamp:1 simpler:1 unbounded:2 mathematical:1 direct:3 become:1 qualitative:1 dan:1 peng:1 inter:1 spine:1 roughly:1 examine:1 nor:1 simulator:3 brain:3 multi:1 discretized:1 little:1 armed:1 window:1 increasing:1 becomes:1 discover:1 underlying:4 monosynaptic:1 circuit:3 rn0:5 advent:1 bounded:3 panel:2 easiest:1 yongxin:1 interpreted:1 depolarization:1 developed:2 lindsten:1 temporal:1 thorough:1 every:2 fellow:1 prakash:1 concave:1 exactly:1 scaled:1 exchangeable:1 normally:1 shenoy:1 understood:1 timing:10 treat:1 modify:1 limit:1 local:1 switching:1 despite:2 bilinear:1 encoding:1 analyzing:1 firing:7 black:2 might:1 studied:1 challenging:1 limited:1 factorization:1 liam:3 jedynak:1 acknowledgment:1 testing:1 block:1 implement:1 jan:1 area:1 maneesh:1 significantly:1 cascade:1 inferential:1 pre:2 induce:1 integrating:1 suggest:1 zoubin:1 cannot:1 close:3 selection:1 mea:2 context:1 live:1 collapsed:1 restriction:1 map:6 center:1 duration:2 convex:2 independently:3 resolution:2 simplicity:1 assigns:1 rule:61 array:2 dw:1 population:7 variation:1 analogous:1 play:1 today:1 strengthen:1 sn0:12 hypothesis:3 harvard:6 trend:1 roy:1 expensive:1 updating:1 mammalian:1 sparsely:2 observed:2 role:1 bottom:2 mike:1 capture:5 connected:6 decrease:1 highest:1 rescaled:1 substantial:2 disease:1 intuition:1 covariates:1 dynamic:10 koerding:2 depend:1 compromise:2 predictive:5 upon:4 efficiency:1 basis:3 joint:1 darpa:1 stock:1 represented:1 various:1 train:12 shortcoming:1 monte:3 detected:1 outside:2 whose:1 larger:2 valued:1 supplementary:1 plausible:1 otherwise:2 optogenetics:1 ability:1 jointly:1 final:1 online:1 biophysical:7 interconnected:3 interaction:3 strengthening:1 product:2 combining:1 realization:1 kapoor:1 flexibility:1 achieve:2 realistically:1 description:1 validate:1 convergence:2 electrode:2 double:1 plethora:1 sea:2 congruent:1 generating:1 adam:3 emery:2 natalia:1 object:1 ben:1 develop:1 propagating:1 pose:2 measured:4 job:1 strong:1 implemented:1 fredrik:1 come:2 correct:2 filter:3 stochastic:1 exploration:1 engineered:1 material:1 adjacency:4 require:1 truccolo:1 biological:5 ryan:1 summation:1 dendritic:2 underdetermined:1 extension:1 rong:1 considered:1 ground:2 stdp:29 maclaurin:1 mapping:1 slab:1 circuitry:1 stabilizes:1 matthew:1 vary:2 early:1 a2:1 omitted:1 resample:1 estimation:7 applicable:1 tool:4 clearly:1 gaussian:7 always:1 gopal:1 zhou:1 lifted:1 varying:24 timevarying:1 broader:1 wilson:1 bernoulli:1 likelihood:14 indicates:1 baseline:1 sense:1 detect:1 inference:17 dependent:6 factoring:1 integrated:1 typically:1 cunningham:1 hidden:2 spurious:1 microbial:1 ancestor:2 rpa:1 interested:1 upward:1 among:4 priori:1 marginal:1 construct:1 never:1 sampling:7 identical:1 biology:1 capitalize:3 look:1 nearly:1 yu:1 mimic:2 future:1 stimulus:7 roman:1 few:2 oja:3 simultaneously:1 packer:2 replaced:1 an0:8 ourselves:1 detection:1 neuroscientist:1 dougal:1 highly:2 investigate:1 evaluation:2 joel:1 saturated:1 yfa:1 regularizers:1 held:1 chain:6 edge:2 capable:2 necessary:2 initialized:1 increased:1 column:1 modeling:2 extensible:1 assignment:1 cost:2 deviation:1 entry:4 hundred:1 delay:1 kn:4 dependency:1 synthetic:2 combined:1 fundamental:2 sensitivity:1 probabilistic:3 michael:1 analogously:1 quickly:1 connectivity:5 reflect:1 recorded:2 central:1 ndseg:1 cognitive:1 tz:1 american:2 zhao:1 leading:1 account:1 potential:9 stevenson:2 lloyd:1 coefficient:1 explicitly:1 mv:5 depends:2 multiplicative:13 performed:2 reached:1 competitive:1 recover:1 complicated:1 parallel:1 rectifying:1 likewise:1 efficiently:1 yield:3 identify:1 ensemble:2 miller:1 bayesian:9 accurately:1 carlo:3 trajectory:29 cybernetics:1 history:3 holenstein:1 simultaneous:1 synapsis:14 explain:1 synaptic:62 definition:1 pp:1 james:1 naturally:1 associated:2 proof:1 static:10 sampled:1 dataset:1 mitchell:1 realm:1 knowledge:1 improves:1 infers:1 segmentation:1 amplitude:4 obtainable:1 sophisticated:1 dt:4 follow:1 methodology:1 response:9 improved:2 synapse:14 erd:1 done:1 evaluated:1 though:3 microcircuit:1 furthermore:1 stage:1 until:2 glms:3 correlation:4 christopher:2 wmax:5 nonlinear:10 o:1 expressive:1 reweighting:1 undergoes:1 indicated:1 impulse:9 stimulate:1 building:1 effect:6 normalized:5 requiring:1 true:5 concept:1 evolution:3 regularization:1 assigned:1 andrieu:1 arnaud:1 brown:2 nonzero:1 iteratively:1 karim:1 biljana:1 konrad:1 self:2 auc:1 m:2 generalized:5 prominent:1 demonstrate:2 l1:2 meaning:1 instantaneous:1 novel:1 recently:3 ari:1 began:1 common:1 functional:1 spiking:8 overview:1 refractory:2 khz:1 extend:1 interpretation:1 occurred:1 approximates:1 interpret:1 discussed:1 relating:1 refer:1 association:2 cambridge:3 gibbs:3 feldman:1 dbn:2 similarly:2 particle:29 nonlinearity:1 analyzer:1 bruno:1 dot:2 funded:1 phenomenological:1 access:2 add:1 align:1 posterior:3 own:1 recent:1 perspective:2 orbanz:1 apart:1 driven:2 certain:1 binary:3 watson:1 christophe:1 joshua:1 peterka:1 krishna:1 minimum:1 greater:1 preceding:1 surely:1 determine:1 period:2 vogelstein:2 semi:1 stephen:1 morrison:1 infer:3 hebbian:2 exceeds:2 match:2 characterized:1 plausibility:1 believed:1 cross:2 compensate:1 pakman:1 post:4 mle:2 estimating:1 a1:1 variant:6 regression:1 poisson:1 iteration:4 normalization:3 represent:1 hochbaum:1 synaptically:2 cell:4 addition:2 background:2 whereas:5 separately:1 interval:2 fellowship:1 harrison:1 grow:1 crucial:1 sch:1 unlike:1 strict:1 subject:1 recording:9 induced:2 hz:3 byron:1 jordan:1 extracting:1 nonstationary:1 consult:1 presence:1 leverage:3 intermediate:1 yang:2 identically:1 wn:2 enough:1 variety:4 psps:2 counterbalanced:1 fit:3 competing:1 identified:1 lasso:1 reduce:1 idea:1 perfectly:1 donoghue:1 whether:2 motivated:1 caporale:1 effort:1 forecasting:1 song:1 peter:1 paige:1 speaking:1 clear:1 detailed:1 ten:2 induces:1 simplest:1 exist:1 canonical:2 inhibitory:4 nsf:2 neuroscience:3 extrinsic:1 popularity:2 rb:2 per:2 correctly:1 discrete:1 santhanam:1 group:1 key:1 four:3 independency:1 threshold:7 pb:1 nevertheless:1 eden:1 changing:1 neither:1 breadth:1 thresholded:1 millivolt:1 abbott:1 n66001:1 imaging:1 graph:1 sum:3 convert:1 year:1 run:1 parameterized:1 powerful:2 extends:1 almost:2 reader:1 missed:1 scaling:1 comparable:1 bound:4 distinguish:5 annual:1 nonnegative:3 activity:4 yielded:2 adapted:1 occur:2 strength:4 precisely:1 eldawlatly:2 constraint:1 incorporation:1 nontrivial:1 erkki:1 markus:1 dominated:1 diesmann:1 nathan:1 simulate:1 min:1 formulating:1 performing:1 optical:6 relatively:2 according:5 combination:3 membrane:5 petreska:2 slightly:1 increasingly:1 postsynaptic:5 across:1 remain:1 em:1 reconstructing:1 biologically:2 invariant:1 pr:3 glm:30 computationally:1 equation:5 previously:1 turn:1 mechanism:1 mind:1 merit:1 end:1 confounded:1 studying:2 available:1 linderman:1 apply:1 enforce:2 appropriate:1 alternative:1 thomas:1 assumes:1 remaining:1 running:3 top:2 opportunity:1 ghahramani:1 especially:1 society:1 question:1 spike:40 parametric:3 dependence:1 traditional:4 exhibit:1 september:1 distance:1 separate:1 link:1 simulated:4 sensible:1 presynaptic:2 considers:1 collected:1 issn:1 modeled:3 insufficient:1 providing:1 innovation:1 mostly:1 robert:1 potentially:8 rise:1 design:4 implementation:2 motivates:1 calcium:1 perform:4 upper:3 neuron:39 convolution:1 markov:4 observation:1 jin:1 truncated:2 january:1 extended:2 looking:2 precise:1 incorporated:1 niklas:1 interacting:1 relational:1 arbitrary:1 august:1 inferred:5 introduced:1 evidenced:1 pair:1 namely:1 specified:1 extensive:1 connection:3 ryu:1 brook:1 able:3 beyond:1 alongside:1 dynamical:7 pattern:2 scott:1 below:4 regime:2 sparsity:2 challenge:2 encompasses:1 saturation:1 built:1 including:1 memory:2 royal:1 belief:1 power:1 regularized:1 mn:2 technology:1 identifies:1 negativity:1 sn:19 sahani:1 text:1 prior:5 understanding:1 literature:3 review:1 marginalizing:1 relative:2 rohit:1 embedded:1 fully:3 interesting:3 limitation:1 filtering:5 yuste:2 oweiss:1 thresholding:1 hn0:4 row:4 karl:1 excitatory:9 penalized:1 pmcmc:2 supported:1 last:1 allow:1 understand:1 fall:1 barrier:1 wmin:4 sparse:7 distributed:2 slice:1 boundary:1 curve:1 transition:2 rich:1 forward:3 made:2 collection:1 qualitatively:1 simplified:1 far:1 approximate:1 implicitly:2 rafael:2 conclusively:1 doucet:1 incoming:2 investigating:1 assumed:1 swl:1 latent:2 decade:2 promising:1 nature:7 expansion:1 warranted:2 excellent:1 complex:2 necessarily:1 investigated:1 domain:1 zou:1 gerstner:1 main:1 linearly:1 motivation:1 noise:7 bounding:1 allowed:1 neuronal:1 west:2 roc:1 darker:1 aid:1 inferring:4 exponential:1 governed:1 third:1 late:1 renyi:1 ian:1 specific:1 covariate:1 showing:1 undergoing:5 explored:1 decay:2 a3:1 evidence:1 sequential:1 effectively:1 importance:1 magnitude:2 illustrates:1 uri:1 led:1 electrophysiology:1 simply:2 likely:1 explore:1 paninski:3 expressed:1 saturating:2 partially:1 collectively:1 truth:2 determines:1 ma:3 conditional:7 formulated:1 jeff:1 absence:1 change:8 hard:3 wulfram:1 except:3 averaging:1 principal:1 called:1 total:1 pas:1 experimental:5 attempted:1 indicating:1 wn0:11 college:1 preand:1 support:1 arises:1 modulated:1 assessed:1 incorporate:1 evaluate:3 mcmc:10 tested:1 |
4,721 | 5,275 | Global Belief Recursive Neural Networks
Romain Paulus, Richard Socher?
MetaMind
Palo Alto, CA
{romain,richard}@metamind.io
Christopher D. Manning
Stanford University
353 Serra Mall
Stanford, CA 94305
[email protected]
Abstract
Recursive Neural Networks have recently obtained state of the art performance on
several natural language processing tasks. However, because of their feedforward
architecture they cannot correctly predict phrase or word labels that are determined by context. This is a problem in tasks such as aspect-specific sentiment
classification which tries to, for instance, predict that the word Android is positive
in the sentence Android beats iOS. We introduce global belief recursive neural
networks (GB-RNNs) which are based on the idea of extending purely feedforward neural networks to include one feedbackward step during inference. This
allows phrase level predictions and representations to give feedback to words. We
show the effectiveness of this model on the task of contextual sentiment analysis. We also show that dropout can improve RNN training and that a combination
of unsupervised and supervised word vector representations performs better than
either alone. The feedbackward step improves F1 performance by 3% over the
standard RNN on this task, obtains state-of-the-art performance on the SemEval
2013 challenge and can accurately predict the sentiment of specific entities.
1
Introduction
Models of natural language need the ability to compose the meaning of words and phrases in order
to understand complex utterances such as facts, multi-word entities, sentences or stories. There has
recently been a lot of work extending single word semantic vector spaces [27, 11, 15] to compositional models of bigrams [16, 29] or phrases of arbitrary length [25, 28, 24, 10]. Work in this
area so far has focused on computing the meaning of longer phrases in purely feedforward types
of architectures in which the meaning of the shorter constituents that are being composed is not
altered. However, a full treatment of semantic interpretation cannot be achieved without taking into
consideration that the meaning of words and phrases can also change once the sentence context is
observed. Take for instance the sentence in Fig. 1: The Android?s screen is better than the iPhone?s.
All current recursive deep learning sentiment models [26] would attempt to classify the phrase The
Android?s screen or than the iPhone?s, both of which are simply neutral. The sentiment of the overall sentence is undefined; it depends on which of the entities the user of the sentiment analysis cares
about. Generally, for many analyses of social media text, users are indeed most interested in the
sentiment directed towards a specific entity or phrase.
In order to solve the contextual classification problem in general and aspect-specific sentiment classification in particular, we introduce global belief recursive neural networks (GB-RNN). These models
generalize purely feedforward recursive neural networks (RNNs) by including a feedbackward step
at inference time. The backward computation uses the representations from both steps in its recursion and allows all phrases, to update their prediction based on the global context of the sentence.
Unlike recurrent neural networks or window-based methods [5] the important context can be many
?
Part of this research was performed while the author was at Stanford University.
1
?
-
0
Android
0
0
beats
iOS
Figure 1: Illustration of the problem of sentiment classification that uses only the phrase to be labeled
and ignores the context. The word Android is neutral in isolation but becomes positive in context.
words away from the phrase that is to be labeled. This will allow models to correctly classify that in
the sentence of Fig. 1, Android is described with positive sentiment and iOS was not. Neither was
possible to determine only from their respective phrases in isolation.
In order to validate the GB-RNN?s ability to contextually disambiguate sentiment on real text, we
use the Twitter dataset and annotations from Semeval Challenge 2013 Task 2.1 The GB-RNN outperforms both the standard RNN and all other baselines, as well the winner of the Sentiment competition of SemEval 2013, showing that it can successfully make use of surrounding context.
2
Related Work
Neural word vectors One common way to represent words is to use distributional word vectors
[27] learned via dimensionality reduction of large co-occurrence matrices over documents (as in
latent semantic analysis [13]), local context windows [15, 18] or combinations of both [11]. Words
with similar meanings are close to each other in the vector space. Since unsupervised word vectors computed from local context windows do not always encode task-specific information, such
as sentiment, word vectors can also be fine-tuned to such specific tasks [5, 24]. We introduce a
hybrid approach where some dimensions are obtained from an unsupervised model and others are
learned for the supervised task. We show that this performs better than both the purely supervised
and unsupervised semantic word vectors.
Recursive Neural Networks The idea of recursive neural networks (RNNs) for natural language
processing (NLP) is to train a deep learning model that can be applied to inputs of any length.
Unlike computer vision tasks, where it is easy to resize an image to a fixed number of pixels, natural sentences do not have a fixed size input. However, phrases and sentences have a grammatical
structure that can be parsed as a binary tree [22].
Following this tree structure, we can assign a fixed-length vector to each word at the leaves of
the tree, and combine word and phrase pairs recursively to create intermediate node vectors of the
same length, eventually having one final vector representing the whole sentence [19, 25]. Multiple
recursive combination functions have been explored, from linear transformation matrices to tensor
products [26]. In this work, we use the simple single matrix RNN to combine node vectors at each
recursive step.
Bidirectional-recurrent and bidirectional-recursive neural networks. Recurrent neural networks
are a special case of recursive neural networks that operate on chains and not trees. Unlike recursive
neural networks, they don?t require a tree structure and are usually applied to time series. In a recurrent neural network, every node is combined with a summarized representation of the past nodes
[8], and then the resulting combination will be forwarded to the next node. Bidirectional recurrent neural network architectures have also been explored [21] and usually compute representations
independently from both ends of a time series.
Bidirectional recursive models [12, 14], developed in parallel with ours, extend the definition of the
recursive neural netword by adding a backward propagation step, where information also flows from
the tree root back to the leaves. We compare our model to theirs theoretically in the model section,
and empirically in the experiments.
1
http://www.cs.york.ac.uk/semeval-2013/task2/
2
Figure 2: Propagation steps of the GB-RNN. Step 1 describes the standard RNN feedforward process, showing that the vector representation of ?Android? is independent of the rest of the document.
Step 2 computes additional vectors at each node (in red), using information from the higher level
nodes in the tree (in blue), allowing ?Android? and ?iOS? to have different representations given the
context.
[20] unfold the same autoencoder multiple times which gives it more representational power with
the same number of parameters. Our model is different in that it takes into consideration more
information at each step and can eventually make better local predictions by using global context.
Sentiment analysis. Sentiment analysis has been the subject of research for some time [4, 2, 3, 6,
1, 23]. Most approaches in sentiment analysis use ?bag of words? representations that do not take
the phrase structure into account but learn from word-level features. We explore our model?s ability
to determine contextual sentiment on Twitter, a social media platform.
3
Global Belief Recursive Neural Networks
In this section, we introduce a new model to compute context-dependent compositional vector representations of variable length phrases. These vectors are trained to be useful as features to classify
each phrase and word. Fig. 2 shows an example phrase computation that we will describe in detail
below. This section begins by motivating compositionality and context-dependence, followed by a
definition of standard recursive neural networks. Next, we introduce our novel global belief model
and hybrid unsupervised-supervised word vectors.
3.1
Context-Dependence as Motivation for Global Belief
A common simplifying assumption when mapping sentences into a feature vector is that word order
does not matter (?bag of words?). However, this will prevent any detailed understanding of language
as exemplified in Fig. 1, where the overall sentiment of the phrase ?Android beats iOS?, is unclear.
Instead, we need an understanding of each phrase which leads us to deep recursive models.
The first step for mapping a sentence into a vector space is to parse them into a binary tree structure
that captures the grammatical relationships between words. Such an input dependent binary tree then
determines the architecture of a recursive neural network which will compute the hidden vectors in a
bottom-up fashion starting with the word vectors. The resulting phrase vectors are given as features
to a classifier. This standard RDL architecture works well for classifying the inherent or contextindependent label of a phrase. For instance, it can correctly classify that a not so beautiful day is
negative in sentiment. However, not all phrases have an inherent sentiment as shown in Fig. 1.
The GB-RNN addresses this issue by propagating information from the root node back to the
leaf nodes as described below. There are other ways context can be incorporated such as with
bi-directional recurrent neural networks or with window-based methods. Both of these methods,
however, cannot incorporate information from words further away from the phrase to be labeled.
3.2
Standard Recursive Neural Networks
We first describe a simple recursive neural network that can be used for context-independent phraselevel classification. It can also be seen as the first step of a GB-RNN.
3
Assume, for now, that each word vector a ? Rn is obtained by sampling each element from a
uniform distribution: ai ? U(?0.001, 0.001). All these vectors are columns of a large embedding
matrix L ? Rn?|V | , where |V | is the size of the vocabulary. All word vectors are learned together
with the model.
For the example word vector sequence (abc) of Fig. 2, the RNN equations become:
b
a
p1 = f W
, p2 = f W
,
c
p1
(1)
where W ? Rn?2n is the matrix governing the composition and f the non-linear activation function. Each node vector is the given as input to a softmax classifier for a classification task such as
sentiment analysis.
3.3
GB-RNN: Global Belief Recursive Neural Networks
Our goal is to include contextual information in the recursive node vector representations. One
simple solution would be to just include the k context words to the left and right of each pair as in
[25]. However, this will only work if the necessary context is at most k words away. Furthermore,
in order to capture more complex linguistic phenomena it may be necessary to allow for multiple
words to compose the contextual shift in meaning. Instead, we will use the feedforward nodes from
a standard RNN architecture and simply move back down the tree. This can also be interpreted as
unfolding the tree and moving up its branches.
Hence, we keep the same Eq. 1 for computing the forward node vectors, but we introduce new
feedbackward vectors, denoted with a down arrow ? , at every level of the parse tree. Unlike the
feedforward vectors, which were computed with a bottom-up recursive function, feedbackward vectors are computed with a top-down recursive function. The backwards pass starts at the root node
and propagates all the way down to the single word vectors. At the root note, in our example the
node p2 , we have:
p?2 = f (V p2 ) ,
(2)
where V ? Vnd ?n so that all ?-node vectors are nd -dimensional. Starting from p?2 , we recursively
get ?-node vectors for every node as we go down the tree:
?
?
p2
p1
a
b
?
?
=
f
W
,
=
f
W
(3)
?
?
c?
p2
p1
p?1
where all ?-vectors, are nd -dimensional and hence W ? ? R(n+nd )?(n+nd ) is a new de-composition
matrix. Figure 2 step 2 illustrates this top-down recursive computation on our example. Once we
have both feedforward and feedbackward vectors for a given node, we concatenate them and employ
the standard softmax classifier
the final prediction. For instance, the classification for word
to make
a
a becomes: ya = softmax Wc
, where we fold the bias into the C-class classifier weights
a?
Wc ? RC?(n+1) .
At the root node, the equation for x?root could be replaced by simply copying x?root = xroot . But
there are two advantages of introducing a transform matrix V . First, it helps clearly differentiating features computed during the forward step and the backward step in multiplication with W ? .
Second, it allows to use a different dimension for the x? vectors, which reduces the number of parameters in the W ? and Wclass matrices, and adds more flexibility to the model. It also performs
better empirically.
3.4
Hybrid Word Vector Representations
There are two ways to initialize the word vectors that are given as inputs to the RNN models. The
simplest one is to initialize them to small random numbers as mentioned above and backpropagate
error signals into them in order to have them capture the necessary information for the task at hand.
This has the advantage of not requiring any other pre-training method and the vectors are sure to
capture domain knowledge. However, the vectors are more likely to overfit and less likely to generalize well to words that have not been in the (usually smaller) labeled training set. Another approach
4
Figure 3: Hybrid unsupervised-supervised vector representations for the most frequent 50 words
of the dataset. For each horizontal vector, the first 100 dimensions are trained on unlabeled twitter
messages, and the last dimensions are trained on labeled contextual sentiment examples.
is to use unsupervised methods that learn semantic word vectors such as [18]. One then has the
option to backpropagate task specific errors into these vectors or keep them at their initialization.
Backpropagating into them still has the potential disadvantage of hurting generalization apart from
slowing down training since it increases the number of parameters by a large amount (there are usually 100, 000 ? 50 many parameters in the embedding matrix L). Without propagating information
however one has to hope that the unsupervised method really captures all the necessary semantic
information which is often not the case for sentiment (which suffers from the antonym problem).
In this paper we propose to combine both ideas by representing each word as a concatenation of both
unsupervised vectors that are kept at their initialization during training and adding a small additional
vector into which we propagate the task specific error signal. This vector representation applies only
to the feedforward word vectors and shold not be confused with the combination of the feedwordard
and feedbackward node vectors in the softmax.
Figure 3.4 shows the resulting word vectors trained on unlabeled documents on one part (the first
100 dimensions), and trained on labeled examples on the other part (the remaining dimensions).
3.5
Training
The GB-RNN is trained by using backpropagation through structure [9]. We train the parameters by
optimizing the regularized cross-entropy error for labeled node vectors with mini-batched AdaGrad
[7]. Since we don?t have labels for every node of the training trees, we decided that unlabeled
nodes do not add an additional error during training. For all models, we use a development set to
cross-validate over regularization of the different weights, word vector size, mini-batch size, dropout
probability and activation function (rectified linear or logistic function).
We also applied the dropout technique to improve training with high dimensional word vectors.
Node vector units are randomly set to zero with a probability of 0.5 at each training step. Our
experiments show that applying dropout in this way helps differentiating word vector units and
hidden units, and leads to better performance. The high-dimensional hybrid word vectors that we
introduced previously have obtained a higher accuracy than other word vectors with the use of
dropout.
3.6
Comparison to Other Models
The idea of unfolding of neural networks is commonly used in autoencoders as well as in a recursive
setting [23], in this setting the unfolding is only used during training and not at inference time to
update the beliefs about the inputs.
Irsoy and Cardie [12] introduced a bidirectional RNN similar to ours. It employs the same standard
feedforward RNN, but a different computation for the backward ? vectors. In practice, their model is
defined by the same forward equations as ours. However, equation 3 which computes the backward
vectors is instead:
?
b
V b + Wlb? p?1
=f
(4)
? ?
c?
V c + Wrb
p1
5
Correct FUSION?s 5th General Meeting is tonight at 7 in ICS 213! Come out and carve pumpkins mid-quarter
with us!
Correct I would rather eat my left foot then to be taking the SATs tomorrow
Correct Special THANKS to EVERYONE for coming out to Taboo Tuesday With DST tonight! It was
FUN&educational!!! :) @XiEtaDST
Correct Tough loss for @statebaseball today. Good luck on Monday with selection Sunday
Correct I got the job at Claytons!(: I start Monday doing Sheetrock(: #MoneyMakin
Correct St Pattys is no big deal for me, no fucks are given, but Cinco De Mayo on the other hand .. thats my
2nd bday .
Incorrect ?@Hannah Sunder: The Walking Dead is just a great tv show? its bad ass just started to watch the
2nd season to catch up with the 3rd
Figure 4: Examples of predictions made by the GB-RNN for twitter documents. In this example,
red phrases are negative and blue phrases are positive. On the last example, the model predicted
incorrectly ?bad ass? as negative.
?
Where Wlb? and Wrb
are two matrices with dimensions nd ? nd . For a better comparison with our
model we rewrite Eq. 3 and make explicit the 4 blocks of W ? :
"
#
"
#!
?
?
?
Wlf
Wlb?
Wlf
p1 + Wlb? p?1
b
?
Let W =
, then
=f
,
(5)
?
?
?
? ?
c?
Wrf
Wrb
Wrf
p1 + Wrb
p1
?
?
?
?
where the dimensions of Wlf
and Wrf
are nd ? n, and the dimensions of Wld
and Wrd
are nd ? nd .
A closer comparison between Eqs. 4 and 5 reveals that both use a left and right forward transfor?
?
mation Wlf
p1 and Wrf
p1 , but the other parts of the sums differ. In the bidirectional-RNN, the
transformation of any children is defined by the forward parent and independent on its position (left
or right node). Whereas our GB-RNN makes uses of both the forward and backward parent node.
The intuition behind our choice is that using both nodes helps to push the model to disentangled
the children from their backward parent vector. We also note that our model does not use the forward node vector for computing the backward node vector, but we find this not necessary since the
softmax function already combines the two vectors.
Our model also has n ? nd more parameters to compute the feedbackward vectors than the
bidirectional-RNN. The W ? matrix of our model has 2n2d + 2n ? nd parameters, while the other
?
?
model has a total of 2n2d + n ? nd parameters with the Wlf
, Wrf
and V matrices. We show in the
next section that GB-RNN outperforms the bidirectional RNN in our experiments.
4
Experiments
We present a qualitative and quantitative analysis of the GB-RNN on a contextual sentiment classification task. The main dataset is provided by the SemEval 2013, Task 2 competition [17]. We
outperform the winners of the 2013 challenge, as well as several baseline and model ablations.
4.1
Evaluation Dataset
The SemEval competition dataset is composed of tweets labeled for 3 different sentiment classes:
positive, neutral and negative. The tweets in this dataset were split into a train (7862 labeled phrases),
development (7862) and development-test (7862) set. The final test set is composed of 10681 examples. Fig. 4 shows example GB-RNN predictions on phrases marked for classification in this dataset.
The development dataset consists only of tweets whereas the final evaluation dataset included also
short text messages (SMS in the tables below).
Tweets were parsed using the Stanford Parser [22] which includes tokenizing of negations (e.g.,
don?t becomes two tokens do and n?t). We constrained the parser to keep each phrase labeled by the
dataset inside its own subtree, so that each labeled example is represented by a single node and can
be classified easily.
6
Classifier
SVM
SVM
SVM
GB-RNN
Feature Sets
stemming, word cluster, SentiWordNet
score, negation
POS, lexicon, negations, emoticons,
elongated words, scores, syntactic dependency, PMI
punctuation, word n-grams, emoticons,
character n-grams, elongated words,
upper case, stopwords, phrase length,
negation, phrase position, large sentiment lexicons, microblogging features
parser, unsupervised word vectors (ensemble)
Twitter 2013 (F1)
85.19
SMS 2013 (F1)
88.37
87.38
85.79
88.93
88.00
89.41
88.40
Table 1: Comparison to the best Semeval 2013 Task 2 systems, their feature sets and F1 results on
each dataset for predicting sentiment of phrases in context. The GB-RNN obtains state of the art
performance on both datasets.
Model
Bigram Naive Bayes
Logistic Regression
SVM
RNN
Bidirectional-RNN (Irsoy and Cardie)
GB-RNN (best single model)
Twitter 2013
80.45
80.91
81.87
82.11
85.77
86.80
SMS 2013
78.53
80.37
81.91
84.07
84.77
87.15
Table 2: Comparison with baselines: F1 scores on the SemEval 2013 test datasets.
4.2 Comparison with Competition Systems
The first comparison is with several highly tuned systems from the SemEval 2013, Task 2 competition. The competition was scored by an average of positive and negative class F1 scores. Table 1
lists results for several methods, together with the resources and features used by each method. Most
systems used a considerable amount of hand-crafted features. In contrast, the GB-RNN only needs
a parser for the tree structure, unsupervised word vectors and training data. Since the competition
allowed for external data we outline below the additional training data we use. Our best model is an
ensemble of the top 5 GB-RNN models trained independently. Their predictions were then averaged
to produce the final output.
4.3 Comparison with Baselines
Next we compare our single best model to several baselines and model ablations. We used the same
hybrid word vectors with dropout training for the RNN, the bidirectional RNN and the GB-RNN.
The best models were selected by cross-validating on the dev set for several hyper-parameters (word
vectors dimension, hidden node vector dimension, number of training epochs, regularization parameters, activation function, training batch size and dropout probability) and we kept the models with
the highest cross-validation accuracy. Table 2 shows these results. The most important comparison
is against the purely feedforward RNN which does not take backward sentence context into account.
This model performs over 5% worse than the GB-RNN.
For the logistic regression and Bigram Naive Bayes classification, each labeled phrase was taken
as a separate example, removing the surrounding context. Another set of baselines used a context
window for classification as well as the entire tweet as input to the classifier.
Optimal performance for the single best GB-RNN was achieved by using vector sizes of 130 dimensions (100 pre-trained, fixed word vectors and 30 trained on sentiment data), a mini-batch size of
30, dropout with p = 0.5 and sigmoid non-linearity. In table 3, we show that the concatenation of
fixed, unsupervised vectors with additional randomly initialized, supervised vectors performs better
than both methods.
4.4 Model Analysis: Additional Training Data
Because the competition allowed the usage of arbitrary resources we included as training data labeled unigrams and bigrams extracted from the NRC-Canada system?s sentiment lexicon. Adding
these additional training examples increased accuracy by 2%. Although this lexicon helps reduc7
Word vectors
supervised word vectors
semantic word vectors
hybrid word vectors
dimension
15
100
100 + 34
Twitter 2013
85.15
85.67
86.80
SMS 2013
85.66
84.70
87.15
Table 3: F1 score comparison of word vectors on the SemEval 2013 Task 2 test dataset.
-
-
-
-
Chelski
-
-
+
-
+
want
-
this
+
Chelski
+
+
+
that it
-
-
so
bad
+
+
+
+
want +
this +
so
+
makes-
+
me
+
even
+
that +
it
-
bad
+
-
-
+
+
+
me
+
+
+
+
+
even
+
+
thinking+
+
we +
may
happier
-
+
-
+
beat
+
makes
+
+
thinking+
we may
happier
+
+
+
-
-
twice
them
+
+
+
+
in
+
+
beat
-
-
-
-
4
days
at
+
SB
+
+
+
+
twice
them
+
in
+
+
+
+
+
+
4
days
at
SB
Figure 5: Change in sentiment predictions in the tweet chelski want this so bad that it makes me even
happier thinking we may beat them twice in 4 days at SB between the RNN (left) and the GB-RNN
(right). In particular, we can see the change for the phrase want this so bad where it is correctly
predicted as positive with context.
ing the number of unknown tokens, it does not do a good job for training recursive composition
functions, because each example is short.
We also included our own dataset composed 176,311 noisily labeled tweets (using heuristics such as
smiley faces) as well as the movie reviews dataset from [26]. In both datasets the labels only denote
the context-independent sentiment of a phrase or full sentence. Hence, we trained the final model in
two steps: train the standard RNN, then train the full GB-RNN model on the smaller context-specific
competition data. Training the GB-RNN jointly in this fashion gave a 1% accuracy improvement.
5
Conclusion
We introduced global belief recursive neural networks, applied to the task of contextual sentiment
analysis. The idea of propagating beliefs through neural networks is a powerful and important piece
for interpreting natural language. The applicability of this idea is more general than RNNs and can
be helpful for a variety of NLP tasks such as word-sense disambiguation.
Acknowledgments
We thank the anonymous reviewers for their valuable comments.
References
[1] B.R. Routledge B. O?Connor, R. Balasubramanyan and N.A. Smith. From tweets to polls:
Linking text sentiment to public opinion time series. International AAAI Conference on Weblogs and Social Media, 2010.
[2] L. Barbosa and J. Feng. Robust sentiment detection on twitter from biased and noisy data.
COLING ?10 Proceedings of the 23rd International Conference on Computational Linguistics:
Posters, pages 36?44, 2010.
[3] A. Bifet and E. Frank. Sentiment knowledge discovery in twitter streaming data. Proceedings
of the 13th international conference on Discovery science, 2010.
[4] K. Sobel B.J. Jansen, M. Zhang and A. Chowdury. Twitter power: Tweets as electronic word
of mouth. Journal of the American Society for Information Science and Technology, 2009.
[5] R. Collobert, J. Weston, L. Bottou, M. Karlen, K. Kavukcuoglu, and P. Kuksa. Natural Language Processing (Almost) from Scratch. JMLR, 12:2493?2537, 2011.
8
[6] O. Tsur D. Davidov and A. Rappoport. Enhanced sentiment learning using twitter hashtags
and smileys. Association for Computational Linguistics, 2010.
[7] J. Duchi, E. Hazan, and Y. Singer. Adaptive subgradient methods for online learning and
stochastic optimization. JMLR, 12, July 2011.
[8] J. L. Elman. Distributed representations, simple recurrent networks, and grammatical structure.
Machine Learning, 7(2-3):195?225, 1991.
[9] C. Goller and A. K?uchler. Learning task-dependent distributed representations by backpropagation through structure. In Proceedings of the International Conference on Neural Networks,
1996.
[10] E. Grefenstette, G. Dinu, Y.-Z. Zhang, M. Sadrzadeh, and M. Baroni. Multi-step regression
learning for compositional distributional semantics. In IWCS, 2013.
[11] E. H. Huang, R. Socher, C. D. Manning, and A. Y. Ng. Improving Word Representations via
Global Context and Multiple Word Prototypes. In ACL, 2012.
[12] O. Irsoy and C. Cardie. Bidirectional recursive neural networks for token-level labeling with
structure. NIPS Deep Learning Workshop, 2013.
[13] T. K. Landauer and S. T. Dumais. A solution to Plato?s problem: the Latent Semantic Analysis theory of acquisition, induction and representation of knowledge. Psychological Review,
104(2):211?240, 1997.
[14] P. Le and W. Zuidema. The inside-outside recursive neural network model for dependency
parsing. EMNLP, 2014.
[15] T. Mikolov, W. Yih, and G. Zweig. Linguistic regularities in continuous spaceword representations. In HLT-NAACL, 2013.
[16] J. Mitchell and M. Lapata. Composition in distributional models of semantics. Cognitive
Science, 34(8):1388?1429, 2010.
[17] Z. Kozareva P. Nakov. Semeval-2013 task 2: Sentiment analysis in twitter. Proceedings of the
Seventh International Workshop on Semantic Evaluation (SemEval 2013), 2013.
[18] J. Pennington, R. Socher, and C. D. Manning. Glove: Global vectors for word representation.
EMNLP, 2014.
[19] J. B. Pollack. Recursive distributed representations. Artificial Intelligence, 46, November
1990.
[20] J.T. Rolfe and Y. LeCun. Discriminative recurrent sparse auto-encoders. arXiv:1301.3775v4,
2013.
[21] M. Schuster and K.K. Paliwal. Bidirectional recurrent neural networks. Signal Processing,
IEEE Transactions, 1997.
[22] R. Socher, J. Bauer, C. D. Manning, and A. Y. Ng. Parsing With Compositional Vector Grammars. In ACL, 2013.
[23] R. Socher, E. H. Huang, J. Pennington, A. Y. Ng, and C. D. Manning. Dynamic Pooling and
Unfolding Recursive Autoencoders for Paraphrase Detection. In NIPS, 2011.
[24] R. Socher, B. Huval, C. D. Manning, and A. Y. Ng. Semantic Compositionality Through
Recursive Matrix-Vector Spaces. In EMNLP, 2012.
[25] R. Socher, C. D. Manning, and A. Y. Ng. Learning continuous phrase representations and syntactic parsing with recursive neural networks. In Proceedings of the NIPS-2010 Deep Learning
and Unsupervised Feature Learning Workshop, 2010.
[26] R. Socher, A. Perelygin, J. Wu, J. Chuang, C. Manning, A. Ng, and C. Potts. Recursive deep
models for semantic compositionality over a sentiment treebank. In EMNLP, 2013.
[27] P. D. Turney and P. Pantel. From frequency to meaning: Vector space models of semantics.
Journal of Artificial Intelligence Research, 37:141?188, 2010.
[28] A. Yessenalina and C. Cardie. Compositional matrix-space models for sentiment analysis. In
EMNLP, 2011.
[29] F.M. Zanzotto, I. Korkontzelos, F. Fallucchi, and S. Manandhar. Estimating linear models for
compositional distributional semantics. In COLING, 2010.
9
| 5275 |@word bigram:4 nd:14 propagate:1 simplifying:1 yih:1 recursively:2 reduction:1 series:3 score:5 tuned:2 document:4 ours:3 outperforms:2 past:1 current:1 contextual:8 activation:3 parsing:3 stemming:1 concatenate:1 update:2 alone:1 intelligence:2 leaf:3 selected:1 slowing:1 smith:1 short:2 node:32 lexicon:4 monday:2 zhang:2 stopwords:1 rc:1 balasubramanyan:1 become:1 tomorrow:1 incorrect:1 qualitative:1 consists:1 compose:2 combine:4 inside:2 introduce:6 theoretically:1 indeed:1 kuksa:1 p1:10 elman:1 multi:2 window:5 xroot:1 becomes:3 begin:1 confused:1 provided:1 linearity:1 alto:1 medium:3 estimating:1 interpreted:1 developed:1 transformation:2 quantitative:1 every:4 fun:1 classifier:6 uk:1 unit:3 positive:7 local:3 io:6 rnns:4 twice:3 initialization:2 acl:2 co:1 contextually:1 bi:1 wrf:5 averaged:1 directed:1 acknowledgment:1 decided:1 lecun:1 recursive:36 practice:1 block:1 backpropagation:2 unfold:1 area:1 rnn:44 got:1 poster:1 word:68 pre:2 get:1 cannot:3 close:1 unlabeled:3 selection:1 context:26 applying:1 bifet:1 www:1 elongated:2 reviewer:1 go:1 educational:1 starting:2 independently:2 focused:1 disentangled:1 embedding:2 enhanced:1 today:1 parser:4 user:2 us:3 romain:2 element:1 walking:1 distributional:4 labeled:14 observed:1 bottom:2 capture:5 barbosa:1 luck:1 highest:1 valuable:1 mentioned:1 intuition:1 dynamic:1 trained:10 rewrite:1 purely:5 uchler:1 sentiwordnet:1 easily:1 po:1 represented:1 surrounding:2 train:5 describe:2 artificial:2 labeling:1 hyper:1 outside:1 heuristic:1 stanford:5 solve:1 forwarded:1 grammar:1 ability:3 syntactic:2 transform:1 jointly:1 noisy:1 final:6 online:1 sequence:1 advantage:2 propose:1 product:1 coming:1 frequent:1 ablation:2 vnd:1 flexibility:1 representational:1 pantel:1 validate:2 competition:9 constituent:1 zuidema:1 parent:3 cluster:1 regularity:1 extending:2 rolfe:1 produce:1 hashtags:1 help:4 recurrent:9 ac:1 propagating:3 job:2 eq:3 p2:5 c:1 predicted:2 come:1 differ:1 foot:1 correct:6 stochastic:1 opinion:1 public:1 require:1 assign:1 f1:7 generalization:1 really:1 anonymous:1 emoticon:2 transfor:1 weblogs:1 ic:1 great:1 mapping:2 predict:3 baroni:1 mayo:1 bag:2 label:4 palo:1 create:1 successfully:1 unfolding:4 hope:1 clearly:1 always:1 sunday:1 mation:1 rather:1 season:1 linguistic:2 encode:1 improvement:1 potts:1 contrast:1 baseline:6 sense:1 helpful:1 inference:3 twitter:12 dependent:3 streaming:1 sb:3 entire:1 hidden:3 interested:1 semantics:4 pixel:1 overall:2 classification:11 issue:1 denoted:1 development:4 jansen:1 art:3 special:2 platform:1 softmax:5 initialize:2 constrained:1 once:2 having:1 ng:6 sampling:1 yessenalina:1 unsupervised:13 thinking:3 others:1 richard:2 inherent:2 employ:2 randomly:2 composed:4 replaced:1 negation:4 attempt:1 detection:2 message:2 wlf:5 highly:1 evaluation:3 punctuation:1 undefined:1 behind:1 tonight:2 sobel:1 chain:1 n2d:2 closer:1 necessary:5 shorter:1 respective:1 tree:15 initialized:1 pollack:1 android:10 psychological:1 instance:4 classify:4 column:1 increased:1 dev:1 disadvantage:1 phrase:36 applicability:1 introducing:1 neutral:3 uniform:1 goller:1 seventh:1 motivating:1 dependency:2 encoders:1 my:2 combined:1 dumais:1 thanks:1 st:1 international:5 v4:1 together:2 aaai:1 huang:2 emnlp:5 worse:1 dead:1 external:1 american:1 cognitive:1 account:2 potential:1 de:2 lapata:1 huval:1 microblogging:1 summarized:1 includes:1 matter:1 depends:1 collobert:1 piece:1 performed:1 try:1 lot:1 root:7 unigrams:1 doing:1 hazan:1 red:2 start:2 bayes:2 option:1 parallel:1 annotation:1 accuracy:4 ensemble:2 directional:1 generalize:2 kavukcuoglu:1 accurately:1 cardie:4 rectified:1 classified:1 suffers:1 hlt:1 definition:2 against:1 acquisition:1 frequency:1 dataset:14 treatment:1 mitchell:1 knowledge:3 improves:1 dimensionality:1 back:3 rappoport:1 bidirectional:12 higher:2 supervised:7 day:4 furthermore:1 governing:1 just:3 autoencoders:2 overfit:1 hand:3 horizontal:1 parse:2 christopher:1 propagation:2 logistic:3 usage:1 naacl:1 requiring:1 hence:3 regularization:2 semantic:11 deal:1 during:5 backpropagating:1 outline:1 performs:5 duchi:1 interpreting:1 meaning:7 image:1 consideration:2 novel:1 recently:2 common:2 sigmoid:1 quarter:1 empirically:2 winner:2 irsoy:3 extend:1 interpretation:1 linking:1 association:1 theirs:1 composition:4 connor:1 hurting:1 ai:1 routledge:1 rd:2 pmi:1 language:6 moving:1 kozareva:1 longer:1 add:2 own:2 noisily:1 optimizing:1 apart:1 paliwal:1 binary:3 meeting:1 seen:1 additional:7 care:1 determine:2 signal:3 july:1 branch:1 full:3 multiple:4 reduces:1 karlen:1 ing:1 feedbackward:8 cross:4 zweig:1 prediction:8 regression:3 vision:1 arxiv:1 represent:1 achieved:2 whereas:2 want:4 fine:1 biased:1 operate:1 unlike:4 rest:1 sure:1 comment:1 subject:1 pooling:1 validating:1 plato:1 flow:1 tough:1 effectiveness:1 backwards:1 feedforward:11 intermediate:1 easy:1 split:1 semeval:12 variety:1 isolation:2 gave:1 architecture:6 idea:6 prototype:1 shift:1 gb:25 sentiment:39 iwcs:1 york:1 compositional:6 deep:6 generally:1 useful:1 detailed:1 amount:2 mid:1 simplest:1 http:1 outperform:1 correctly:4 blue:2 poll:1 prevent:1 neither:1 kept:2 backward:9 subgradient:1 tweet:9 sum:1 powerful:1 dst:1 almost:1 electronic:1 wu:1 disambiguation:1 resize:1 dropout:8 followed:1 fold:1 smiley:2 wc:2 aspect:2 carve:1 metamind:2 mikolov:1 eat:1 tsur:1 tv:1 combination:5 manning:9 describes:1 smaller:2 character:1 pumpkin:1 taken:1 equation:4 resource:2 previously:1 eventually:2 singer:1 end:1 away:3 occurrence:1 batch:3 chuang:1 top:3 remaining:1 include:3 nlp:2 linguistics:2 parsed:2 rdl:1 society:1 iphone:2 feng:1 tensor:1 move:1 already:1 dependence:2 unclear:1 separate:1 thank:1 entity:4 concatenation:2 me:4 induction:1 length:6 copying:1 relationship:1 illustration:1 mini:3 frank:1 perelygin:1 negative:5 unknown:1 allowing:1 upper:1 datasets:3 sm:4 november:1 beat:6 incorrectly:1 incorporated:1 rn:3 wlb:4 arbitrary:2 paraphrase:1 canada:1 compositionality:3 introduced:3 pair:2 sentence:14 learned:3 nip:3 address:1 usually:4 below:4 exemplified:1 challenge:3 including:1 belief:10 mall:1 power:2 everyone:1 mouth:1 natural:6 hybrid:7 beautiful:1 regularized:1 task2:1 predicting:1 recursion:1 representing:2 improve:2 altered:1 movie:1 technology:1 started:1 catch:1 autoencoder:1 utterance:1 naive:2 auto:1 text:4 epoch:1 understanding:2 review:2 discovery:2 multiplication:1 adagrad:1 loss:1 validation:1 propagates:1 treebank:1 story:1 classifying:1 token:3 last:2 bias:1 allow:2 understand:1 taking:2 face:1 differentiating:2 sparse:1 serra:1 distributed:3 grammatical:3 feedback:1 dimension:13 vocabulary:1 gram:2 bauer:1 computes:2 ignores:1 author:1 forward:7 commonly:1 made:1 adaptive:1 far:1 social:3 transaction:1 obtains:2 keep:3 global:12 happier:3 reveals:1 sat:1 discriminative:1 landauer:1 don:3 continuous:2 latent:2 table:7 disambiguate:1 scratch:1 learn:2 robust:1 ca:2 improving:1 tuesday:1 as:2 bottou:1 complex:2 domain:1 paulus:1 main:1 arrow:1 whole:1 motivation:1 big:1 scored:1 nrc:1 child:2 allowed:2 fig:7 thats:1 crafted:1 screen:2 fashion:2 batched:1 contextindependent:1 position:2 explicit:1 jmlr:2 coling:2 hannah:1 down:7 tokenizing:1 removing:1 bad:6 specific:9 showing:2 explored:2 list:1 svm:4 fusion:1 workshop:3 socher:8 adding:3 pennington:2 subtree:1 illustrates:1 antonym:1 push:1 backpropagate:2 entropy:1 simply:3 explore:1 likely:2 watch:1 applies:1 determines:1 abc:1 extracted:1 weston:1 grefenstette:1 goal:1 marked:1 towards:1 considerable:1 change:3 included:3 determined:1 glove:1 total:1 pas:1 ya:1 turney:1 phenomenon:1 incorporate:1 schuster:1 |
4,722 | 5,276 | Deep Networks with Internal Selective Attention
through Feedback Connections
Marijn F. Stollenga?, Jonathan Masci? , Faustino Gomez, Juergen Schmidhuber
IDSIA, USI-SUPSI
Manno-Lugano, Switzerland
{marijn,jonathan,tino,juergen}@idsia.ch
Abstract
Traditional convolutional neural networks (CNN) are stationary and feedforward.
They neither change their parameters during evaluation nor use feedback from
higher to lower layers. Real brains, however, do. So does our Deep Attention
Selective Network (dasNet) architecture. DasNet?s feedback structure can dynamically alter its convolutional filter sensitivities during classification. It harnesses the power of sequential processing to improve classification performance,
by allowing the network to iteratively focus its internal attention on some of its
convolutional filters. Feedback is trained through direct policy search in a huge
million-dimensional parameter space, through scalable natural evolution strategies (SNES). On the CIFAR-10 and CIFAR-100 datasets, dasNet outperforms the
previous state-of-the-art model on unaugmented datasets.
1
Introduction
Deep convolutional neural networks (CNNs) [1] with max-pooling layers [2] trained by backprop
[3] on GPUs [4] have become the state-of-the-art in object recognition [5, 6, 7, 8], segmentation/detection [9, 10], and scene parsing [11, 12] (for an extensive review see [13]). These architectures consist of many stacked feedforward layers, mimicking the bottom-up path of the human
visual cortex, where each layer learns progressively more abstract representations of the input data.
Low-level stages tend to learn biologically plausible feature detectors, such as Gabor filters [14].
Detectors in higher layers learn to respond to concrete visual objects or their parts, e.g., [15]. Once
trained, the CNN never changes its weights or filters during evaluation.
Evolution has discovered efficient feedforward pathways for recognizing certain objects in the blink
of an eye. However, an expert ornithologist, asked to classify a bird belonging to one of two very
similar species, may have to think for more than a few milliseconds before answering [16, 17],
implying that several feedforward evaluations are performed, where each evaluation tries to elicit
different information from the image. Since humans benefit greatly from this strategy, we hypothesise CNNs can too. This requires: (1) the formulation of a non-stationary CNN that can adapt its
own behaviour post-training, and (2) a process that decides how to adapt the CNNs behaviour.
This paper introduces Deep Attention Selective Networks (dasNet) which model selective attention
in deep CNNs by allowing each layer to influence all other layers on successive passes over an image
through special connections (both bottom-up and top-down), that modulate the activity of the convolutional filters. The weights of these special connections implement a control policy that is learned
through reinforcement learning after the CNN has been trained in the usual way via supervised
learning. Given an input image, the attentional policy can enhance or suppress features over multiple passes to improve the classification of difficult cases not captured by the initially supervised
?
Shared first author.
1
training. Our aim is to let the system check the usefulness of internal CNN filters automatically,
omitting manual inspection [18].
In our current implementation, the attentional policy is evolved using Separable Natural Evolution
Strategies (SNES; [19]), instead of a conventional, single agent reinforcement learning method
(e.g. value iteration, temporal difference, policy gradients, etc.) due to the large number of parameters (over 1 million) required to control CNNs of the size typically used in image classification.
Experiments on CIFAR-10 and CIFAR100 [20] show that on difficult classification instances, the
network corrects itself by emphasising and de-emphasising certain filters, outperforming a previous
state-of-the-art CNN.
2
Maxout Networks
In this work we use the Maxout networks [7], combined with dropout [21], as the underlying model
for dasNet. Maxout networks represent the state-of-the-art for object recognition in various tasks and
have only been outperformed (by a small margin) by averaging committees of several convolutional
neural networks. A similar approach, which does not reduce dimensionality in favor of sparsity
in the representation has also been recently presented [22]. Maxout CNNs consist of a stack of
alternating convolutional and maxout layers, with a final classification layer on top:
Convolutional Layer. The input to this layer can be an image or the output of a previous layer,
consisting of c input maps of width m and height n: x ? Rc?m?n . The output consists of a set of c?
?
?
?
output maps: y ? Rc ?m ?n . The convolutional layer is parameterised by c ? c? filters of size k ? k.
?
We denote the filters by Fi,j ? Rk?k , where i and j are indexes of the input and output maps and ?
denotes the layer.
yj? =
i=c
?
?
?(xi ? Fi,j
)
(1)
i=0
where i and j index the input and output map respectively, ? is the convolutional operator, ? is an
element-wise nonlinear function, and ? is used to index the layer. The size of the output is determined
by the kernel size and the stride used for the convolution (see [7]).
Pooling Layer. A pooling layer is used to reduced the dimensionality of the output from a convolutional layer. The usual approach is to take the maximum value among non- or partially-overlapping
patches in every map, therefore reducing dimensionality along the height and width [2]. Instead, a
Maxout pooling layer reduces every b consecutive maps to one map, by keeping only the maximum
value for every pixel-position, where b is called the block size. Thus the map reduces c input maps
to c? = c/b output maps.
b
??1
?
yj,x,y
= max yj?b+i,x,y
i=0
?
?
(2)
?
where y ? ? Rc ?m ?n , and ? again is used to index the layer. The output of the pooling layer can
either be used as input to another pair of convolutional- and pooling layers, or form input to a final
classification layer.
Classification Layer. Finally, a classification step is performed. First the output of the last pooling
layer is flattened into one large vector ?x, to form the input to the following equations:
?
y?j? = max Fj?b+i
?x
(3)
v = ?(F ?+1 y?? )
(4)
i=0..b
where F ? ? RN ?|?x| (N is chosen), and ?(?) is the softmax activation function which produces the
class probabilities v. The input is projected by F and then reduced using a maxout, similar to the
pooling layer (3).
2
3
Reinforcement Learning
Reinforcement learning (RL) is a general framework for learning to make sequential decisions order
to maximise an external reward signal [23, 24]. The learning agent can be anything that has the
ability to act and perceive in a given environment.
At time t, the agent receives an observation ot ? O of the current state of the environment st ? S,
and selects an action, at ? A, chosen by a policy ? : O ? A, where S, O and A the spaces
of all possible states, observations, and action, respectively.1 The agent then enters state st+1 and
receives a reward rt ??
R. The objective is to find the policy, ?, that maximises the expected future
discounted reward, E[ t ? t rt ], where ? ? [0, 1] discounts the future, modeling the ?farsightedness?
of the agent.
In dasNet, both the observation and action spaces are real valued O = Rdim(O) , A = Rdim(A) .
Therefore, policy ?? must be represented by a function approximator, e.g. a neural network, parameterised by ?. Because the policies used to control the attention of the dasNet have state and
actions spaces of close to a thousand dimensions, the policy parameter vector, ?, will contain close
to a million weights, which is impractical for standard RL methods. Therefore, we instead evolve
the policy using a variant for Natural Evolution Strategies (NES; [25, 26]), called Separable NES
(SNES; [19]). The NES family of black-box optimization algorithms use parameterised probability
distributions over the search space, instead of an explicit population (i.e., a conventional ES [27]).
Typically, the distribution is a multivariate Gaussian parameterised by mean ? and covariance matrix
?. Each epoch a generation is sampled from the distribution, which is then updated the direction
of the natural gradient of the expected fitness of the distribution. SNES differs from standard NES
in that instead of maintaining the full covariance matrix of the search distribution, uses only the
diagonal entries. SNES is theoretically less powerful than standard NES, but is substantially more
efficient.
4
Deep Attention Selective Networks (dasNet)
The idea behind dasNet is to harness the power of sequential processing to improve classification
performance by allowing the network to iteratively focus the attention of its filters. First, the standard
Maxout net (see Section 2) is augmented to allow the filters to be weighted differently on different
passes over the same image (compare to equation 1):
yj?
=
a?j
i=c
?
?
?(xi ? Fi,j
),
(5)
i=0
where a?j is the weight of the j-th output map in layer ?, changing the strength of its activation,
before applying the maxout pooling operator. The vector a = [a00 , a01 , ? ? ? , a0c? , a10 , ? ? ? , a1c? , ? ? ? ]
represents the action that the learned policy must select in order to sequentially focus the attention
of the Maxout net on the most discriminative features in the image being processed. Changing action
a will alter the behaviour of the CNN, resulting in different outputs, even when the image x does
not change. We indicate this with the following notation:
vt = Mt (?, x)
(6)
where ? is the parameter vector of the policy, ?? , and vt is the output of the network on pass t.
Algorithm 1 describes the dasNet training algorithm. Given a Maxout net, M, that has already
been trained to classify images using training set, X, the policy, ?, is evolved using SNES to focus
the attention of M. Each pass through the while loop represents one generation of SNES. Each
generation starts by selecting a subset of n images from X at random.
Then each of the p samples drawn from the SNES search distribution (with mean ? and covariance
?) representing the parameters, ?i , of a candidate policy, ??i , undergoes n trials, one for each image
in the batch. During a trial, the image is presented to the Maxout net T times. In the first pass, t = 0,
the action, a0 , is set to ai = 1, ?i, so that the Maxout network functions as it would normally ?
1
In this work ? : O ? A is a deterministic policy; given an observation it will always output the same
action. However, ? could be extended to stochastic policies.
3
Algorithm 1 T RAIN DAS N ET (M, ?, ?, p, n)
1: while True do
2:
images ? N EXT BATCH(n)
3:
for i = 0 ? p do
4:
?i ? N(?, ?)
5:
for j = 0 ? n do
6:
a0 ? 1 {Initialise gates a with identity activation}
7:
for t = 0 ? T do
8:
vt = Mt (?i , xi )
9:
ot ? h(Mt )
10:
at+1 ? ??i (ot )
11:
end for
12:
Li = ??boost d log(vT )
13:
end for
14:
F[i] ? f (?i )
15:
?[i] ? ?i
16:
end for
17:
U PDATE SNES(F, ?) {Details in supplementary material.}
18: end while
the action has no effect. Once the image is propagated through the net, an observation vector, o0 , is
constructed by concatenating the following values extracted from M, by h(?):
1. the average activation of every output map Avg(yj ) (Equation 2), of each Maxout layer.
2. the intermediate activations y?j of the classification layer.
3. the class probability vector, vt .
While averaging map activations provides only partial state information, these values should still be
meaningful enough to allow for the selection of good actions. The candidate policy then maps the
observation to an action:
??i (o) = dim(A)?(? i ot ) = at ,
(7)
where ? ? Rdim(A)?dim(O) is the weight matrix of the neural network, and ? is the softmax. Note
that the softmax function is scaled by the dimensionality of the action space so that elements in the
action vector average to 1 (instead of regular softmax which sums to 1), ensuring that all network
outputs are positive, thereby keeping the filter activations stable.
On the next pass, the same image is processed again, but this time using the filter weighting, a1 .
This cycle is repeated until pass T (see figure 1 for a illustration of the process), at which time the
performance of the network is scored by:
Li = ??boost d log(vT )
vT = MT (?i , xi )
?boost =
{
?correct
?misclassified
if d = ?vT ??
otherwise,
(8)
(9)
(10)
where v is the output of M at the end of the pass T , d is the correct classification, and ?correct
and ?misclassif ied are constants. Li measures the weighted loss, where misclassified samples are
weighted higher than correctly classified samples ?misclassif ied > ?correct . This simple form of
boosting is used to focus on the ?difficult? misclassified images. Once all of the input images have
been processed, the policy is assigned the fitness:
cumulative score
f (?i ) =
z }| {
n
?
Li
i=1
4
regularization
z }| {
+ ?L2 ??i ?2
(11)
Classes
Classes
Softmax
Softmax
ma
era p
ge
s
error
Softmax
ma
era p
ge
av
av
s
Maps
Maps
Maps
gates
Filters
gates
Filters
ma
era p
ge
s
av
Filters
m
av ap
era
ge
s
policy
Maps
Maps
gates
Filters
Maps
gates
Filters
ag
B
RG
Observation Action
t=1
t=2
Im
Filters
e
ag
B
RG
Action
e
Im
t=T
Figure 1: The dasNet Network. Each image in classified after T passes through the network. After each forward propagation through the Maxout net, the output classification vector, the output of the second to last layer,
and the averages of all feature maps, are combined into an observation vector that is used by a deterministic
policy to choose an action that changes the weights of all the feature maps for the next pass of the same image.
After pass T , the output of the Maxout net is finally used to classify the image.
where ?L2 is a regularization parameter. Once all of the candidate policies have been evaluated,
SNES updates its distribution parameters (?, ?) according the natural gradient calculated from the
sampled fitness values, F. As SNES repeatedly updates the distribution over the course of many
generations, the expected fitness of the distribution improves, until the stopping criterion is met
when no improvement is made for several consecutive epochs.
5
Related Work
Human vision is still the most advanced and flexible perceptual system known. Architecturally,
visual cortex areas are highly connected, including direct connections over multiple levels and topdown connections. Felleman and Essen [28] constructed a (now famous) hierarchy diagram of
32 different visual cortical areas in macaque visual cortex. About 40% of all pairs of areas were
considered connected, and most connected areas were connected bidirectionally. The top-down
connections are more numerous than bottom-up connections, and generally more diffuse [29]. They
are thought to play primarily a modulatory role, while feedforward connections serve as directed
information carriers [30].
Analysis of response latencies to a newly-presented image lends credence to the theory that there are
two stages of visual processing: a fast, pre-attentive phase, due to feedforward processing, followed
by an attentional phase, due to the influence of recurrent processing [31]. After the feedforward pass,
we can recognise and localise simple salient stimuli, which can ?pop-out? [32], and response times
do not increase regardless of the number of distractors. However, this effect has only been conclusively shown for basic features such as colour or orientation; for categorical stimuli or faces, whether
there is a pop-out effect remains controversial [33]. Regarding the attentional phase, feedback connections are known to play important roles, such as in feature grouping [34], in differentiating a
foreground from its background, (especially when the foreground is not highly salient [35]), and
perceptual filling in [36]. Work by Bar et al. [37] supports the idea that top-down projections from
prefrontal cortex play an important role in object recognition by quickly extracting low-level spatial
frequency information to provide an initial guess about potential categories, forming a top-down
expectation that biases recognition. Recurrent connections seem to rely heavily on competitive inhibition and other feedback to make object recognition more robust [38, 39].
In the context of computer vision, RL has been shown to be able to learn saccades in visual scenes
to learn selective attention [40, 41], learn feedback to lower levels [42, 43], and improve face recognition [44, 45]. It has been shown to be effective for object recognition [46], and has also been
5
0.452
Method
Dropconnect [51]
Stochastic Pooling [52]
Multi-column CNN [5]
Maxout [7]
Maxout (trained by us)
dasNet
NiN [8]
NiN (augmented)
CIFAR-10
CIFAR-100
9.32%
15.13%
11.21%
9.38%
9.61%
9.22%
10.41%
8.81%
38.57%
34.54%
33.78%
35.68%
-
0.45
% Correct
Table 1: Classification results on CIFAR-10 and CIFAR-100
datasets. The error on the test-set is shown for several methods. Note that the result for Dropconnect is the average of
12 models. Our method improves over the state-of-the-art
reference implementation to which feedback connections are
added. The recent Network in Network architecture [8] has
better results when data-augmentation is applied.
0.448
0 steps
1 step
2 steps
3 steps
0.446
0.444
0.442
0.44
0
1
2 3 4 5 6 7 8 9
Number of steps evaluated
Figure 2: Two dasNets were trained on
CIFAR-100 for different values of T . Then
they were allowed to run for [0..9] iterations for each image. The performance
peeks at the number of steps that the network is trained on, after which the performance drops, but does not explode, showing
the dynamics are stable.
combined with traditional computer vision primitives [47]. Iterative processing of images using
recurrency has been successfully used for image reconstruction [48], face-localization [49] and
compression [50]. All these approaches show that recurrency in processing and an RL perspective can lead to novel algorithms that improve performance. However, this research is often applied to simplified datasets for demonstration purposes due to computation constraints, and are not
aimed at improving the state-of-the-art. In contrast, we apply this perspective directly to the known
state-of-the-art neural networks to show that this approach is now feasible and actually increases
performance.
6
Experiments on CIFAR-10/100
The experimental evaluation of dasNet focuses on ambiguous classification cases in the CIFAR-10
and CIFAR-100 data sets where, due to a high number of common features, two classes are often
mistaken for each other. These are the most interesting cases for our approach. By learning on top of
an already trained model, dasNet must aim at fixing these erroneous predictions without disrupting,
or forgetting, what has been learned. The CIFAR-10 dataset [20] is composed of 32 ? 32 colour
images split into 5?104 training and 104 testing samples, where each image is assigned to one of 10
classes. The CIFAR-100 is similarly composed, but contains 100 classes. The number of steps was
experimentally determined and fixed at T = 5; small enough to be computationally tractable while
still allowing for enough interaction. In all experiments we set ?correct = 0.005, ?misclassified = 1 and
?L2 = 0.005.
The Maxout network, M, was trained with data augmentation following global contrast normalization and ZCA normalization. The model consists of three convolutional maxout layers followed by
a fully connected maxout and softmax outputs. Dropout of 0.5 was used in all layers except the
input layer, and 0.2 for the input layer. The population size for SNES was set to 50. Training took
of dasNet took around 4 days on a GTX 560 Ti GPU, excluding the original time used to train M.
Table 1 shows the performance of dasNet vs. other methods, where it achieves a relative improvement of 6% with respect to the vanilla CNN. This establishes a new state-of-the-art result for this
challenging dataset, for unaugmented data. Figure 3 shows the classification of a cat-image from the
test-set. All output map activations in the final step are shown at the top. The difference in activations compared to the first step, i.e., the (de-)emphasis of each map, is shown on the bottom. On the
left are the class probabilities for each time-step. At the first step, the classification is ?dog?, and the
cat could indeed be mistaken for a puppy. Note that in the first step, the network has not yet received
any feedback. In the next step, the probability for ?cat? goes up dramatically, and subsequently drops
a bit in the following steps. The network has successfully disambiguated a cat from a dog. If we
investigate the filters, we see that in the lower layer emphasis changes significantly (see ?change
of layer 0?). Some filters focus more on surroundings whilst others de-emphasise the eyes. In the
6
layer 0
layer 1
layer 2
change of layer 0
change of layer 1
change of layer 2
airplane
automobile
bird
cat
deer
dog
frog
horse
ship
truck
1 2 3 4 5
Timesteps
class probabilities
Figure 3: The classification of a cat by the dasNet is shown. All output map activations in the final step
are shown on the top. Their changes relative to initial activations in the first step are shown at the bottom
(white = emphasis, black = suppression). The changes are normalised to show the effects more clearly. The
class probabilities over time are shown on the left. The network first classifies the image as a dog (wrong)
but corrects itself by emphasising its convolutional filters to see it is actually a cat. Two more examples are
included in the supplementary material.
second layer, almost all output maps are emphasised. In the third and highest convolutional layer,
the most complex changes to the network can be seen. At this level the positional correspondence
is largely lost, and the filters are known to code for ?higher level? features. It is in this layer that
changes are the most influential because they are closest to the final output layers.
It is hard to qualitatively analyze the effect of the alterations. If we compare each final activation in
layer 2 to its corresponding change (see Figure 3, right), we see that the activations are not simply
uniformly enhanced. Instead, complex suppression and enhancement patterns are found, increasing
and decreasing activation of specific pixels. Visualising what these high-level actually do is an open
problem in deep learning.
Dynamics To investigate the dynamics, a small 2-layer dasNet network was trained for different
values of T . Then they were evaluated by allowing them to run for [0..9] steps. Figure 2 shows
results of training dasNet on CIFAR-100 for T = 1 and T = 2. The performance goes up from the
vanilla CNN, peaks at the step = T as expected, and reduces but stays stable after that. So even
though the dasNet was trained using only a small number of steps, the dynamics stay stable when
these are evaluated for as many as 10 steps.
To verify whether the dasNet policy is actually making good use of its gates, we estimate their
information content in the following way: The gate values in the final step are used directly for
classification. The hypothesis is that if the gates are used properly, then their activation should
contain information that is relevant for classification. For this purpose, a dasNet that was trained with
T = 2. Then using only the final gate-values (so without e.g. the output of the classification layer),
a classification using 15-nearest neighbour and logistic regression was performed. This resulted in
a performance of 40.70% and 45.74% correct respectively, similar to the performance of dasNet,
confirming that they contain significant information.
7
Conclusion
DasNet is a deep neural network with feedback connections that are learned by through reinforcement learning to direct selective internal attention to certain features extracted from images. After
a rapid first shot image classification through a standard stack of feedforward filters, the feedback
can actively alter the importance of certain filters ?in hindsight?, correcting the initial guess via
additional internal ?thoughts?.
7
DasNet successfully learned to correct image misclassifications produced by a fully trained feedforward Maxout network. Its active, selective, internal spotlight of attention enabled state-of-the-art
results. Future research will also consider more complex actions that spatially focus on (or alter)
parts of observed images.
Acknowledgments
We acknowledge Matthew Luciw for his discussions and for providing a short literature review,
included in the Related Work section.
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[23]
K. Fukushima. ?Neural network model for a mechanism of pattern recognition unaffected by
shift in position - Neocognitron?. In: Trans. IECE J62-A(10) (1979), pp. 658?665.
J. Weng, N. Ahuja, and T. S. Huang. ?Cresceptron: a self-organizing neural network which
grows adaptively?. In: IJCNN. Vol. 1. IEEE. 1992, pp. 576?581.
Y. LeCun et al. ?Back-Propagation Applied to Handwritten Zip Code Recognition?. In: Neural Computation 1.4 (1989), pp. 541?551.
D. C. Ciresan et al. ?Flexible, High Performance Convolutional Neural Networks for Image
Classification?. In: IJCAI. 2011, pp. 1237?1242.
D. C. Ciresan, U. Meier, and J. Schmidhuber. ?Multi-Column Deep Neural Networks for
Image Classification?. In: CVPR. Long preprint arXiv:1202.2745v1 [cs.CV]. 2012.
A. Krizhevsky, I Sutskever, and G. E Hinton. ?ImageNet Classification with Deep Convolutional Neural Networks?. In: NIPS. 2012, p. 4.
I. J. Goodfellow et al. ?Maxout networks?. In: ICML. 2013.
Min Lin, Qiang Chen, and Shuicheng Yan. ?Network In Network?. In: CoRR abs/1312.4400
(2013).
D. C. Ciresan et al. ?Mitosis Detection in Breast Cancer Histology Images with Deep Neural
Networks?. In: MICCAI. Vol. 2. 2013, pp. 411?418.
D. C. Ciresan et al. ?Deep Neural Networks Segment Neuronal Membranes in Electron Microscopy Images?. In: NIPS. 2012, pp. 2852?2860.
C. Farabet et al. ?Learning hierarchical features for scene labeling?. In: PAMI 35.8 (2013),
pp. 1915?1929.
P. Sermanet et al. ?Pedestrian Detection with Unsupervised Multi-Stage Feature Learning?.
In: CVPR. IEEE, 2013.
J. Schmidhuber. Deep Learning in Neural Networks: An Overview. Tech. rep. IDSIA-03-14 /
arXiv:1404.7828v1 [cs.NE]. The Swiss AI Lab IDSIA, 2014.
D. Gabor. ?Theory of communication. Part 1: The analysis of information?. In: Electrical
Engineers-Part III: Journal of the Institution of Radio and Communication Engineering 93.26
(1946), pp. 429?441.
Matthew D Zeiler and Rob Fergus. ?Visualizing and understanding convolutional networks?.
In: Computer Vision?ECCV. Springer, 2014, pp. 818?833.
S. Branson et al. ?Visual recognition with humans in the loop?. In: Computer Vision?ECCV
2010. Springer, 2010, pp. 438?451.
P. Welinder et al. Caltech-UCSD Birds 200. Tech. rep. CNS-TR-2010-001. California Institute
of Technology, 2010.
M. D. Zeiler and R. Fergus. Visualizing and Understanding Convolutional Networks. Tech.
rep. arXiv:1311.2901 [cs.CV]. NYU, 2013.
T. Schaul, T. Glasmachers, and J. Schmidhuber. ?High dimensions and heavy tails for natural
evolution strategies?. In: GECCO. ACM. 2011, pp. 845?852.
A. Krizhevsky. ?Learning multiple layers of features from tiny images?. MA thesis. Computer
Science Department, University of Toronto, 2009.
G. E. Hinton et al. ?Improving neural networks by preventing co-adaptation of feature detectors?. In: arXiv preprint arXiv:1207.0580 (2012).
R. K. Srivastava et al. ?Compete to Compute?. In: NIPS. 2013.
L. P. Kaelbling, M. L. Littman, and A. W. Moore. ?Reinforcement learning: a survey?. In:
Journal of Artificial Intelligence Research 4 (1996), pp. 237?285.
8
[24] R. S. Sutton and A. G. Barto. Reinforcement Learning I: Introduction. 1998.
[25] D. Wierstra et al. ?Natural evolution strategies?. In: IEEE Congress on Evolutionary Computation. IEEE. 2008, pp. 3381?3387.
[26] T. Glasmachers et al. ?Exponential natural evolution strategies?. In: GECCO. ACM. 2010,
pp. 393?400.
[27] I. Rechenberg. Evolutionsstrategie - Optimierung technischer Systeme nach Prinzipien der
biologischen Evolution. Dissertation. Published 1973 by Fromman-Holzboog. 1971.
[28] D. J. Felleman and D. C. van Essen. ?Distributed hierarchical processing in the primate cerebral cortex?. In: Cerebral cortex 1.1 (1991), pp. 1?47.
[29] R. J. Douglas et al. ?Recurrent excitation in neocortical circuits?. In: Science 269.5226 (1995),
pp. 981?985.
[30] J. Bullier. ?Hierarchies of Cortical Areas?. In: The Primate Visual System. Ed. by J.H. Kaas
and C.E. Collins. New York: CRC Press, 2004, pp. 181?204.
[31] V. A. F. Lamme and P. R. Roelfsema. ?The distinct modes of vision offered by feedforward
and recurrent processing?. In: Trends in neurosciences 23.11 (2000), pp. 571?579.
[32] L. Itti. ?Visual salience?. In: 2.9 (2007), p. 3327.
[33] R. VanRullen. ?On second glance: Still no high-level pop-out effect for faces?. In: Vision
research 46.18 (2006), pp. 3017?3027.
[34] C. D. Gilbert and M. Sigman. ?Brain states: top-down influences in sensory processing?. In:
Neuron 54.5 (2007), pp. 677?696.
[35] J. M. Hupe et al. ?Cortical feedback improves discrimination between figure and background
by V1, V2 and V3 neurons?. In: Nature 394.6695 (1998), pp. 784?787.
[36] V. A. F. Lamme. ?Blindsight: the role of feedforward and feedback corticocortical connections?. In: Acta psychologica 107.1 (2001), pp. 209?228.
[37] M Bar et al. ?Top-down facilitation of visual recognition?. In: Proceedings of the National
Academy of Sciences of the United States of America 103.2 (2006), pp. 449?454.
[38] D. Wyatte et al. ?The Role of Competitive Inhibition and Top-Down Feedback in Binding
during Object Recognition?. In: Frontiers in Psychology 3 (2012), p. 182.
[39] D. Wyatte, T. Curran, and R. O?Reilly. ?The limits of feedforward vision: Recurrent processing promotes robust object recognition when objects are degraded?. In: Journal of Cognitive
Neuroscience 24.11 (2012), pp. 2248?2261.
[40] J. Schmidhuber and R. Huber. ?Learning to Generate Artificial Fovea Trajectories for Target
Detection?. In: International Journal of Neural Systems 2.1 & 2 (1991), pp. 135?141.
[41] Misha Denil et al. ?Learning where to attend with deep architectures for image tracking?. In:
Neural computation 24.8 (2012), pp. 2151?2184.
[42] R. C. O?Reilly. ?Biologically plausible error-driven learning using local activation differences: The generalized recirculation algorithm?. In: Neural Computation 8.5 (1996), pp. 895?
938.
[43] K. Fukushima. ?Restoring partly occluded patterns: a neural network model with backward
paths?. In: ICANN/ICONIP. Springer, 2003, pp. 393?400.
[44] H. Larochelle and G. Hinton. ?Learning to combine foveal glimpses with a third-order Boltzmann machine?. In: Image 1 (2010), p. x2.
[45] M. F. Stollenga, M. A. Wiering, and L. R. B. Schomaker. ?Using Guided Autoencoders on
Face Recognition?. In: Master?s thesis. University of Groningen. 2011.
[46] C OReilly R et al. ?Recurrent processing during object recognition?. In: Frontiers in Psychology 4 (2013), p. 124.
[47] S. D. Whitehead. ?Reinforcement Learning for the adaptive control of perception and action?.
PhD thesis. University of Rochester, 1992.
[48] S. Behnke. ?Learning iterative image reconstruction in the Neural Abstraction Pyramid?. In:
IJCIA 1.04 (2001), pp. 427?438.
[49] S. Behnke. ?Face localization and tracking in the neural abstraction pyramid?. In: Neural
Computing & Applications 14.2 (2005), pp. 97?103.
[50] B. Knoll and N. de Freitas. ?A Machine Learning Perspective on Predictive Coding with
PAQ8?. In: Data Compression Conference. 2012, pp. 377?386.
[51] L. Wan et al. ?Regularization of neural networks using dropconnect?. In: ICML. 2013,
pp. 1058?1066.
[52] Matthew D Zeiler and Rob Fergus. ?Stochastic pooling for regularization of deep convolutional neural networks?. In: arXiv preprint arXiv:1301.3557 (2013).
9
| 5276 |@word trial:2 cnn:10 compression:2 open:1 shuicheng:1 schomaker:1 covariance:3 systeme:1 thereby:1 tr:1 shot:1 initial:3 contains:1 score:1 selecting:1 united:1 foveal:1 outperforms:1 freitas:1 current:2 activation:16 yet:1 must:3 parsing:1 gpu:1 confirming:1 localise:1 drop:2 progressively:1 update:2 v:1 stationary:2 implying:1 credence:1 guess:2 intelligence:1 discrimination:1 inspection:1 short:1 dissertation:1 institution:1 provides:1 boosting:1 toronto:1 successive:1 height:2 rc:3 along:1 constructed:2 direct:3 become:1 wierstra:1 consists:2 pathway:1 combine:1 theoretically:1 huber:1 indeed:1 forgetting:1 expected:4 rapid:1 nor:1 multi:3 brain:2 discounted:1 decreasing:1 automatically:1 increasing:1 classifies:1 underlying:1 notation:1 circuit:1 what:2 evolved:2 substantially:1 whilst:1 ag:2 hindsight:1 impractical:1 temporal:1 every:4 j62:1 act:1 ti:1 scaled:1 wrong:1 control:4 normally:1 before:2 maximise:1 positive:1 carrier:1 engineering:1 congress:1 limit:1 local:1 attend:1 ext:1 era:4 sutton:1 path:2 ap:1 pami:1 black:2 bird:3 emphasis:3 frog:1 acta:1 dynamically:1 challenging:1 co:1 branson:1 directed:1 acknowledgment:1 lecun:1 yj:5 testing:1 lost:1 block:1 implement:1 differs:1 swiss:1 restoring:1 area:5 elicit:1 yan:1 gabor:2 thought:2 projection:1 significantly:1 pre:1 reilly:2 regular:1 close:2 selection:1 operator:2 context:1 influence:3 applying:1 gilbert:1 conventional:2 map:26 deterministic:2 primitive:1 attention:13 regardless:1 go:2 survey:1 correcting:1 perceive:1 usi:1 marijn:2 facilitation:1 his:1 initialise:1 enabled:1 population:2 cifar100:1 updated:1 hierarchy:2 play:3 heavily:1 enhanced:1 target:1 us:1 curran:1 hypothesis:1 goodfellow:1 element:2 idsia:4 recognition:15 trend:1 corticocortical:1 bottom:5 role:5 observed:1 preprint:3 enters:1 visualising:1 electrical:1 thousand:1 wiering:1 cycle:1 connected:5 highest:1 environment:2 architecturally:1 reward:3 asked:1 littman:1 occluded:1 dynamic:4 trained:14 segment:1 predictive:1 serve:1 localization:2 manno:1 differently:1 various:1 represented:1 pdate:1 cat:7 america:1 stacked:1 train:1 distinct:1 fast:1 effective:1 artificial:2 horse:1 deer:1 labeling:1 supplementary:2 plausible:2 valued:1 cvpr:2 otherwise:1 favor:1 ability:1 think:1 itself:2 final:8 net:7 took:2 reconstruction:2 interaction:1 adaptation:1 relevant:1 loop:2 rdim:3 organizing:1 academy:1 schaul:1 sutskever:1 ijcai:1 enhancement:1 nin:2 produce:1 object:11 recurrent:6 fixing:1 nearest:1 received:1 c:3 indicate:1 larochelle:1 met:1 switzerland:1 direction:1 puppy:1 guided:1 correct:8 filter:25 cnns:6 stochastic:3 subsequently:1 human:4 material:2 backprop:1 glasmachers:2 crc:1 behaviour:3 emphasising:3 ied:2 im:2 frontier:2 around:1 considered:1 snes:12 electron:1 matthew:3 achieves:1 consecutive:2 purpose:2 faustino:1 outperformed:1 radio:1 successfully:3 establishes:1 weighted:3 clearly:1 gaussian:1 always:1 aim:2 denil:1 barto:1 focus:8 groningen:1 improvement:2 properly:1 check:1 greatly:1 contrast:2 tech:3 zca:1 suppression:2 a01:1 dim:2 abstraction:2 stopping:1 typically:2 a0:2 initially:1 selective:8 misclassified:4 selects:1 mimicking:1 pixel:2 classification:26 among:1 flexible:2 orientation:1 art:9 special:2 softmax:8 spatial:1 once:4 never:1 qiang:1 represents:2 icml:2 filling:1 unsupervised:1 alter:4 future:3 foreground:2 others:1 stimulus:2 few:1 primarily:1 surroundings:1 neighbour:1 composed:2 national:1 resulted:1 fitness:4 phase:3 consisting:1 cns:1 fukushima:2 ab:1 detection:4 huge:1 highly:2 essen:2 investigate:2 evaluation:5 introduces:1 weng:1 misha:1 behind:1 stollenga:2 partial:1 glimpse:1 recirculation:1 instance:1 classify:3 modeling:1 column:2 hypothesise:1 juergen:2 kaelbling:1 entry:1 subset:1 usefulness:1 recognizing:1 krizhevsky:2 welinder:1 supsi:1 too:1 combined:3 adaptively:1 st:2 peak:1 sensitivity:1 bullier:1 international:1 stay:2 corrects:2 enhance:1 quickly:1 concrete:1 again:2 augmentation:2 thesis:3 wan:1 choose:1 prefrontal:1 huang:1 dropconnect:3 external:1 cognitive:1 expert:1 itti:1 li:4 actively:1 potential:1 de:4 stride:1 alteration:1 coding:1 pedestrian:1 performed:3 try:1 lab:1 kaas:1 analyze:1 start:1 competitive:2 rochester:1 degraded:1 convolutional:20 largely:1 blink:1 mitosis:1 famous:1 handwritten:1 produced:1 trajectory:1 unaffected:1 published:1 classified:2 detector:3 manual:1 farabet:1 ed:1 a10:1 attentive:1 frequency:1 pp:32 propagated:1 sampled:2 newly:1 dataset:2 distractors:1 dimensionality:4 improves:3 segmentation:1 oreilly:1 actually:4 back:1 higher:4 supervised:2 day:1 harness:2 response:2 formulation:1 evaluated:4 box:1 though:1 parameterised:4 stage:3 miccai:1 until:2 autoencoders:1 receives:2 nonlinear:1 overlapping:1 propagation:2 glance:1 undergoes:1 logistic:1 mode:1 grows:1 omitting:1 effect:6 verify:1 contain:3 true:1 gtx:1 evolution:8 regularization:4 assigned:2 alternating:1 spatially:1 iteratively:2 moore:1 white:1 visualizing:2 tino:1 during:6 width:2 self:1 ambiguous:1 anything:1 excitation:1 criterion:1 generalized:1 neocognitron:1 iconip:1 disrupting:1 neocortical:1 felleman:2 fj:1 image:40 wise:1 novel:1 recently:1 fi:3 common:1 mt:4 rl:4 overview:1 cerebral:2 million:3 tail:1 a00:1 significant:1 spotlight:1 ai:2 cv:2 mistaken:2 vanilla:2 similarly:1 stable:4 cortex:6 inhibition:2 etc:1 multivariate:1 own:1 recent:1 closest:1 perspective:3 histology:1 driven:1 ship:1 schmidhuber:5 certain:4 outperforming:1 rep:3 vt:8 der:1 caltech:1 captured:1 seen:1 additional:1 zip:1 v3:1 signal:1 multiple:3 full:1 reduces:3 adapt:2 long:1 cifar:14 lin:1 post:1 promotes:1 a1:1 ensuring:1 prediction:1 scalable:1 variant:1 basic:1 breast:1 vision:8 expectation:1 regression:1 arxiv:7 iteration:2 represent:1 kernel:1 normalization:2 pyramid:2 microscopy:1 background:2 diagram:1 peek:1 ot:4 pass:4 pooling:11 tend:1 seem:1 extracting:1 feedforward:12 intermediate:1 enough:3 split:1 iii:1 vanrullen:1 timesteps:1 misclassifications:1 architecture:4 ciresan:4 psychology:2 behnke:2 reduce:1 idea:2 regarding:1 airplane:1 shift:1 whether:2 o0:1 colour:2 york:1 action:18 repeatedly:1 deep:15 dramatically:1 generally:1 modulatory:1 latency:1 aimed:1 discount:1 processed:3 category:1 reduced:2 generate:1 millisecond:1 neuroscience:2 correctly:1 vol:2 salient:2 drawn:1 changing:2 neither:1 douglas:1 backward:1 v1:3 sum:1 run:2 compete:1 unaugmented:2 powerful:1 respond:1 master:1 family:1 almost:1 roelfsema:1 patch:1 recognise:1 decision:1 bit:1 dropout:2 layer:50 followed:2 gomez:1 correspondence:1 truck:1 activity:1 strength:1 ijcnn:1 constraint:1 scene:3 x2:1 diffuse:1 explode:1 disambiguated:1 min:1 separable:2 gpus:1 influential:1 department:1 according:1 belonging:1 membrane:1 describes:1 rob:2 biologically:2 making:1 primate:2 computationally:1 equation:3 remains:1 committee:1 mechanism:1 ge:4 tractable:1 end:5 whitehead:1 sigman:1 apply:1 hierarchical:2 v2:1 recurrency:2 batch:2 gate:9 original:1 top:11 denotes:1 rain:1 zeiler:3 maintaining:1 emphasised:1 especially:1 objective:1 already:2 added:1 strategy:7 rt:2 usual:2 traditional:2 diagonal:1 evolutionary:1 gradient:3 lends:1 fovea:1 attentional:4 gecco:2 code:2 index:4 illustration:1 providing:1 demonstration:1 sermanet:1 difficult:3 suppress:1 implementation:2 policy:23 boltzmann:1 allowing:5 maximises:1 av:4 convolution:1 observation:8 datasets:4 neuron:2 acknowledge:1 extended:1 excluding:1 hinton:3 communication:2 discovered:1 rn:1 stack:2 ucsd:1 pair:2 required:1 dog:4 extensive:1 connection:13 meier:1 imagenet:1 california:1 learned:5 boost:3 pop:3 macaque:1 trans:1 nip:3 able:1 bar:2 topdown:1 pattern:3 perception:1 sparsity:1 max:3 including:1 dasnet:25 power:2 natural:8 rely:1 advanced:1 representing:1 improve:5 technology:1 eye:2 ne:6 numerous:1 categorical:1 review:2 epoch:2 l2:3 literature:1 understanding:2 evolve:1 relative:2 loss:1 fully:2 generation:4 interesting:1 approximator:1 agent:5 controversial:1 offered:1 tiny:1 a1c:1 heavy:1 eccv:2 cancer:1 course:1 last:2 keeping:2 salience:1 bias:1 allow:2 normalised:1 institute:1 face:6 differentiating:1 emphasise:1 benefit:1 van:1 feedback:14 dimension:2 calculated:1 cortical:3 cumulative:1 distributed:1 preventing:1 author:1 forward:1 reinforcement:8 projected:1 avg:1 made:1 simplified:1 qualitatively:1 sensory:1 adaptive:1 conclusively:1 global:1 decides:1 sequentially:1 active:1 xi:4 discriminative:1 fergus:3 search:4 iterative:2 table:2 learn:5 nature:1 robust:2 improving:2 automobile:1 complex:3 da:1 icann:1 scored:1 repeated:1 allowed:1 augmented:2 neuronal:1 ahuja:1 position:2 explicit:1 lugano:1 concatenating:1 candidate:3 exponential:1 answering:1 perceptual:2 psychologica:1 weighting:1 third:2 learns:1 masci:1 down:7 rk:1 erroneous:1 specific:1 showing:1 lamme:2 nyu:1 grouping:1 consist:2 sequential:3 corr:1 flattened:1 importance:1 phd:1 margin:1 chen:1 rg:2 simply:1 forming:1 bidirectionally:1 visual:11 positional:1 tracking:2 partially:1 saccade:1 binding:1 springer:3 ch:1 extracted:2 ma:4 acm:2 modulate:1 identity:1 maxout:23 shared:1 content:1 feasible:1 change:14 experimentally:1 included:2 determined:2 except:1 reducing:1 hard:1 averaging:2 uniformly:1 engineer:1 called:2 specie:1 pas:9 partly:1 e:1 experimental:1 meaningful:1 select:1 internal:6 support:1 jonathan:2 collins:1 srivastava:1 |
4,723 | 5,277 | Iterative Neural Autoregressive Distribution
Estimator (NADE-k)
Tapani Raiko
Aalto University
Li Yao
Universit?e de Montr?eal
KyungHyun Cho
Universit?e de Montr?eal
Yoshua Bengio
Universit?e de Montr?eal,
CIFAR Senior Fellow
Abstract
Training of the neural autoregressive density estimator (NADE) can be viewed as
doing one step of probabilistic inference on missing values in data. We propose
a new model that extends this inference scheme to multiple steps, arguing that
it is easier to learn to improve a reconstruction in k steps rather than to learn to
reconstruct in a single inference step. The proposed model is an unsupervised
building block for deep learning that combines the desirable properties of NADE
and multi-prediction training: (1) Its test likelihood can be computed analytically,
(2) it is easy to generate independent samples from it, and (3) it uses an inference
engine that is a superset of variational inference for Boltzmann machines. The
proposed NADE-k is competitive with the state-of-the-art in density estimation
on the two datasets tested.
1
Introduction
Traditional building blocks for deep learning have some unsatisfactory properties. Boltzmann machines are, for instance, difficult to train due to the intractability of computing the statistics of the
model distribution, which leads to the potentially high-variance MCMC estimators during training
(if there are many well-separated modes (Bengio et al., 2013)) and the computationally intractable
objective function. Autoencoders have a simpler objective function (e.g., denoising reconstruction
error (Vincent et al., 2010)), which can be used for model selection but not for the important choice
of the corruption function. On the other hand, this paper follows up on the Neural Autoregressive
Distribution Estimator (NADE, Larochelle and Murray, 2011), which specializes previous neural
auto-regressive density estimators (Bengio and Bengio, 2000) and was recently extended (Uria et al.,
2014) to deeper architectures. It is appealing because both the training criterion (just log-likelihood)
and its gradient can be computed tractably and used for model selection, and the model can be
trained by stochastic gradient descent with backpropagation. However, it has been observed that the
performance of NADE has still room for improvement.
The idea of using missing value imputation as a training criterion has appeared in three recent papers. This approach can be seen either as training an energy-based model to impute missing values
well (Brakel et al., 2013), as training a generative probabilistic model to maximize a generalized
pseudo-log-likelihood (Goodfellow et al., 2013), or as training a denoising autoencoder with a masking corruption function (Uria et al., 2014). Recent work on generative stochastic networks (GSNs),
which include denoising auto-encoders as special cases, justifies dependency networks (Heckerman et al., 2000) as well as generalized pseudo-log-likelihood (Goodfellow et al., 2013), but have
the disadvantage that sampling from the trained ?stochastic fill-in? model requires a Markov chain
(repeatedly resampling some subset of the values given the others). In all these cases, learning
progresses by back-propagating the imputation (reconstruction) error through inference steps of the
model. This allows the model to better cope with a potentially imperfect inference algorithm. This
learning-to-cope was introduced recently in 2011 by Stoyanov et al. (2011) and Domke (2011).
1
<1>
<2>
h
[2]
h
v
h
<1>
W
V W
<0>
v
<1>
h
<1>
V
W
v
<2>
v
h
<1>
[1]
U
<1>
[2]
h
V
<0>
W
v
h
<2>
[1]
U
V
<2>
[2]
V
<1>
h
V
v
<1>
<2>
v
h
[2]
T
V
V
<2>
h
[1]
T
<3>
h
[1]
T
W
W W
W W
<0>
v
v
<1>
T
<2>
[1]
T
W
v
<3>
Figure 1: The choice of a structure for NADE-k is very flexible. The dark filled halves indicate that
a part of the input is observed and fixed to the observed values during the iterations. Left: Basic
structure corresponding to Equations (6?7) with n = 2 and k = 2. Middle: Depth added as in
NADE by Uria et al. (2014) with n = 3 and k = 2. Right: Depth added as in Multi-Prediction Deep
Boltzmann Machine by Goodfellow et al. (2013) with n = 2 and k = 3. The first two structures are
used in the experiments.
The NADE model involves an ordering over the components of the data vector. The core of the
model is the reconstruction of the next component given all the previous ones. In this paper we
reinterpret the reconstruction procedure as a single iteration in a variational inference algorithm,
and we propose a version where we use k iterations instead, inspired by (Goodfellow et al., 2013;
Brakel et al., 2013). We evaluate the proposed model on two datasets and show that it outperforms
the original NADE (Larochelle and Murray, 2011) as well as NADE trained with the order-agnostic
training algorithm (Uria et al., 2014).
2
Proposed Method: NADE-k
We propose a probabilistic model called NADE-k for D-dimensional binary data vectors x. We start
by defining p? for imputing missing values using a fully factorial conditional distribution:
Y
p? (xmis | xobs ) =
p? (xi | xobs ),
(1)
i?mis
where the subscripts mis and obs denote missing and observed components of x. From the conditional distribution p? we compute the joint probability distribution over x given an ordering o (a
permutation of the integers from 1 to D) by
p? (x | o) =
D
Y
p? (xod | xo<d ),
(2)
d=1
where o<d stands for indices o1 . . . od?1 .
The model is trained to minimize the negative log-likelihood averaged over all possible orderings o
L(?) = Eo?D! [Ex?data [? log p? (x | o)]] .
(3)
using an unbiased, stochastic estimator of L(?)
D
log p? (xo?d | xo<d )
(4)
D?d+1
by drawing o uniformly from all D! possible orderings and d uniformly from 1 . . . D (Uria et al.,
2014). Note that while the model definition in Eq. (2) is sequential in nature, the training criterion
(4) involves reconstruction of all the missing values in parallel. In this way, training does not involve
picking or following specific orders of indices.
?
L(?)
=?
In this paper, we define the conditional model p? (xmis | xobs ) using a deep feedforward neural
network with nk layers, where we use n weight matrices k times. This can also be interpreted as
running k successive inference steps with an n-layer neural network.
The input to the network is
vh0i = m Ex?data [x] + (1 ? m) x
(5)
where m is a binary mask vector indicating missing components with 1, and is an elementwise multiplication. Ex?data [x] is an empirical mean of the observations. For simplicity, we give
2
Figure 2: The inner working mechanism of NADE-k. The left most column shows the data vectors x,
the second column shows their masked version and the subsequent columns show the reconstructions
vh0i . . . vh10i (See Eq. (7)).
equations for a simple structure with n = 2. See Fig. 1 (left) for the illustration of this simple
structure.
In this case, the activations of the layers at the t-th step are
hhti = ?(Wvht?1i + c)
(6)
vhti = m ?(Vhhti + b) + (1 ? m) x
(7)
where ? is an element-wise nonlinearity, ? is a logistic sigmoid function, and the iteration index t
runs from 1 to k. The conditional probabilities of the variables (see Eq. (1)) are read from the output
vhki as
hki
p? (xi = 1 | xobs ) = vi .
(8)
Fig. 2 shows examples of how vhti evolves over iterations, with the trained model.
The parameters ? = {W, V, c, b} can be learned by stochastic gradient descent to minimize ?L(?)
?
in Eq. (3), or its stochastic approximation ?L(?)
in Eq. (4), with the stochastic gradient computed
by back-propagation.
Once the parameters ? are learned, we can define a mixture model by using a uniform probability
over a set of orderings O. We can compute the probability of a given vector x as a mixture model
1 X
pmixt (x | ?, O) =
p? (x | o)
(9)
|O|
o?O
with Eq. (2). We can draw independent samples from the mixture by first drawing an ordering o and
then sequentially drawing each variable using xod ? p? (xod | xo<d ). Furthermore, we can draw
samples from the conditional p(xmis | xobs ) easily by considering only orderings where the observed
indices appear before the missing ones.
Pretraining It is well known that training deep networks is difficult without pretraining, and in our
experiments, we train networks up to kn = 7 ? 3 = 21 layers. When pretraining, we train the model
to produce good reconstructions vhti at each step t = 1 . . . k. More formally, in the pretraining
phase, we replace Equations (4) and (8) by
L?pre (?) = ?
k
Y hti
1X
D
log
p? (xi | xo<d )
D ? d + 1 k t=1
i?o
(10)
?d
hti
hti
p? (xi = 1 | xobs ) = vi .
2.1
(11)
Related Methods and Approaches
Order-agnostic NADE The proposed method follows closely the order-agnostic version of
NADE (Uria et al., 2014), which may be considered as the special case of NADE-k with k = 1. On
the other hand, NADE-k can be seen as a deep NADE with some specific weight sharing (matrices
W and V are reused for different depths) and gating in the activations of some layers (See Equation
(7)).
3
Additionally, Uria et al. (2014) found it crucial to give the mask m as an auxiliary input to the
network, and initialized missing values to zero instead of the empirical mean (See Eq. (5)). Due to
these differences, we call their approach NADE-mask. One should note that NADE-mask has more
parameters due to using the mask as a separate input to the network, whereas NADE-k is roughly k
times more expensive to compute.
Probabilistic Inference Let us consider the task of missing value imputation in a probabilistic
latent variable model. We get the conditional probability of interest by marginalizing out the latent
variables from the posterior distribution:
Z
p(xmis | xobs ) =
p(h, xmis | xobs )dh.
(12)
h
Accessing the joint distribution p(h, xmis | xobs ) directly is often harder than alternatively updating
h and xmis based on the conditional distributions p(h | xmis , xobs ) and p(xmis | h).1 Variational
inference is one of the representative examples that exploit this.
In variational inference, a factorial distribution q(h, xmis ) = q(h)q(xmis ) is iteratively fitted to
p(h, xmis | xobs ) such that the KL-divergence between q and p
Z
p(h, xmis | xobs )
KL[q(h, xmis )||p(h, xmis | xobs )] = ?
q(h, xmis ) log
dhdxmis
(13)
q(h, xmis )
h,xmis
is minimized. The algorithm alternates between updating q(h) and q(xmis ), while considering the
other one fixed.
As an example, let us consider a restricted Boltzmann machine (RBM) defined by
p(v, h) ? exp(b> v + c> h + h> Wv).
(14)
We can fit an approximate posterior distribution parameterized as q(vi = 1) = v?i and q(hj = 1) =
? j to the true posterior distribution by iteratively computing
h
? ? ?(W?
h
v + c)
(15)
? ? m ?(W> h + b) + (1 ? m) v.
v
(16)
>
We notice the similarity to Eqs. (6)?(7): If we assume ? = ? and V = W , the inference in the
NADE-k is equivalent to performing k iterations of variational inference on an RBM for the missing
values (Peterson and Anderson, 1987). We can also get variational inference on a deep Boltzmann
machine (DBM) using the structure in Fig. 1 (right).
Multi-Prediction Deep Boltzmann Machine Goodfellow et al. (2013) and Brakel et al. (2013)
use backpropagation through variational inference steps to train a deep Boltzmann machine. This
is very similar to our work, except that they approach the problem from the view of maximizing
the generalized pseudo-likelihood (Huang and Ogata, 2002). Also, the deep Boltzmann machine
lacks the tractable probabilistic interpretation similar to NADE-k (See Eq. (2)) that would allow
to compute a probability or to generate independent samples without resorting to a Markov chain.
Also, our approach is somewhat more flexible in the choice of model structures, as can be seen in
Fig. 1. For instance, in the proposed NADE-k, encoding and decoding weights do not have to be
shared and any type of nonlinear activations, other than a logistic sigmoid function, can be used.
Product and Mixture of Experts One could ask what would happen if we would define an ensemble
likelihood along the line of the training criterion in Eq. (3). That is,
? log pprod (x | ?) ? Eo?D! [? log p(x | ?, o)] .
(17)
Maximizing this ensemble likelihood directly will correspond to training a product-of-experts
model (Hinton, 2000). However, this requires us to evaluate the intractable normalization constant
during training as well as in the inference, making the model not tractable anymore.
On the other hand, we may consider using the log-probability of a sample under the mixture-ofexperts model as the training criterion
? log pmixt (x | ?) = ? log Eo?D! [p(x | ?, o)] .
(18)
This criterion resembles clustering, where individual models may specialize in only a fraction of the
data. In this case, however, the simple estimator such as in Eq. (4) would not be available.
1
We make a typical assumption that observations are mutually independent given the latent variables.
4
Model
NADE 1HL(fixed order)
NADE 1HL
NADE 2HL
NADE-mask 1HL
NADE-mask 2HL
NADE-mask 4HL
EoNADE-mask 1HL(128 Ords)
EoNADE-mask 2HL(128 Ords)
Log-Prob.
-88.86
-99.37
-95.33
-92.17
-89.17
-89.60
-87.71
-85.10
Model
RBM (500h, CD-25)
DBN (500h+2000h)
DARN (500h)
DARN (500h, adaNoise)
NADE-5 1HL
NADE-5 2HL
EoNADE-5 1HL(128 Ords)
EoNADE-5 2HL(128 Ords)
Log-Prob.
? -86.34
? -84.55
? -84.71
? -84.13
-90.02
-87.14
-86.23
-84.68
Table 1: Results obtained on MNIST using various models and number of hidden layers (1HL
or 2HL). ?Ords? is short for ?orderings?. These are the average log-probabilities of the test set.
EoNADE refers to the ensemble probability (See Eq. (9)). From here on, in all figures and tables we
use ?HL? to denote the number of hidden layers and ?h? for the number of hidden units.
3
Experiments
We study the proposed model with two datasets: binarized MNIST handwritten digits and Caltech
101 silhouettes.
We train NADE-k with one or two hidden layers (n = 2 and n = 3, see Fig. 1, left and middle)
with a hyperbolic tangent as the activation function ?(?). We use stochastic gradient descent on
the training set with a minibatch size fixed to 100. We use AdaDelta (Zeiler, 2012) to adaptively
choose a learning rate for each parameter update on-the-fly. We use the validation set for earlystopping and to select the hyperparameters. With the best model on the validation set, we report the
log-probability computed on the test set. We have made our implementation available2 .
3.1
MNIST
We closely followed the procedure used by Uria et al. (2014), including the split of the dataset into
50,000 training samples, 10,000 validation samples and 10,000 test samples. We used the same
version where the data has been binarized by sampling.
We used a fixed width of 500 units per hidden layer. The number of steps k was selected among
{1, 2, 4, 5, 7}. According to our preliminary experiments, we found that no separate regularization
was needed when using a single hidden layer, but in case of two
hidden layers, we used weight
decay with the regularization constant in the interval e?5 , e?2 . Each model was pretrained for
1000 epochs and fine-tuned for 1000 epochs in the case of one hidden layer and 2000 epochs in the
case of two.
For both NADE-k with one and two hidden layers, the validation performance was best with k = 5.
The regularization constant was chosen to be 0.00122 for the two-hidden-layer model.
Results We report in Table 1 the mean of the test log-probabilities averaged over randomly selected
orderings. We also show the experimental results by others from (Uria et al., 2014; Gregor et al.,
2014). We denote the model proposed in (Uria et al., 2014) as a NADE-mask.
From Table 1, it is clear that NADE-k outperforms the corresponding NADE-mask both with the
individual orderings and ensembles over orderings using both 1 or 2 hidden layers. NADE-k with
two hidden layers achieved the generative performance comparable to that of the deep belief network
(DBN) with two hidden layers.
Fig. 3 shows training curves for some of the models. We can see that the NADE-1 does not perform
as well as NADE-mask. This confirms that in the case of k = 1, the auxiliary mask input is indeed
useful. Also, we can note that the performance of NADE-5 is still improving at the end of the
preallocated 2000 epochs, further suggesting that it may be possible to obtain a better performance
simply by training longer.
2
[email protected]:yaoli/nade k.git
5
120
?90
115
training cost
110
105
100
95
90
?94
end of pretrain
?96
NADE-mask 1HL
NADE-5 1HL
NADE-1 1HL
?98
end of pretrain
85
80
?92
testset log-probability
NADE-mask 1HL
NADE-5 1HL
NADE-1 1HL
0
500
1000
?100
200
1500
training epochs
400
600
(a)
800
1000
1200
1400
training epochs
1600
1800
2000
(b)
Figure 3: NADE-k with k steps of variational inference helps to reduce the training cost (a) and to
generalize better (b). NADE-mask performs better than NADE-1 without masks both in training and
test.
?87
?85
?90
?89
?90
?91
?92
?93
NADE-k 1HL
NADE-k 2HL
NADE-mask 1HL
NADE-mask 2HL
?94
?95
?96
1
2
4
5
trained with k steps of iterations
testset log-probability
testset log-probability
?88
?95
?100
?105
?110
?115
7
(a)
NADE-5 2HL
NADE-mask 2HL
0
5
10
15
perform k steps of iterations at test time
20
(b)
Figure 4: (a) The generalization performance of different NADE-k models trained with different k.
(b) The generalization performance of NADE-5 2h, trained with k=5, but with various k in test time.
Fig. 4 (a) shows the effect of the number of iterations k during training. Already with k = 2, we can
see that the NADE-k outperforms its corresponding NADE-mask. The performance increases until
k = 5. We believe the worse performance of k = 7 is due to the well known training difficulty of a
deep neural network, considering that NADE-7 with two hidden layers effectively is a deep neural
network with 21 layers.
At inference time, we found that it is important to use the exact k that one used to train the model.
As can be seen from Fig. 4 (b), the assigned probability increases up to the k, but starts decreasing
as the number of iterations goes over the k. 3
3.1.1
Qualitative Analysis
In Fig. 2, we present how each iteration t = 1 . . . k improves the corrupted input (vhti from Eq. (5)).
We also investigate what happens with test-time k being larger than the training k = 5. We can see
that in all cases, the iteration ? which is a fixed point update ? seems to converge to a point that is
in most cases close to the ground-truth sample. Fig. 4 (b) shows however that the generalization
performance drops after k = 5 when training with k = 5. From Fig. 2, we can see that the
reconstruction continues to be sharper even after k = 5, which seems to be the underlying reason
for this phenomenon.
3
In the future, one could explore possibilities for helping better converge beyond step k, for instance by
using costs based on reconstructions at k ? 1 and k even in the fine-tuning phase.
6
(b) Caltech-101 Silhouettes
(a) MNIST
Figure 5: Samples generated from NADE-k trained on (a) MNIST and (b) Caltech-101 Silhouettes.
(b)
(a)
Figure 6: Filters learned from NADE-5 2HL. (a) A random subset of the encodering filters. (b) A
random subset of the decoding filters.
From the samples generated from the trained NADE-5 with two hidden layers shown in Fig. 5 (a),
we can see that the model is able to generate digits. Furthermore, the filters learned by the model
show that it has learned parts of digits such as pen strokes (See Fig. 6).
3.1.2
Variability over Orderings
In Section 2, we argued that we can perform any inference task p(xmis | xobs ) easily and efficiently
by restricting the set of orderings O in Eq. (9) to ones where xobs is before xmis . For this to work
well, we should investigate how much the different orderings vary.
To measure the variability over orderings, we computed the variance of log p(x | o) for 128 randomly chosen orderings o with the trained NADE-k?s and NADE-mask with a single hidden layer.
For comparison, we computed the variance of log p(x | o) over the 10,000 test samples.
p
p
log p(x | o)
Eo,x [?]
Ex Varo [?]
Eo Varx [?]
Table 2: The variance of
NADE-mask 1HL -92.17
3.5
23.5
log p(x | o) over orderings o
NADE-5 1HL
-90.02
3.1
24.2
and over test samples x.
NADE-5 2HL
-87.14
2.4
22.7
In Table 2, the variability over the orderings is clearly much smaller than that over the samples.
Furthermore, the variability over orderings tends to decrease with the better models.
3.2
Caltech-101 silhouettes
We also evaluate the proposed NADE-k on Caltech-101 Silhouettes (Marlin et al., 2010), using
the standard split of 4100 training samples, 2264 validation samples and 2307 test samples. We
demonstrate the advantage of NADE-k compared with NADE-mask under the constraint that they
have a matching number of parameters. In particular, we compare NADE-k with 1000 hidden
units with NADE-mask with 670 hiddens. We also compare NADE-k with 4000 hidden units with
NADE-mask with 2670 hiddens.
We optimized the hyper-parameter k ? {1, 2, . . . , 10} in the case of NADE-k. In both NADE-k
and NADE-mask, we experimented without regularizations, with weight decays, or with dropout.
Unlike the previous experiments, we did not use the pretraining scheme (See Eq. (10)).
7
Table 3: Average log-probabilities of test samples of Caltech-101 Silhouettes. (?) The results are
from Cho et al. (2013). The terms in the parenthesis indicate the number of hidden units, the total
number of parameters (M for million), and the L2 regularization coefficient. NADE-mask 670h
achieves the best performance without any regularizations.
Model
RBM?
(2000h, 1.57M)
NADE-mask
(670h, 1.58M)
NADE-2
(1000h, 1.57M, L2=0.0054)
Test LL
-108.98
Model
RBM ?
(4000h, 3.14M)
NADE-mask
(2670h, 6.28M, L2=0.00106)
NADE-5
(4000h, 6.28M, L2=0.0068)
-112.51
-108.81
Test LL
-107.78
-110.95
-107.28
As we can see from Table 3, NADE-k outperforms the NADE-mask regardless of the number of
parameters. In addition, NADE-2 with 1000 hidden units matches the performance of an RBM with
the same number of parameters. Futhermore, NADE-5 has outperformed the previous best result
obtained with the RBMs in (Cho et al., 2013), achieving the state-of-art result on this dataset. We
can see from the samples generated by the NADE-k shown in Fig. 5 (b) that the model has learned
the data well.
4
Conclusions and Discussion
In this paper, we proposed a model called iterative neural autoregressive distribution estimator
(NADE-k) that extends the conventional neural autoregressive distribution estimator (NADE) and its
order-agnostic training procedure. The proposed NADE-k maintains the tractability of the original
NADE while we showed that it outperforms the original NADE as well as similar, but intractable
generative models such as restricted Boltzmann machines and deep belief networks.
The proposed extension is inspired from the variational inference in probabilistic models such as
restricted Boltzmann machines (RBM) and deep Boltzmann machines (DBM). Just like an iterative
mean-field approximation in Boltzmann machines, the proposed NADE-k performs multiple iterations through hidden layers and a visible layer to infer the probability of the missing value, unlike
the original NADE which performs the inference of a missing value in a single iteration through
hidden layers.
Our empirical results show that this approach of multiple iterations improves the performance of
a model that has the same number of parameters, compared to performing a single iteration. This
suggests that the inference method has significant effect on the efficiency of utilizing the model
parameters. Also, we were able to observe that the generative performance of NADE can come
close to more sophisticated models such as deep belief networks in our approach.
In the future, more in-depth analysis of the proposed NADE-k is needed. For instance, a relationship between NADE-k and the related models such as the RBM need to be both theoretically and
empirically studied. The computational speed of the method could be improved both in training (by
using better optimization algorithms. See, e.g., (Pascanu and Bengio, 2014)) and in testing (e.g. by
handling the components in chunks rather than fully sequentially). The computational efficiency of
sampling for NADE-k can be further improved based on the recent work of Yao et al. (2014) where
an annealed Markov chain may be used to efficiently generate samples from the trained ensemble.
Another promising idea to improve the model performance further is to let the model adjust its own
confidence based on d. For instance, in the top right corner of Fig. 2, we see a case with lots of missing values values (low d), where the model is too confident about the reconstructed digit 8 instead
of the correct digit 2.
Acknowledgements
The authors would like to acknowledge the support of NSERC, Calcul Qu?ebec, Compute Canada,
the Canada Research Chair and CIFAR, and developers of Theano (Bergstra et al., 2010; Bastien
et al., 2012).
8
References
Bastien, F., Lamblin, P., Pascanu, R., Bergstra, J., Goodfellow, I. J., Bergeron, A., Bouchard, N.,
and Bengio, Y. (2012). Theano: new features and speed improvements. Deep Learning and
Unsupervised Feature Learning NIPS 2012 Workshop.
Bengio, Y. and Bengio, S. (2000). Modeling high-dimensional discrete data with multi-layer neural
networks. In NIPS?99, pages 400?406. MIT Press.
Bengio, Y., Mesnil, G., Dauphin, Y., and Rifai, S. (2013). Better mixing via deep representations.
In Proceedings of the 30th International Conference on Machine Learning (ICML?13). ACM.
Bergstra, J., Breuleux, O., Bastien, F., Lamblin, P., Pascanu, R., Desjardins, G., Turian, J., WardeFarley, D., and Bengio, Y. (2010). Theano: a CPU and GPU math expression compiler. In
Proceedings of the Python for Scientific Computing Conference (SciPy). Oral Presentation.
Brakel, P., Stroobandt, D., and Schrauwen, B. (2013). Training energy-based models for time-series
imputation. The Journal of Machine Learning Research, 14(1), 2771?2797.
Cho, K., Raiko, T., and Ilin, A. (2013). Enhanced gradient for training restricted boltzmann machines. Neural computation, 25(3), 805?831.
Domke, J. (2011). Parameter learning with truncated message-passing. In Computer Vision and
Pattern Recognition (CVPR), 2011 IEEE Conference on, pages 2937?2943. IEEE.
Goodfellow, I., Mirza, M., Courville, A., and Bengio, Y. (2013). Multi-prediction deep boltzmann
machines. In Advances in Neural Information Processing Systems, pages 548?556.
Gregor, K., Danihelka, I., Mnih, A., Blundell, C., and Wierstra, D. (2014). Deep autoregressive
networks. In International Conference on Machine Learning (ICML?2014).
Heckerman, D., Chickering, D. M., Meek, C., Rounthwaite, R., and Kadie, C. (2000). Dependency networks for inference, collaborative filtering, and data visualization. Journal of Machine
Learning Research, 1, 49?75.
Hinton, G. E. (2000). Training products of experts by minimizing contrastive divergence. Technical
Report GCNU TR 2000-004, Gatsby Unit, University College London.
Huang, F. and Ogata, Y. (2002). Generalized pseudo-likelihood estimates for Markov random fields
on lattice. Annals of the Institute of Statistical Mathematics, 54(1), 1?18.
Larochelle, H. and Murray, I. (2011). The neural autoregressive distribution estimator. Journal of
Machine Learning Research, 15, 29?37.
Marlin, B., Swersky, K., Chen, B., and de Freitas, N. (2010). Inductive principles for restricted
Boltzmann machine learning. In Proceedings of The Thirteenth International Conference on
Artificial Intelligence and Statistics (AISTATS?10), volume 9, pages 509?516.
Pascanu, R. and Bengio, Y. (2014). Revisiting natural gradient for deep networks. In International
Conference on Learning Representations 2014(Conference Track).
Peterson, C. and Anderson, J. R. (1987). A mean field theory learning algorithm for neural networks.
Complex Systems, 1(5), 995?1019.
Stoyanov, V., Ropson, A., and Eisner, J. (2011). Empirical risk minimization of graphical model
parameters given approximate inference, decoding, and model structure. In International Conference on Artificial Intelligence and Statistics, pages 725?733.
Uria, B., Murray, I., and Larochelle, H. (2014). A deep and tractable density estimator. In Proceedings of the 30th International Conference on Machine Learning (ICML?14).
Vincent, P., Larochelle, H., Lajoie, I., Bengio, Y., and Manzagol, P.-A. (2010). Stacked denoising
autoencoders: Learning useful representations in a deep network with a local denoising criterion.
J. Machine Learning Res., 11.
Yao, L., Ozair, S., Cho, K., and Bengio, Y. (2014). On the equivalence between deep nade and generative stochastic networks. In European Conference on Machine Learning (ECML/PKDD?14).
Springer.
Zeiler, M. D. (2012). ADADELTA: an adaptive learning rate method. Technical report, arXiv
1212.5701.
9
| 5277 |@word version:4 middle:2 seems:2 reused:1 confirms:1 git:2 contrastive:1 tr:1 harder:1 series:1 tuned:1 yaoli:1 outperforms:5 freitas:1 com:1 od:1 varx:1 activation:4 gpu:1 uria:11 subsequent:1 happen:1 visible:1 drop:1 update:2 resampling:1 generative:6 half:1 selected:2 intelligence:2 core:1 short:1 regressive:1 pascanu:4 math:1 successive:1 simpler:1 wierstra:1 along:1 qualitative:1 specialize:1 ilin:1 combine:1 theoretically:1 mask:32 indeed:1 roughly:1 pkdd:1 multi:5 inspired:2 decreasing:1 cpu:1 xobs:15 considering:3 underlying:1 agnostic:4 what:2 interpreted:1 developer:1 marlin:2 pseudo:4 fellow:1 reinterpret:1 binarized:2 ebec:1 universit:3 unit:7 appear:1 danihelka:1 before:2 local:1 tends:1 encoding:1 subscript:1 resembles:1 studied:1 equivalence:1 suggests:1 averaged:2 arguing:1 testing:1 block:2 backpropagation:2 digit:5 procedure:3 empirical:4 eonade:5 hyperbolic:1 matching:1 pre:1 confidence:1 refers:1 bergeron:1 get:2 close:2 selection:2 risk:1 equivalent:1 conventional:1 missing:14 maximizing:2 annealed:1 go:1 regardless:1 simplicity:1 scipy:1 estimator:11 utilizing:1 fill:1 lamblin:2 ropson:1 annals:1 enhanced:1 exact:1 us:1 goodfellow:7 element:1 adadelta:2 expensive:1 recognition:1 updating:2 continues:1 observed:5 fly:1 revisiting:1 ordering:19 decrease:1 mesnil:1 accessing:1 trained:12 oral:1 efficiency:2 easily:2 joint:2 various:2 train:6 separated:1 stacked:1 london:1 artificial:2 hyper:1 larger:1 cvpr:1 drawing:3 reconstruct:1 statistic:3 advantage:1 propose:3 reconstruction:10 product:3 mixing:1 produce:1 help:1 propagating:1 ex:4 progress:1 eq:15 auxiliary:2 involves:2 indicate:2 larochelle:5 come:1 xod:3 closely:2 correct:1 filter:4 stochastic:9 argued:1 generalization:3 preliminary:1 extension:1 helping:1 considered:1 ground:1 exp:1 dbm:2 desjardins:1 vary:1 achieves:1 estimation:1 outperformed:1 minimization:1 mit:1 clearly:1 rather:2 hj:1 improvement:2 unsatisfactory:1 likelihood:9 aalto:1 pretrain:2 inference:25 hidden:22 among:1 flexible:2 dauphin:1 available2:1 art:2 special:2 field:3 once:1 sampling:3 unsupervised:2 icml:3 future:2 minimized:1 yoshua:1 others:2 report:4 mirza:1 randomly:2 divergence:2 individual:2 phase:2 montr:3 interest:1 message:1 investigate:2 possibility:1 mnih:1 adjust:1 mixture:5 chain:3 futhermore:1 filled:1 initialized:1 re:1 fitted:1 instance:5 eal:3 column:3 modeling:1 disadvantage:1 lattice:1 cost:3 tractability:1 subset:3 masked:1 uniform:1 too:1 encoders:1 dependency:2 kn:1 corrupted:1 cho:5 adaptively:1 chunk:1 density:4 hiddens:2 confident:1 international:6 probabilistic:7 decoding:3 picking:1 yao:3 schrauwen:1 huang:2 choose:1 worse:1 corner:1 expert:3 li:1 suggesting:1 de:4 bergstra:3 kadie:1 coefficient:1 preallocated:1 vi:3 view:1 lot:1 doing:1 compiler:1 competitive:1 start:2 maintains:1 parallel:1 bouchard:1 masking:1 collaborative:1 minimize:2 variance:4 efficiently:2 ensemble:5 correspond:1 generalize:1 handwritten:1 vincent:2 corruption:2 stroke:1 sharing:1 definition:1 energy:2 rbms:1 gcnu:1 mi:2 rbm:8 dataset:2 ask:1 improves:2 sophisticated:1 ords:5 back:2 improved:2 anderson:2 furthermore:3 just:2 autoencoders:2 until:1 hand:3 working:1 nonlinear:1 propagation:1 lack:1 minibatch:1 mode:1 logistic:2 scientific:1 believe:1 building:2 effect:2 true:1 unbiased:1 inductive:1 analytically:1 regularization:6 kyunghyun:1 read:1 assigned:1 iteratively:2 ll:2 during:4 impute:1 width:1 criterion:7 generalized:4 darn:2 demonstrate:1 performs:3 variational:9 wise:1 recently:2 sigmoid:2 imputing:1 empirically:1 hki:1 volume:1 million:1 interpretation:1 elementwise:1 significant:1 tuning:1 resorting:1 dbn:2 mathematics:1 nonlinearity:1 similarity:1 longer:1 posterior:3 own:1 recent:3 showed:1 binary:2 wv:1 caltech:6 seen:4 tapani:1 somewhat:1 eo:5 converge:2 maximize:1 multiple:3 desirable:1 stoyanov:2 infer:1 technical:2 match:1 cifar:2 parenthesis:1 prediction:4 basic:1 vision:1 arxiv:1 iteration:16 normalization:1 achieved:1 whereas:1 addition:1 fine:2 thirteenth:1 interval:1 crucial:1 breuleux:1 unlike:2 integer:1 call:1 feedforward:1 bengio:14 easy:1 superset:1 split:2 fit:1 architecture:1 ofexperts:1 imperfect:1 idea:2 inner:1 reduce:1 rifai:1 blundell:1 expression:1 passing:1 pretraining:5 repeatedly:1 deep:24 useful:2 clear:1 involve:1 factorial:2 dark:1 generate:4 notice:1 per:1 track:1 discrete:1 achieving:1 imputation:4 fraction:1 run:1 prob:2 parameterized:1 swersky:1 extends:2 draw:2 ob:1 comparable:1 dropout:1 layer:25 followed:1 meek:1 courville:1 constraint:1 speed:2 chair:1 performing:2 according:1 alternate:1 heckerman:2 smaller:1 appealing:1 qu:1 evolves:1 making:1 happens:1 hl:31 restricted:5 xo:5 theano:3 computationally:1 equation:4 mutually:1 visualization:1 mechanism:1 needed:2 tractable:3 end:3 available:1 observe:1 anymore:1 original:4 top:1 running:1 include:1 clustering:1 zeiler:2 graphical:1 exploit:1 eisner:1 murray:4 gregor:2 objective:2 added:2 already:1 traditional:1 gradient:7 separate:2 lajoie:1 reason:1 ozair:1 o1:1 index:4 relationship:1 illustration:1 manzagol:1 minimizing:1 difficult:2 potentially:2 sharper:1 negative:1 implementation:1 boltzmann:15 perform:3 observation:2 datasets:3 markov:4 acknowledge:1 descent:3 ecml:1 truncated:1 defining:1 extended:1 hinton:2 variability:4 canada:2 introduced:1 kl:2 optimized:1 engine:1 learned:6 tractably:1 nip:2 beyond:1 able:2 wardefarley:1 pattern:1 appeared:1 including:1 belief:3 difficulty:1 natural:1 scheme:2 improve:2 github:1 raiko:2 specializes:1 auto:2 autoencoder:1 rounthwaite:1 epoch:6 l2:4 tangent:1 acknowledgement:1 multiplication:1 marginalizing:1 calcul:1 python:1 fully:2 permutation:1 filtering:1 validation:5 principle:1 intractability:1 cd:1 senior:1 deeper:1 allow:1 institute:1 peterson:2 curve:1 depth:4 stand:1 autoregressive:7 author:1 made:1 adaptive:1 testset:3 cope:2 brakel:4 reconstructed:1 approximate:2 silhouette:6 sequentially:2 xi:4 alternatively:1 iterative:3 latent:3 pen:1 table:8 additionally:1 promising:1 learn:2 gsns:1 nature:1 improving:1 complex:1 european:1 did:1 aistats:1 hyperparameters:1 turian:1 fig:15 nade:103 representative:1 gatsby:1 stroobandt:1 chickering:1 hti:3 ogata:2 specific:2 bastien:3 gating:1 decay:2 experimented:1 intractable:3 workshop:1 mnist:5 restricting:1 sequential:1 effectively:1 justifies:1 nk:1 chen:1 easier:1 simply:1 explore:1 nserc:1 pretrained:1 springer:1 truth:1 dh:1 acm:1 conditional:7 viewed:1 presentation:1 room:1 replace:1 shared:1 typical:1 except:1 uniformly:2 domke:2 denoising:5 called:2 total:1 experimental:1 indicating:1 formally:1 select:1 college:1 support:1 evaluate:3 mcmc:1 tested:1 phenomenon:1 handling:1 |
4,724 | 5,278 | General Stochastic Networks for Classification
Matthias Z?ohrer and Franz Pernkopf
Signal Processing and Speech Communication Laboratory
Graz University of Technology
[email protected], [email protected]
Abstract
We extend generative stochastic networks to supervised learning of representations. In particular, we introduce a hybrid training objective considering a generative and discriminative cost function governed by a trade-off parameter ?. We use
a new variant of network training involving noise injection, i.e. walkback training, to jointly optimize multiple network layers. Neither additional regularization
constraints, such as `1, `2 norms or dropout variants, nor pooling- or convolutional layers were added. Nevertheless, we are able to obtain state-of-the-art performance on the MNIST dataset, without using permutation invariant digits and
outperform baseline models on sub-variants of the MNIST and rectangles dataset
significantly.
1
Introduction
Since 2006 there has been a boost in machine learning due to improvements in the field of unsupervised learning of representations. Most accomplishments originate from variants of restricted
Boltzmann machines (RBMs) [1], auto-encoders (AE) [2, 3] and sparse-coding [4, 5, 6]. Deep models in representation learning, also obtain impressive results in supervised learning problems, such
as speech recognition, e.g. [7, 8, 9] and computer vision tasks [10].
If no a-priori knowledge is modeled in the architecture, cf. convolutional layers or pooling layers
[11], generatively pre-trained networks are among the best when applied to supervised learning tasks
[12]. Usually, a generative representation is obtained through a greedy-layerwise training procedure
called contrastive divergence (CD) [1]. In this case, the network layer learns the representation from
the layer below by treating the latter as static input. Despite of the impressive results achieved with
CD, we identify two (minor) drawbacks when used for supervised learning: Firstly, after obtaining
a representation by pre-training a network, a new discriminative model is initialized with the trained
weights, splitting the training into two separate models. This seems to be neither biologically plausible, nor optimal when it comes to optimization, as carefully designed early stopping criteria have to
be implemented to prevent over- or under-fitting. Secondly, generative and discriminative objectives
might influence each other beneficially when combined during training. CD does not take this into
account.
In this work, we introduce a new training procedure for supervised learning of representations. In
particular we define a hybrid training objective for general stochastic networks (GSN), dividing the
cost function into a generative and discriminative part, controlled by a trade-off parameter ?. It turns
out that by annealing ?, when solving this unconstrained non-convex multi-objective optimization
problem, we do not suffer from the shortcomings described above. We are able to obtain stateof-the-art performance on the MNIST [13] dataset, without using permutation invariant digits and
significantly outperform baseline models on sub-variants of the MNIST and rectangle database [14].
Our approach is related to the generative-discriminative training approach of RBMs [15]. However
a different model and a new variant of network training involving noise injection, i.e. walkback
training [16, 17], is used to jointly optimize multiple network layers. Most notably, we did not
1
apply any additional regularization constraints, such as `1, `2 norms or dropout variants [12], [18],
unlocking further potential for possible optimizations. The model can be extended to learn multiple
tasks at the same time using jointly trained weights and by introducing multiple objectives. This
might also open a new prospect in the field of transfer learning [19] and multi-task learning [20]
beyond classification.
This paper is organized as follows: Section 2 presents mathematical background material i.e. the
GSN and a hybrid learning criterion. In Section 3 we empirically study the influence of hyper
parameters of GSNs and present experimental results. Section 4 concludes the paper and provides a
perspective on future work.
2
General Stochastic Networks
Recently, a new supervised learning algorithm called walkback training for generalized autoencoders (GAE) was introduced [16]. A follow-up study [17] defined a new network model ?
generative stochastic networks, extending the idea of walkback training to multiple layers. When
applied to image reconstruction, they were able to outperform various baseline systems, due to its
ability to learn multi-modal representations [17, 21]. In this paper, we extend the work of [17].
First, we provide mathematical background material for generative stochastic networks. Then, we
introduce modifications to make the model suitable for supervised learning. In particular we present
a hybrid training objective, dividing the cost into a generative and discriminative part. This paves
the way for any multi-objective learning of GSNs. We also introduce a new terminology, i.e. general stochastic networks, a model class including generative-, discriminative- and hybrid stochastic
network variants.
General Stochastic Networks for Unsupervised Learning
Restricted Boltzmann machines (RBM) [22] and denoising autoencoders (DAE) [3] share the following commonality; The input distribution P (X) is sampled to convergence in a Markov chain.
In the case of the DAE, the transition operator first samples the hidden state Ht from a corruption
distribution C(H|X), and generates a reconstruction from the parametrized model, i.e the density
P?2 (X|H).
Ht+1
P? 1
Ht+2
P?1
P?2
Xt+0
Ht+3
P?1
P?2
Xt+1
Ht+4
P?2
Xt+2
P? 1
P?1
P?2
Xt+3
Xt+4
Figure 1: DAE Markov chain.
The resulting DAE Markov chain, shown in Figure 1, is defined as
Ht+1 ? P?1 (H|Xt+0 ) and Xt+1 ? P?2 (X|Ht+1 ),
(1)
where Xt+0 is the input sample X, fed into the chain at time step 0 and Xt+1 is the reconstruction
of X at time step 1. In the case of a GSN, an additional dependency between the latent variables Ht
over time is introduced to the network graph. The GSN Markov chain is defined as follows:
Ht+1 ? P?1 (H|Ht+0 , Xt+0 ) and Xt+1 ? P?2 (X|Ht+1 ).
(2)
Figure 2 shows the corresponding network graph.
This chain can be expressed with deterministic functions of random variables f? ? {f?? , f?? }. In
particular, the density f? is used to model Ht+1 = f? (Xt+0 , Zt+0 , Ht+0 ), specified for some independent noise source Zt+0 , with the condition that Xt+0 cannot be recovered exactly from Ht+1 .
2
Ht+0
Ht+1
Ht+2
P?1
Ht+3
P?1
P?2
Xt+0
Ht+4
P?1
P?2
Xt+1
P?1
P?1
P?2
Xt+2
P? 2
Xt+3
Xt+4
Figure 2: GSN Markov chain.
We introduce f??i as a back-probable stochastic non-linearity of the form f??i = ?out + g(?in + a
?i )
with noise processes Zt ? {?in , ?out } for layer i. The variable a
?i is the activation for unit i, where
a
?i = W i Iti + bi with a weight matrix W i and bias bi , representing the parametric distribution. It is
embedded in a non-linear activation function g. The input Iti is either the realization xit of observed
sample Xti or the hidden realization hit of Hti . In general, f??i (Iti ) specifies an upward path in a GSN
i
for a specific layer i. In the case of Xt+1
= f??i (Zt+0 , Ht+1 ) we define f??i (Hti ) = ?out + g(?in + a
?i )
i
i
i T
i
as a downward path in the network i.e. a
? = (W ) Ht + b , using the transpose of the weight
matrix W i and the bias bi . This formulation allows to directly back-propagate the reconstruction log-likelihood P (X|H) for all parameters ? ? {W 0 , ..., W d , b0 , ..., bd } where d is the
number of hidden layers. In Figure 2 the GSN includes a simple hidden layer. This can be
extended to multiple hidden layers requiring multiple deterministic functions of random variables
f? ? {f??0 , ..., f??d , f??0 , ...f??d }.
Figure 3 visualizes the Markov chain for a multi-layer GSN, inspired by the unfolded computational
graph of a deep Boltzmann machine Gibbs sampling process.
3
Ht+3
3
Ht+4
f??2
f??2
2
Ht+2
f??1
f??1
f??1
f??1
1
Ht+3
f??0
f??2
2
Ht+4
f??1
1
Ht+2
f??0
f??0
f??2
2
Ht+3
f??1
1
Ht+1
f??0
f??2
f??0
f??1
1
Ht+4
f??0
f??0
f??0
f??0
0
Xt+0
0
Xt+1
0
Xt+2
0
Xt+3
0
Xt+4
Xt+0
0
Lt {Xt+1
, Xt+0 }
0
Lt {Xt+2
, Xt+0 }
0
Lt {Xt+3
, Xt+0 }
0
Lt {Xt+4
, Xt+0 }
Figure 3: GSN Markov chain with multiple layers and backprop-able stochastic units.
In the training case, alternatively even or odd layers are updated at the same time. The information
is propagated both upwards and downwards for K steps allowing the network to build higher order
representations. An example for this update process is given in Figure 3. In the even update (marked
2
1
0
0
1
= f??0 (Ht+1
) and Ht+2
=
in red) Ht+1
= f??0 (Xt+0
). In the odd update (marked in blue) Xt+1
?
?
?
?
2
1
1
0
2
3
0
1
2
1
f? (Ht+1 ) for k = 0. In the case of k = 1, Ht+2 = f? (Xt+1 ) + f? (Ht+2 ) and Ht+3 = f? (Ht+2 )
1
2
1
3
0
= f??0 (Ht+2
) and Ht+3
= f??1 (Ht+2
) + f??2 (Ht+3
) in the odd update.
in the even update and Xt+2
?
?
?
1
0
2
3
2
0
1
2
In case of k = 2, Ht+3 = f? (Xt+2 ) + f? (Ht+3 ) and Ht+4 = f? (Ht+3 ) in the even update and
0
1
2
1
3
Xt+3
= f??0 (Ht+3
) and Ht+4
= f??1 (Ht+3
) + f??2 (Ht+4
) in the odd update.
The cost function of a generative GSN can be written as:
C=
K
X
0
Lt {Xt+k
, Xt+0 },
k=1
3
(3)
Lt is a specific loss-function such as the mean squared error (MSE) at time step t. In general any
0
arbitrary loss function could be used (as long as they can be seen as a log-likelihood) [16]. Xt+k
0
is the reconstruction of the input Xt+0 at layer 0 after k steps. Optimizing the loss function by
building the sum over the costs of multiple corrupted reconstructions is called walkback training
[16, 17]. This form of network training leads to a significant performance boost when used for input
reconstruction. The network is able to handle multi-modal input representations and is therefore
considerably more favorable than standard generative models [16].
General Stochastic Networks for Supervised Learning
In order to make a GSN suitable for a supervised learning task we introduce the output Y to the
network graph. In this case L = log P (X) + log P (Y |X). Although the target Y is not fed into the
network, it is introduced as an additional cost term. The layer update-process stays the same.
3
Lt {Ht+1
, Yt+0 }
3
Lt {Ht+2
, Yt+0 }
3
Ht+3
3
Ht+4
f??2
f??2
2
Ht+2
f??1
f??1
f??0
f??1
f??1
1
Ht+3
f??0
f??2
2
Ht+4
f??1
1
Ht+2
f??0
f??2
2
Ht+3
f??1
1
Ht+1
f??0
f??2
f??0
f??1
1
Ht+4
f??0
f??0
f??0
f??0
0
Xt+0
0
Xt+1
0
Xt+2
0
Xt+3
0
Xt+4
Xt+0
0
Lt {Xt+1
, Xt+0 }
0
Lt {Xt+2
, Xt+0 }
0
Lt {Xt+3
, Xt+0 }
0
Lt {Xt+4
, Xt+0 }
Figure 4: GSN Markov chain for input Xt+0 and target Yt+0 with backprop-able stochastic units.
We define the following cost function for a 3-layer GSN:
C=
K
K
1?? X
? X
3
Lt {Xt+k , Xt+0 } +
Lt {Ht+k
, Yt+0 }
K
K ?d+1
k=1
k=d
|
{z
} |
{z
}
generative
discriminative
(4)
This is a non-convex multi-objective optimization problem, where ? weights the generative and
discriminative part of C. The parameter d specifies the number of network layers i.e. depth of the
network. Scaling the mean loss in (4) is not mandatory, but allows to equally balance both loss terms
with ? = 0.5 for input Xt+0 and target Yt+0 scaled to the same range. Again Figure 4 shows the
corresponding network graph for supervised learning with red and blue edges denoting the even and
odd network updates.
In general the hybrid objective optimization criterion is not restricted to hX, Y i, as additional input
and output terms could be introduced to the network. This setup might be useful for transfer-learning
[19] or multi-task scenarios [20], which is not discussed in this paper.
3
Experimental Results
In order to evaluate the capabilities of GSNs for supervised learning, we studied MNIST digits
[13], variants of MNIST digits [14] and the rectangle datasets [14]. The first database consists of
60.000 labeled training and 10.000 labeled test images of handwritten digits. The second dataset includes 6 variants of MNIST digits, i.e. { mnist-basic, mnist-rot, mnist-back-rand, mnist-back-image,
mnist-rot-back-image }, with additional factors of variation added to the original data. Each variant
includes 10.000 labeled training, 2000 labeled validation, and 50.000 labeled test images. The third
dataset involves two subsets, i.e. { rectangle, rectangle-image }. The dataset rectangle consists of
4
1000 labeled training, 200 labeled validation, and 50.000 labeled test images. The dataset rectangleimage includes 10.000 labeled train, 2000 labeled validation and 50.000 labeled test images.
In a first experiment we focused on the multi-objective optimization problem defined in (4). Next we
evaluated the number of walkback steps in a GSN, necessary for convergence. In a third experiment
we analyzed the influence of different Gaussian noise settings during walkback training, improving
the generalization capabilities of the network. Finally we summarize classification results for all
datasets and compare to baseline systems [14].
3.1
Multi-Objective Optimization in a Hybrid Learning Setup
In order to solve the non-convex multi-objective optimization problem, variants of stochastic gradient descent (SGD) can be used. We applied a search over fixed ? values on all problems. Furthermore, we show that the use of an annealed ? factor, during training works best in practice.
In all experiments a three layer GSN, i.e. GSN-3, with 2000 neurons in each layer, randomly initialized with small Gaussian noise, i.e. 0.01 ? N (0, 1), and an MSE loss function for both inputs and
targets was used. Regarding optimization we applied SGD with a learning rate ? = 0.1, a momentum term of 0.9 and a multiplicative annealing factor ?n+1 = ?n ? 0.99 per epoch n for the learning
rate. A rectifier unit [23] was chosen as activation function. Following the ideas of [24] no explicit
sampling was applied at the input and output layer. In the test case the zero-one loss was computed
averaging the network?s output over k walkback steps.
Analysis of the Hybrid Learning Parameter ?
Concerning the influence of the trade-off parameter ?, we tested fixed ? values in the range
? ? {0.01, 0.1, 0.2, ..., 0.9, 0.99}, where low values emphasize the discriminative part in the objective and vice versa. Walkback training with K = 6 steps using zero-mean pre- and postactivation Gaussian noise with zero mean and variance ? = 0.1 was performed for 500 training epochs. In a more dynamic scenario ?n=1 = 1 was annealed by ?n+1 = ?n ? ? to reach
?n=500 ? {0.01, 0.1, 0.2, ..., 0.9, 0.99} within 500 epochs, simulating generative pre-training to a
certain extend.
Figure 5: Influence of dynamic and static ? on MNIST variants basic (left), rotated (middle) and
background (right) where ? denotes the training-, 4 the validation- and 5 the test-set. The dashed
line denotes the static setup, the bold line the dynamic setup.
Figure 5 compares the results of both GSNs, using static and dynamic ? setups on the MNIST
variants basic, rotated and background. The use of a dynamic i.e. annealed ?n=500 = 0.01, achieved
the best validation and test error in all experiments. In this case, more attention was given to the
generative proportion P (X) of the objective (4) in the early stage of training. After approximately
400 epochs discriminative training i.e. fine-tuning, dominates. This setup is closely related to DBN
training, where emphasis is on optimizing P (X) at the beginning of the optimization, whereas
P (Y |X) is important at the last stages. In case of the GSN, the annealed ? achieves a more smooth
transition by shifting the weight in the optimization criterion from P (X) to P (Y |X) within one
model.
5
Analysis of Walkback Steps K
In a next experiment we tested the influence of K walkback steps for GSNs. Figure 6 shows the
results for different GSNs, trained with K ? {6, 7, 8, 9, 10} walkback steps and annealed ? with
? = 0.99. In all cases the information was at least propagated once up and once downwards in the
d = 3 layer network using fixed Gaussian pre- and post-activation noise with ? = 0 and ? = 0.1.
Figure 6: Evaluating the number of walkback steps on MNIST variants basic (left), rotated (middle)
and background (right) where ? denotes the training-, 4 the validation- and 5 the test-set.
Figure 6 shows that increasing the walkback steps, does not improve the generalization capabilities
of the used GSNs. The setup K = 2 ? d is sufficient for convergence and achieves the best validation
and test result in all experiments.
Analysis of Pre- and Post-Activation Noise
Injecting noise during the training process of GSNs serves as a regularizer and improves the generalization capabilities of the model [17]. In this experiment the influence of Gaussian pre- and
post-activation noise with ? = 0 and ? ? {0.05, 0.1, 0.15, 0.2, 0.25, 0.3} and deactivated noise
during training, was tested on a GSN-3 trained for K = 6 walkback steps. The trade-off factor
? was annealed with ? = 0.99. Figure 7 summarizes the results of the different GSNs for the
MNIST variants basic, rotated and background. Setting ? = 0.1 achieved the best overall result
on the validation- and test-set for all three experiments. In all other cases the GSNs either over- or
underfitted the data.
Figure 7: Evaluating noise injections during training on MNIST variants basic (left), rotated (middle)
and background (right) where ? denotes the training-, 4 the validation- and 5 the test-set.
3.2
MNIST results
Table 1 presents the average classification error of three runs of all MNIST variation datasets obtained by a GSN-3, using fixed Gaussian pre- and post-activation noise with ? = 0, ? = 0.1 and
K = 6 walkback steps. The hybrid learning parameter ? was annealed with ? = 0.99 and ?n=1 = 1.
A small grid test was performed in the range of N ? d with N ? {1000, 2000, 3000} neurons per
layer for d ? {1, 2, 3} layers to find the optimal network configuration.
6
Dataset
SVMrbf
SVMpoly NNet
DBN-1
SAA-3
DBN-3
GSN-3
mnist-basic
3.03
?0.15
3.69
?0.17
4.69
?0.19
3.94
?0.17
3.46
?0.16
3.11
?0.15
2.40
?0.04
mnist-rot*
11.11
?0.28
15.42
?0.32
18.11
?0.34
10.30
?0.27
10.30
?0.27
14.69
?0.31
8.66
?0.08
mnist-back-rand
14.58
?0.31
16.62
?0.33
20.04
?0.35
9.80
?0.26
11.28
?0.28
6.73
?0.22
9.38
?0.03
mnist-back-image
22.61
?0.37
24.01
?0.37
27.41
?0.39
16.15
?0.32
23.00
?0.37
16.31
?0.32
16.04
?0.04
mnist-rot-back-image*
55.18
?0.44
2.15
?0.13
56.41
?0.43
2.15
?0.13
62.16
?0.43
7.16
?0.23
47.39
?0.44
4.71
?0.19
51.93
?0.44
2.41
?0.13
52.21
?0.44
2.60
?0.14
43.86
?0.05
2.04
?0.04
24.04
?0.37
24.05
?0.37
33.20
?0.41
23.69
?0.37
24.05
?0.37
22.50
?0.37
22.10
?0.03
rectangles
rectangles-image
Table 1: MNIST variations and recangle results [14]; For datasets marked by (*) updated results are
shown [25].
Table 1 shows that a three layer GSN clearly outperforms all other models, except for the MNIST
random-background dataset. In particular, when comparing the GSN-3 to the radial basis function
support vector machine (SVMrbf), i.e. the second best model on MNIST basic, the GSN-3 achieved
an relative improvement of 20.79% on the test set. On the MNIST rotated dataset the GSN-3 was
able to beat the second best model i.e. DBN-1, by 15.92% on the test set. On the MNIST rotatedbackground there is an relative improvement of 7.25% on the test set between the second best model,
i.e. DBN-1, and the GSN-3. All results are statistically significant. Regarding the number of model
parameters, although we cannot directly compare the models in terms of network parameters, it is
worth to mention that a far smaller grid test was used to generate the results for all GSNs, cf. [14].
When comparing the classification error of the GSN-3 trained without noise, obtained in the previous
experiments (7) with Table 1, the GSN-3 achieved the test error of 2.72% on the MNIST variant
basic, outperforming all other models on this task. On the MNIST variant rotated, the GSN-3 also
outperformed the DBN-3, obtaining a test error of 11.2%. This indicates that not only the Gaussian
regularizer in the walkback training improves the generalization capabilities of the network, but also
the hybrid training criterion of the GSN.
Table 2 lists the results for the MNIST dataset without additional affine transformations applied to
the data i.e. permutation invariant digits. A three layer GSN achieved the state-of-the-art test error
of 0.80%.
Network
Result
Rectifier MLP + dropout [12]
DBM [26]
Maxout MLP + dropout [27]
MP-DBM [28]
Deep Convex Network [29]
Manifold Tangent Classifier [30]
DBM + dropout [12]
GSN-3
1.05%
0.95%
0.94%
0.91%
0.83%
0.81%
0.79%
0.80%
Table 2: MNIST results.
7
It might be worth noting that in addition to the noise process in walkback training, no other regularizers, such as `1, `2 norms and dropout variants [12], [18] were used in the GSNs. In general ? 800
training epochs with early-stopping are necessary for GSN training.
All simulations1 were executed on a GPU with the help of the mathematical expression compiler
Theano [31].
4
Conclusions and Future Work
We have extended GSNs for classification problems. In particular we defined an hybrid multiobjective training criterion for GSNs, dividing the cost function into a generative and discriminative
part. This renders the need for generative pre-training unnecessary. We analyzed the influence of
the objective?s trade-off parameter ? empirically, showing that by annealing ? we outperform a
static choice of ?. Furthermore, we discussed effects of noise injections and sampling steps during
walkback training. As a conservative starting point we restricted the model to use only rectifier
units. Neither additional regularization constraints, such as `1, `2 norms or dropout variants [12],
[18], nor pooling- [11, 32] or convolutional layers [11] were added. Nevertheless, the GSN was
able to outperform various baseline systems, in particular a deep belief network (DBN), a multi
layer perceptron (MLP), a support vector machine (SVM) and a stacked auto-associator (SSA), on
variants of the MNIST dataset. Furthermore, we also achieved state-of-the-art performance on the
original MNIST dataset without permutation invariant digits. The model not only converges faster
in terms of training iterations, but also show better generalization behavior in most cases. Our
approach opens a wide field of new applications for GSNs. In future research we explore adaptive
noise injection methods for GSNs and non-convex multi-objective optimization strategies.
References
[1] G. E. Hinton, S. Osindero, and Y. Teh, ?A fast learning algorithm for deep belief nets.? Neural computation, vol. 18, no. 7, pp. 1527?1554, 2006.
[2] Y. Bengio, P. Lamblin, D. Popovici, and H. Larochelle, ?Greedy layer-wise training of deep networks,? in
Advances in Neural Information Processing Systems (NIPS), 2007, pp. 153?160.
[3] P. Vincent, H. Larochelle, Y. Bengio, and P. Manzagol, ?Extracting and composing robust features with
denoising autoencoders,? in International Conference on Machine Learning (ICML), 2008, pp. 1096?
1103.
[4] H. Lee, A. Battle, R. Raina, and A. Y. Ng, ?Efficient sparse coding algorithms,? in Advances in Neural
Information Processing Systems (NIPS), 2007, pp. 801?808.
[5] J. Ngiam, Z. Chen, S. A. Bhaskar, P. W. Koh, and A. Y. Ng, ?Sparse filtering,? in Advances in Neural
Information Processing Systems (NIPS), 2011, pp. 1125?1133.
[6] M. Ranzato, M. Poultney, S. Chopra, and Y. LeCun, ?Efficient learning of sparse representations with an
energy-based model,? in Advances in Neural Information Processing Systems (NIPS), 2006, pp. 1137?
1144.
[7] G. E. Dahl, M. Ranzato, A. Mohamed, and G. E. Hinton, ?Phone recognition with the mean-covariance
restricted Boltzmann machine,? in Advances in Neural Information Processing Systems (NIPS), 2010, pp.
469?477.
[8] L. Deng, M. L. Seltzer, D. Yu, A. Acero, A. Mohamed, and G. E. Hinton, ?Binary coding of speech
spectrograms using a deep auto-encoder.? in Interspeech, 2010, pp. 1692?1695.
[9] F. Seide, G. Li, and D. Yu, ?Conversational speech transcription using context-dependent deep neural
networks.? in Interspeech, 2011, pp. 437?440.
[10] A. Krizhevsky, I. Sutskever, and G. E. Hinton, ?Imagenet classification with deep convolutional neural
networks,? in Advances in Neural Information Processing Systems (NIPS), 2012, pp. 1097?1105.
[11] Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, ?Gradient-based learning applied to document recognition,? Proceedings of the IEEE, vol. 86, no. 11, 1998.
[12] G. E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, ?Improving neural networks by preventing co-adaptation of feature detectors,? CoRR, vol. abs/1207.0580, 2012.
1
The code will be made publicly available for reproducing the results.
8
[13] Y. Lecun and C. Cortes, ?The MNIST database of handwritten digits,? 2014. [Online]. Available:
http://yann.lecun.com/exdb/mnist/
[14] H. Larochelle, D. Erhan, A. Courville, J. Bergstra, and Y. Bengio, ?An empirical evaluation of deep architectures on problems with many factors of variation,? in International Conference on Machine Learning
(ICML), 2007, pp. 473?480.
[15] H. Larochelle, M. Mandel, R. Pascanu, and Y. Bengio, ?Learning algorithms for the classification restricted Boltzmann machine,? Journal of Machine Learning Research (JMLR), vol. 13, pp. 643?669,
2012.
[16] Y. Bengio, L. Yao, G. Alain, and P. Vincent, ?Generalized denoising auto-encoders as generative models,?
in Advances in Neural Information Processing Systems (NIPS), 2013, pp. 899?907.
[17] Y. Bengio, E. Thibodeau-Laufer, and J. Yosinski, ?Deep generative stochastic networks trainable by backprop,? CoRR, vol. abs/1306.1091, 2013.
[18] L. Wan and M. Zeiler, ?Regularization of neural networks using dropconnect,? in International Conference on Machine Learning (ICML), 2013, pp. 109?111.
[19] G. Mesnil, Y. Dauphin, X. Glorot, S. Rifai, Y. Bengio, I. J. Goodfellow, E. Lavoie, X. Muller, G. Desjardins, D. Warde-Farley, P. Vincent, A. Courville, and J. Bergstra, ?Unsupervised and transfer learning
challenge: a deep learning approach,? in Unsupervised and Transfer Learning challenge and workshop
(JMLR W& CP), 2012, pp. 97?110.
[20] K. Abhishek and D. Hal, ?Learning task grouping and overlap in multi-task learning,? in International
Conference on Machine Learning (ICML), 2012.
[21] S. Ozair, L. Yao, and Y. Bengio, ?Multimodal transitions for generative stochastic networks.? CoRR, vol.
abs/1312.5578, 2013.
[22] P. Smolensky, Information processing in dynamical systems: Foundations of harmony theory.
Press, 1986, vol. 1, no. 1, pp. 194?281.
MIT
[23] X. Glorot, A. Bordes, and Y. Bengio, ?Deep sparse rectifier neural networks,? in International Conference
on Artificial Intelligence and Statisitics (AISTATS), 2011, pp. 315?323.
[24] G. E. Hinton, ?A practical guide to training restricted boltzmann machines,? in Neural Networks: Tricks
of the Trade (2nd ed.), ser. Lecture Notes in Computer Science. Springer, 2012, pp. 599?619.
[25] H. Larochelle, D. Erhan, A. Courville, J. Bergstra, and Y. Bengio, ?Online companion for
the paper an empirical evaluation of deep architectures on problems with many factors of
variation,? 2014. [Online]. Available: http://www.iro.umontreal.ca/?lisa/twiki/bin/view.cgi/Public/
DeepVsShallowComparisonICML2007
[26] R. Salakhutdinov and G. E. Hinton, ?Deep boltzmann machines,? in International Conference on Artificial
Intelligence and Statistics (AISTATS), 2009, pp. 448?455.
[27] I. J. Goodfellow, D. Warde-Farley, M. Mirza, A. Courville, and Y. Bengio, ?Maxout networks,? in International Conference on Machine Learning (ICML), 2013, pp. 1319?1327.
[28] I. J. Goodfellow, A. C. Courville, and Y. Bengio, ?Joint training deep boltzmann machines for classification,? CoRR, vol. abs/1301.3568, 2013.
[29] D. Yu and L. Deng, ?Deep convex net: A scalable architecture for speech pattern classification.? in Interspeech, 2011, pp. 2285?2288.
[30] S. Rifai, Y. Dauphin, P. Vincent, Y. Bengio, and X. Muller, ?The manifold tangent classifier,? in Advances
in Neural Information Processing Systems (NIPS), 2012, pp. 2294?2302.
[31] J. Bergstra, O. Breuleux, F. Bastien, P. Lamblin, R. Pascanu, G. Desjardins, J. Turian, D. Warde-Farley,
and Y. Bengio, ?Theano: a CPU and GPU math expression compiler,? in Python for Scientific Computing
Conference (SciPy), 2010.
[32] M. Zeiler and R. Fergus, ?Stochastic pooling for regularization of deep convolutional neural networks,?
CoRR, vol. abs/1301.3557, 2013.
9
| 5278 |@word middle:3 proportion:1 norm:4 nd:1 seems:1 open:2 propagate:1 covariance:1 contrastive:1 sgd:2 mention:1 generatively:1 configuration:1 denoting:1 document:1 outperforms:1 recovered:1 comparing:2 com:1 activation:7 bd:1 written:1 gpu:2 treating:1 designed:1 update:9 generative:21 greedy:2 intelligence:2 beginning:1 provides:1 pascanu:2 math:1 firstly:1 mathematical:3 consists:2 seide:1 fitting:1 introduce:6 notably:1 behavior:1 nor:3 multi:14 inspired:1 salakhutdinov:2 unfolded:1 xti:1 cpu:1 considering:1 increasing:1 linearity:1 transformation:1 exactly:1 scaled:1 hit:1 classifier:2 ser:1 unit:5 multiobjective:1 laufer:1 despite:1 path:2 approximately:1 might:4 emphasis:1 studied:1 co:1 bi:3 range:3 statistically:1 practical:1 lecun:4 practice:1 digit:9 procedure:2 empirical:2 significantly:2 pre:9 radial:1 mandel:1 cannot:2 operator:1 acero:1 context:1 influence:8 optimize:2 www:1 deterministic:2 yt:5 annealed:7 attention:1 starting:1 convex:6 focused:1 splitting:1 scipy:1 lamblin:2 gsn:33 handle:1 variation:5 updated:2 target:4 goodfellow:3 trick:1 recognition:3 database:3 labeled:11 observed:1 graz:1 ranzato:2 trade:6 prospect:1 mesnil:1 warde:3 dynamic:5 trained:6 solving:1 basis:1 multimodal:1 joint:1 various:2 regularizer:2 train:1 stacked:1 fast:1 shortcoming:1 artificial:2 hyper:1 walkback:19 plausible:1 solve:1 tested:3 encoder:1 ability:1 statistic:1 jointly:3 online:3 matthias:2 net:2 reconstruction:7 adaptation:1 realization:2 beneficially:1 sutskever:2 convergence:3 extending:1 converges:1 rotated:7 help:1 odd:5 minor:1 b0:1 dividing:3 implemented:1 involves:1 come:1 larochelle:5 drawback:1 closely:1 stochastic:17 material:2 seltzer:1 bin:1 backprop:3 public:1 hx:1 generalization:5 probable:1 secondly:1 dbm:3 desjardins:2 achieves:2 early:3 commonality:1 favorable:1 injecting:1 outperformed:1 harmony:1 vice:1 mit:1 clearly:1 gaussian:7 xit:1 improvement:3 likelihood:2 indicates:1 baseline:5 dependent:1 stopping:2 hidden:5 upward:1 overall:1 classification:10 among:1 dauphin:2 stateof:1 priori:1 art:4 field:3 once:2 ng:2 sampling:3 yu:3 unsupervised:4 icml:5 future:3 mirza:1 randomly:1 divergence:1 ab:5 mlp:3 nnet:1 evaluation:2 analyzed:2 farley:3 regularizers:1 chain:10 edge:1 necessary:2 initialized:2 dae:4 cost:8 introducing:1 subset:1 krizhevsky:2 osindero:1 encoders:2 dependency:1 corrupted:1 thibodeau:1 considerably:1 combined:1 density:2 international:7 stay:1 lee:1 off:5 yao:2 squared:1 again:1 wan:1 dropconnect:1 li:1 account:1 potential:1 bergstra:4 coding:3 bold:1 includes:4 mp:1 multiplicative:1 performed:2 view:1 red:2 compiler:2 capability:5 publicly:1 convolutional:5 variance:1 identify:1 handwritten:2 vincent:4 worth:2 corruption:1 visualizes:1 detector:1 reach:1 ed:1 rbms:2 energy:1 pp:22 mohamed:2 rbm:1 static:5 propagated:2 sampled:1 dataset:13 knowledge:1 improves:2 organized:1 carefully:1 back:8 higher:1 supervised:11 follow:1 modal:2 rand:2 formulation:1 evaluated:1 furthermore:3 stage:2 autoencoders:3 scientific:1 hal:1 building:1 effect:1 requiring:1 regularization:5 laboratory:1 during:7 interspeech:3 criterion:6 generalized:2 exdb:1 cp:1 upwards:1 image:11 wise:1 recently:1 umontreal:1 empirically:2 extend:3 discussed:2 yosinski:1 significant:2 versa:1 gibbs:1 tuning:1 unconstrained:1 dbn:7 grid:2 rot:4 impressive:2 perspective:1 optimizing:2 phone:1 mandatory:1 scenario:2 certain:1 outperforming:1 binary:1 muller:2 seen:1 additional:8 spectrogram:1 deng:2 signal:1 dashed:1 multiple:9 smooth:1 faster:1 long:1 concerning:1 post:4 equally:1 controlled:1 variant:22 involving:2 basic:9 ae:1 vision:1 scalable:1 iteration:1 achieved:7 background:8 whereas:1 fine:1 addition:1 annealing:3 source:1 breuleux:1 pooling:4 bhaskar:1 extracting:1 chopra:1 noting:1 bengio:15 architecture:4 idea:2 regarding:2 haffner:1 rifai:2 accomplishment:1 expression:2 suffer:1 render:1 speech:5 deep:18 useful:1 svmrbf:2 generate:1 specifies:2 outperform:5 http:2 per:2 blue:2 vol:9 terminology:1 nevertheless:2 prevent:1 neither:3 dahl:1 ht:62 rectangle:8 lavoie:1 graph:5 sum:1 run:1 yann:1 summarizes:1 scaling:1 dropout:7 layer:31 courville:5 constraint:3 generates:1 layerwise:1 conversational:1 injection:5 battle:1 smaller:1 biologically:1 modification:1 invariant:4 restricted:7 theano:2 koh:1 turn:1 fed:2 serf:1 available:3 apply:1 simulating:1 original:2 denotes:4 cf:2 tugraz:2 zeiler:2 build:1 objective:16 added:3 parametric:1 strategy:1 pave:1 gradient:2 separate:1 parametrized:1 cgi:1 originate:1 manifold:2 iro:1 ozair:1 code:1 modeled:1 manzagol:1 balance:1 setup:7 executed:1 zt:4 boltzmann:8 allowing:1 teh:1 neuron:2 markov:8 datasets:4 iti:3 descent:1 beat:1 unlocking:1 extended:3 communication:1 hinton:7 pernkopf:2 reproducing:1 arbitrary:1 introduced:4 specified:1 imagenet:1 boost:2 nip:8 able:8 beyond:1 usually:1 below:1 dynamical:1 pattern:1 smolensky:1 summarize:1 challenge:2 poultney:1 including:1 deactivated:1 belief:2 shifting:1 suitable:2 overlap:1 hybrid:11 raina:1 representing:1 improve:1 technology:1 concludes:1 auto:4 epoch:5 popovici:1 tangent:2 python:1 relative:2 embedded:1 loss:7 lecture:1 permutation:4 filtering:1 validation:9 foundation:1 affine:1 sufficient:1 share:1 cd:3 bordes:1 last:1 transpose:1 alain:1 bias:2 guide:1 lisa:1 perceptron:1 wide:1 sparse:5 depth:1 transition:3 evaluating:2 preventing:1 made:1 adaptive:1 franz:1 saa:1 far:1 erhan:2 emphasize:1 transcription:1 gae:1 unnecessary:1 discriminative:12 abhishek:1 alternatively:1 fergus:1 search:1 latent:1 table:6 learn:2 transfer:4 robust:1 ca:1 gsns:16 associator:1 obtaining:2 composing:1 improving:2 mse:2 ngiam:1 bottou:1 did:1 aistats:2 noise:18 turian:1 downwards:2 sub:2 momentum:1 explicit:1 governed:1 jmlr:2 third:2 hti:2 learns:1 companion:1 xt:61 specific:2 rectifier:4 showing:1 bastien:1 list:1 svm:1 cortes:1 ssa:1 dominates:1 glorot:2 workshop:1 grouping:1 mnist:37 corr:5 downward:1 chen:1 lt:14 explore:1 expressed:1 springer:1 marked:3 maxout:2 except:1 averaging:1 denoising:3 conservative:1 called:3 experimental:2 twiki:1 support:2 latter:1 evaluate:1 trainable:1 srivastava:1 |
4,725 | 5,279 | Improved Multimodal Deep Learning
with Variation of Information
Kihyuk Sohn, Wenling Shang and Honglak Lee
University of Michigan Ann Arbor, MI, USA
{kihyuks,shangw,honglak}@umich.edu
Abstract
Deep learning has been successfully applied to multimodal representation learning problems, with a common strategy to learning joint representations that are
shared across multiple modalities on top of layers of modality-specific networks.
Nonetheless, there still remains a question how to learn a good association between data modalities; in particular, a good generative model of multimodal data
should be able to reason about missing data modality given the rest of data modalities. In this paper, we propose a novel multimodal representation learning framework that explicitly aims this goal. Rather than learning with maximum likelihood, we train the model to minimize the variation of information. We provide a
theoretical insight why the proposed learning objective is sufficient to estimate the
data-generating joint distribution of multimodal data. We apply our method to restricted Boltzmann machines and introduce learning methods based on contrastive
divergence and multi-prediction training. In addition, we extend to deep networks
with recurrent encoding structure to finetune the whole network. In experiments,
we demonstrate the state-of-the-art visual recognition performance on MIR-Flickr
database and PASCAL VOC 2007 database with and without text features.
1
Introduction
Different types of multiple data modalities can be used to describe the same event. For example,
images, which are often represented with pixels or image descriptors, can also be described with
accompanying text (e.g., user tags or subtitles) or audio data (e.g., human voice or natural sound).
There have been several applications of multimodal learning from multiple domains such as emotion [13] and speech [10] recognition with audio-visual data, robotics applications with visual and
depth data [15, 17, 32, 23], or medical applications with visual and temporal data [26]. These data
from multiple sources are semantically correlated, and sometimes provide complementary information to each other. In order to exchange such information, it is important to capture a high-level
association between data modalities with a compact set of latent variables. However, learning associations between multiple heterogeneous data distributions is a challenging problem.
A naive approach is to concatenate the data descriptors from different sources of input to construct a
single high-dimensional feature vector and use it to solve a unimodal representation learning problem. Unfortunately, this approach has been unsuccessful since the correlation between features in
each data modality is much stronger than that between data modalities [21]. As a result, the learning
algorithms are easily tempted to learn dominant patterns in each data modality separately while giving up learning patterns that occur simultaneously in multiple data modalities. To resolve this issue,
deep learning methods, such as deep autoencoders [9] or deep Boltzmann machines (DBM) [24],
have been used to this problem [21, 27], with a common strategy to learning joint representations that
are shared across multiple modalities at the higher layer of the deep network after learning layers of
modality-specific networks. The rationale is that the learned features may have less within-modality
correlation than raw features, and this makes it easier to capture patterns across data modalities. Despite the promise, there still remains a challenging question how to learn a good association between
multiple data modalities that can effectively deal with missing data modalities in the testing time.
One necessary condition of being a good generative model of multimodal data is to have an ability
to predict or reason about missing data modalities given partial observation. To this end, we propose
1
a novel multimodal representation learning framework that explicitly aims this goal. The key idea
is to minimize the information distance between data modalities through the shared latent representations. More concretely, we train the model to minimize the variation of information (VI), an
information theoretic measure that computes the distance between random variables, i.e., multiple
data modalities. Note that this is in contrast to the previous approaches on multimodal deep learning, which are based on maximum (joint) likelihood (ML) learning [21, 27]. We provide an intuition
how our method could be more effective in learning the joint representation of multimodal data than
ML learning, and show theoretical insights why the proposed learning objective is sufficient to estimate the data-generating joint distribution of multimodal data. We apply the proposed framework to
multimodal restricted Boltzmann machine (MRBM). We introduce two learning algorithms, based
on contrastive divergence [19] and multi-prediction training [6]. Finally, we extend to multimodal
deep recurrent neural network (MDRNN) for unsupervised finetuning of whole network. In experiments, we demonstrate the state-of-the-art visual recognition performance on MIR-Flickr database
and PASCAL VOC 2007 database with and without text features.
2
Multimodal Learning with Variation of Information
In this section, we propose a novel training objective based on the VI. We make a comparison to
the ML objective, a typical learning objective for training models of multimodal data, to give an
insight how our proposal outperforms the baseline. Finally, we establish a theorem showing that the
proposed learning objective is sufficient to obtain a good generative model that fully recovers the
joint data-generating distribution of multimodal data.
Notation. We use uppercase letters X, Y to denote random variables, lowercase letters x, y for
realizations. Let PD be the data-generating distribution and P? the model distribution parameterized by ?. For presentation clarity, we slightly abuse the notation for Q to denote conditional
(Q(x|y), Q(y|x)), marginal (Q(x), Q(y)), as well as joint distributions (Q(x, y)) that are derived
from the joint distribution Q(x, y). The type of distribution for Q should be clear from the context.
2.1
Minimum Variation of Information Learning
Motivated from the necessary condition of good generative models to reason about the missing data
modality, it seems natural to learn to maximize the amount of information that one data modality has
about the others. We quantify such an amount of information between data modalities using variation
of information (VI). The VI is an information theoretic measure that computes the information
distance between two random variables (e.g., data modalities), and is written as follows:1
(1)
VIQ (X, Y ) = ?EQ(X,Y ) log Q(X|Y ) + log Q(Y |X)
where Q(X, Y ) = P? (X, Y ) is any joint distribution on random variables (X, Y ) parametrized by
?. Informally, VI is small when the conditional LLs Q(X|Y ) and Q(Y |X) are ?peaked?, meaning
that X has low entropy conditioned on Y and vice versa. Following the intuition, we define new
multimodal learning criteria, a minimum variation of information (MinVI) learning, as follows:
(2)
MinVI: min? LVI (?), LVI (?) = ?EPD (X,Y ) log P? (X|Y ) + log P? (Y |X)
Note the difference in LVI (?) that we take the expectation over PD in LVI (?). Furthermore, we
observe that the MinVI objective can be decomposed into a sum of two negative conditional LLs.
This indeed well aligns with our initial motivation about reasoning missing data modality. In the
following, we provide a more insight of our MinVI objective in relation to the ML objective, which
is a standard learning objective in generative models.
2.2 Relation to Maximum Likelihood Learning
The ML objective function can be written as a minimization of the negative LL (NLL) as follows:
ML: min? LNLL (?), LNLL (?) = ?EPD (X,Y ) log P? (X, Y ) ,
(3)
and we can show that the NLL objective function is reformulated as follows:
2LNLL (?) = KL (PD (X)kP? (X)) + KL (PD (Y )kP? (Y )) +
|
{z
}
(a)
EPD (X) KL (PD (Y |X)kP? (Y |X)) + EPD (Y ) KL (PD (X|Y )kP? (X|Y )) + C,
|
{z
}
(4)
(b)
1
In practice, we use finite samples of the training data and use a regularizer (e.g., l2 regularizer) to avoid
overfitting to the finite sample distribution.
2
where C is a constant which is irrelevant to ?. Note that (b) is equivalent to LVI (?) in Equation (2)
up to a constant. We provide a full derivation of Equation (4) in supplementary material.
Ignoring the constant, the NLL objective is composed of four terms of KL divergence. Since KL
divergence is non-negative and is 0 only when two distributions match, the ML learning in Equation (3) can be viewed as a distribution matching problem involving (a) marginal likelihoods and (b)
conditional likelihoods. Here, we argue that (a) is more difficult to optimize than (b) because there
are often too many modes in the marginal distribution. Compared to that, the number of modes can
be dramatically reduced in the conditional distribution since the conditioning variables may restrict
the support of random variable effectively. Therefore, (a) may become a dominant factor to be minimized during the optimization process and as a trade-off, (b) will be easily compromised, which
makes it difficult to learn a good association between data modalities. On the other hand, the MinVI
objective focuses on modelling the conditional distributions (Equation (4)), which is arguably easier to optimize. Indeed, similar argument has been made for generalized denoising autoencoders
(DAEs) [1] and generative stochastic networks (GSNs) [2], which focus on learning the transition
? where X
? is a corrupted version of data X, or P? (X|H), where H can be
operators (e.g., P? (X|X),
arbitrary latent variables) to bypass an intractable problem of learning density model P? (X).
2.3
Theoretical Results
Bengio et al. [1, 2] proved that learning transition operators of DAEs or GSNs is sufficient to learn
a good generative model that estimates a data-generating distribution. Under similar assumptions,
we establish a theoretical result that we can obtain a good density estimator for joint distribution
of multimodal data by learning the transition operators derived from the conditional distributions of
one data modality given the other. In multimodal learning framework, the transition operators TnX
and TnY with model distribution P?n (X, Y ) are defined
P for Markov chains of data modalities X and
Y , respectively. Specifically, TnX (x[t]|x[t ? 1]) = y?Y P?n (x[t]|y) P?n (y|x[t ? 1]) and TnY is
defined in a similar way. Now, we formalize the theorem as follows:
Theorem 2.1 For finite state space X , Y, if, ?x ? X , ?y ? Y, P?n (?|y) and P?n (?|x) converges in
probability to PD (?|y) and PD (?|x), respectively, and TnX and TnY are ergodic Markov chains, then,
as the number of examples n ? ?, the asymptotic distribution ?n (X) and ?n (Y ) converge to datagenerating marginal distributions PD (X) and PD (Y ), respectively. Moreover, the joint probability
distribution P?n (x, y) converges to PD (x, y) in probability.
The proof is provided in supplementary material. The theorem ensures that the MinVI objective
can lead to a good generative model estimating the joint data-generating distribution of multimodal
data. The theorem holds under two assumptions, consistency of density estimators and ergodicity
of transition operators. The ergodicity of transition operators are satisfied for wide variety of neural
networks, such as an RBM or DBM. 2 The consistency assumption is more difficult to satisfy and
the aforementioned deep energy-based models nor RNN may not satisfy the condition due to the
approximated posteriors using factorized distribution. Probably, deep networks that allow exact
posterior inference, such as stochastic feedforward neural networks [20, 29], could be a better model
in our multimodal learning framework, but we leave this as a future work.
3
Application to Multimodal Deep Learning
In this section, we describe the MinVI learning in multimodal deep learning framework. To overview
our pipeline, we use the commonly used network architecture that consists of layers of modalityspecific deep networks followed by a layer of neural network that jointly models the multiple modalities [21, 27]. The network is trained in two steps: In layer-wise pretraining, each layer of modalityspecific deep network is trained using restricted Boltzmann machines (RBMs). For the top-layer
shared network, we train MRBM with MinVI objective (Section 3.2). Then, we finetune the whole
deep network by constructing multimodal deep recurrent neural network (MDRNN) (Section 3.3).
3.1
Restricted Boltzmann Machines for Multimodal Learning
The restricted Boltzmann machine (RBM) is an undirected graphical model that defines the distribution of visible units using hidden units. For multimodal input, we define the joint distribution of
2
For energy-based models like RBM and DBM, it is straightforward to see that every state has non-zero
probability and can be reached from any other state. However, the mixing of the chain might be slow in practice.
3
multimodal RBM (MRBM) [21, 27] as P (x, y, h) =
E(x, y, h) = ?
Nx X
K
X
x
xi Wik
hk ?
i=1 k=1
Ny K
X
X
1
Z
exp ?E(x, y, h) with the energy function:
y
yj Wjk
hk ?
j=1 k=1
K
X
bk hk ?
Nx
X
i=1
k=1
cxi xi ?
Ny
X
cyj yj ,
(5)
j=1
where Z is the normalizing constant, x ? {0, 1}Nx , y ? {0, 1}Ny are the binary visible (i.e.,
observation) variables of multimodal input, and h ? {0, 1}K are the binary hidden (i.e., latent)
variables. W x ? RNx ?K defines the weights between x and h, and W y ? RNy ?K defines the
weights between y and h. cx ? RNx , cy ? RNy , and b ? RK are bias vectors corresponding to x,
y, and h, respectively. Note that the MRBM is equivalent to an RBM whose visible variables are
constructed by concatenating the visible variables of multiple input modalities, i.e., v = [x ; y].
Due to bipartite structure, variables in the same layer are conditionally independent given the variables of the other layer, and the conditional probabilities are written as follows:
X
X y
x
P (hk = 1 | x, y) = ?
Wik
xi +
Wjk yj + bk ,
(6)
i
P (xi = 1 | h) = ?
X
j
x
Wik
hk
X y
+ cxi , P (yj = 1 | h) = ?
Wjk hk + cyj ,
k
(7)
k
1
. Similarly to the standard RBM, the MRBM can be trained to maximize
where ?(x) = 1+exp(?x)
the joint LL (log P (x, y)) using stochastic gradient descent (SGD) while approximating the gradient
with contrastive divergence (CD) [8] or persistent CD (PCD) [30]. In our case, however, we train
the MRBM in MinVI criteria. We will discuss the inference and training algorithms in Section 3.2.
When we have access to all data modalities, we can use Equation (6) for exact posterior inference.
On the other hand, when some of the input modalities are missing, the inference is intractable, and
we resort to the variational method. For example, when we are given
x but no y, the true posterior can
Q Q
be approximated with a fully factorized distribution Q(y, h) = j k Q(yj )Q(hk ) by minimizing
the KL Q(y, h)kP? (y, h|x) . This leads to the following fixed-point equations:
X
X y
X y
x
?k = ?
? k + cy ,
h
Wik
xi +
Wjk y?j + bk , y?j = ?
Wjk h
(8)
j
i
j
k
? k = Q(hk ) and y?j = Q(yj ). The variational inference proceeds by alternately updating the
where h
? and y? that are initialized with all 0?s.
mean-field parameters h
3.2
Training Algorithms
CD-PercLoss. As in Equation (2), the objective function can be decomposed into two conditional
LLs, and the MRBM with MinVI objective can be trained equivalently by training the two conditional RBMs (CRBMs) while sharing the weights. Since the objective functions are the sum of
two conditional LLs, we compute the (approximate) gradient of each CRBM separately using CDPercLoss [19] and accumulate them to update parameters.3
Multi-Prediction. We found a few practical issues of CD-PercLoss training: First, the gradient
estimates are inaccurate. Second, there exists a difference between encoding process of training and
testing, especially when the unimodal query (e.g., one of the data modality is missing) is considered for testing. As an alternative objective, we propose multi-prediction (MP) training of MRBM
in MinVI criteria. The MP training was originally proposed to train deep Boltzmann machines
(DBMs) [6] as an alternative to the stochastic approximation procedure learning [24]. The idea is to
train the model good at predicting any subset of input variables given the rest of them by constructing
the recurrent network with encoding function derived from the variational inference problem.
The MP training can be adapted to train MRBM with MinVI objective with some modifications.
For example, the CRBM with an objective log P (y|x) can be trained by randomly selecting the
subset of variables to be predicted only from the target modality y, but the conditioning modality x
3
In CD-PercLoss learning, we run separate Gibbs chains for different conditioning variables and select the
negative particles with the lowest free energy among sampled particles. We refer [19] for further details.
4
Wx(1)
x=hx(0)
Wx(2)
hx(1)
Wx(3)
hx(2)
Wy(3)
h(3)
Wy(2)
hy(2)
Wy(1)
hy(1)
y=hy(0)
Figure 1: An instance of MDRNN with target y given x. Multiple iterations of bottom-up updates
(y ? h(3) ; Equation (11)) and top-down updates (h(3) ? y; Equation (13)) are performed. The
arrow indicates encoding direction.
is assumed to be given in all cases. Specifically, given an arbitrary subset s ? {1, ? ? ? , Ny } drawn
from the independent Bernoulli distribution PS , the MP algorithm predicts ys = {yj : j ? s} given
x and y\s = {yj : j ?
/ s} through the iterative encoding function derived from fixed-point equations
?k = ?
h
X
x
Wik
xi +
i
X
j?s
y
Wjk
y?j +
X
X y
y
? k + cy , j ? s,
Wjk
yj + bk , y?j = ?
Wjk h
j
j ?s
/
(9)
k
which is a solution to the variationalQ
inference
Q problem minQ KL Q(ys , h)kP? (ys , h|x, y\s ) with
factorized distribution Q(ys , h) = j?s k Q(yj )Q(hk ). Note that Equation (9) is similar to the
Equation (8) except that only yj , j ? s are updated. Using an iterative encoding function, the
network parameters are trained using SGD while computing the gradient by backpropagating the
error between the prediction and the ground truth of ys through the derived recurrent network. The
MP formulation (e.g., encoding function) of the CRBM with log P (x|y) can be derived similarly,
and the gradients are simply the addition of two gradients that are computed individually.
We have two additional hyper parameters, the number of mean-field updates and the sampling ratio
of a subset s to be predicted from the target data modality. In our experiments, it was sufficient to
use 10 ? 20 iterations until convergence. We used the sampling ratio of 1 (i.e., all the variables in
the target data modality are to be predicted) since we are already conditioned on one data modality,
which is sufficient to make a good prediction of variables in the target data modality.
3.3
Finetuning Multimodal Deep Network with Recurrent Neural Network
Motivated from the MP training of MRBM, we propose multimodal deep recurrent neural network
(MDRNN) that tries to predict the target modality given the input modality through the recurrent
encoding function, which iteratively performs a full pass of bottom-up and top-down encoding from
bottom-layer visible variables to top-layer joint representation back to bottom-layer through the
modality-specific deep networks. We show an instance of L = 3 layer MDRNN in Figure 1, and the
encoding functions are written as follows:4
(l)
x,(l)> (l?1)
x,(l)
x ? h(L?1)
:
h
=
?
W
h
+
b
, l = 1, ? ? ? , L ? 1 (10)
x
x
x
(l)
y ? h(L?1)
:
hy
= ? W y,(l)> h(l?1)
+ by,(l) , l = 1, ? ? ? , L ? 1 (11)
y
y
h(L?1)
, h(L?1)
? h(L) : h(L) = ? W x,(L)> h(L?1)
+ W y,(L)> h(L?1)
+ b(L)
(12)
x
y
x
y
(l?1)
y,(l?1)
h(L) ? y : hy
= ? W y,(l) h(l)
, l = L, ? ? ? , 1
(13)
y +b
(0)
(0)
where hx = x and hy = y. The visible variables of the target modality are initialized with 0?s.
In other words, in the initial bottom-up update, we compute h(L) only from x while setting y = 0
using Equation (10),(11),(12). Then, we run multiple iterations of top-down (Equation (13)) and
bottom-up updates (Equation (11), (12)). Finally, we compute the gradient by backpropagating the
reconstruction error of target modality through the network.
4
There could be different ways of constructing MDRNN; for instance, one can construct the RNN with
DBM-style mean-field updates. In our empirical evaluation, however, running full pass of bottom-up and topdown updates performed the best, and DBM-style updates didn?t give competitive results.
5
Ground Truth
Query
ML (PCD)
MinVI (CDPercLoss)
MinVI (MP)
Figure 2: Visualization of samples with inferred missing modality. From top to bottom, we visualize
ground truth, left or right halves of digits, generated samples with inferred missing modality using
MRBM with ML objective, MinVI objective using CD-PercLoss and MP training methods.
Input modalities at test time
ML (PCD)
MinVI (CD-PercLoss)
MinVI (MP)
Left+Right
1.57%
1.71%
1.73%
Left
14.98%
9.42%
6.58%
Right
18.88%
11.02%
7.27%
Table 1: Test set handwritten digit recognition errors of MRBMs trained with different objectives
and learning algorithms. Linear SVM was used for classification with joint feature representations.
4
Experiments
4.1
Toy Example on MNIST
In our first experiment, we evaluate the proposed learning algorithm on the MNIST handwritten
digit recognition dataset [16]. We consider left and right halves of the digit images as two input
modalities and report the recognition performance with different combinations of input modalities
at the test time, such as full (left + right) or missing (left or right) data modalities. We compare
the performance of the MRBM trained with 1) ML objective using PCD [30], or MinVI objectives
with 2) CD-PercLoss or 3) MP training. The recognition errors are provided in Table 1. Compared
to ML training, the recognition errors for unimodal queries are reduced by more than a half with
MP training of MinVI objective. For multimodal queries, the model trained with ML objective
performed the best, although the performance gain was incremental. CD-PercLoss training of MinVI
objective also showed significant improvement over ML training, but the errors were not as low
as those obtained with MP training. We believe that, although it is an approximation of MinVI
objective, the exact gradient for MP algorithm makes learning more efficient than CD-PercLoss.
For the rest of the paper, we focus on MP training method.
In Figure 2, we visualize the generated samples conditioned on one input modality (e.g., left or right
halves of digits). There are many samples generated by the models with MinVI objective that look
clearly better than those generated by the model with ML objective.
4.2
MIR-Flickr Database
In this section, we evaluate our methods on MIR-Flickr database [11], which is composed of 1 million examples of image and their user tags collected from the social photo-sharing website Flickr.5
Among those, 25000 examples are annotated with 24 potential topics and 14 regular topics, which
leads to 38 classes in total with distributed class membership. The topics include object categories
such as dog, flower, and people, or scenic concepts such as sky, sea, and night.
We used the same visual and text features as in [27].6 Specifically, the image feature is 3857 dimensional vector composed of Pyramid Histogram of Words (PHOW) features [3], GIST [22], and
MPEG-7 descriptors [18]. We preprocessed the image features to have zero mean and unit variance for each dimension across all examples. The text feature is a word count vector of 2000 most
frequent tags. The number of tags varies from 0 to 72, with 5.15 tags per example in average.
Following the experimental protocol [12, 27], we randomly split the labeled data into 15000 for
training and 10000 for testing, and used 5000 from training set for validation. We iterate the procedure for 5 times and report the mean average precision (mAP) over 38 classes.
Model Architecture. As used in [27], the network is composed of [3857, 1024, 1024] variables
for visual pathway, [2000, 1024, 1024] variables for text pathway, and 2048 variables for top-layer
MRBM. As described in Section 3, we pretrain the modality-specific deep networks in a greedy
5
6
http://www.flickr.com
http://www.cs.toronto.edu/?nitish/multimodal/index.html
6
layerwise manner, and finetune the whole network by initializing MDRNN with the pretrained network. Specifically, we used gaussian RBM for the bottom layer of visual pathway and binary RBM
for text pathway.7 The intermediate layers are trained with binary RBMs, and the top-layer MRBM
is trained using MP training algorithm. For the layer-wise pretraining of RBMs, we used PCD [30]
to approximate gradient. Since our algorithm requires both data modalities during the training, we
excluded examples with too sparse or no tags from unlabeled dataset and used about 750K examples with at least 2 tags. After unsupervised training, we extract joint feature representations of the
labeled training data and use them to train multiclass logistic regression classifiers.
Recognition Tasks. For recognition tasks,
Model
Multimodal query
we train multiclass logistic regression classiAutoencoder
0.610
fiers using joint representations as input feaMultimodal DBM [27]
0.609
tures. Depending on the availability of data
Multimodal DBM? [28]
0.641
modalities at testing time, we evaluate the perMK-SVM [7]
0.623
formance using multimodal queries (i.e., both
TagProp [31]
0.640
MDRNN
0.686 ? 0.003
visual and text data are available) and unimodal
queries (i.e., visual data is available while the
Model
Unimodal query
text data is missing). The summary results are
Autoencoder
0.495
Multimodal DBM [27]
0.531
in Table 2. We report the test set mAPs of our
0.530
MK-SVM [7]
proposed model and compared to other methMDRNN
0.607 ? 0.005
ods. The proposed MDRNN outperformed the
previous state-of-the-art in multimodal queries
Table 2: Test set mAPs on MIR-Flickr database.
by 4.5% in mAP. The performance improveWe implemented autoencoder following the dement becomes more significant for unimodal
scription in [21]. Multimodal DBM? is supervised
queries, achieving 7.6% improvement in mAP
finetuned model. See [28] for details.
over the best published result. As we used the
same input features in [27], the results suggest that our proposed algorithm learns better representations shared across multiple modalities.
To take a closer look into our model, we performed additional control experiment. In particular, we
explore the benefit of recurrent encoding network structure of MDRNN. We compare the performance of the models with different number of mean-field iterations.8 We report the validation set
mAPs of models with different number of iterations (0 ? 10) in Table 3. For multimodal query, the
MDRNN with 10 iterations improves the recognition performance by only 0.8% compared to the
model with 0 iterations. However, the improvement becomes significant for unimodal query, achieving 5.0% performance gain. In addition, we note that the largest improvement was made when we
have at least one iteration (from 0 to 1 iteration, 3.4% gain; from 1 to 10 iteration, 1.6% gain). This
suggests that the most crucial factor of improvement comes from the inference with reconstructed
missing data modality (e.g., text features), and the quality of inferred missing modality improves as
we increase the number of iterations.
# iterations
Multimodal query
Unimodal query
0
0.677
0.557
1
0.678
0.591
2
0.679
0.599
3
0.680
0.602
5
0.682
0.605
10
0.685
0.607
Table 3: Validation set mAPs on MIR-Flickr database with different number of mean-field iterations.
Retrieval Tasks. We perform retrieval tasks using multimodal and unimodal input queries. Following the experimental setting in [27], we select 5000 image-text pairs from the test set to form a
database and use 1000 disjoint set of examples from the test set as queries. For each query example,
we compute the relevance score to the data points as a cosine similarity of joint representations.
The binary relevance label between query and the data points are determined 1 if any of the 38
class labels are overlapped. Our proposed model achieves 0.633 mAP with multimodal query and
0.638 mAP with unimodal query. This significantly outperforms the performance of multimodal
DBM [27], which reported 0.622 mAP with multimodal query and 0.614 mAP with unimodal query.
7
We assume text features as binary, which is different from [27] where they modeled using replicatedsoftmax RBM [25]. The rationale is that the tags are not likely to be assigned more than once for single image.
8
In [21], they proposed the ?video-only? deep autoencoder whose objective is to predict audio data and
reconstruct video data when only video data is given as an input during the training. Our baseline model
(MDRNN with 0 iterations) is similar, but different since we don?t have a reconstruction training objective.
7
night, city, river,
dark, buildings, skyline
night, long exposure,
reflection, buildings,
massachusetts, boston
sunset, explore, sun
sunset, platinumphoto,
trees, silhouette
toys
lego
skyline, indiana, 1855mm
city, lights, buildings,
fireworks, skyscrapers
nikon, night, d80, asia,
skyline, hongkong, harbour
sunset, sol, searchthebest,
atardecer, nubes, abigfave
sunset
canon, naturesfinest, 30d
diy, robot
toy, plastic,
kitty, miniature
lego
Figure 3: Retrieval results with multimodal queries. The leftmost image-text pairs are multimodal
query samples and those in the right side of the bar are retrieved samples with the highest similarities
to the query sample from the database. We include more results in supplementary material.
4.3
PASCAL VOC 2007
We evaluate the proposed algorithm on PASCAL VOC 2007 database. The original dataset doesn?t
contain user tags, but Guillaumin et al. [7] has collected the user tags from Flickr website.9
Motivated from the success of convolutional neural networks (CNNs) on large-scale visual object
recognition [14], we used the DeCAF7 features [5] as an input features for visual pathway, where
DeCAF7 is 4096 dimensional feature extracted from the CNN trained on ImageNet [4]. For text
features, we used the vocabulary of size 804 suggested by [7]. For unsupervised feature learning of
MDRNN, we used unlabeled data of MIR-Flickr database while converting the text features using
the new vocabulary from PASCAL database. The network architecture used in this experiment is as
follows: [4096, 1536, 1536] variables for the visual pathway, [804, 512, 1536] variables for the text
pathway, and 2048 variables for top-layer joint network.
Following the standard practice, we report the mAP over 20 object classes. The performance improvement of our proposed method was significant, achieving 81.5% mAP with multimodal queries
and 76.2% mAP with unimodal queries, whereas the performance of baseline model was 74.5%
mAP with multimodal queries (DeCAF7 + Text) and 74.3% mAP with unimodal queries (DeCAF7 ).
5
Conclusion
Motivated from the property of good generative models of multimodal data, we proposed a novel
multimodal deep learning framework based on variation of information. The minimum variation of
information objective enables to learn a good shared representations of multiple heterogeneous data
modalities with a better prediction of missing input modality. We demonstrated the effectiveness
of our proposed method on multimodal RBM and its deep extensions and showed state-of-the-art
recognition performance on MIR-Flickr database and competitive performance on PASCAL VOC
2007 database with multimodal (visual + text) and unimodal (visual only) queries.
Acknowledgments
This work was supported in part by ONR N00014-13-1-0762, Toyota, and Google Faculty Research
Award.
References
[1] Y. Bengio, L. Yao, G. Alain, and P. Vincent. Generalized denoising auto-encoders as generative models.
In NIPS, 2013.
[2] Y. Bengio, E. Thibodeau-Laufer, G. Alain, and J. Yosinski. Deep generative stochastic networks trainable
by backprop. In ICML, 2014.
9
http://lear.inrialpes.fr/people/guillaumin/data.php
8
[3] A. Bosch, A. Zisserman, and X. Munoz. Image classification using random forests and ferns. In ICCV,
2007.
[4] J. Deng, W. Dong, R. Socher, L. Li, K. Li, and L. Fei-Fei. ImageNet: A large-scale hierarchical image
database. In CVPR, 2009.
[5] J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng, and T. Darrell. DeCAF: A deep convolutional activation feature for generic visual recognition. CoRR, abs/1310.1531, 2013.
[6] I. Goodfellow, M. Mirza, A. Courville, and Y. Bengio. Multi-prediction deep Boltzmann machines. In
NIPS, 2013.
[7] M. Guillaumin, J. Verbeek, and C. Schmid. Multimodal semi-supervised learning for image classification.
In CVPR, 2010.
[8] G. E. Hinton. Training products of experts by minimizing contrastive divergence. Neural Computation,
14(8):1771?1800, 2002.
[9] G. E. Hinton and R. Salakhutdinov. Reducing the dimensionality of data with neural networks. Science,
313(5786):504?507, 2006.
[10] J. Huang and B. Kingsbury. Audio-visual deep learning for noise robust speech recognition. In ICASSP,
2013.
[11] M. J. Huiskes and M. S. Lew. The MIR Flickr retrieval evaluation. In ICMIR, 2008.
[12] M. J. Huiskes, B. Thomee, and M. S. Lew. New trends and ideas in visual concept detection: The MIR
Flickr retrieval evaluation initiative. In ICMIR, 2010.
[13] Y. Kim, H. Lee, and E. M. Provost. Deep learning for robust feature generation in audiovisual emotion
recognition. In ICASSP, 2013.
[14] A. Krizhevsky, I. Sutskever, and G. E. Hinton. ImageNet classification with deep convolutional neural
networks. In NIPS, 2012.
[15] K. Lai, L. Bo, X. Ren, and D. Fox. RGB-D object recognition: Features, algorithms, and a large scale
benchmark. In Consumer Depth Cameras for Computer Vision, pages 167?192. Springer, 2013.
[16] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition.
Proceedings of the IEEE, 86(11):2278?2324, 1998.
[17] I. Lenz, H. Lee, and A. Saxena. Deep learning for detecting robotic grasps. In RSS, 2013.
[18] B. S. Manjunath, J-R. Ohm, V. V. Vasudevan, and A. Yamada. Color and texture descriptors. IEEE
Transactions on Circuits and Systems for Video Technology, 11(6):703?715, 2001.
[19] V. Mnih, H. Larochelle, and G. E. Hinton. Conditional restricted boltzmann machines for structured
output prediction. In UAI, 2011.
[20] R. M. Neal. Learning stochastic feedforward networks. Department of Computer Science, University of
Toronto, 1990.
[21] J. Ngiam, A. Khosla, M. Kim, J. Nam, H. Lee, and A. Y. Ng. Multimodal deep learning. In ICML, 2011.
[22] A. Oliva and A. Torralba. Modeling the shape of the scene: A holistic representation of the spatial
envelope. International Journal of Computer Vision, 42(3):145?175, 2001.
[23] D. Rao, M. De Deuge, N. Nourani-Vatani, B. Douillard, S. B. Williams, and O. Pizarro. Multimodal
learning for autonomous underwater vehicles from visual and bathymetric data. In ICRA, 2014.
[24] R. Salakhutdinov and G. E. Hinton. Deep Boltzmann machines. In AISTATS, 2009.
[25] R. Salakhutdinov and G. E. Hinton. Replicated softmax: an undirected topic model. In NIPS, 2009.
[26] H-C. Shin, M. R. Orton, D. J. Collins, S. J. Doran, and M. O. Leach. Stacked autoencoders for unsupervised feature learning and multiple organ detection in a pilot study using 4D patient data. IEEE
Transactions on Pattern Analysis and Machine Intelligence, 35(8):1930?1943, 2013.
[27] N. Srivastava and R. Salakhutdinov. Multimodal learning with deep Boltzmann machines. In NIPS, 2012.
[28] N. Srivastava and R. Salakhutdinov. Discriminative transfer learning with tree-based priors. In NIPS,
2013.
[29] Y. Tang and R. Salakhutdinov. Learning stochastic feedforward neural networks. In NIPS, 2013.
[30] T. Tieleman. Training restricted Boltzmann machines using approximations to the likelihood gradient. In
ICML, 2008.
[31] J. Verbeek, M. Guillaumin, T. Mensink, and C. Schmid. Image annotation with tagprop on the MIR Flickr
set. In ICMIR, 2010.
[32] A. Wang, J. Lu, G. Wang, J. Cai, and T-J. Cham. Multi-modal unsupervised feature learning for RGB-D
scene labeling. In ECCV. Springer, 2014.
9
| 5279 |@word cnn:1 faculty:1 version:1 stronger:1 seems:1 r:1 rgb:2 crbms:1 contrastive:4 datagenerating:1 sgd:2 initial:2 score:1 selecting:1 document:1 outperforms:2 com:1 od:1 activation:1 written:4 concatenate:1 visible:6 wx:3 shape:1 enables:1 gist:1 update:9 generative:11 half:4 website:2 greedy:1 intelligence:1 yamada:1 detecting:1 toronto:2 zhang:1 kingsbury:1 constructed:1 become:1 persistent:1 initiative:1 consists:1 pathway:7 manner:1 introduce:2 crbm:3 indeed:2 nor:1 multi:6 salakhutdinov:6 voc:5 decomposed:2 audiovisual:1 resolve:1 becomes:2 provided:2 estimating:1 notation:2 moreover:1 circuit:1 factorized:3 didn:1 lowest:1 indiana:1 temporal:1 sky:1 every:1 saxena:1 classifier:1 control:1 unit:3 medical:1 arguably:1 laufer:1 despite:1 encoding:11 abuse:1 might:1 suggests:1 challenging:2 rnx:2 practical:1 acknowledgment:1 camera:1 testing:5 yj:11 practice:3 lecun:1 skyline:3 digit:5 procedure:2 huiskes:2 shin:1 rnn:2 empirical:1 significantly:1 matching:1 word:3 regular:1 suggest:1 unlabeled:2 operator:6 thomee:1 context:1 optimize:2 equivalent:2 map:16 tny:3 missing:14 www:2 demonstrated:1 straightforward:1 exposure:1 williams:1 minq:1 ergodic:1 insight:4 estimator:2 nam:1 variation:9 autonomous:1 underwater:1 updated:1 target:8 user:4 exact:3 goodfellow:1 overlapped:1 trend:1 recognition:18 approximated:2 updating:1 finetuned:1 predicts:1 database:16 labeled:2 bottom:9 sunset:4 subtitle:1 initializing:1 capture:2 wang:2 cy:3 ensures:1 sun:1 trade:1 sol:1 highest:1 intuition:2 pd:11 trained:12 bipartite:1 multimodal:56 joint:21 easily:2 finetuning:2 icassp:2 represented:1 regularizer:2 derivation:1 train:9 stacked:1 describe:2 effective:1 kp:6 query:30 labeling:1 hyper:1 whose:2 supplementary:3 solve:1 cvpr:2 reconstruct:1 ability:1 jointly:1 nll:3 cai:1 propose:5 reconstruction:2 product:1 fr:1 frequent:1 realization:1 holistic:1 mixing:1 wjk:8 sutskever:1 convergence:1 p:1 darrell:1 sea:1 generating:6 incremental:1 converges:2 leave:1 object:4 mpeg:1 depending:1 recurrent:9 bosch:1 eq:1 implemented:1 predicted:3 c:1 come:1 larochelle:1 quantify:1 direction:1 annotated:1 cnns:1 stochastic:7 human:1 dbms:1 material:3 backprop:1 exchange:1 hx:4 cyj:2 extension:1 accompanying:1 hold:1 mm:1 viq:1 considered:1 ground:3 exp:2 dbm:10 predict:3 visualize:2 miniature:1 achieves:1 torralba:1 lenz:1 outperformed:1 label:2 individually:1 largest:1 organ:1 vice:1 successfully:1 city:2 hoffman:1 minimization:1 clearly:1 gaussian:1 aim:2 rather:1 avoid:1 derived:6 focus:3 improvement:6 modelling:1 likelihood:6 indicates:1 bernoulli:1 hk:9 contrast:1 pretrain:1 baseline:3 kim:2 inference:8 lowercase:1 membership:1 mdrnn:13 inaccurate:1 hidden:2 relation:2 pixel:1 issue:2 aforementioned:1 among:2 pascal:6 classification:4 html:1 scription:1 art:4 spatial:1 tzeng:1 softmax:1 marginal:4 emotion:2 construct:2 field:5 once:1 ng:1 sampling:2 hongkong:1 look:2 unsupervised:5 icml:3 peaked:1 future:1 minimized:1 report:5 others:1 mirza:1 lls:4 few:1 randomly:2 composed:4 simultaneously:1 divergence:6 decaf7:4 firework:1 ab:1 detection:2 mnih:1 evaluation:3 grasp:1 light:1 uppercase:1 chain:4 closer:1 partial:1 necessary:2 fox:1 tree:2 initialized:2 theoretical:4 mk:1 instance:3 modeling:1 rao:1 subset:4 epd:4 krizhevsky:1 too:2 ohm:1 reported:1 encoders:1 varies:1 corrupted:1 thibodeau:1 density:3 international:1 river:1 lee:4 off:1 dong:1 diy:1 yao:1 satisfied:1 huang:1 resort:1 expert:1 style:2 toy:3 li:2 potential:1 de:1 availability:1 satisfy:2 explicitly:2 mp:15 vi:5 performed:4 try:1 vehicle:1 reached:1 competitive:2 annotation:1 jia:1 minimize:3 php:1 convolutional:3 descriptor:4 variance:1 formance:1 lew:2 raw:1 handwritten:2 plastic:1 vincent:1 fern:1 ren:1 lu:1 published:1 flickr:14 lnll:3 sharing:2 aligns:1 guillaumin:4 nonetheless:1 energy:4 rbms:4 proof:1 mi:1 recovers:1 rbm:10 sampled:1 gain:4 proved:1 dataset:3 massachusetts:1 pilot:1 color:1 improves:2 dimensionality:1 formalize:1 back:1 finetune:3 higher:1 originally:1 supervised:2 asia:1 zisserman:1 improved:1 modal:1 formulation:1 mensink:1 furthermore:1 ergodicity:2 correlation:2 autoencoders:3 hand:2 until:1 night:4 google:1 defines:3 mode:2 logistic:2 quality:1 believe:1 usa:1 building:3 concept:2 true:1 contain:1 vasudevan:1 assigned:1 excluded:1 iteratively:1 neal:1 deal:1 conditionally:1 ll:2 during:3 daes:2 backpropagating:2 cosine:1 criterion:3 generalized:2 leftmost:1 theoretic:2 demonstrate:2 performs:1 reflection:1 reasoning:1 image:13 meaning:1 wise:2 novel:4 variational:3 inrialpes:1 common:2 overview:1 conditioning:3 million:1 association:5 extend:2 yosinski:1 accumulate:1 refer:1 significant:4 honglak:2 versa:1 gibbs:1 munoz:1 consistency:2 similarly:2 particle:2 minvi:22 access:1 robot:1 similarity:2 dominant:2 posterior:4 showed:2 retrieved:1 irrelevant:1 n00014:1 binary:6 success:1 onr:1 leach:1 cham:1 minimum:3 additional:2 canon:1 deng:1 converting:1 converge:1 maximize:2 semi:1 multiple:16 sound:1 unimodal:14 full:4 match:1 long:1 retrieval:5 lai:1 y:5 award:1 prediction:9 involving:1 regression:2 verbeek:2 heterogeneous:2 patient:1 expectation:1 vision:2 oliva:1 iteration:14 sometimes:1 histogram:1 fiers:1 pyramid:1 robotics:1 proposal:1 addition:3 whereas:1 separately:2 harbour:1 source:2 modality:59 crucial:1 envelope:1 rest:3 probably:1 mir:11 undirected:2 effectiveness:1 lego:2 feedforward:3 bengio:5 split:1 intermediate:1 variety:1 iterate:1 architecture:3 restrict:1 idea:3 haffner:1 multiclass:2 motivated:4 manjunath:1 reformulated:1 speech:2 pretraining:2 deep:35 dramatically:1 wenling:1 clear:1 informally:1 amount:2 dark:1 sohn:1 category:1 reduced:2 http:3 disjoint:1 per:1 promise:1 key:1 four:1 achieving:3 drawn:1 clarity:1 preprocessed:1 nikon:1 sum:2 run:2 letter:2 parameterized:1 layer:20 followed:1 courville:1 adapted:1 occur:1 pcd:5 fei:2 scene:2 hy:6 tag:10 layerwise:1 argument:1 min:2 nitish:1 structured:1 department:1 combination:1 across:5 slightly:1 modification:1 restricted:7 iccv:1 pipeline:1 equation:15 visualization:1 remains:2 discus:1 count:1 end:1 umich:1 photo:1 available:2 apply:2 observe:1 hierarchical:1 generic:1 alternative:2 voice:1 original:1 top:10 running:1 include:2 graphical:1 giving:1 especially:1 establish:2 approximating:1 icra:1 objective:36 question:2 already:1 strategy:2 gradient:12 distance:3 separate:1 mrbm:14 parametrized:1 nx:3 topic:4 argue:1 collected:2 reason:3 consumer:1 index:1 modeled:1 ratio:2 minimizing:2 equivalently:1 difficult:3 unfortunately:1 tagprop:2 negative:4 boltzmann:12 perform:1 observation:2 markov:2 benchmark:1 finite:3 descent:1 phow:1 hinton:6 provost:1 arbitrary:2 inferred:3 bk:4 dog:1 pair:2 kl:8 imagenet:3 learned:1 alternately:1 nip:7 able:1 bar:1 proceeds:1 wy:3 pattern:4 topdown:1 flower:1 suggested:1 unsuccessful:1 video:4 event:1 natural:2 predicting:1 wik:5 technology:1 naive:1 extract:1 autoencoder:3 auto:1 schmid:2 text:18 prior:1 l2:1 asymptotic:1 fully:2 rationale:2 generation:1 tures:1 skyscraper:1 validation:3 sufficient:6 bypass:1 cd:10 eccv:1 summary:1 supported:1 free:1 alain:2 bias:1 allow:1 side:1 wide:1 sparse:1 distributed:1 benefit:1 depth:2 dimension:1 transition:6 vocabulary:2 computes:2 doesn:1 concretely:1 made:2 commonly:1 replicated:1 social:1 transaction:2 reconstructed:1 approximate:2 compact:1 silhouette:1 ml:15 overfitting:1 robotic:1 uai:1 assumed:1 xi:6 discriminative:1 don:1 latent:4 compromised:1 iterative:2 khosla:1 why:2 table:6 learn:7 gsns:2 robust:2 transfer:1 ignoring:1 forest:1 kitty:1 ngiam:1 bottou:1 constructing:3 domain:1 protocol:1 lvi:5 aistats:1 arrow:1 whole:4 motivation:1 noise:1 complementary:1 cxi:2 slow:1 ny:4 precision:1 concatenating:1 toyota:1 learns:1 donahue:1 tang:1 theorem:5 rk:1 down:3 specific:4 showing:1 svm:3 normalizing:1 intractable:2 exists:1 mnist:2 socher:1 effectively:2 decaf:1 corr:1 texture:1 conditioned:3 easier:2 boston:1 entropy:1 cx:1 michigan:1 simply:1 explore:2 likely:1 visual:19 vinyals:1 doran:1 bo:1 pretrained:1 srivastava:2 springer:2 truth:3 rny:2 tieleman:1 extracted:1 conditional:12 goal:2 presentation:1 viewed:1 ann:1 lear:1 tempted:1 shared:6 typical:1 specifically:4 except:1 semantically:1 determined:1 reducing:1 denoising:2 shang:1 total:1 pas:2 arbor:1 experimental:2 select:2 support:1 people:2 kihyuk:1 collins:1 relevance:2 evaluate:4 audio:4 trainable:1 correlated:1 |
4,726 | 528 | Direction Selective Silicon Retina
that uses N uIl Inhibition
Ronald G. Benson and Tobi Delbriick
Computation and Neural Systems Program, 139-74
California Institute of Technology
Pasadena CA 91125
email: [email protected] and [email protected]
Abstract
Biological retinas extract spatial and temporal features in an attempt to
reduce the complexity of performing visual tasks. We have built and tested
a silicon retina which encodes several useful temporal features found in vertebrate retinas. The cells in our silicon retina are selective to direction,
highly sensitive to positive contrast changes around an ambient light level,
and tuned to a particular velocity. Inhibitory connections in the null direction perform the direction selectivity we desire. This silicon retina is
on a 4.6 x 6.8mm die and consists of a 47 x 41 array of photoreceptors.
1
INTRODUCTION
The ability to sense motion in the visual world is essential to survival in animals.
Visual motion processing is indispensable; it tells us about predators and prey, our
own motion and image stablization on the retina. Many algorithms for performing
early visual motion processing have been proposed [HK87] [Nak85]. A key salient
feature of motion is direction selectivity, ie the ability to detect the direction of
moving features. We have implemented Barlow and Levick's model, [BHL64], which
hypothesizes inhibition in the null direction to accomplish direction selectivity.
In contrast to our work, Boahen, [BA91], in these proceedings, describes a silicon
retina that is specialized to do spatial filtering of the image. Mahowald, [Mah91],
describes a silicon retina that has surround interactions and adapts over mulitiple
time scales. Her silicon retina is designed to act as an analog preprocessor and
756
Direction Selective Silicon Retina that uses Null Inhibition
Pixels inhibit to the left
Preferred
.Null ~
Preferred dPrection
~Photoreceptor
L~ DS cell
(a)
. Inhibition
(b)
Figure 1: Barlow and Levick model of direction selectivity (DS). (a) Shows
how two cells are connected in an inhibitory fashion and (b) a mosaic of such
cells.
so the gain of the output stage is rather low. In addition there is no rectification
into on- and off-pathways. This and earlier work on silicon early vision systems
have stressed spatial processing performed by biological retinas at the expense of
temporal processing.
The work we describe here and the work described by Delbriick, [DM9l], emphasizes
temporal processing. Temporal differentiation and separation of intensity changes
into on- and off-pathways are important computations performed by vertebrate
retinas. Additionally, specialized vertebrate retinas, [BHL64], have cells which are
sensitive to moving stimuli and respond maximally to a preferred direction; they
have almost zero response in the opposite or null direction. We have designed and
tested a silicon retina that models these direction selective velocity tuned cells.
These receptors excite cells which respond to positive contrast changes only and
are selective for a particular direction of stimuli. Our silicon retina may be useful
as a preprocessor for later visual processing and certainly as an enhancement for
the already existing spatial retinas. It is a striking demonstration of the perceptual
saliency of contrast changes and directed motion in the visual world.
2
INHIBITION IN THE NULL DIRECTION
Barlow and Levick, [BHL64]' described a mechanism for direction selectivity found
in the rabbit retina which postulates inhibitory connections to achieve the desired
direction selectivity. Their model is shown in Figure l(a). As a moving edge
passes over the photoreceptors from left to right, the left photoreceptor is excited
first, causing its direction selective (DS) cell to fire. The right photoreceptor fires
when the edge reaches it and since it has an inhibitory connection to the left DS
cell, the right photoreceptor retards further output from the left DS cell. If an edge
is moving in the opposite or null direction (right to left), the activity evoked in the
right photoreceptor completely inhibits the left DS cell from firing, thus creating a
direction selective cell.
757
758
Benson and Delbriick
Inhibition from right
Inhibition to left
Ir
Q
r
Photoreceptor
~
Preferred Direction
DS cell
Figure 2: Photoreceptor and direction selective (DS) cell. The output of the
high-gain, adaptive photoreceptor is fed capacitively to the input of the DS
cell. The output of the photoreceptor sends inhibition to the left. Inhibition
from the right photoreceptors connect to the input of the DS cell.
In the above explanation with the edge moving in the preferred direction (left to
right), as the edge moves faster, the inhibition from leading photoreceptors truncates
the output of the DS cell ever sooner. In fact, it is this inhibitory connection which
leads to velocity tuning in the preferred direction.
By tiling these cells as shown in Figure l(b), it is possible to obtain an array of
directionally tuned cells. This is the architecture we used in our chip. Direction
selectivity is inherent in the connections of the mosaic, ie the hardwiring of the
inhibitory connections leads to directionally tuned cells.
3
PIXEL OPERATION
A pixel consists of a photoreceptor, a direction selective (DS) cell and inhibition to
and from other pixels as shown in Figure 2. The photoreceptor has high-gain and
is adaptive [Mah91, DM91]. The output from this receptor, Vp , is coupled into the
DS cell which acts as a rectifying gain element, [MS91], that is only sensitive to
positive-going transitions due to increases in light intensity at the receptor input.
Additionally, the output from the photoreceptor is capacitively coupled to the inhibitory synapses which send their inhibition to the left and are coupled into the
DS cell of the neighboring cells.
A more detailed analysis of the DS cell yields several insights into this cell's functionality. A step increase of 6. V at Vp , caused by a step increase in light intensity
incident upon the phototransistor, results in a charge injection of Cc 6. V at Vi. This
charge is leaked away by QT at a rate IT, set by voltage VT. Hence, to first order,
the output pulse width T is simply
T
= Cc 6.V.
IT
There is also a threshold minimum step input size that will result in enough change
Direction Selective Silicon Retina that uses Null Inhibition
1.6
-..
>
1.2
Output
'-'
~
<Il
0.8
= 0.4
c.
<Il
0
~
~
0.0
Input intensity
0 40 80 120 160 200
Time (msec)
Figure 3: Pixel response to intensity step. Bottom trace is intensity; top trace
is pixel output.
in Vi to pull Vout all the way to ground. This threshold is set by Cc and the gain of
the photoreceptor.
When the input to the rectifying gain element is not a step, but instead a steady
increase in voltage, the current lin flowing into node Vi is
= CcVp.
When this current exceeds IT there is a net increase in the voltage Vi, and the
output Vout will quickly go low. The condition lin = IT defines the threshold
limit for stimuli detection, i.e. input stimuli resu~ting in an lin < IT are not
perceptible to the pixel. For a changing intensity I, the adaptive photoreceptor
stage outputs a voltage Vp proportional to j / I, where I is the input light intensity.
This photoreceptor behavior means that the pixel threshold will occur at whatever
j / I causes Cc Vp to exceed the constant current I r .
lin
The inhibitory synapses (shown as Inhibition from right in Figure 2) provide additionalleakage from Vi resulting in a shortened response width from the DS cell.
This analysis suggests that a characterization of the pixel should investigate both
the response amplitude, measured as pulse width versus input intensity step size,
and the response threshold, measured with temporal intensity contrast. In the next
section we show such measurements.
4
CHARACTERIZATION OF THE PIXEL
We have tested both an isolated pixel and a complete 2-dimensional retina of 47 x 41
pixels. Both circuits were fabricated in a 2J.tm p-well CMOS double poly process
available through the MOSIS facility. The retina is scanned out onto a monitor using
a completely integrated on-chip scanner[MD91]. The only external components are
a video amplifier and a crystal.
We show a typical response of the isolated pixel to an input step of intensity in
Figure 3. In response to the input step increase of intensity, the pixel output goes
low and saturates for a time set by the bias Vr in Figure 2. Eventually the pixel
recovers and the output returns to its quiescent level. In response to the step
decrease of intensity there is almost no response as seen in Figure 3.
759
760
Benson and Delbriick
~~
16O
UIII
rIJ
!,120
..c=
....be
80
=
III
/
- 40
III
rIJ
::l
A..
1.8
Step Contrast
(a)
2.2
Temporal Frequency (Hz)
(b)
Figure 4: (a) Pulse width of response as function of input contrast step size.
The abscissa is measured in units of ratio-intensity, i.e., a value of 1 means
no intensity step, a value of 1.1 means a step from a normalized intensity of 1
to a normalized intensity of 1.1, and so forth. The different curves show the
response at different absolute light levels; the number in the figure legend is
the log of the absolute intensity. (b) Receptor threshold measurements. At
each temporal frequency, we determined the minimum necessary amplitude of
triangular intensity variations to make the pixel respond. The different curves
were taken at different background intensity levels, shown to the left of each
curve. For example, the bottom curve was taken at a background level of 1
unit of intensity; at 8 Hz, the threshold occurred at a variation of 0.2 units of
intensity.
The output from the pixel is essentially quantized in amplitude, but the resulting
pulse has a finite duration related to the input intensity step. The analysis in
Section 3 showed that the output pulse width, T, should be linear in the input
intensity contrast step. In Figure 4{ a), we show the measured pulse-width as a
function of input contrast step. To show the adaptive nature of the receptor, we
did this same measurement at several different absolute intensity levels.
Our silicon retina sees some features of a moving image and not others. Detection
of a moving feature depends on its contrast and velocity. To characterize this
behavior, we measured a receptor's thresholds for intensity variations, as a function
of temporal frequency.
These measurements are shown in Figure 4(b); the curves define "zones of visibility"; if stimuli lie below a curve, they are visible, if they fall above a curve they
are not. (The different curves are for different absolute intensity levels.) For low
temporal frequencies stimuli are visible only if they are high contrast; at higher
temporal frequencies, but still below the photoreceptor cutoff frequency, lower contrast stimuli are visible. Simply put, if the input image has low contrast and is
slowly moving, it is not seen. Only high contrast or quickly moving features are
salient stimuli. More precisely, for temporal frequencies below the photoreceptor
cutoff frequency, the threshold occurs at a constant value of the temporal intensity
contrast j / I.
Direction Selective Silicon Retina that uses Null Inhibition
761
Preferred
L
R
Inhib
Preferred
fN uu -
Photoreceptors
"'----- DS
Excitatio
L
-
R
- - - - Inhib
' - - - - - DS
(a)
0.1 sec
(b)
Figure 5: (a) shows the basic connectivity of the tested cell. (b) top trace is
the response due to an edge moving in the preferred direction (left to right).
(b) bottom trace is the response due to an edged moving in the null direction
(right to left).
5
NULL DIRECTION INHIBITION PROPERTIES
We performed a series of tests to characterize the inhibition for various orientations
and velocities. The data in Figure 5(b) shows the outputs of two photo receptors,
the inhibitory signal and the output of a DS cell. The top panel in Figure 5(b) shows
the outputs in the preferred direction and the bottom panel shows them in the null
direction. Notice that the out pu t of the left photoreceptor (L in Figure 5 (b) top
panel) precedes the right (R). The output of the DS cell is quite pronounced, but is
truncated by the inhibition from the right photoreceptor. On the other hand, the
bottom panel shows that the output of the DS cell is almost completely truncated
by the inhibitory input.
A DS cell receives most inhibition when the stimulus is travelling exactly in the null
direction. As seen in Figure 6(a) as the angle of stimulus is rotated, the maximum
response from the DS cell is obtained when the stimulus is moving in the preferred
direction (directly opposite to the null direction). As the bar is rotated toward the
null direction, the response of the cell is reduced due to the increasing amount of
inhibition received from the neighboring photo receptors.
If a bar is moving in the preferred direction with varying velocity, there is a velocity,
Vmaz , for which the DS cell responds maximally as shown in Figure 6(b). As the
bar is moved faster than Vmaz , inhibition arrives at the cell sooner, thus truncating
the response. As the cell is moved slower than Vmaz, less input is provided to the
DS cell as described in Section 3. In the null direction (negative in Figure 6(b?) the
cell does not respond, as expected, until the bar is travelling fast enough to beat
the inhibition's onset (recall delay from Figure 5).
In Figure 7 we show the response of the entire silicon retina to a rotating fan. When
the fan blades are moving to the left the retina does not respond, but when moving
to the right, note the large response. Note the largest response when the blades are
moving exactly in the preferred direction.
762
Benson and Delbruck
160
-;;-120
8
~
Q)
rn
? 80
0..
~
~ 40
-0.8 -0.4 0.0
0.4
0.8
Velocity (arbitrary units)
(a)
(b)
Figure 6: (a) polar plot which shows the pixels are directionally tuned.
(b) shows velocity tuning of the DS cell (positive velocities are in the preferred direction).
(a)
(b)
Figure 7: (a) Rotating fan used as stimulus to the retina. (b) Output of the
retina.
Direction Selective Silicon Retina that uses Null Inhibition
6
CONCLUSION
We have designed and tested a silicon retina that detects temporal changes in an
image. The salient image features are sufficiently high contrast stimuli, relatively
fast increase in intensity (measured with respect to the recent past history of the
intensity), direction and velocity of moving stimuli. These saliency measures result
in a large compression of information, which will be useful in later processing stages.
Acknowledgments
Our thanks to Carver Mead and John Hopfield for their guidance and encouragement, to the Office of Naval Research for their support under grant NAV N0001489-J-1675, and, of course, to the MOSIS fabrication service.
References
[BA91]
K. Boahen and A. Andreou. A contrast sensitive silicon retina with reciprocal synapses. In S. Hanson J. Moody and R. Lippmann, editors,
Advances in Neural Information Processing Systems, Volume 4. Morgan
Kaufmann, Palo Alto, CA, 1991.
[BHL64] H.B. Barlow, M.R. Hill, and W.R. Levick. Retinal ganglion cells responding selectively to direction and speed of image motion in the rabbit. J.
Physiol., 173:377-407, 1964.
[DM91] T. Delbriick and Carver Mead. Silicon adaptive photoreceptor array that
computes temporal intensity derivatives. In Proc. SPIE 1541, volume
1541-12, pages 92-99, San Diego, CA, July 1991. Infrared Sensors: Detectors, Electronics, and Signal Processing.
E. Hildreth and C. Koch. The analysis of visual motion: From computational theory to neuronal mechanisms. Annual Review in Neuroscience,
10:477-533, 1987.
[Mah91] M.A. Mahowald. Silicon retina with adaptive photoreceptor. In SPIE
Technical Symposia on Optical Engineering and Photonics in Aerospace
Sensing, Orlando, FL, April 1991. Visual Information Processing: From
Neurons to Chips.
[MD91] C.A. Mead and T. Delbriick. Scanners for use in visualizing analog VLSI
circuitry. Analog Integrated Circuits and Signal Processing, 1:93-106, 1991.
[MS91] C.A. Mead and R. Sarpeshkar. An axon circuit. Internal Memo, Physics
of Computation Laboratory, Caltech, 1991.
[HK87]
[Nak85] K. Nakayama. Biological image motion processing: A review.
Research, 25(5):625-660, 1985.
Vision
763
| 528 |@word compression:1 pulse:6 excited:1 blade:2 electronics:1 series:1 tuned:5 past:1 existing:1 current:3 john:1 fn:1 physiol:1 ronald:1 visible:3 visibility:1 designed:3 plot:1 reciprocal:1 characterization:2 quantized:1 node:1 symposium:1 consists:2 pathway:2 expected:1 behavior:2 abscissa:1 retard:1 detects:1 vertebrate:3 increasing:1 provided:1 circuit:3 panel:4 alto:1 null:17 nav:1 differentiation:1 fabricated:1 temporal:15 act:2 charge:2 exactly:2 whatever:1 unit:4 grant:1 positive:4 service:1 engineering:1 limit:1 receptor:8 shortened:1 mead:4 firing:1 evoked:1 suggests:1 directed:1 acknowledgment:1 onto:1 put:1 phototransistor:1 send:1 go:2 duration:1 rabbit:2 truncating:1 insight:1 array:3 pull:1 variation:3 diego:1 us:5 mosaic:2 velocity:11 element:2 infrared:1 bottom:5 rij:2 connected:1 decrease:1 inhibit:1 boahen:2 complexity:1 upon:1 completely:3 hopfield:1 chip:3 various:1 sarpeshkar:1 fast:2 describe:1 precedes:1 tell:1 quite:1 triangular:1 ability:2 directionally:3 net:1 interaction:1 causing:1 neighboring:2 achieve:1 adapts:1 forth:1 moved:2 pronounced:1 enhancement:1 double:1 cmos:1 rotated:2 measured:6 qt:1 received:1 implemented:1 uu:1 direction:44 functionality:1 orlando:1 biological:3 mm:1 scanner:2 around:1 sufficiently:1 ground:1 koch:1 uiii:1 uil:1 circuitry:1 early:2 polar:1 proc:1 palo:1 sensitive:4 largest:1 sensor:1 rather:1 varying:1 voltage:4 office:1 naval:1 contrast:17 sense:1 detect:1 integrated:2 entire:1 pasadena:1 her:1 vlsi:1 selective:12 going:1 pixel:18 orientation:1 animal:1 spatial:4 others:1 stimulus:14 inherent:1 retina:31 cns:1 fire:2 attempt:1 amplifier:1 detection:2 highly:1 investigate:1 certainly:1 photonics:1 arrives:1 light:5 ambient:1 edge:6 necessary:1 carver:2 capacitively:2 sooner:2 rotating:2 desired:1 isolated:2 guidance:1 earlier:1 delbruck:1 mahowald:2 delay:1 fabrication:1 characterize:2 connect:1 accomplish:1 thanks:1 ie:2 off:2 physic:1 quickly:2 excitatio:1 moody:1 connectivity:1 postulate:1 slowly:1 external:1 creating:1 derivative:1 leading:1 return:1 retinal:1 sec:1 hypothesizes:1 caused:1 vi:5 depends:1 onset:1 performed:3 later:2 predator:1 rectifying:2 il:2 ir:1 kaufmann:1 yield:1 saliency:2 vp:4 vout:2 emphasizes:1 cc:4 history:1 detector:1 synapsis:3 reach:1 email:1 frequency:8 spie:2 recovers:1 gain:6 recall:1 amplitude:3 levick:4 higher:1 response:19 maximally:2 flowing:1 april:1 stage:3 until:1 d:26 hand:1 receives:1 defines:1 hildreth:1 normalized:2 barlow:4 facility:1 hence:1 laboratory:1 leaked:1 visualizing:1 width:6 die:1 steady:1 hill:1 crystal:1 complete:1 motion:9 image:8 specialized:2 inhib:2 volume:2 analog:3 occurred:1 silicon:20 measurement:4 surround:1 edged:1 encouragement:1 tuning:2 moving:17 inhibition:23 pu:1 own:1 showed:1 recent:1 selectivity:7 indispensable:1 vt:1 caltech:3 seen:3 minimum:2 morgan:1 july:1 signal:3 exceeds:1 technical:1 faster:2 lin:4 basic:1 vision:2 essentially:1 cell:41 addition:1 background:2 sends:1 pass:1 hz:2 legend:1 exceed:1 iii:2 enough:2 architecture:1 opposite:3 reduce:1 tm:1 cause:1 useful:3 detailed:1 amount:1 reduced:1 inhibitory:10 notice:1 neuroscience:1 key:1 salient:3 threshold:9 monitor:1 changing:1 cutoff:2 prey:1 mosis:2 angle:1 respond:5 striking:1 almost:3 separation:1 fl:1 fan:3 annual:1 activity:1 occur:1 scanned:1 precisely:1 encodes:1 speed:1 performing:2 injection:1 optical:1 inhibits:1 relatively:1 describes:2 perceptible:1 benson:5 taken:2 rectification:1 eventually:1 mechanism:2 fed:1 photo:2 travelling:2 tiling:1 available:1 operation:1 away:1 slower:1 top:4 responding:1 ting:1 move:1 already:1 occurs:1 responds:1 toward:1 ratio:1 demonstration:1 truncates:1 expense:1 trace:4 negative:1 memo:1 perform:1 neuron:1 finite:1 truncated:2 beat:1 saturates:1 ever:1 rn:1 delbriick:6 arbitrary:1 intensity:31 connection:6 hanson:1 andreou:1 aerospace:1 california:1 bar:4 below:3 program:1 built:1 explanation:1 video:1 technology:1 extract:1 coupled:3 review:2 filtering:1 resu:1 proportional:1 versus:1 incident:1 editor:1 course:1 bias:1 institute:1 tobi:1 fall:1 absolute:4 curve:8 world:2 transition:1 computes:1 adaptive:6 san:1 lippmann:1 preferred:14 photoreceptors:5 excite:1 quiescent:1 additionally:2 nature:1 ca:3 nakayama:1 poly:1 did:1 neuronal:1 fashion:1 vr:1 axon:1 msec:1 lie:1 perceptual:1 preprocessor:2 sensing:1 survival:1 essential:1 simply:2 ganglion:1 visual:8 desire:1 change:6 typical:1 determined:1 photoreceptor:21 zone:1 selectively:1 internal:1 support:1 stressed:1 tested:5 |
4,727 | 5,280 | Restricted Boltzmann machines modeling human
choice
Makoto Otsuka
IBM Research - Tokyo
[email protected]
Takayuki Osogami
IBM Research - Tokyo
[email protected]
Abstract
We extend the multinomial logit model to represent some of the empirical phenomena that are frequently observed in the choices made by humans. These phenomena include the similarity effect, the attraction effect, and the compromise
effect. We formally quantify the strength of these phenomena that can be represented by our choice model, which illuminates the flexibility of our choice model.
We then show that our choice model can be represented as a restricted Boltzmann
machine and that its parameters can be learned effectively from data. Our numerical experiments with real data of human choices suggest that we can train our
choice model in such a way that it represents the typical phenomena of choice.
1
Introduction
Choice is a fundamental behavior of humans and has been studied extensively in Artificial Intelligence and related areas. The prior work suggests that the choices made by humans can significantly
depend on available alternatives, or the choice set, in rather complex but systematic ways [13]. The
empirical phenomena that result from such dependency on the choice set include the similarity effect, the attraction effect, and the compromise effect. Informally, the similarity effect refers to the
phenomenon that a new product, S, reduces the share of a similar product, A, more than a dissimilar
product, B (see Figure 1 (a)). With the attraction effect, a new dominated product, D, increases the
share of the dominant product, A (see Figure 1 (b)). With the compromise effect, a product, C, has
a relatively larger share when two extreme products, A and B, are in the market than when only
one of A and B is in the market (see Figure 1 (c)). We call these three empirical phenomena as the
typical choice phenomena.
However, the standard choice model of the multinomial logit model (MLM) and its variants cannot
represent at least one of the typical choice phenomena [13]. More descriptive models have been
proposed to represent the typical choice phenomena in some representative cases [14, 19]. However,
it is unclear when and to what degree the typical choice phenomena can be represented. Also, no
algorithms have been proposed for training these descriptive models from data.
S
A
A
A
D
C
B
(a) Similarity
B
B
(b) Attraction
(c) Compromise
Figure 1: Choice sets that cause typical choice phenomena.
1
We extend the MLM to represent the typical choice phenomena, which is our first contribution.
We show that our choice model can be represented as a restricted Boltzmann machine (RBM). Our
choice model is thus called the RBM choice model. An advantage of this representation as an RBM
is that training algorithms for RBMs are readily available. See Section 2.
We then formally define the measure of the strength for each typical choice phenomenon and quantify the strength of each typical choice phenomenon that the RBM choice model can represent. Our
analysis not only gives a guarantee on the flexibility of the RBM choice model but also illuminates
why the RBM choice model can represent the typical choice phenomena. These definitions and
analysis constitute our second contribution and are presented in Section 3.
Our experiments suggest that we can train the RBM choice model in such a way that it represents
the typical choice phenomena. We show that the trained RBM choice model can then adequately
predict real human choice on the means of transportation [2]. These experimental results constitute
our third contribution and are presented in Section 4.
2
Choice model with restricted Boltzmann machine
We extend the MLM to represent the typical choice phenomena. Let I be the set of items. For A ?
X ? I, we study the probability that an item, A, is selected from a choice set, X . This probability
is called the choice probability. The model of choice, equipped with the choice probability, is called
a choice model. We use A, B, C, D, S, or X to denote an item and X , Y, or a set such as {A, B} to
denote a choice set.
For the MLM, the choice probability of A from X can be represented by
?(A|X )
p(A|X ) = P
,
X?X ?(X|X )
(1)
where we refer to ?(X|X ) as the choice rate of X from X . The choice rate of the MLM is given by
?MLM (X|X ) = exp(bX ),
(2)
where bX can be interpreted as the attractiveness of X. One could define bX through uX , the
vector of the utilities of the attributes for X, and ?, the vector of the weight on each attribute (i.e.,
bX ? ??uX ). Observe that ?MLM (X|X ) is independent of X as long as X ? X . This independence
causes the incapability of the MLM in representing the typical choice phenomena.
We extend the choice rate of (2) but keep the choice probability in the form of (1). Specifically, we
consider the following choice rate:
Y
k
?(X|X ) ? exp(bX )
1 + exp TXk + UX
,
(3)
k?K
where we define
TXk ?
X
TYk .
(4)
Y ?X
k
Our choice model has parameters, bX , TXk , UX
for X ? X , k ? K, that take values in (??, ?).
Equation (3) modifies exp(bX ) by multiplying factors. Each factor is associated with an index, k,
k
and has parameters, TXk and UX
, that depend on k. The set of these indices is denoted by K.
We now show that our choice model can be represented as a restricted Boltzmann machine (RBM).
This means that we can use existing algorithms for RBMs to learn the parameters of the RBM choice
model (see Appendix A.1).
An RBM consists of a layer of visible units, i ? V, and a layer of hidden units, k ? H. A visible unit,
i, and a hidden unit, k, are connected with weight, Wik . The units within each layer are disconnected
from each other. Each unit is associated with a bias. The bias of a visible unit, i, is denoted by bvis
i .
The bias of a hidden unit, k, is denoted by bhid
k . A visible unit, i, is associated with a binary variable,
zi , and a hidden unit, k, is associated with a binary variable, hk , which takes a value in {0, 1}.
For a given configuration of binary variables, the energy of the RBM is defined as
XX
hid
E? (z, h) ? ?
zi Wik hk + bvis
i zi + bk hk ,
i?V k?H
2
(5)
...
Hidden
k
...
UAk
TXk
Choice set
...
...
X
...
A
Selected item
...
bA
Figure 2: RBM choice model
where ? ? {W, bvis , bhid } denotes the parameters of the RBM. The probability of realizing a particular configuration of (z, h) is given by
exp(?E? (z, h))
P? (z, h) ? P P
.
(6)
0
0
z0
h0 exp(?E? (z , h ))
P
P
The summation with respect to a binary vector (i.e., z0 or h0 ) denotes the summation over all of
the possible binary vectors of a given length. The length of z 0 is |V|, and the length of h0 is |H|.
The RBM choice model can be represented as an RBM having the structure in Figure 2. Here, the
layer of visible units is split into two parts: one for the choice set and the other for the selected item.
The corresponding binary vector is denoted by z = (v, w). Here, v is a binary vector associated
with the part for the choice set. Specifically, v has length |I|, and vX = 1 denotes that X is in the
k
choice set. Analogously, w has length |I|, and wA = 1 denotes that A is selected. We use TX
to
denote the weight between a hidden unit, k, and a visible unit, X, for the choice set. We use UAk to
denote the weight between a hidden unit, k, and a visible unit, A, for the selected item. The bias is
zero for all of the hidden units and for all of the visible units for the choice set. The bias for a visible
unit, A, for the selected item is denoted by bA . Finally, let H = K.
The choice rate (3) of the RBM choice model can then be represented by
X
?(A|X ) =
exp ?E? v X , wA , h ,
(7)
h
X
where we define the binary vectors, v , wA , such that viX = 1 iff i ? X and wjA = 1 iff j = A.
Observe that the right-hand side of (7) is
!
X
X
X X
X
X
A
k
k
exp(?E? ((v , w ), h)) =
exp
TX hk +
UA hk + bA
(8)
h
X?X
h
=
exp(bA )
h
=
exp(bA )
k
XY
Y
exp
k
TXk
+ UAk hk
(9)
k
X
exp
TXk + UAk hk ,
(10)
k hk ?{0,1}
which is equivalent to (3).
The RBM choice model assumes that one item from a choice set is selected. In the context of the
RBM, this means that wA = 1 for only one A ? X ? I. Using (6), our choice probability (1) can
be represented by
P
P? ((v X , wA ), h)
hP
.
(11)
p(A|X ) = P
X
X
X?X
h P? ((v , w ), h)
This is the conditional probability of realizing the configuration, (v X , wA ), given that the realized
configuration is either of the (v X , wX ) for X ? X . See Appendix A.2for an extension of the RBM
choice model.
3
Flexibility of the RBM choice model
In this section, we formally study the flexibility of the RBM choice model. Recall that ?(X|X ) in
(3) is modified from ?MLM (X|X ) in (2) by a factor,
k
1 + exp TXk + UX
,
(12)
3
for each k in K, so that ?(X|X ) can depend on X through TXk . We will see how this modification
allows the RBM choice model to represent each of the typical choice phenomena.
The similarity effect refers to the following phenomenon [14]:
p(A|{A, B}) > p(B|{A, B})
p(A|{A, B, S}) < p(B|{A, B, S}).
and
(13)
Motivated by (13), we define the strength of the similarity effect as follows:
Definition 1. For A, B ? X , the strength of the similarity effect of S on A relative to B with X is
defined as follows:
(sim)
?A,B,S,X
?
p(A|X ) p(B|X ? {S})
.
p(B|X ) p(A|X ? {S})
(14)
(sim)
When ?A,B,S,X = 1, adding S into X does not change the ratio between p(A|X ) and p(B|X ).
(sim)
Namely, there is no similarity effect. When ?A,B,S,X > 1, we can increase
(sim)
of ?A,B,S,X
(sim)
?A,B,S,X <
p(B|X )
p(A|X )
by a factor
by the addition of S into X . This corresponds to the similarity effect of (13). When
1, this ratio decreases by an analogous factor. We will study the strength of this (rather
general) similarity effect without the restriction that S is ?similar? to A (see Figure 1 (a)).
Because p(X|X ) has a common denominator for X = A and X = B, we have
(sim)
?A,B,S,X
=
?(A|X ) ?(B|X ? {S})
.
?(B|X ) ?(A|X ? {S})
(15)
The MLM cannot represent the similarity effect, because the ?MLM (X|X ) in (2) is independent of
X . For any choice sets, X and Y, we must have
?MLM (A|X )
?MLM (B|X )
=
?MLM (A|Y)
.
?MLM (B|Y)
(16)
The equality (16) is known as the independence from irrelevant alternatives (IIA).
The RBM choice model can represent an arbitrary strength of the similarity effect. Specifically, by
? into K of (3), we can set ?(A|X ?{S}) at an arbitrary value without affecting
adding an element, k,
?(A|X )
the value of ?(B|Y), ?B 6= A, for any Y. We prove the following theorem in Appendix C:
Theorem 1. Consider an RBM choice model where the choice rate of X from X is given by (2). Let
?
?(X|X
) be the corresponding choice rate after adding k? into K. Namely,
?
?
k
?
.
(17)
?(X|X
) = ?(X|X ) 1 + exp TXk + UX
Consider an item A ? X and an item S 6? X . For any c ? (0, ?) and ? > 0, we can then choose
?
?
T?k and U?k such that
?(B|Y)
?
?(A|X
? {S})
?
c=
;
?>
? 1 , ?Y, B s.t. B 6= A.
(18)
?
?(B|Y)
?(A|X
)
By (15) and Theorem 1, the strength of the similarity effect after adding k? into K is
(sim)
??A,B,S,X =
?
?
?(A|X
)
?(B|X
? {S})
1 ?(B|X ? {S})
.
?
?
?
c
?(B|X )
?(A|X ? {S})
?(B|X )
(19)
? indeed allows
Because c can take an arbitrary value in (0, ?), the additional factor, (12) with k = k,
(sim)
??
to take any positive value without affecting the value of ?(B|Y), ?B 6= A, for any Y. The
A,B,S,X
first part of (18) guarantees that this additional factor does not change p(X|Y) for any X if A ?
/ Y.
Note that what we have shown is not limited to the similarity effect of (13). The RBM choice model
can represent an arbitrary phenomenon where the choice set affects the ratio of the choice rate.
4
According to [14], the attraction effect is represented by
p(A|{A, B}) < p(A|{A, B, D}).
(20)
MLM
The MLM cannot represent
the
attraction
effect,
because
the
?
(X|Y)
in
(2)
is
independent
P
P
of Y, and we must have X?X ?MLM (X|X ) ? X?Y ?MLM (X|Y) for X ? Y, which in turn
implies the regularity principle: p(X|X ) ? p(X|Y) for X ? Y.
Motivated by (20), we define the strength of the attraction effect as the magnitude of the change in
the choice probability of an item when another item is added into the choice set. Formally,
Definition 2. For A ? X , the strength of the attraction effect of D on A with X is defined as follows:
p(A|X ? {D})
(att)
?A,D,X ?
.
(21)
p(A|X )
(att)
When there is no attraction effect, adding D into X can only decrease p(A|X ); hence, ?A,D,X ? 1.
(att)
The standard definition of the attraction effect (20) implies ?A,D,X > 1. We study the strength of
this attraction effect without the restriction that A ?dominates? D (see Figure 1 (b)).
We prove the following theorem in Appendix C:
Theorem 2. Consider the two RBM choice models in Theorem 1. The first RBM choice model has
the choice rate given by (3), and the second RBM choice model has the choice rate given by (17).
Let p(?|?) denote the choice probability for the first RBM choice model and p?(?|?) denote the choice
probability for the second RBM choice model. Consider an item A ? X and an item D 6? X . For
?
?
any r ? (p(A|X ? {D}), 1/p(A|X )) and ? > 0, we can choose T?k , U?k such that
?(B|Y)
p?(A|X ? {D})
?
r=
;
?>
? 1 , ?Y, B s.t. B 6= A.
(22)
?(B|Y)
p?(A|X )
We expect that the range, (p(A|X ? {D}), 1/p(A|X )), of r in the theorem covers the attraction
effect in practice. Also, this range is the widest possible in the following sense. The factor (12) can
only increase ?(X|Y) for any X, Y. The form of (1) then implies that, to decrease p(A|Y), we must
increase ?(X|Y) for X 6= A. However, increasing ?(X|Y) for X 6= A is not allowed due to the
? can only increase
second part of (22) with ? ? 0. Namely, the additional factor, (12) with k = k,
p(A|Y) for any Y under the condition of the second part of (22). The lower limit, p(A|X ? {D}),
is achieved when p?(A|X ) ? 1, while keeping p?(A|X ? {D}) ? p(A|X ? {D}). The upper limit,
1/p(A|X ), is achieved when p?(A|X ? {D}) ? 1, while keeping p?(A|X ) ? p(A|X ).
According to [18], the compromise effect is formally represented by
p(C|{A, B, C})
p(C|{A, B, C})
X
X
> p(C|{A, C}) and
> p(C|{B, C}). (23)
p(X|{A, B, C})
p(X|{A, B, C})
X?{A,C}
X?{B,C}
The MLM cannot represent the compromise effect, because the ?MLM (X|Y) in (2) is independent
of Y, which in turn makes the inequalities in (23) equalities.
Motivated by (23), we define the strength of the compromise effect as the magnitude of the change
in the conditional probability of selecting an item, C, given that either C or another item, A, is
selected when yet another item, B, is added into the choice set. More precisely, we also exchange
the roles of A and B, and study the minimum magnitude of those changes:
Definition 3. For a choice set, X , and items, A, B, C, such that A, B, C ? X , let
qAC (C|X )
?A,B,C,X ?
,
(24)
qAC (C|X \ {B})
where, for Y such that A, C ? Y, we define
p(C|Y)
qAC (C|Y) ? P
.
(25)
X?{A,C} p(X|Y)
The strength of the compromise effect of A and B on C with X is then defined as
(com)
?A,B,C,X
?
min {?A,B,C,X , ?B,A,C,X } .
5
(26)
Here, we do not have the restriction that C is a ?compromise? between A and B (see Figure 1 (c)).
In Appendix C:we prove the following theorem:
Theorem 3. Consider a choice set, X , and three items, A, B, C ? X . Consider the two RBM choice
(com)
models in Theorem 2. Let ??A,B,C,X be defined analogously to (26) but with p?(?|?). Let
q ? max {qAC (C|X \ {B}), qBC (C|X \ {A})}
q ? min {qAC (C|X ), qBC (C|X )} .
Then, for any r ? (q, 1/q) and ? > 0, we can choose T?k , U?k such that
?(X|Y)
?
(com)
? 1 , ?Y, X s.t. X 6= C.
r = ??A,B,C,X ;
?>
?(X|Y)
(27)
(28)
(29)
We expect that the range of r in the theorem covers the compromising effect in practice. Also,
this range is best possible in the sense analogous to what we have discussed with the range in
? can only increase p(C|Y) for any Y
Theorem 2. Because the additional factor, (12) with k = k,
under the condition of the second part of (29), it can only increase qXC (C|Y) for X ? {A, B}. The
lower limit, q, is achieved when qXC (C|X \ {X}) ? 1, while keeping qXC (C|X ) approximately
unchanged, for X ? {A, B}. The upper limit, 1/q, is achieved when qXC (C|X ) ? 1, while
keeping qXC (C|X \ {X}) approximately unchanged, for X ? {A, B}.
4
Numerical experiments
We now validate the effectiveness of the RBM choice model in predicting the choices made by
humans. Here we use the dataset from [2], which is based on the survey conducted in Switzerland,
where people are asked to choose a means of transportation from given options. A subset of the
dataset is used to train the RBM choice model, which is then used to predict the choice in the
remaining dataset. In Appendix B.2,we also conduct an experiment with artificial dataset and show
that the RBM choice model can indeed be trained to represent each of the typical choice phenomena.
This flexibility in the representation is the basis of the predictive accuracy of the RBM choice model
to be presented in this section. All of our experiments are run on a single core of a Windows PC
with main memory of 8 GB and Core i5 CPU of 2.6 GHz.
The dataset [2] consists of 10,728 choices that 1,192 people have made from a varying choice set.
For those who own a car, the choice set has three items: a train, a maglev, and a car. For those who
do not own a car, the choice set consists of a train and a maglev. The train can operate at the interval
of 30, 60, or 120 minutes. The maglev can operate at the interval of 10, 20, or 30 minutes. The
trains (or maglevs) with different intervals are considered to be distinct items in our experiment.
Figure 3 (a) shows the empirical choice probability for each choice set. Each choice set consists of
a train with a particular interval (blue, shaded) and a maglev with a particular interval (red, mesh)
possibly with a car (yellow, circles). The interval of the maglev varies as is indicated at the bottom
of the figure. The interval of the train is indicated at the left side of the figure. For each combination
of the intervals of the train and the maglev, there are two choice sets, with or without a car.
We evaluate the accuracy of the RBM choice model in predicting the choice probability for an
arbitrary choice set, when the RBM choice model is trained with the data of the choice for the
remaining 17 choice sets (i.e., we have 18 test cases). We train the RBM choice model (or the
MLM) by the use of discriminative training with stochastic gradient descent using the mini-batch
of size 50 and the learning rate of ? = 0.1 (see Appendix A.1).Each run of the evaluation uses the
entire training dataset 50 times for training, and the evaluation is repeated five times by varying the
initial values of the parameters. The elements
independently with samples
p
p of T and U are initialized
from the uniform distribution on [?10/ max(|I|, |K|), ?10/ max(|I|, |K|)], where |I| = 7 is
the number of items under consideration, and |K| is the number of hidden nodes. The elements of b
are initialized with samples from the uniform distribution on [?1, 1].
Figure 3 (b) shows the Kullback-Leibler (KL) divergence between the predicted distribution of the
choice and the corresponding true distribution. The dots connected with a solid line show the the
6
Train120
0.20
0.15
0.10
0.05
0.00
Maglev10
Maglev20
Maglev30
Train30
Train60
Train30
Train60
Train120
Maglev10
Maglev20
Maglev30
Car
Maglev10
Maglev20
0
1
2
4
8
Number of hidden units
16
(b) Error
Train120
Train120
Train60
Train30
(a) Dataset
1.0
0.8
0.6
0.4
0.2
0.0
1.0
0.8
0.6
0.4
0.2
0.0
1.0
0.8
0.6
0.4
0.2
0.0
Training
Test
0.25
Train30
Train60
Train120
Maglev10
Maglev20
Maglev30
Car
Average KL divergence
Train60
Train30
0.30
1.0
0.8
0.6
0.4
0.2
0.0
1.0
0.8
0.6
0.4
0.2
0.0
1.0
0.8
0.6
0.4
0.2
0.0
Maglev30
(c) RBM
1.0
0.8
0.6
0.4
0.2
0.0
1.0
0.8
0.6
0.4
0.2
0.0
1.0
0.8
0.6
0.4
0.2
0.0
Train30
Train60
Train120
Maglev10
Maglev20
Maglev30
Car
Maglev10
Maglev20
Maglev30
(d) MLM
Figure 3: Dataset (a), the predictive error of the RBM choice model against the number of hidden
units (b), and the choice probabilities learned by the RBM choice model (c) and the MLM (d).
average KL divergence over all of the 18 test cases and five runs with varying initialization. The
average KL divergence is also evaluated for training data and is shown with a dashed line. The
confidence interval represents the corresponding standard deviation. The wide confidence interval is
largely due to the variance between test instances (see Figure 4 in the appendix.The horizontal axis
shows the number of the hidden units in the RBM choice model, where zero hidden units correspond
to the MLM. The average KL divergence is reduced from 0.12 for the MLM to 0.02 for the RBM
choice model with 16 hidden units, an improvement by a factor of six.
Figure 3 (c)-(d) shows the choice probabilities given by (a) the RBM choice model with 16 hidden
units and (b) the MLM, after these models are trained for the test case where the choice set consists
of the train with 30-minute interval (Train30) and the maglev with 20-minute interval (Maglev20).
Observe that the RBM choice model gives the choice probabilities that are close to the true choice
probabilities shown in Figure 3 (a), while the MLM has difficulty in fitting these choice probabilities.
Taking a closer look at Figure 3 (a), we can observe that the MLM is fundamentally incapable of
learning this dataset. For example, Train30 is more popular than Maglev20 for people who do not
own cars, while the preference is reversed for car owners (i.e., the attraction effect). The attraction
effect can also be seen for the combination of Maglev30 and Train60. As we have discussed in
Section 3, the MLM cannot represent such attraction effects, but the RBM choice model can.
5
Related work
We now review the prior work related to our contributions. We will see that all of the existing
choice models either cannot represent at least one of the typical choice phenomena or do not have
systematic training algorithms. We will also see that the prior work has analyzed choice models
with respect to whether those choice models can represent typical choice phenomena or others but
only in specific cases of specific strength. On the contrary, our analysis shows that the RBM choice
model can represent the typical choice phenomena for all cases of the specified strength.
A majority of the prior work on the choice model is about the MLM and its variants such as the
hierarchical MLM [5], the multinomial probit model [6], and, generally, random utility models [17].
7
In particular, the attraction effect cannot be represented by these variants of the MLM [13]. In
general, when the choice probability depends only on the values that are determined independently
for each item (e.g., the models of [3, 7]), none of the typical choice phenomena can be represented
[18]. Recently, Hruschka has proposed a choice model based on an RBM [9], but his choice model
cannot represent any of the typical choice phenomena, because the corresponding choice rate is
independent of the choice set. It is thus nontrivial how we use the RBM as a choice model in such a
way that the typical choice phenomena can be represented. In [11], a hierarchical Bayesian choice
model is shown to represent the attraction effect in a specific case.
There also exist choice models that have been numerically shown to represent all of the typical
choice phenomena for some specific cases. For example, sequential sampling models, including
the decision field theory [4] and the leaky competing accumulator model [19], are meant to directly
mimic the cognitive process of the human making a choice [12]. However, no paper has shown an
algorithm that can train a sequential sampling model in such a way that the trained model exhibits the
typical choice phenomena. Shenoy and Yu propose a hierarchical Bayesian model to represent the
three typical choice phenomena [16]. Although they perform inferences of the posterior distributions
that are needed to compute the choice probabilities with their model, they do not show how to train
their model to fit the choice probabilities to given data. Their experiments show that their model
represents the typical choice phenomena in particular cases, where the parameters of the model are
set manually. Rieskamp et al. classify choice models according to whether a choice model can never
represent a certain phenomenon or can do so in some cases to some degree [13]. The phenomena
studied in [13] are not limited to the typical choice phenomena, but they list the typical choice
phenomena as the ones that are robust and significant. Also, Otter et al. exclusively study all of the
typical choice phenomena [12].
Luce is a pioneer of the formal analysis of choice models, which however is largely qualitative [10].
For example, Lemma 3 of [10] can tell us whether a given choice model satisfies the IIA in (16)
for all cases or it violates the IIA for some cases to some degree. We address the new question of
to what degree a choice model can represent each of the typical choice phenomena (e.g., to what
degree the RBM choice model can violate the IIA).
Finally, our theorems can be contrasted with the universal approximation theorem of RBMs, which
states that an arbitrary distribution can be approximated arbitrarily closely with a sufficient number
of hidden units [15, 8]. This is in contrast to our theorems, which show that a single hidden unit
suffices to represent the typical choice phenomena of the strength that is specified in the theorems.
6
Conclusion
The RBM choice model is developed to represent the typical choice phenomena that have been
reported frequently in the literature of cognitive psychology and related areas. Our work motivates a
new direction of research on using RBMs to model such complex behavior of humans. Particularly
interesting behavior includes the one that is considered to be irrational or the one that results from
cognitive biases (see e.g. [1]). The advantages of the RBM choice model that are demonstrated in
this paper include their flexibility in representing complex behavior and the availability of effective
training algorithms.
The RBM choice model can incorporate the attributes of the items in its parameters. Specifically,
one can represent the parameters of the RBM choice model as functions of uX , the attributes of
X ? I analogously to the MLM, where bX can be represented as bX = ? ? uX as we have discussed
after (2). The focus of this paper is in designing the fundamental structure of the RBM choice model
and analyzing its fundamental properties, and the study about the RBM choice model with attributes
will be reported elsewhere. Although the attributes are important for generalization of the RBM
model to unseen items, our experiments suggest that the RBM choice model, without attributes, can
learn the typical choice phenomena from a given choice set and generalize it to unseen choice sets.
Acknowledgements
A part of this research is supported by JST, CREST.
8
References
[1] D. Ariely. Predictably Irrational: The Hidden Forces That Shape Our Decisions. Harper
Perennial, revised and expanded edition, 2010.
[2] M. Bierlaire, K. Axhausen, and G. Abay. The acceptance of modal innovation: The case of
Swissmetro. In Proceedings of the First Swiss Transportation Research Conference, March
2001.
[3] E. Bonilla, S. Guo, and S. Sanner. Gaussian process preference elicitation. In J. Lafferty,
C. K. I. Williams, J. Shawe-Taylor, R. S. Zemel, and A. Culotta, editors, Advances in Neural
Information Processing Systems 23, pages 262?270. 2010.
[4] J. R. Busemeyer and J. T. Townsend. Decision field theory: A dynamic cognition approach to
decision making. Psychological Review, 100:432?459, 1993.
[5] O. Chapelle and Z. Harchaoui. A machine learning approach to conjoint analysis. In L. K.
Saul, Y. Weiss, and L. Bottou, editors, Advances in Neural Information Processing Systems 17,
pages 257?264. 2005.
[6] B. Eric, N. de Freitas, and A. Ghosh. Active preference learning with discrete choice data.
In J. C. Platt, D. Koller, Y. Singer, and S. Roweis, editors, Advances in Neural Information
Processing Systems 20, pages 409?416. 2008.
[7] V. F. Farias, S. Jagabathula, and D. Shah. A nonparametric approach to modeling choice with
limited data. Management Science, 59(2):305?322, 2013.
[8] Y. Freund and D. Haussler. Unsupervised learning of distributions on binary vectors using two
layer networks. Technical Report UCSC-CRL-94-25, University of California, Santa Cruz,
June 1994.
[9] H. Hruschka. Analyzing market baskets by restricted Boltzmann machines. OR Spectrum,
pages 1?22, 2012.
[10] R. D. Luce. Individual choice behavior: A theoretical analysis. John Wiley and Sons, New
York, NY, 1959.
[11] T. Osogami and T. Katsuki. A hierarchical Bayesian choice model with visibility. In Proceedings of the 22nd International Conference on Pattern Recognition (ICPR 2014), pages
3618?3623, August 2014.
[12] T. Otter, J. Johnson, J. Rieskamp, G. M. Allenby, J. D. Brazell, A. Diederich, J. W. Hutchinson,
S. MacEachern, S. Ruan, and J. Townsend. Sequential sampling models of choice: Some recent
advances. Marketing Letters, 19(3-4):255?267, 2008.
[13] J. Rieskamp, J. R. Busemeyer, and B. A. Mellers. Extending the bounds of rationality: Evidence and theories of preferential choice. Journal of Economic Literature, 44:631?661, 2006.
[14] R. M. Roe, J. R. Busemeyer, and J. T. Townsend. Multialternative decision field theory: A
dynamic connectionist model of decision making. Psychological Review, 108(2):370?392,
2001.
[15] N. L. Roux and Y. Bengio. Representational power of restricted Boltzmann machines and deep
belief networks. Neural Computation, 20(6):1631?1649, 2008.
[16] P. Shenoy and A. J. Yu. Rational preference shifts in multi-attribute choice: What is fair?
In Proceedings of the Annual Meeting of the Cognitive Science Society (CogSci 2013), pages
1300?1305, 2013.
[17] K. Train. Discrete Choice Methods with Simulation. Cambridge University Press, second
edition, 2009.
[18] A. Tversky and I. Simonson.
39(10):1179?1189, 1993.
Context-dependent preferences.
Management Science,
[19] M. Usher and J. L. McClelland. Loss aversion and inhibition in dynamical models of multialternative choice. Psychological Review, 111(3):757?769, 2004.
9
| 5280 |@word logit:2 nd:1 simulation:1 solid:1 initial:1 configuration:4 att:3 selecting:1 exclusively:1 existing:2 freitas:1 com:4 yet:1 must:3 readily:1 pioneer:1 cruz:1 mesh:1 visible:9 numerical:2 wx:1 john:1 uak:4 shape:1 visibility:1 rieskamp:3 intelligence:1 selected:8 item:25 realizing:2 core:2 node:1 preference:5 five:2 txk:10 ucsc:1 qualitative:1 consists:5 prove:3 fitting:1 owner:1 indeed:2 market:3 behavior:5 frequently:2 multi:1 cpu:1 equipped:1 window:1 ua:1 increasing:1 xx:1 what:6 interpreted:1 developed:1 ghosh:1 guarantee:2 platt:1 unit:26 shenoy:2 positive:1 limit:4 analyzing:2 approximately:2 initialization:1 studied:2 suggests:1 shaded:1 limited:3 range:5 accumulator:1 busemeyer:3 practice:2 swiss:1 area:2 universal:1 empirical:4 significantly:1 confidence:2 refers:2 suggest:3 cannot:8 close:1 context:2 restriction:3 equivalent:1 demonstrated:1 transportation:3 modifies:1 williams:1 independently:2 survey:1 roux:1 attraction:17 haussler:1 his:1 analogous:2 rationality:1 us:1 designing:1 element:3 approximated:1 particularly:1 recognition:1 observed:1 role:1 bottom:1 culotta:1 connected:2 decrease:3 asked:1 dynamic:2 irrational:2 trained:5 depend:3 tversky:1 compromise:9 predictive:2 eric:1 basis:1 farias:1 perennial:1 represented:15 tx:2 simonson:1 train:15 distinct:1 effective:1 cogsci:1 artificial:2 zemel:1 tell:1 h0:3 larger:1 unseen:2 descriptive:2 advantage:2 propose:1 product:7 hid:1 iff:2 flexibility:6 roweis:1 representational:1 validate:1 regularity:1 extending:1 sim:8 predicted:1 implies:3 quantify:2 switzerland:1 direction:1 closely:1 tokyo:2 attribute:8 compromising:1 stochastic:1 human:9 vx:1 jst:1 violates:1 exchange:1 suffices:1 generalization:1 summation:2 extension:1 considered:2 exp:15 cognition:1 predict:2 makoto:1 gaussian:1 modified:1 rather:2 varying:3 focus:1 june:1 improvement:1 hk:8 contrast:1 sense:2 inference:1 dependent:1 entire:1 hidden:18 koller:1 denoted:5 ruan:1 field:3 never:1 having:1 sampling:3 manually:1 represents:4 look:1 yu:2 unsupervised:1 mimic:1 others:1 report:1 fundamentally:1 connectionist:1 divergence:5 individual:1 acceptance:1 evaluation:2 analyzed:1 extreme:1 pc:1 closer:1 preferential:1 xy:1 conduct:1 taylor:1 initialized:2 circle:1 theoretical:1 psychological:3 instance:1 classify:1 modeling:2 cover:2 deviation:1 subset:1 uniform:2 conducted:1 johnson:1 reported:2 dependency:1 varies:1 hutchinson:1 fundamental:3 international:1 systematic:2 analogously:3 management:2 choose:4 possibly:1 cognitive:4 bx:9 de:1 includes:1 availability:1 bonilla:1 depends:1 red:1 wja:1 qac:5 option:1 contribution:4 accuracy:2 variance:1 who:3 largely:2 correspond:1 yellow:1 generalize:1 bayesian:3 none:1 multiplying:1 basket:1 diederich:1 definition:5 against:1 rbms:4 energy:1 associated:5 rbm:59 rational:1 dataset:9 popular:1 recall:1 car:10 modal:1 wei:1 evaluated:1 marketing:1 hand:1 horizontal:1 indicated:2 effect:37 true:2 adequately:1 hence:1 equality:2 leibler:1 illuminates:2 consideration:1 recently:1 common:1 multinomial:3 jp:1 extend:4 discussed:3 numerically:1 refer:1 significant:1 cambridge:1 hp:1 iia:4 shawe:1 dot:1 chapelle:1 similarity:14 otsuka:1 inhibition:1 dominant:1 posterior:1 own:3 recent:1 irrelevant:1 certain:1 incapable:1 inequality:1 binary:9 arbitrarily:1 meeting:1 seen:1 minimum:1 additional:4 dashed:1 violate:1 harchaoui:1 reduces:1 technical:1 long:1 variant:3 denominator:1 roe:1 represent:27 achieved:4 addition:1 affecting:2 interval:12 operate:2 usher:1 contrary:1 lafferty:1 effectiveness:1 call:1 split:1 bengio:1 independence:2 affect:1 zi:3 fit:1 psychology:1 competing:1 economic:1 luce:2 shift:1 whether:3 motivated:3 six:1 utility:2 gb:1 york:1 cause:2 constitute:2 deep:1 generally:1 santa:1 informally:1 incapability:1 nonparametric:1 extensively:1 mcclelland:1 reduced:1 exist:1 blue:1 discrete:2 run:3 letter:1 i5:1 decision:6 appendix:8 layer:5 bound:1 annual:1 nontrivial:1 strength:16 precisely:1 ucla:1 dominated:1 min:2 expanded:1 relatively:1 according:3 icpr:1 combination:2 disconnected:1 march:1 son:1 osogami:3 modification:1 making:3 restricted:7 jagabathula:1 equation:1 turn:2 needed:1 singer:1 available:2 observe:4 hierarchical:4 hruschka:2 alternative:2 batch:1 shah:1 denotes:4 assumes:1 include:3 remaining:2 widest:1 society:1 unchanged:2 added:2 realized:1 question:1 unclear:1 exhibit:1 gradient:1 reversed:1 majority:1 length:5 index:2 mini:1 ratio:3 innovation:1 ba:5 motivates:1 boltzmann:7 perform:1 takayuki:1 upper:2 revised:1 descent:1 arbitrary:6 august:1 bk:1 namely:3 kl:5 specified:2 california:1 learned:2 address:1 elicitation:1 dynamical:1 pattern:1 max:3 memory:1 including:1 belief:1 tyk:1 power:1 difficulty:1 force:1 predicting:2 townsend:3 sanner:1 representing:2 wik:2 axis:1 prior:4 review:4 literature:2 acknowledgement:1 relative:1 freund:1 probit:1 expect:2 loss:1 interesting:1 conjoint:1 aversion:1 degree:5 sufficient:1 principle:1 editor:3 share:3 ibm:3 elsewhere:1 supported:1 keeping:4 bias:6 side:2 formal:1 wide:1 saul:1 taking:1 leaky:1 ghz:1 made:4 crest:1 kullback:1 keep:1 otter:2 active:1 predictably:1 discriminative:1 spectrum:1 why:1 learn:2 robust:1 ariely:1 mlm:34 bottou:1 complex:3 main:1 edition:2 allowed:1 repeated:1 fair:1 representative:1 attractiveness:1 ny:1 wiley:1 third:1 z0:2 theorem:16 minute:4 specific:4 list:1 dominates:1 evidence:1 adding:5 effectively:1 sequential:3 magnitude:3 ux:9 corresponds:1 satisfies:1 qbc:2 conditional:2 crl:1 change:5 typical:32 specifically:4 determined:1 contrasted:1 lemma:1 called:3 experimental:1 formally:5 people:3 guo:1 maceachern:1 harper:1 meant:1 dissimilar:1 incorporate:1 evaluate:1 phenomenon:42 |
4,728 | 5,281 | Deep Fragment Embeddings for Bidirectional Image
Sentence Mapping
Andrej Karpathy
Armand Joulin
Li Fei-Fei
Department of Computer Science, Stanford University, Stanford, CA 94305, USA
{karpathy,ajoulin,feifeili}@cs.stanford.edu
Abstract
We introduce a model for bidirectional retrieval of images and sentences through
a deep, multi-modal embedding of visual and natural language data. Unlike previous models that directly map images or sentences into a common embedding
space, our model works on a finer level and embeds fragments of images (objects) and fragments of sentences (typed dependency tree relations) into a common space. We then introduce a structured max-margin objective that allows our
model to explicitly associate these fragments across modalities. Extensive experimental evaluation shows that reasoning on both the global level of images and
sentences and the finer level of their respective fragments improves performance
on image-sentence retrieval tasks. Additionally, our model provides interpretable
predictions for the image-sentence retrieval task since the inferred inter-modal
alignment of fragments is explicit.
1
Introduction
There is significant value in the ability to associate natural language descriptions with images. Describing the contents of images is useful for automated image captioning and conversely, the ability
to retrieve images based on natural language queries has immediate image search applications. In
particular, in this work we are interested in training a model on a set of images and their associated
natural language descriptions such that we can later rank a fixed set of withheld sentences given an
image query, and vice versa.
This task is challenging because it requires detailed understanding of the content of images, sentences and their inter-modal correspondence. Consider an example sentence query, such as ?A dog
with a tennis ball is swimming in murky water? (Figure 1). In order to successfully retrieve a corresponding image, we must accurately identify all entities, attributes and relationships present in the
sentence and ground them appropriately to a complex visual scene.
Our primary contribution is in formulating a structured, max-margin objective for a deep neural network that learns to embed both visual and language data into a common, multimodal space. Unlike
previous work that embeds images and sentences, our model breaks down and embeds fragments of
images (objects) and fragments of sentences (dependency tree relations [1]) in a common embedding space and explicitly reasons about their latent, inter-modal correspondences. Extensive empirical evaluation validates our approach. In particular, we report dramatic improvements over state of
the art methods on image-sentence retrieval tasks on Pascal1K [2], Flickr8K [3] and Flickr30K [4]
datasets. We make our code publicly available.
2
Related Work
Image Annotation and Image Search. There is a growing body of work that associates images and
sentences. Some approaches focus on the direction of describing the contents of images, formulated
either as a task of mapping images to a fixed set of sentences written by people [5, 6], or as a task of
automatically generating novel captions [7, 8, 9, 10, 11, 12]. More closely related to our motivation
are methods that allow natural bi-drectional mapping between the two modalities. Socher and FeiFei [13] and Hodosh et al. [3] use Kernel Canonical Correlation Analysis to align images and
sentences, but their method is not easily scalable since it relies on computing kernels quadratic in
1
Figure 1: Our model takes a dataset of
images and their sentence descriptions
and learns to associate their fragments.
In images, fragments correspond to object detections and scene context. In sentences, fragments consist of typed dependency tree [1] relations.
number of images and sentences. Farhadi et al. [5] learn a common meaning space, but their method
is limited to representing both images and sentences with a single triplet of (object, action, scene).
Zitnick et al. [14] use a Conditional Random Field to reason about spatial relationships in cartoon
scenes and their relation to natural language descriptions. Finally, joint models of language and
perception have also been explored in robotics settings [15].
Multimodal Representation Learning. Our approach falls into a general category of learning
from multi-modal data. Several probabilistic models for representing joint multimodal probability
distributions over images and sentences have been developed, using Deep Boltzmann Machines [16],
log-bilinear models [17], and topic models [18, 19]. Ngiam et al. [20] described an autoencoder
that learns audio-video representations through a shared bottleneck layer. More closely related to
our task and approach is the work of Frome et al. [21], who introduced a model that learns to
map images and words to a common semantic embedding with a ranking cost. Adopting a similar
approach, Socher et al. [22] described a Dependency Tree Recursive Neural Network that puts
entire sentences into correspondence with visual data. However, these methods reason about the
image only on the global level using a single, fixed-sized representation from the top layer of a
Convolutional Neural Network as a description for the entire image. In contrast, our model explicitly
reasons about objects that make up a complex scene.
Neural Representations for Images and Natural Language. Our model is a neural network
that is connected to image pixels on one side and raw 1-of-k word representations on the other.
There have been multiple approaches for learning neural representations in these data domains. In
Computer Vision, Convolutional Neural Networks (CNNs) [23] have recently been shown to learn
powerful image representations that support state of the art image classification [24, 25, 26] and
object detection [27, 28]. In language domain, several neural network models have been proposed
to learn word/n-gram representations [29, 30, 31, 32, 33, 34], sentence representations [35] and
paragraph/document representations [36].
3
Proposed Model
Learning and Inference. Our task is to retrieve relevant images given a sentence query, and conversely, relevant sentences given an image query. We train our model on a set of N images and N
corresponding sentences that describe their content (Figure 2). Given this set of correspondences,
we learn the weights of a neural network with a structured loss to output a high score when a compatible image-sentence pair is fed through the network, and low score otherwise. Once the training is
complete, all training data is discarded and the network is evaluated on a withheld set of images and
sentences. The evaluation scores all image-sentence pairs in the test set, sorts the images/sentences
in order of decreasing score and records the location of a ground truth result in the list.
Fragment Embeddings. Our core insight is that images are complex structures that are made
up of multiple entities that the sentences make explicit references to. We capture this intuition
directly in our model by breaking down both images and sentences into fragments and reason about
their alignment. In particular, we propose to detect objects as image fragments and use sentence
dependency tree relations [1] as sentence fragments (Figure 2).
Objective. We will compute the representation of both image and sentence fragments with a neural
network, and interpret the top layer as high-dimensional vectors embedded in a common multimodal space. We will think of the inner product between these vectors as a fragment compatibility
score, and compute the global image-sentence score as a fixed function of the scores of their respective fragments. Intuitively, an image-sentence pair will obtain a high global score if the sentence
fragments can each be confidently matched to some fragment in the image. Finally, we will learn
the weights of the neural networks such that the true image-sentence pairs achieve a score higher
(by a margin) than false image-sentence pairs.
2
Figure 2: Computing the Fragment and image-sentence similarities. Left: CNN representations (green) of
detected objects are mapped to the fragment embedding space (blue, Section 3.2). Right: Dependency tree
relations in the sentence are embedded (Section 3.1). Our model interprets inner products (shown as boxes)
between fragments as a similarity score. The alignment (shaded boxes) is latent and inferred by our model
(Section 3.3.1). The image-sentence similarity is computed as a fixed function of the pairwise fragment scores.
We first describe the neural networks that compute the Image and Sentence Fragment embeddings.
Then we discuss the objective function, which is composed of the two aforementioned objectives.
3.1 Dependency Tree Relations as Sentence Fragments
We would like to extract and represent the set of visually identifiable entities described in a sentence.
For instance, using the example in Figure 2, we would like to identify the entities (dog, child)
and characterise their attributes (black, young) and their pairwise interactions (chasing). Inspired
by previous work [5, 22] we observe that a dependency tree of a sentence provides a rich set of
typed relationships that can serve this purpose more effectively than individual words or bigrams.
We discard the tree structure in favor of a simpler model and interpret each relation (edge) as an
individual sentence fragment (Figure 2, right shows 5 example dependency relations). Thus, we
represent every word using 1-of-k encoding vector w using a dictionary of 400,000 words and map
every dependency triplet (R, w1 , w2 ) into the embedding space as follows:
W e w1
s = f WR
+ bR .
(1)
W e w2
Here, We is a d ? 400, 000 matrix that encodes a 1-of-k vector into a d-dimensional word vector
representation (we use d = 200). We fix We to weights obtained through an unsupervised objective
described in Huang et al. [34]. Note that every relation R can have its own set of weights WR and
biases bR . We fix the element-wise nonlinearity f (.) to be the Rectified Linear Unit (ReLU), which
computes x ? max(0, x). The size of the embedded space is cross-validated, and we found that
values of approximately 1000 generally work well.
3.2 Object Detections as Image Fragments
Similar to sentences, we wish to extract and describe the set of entities that images are composed of.
Inspired by prior work [7], as a modeling assumption we observe that the subject of most sentence
descriptions are attributes of objects and their context in a scene. This naturally motivates the use of
objects and the global context as the fragments of an image. In particular, we follow Girshick et al.
[27] and detect objects in every image with a Region Convolutional Neural Network (RCNN). The
CNN is pre-trained on ImageNet [37] and finetuned on the 200 classes of the ImageNet Detection
Challenge [38]. We use the top 19 detected locations and the entire image as the image fragments
and compute the embedding vectors based on the pixels Ib inside each bounding box as follows:
v = Wm [CNN?c (Ib )] + bm ,
(2)
where CNN(Ib ) takes the image inside a given bounding box and returns the 4096-dimensional
activations of the fully connected layer immediately before the classifier. The CNN architecture is
identical to the one described in Girhsick et al. [27]. It contains approximately 60 million parameters
?c and closely resembles the architecture of Krizhevsky et al [25].
3.3 Objective Function
We are now ready to formulate the objective function. Recall that we are given a training set of N
images and corresponding sentences. In the previous sections we described parameterized functions
that map every sentence and image to a set of fragment vectors {s} and {v}, respectively. All
parameters of our model are contained in these two functions. As shown in Figure 2, our model
3
Figure 3: The two objectives for a
batch of 2 examples. Left: Rows represent fragments vi , columns sj . Every square shows an ideal scenario of
yij = sign(viT sj ) in the MIL objective. Red boxes are yij = ?1.
Yellow indicates members of positive bags that happen to currently
be yij = ?1. Right: The scores
are accumulated with Equation 6 into
image-sentence score matrix Skl .
then interprets the inner product viT sj between an image fragment vi and a sentence fragment sj as
a similarity score, and computes the image-sentence similarity as a fixed function of the scores of
their respective fragments.
We are motivated by two criteria in designing the objective function. First, the image-sentence
similarities should be consistent with the ground truth correspondences. That is, corresponding
image-sentence pairs should have a higher score than all other image-sentence pairs. This will
be enforced by the Global Ranking Objective. Second, we introduce a Fragment Alignment
Objective that explicitly learns the appearance of sentence fragments (such as ?black dog?) in the
visual domain. Our full objective is the sum of these two objectives and a regularization term:
C(?) = CF (?) + ?CG (?) + ?||?||22 ,
(3)
where ? is a shorthand for parameters of our neural network (? = {We , WR , bR , Wm , bm , ?c }) and
? and ? are hyperparameters that we cross-validate. We now describe both objectives in more detail.
3.3.1 Fragment Alignment Objective
The Fragment Alignment Objective encodes the intuition that if a sentence contains a fragment
(e.g.?blue ball?, Figure 3), at least one of the boxes in the corresponding image should have a high
score with this fragment, while all the other boxes in all the other images that have no mention of
?blue ball? should have a low score. This assumption can be violated in multiple ways: a triplet
may not refer to anything visually identifiable in the image. The box that the triplet refers to may
not be detected by the RCNN. Lastly, other images may contain the described visual concept but
its mention may omitted in the associated sentence descriptions. Nonetheless, the assumption is
still satisfied in many cases and can be used to formulate a cost function. Consider an (incomplete)
Fragment Alignment Objective that assumes a dense alignment between every corresponding image
and sentence fragments:
XX
C0 (?) =
max(0, 1 ? yij viT sj ).
(4)
i
j
Here, the sum is over all pairs of image and sentence fragments in the training set. The quantity viT sj
can be interpreted as the alignment score of visual fragment vi and sentence fragment sj . In this
incomplete objective, we define yij as +1 if fragments vi and sj occur together in a corresponding
image-sentence pair, and ?1 otherwise. Intuitively, C0 (?) encourages scores in red regions of Figure
3 to be less than -1 and scores along the block diagonal (green and yellow) to be more than +1.
Multiple Instance Learning extension. The problem with the objective C0 (?) is that it assumes
dense alignment between all pairs of fragments in every corresponding image-sentence pair. However, this is hardly ever the case. For example, in Figure 3, the ?boy playing? triplet refers to only
one of the three detections. We now describe a Multiple Instance Learning (MIL) [39] extension
of the objective C0 that attempts to infer the latent alignment between fragments in corresponding
image-sentence pairs. Concretely, for every triplet we put image fragments in the associated image into a positive bag, while image fragments in every other image become negative examples.
Our precise formulation is inspired by the mi-SVM [40], which is a simple and natural extension
of a Support Vector Machine to a Multiple Instance Learning setting. Instead of treating the yij as
constants, we minimize over them by wrapping Equation 4 as follows:
4
CF (?) = min C0 (?)
yij
s.t.
X yij + 1
? 1 ?j
2
i?p
(5)
j
yij = ?1 ?i, j s.t. mv (i) 6= ms (j) and yij ? {?1, 1}
Here, we define pj to be the set of image fragments in the positive bag for sentence fragment j.
mv (i) and ms (j) return the index of the image and sentence (an element of {1, . . . , N }) that the
fragments vi and sj belong to. Note that the inequality simply states that at least one of the yij
should be positive for every sentence fragment j (i.e. at least one green box in every column in
Figure 3). This objective cannot be solved efficiently [40] but a commonly used heuristic is to set
yij = sign(viT sj ). If the constraint is not satisfied for any positive bag (i.e. all scores were below
zero), the highest-scoring item in the positive bag is set to have a positive label.
3.3.2 Global Ranking Objective
Recall that the Global Ranking Objective ensures that the computed image-sentence similarities are
consistent with the ground truth annotation. First, we define the image-sentence alignment score to
be the average thresholded score of their pairwise fragment scores:
XX
1
Skl =
max(0, viT sj ).
(6)
|gk |(|gl | + n) i?g j?g
k
l
Here, gk is the set of image fragments in image k and gl is the set of sentence fragments in sentence
l. Both k, l range from 1, . . . , N . We truncate scores at zero because in the mi-SVM objective, scores
greater than 0 are considered correct alignments and scores less than 0 are considered to be incorrect
alignments (i.e. false members of a positive bag). In practice, we found that it was helpful to add
a smoothing term n, since short sentences can otherwise have an advantage (we found that n = 5
works well and that this setting is not very sensitive). The Global Ranking Objective then becomes:
i
X
XhX
max(0, Slk ? Skk + ?) .
(7)
CG (?) =
max(0, Skl ? Skk + ?) +
k
l
l
|
{z
}
rank images
|
{z
rank sentences
}
Here, ? is a hyperparameter that we cross-validate. The objective stipulates that the score for true
image-sentence pairs Skk should be higher than Skl or Slk for any l 6= k by at least a margin of ?.
3.4 Optimization
We use Stochastic Gradient Descent (SGD) with mini-batches of 100, momentum of 0.9 and make
20 epochs through the training data. The learning rate is cross-validated and annealed by a fraction
of ?0.1 for the last two epochs. Since both Multiple Instance Learning and CNN finetuning benefit
from a good initialization, we run the first 10 epochs with the fragment alignment objective C0
and CNN weights ?c fixed. After 10 epochs, we switch to the full MIL objective CF and begin
finetuning the CNN. The word embedding matrix We is kept fixed due to overfitting concerns. Our
implementation runs at approximately 1 second per batch on a standard CPU workstation.
4
Experiments
Datasets. We evaluate our image-sentence retrieval performance on Pascal1K [2], Flickr8K [3] and
Flickr30K [4] datasets. The datasets contain 1,000, 8,000 and 30,000 images respectively and each
image is annotated using Amazon Mechanical Turk with 5 independent sentences.
Sentence Data Preprocessing. We did not explicitly filter, spellcheck or normalize any of the
sentences for simplicity. We use the Stanford CoreNLP parser to compute the dependency trees
for every sentence. Since there are many possible relation types (as many as hundreds), due to
overfitting concerns and practical considerations we remove all relation types that occur less than
1% of the time in each dataset. In practice, this reduces the number of relations from 136 to 16 in
Pascal1K, 170 to 17 in Flickr8K, and from 212 to 21 in Flickr30K. Additionally, all words that are
not found in our dictionary of 400,000 words [34] are discarded.
Image Data Preprocessing. We use the Caffe [41] implementation of the ImageNet Detection
RCNN model [27] to detect objects in all images. On our machine with a Tesla K40 GPU,
the RCNN processes one image in approximately 25 seconds. We discard the predictions for
200 ImageNet detection classes and only keep the 4096-D activations of the fully connect layer
immediately before the classifier at all of the top 19 detected locations and from the entire image.
5
Model
Random Ranking
Socher et al. [22]
kCCA. [22]
DeViSE [21]
SDT-RNN [22]
Our model
R@1
4.0
23.0
21.0
17.0
25.0
39.0
Pascal1K
Image Annotation
R@5 R@10 Mean r
9.0
12.0
71.0
45.0
63.0
16.9
47.0
61.0
18.0
57.0
68.0
11.9
56.0
70.0
13.4
68.0
79.0
10.5
R@1
1.6
16.4
16.4
21.6
25.4
23.6
Image Search
R@5 R@10
5.2
10.6
46.6
65.6
41.4
58.0
54.6
72.4
65.2
84.4
65.2
79.8
Mean r
50.0
12.5
15.9
9.5
7.0
7.6
Table 1: Pascal1K ranking experiments. R@K is Recall@K (high is good). Mean r is the mean rank (low is
good). Note that the test set only consists of 100 images.
Model
Random Ranking
Socher et al. [22]
DeViSE [21]
SDT-RNN [22]
Fragment Alignment Objective
Global Ranking Objective
(?) Fragment + Global
? ? Images: Fullframe Only
? ? Sentences: BOW
? ? Sentences: Bigrams
Our model (? + MIL)
* Hodosh et al. [3]
* Our model (? + MIL)
R@1
0.1
4.5
4.8
6.0
7.2
5.8
12.5
5.9
9.1
8.7
12.6
8.3
9.3
Flickr8K
Image Annotation
R@5 R@10 Med r
0.6
1.1
631
18.0
28.6
32
16.5
27.3
28
22.7
34.0
23
21.9
31.8
25
21.8
34.8
20
29.4
43.8
14
19.2
27.3
34
25.9
40.7
17
28.5
41.0
16
32.9
44.0
14
21.6
30.3
34
24.9
37.4
21
R@1
0.1
6.1
5.9
6.6
5.9
7.5
8.6
5.2
6.9
8.5
9.7
7.6
8.8
Image Search
R@5 R@10
0.5
1.0
18.5
29.0
20.1
29.6
21.6
31.7
20.0
30.3
23.4
35.0
26.7
38.7
17.6
26.5
22.4
34.0
25.2
37.0
29.6
42.5
20.7
30.1
27.9
41.3
Med r
500
29
29
25
26
21
17
32
23
20
15
38
17
Table 2: Flickr8K experiments. R@K is Recall@K (high is good). Med r is the median rank (low is good).
The starred evaluation criterion (*) in [3] is slightly different since it discards 4,000 out of 5,000 test sentences.
Evaluation Protocol for Bidirectional Retrieval. For Pascal1K we follow Socher et al. [22] and
use 800 images for training, 100 for validation and 100 for testing. For Flickr datasets we use
1,000 images for validation, 1,000 for testing and the rest for training (consistent with [3]). We
compute the dense image-sentence similarity Skl between every image-sentence pair in the test set
and rank images and sentences in order of decreasing score. For both Image Annotation and Image
Search, we report the median rank of the closest ground truth result in the list, as well as Recall@K
which computes the fraction of times the correct result was found among the top K items. When
comparing to Hodosh et al. [3] we closely follow their evaluation protocol and only keep a subset
of N sentences out of total 5N (we use the first sentence out of every group of 5).
4.1 Comparison Methods
SDT-RNN. Socher et al. [22] embed a fullframe CNN representation with the sentence representation from a Semantic Dependency Tree Recursive Neural Network (SDT-RNN). Their loss matches
our global ranking objective. We requested the source code of Socher et al. [22] and substituted the
Flickr8K and Flick30K datasets. To better understand the benefits of using our detection CNN and
for a more fair comparison we also train their method with our CNN features. Since we have multiple objects per image, we average representations from all objects with detection confidence above a
(cross-validated) threshold. We refer to the exact method of Socher et al. [22] with a single fullframe
CNN as ?Socher et al?, and to their method with our combined CNN features as ?SDT-RNN?.
DeViSE. The DeViSE [21] source code is not publicly available but their approach is a special case
of our method with the following modifications: we use the average (L2-normalized) word vectors
as a sentence fragment, the average CNN activation of all objects above a detection threshold (as
discussed in case of SDT-RNN) as an image fragment and only use the global ranking objective.
4.2 Quantitative Evaluation
Our model outperforms previous methods. Our full method consistently outperforms previous
methods on Flickr8K (Table 2) and Flickr30K (Table 3) datasets. On Pascal1K (Table 1) the
SDT-RNN appears to be competitive on Image Search.
Fragment and Global Objectives are complementary. As seen in Tables 2 and 3, both objectives
perform well independently, but benefit from the combination. Note that the Global Objective
performs consistently better, possibly because it directly minimizes the evaluation criterion (ranking
6
Model
Random Ranking
DeViSE [21]
SDT-RNN [22]
Fragment Alignment Objective
Global Ranking Objective
(?) Fragment + Global
Our model (? + MIL)
Our model + Finetune CNN
R@1
0.1
4.5
9.6
11
11.5
12.0
14.2
16.4
Flickr30K
Image Annotation
R@5 R@10 Med r
0.6
1.1
631
18.1
29.2
26
29.8
41.1
16
28.7
39.3
18
33.2
44.9
14
37.1
50.0
10
37.7
51.3
10
40.2
54.7
8
R@1
0.1
6.7
8.9
7.6
8.8
9.9
10.2
10.3
Image Search
R@5 R@10
0.5
1.0
21.9
32.7
29.8
41.1
23.8
34.5
27.6
38.4
30.5
43.2
30.8
44.2
31.4
44.5
Med r
500
25
16
22
17
14
14
13
Table 3: Flickr30K experiments. R@K is Recall@K (high is good). Med r is the median rank (low is good).
Figure 4: Qualitative Image Annotation results. Below each image we show the top 5 sentences (among a set
of 5,000 test sentences) in descending confidence. We also show the triplets for the top sentence and connect
each to the detections with the highest compatibility score (indicated by lines). The numbers next to each triplet
indicate the matching fragment score. We color a sentence green if it correct and red otherwise.
cost), while the Fragment Alignment Objective only does so indirectly.
Extracting object representations is important. Using only the global scene-level CNN representation as a single fragment for every image leads to a consistent drop in performance, suggesting
that a single fullframe CNN alone is inadequate in effectively representing the images. (Table 2)
Dependency tree relations outperform BoW/bigram representations. We compare to a simpler
Bag of Words (BoW) baseline to understand the contribution of dependency relations. In BoW
baseline we iterate over words instead of dependency triplets when creating bags of sentence
fragments (set w1 = w2 in Equation1). As can be seen in the Table 2, this leads to a consistent drop
in performance. This drop could be attributed to a difference between using one word or two words
at a time, so we also compare to a bigram baseline where the words w1 , w2 in Equation 1 refer to
consecutive words in a sentence, not nodes that share an edge in the dependency tree. Again, we
observe a consistent performance drop, which suggests that the dependency relations provide useful
structure that the neural network takes advantage of.
Finetuning the CNN helps on Flickr30K. Our end-to-end Neural Network approach allows us to
backpropagate gradients all the way down to raw data (pixels or 1-of-k word encodings). In particular, we observed additional improvements on the Flickr30K dataset (Table 3) when we finetune the
CNN. Training the CNN improves the validation error for a while but the model eventually starts to
overfit. Thus, we found it critical to keep track of the validation error and freeze the model before it
overfits. We were not able to improve the validation performance on Pascal1K and Flickr8K datasets
and suspect that there is an insufficient amount of training data.
4.3
Qualitative Experiments
Interpretable Predictions. We show some example sentence retrieval results in Figure 4. The
alignment in our model is explicitly inferred on the fragment level, which allows us to interpret the
scores between images and sentences. For instance, in the last image it is apparent that the model
retrieved the top sentence because it erroneously associated a mention of a blue person to the blue
flag on the bottom right of the image.
Fragment Alignment Objective trains attribute detectors. The detection CNN is trained to
predict one of 200 ImageNet Detection classes, so it is not clear if the representation is powerful
enough to support learning of more complex attributes of the objects or generalize to novel classes.
To see whether our model successfully learns to predict sentence triplets, we fix a triplet vector and
7
Figure 5: We fix a triplet and retrieve the highest scoring image fragments in the test set. Note that ball, person
and dog are ImageNet Detection classes but their visual properties (e.g. soccer, standing, snowboarding, black)
are not. Jackets and rocky scenes are not ImageNet Detection classes. Find more in supplementary material.
search for the highest scoring boxes in the test set. Qualitative results shown in Figure 5 suggest
that the model is indeed capable of generalizing to more fine-grained subcategories (such as ?black
dog?, ?soccer ball?) and to out of sample classes such as ?rocky terrain? and ?jacket?.
Limitations. Our model is subject to multiple limitations. From a modeling perspective, the use of
edges from a dependency tree is simple, but not always appropriate. First, a single complex phrase
that describes a single visual entity can be split across multiple sentence fragments. For example,
?black and white dog? is parsed as two relations (CONJ, black, white) and (AMOD, white, dog).
Conversely, there are many dependency relations that don?t have a clear grounding in the image (for
example ?each other?). Furthermore, phrases such as ?three children playing? that describe some
particular number of visual entiries are not modeled. While we have shown that the relations give
rise to more powerful representations than words or bigrams, a more careful treatment of sentence
fragments will likely lead to further improvements. On the image side, the non-maximum suppression in the RCNN can sometimes detect, for example, multiple people inside one person. Since the
model does not take into account any spatial information associated with the detections, it is hard
for it to disambiguate between two distinct people or spurious detections of one person.
5
Conclusions
We addressed the problem of bidirectional retrieval of images and sentences. Our neural network
learns a multi-modal embedding space for fragments of images and sentences and reasons about
their latent, inter-modal alignment. We have shown that our model significantly improves the retrieval performance on image sentence retrieval tasks compared to previous work. Our model also
produces interpretable predictions. In future work we hope to develop better sentence fragment
representations, incorporate spatial reasoning, and move beyond bags of fragments.
Acknowledgments. We thank Justin Johnson and Jon Krause for helpful comments and discussions.
We gratefully acknowledge the support of NVIDIA Corporation with the donation of the GPUs used
for this research. This research is supported by an ONR MURI grant, and NSF ISS-1115313.
References
[1] De Marneffe, M.C., MacCartney, B., Manning, C.D., et al.: Generating typed dependency parses from
phrase structure parses. In: Proceedings of LREC. Volume 6. (2006) 449?454
[2] Rashtchian, C., Young, P., Hodosh, M., Hockenmaier, J.: Collecting image annotations using amazon?s
mechanical turk. In: Proceedings of the NAACL HLT 2010 Workshop on Creating Speech and Language
Data with Amazon?s Mechanical Turk, Association for Computational Linguistics (2010) 139?147
[3] Hodosh, M., Young, P., Hockenmaier, J.: Framing image description as a ranking task: data, models and
evaluation metrics. Journal of Artificial Intelligence Research (2013)
[4] Young, P., Lai, A., Hodosh, M., Hockenmaier, J.: From image descriptions to visual denotations: New
similarity metrics for semantic inference over event descriptions. TACL (2014)
[5] Farhadi, A., Hejrati, M., Sadeghi, M.A., Young, P., Rashtchian, C., Hockenmaier, J., Forsyth, D.: Every
picture tells a story: Generating sentences from images. In: ECCV. (2010)
[6] Ordonez, V., Kulkarni, G., Berg, T.L.: Im2text: Describing images using 1 million captioned photographs.
In: NIPS. (2011)
[7] Kulkarni, G., Premraj, V., Dhar, S., Li, S., Choi, Y., Berg, A.C., Berg, T.L.: Baby talk: Understanding
and generating simple image descriptions. In: CVPR. (2011)
8
[8] Yao, B.Z., Yang, X., Lin, L., Lee, M.W., Zhu, S.C.: I2t: Image parsing to text description. Proceedings
of the IEEE 98(8) (2010) 1485?1508
[9] Yang, Y., Teo, C.L., Daum?e III, H., Aloimonos, Y.: Corpus-guided sentence generation of natural images.
In: EMNLP. (2011)
[10] Li, S., Kulkarni, G., Berg, T.L., Berg, A.C., Choi, Y.: Composing simple image descriptions using
web-scale n-grams. In: CoNLL. (2011)
[11] Mitchell, M., Han, X., Dodge, J., Mensch, A., Goyal, A., Berg, A., Yamaguchi, K., Berg, T., Stratos,
K., Daum?e, III, H.: Midge: Generating image descriptions from computer vision detections. In: EACL.
(2012)
[12] Kuznetsova, P., Ordonez, V., Berg, A.C., Berg, T.L., Choi, Y.: Collective generation of natural image
descriptions. In: ACL. (2012)
[13] Socher, R., Fei-Fei, L.: Connecting modalities: Semi-supervised segmentation and annotation of images
using unaligned text corpora. In: CVPR. (2010)
[14] Zitnick, C.L., Parikh, D., Vanderwende, L.: Learning the visual interpretation of sentences. ICCV (2013)
[15] Matuszek*, C., FitzGerald*, N., Zettlemoyer, L., Bo, L., Fox, D.: A Joint Model of Language and
Perception for Grounded Attribute Learning. In: Proc. of the 2012 International Conference on Machine
Learning, Edinburgh, Scotland (June 2012)
[16] Srivastava, N., Salakhutdinov, R.: Multimodal learning with deep boltzmann machines. In: NIPS. (2012)
[17] Kiros, R., Zemel, R.S., Salakhutdinov, R.: Multimodal neural language models. ICML (2014)
[18] Jia, Y., Salzmann, M., Darrell, T.: Learning cross-modality similarity for multinomial data. In: ICCV.
(2011)
[19] Barnard, K., Duygulu, P., Forsyth, D., De Freitas, N., Blei, D.M., Jordan, M.I.: Matching words and
pictures. JMLR (2003)
[20] Ngiam, J., Khosla, A., Kim, M., Nam, J., Lee, H., Ng, A.Y.: Multimodal deep learning. In: ICML. (2011)
[21] Frome, A., Corrado, G.S., Shlens, J., Bengio, S., Dean, J., Mikolov, T., et al.: Devise: A deep visualsemantic embedding model. In: NIPS. (2013)
[22] Socher, R., Karpathy, A., Le, Q.V., Manning, C.D., Ng, A.Y.: Grounded compositional semantics for
finding and describing images with sentences. TACL (2014)
[23] LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition.
Proceedings of the IEEE 86(11) (1998) 2278?2324
[24] Le, Q.V.: Building high-level features using large scale unsupervised learning. In: Acoustics, Speech and
Signal Processing (ICASSP), 2013 IEEE International Conference on, IEEE (2013) 8595?8598
[25] Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: NIPS. (2012)
[26] Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional neural networks. arXiv preprint
arXiv:1311.2901 (2013)
[27] Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and
semantic segmentation. In: CVPR. (2014)
[28] Sermanet, P., Eigen, D., Zhang, X., Mathieu, M., Fergus, R., LeCun, Y.: Overfeat: Integrated recognition,
localization and detection using convolutional networks. In: ICLR. (2014)
[29] Bengio, Y., Schwenk, H., Sen?ecal, J.S., Morin, F., Gauvain, J.L.: Neural probabilistic language models.
In: Innovations in Machine Learning. Springer (2006)
[30] Mnih, A., Hinton, G.: Three new graphical models for statistical language modelling. In: ICML. (2007)
[31] Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and
phrases and their compositionality. In: NIPS. (2013)
[32] Turian, J., Ratinov, L., Bengio, Y.: Word representations: a simple and general method for semisupervised learning. In: ACL. (2010)
[33] Collobert, R., Weston, J.: A unified architecture for natural language processing: Deep neural networks
with multitask learning. In: ICML. (2008)
[34] Huang, E.H., Socher, R., Manning, C.D., Ng, A.Y.: Improving word representations via global context
and multiple word prototypes. In: ACL. (2012)
[35] Socher, R., Lin, C.C., Manning, C., Ng, A.Y.: Parsing natural scenes and natural language with recursive
neural networks. In: ICML. (2011)
[36] Le, Q.V., Mikolov, T.: Distributed representations of sentences and documents. ICML (2014)
[37] Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: A large-scale hierarchical image
database. In: CVPR. (2009)
[38] Russakovsky, O., Deng, J., Krause, J., Berg, A., Fei-Fei, L.: Large scale visual recognition challenge
2013. http://image-net.org/challenges/LSVRC/2013/ (2013)
[39] Chen, Y., Bi, J., Wang, J.Z.: Miles: Multiple-instance learning via embedded instance selection. CVPR
28(12) (2006)
[40] Andrews, S., Hofmann, T., Tsochantaridis, I.: Multiple instance learning with generalized support vector
machines. In: AAAI/IAAI. (2002) 943?944
[41] Jia, Y.:
Caffe: An open source convolutional architecture for fast feature embedding.
http://caffe.berkeleyvision.org/ (2013)
9
| 5281 |@word multitask:1 armand:1 cnn:21 bigram:5 c0:6 open:1 sgd:1 dramatic:1 mention:3 contains:2 fragment:78 score:33 salzmann:1 document:3 outperforms:2 freitas:1 comparing:1 gauvain:1 activation:3 must:1 written:1 gpu:1 parsing:2 happen:1 hofmann:1 remove:1 treating:1 interpretable:3 drop:4 alone:1 intelligence:1 item:2 scotland:1 core:1 short:1 record:1 blei:1 provides:2 node:1 location:3 org:2 simpler:2 zhang:1 along:1 become:1 incorrect:1 shorthand:1 consists:1 qualitative:3 paragraph:1 inside:3 introduce:3 pairwise:3 inter:4 indeed:1 growing:1 multi:3 kiros:1 inspired:3 salakhutdinov:2 decreasing:2 automatically:1 cpu:1 farhadi:2 becomes:1 begin:1 xx:2 matched:1 interpreted:1 minimizes:1 developed:1 unified:1 finding:1 corporation:1 hejrati:1 quantitative:1 every:17 flickr8k:8 collecting:1 classifier:2 unit:1 grant:1 before:3 positive:8 bilinear:1 encoding:2 approximately:4 black:6 acl:3 initialization:1 resembles:1 conversely:3 challenging:1 shaded:1 suggests:1 jacket:2 limited:1 bi:2 range:1 snowboarding:1 practical:1 acknowledgment:1 lecun:2 testing:2 recursive:3 block:1 chasing:1 practice:2 goyal:1 empirical:1 rnn:8 significantly:1 matching:2 word:24 pre:1 refers:2 confidence:2 morin:1 suggest:1 cannot:1 selection:1 andrej:1 tsochantaridis:1 put:2 context:4 descending:1 map:4 dean:2 annealed:1 vit:6 independently:1 formulate:2 amazon:3 simplicity:1 immediately:2 insight:1 shlens:1 nam:1 retrieve:4 embedding:11 i2t:1 hierarchy:1 parser:1 caption:1 exact:1 premraj:1 designing:1 associate:4 element:2 finetuned:1 recognition:3 muri:1 database:1 observed:1 bottom:1 preprint:1 solved:1 capture:1 wang:1 region:2 ensures:1 connected:2 k40:1 highest:4 intuition:2 fitzgerald:1 trained:2 eacl:1 serve:1 dodge:1 localization:1 multimodal:7 easily:1 joint:3 finetuning:3 icassp:1 schwenk:1 talk:1 train:3 distinct:1 fast:1 describe:6 query:5 detected:4 artificial:1 tell:1 zemel:1 caffe:3 apparent:1 heuristic:1 stanford:4 supplementary:1 cvpr:5 otherwise:4 ability:2 favor:1 think:1 validates:1 advantage:2 net:1 sen:1 propose:1 interaction:1 product:3 unaligned:1 relevant:2 bow:4 starred:1 achieve:1 rashtchian:2 description:15 validate:2 normalize:1 sutskever:2 darrell:2 captioning:1 generating:5 produce:1 object:19 help:1 donation:1 develop:1 andrew:1 c:1 frome:2 indicate:1 direction:1 guided:1 closely:4 correct:3 attribute:6 cnns:1 stochastic:1 annotated:1 xhx:1 filter:1 material:1 fix:4 feifeili:1 yij:12 extension:3 considered:2 ground:5 visually:2 mapping:3 predict:2 dictionary:2 consecutive:1 omitted:1 purpose:1 proc:1 bag:9 label:1 currently:1 sensitive:1 teo:1 vice:1 successfully:2 hope:1 always:1 mil:6 validated:3 focus:1 june:1 improvement:3 consistently:2 rank:8 indicates:1 modelling:1 contrast:1 cg:2 baseline:3 yamaguchi:1 detect:4 helpful:2 inference:2 suppression:1 kim:1 accumulated:1 entire:4 integrated:1 spurious:1 captioned:1 relation:19 interested:1 semantics:1 pixel:3 compatibility:2 classification:2 aforementioned:1 among:2 overfeat:1 art:2 spatial:3 smoothing:1 special:1 field:1 once:1 ng:4 cartoon:1 identical:1 unsupervised:2 icml:6 jon:1 future:1 report:2 composed:2 individual:2 attempt:1 detection:20 mnih:1 evaluation:9 alignment:21 accurate:1 edge:3 capable:1 conj:1 respective:3 fox:1 tree:14 incomplete:2 girshick:2 instance:9 column:2 modeling:2 phrase:4 cost:3 subset:1 hundred:1 krizhevsky:2 inadequate:1 johnson:1 dependency:20 connect:2 combined:1 person:4 international:2 standing:1 probabilistic:2 lee:2 dong:1 together:1 corenlp:1 connecting:1 yao:1 w1:4 again:1 aaai:1 satisfied:2 huang:2 possibly:1 emnlp:1 creating:2 return:2 li:5 suggesting:1 account:1 sdt:8 de:2 skl:5 forsyth:2 explicitly:6 ranking:15 vi:5 mv:2 collobert:1 later:1 break:1 overfits:1 red:3 wm:2 sort:1 competitive:1 start:1 annotation:9 jia:2 contribution:2 minimize:1 square:1 publicly:2 convolutional:7 who:1 efficiently:1 correspond:1 identify:2 yellow:2 generalize:1 raw:2 accurately:1 rectified:1 finer:2 russakovsky:1 detector:1 flickr:1 hlt:1 nonetheless:1 typed:4 turk:3 naturally:1 associated:5 mi:2 attributed:1 workstation:1 dataset:3 treatment:1 iaai:1 mitchell:1 recall:6 color:1 improves:3 murky:1 segmentation:2 appears:1 bidirectional:4 finetune:2 higher:3 supervised:1 follow:3 modal:7 formulation:1 evaluated:1 box:10 furthermore:1 lastly:1 correlation:1 overfit:1 tacl:2 web:1 ordonez:2 indicated:1 semisupervised:1 building:1 usa:1 grounding:1 contain:2 true:2 concept:1 normalized:1 naacl:1 regularization:1 semantic:4 mile:1 white:3 visualizing:1 encourages:1 berkeleyvision:1 anything:1 soccer:2 criterion:3 m:2 generalized:1 complete:1 performs:1 reasoning:2 image:145 meaning:1 wise:1 novel:2 recently:1 consideration:1 parikh:1 common:7 multinomial:1 volume:1 million:2 belong:1 discussed:1 association:1 interpretation:1 interpret:3 significant:1 refer:3 freeze:1 versa:1 nonlinearity:1 language:16 gratefully:1 tennis:1 similarity:10 han:1 align:1 add:1 closest:1 own:1 retrieved:1 perspective:1 discard:3 scenario:1 nvidia:1 inequality:1 onr:1 baby:1 devise:6 scoring:3 seen:2 greater:1 additional:1 deng:2 feifei:1 corrado:2 signal:1 semi:1 multiple:14 full:3 infer:1 reduces:1 match:1 cross:6 retrieval:10 lin:2 lai:1 prediction:4 scalable:1 kcca:1 vision:2 metric:2 arxiv:2 kernel:2 adopting:1 represent:3 sometimes:1 robotics:1 grounded:2 ajoulin:1 zettlemoyer:1 fine:1 krause:2 addressed:1 median:3 source:3 modality:4 appropriately:1 w2:4 rest:1 unlike:2 comment:1 subject:2 med:6 suspect:1 member:2 jordan:1 extracting:1 yang:2 ideal:1 split:1 embeddings:3 enough:1 automated:1 switch:1 iterate:1 relu:1 iii:2 bengio:4 architecture:4 interprets:2 inner:3 prototype:1 haffner:1 br:3 bottleneck:1 whether:1 motivated:1 speech:2 hardly:1 action:1 compositional:1 deep:9 useful:2 generally:1 detailed:1 clear:2 characterise:1 karpathy:3 slk:2 amount:1 category:1 http:2 outperform:1 canonical:1 nsf:1 sign:2 wr:3 per:2 track:1 blue:5 stipulates:1 hyperparameter:1 group:1 threshold:2 pj:1 thresholded:1 kept:1 dhar:1 swimming:1 fraction:2 sum:2 enforced:1 run:2 ratinov:1 parameterized:1 powerful:3 conll:1 matuszek:1 layer:5 lrec:1 correspondence:5 quadratic:1 identifiable:2 occur:2 denotation:1 constraint:1 fei:8 scene:9 vanderwende:1 encodes:2 erroneously:1 min:1 formulating:1 mensch:1 duygulu:1 mikolov:3 gpus:1 department:1 structured:3 truncate:1 ball:5 combination:1 manning:4 across:2 hodosh:6 slightly:1 describes:1 hockenmaier:4 modification:1 intuitively:2 iccv:2 equation:3 describing:4 discus:1 eventually:1 ecal:1 fed:1 end:2 available:2 flickr30k:8 observe:3 hierarchical:1 indirectly:1 appropriate:1 batch:3 eigen:1 top:8 assumes:2 cf:3 linguistics:1 zeiler:1 graphical:1 daum:2 parsed:1 objective:41 move:1 malik:1 quantity:1 wrapping:1 primary:1 diagonal:1 gradient:3 iclr:1 thank:1 mapped:1 entity:6 topic:1 water:1 reason:6 code:3 index:1 relationship:3 mini:1 insufficient:1 modeled:1 sermanet:1 innovation:1 boy:1 gk:2 negative:1 skk:3 rise:1 implementation:2 motivates:1 boltzmann:2 collective:1 perform:1 im2text:1 datasets:8 discarded:2 withheld:2 acknowledge:1 descent:1 immediate:1 hinton:2 ever:1 precise:1 inferred:3 compositionality:1 introduced:1 dog:7 pair:14 mechanical:3 extensive:2 sentence:114 imagenet:9 acoustic:1 framing:1 nip:5 aloimonos:1 able:1 beyond:1 justin:1 below:2 perception:2 confidently:1 challenge:3 max:7 green:4 video:1 critical:1 event:1 natural:13 zhu:1 representing:3 sadeghi:1 improve:1 rocky:2 picture:2 mathieu:1 ready:1 autoencoder:1 extract:2 text:2 prior:1 understanding:3 epoch:4 l2:1 embedded:4 loss:2 fully:2 subcategories:1 par:2 generation:2 limitation:2 validation:5 rcnn:5 consistent:6 story:1 playing:2 share:1 row:1 eccv:1 compatible:1 gl:2 last:2 supported:1 side:2 allow:1 bias:1 understand:2 fall:1 benefit:3 edinburgh:1 distributed:2 gram:2 rich:2 computes:3 concretely:1 made:1 commonly:1 preprocessing:2 bm:2 sj:11 keep:3 global:19 overfitting:2 corpus:2 fergus:2 terrain:1 don:1 search:8 latent:4 triplet:12 khosla:1 table:10 additionally:2 disambiguate:1 learn:5 ca:1 composing:1 improving:1 requested:1 ngiam:2 bottou:1 complex:5 zitnick:2 domain:3 protocol:2 did:1 joulin:1 dense:3 substituted:1 motivation:1 bounding:2 hyperparameters:1 turian:1 child:2 tesla:1 fair:1 complementary:1 body:1 i:1 embeds:3 momentum:1 explicit:2 wish:1 breaking:1 ib:3 jmlr:1 learns:7 young:5 grained:1 donahue:1 down:3 choi:3 embed:2 explored:1 list:2 svm:2 concern:2 consist:1 socher:14 workshop:1 false:2 effectively:2 margin:4 stratos:1 chen:2 backpropagate:1 generalizing:1 photograph:1 simply:1 appearance:1 likely:1 visual:13 contained:1 bo:1 springer:1 truth:4 relies:1 weston:1 conditional:1 sized:1 formulated:1 careful:1 shared:1 barnard:1 content:4 hard:1 lsvrc:1 flag:1 total:1 experimental:1 berg:10 people:3 support:5 violated:1 kulkarni:3 kuznetsova:1 incorporate:1 evaluate:1 audio:1 srivastava:1 |
4,729 | 5,282 | Recursive Context Propagation Network for Semantic
Scene Labeling
Abhishek Sharma
University of Maryland
College Park, MD
[email protected]
Oncel Tuzel
Ming-Yu Liu
Mitsubishi Electric Research Labs (MERL)
Cambridge, MA
{oncel,mliu}@merl.com
Abstract
We propose a deep feed-forward neural network architecture for pixel-wise semantic scene labeling. It uses a novel recursive neural network architecture for
context propagation, referred to as rCPN. It first maps the local visual features
into a semantic space followed by a bottom-up aggregation of local information
into a global representation of the entire image. Then a top-down propagation of
the aggregated information takes place that enhances the contextual information of
each local feature. Therefore, the information from every location in the image is
propagated to every other location. Experimental results on Stanford background
and SIFT Flow datasets show that the proposed method outperforms previous approaches. It is also orders of magnitude faster than previous methods and takes
only 0.07 seconds on a GPU for pixel-wise labeling of a 256 ? 256 image starting
from raw RGB pixel values, given the super-pixel mask that takes an additional
0.3 seconds using an off-the-shelf implementation.
1
Introduction
Semantic labeling aims at getting pixel-wise dense labeling of an image in terms of semantic concepts such as tree, road, sky, water, foreground objects etc. Mathematically, the problem can be
framed as a mapping from a set of nodes arranged on a 2D grid (pixels) to the semantic categories.
Typically, this task is broken down into two steps - feature extraction and inference. Feature extraction involves retrieving descriptive information useful for semantic labeling under varying illumination and view-point conditions. These features are generally color, texture or gradient based and
extracted from a local patch around each pixel. Inference step consists of predicting the labels of
the pixels using the extracted features. The rich diversity in the appearance of even simple concepts
(sky, water, grass) makes the semantic labeling very challenging. Surprisingly, human performance
is almost close to perfect on this task. This striking difference of performance has been a heated field
of research in vision community. Past experiences and recent research [1, 2, 3] have conclusively
established that the ability of humans to utilize the information from the entire image is the main
reason behind the large performance gap. Interestingly, [2, 3] have shown that human performance
in labeling a small local region (super-pixel) is worse than a computer when both are looking at only
that region of the image. Motivated from these observations, increasingly sophisticated inference
algorithms have been developed to utilize the information from the entire image. Conditional Random Fields (CRFs) [4] and Structured Support Vector Machines (SVMs) [5] are among the most
successful and widely used algorithms for inference.
We model the semantic labeling task as a mapping from the set of all pixels in an image I to the
corresponding label image Y. We have several design considerations: (1) the mapping should be
evaluated fast, (2) it should utilize the entire image such that every location influences the labeling of
every other location, (3) mapping parameters should be learned from the training data, (4) it should
scale to different image sizes. In addition, good generalization requires limiting the capacity of
1
Semantic labels:
SKY
WATER
BLDG
BOAT
TREE
Figure 1: Conceptual illustration of recursive context propagation network (rCPN). rCPN recursively aggregates contextual information from local neighborhoods to the entire image and then
disseminates global context information back to individual local features. In this example, starting
from confusion between boat and building, the propagated context information helps resolve the
confusion by using the feature of the water segment.
the mapping while still utilizing the entire image information at once. For example, a simple fullyconnected-linear mapping from I to Y requires 4 Trillion parameters for an image of size 256?256,
but it will fail to generalize under practical conditions of limited training data.
Considering the requirements discussed above, we designed the mapping as a single feed-forward
neural network with carefully controlled capacity by parameter sharing. All the network parameters
are learned from the data and the feed-forward structure allows fast inference. The proposed network
can be functionally partitioned into two sub-networks: local feature extraction and recursive context
propagation.
As the name implies, local-feature extraction refers to the extraction of pixel- or region-wise visual
features for semantic labeling. We used the multi scale convolutional neural network (Multi-CNN)
architecture proposed in [6] to get pixel-wise features. Convolutional structure with shared parameters brings down the number of parameters for local feature extraction.
We propose a novel recursive context propagation network (rCPN), which, starting from the local
features, recursively aggregates contextual information from local neighborhoods up to the entire
image and then disseminates the aggregated information back to individual local features for better
semantic classification. rCPN is a recursive neural network with shared parameters through the parse
tree hierarchy. A conceptual illustration of this network is given in Figure 1. The scene consists of
three segments corresponding to a boat, a tree and a water/sky region. The nodes of the graph
(formed by a binary parse tree and its inversion) represent semantic description of the segments.
The distributions on the left are probable label distributions for the adjacent segments based on their
appearance. Initially (at the bottom), the boat can be confused as a white building, while looking
only at the bottom-left segment. The rCPN recursively combines two segment descriptions and
produces the semantic description of the combined segment. For example, as the tree is combined
with the boat, the belief that the combined segment includes a building increased since usually they
appear together in the images. Similarly, after we merge the water/sky segment description with
this segment description, the probability of the boat increased since the simultaneous occurrence
of water and building is rare. The middle node in the graph (root node of the segmentation tree)
2
corresponds to the semantic description of the entire image. After all the segment descriptions are
merged into a single holistic description of the entire image, this information is propagated to the
local regions. It is achieved by recursive updates of the semantic descriptions of the segments given
the descriptions of their parent segments. Finally, contextually enhanced descriptions of the leaf
nodes are used to label the segments. Note that, rCPN propagates segment semantic descriptions but
not the label distributions shown in the illustration.
Our work is influenced by Socher et al.?s work [7] that learns a non-linear mapping from feature
space to a semantic space, termed as semantic mapping. It is learned by optimizing a structure
prediction cost on the ground-truth parse trees of training images or sentences. Next, a classifier
is learned on the semantic mappings of the individual local features from the training images. At
test time, local features are projected to the semantic space using the learned semantic mapping
followed by classification. Therefore, only the information contained in each individual local feature
is used for labeling. In contrast, we use recursive bottom-top-bottom paths on randomly generated
parse trees to propagate contextual information from local regions to all other regions in the image.
Therefore, our approach uses entire image information for labeling each local region. Please see
experiments section for detailed comparison.
The main contributions of the proposed approach are:
? The proposed model is scalable. It is a combination of a CNN and a recursive neural network which is trained without using any human-designed features. In addition, convolution
and recursive structure allows scaling to arbitrary image sizes while still utilizing the entire
image content at once.
? We achieved state-of-the-art labeling accuracy on two important benchmarks while being
an order of magnitude faster than the existing methods due to feed-forward operations. It
takes only 0.07 seconds on GPU and 0.8 seconds on CPU for pixel-wise semantic labeling
of a 256 ? 256 image, with a given super-segmentation mask, that can be computed using
an off-the-shelf algorithm within 0.3 second.
? Proposed rCPN module can be used in conjunction with pre-computed features to propagate context information through the structure of an image (see experiments section) and
potentially for other structured prediction tasks.
2
Semantic labeling architecture
In this section we describe our semantic labeling architecture and discuss the design choices for
practical considerations. An illustration of this architecture is shown in Figure 2. The input image is
fed to a CNN, FCN N , which extracts local features per pixel. We then use a super-pixel tessellation
of the input image and average pool the local features within the same super-pixel. Following, we
use the proposed rCPN to recursively propagate the local information throughout the image using a
parse tree hierarchy and finally label the super-pixels.
2.1
Local feature extraction
We used the multi scale CNN architecture proposed in Farabet et al. [6] for extracting per pixel local
features. This network has three convolutional stages which are organized as 8 ? 8 ? 16 conv ?
2 ? 2 maxpool ? 7 ? 7 ? 64 conv ? 2 ? 2 maxpool ? 7 ? 7 ? 256 conv configuration, each maxpooling is non-overlapping. After each convolution we apply a rectified linear (ReLU) nonlinearity.
Unlike [6], we do not preprocess the input raw RGB images other than scaling it between 0 to 1,
and centering by subtracting 0.5. Tied filters are applied separately at three scales of the Gaussian
pyramid. The final feature maps at lower scales are spatially scaled up to the size of the feature map
at the highest scale and concatenated to get 256 ? 3 = 768 dimensional features per pixel. The
obtained pixel features are fed to a Softmax classifier for final classification. Please refer to [6] for
more details. After training, we drop the final Softmax layer and use the 768 dimensional features
as local features.
Note that the 768 dimensional concatenated output feature map is still 1/4th of the height and width
of the input image due to the max-pooling operations. To obtain the input size per pixel feature map
we either (1) shift the input image by one pixel on a 4 ? 4 grid to get 16 output feature maps that are
3
?
?
??
????
????
+
?1
????
??
????
??
????
?2
??
?12
????
????
?34
????
????
?????
????
?1
????
?12
SKY
BLDG
TREE
?34
??
WATER
BOAT
????
super-pixels
Local Feature Extraction
Recursive Context Propagation Network (rCPN)
Figure 2: Overview of semantic scene labeling architecture
combined to get the full-resolution image, or (2) scale-up each feature map by a factor of 4 using
bilinear interpolation. We refer to the later strategy as fast feature map generation in experiments
section.
Super-pixel representation: Although it is possible to do per pixel classification using the rCPN,
learning such a model is computationally intensive and the resulting network is too deep to propagate the gradients efficiently due to recursion. To reduce the complexity, we utilize a super-pixel
segmentation algorithm [8], which provides the desired number of super-pixels per image. This algorithm uses pairwise color similarity together with an entropy rate criteria to produce homogenous
super-pixels with roughly equal sizes. We average pool the local features within the same superpixel and retrieve s local features, {vi }i=1...s , one per super-pixel. In our experiments we used
s = 100 super-pixels per image.
2.2
Recursive context propagation network
rCPN consists of four neural networks: Fsem maps local features to the semantic space in which
the local information is propagated to other segments; Fcom recursively aggregates local information from smaller segments to larger segments through a parse tree hierarchy to capture a holistic
description of the image; Fdec recursively disseminates the holistic description to smaller segments
using the same parse tree; and Flab classifies the super-pixels utilizing the contextually enhanced
features.
Parse tree synthesis: Both for training and inference, the binary parse trees that are used for propagating information through the network are synthesized at random. We used a simple agglomerative
algorithm to synthesize the trees by combining sub-trees (starting from a single node) according to
the neighborhood information. To reduce the complexity and avoid degenerate solutions, the synthesis algorithm favors roughly balanced parse trees by greedily selecting sub-trees with smaller
heights at random. Note that, we use parse trees only as a tool to propagate the contextual information throughout the image. Therefore, we are not limited to the parse trees that represent an accurate
hierarchical segmentation of the image unlike [9, 7].
Semantic mapping network is a feed-forward neural network which maps the local features to the
dsem dimensional semantic vector space
xi = Fsem (vi ; ?sem ),
(1)
where ?sem is the model parameter. The aim of the semantic features is to capture a joint representation of the local features and the context, and being able to propagate this information through a
parse tree hierarchy to other super-pixels.
Combiner network is a recursive neural network which recursively maps the semantic features of
two child nodes (super-pixels) in the parse tree to obtain the semantic feature of the parent node
(combination of the two child nodes)
xi,j = Fcom ([xi , xj ]; ?com ).
4
(2)
Intuitively, combiner network attempts to aggregate the semantic content of the children nodes such
that the parent node becomes representative of its children. The information is recursively aggregated bottom-up from super-pixels to the root node through the parse tree. The semantic features of
the root node correspond to the holistic description of the entire image.
Decombiner network is a recursive neural network which recursively disseminates the context
information from the parent nodes to the children nodes throughout the parse tree hierarchy. This
network maps the semantic features of the child node and its parent to the contextually enhanced
feature of the child node
? i = Fdec ([?
x
xi,j , xi ]; ?dec ).
(3)
Since we start from the root feature of the entire image and apply the decombiner network top-down
recursively until we reach the super-pixel features, every super-pixel feature contains the contextual
information aggregated from the entire image. Therefore, it is influenced by every other super-pixel
in the image.
Labeler network is the final feed forward network which maps the contextually enhanced semantic
features (?
xi ) of each super-pixel to one of the semantic category labels
yj = Flab (?
xi ; ?lab ).
(4)
Contextually enhanced features contain both local and global context information, thereby leading
to better classification.
Side information: It is possible to input information to the recursive networks not only at the leaf
nodes but also at any level of the parse tree. The side information can encode the static knowledge
about the parse tree nodes and is not recurred through the tree. In our implementation we used
average x and y locations of the nodes and their sizes as the side information.
3
Learning
Proposed labeling architecture is a feed-forward neural network that can be trained end-to-end.
However, the recursion makes the depth of the neural network too deep for an efficient joint training.
Therefore, we first learn the CNN parameters (?CN N ) using the raw image and the ground truth per
pixel labels. The trained CNN model is used to extract super-pixel features followed by training of
rCPN (?rCP N = [?sem , ?com , ?dec , ?lab ]) to predict the ground truth super-pixel labels.
Feature extractor CNN is trained on a GPU using a publicly available implementation CAFFE [10].
In order to avoid over-fitting we used data augmentation and dropout. All the training images were
flipped horizontally to get twice the original images. We used dropout in the last layer with dropout
ratio equal to 0.5. Standard back-propagation for CNN is used with stochastic gradient descent
update scheme on mini-batches of 6 images, with weight decay (? = 5 ? 10?5 ) and momentum
(? = 0.9). It typically took 6-8 hours of training on a GPU as compared to 3-5 days training on a
CPU as reported in [6]. We found that simply using RGB images with ReLU units and dropout gave
slightly better pixel-wise accuracy as compared to [6].
rCPN parameters are trained using back-propagation through structure [11], which back-propagates
the error through the parse tree, from Flab to Fsem . The basic idea is to split the error message
at each node and propagate it to the children nodes. Limited memory BFGS [12] with line-search
is used for parameter updates using publicly available implementation1 . From each super-pixel we
obtained 5 different features by average pooling a random subset of pixels within the super-pixel (as
opposed to average pooling all the pixels), and used a different random parse tree for each set of
random feature, thus we increased our training data by a factor of 5. It typically took 600 to 1000
iterations for complete training.
4
Experiments
We extensively tested the proposed model on two widely used datasets for semantic scene labeling Stanford background [13] and SIFT Flow [14]. Stanford background dataset contains 715 color
images of outdoor scenes, it has 8 classes and the images are approximately 240 ? 320 pixels.
1
http://www.di.ens.fr/?mschmidt/Software/minFunc.html
5
We used the 572 train and 143 test image split provided by [7] for reporting the results. SIFT
Flow contains 2688, 256 ? 256 color images with 33 semantic classes. We experimented with the
train/test (2488/200) split provided by the authors of [15]. We have used three evaluation metrics Per pixel accuracy (PPA): ratio of the correct pixels to the total pixels in the test images; Mean
class accuracy (MCA): mean of the category-wise pixel accuracy; Time per image (Time): time
required to label an input image starting from the raw image input, we report our time on both GPU
and CPU.
The local feature extraction through Multi-CNN [6] encodes contextual information due to large
field of view (FOV); the FOV for 1, 1/2 and 1/4 scaled input images is 47 ? 47, 94 ? 94 and
188 ? 188 pixels, respectively. Therefore, we designed the experiments under single and multi scale
settings to assess rCPN?s contribution. Mutli-CNN + rCPN refers to the case where feature maps
from all the three scales (1,1/2 and 1/4), 3 ? 256 = 768 dimensional local feature, for each pixel
are used. Single-CNN + rCPN refers to the case where only the 256 feature maps corresponding
to the original resolution image are used. Evidently, the amount of contextual information in the
local features of Single-CNN is significantly lesser than that of Multi-CNN because of smaller FOV.
All the individual modules in rCPN, Fsem , Fcom , Fdec and Flab , are single layer neural networks
with ReLU non-linearity and dsem = 60 for all the experiments. We used 20 randomly generated
parse trees for each image and used voting for the final super-pixel labels. We did not optimize
these hyper-parameters and believe that parameter-tuning can further increase the performance. The
baseline is two-layer neural network with 60 hidden neurons classifier with Single-CNN or MultiCNN features of super-pixels and referred to as Multi/Single-CNN + Plain NN.
4.1
SIFT Flow dataset
We used 100 super-pixels per image obtained by method of [8]. The result on SIFT Flow database is
shown in Table 1. From the comparison it is clear that we outperform all the other previous methods
on pixel accuracy while being an order of magnitude faster. Farabet et al. [6] improved the mean
class accuracy by training a model based on the balanced class frequency. Since some of the classes
in SIFT Flow dataset are under represented, the class accuracies for them are very low. Therefore,
following [6], we also trained a balanced rCPN model that puts more weights on the errors for rare
classes as compared to the dominant ones, referred to as Multi-CNN + rCPN balanced. Smoothed
inverse frequency of the pixels of each category is used as the weights. Balanced training helped
improve our mean class accuracy from 33.6 % to 48.0 %, which is still slightly worse than [6] (48.0
% vs 50.8 %), but our pixel accuracy is higher (75.5 % vs 72.3 %). Multi-CNN + rCPN performed
better than Single-CNN + rCPN and both performed significantly better than Plain NN approaches,
because the later approaches do not utilize global contextual information. We also observed that the
relative improvement over Plain NN was more with Single-CNN features which uses less context
information than that of Multi-CNN.
4.2
Stanford background dataset
We used publicly available super-pixels provided by [7] with our CNN based local features to obtain
super-pixel features. A comparison of our results with previous approaches on Stanford background
database is shown in Table 2. We outperform previous approaches on all the performance metrics.
Interestingly, we observe that Single-CNN + rCPN performs better than Multi-CNN + rCPN for
pixel accuracy. We believe that it is due to over-fitting on high-dimensional Multi-CNN features
and relatively smaller training data size with only 572 images. Once again the improvement due to
rCPN over plain NN is more prominent in the case of Single-CNN features.
Model analysis: In this section, we analyze the performance of individual components of the proposed model. First, we use rCPN with hand-designed features to evaluate the performance of context
model alone, beyond the learned local features using CNN. We utilize the visual features and superpixels used in semantic mapping and CRF labeling framework [7, 13], and trained our rCPN module.
The results are presented in Table 3. We see that rCPN module significantly improves upon the existing context models, namely a CRF model used in [13] and semantic space proposed in [7]. In
addition, CNN based visual features improve over the hand-designed features.
Next, we analyze the performance of combiner and decombiner networks separately. To evaluate combiner network in isolation, we first obtain the semantic mapping (xi ) of each super-pixel?s
6
Table 1: SIFT Flow result
Table 2: Stanford background result
Method
PPA
MCA
Tighe, [15]
Liu, [14]
Singh, [16]
Eigen, [17]
Farabet, [6]
(Balanced), [6]
Tighe, [18]
Pinheiro, [19]
Single-CNN +
Plain NN
Multi-CNN +
Plain NN
Single-CNN +
rCPN
Multi-CNN +
rCPN
Multi-CNN +
rCPN Balanced
Multi-CNN +
rCPN Fast
77.0
76.7
79.2
77.1
78.5
72.3
78.6
77.7
30.1
NA
33.8
32.5
29.6
50.8
39.2
29.8
Time (s)
CPU/GPU
8.4 / NA
31 / NA
20 / NA
16.6 / NA
NA / NA
NA / NA
? 8.4 / NA
NA / NA
72.8
25.5
5.1/0.5
76.3
32.1
13.1/1.4
77.2
25.5
5.1/0.5
79.6
33.6
13.1/1.4
75.5
48.0
13.1/1.4
79.5
33.4
1.1/0.37
Method
PPA
MCA
Gould, [13]
Munoz, [20]
Tighe, [15]
Kumar, [21]
Socher, [7]
Lempitzky, [9]
Singh, [16]
Farabet, [6]
Eigen, [17]
Pinheiro, [19]
Single-CNN +
Plain NN
Multi-CNN +
Plain NN
Single-CNN +
rCPN
Multi-CNN +
rCPN
Multi-CNN +
rCPN Fast
76.4
76.9
77.5
79.4
78.1
81.9
74.1
81.4
75.3
80.2
NA
NA
NA
NA
NA
72.4
62.2
76.0
66.5
69.9
Time (s)
CPU/GPU
30 to 600 / NA
12 / NA
4 / NA
? 600 / NA
NA / NA
? 60 / NA
20 / NA
60.5 / NA
16.6 / NA
10.6 / NA
80.1
69.7
5.1/0.5
80.9
74.4
13.1/1.4
81.9
73.6
5.1/0.5
81.0
78.8
13.1/1.4
80.9
78.8
1.1/0.37
Table 3: Stanford hand-designed local feature
Method
PPA
2-layer NN [7]
76.1
CRF [13]
76.4
Semantic space [7]
78.1
proposed rCPN
81.4
visual feature using rCPN?s Fsem and append to it the root feature of the entire image to obtain
xcom
= [xi , xroot ]. Then we train a separate Softmax classifier on xcom
. This resulted in better peri
i
formance for both Single-scale (PPA: 80.4 & MCA: 71.5) and Multi-scale (PPA: 80.8 & MCA: 79.1)
CNN feature settings over (Single/Multi)-CNN + Plain NN. As was previously shown in Table 2,
decombiner network further improves this model.
Computation speed: Our fast method (Section 2.1) takes only 0.37 seconds (0.3 for super-pixel
segmentation, 0.06 for feature extraction and 0.01 for rCPN and labeling) to label a 256 ? 256 image
starting from the raw RGB image on a GTX Titan GPU and 1.1 seconds on a Intel core i7 CPU. In
both of the experiments the performance loss is negligible using the fast method. Interestingly, the
time bottleneck of our approach on a GPU is the super-pixel segmentation time.
Several typical labeling results on Stanford background dataset using the proposed semantic scene
labeling algorithm are shown in Figure 3.
5
Related Work
Scene labeling has two broad categories of approaches - non-parametric and model-based. Recently,
many non-parametric approaches for natural scene parsing have been proposed [15, 14, 16, 17, 18].
The underlying theme is to find similar looking images to the query image from a database of pixelwise labeled images, followed by super-pixel label transfer from the retrieved images to the query
image. Finally, a structured prediction model such as CRF is used to integrate contextual information
to obtain the final labeling. These approaches mainly differ in the retrieval of candidate images or
super-pixels, transfer of label from the retrieved candidates to the query image, and the form of the
structured prediction model used for final labeling. They are based on nearest-neighbor retrieval that
introduces a performance/accuracy tradeoff. The variations present in natural scene images are large
and it is very difficult to cover this entire space of variation with a reasonable size database, which
limits the accuracy of these methods. On the other extreme, a very large database would require
large retrieval-time, which limits the scalability of these methods.
7
sky
tree
road
grass
water
bldg
mntn
fig obj
Figure 3: Typical labeling results on Stanford background dataset using our method
Model-based approaches learn the appearance of semantic categories and relations among them
using a parametric model. In [13, 20, 2, 3, 22], CRF models are used to combine unary potentials
devised through the visual features extracted from super-pixels with the neighborhood constraints.
The differences are mainly in terms of the visual features, unary potentials and the structure of
the CRF model. Lempitsky et al. [9] have formulated a joint-CRF on multiple levels of an image
segmentation hierarchy to achieve better results than a flat-CRF on the image super-pixels only.
Socher et al. [7] learnt a mapping from the visual features to a semantic space followed by a twolayer neural network for classification. Better use of contextual information, with the same superpixels and features, increased the performance on Stanford background dataset from CRF based
method of Gould et al. to semantic mapping of Socher et al. to the proposed work (76.4% ?
78.1% ? 81.4%). It indicates that neural network based models have the potential to learn more
complicated interactions than a CRF. Moreover, they have the advantage of being extremely fast, due
to the feed-forward nature. Farabet et al. [6] proposed to learn the visual features from image/label
training pairs using a multi-scale convolutional neural network. They reported state-of-the-art results
on various datasets using gPb, purity-cover and CRF on top of their learned features. Pinheiro et al.
[19] extended their work by feeding in the per-pixel predicted labels using a CNN classifier to the
next stage of the same CNN classifier. However, their propagation structure is not adaptive to the
image content and only propagating label information did not improve much over the prior work.
Similar to these methods, we also make use of the Multi-CNN module to extract local features in
our pipeline. However, our novel context propagation network shows that propagating semantic
representation bottom-up and top-down using a parse three hierarchy is a more effective way to
aggregate global context information. Please see Tables 1 and 2 for a detailed comparison of our
method with the methods discussed above.
CRFs model the joint distribution of the output variables given the observations and can include
higher order potentials in addition to the unary potentials. Higher order potentials allow these models to represent the dependencies between the output variables, which is important for structured
prediction tasks. On the downside, except for a few exceptions such as non-loopy models, inference
in these models is NP-Hard that can be only approximately solved and is time consuming. Moreover,
parameter learning procedures that are tractable usually limit the form of the potential functions to
simple forms such as linear models. In contrast, in our model, we can efficiently learn complex
relations between a single output variable and all the observations from an image, allowing a large
context to be considered effectively. Additionally, the inference procedure is a simple feed-forward
pass that can be performed very fast. However, the form of our function is still a unary term and our
model cannot represent higher order output dependencies. Our model can also be used to obtain the
unary potential for a structured inference model.
6
Conclusion
We introduced a novel deep neural network architecture, which is a combination of a convolutional
neural network and recursive neural network, for semantic scene labeling. The key contribution
is the recursive context propagation network, which effectively propagates contextual information
from one location of the image to other locations in a feed-forward manner. This structure led to
the state-of-the-art semantic scene labeling results on Stanford background and SIFT Flow datasets
with very fast processing speed. Next we plan to scale-up our model for recently introduced large
scale learning task [23].
8
References
[1] A. Torralba, K.P. Murphy, W.T. Freeman, and M.A. Rubin. Context-based vision system for
place and object recognition. IEEE CVPR, 2003.
[2] Roozbeh Mottaghi, Sanja Fidler, Jian Yao, Raquel Urtasun, and Devi Parikh. Analyzing semantic segmentation using hybrid human-machine crfs. IEEE CVPR, 2013.
[3] Roozbeh Mottaghi, Xianjie Chen, Xiaobai Liu, Nam-Gyu Cho, Seong-Whan Lee, Sanja Fidler, Raquel Urtasun, and Alan Yuille. The role of context for object detection and semantic
segmentation in the wild. IEEE CVPR, 2014.
[4] John D. Lafferty, Andrew McCallum, and Fernando C. N. Pereira. Conditional random fields:
Probabilistic models for segmenting and labeling sequence data. ICML, pages 282?289, 2001.
[5] Ioannis Tsochantaridis, Thorsten Joachims, Thomas Hofmann, Yasemin Altun, and Yoram
Singer. Large margin methods for structured and interdependent output variables. Journal of
Machine Learning Research, 6(2):1453, 2006.
[6] Clement Farabet, Camille Couprie, Laurent Najman, and Yann LeCun. Learning hierarchical
features for scene labeling. IEEE TPAMI, August 2013.
[7] Richard Socher, Cliff Chiung-Yu Lin, Andrew Y. Ng, and Christopher D. Manning. Parsing
natural scenes and natural language with recursive neural networks. ICML, 2011.
[8] Ming-Yu Liu, Oncel Tuzel, Srikumar Ramalingam, and Rama Chellappa. Entropy rate superpixel segmentation. IEEE CVPR, 2011.
[9] V. Lempitsky, A. Vedaldi, and A. Zisserman. A pylon model for semantic segmentation. NIPS,
2011.
[10] Yangqing Jia. Caffe: An open source convolutional architecture for fast feature embedding.
http://caffe.berkeleyvision.org/, 2013.
[11] Christoph Goller and Andreas Kchler. Learning task-dependent distributed representations by
backpropagation through structure. Int Conf. on Neural Network, 1995.
[12] Dong C. Liu, Jorge Nocedal, and Dong C. On the limited memory bfgs method for large scale
optimization. Mathematical Programming, 45:503?528, 1989.
[13] Stephen Gould, Richard Fulton, and Daphne Koller. Decomposing a scene into geometric and
semantically consistent regions. IEEE ICCV, 2009.
[14] Ce Liu, Jenny Yuen, and Antonio Torralba. Nonparametric scene parsing via label transfer.
IEEE TPAMI, 33(12), Dec 2011.
[15] Joseph Tighe and Svetlana Lazebnik. Superparsing: Scalable nonparametric image parsing
with superpixels. IJCV, 101:329?349, 2013.
[16] Gautam Singh and Jana Kosecka. Nonparametric scene parsing with adaptive feature relevance
and semantic context. IEEE CVPR, 2013.
[17] R. Fergus and D. Eigen. Nonparametric image parsing using adaptive neighbor sets. IEEE
CVPR, 2012.
[18] Joseph Tighe and Svetlana Lazebnik. Finding things: Image parsing with regions and perexemplar detectors. IEEE CVPR, 2013.
[19] Pedro H. O. Pinheiro and Ronan Collobert. Recurrent convolutional neural networks for scene
parsing. ICML, 2014.
[20] Daniel Munoz, J. Andrew Bagnell, and Martial Hebert. Stacked hierarchical labeling. ECCV,
2010.
[21] M. Pawan Kumar and Daphne Koller. Efficiently selecting regions for scene understanding.
IEEE CVPR, 2010.
[22] Gungor Polatkan and Oncel Tuzel. Compressed inference for probabilistic sequential models.
UAI, pages 609?618, 2011.
[23] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan,
Piotr Doll?ar, and C Lawrence Zitnick. Microsoft coco: Common objects in context. ECCV,
2014.
9
| 5282 |@word cnn:43 middle:1 inversion:1 open:1 mitsubishi:1 rgb:4 propagate:7 twolayer:1 thereby:1 recursively:10 liu:6 configuration:1 contains:3 selecting:2 daniel:1 interestingly:3 outperforms:1 past:1 existing:2 com:3 contextual:12 gpu:9 parsing:8 john:1 ronan:1 hofmann:1 designed:6 drop:1 update:3 grass:2 v:2 alone:1 leaf:2 mccallum:1 core:1 provides:1 node:22 location:7 gautam:1 org:1 daphne:2 height:2 mathematical:1 retrieving:1 consists:3 ijcv:1 combine:2 fullyconnected:1 fitting:2 wild:1 manner:1 pairwise:1 mask:2 roughly:2 multi:23 freeman:1 ming:2 resolve:1 cpu:6 xroot:1 considering:1 conv:3 confused:1 classifies:1 becomes:1 provided:3 linearity:1 underlying:1 moreover:2 developed:1 finding:1 sky:7 every:6 rcp:1 voting:1 classifier:6 scaled:2 unit:1 ramanan:1 appear:1 segmenting:1 negligible:1 local:41 limit:3 bilinear:1 analyzing:1 cliff:1 laurent:1 path:1 interpolation:1 merge:1 approximately:2 twice:1 fov:3 challenging:1 christoph:1 contextually:5 limited:4 practical:2 lecun:1 yj:1 recursive:18 backpropagation:1 procedure:2 maire:1 tuzel:3 polatkan:1 significantly:3 vedaldi:1 pre:1 road:2 refers:3 altun:1 get:5 cannot:1 close:1 tsochantaridis:1 put:1 context:24 influence:1 heated:1 www:1 optimize:1 map:15 crfs:3 starting:6 resolution:2 utilizing:3 nam:1 pylon:1 retrieve:1 embedding:1 variation:2 limiting:1 hierarchy:7 enhanced:5 programming:1 us:4 superpixel:2 ppa:6 synthesize:1 recognition:1 srikumar:1 database:5 labeled:1 bottom:7 observed:1 module:5 role:1 solved:1 capture:2 region:11 highest:1 balanced:7 broken:1 complexity:2 gpb:1 trained:7 singh:3 deva:1 segment:19 yuille:1 upon:1 joint:4 represented:1 various:1 train:3 stacked:1 fast:11 describe:1 effective:1 chellappa:1 query:3 labeling:33 aggregate:5 hyper:1 neighborhood:4 caffe:3 stanford:11 widely:2 larger:1 cvpr:8 compressed:1 ability:1 favor:1 final:7 descriptive:1 advantage:1 evidently:1 sequence:1 tpami:2 took:2 propose:2 subtracting:1 interaction:1 fr:1 combining:1 holistic:4 degenerate:1 achieve:1 description:15 scalability:1 getting:1 parent:5 requirement:1 produce:2 perfect:1 object:4 help:1 rama:1 andrew:3 recurrent:1 propagating:3 nearest:1 c:1 involves:1 implies:1 predicted:1 differ:1 merged:1 correct:1 filter:1 stochastic:1 human:5 require:1 feeding:1 generalization:1 yuen:1 probable:1 mathematically:1 around:1 considered:1 ground:3 lawrence:1 mapping:16 predict:1 torralba:2 label:19 tool:1 gaussian:1 super:37 aim:2 avoid:2 shelf:2 varying:1 conjunction:1 combiner:4 encode:1 joachim:1 improvement:2 indicates:1 mainly:2 superpixels:3 contrast:2 greedily:1 baseline:1 inference:10 dependent:1 nn:10 unary:5 entire:16 typically:3 initially:1 hidden:1 relation:2 perona:1 koller:2 pixel:70 among:2 classification:6 html:1 plan:1 art:3 softmax:3 homogenous:1 field:4 once:3 equal:2 extraction:10 ng:1 piotr:1 labeler:1 flipped:1 park:1 yu:3 broad:1 icml:3 foreground:1 fcn:1 report:1 np:1 richard:2 few:1 randomly:2 resulted:1 individual:6 murphy:1 pawan:1 microsoft:1 attempt:1 detection:1 message:1 evaluation:1 introduces:1 extreme:1 behind:1 accurate:1 experience:1 tree:32 desired:1 minfunc:1 merl:2 increased:4 fdec:3 downside:1 cover:2 ar:1 tessellation:1 loopy:1 cost:1 subset:1 rare:2 successful:1 goller:1 too:2 pixelwise:1 reported:2 dependency:2 learnt:1 combined:4 cho:1 peri:1 lee:1 off:2 probabilistic:2 dong:2 pool:2 maxpool:2 together:2 synthesis:2 michael:1 yao:1 na:28 augmentation:1 again:1 opposed:1 worse:2 conf:1 leading:1 potential:8 diversity:1 bfgs:2 ioannis:1 includes:1 int:1 titan:1 tighe:5 vi:2 collobert:1 tsung:1 later:2 view:2 root:5 lab:3 helped:1 performed:3 analyze:2 start:1 aggregation:1 complicated:1 jia:1 kosecka:1 ass:1 contribution:3 formed:1 accuracy:13 convolutional:7 publicly:3 formance:1 efficiently:3 implementation1:1 correspond:1 preprocess:1 serge:1 generalize:1 raw:5 disseminates:4 rectified:1 simultaneous:1 detector:1 influenced:2 reach:1 sharing:1 farabet:6 centering:1 frequency:2 james:1 di:1 static:1 propagated:4 dataset:7 color:4 knowledge:1 improves:2 segmentation:11 organized:1 sophisticated:1 carefully:1 back:5 feed:10 higher:4 day:1 zisserman:1 improved:1 roozbeh:2 arranged:1 evaluated:1 stage:2 until:1 hand:3 parse:22 christopher:1 overlapping:1 propagation:13 brings:1 believe:2 name:1 building:4 concept:2 contain:1 gtx:1 fidler:2 spatially:1 semantic:54 white:1 adjacent:1 width:1 mschmidt:1 please:3 berkeleyvision:1 jana:1 mutli:1 criterion:1 prominent:1 ramalingam:1 complete:1 crf:11 confusion:2 performs:1 image:80 wise:8 consideration:2 novel:4 recently:2 parikh:1 lazebnik:2 common:1 overview:1 discussed:2 functionally:1 synthesized:1 refer:2 cambridge:1 munoz:2 framed:1 tuning:1 clement:1 grid:2 similarly:1 nonlinearity:1 language:1 sanja:2 similarity:1 maxpooling:1 etc:1 dominant:1 recent:1 retrieved:2 optimizing:1 coco:1 termed:1 hay:1 binary:2 jorge:1 yi:1 mottaghi:2 yasemin:1 additional:1 mca:5 purity:1 sharma:1 aggregated:4 mntn:1 fernando:1 jenny:1 stephen:1 full:1 multiple:1 fcom:3 alan:1 faster:3 retrieval:3 lin:2 devised:1 controlled:1 prediction:5 scalable:2 basic:1 vision:2 metric:2 iteration:1 represent:4 pyramid:1 achieved:2 dec:3 background:10 addition:4 separately:2 jian:1 source:1 pinheiro:4 umd:1 unlike:2 pooling:3 thing:1 flow:8 lafferty:1 obj:1 extracting:1 split:3 xj:1 relu:3 gave:1 isolation:1 architecture:11 reduce:2 idea:1 cn:1 lesser:1 tradeoff:1 andreas:1 intensive:1 shift:1 i7:1 bottleneck:1 motivated:1 rcpn:38 deep:4 antonio:1 useful:1 generally:1 detailed:2 clear:1 amount:1 nonparametric:4 extensively:1 svms:1 category:6 http:2 outperform:2 per:13 key:1 four:1 yangqing:1 ce:1 utilize:6 nocedal:1 graph:2 pietro:1 inverse:1 bldg:3 striking:1 raquel:2 svetlana:2 place:2 almost:1 throughout:3 reporting:1 reasonable:1 yann:1 patch:1 scaling:2 dropout:4 layer:5 followed:5 constraint:1 scene:19 software:1 encodes:1 flat:1 seong:1 speed:2 extremely:1 kumar:2 relatively:1 gould:3 structured:7 according:1 combination:3 manning:1 smaller:5 slightly:2 oncel:4 increasingly:1 partitioned:1 joseph:2 intuitively:1 iccv:1 thorsten:1 pipeline:1 computationally:1 previously:1 discus:1 fail:1 singer:1 fed:2 tractable:1 end:2 available:3 operation:2 decomposing:1 doll:1 apply:2 observe:1 hierarchical:3 occurrence:1 batch:1 eigen:3 original:2 thomas:1 top:5 include:1 yoram:1 concatenated:2 strategy:1 parametric:3 md:1 fulton:1 bagnell:1 enhances:1 gradient:3 separate:1 maryland:1 capacity:2 agglomerative:1 urtasun:2 water:9 reason:1 illustration:4 ratio:2 mini:1 difficult:1 potentially:1 append:1 implementation:3 design:2 allowing:1 observation:3 convolution:2 datasets:4 neuron:1 benchmark:1 descent:1 najman:1 extended:1 looking:3 smoothed:1 arbitrary:1 august:1 camille:1 community:1 introduced:2 namely:1 required:1 pair:1 sentence:1 learned:7 established:1 hour:1 nip:1 able:1 beyond:1 usually:2 max:1 memory:2 belief:1 natural:4 hybrid:1 predicting:1 boat:7 recursion:2 scheme:1 improve:3 martial:1 extract:3 prior:1 interdependent:1 geometric:1 understanding:1 relative:1 loss:1 generation:1 integrate:1 consistent:1 propagates:3 rubin:1 eccv:2 surprisingly:1 last:1 hebert:1 side:3 allow:1 neighbor:2 distributed:1 depth:1 plain:9 rich:1 forward:10 author:1 adaptive:3 projected:1 conclusively:1 global:5 uai:1 conceptual:2 belongie:1 consuming:1 abhishek:1 xi:9 fergus:1 search:1 table:8 additionally:1 learn:5 transfer:3 nature:1 sem:3 complex:1 electric:1 zitnick:1 did:2 dense:1 main:2 child:8 fig:1 referred:3 representative:1 en:1 intel:1 sub:3 momentum:1 theme:1 pereira:1 candidate:2 outdoor:1 tied:1 extractor:1 learns:1 down:5 sift:8 decay:1 experimented:1 socher:5 sequential:1 effectively:2 texture:1 magnitude:3 illumination:1 superparsing:1 margin:1 gap:1 chen:1 entropy:2 led:1 simply:1 appearance:3 devi:1 visual:9 horizontally:1 contained:1 pedro:1 corresponds:1 truth:3 extracted:3 ma:1 trillion:1 conditional:2 lempitsky:2 formulated:1 couprie:1 shared:2 content:3 hard:1 typical:2 except:1 semantically:1 total:1 pas:1 experimental:1 exception:1 college:1 support:1 relevance:1 evaluate:2 tested:1 |
4,730 | 5,283 | Multiscale Fields of Patterns
John G. Oberlin
Brown University
Providence, RI 02906
john [email protected]
Pedro F. Felzenszwalb
Brown University
Providence, RI 02906
[email protected]
Abstract
We describe a framework for defining high-order image models that can be used
in a variety of applications. The approach involves modeling local patterns in a
multiscale representation of an image. Local properties of a coarsened image reflect non-local properties of the original image. In the case of binary images local
properties are defined by the binary patterns observed over small neighborhoods
around each pixel. With the multiscale representation we capture the frequency
of patterns observed at different scales of resolution. This framework leads to
expressive priors that depend on a relatively small number of parameters. For inference and learning we use an MCMC method for block sampling with very large
blocks. We evaluate the approach with two example applications. One involves
contour detection. The other involves binary segmentation.
1
Introduction
Markov random fields are widely used as priors for solving a variety of vision problems such as
image restoration and stereo [5, 8]. Most of the work in the area has concentrated on low-order
models involving pairs of neighboring pixels. However, it is clear that realistic image priors need to
capture higher-order properties of images.
In this paper we describe a general framework for defining high-order image models that can be used
in a variety of applications. The approach involves modeling local properties in a multiscale representation of an image. This leads to a natural low-dimensional representation of a high-order model.
We concentrate on the problem of estimating binary images. In this case local image properties can
be captured by the binary patterns in small neighborhoods around each pixel.
We define a Field of Patterns (FoP) model using an energy function that assigns a cost to each 3x3
pattern observed in an image pyramid. The cost of a pattern depends on the scale where it appears.
Figure 1 shows a binary image corresponding to a contour map from the Berkeley segmentation
dataset (BSD) [12, 2] and a pyramid representation obtained by repeated coarsening. The 3x3 patterns we observe after repeated coarsening depend on large neighborhoods of the original image.
These coarse 3x3 patterns capture non-local image properties. We train models using a maximumlikelihood criteria. This involves selecting pattern costs making the expected frequency of patterns
in a random sample from the model match the average frequency of patterns in the training images.
Using the pyramid representation the model matches frequencies of patterns at each resolution.
In practice we use MCMC methods for inference and learning. In Section 3 we describe an MCMC
sampling algorithm that can update a very large area of an image (a horizontal or vertical band of
pixels) in a single step, by combining the forward-backward algorithm for one-dimensional Markov
models with a Metropolis-Hastings procedure.
We evaluated our models and algorithms on two different applications. One involves contour detection. The other involves binary segmentation. These two applications require very different image
priors. For contour detection the prior should encourage a network of thin contours, while for bi1
(a)
(b)
(c)
Figure 1: (a) Multiscale/pyramid representation of a contour map. (b) Coarsest image scaled up
for better visualization, with a 3x3 pattern highlighted. The leftmost object in the original image
appears as a 3x3 ?circle? pattern in the coarse image. (c) Patches of contour maps (top) that coarsen
to a particular 3x3 pattern (bottom) after reducing their resolution by a factor of 8.
nary segmentation the prior should encourage spatially coherent masks. In both cases we can design
effective models using maximum-likelihood estimation.
1.1
Related Work
FRAME models [24] and more recently Fields of Experts (FoE) [15] defined high-order energy
models using the response of linear filters. FoP models are closely related. The detection of 3x3
patterns at different resolutions corresponds to using non-linear filters of increasing size. In FoP we
have a fixed set of pre-defined non-linear filters that detect common patterns at different resolutions.
This avoids filter learning, which leads to a non-convex optimization problem in FoE.
A restricted set of 3x3 binary patterns was considered in [6] to define priors for image restoration.
Binary patterns were also used in [17] to model curvature of a binary shape. There has been recent
work on inference algorithms for CRFs defined by binary patterns [19] and it may be possible to
develop efficient inference algorithms for FoP models using those techniques.
The work in [23] defined a variety of multiresolution models for images based on a quad-tree representation. The quad-tree leads to models that support efficient learning and inference via dynamic
programming, but such models also suffer from artifacts due to the underlying tree-structure. The
work in [7] defined binary image priors using deep Boltzmann machines. Those models are based
on a hierarchy of hidden variables that is related to our multiscale representation. However in our
case the multiscale representation is a deterministic function of the image and does not involve extra
hidden variables as [7]. The approach we take to define a multiscale model is similar to [9] where
local properties of subsampled signals where used to model curves.
One of our motivating applications involves detecting contours in noisy images. This problem has a
long history in computer vision, going back at least to [16], who used a type of Markov model for
detecting salient contours. Related approaches include the stochastic completion field in [22, 21],
spectral methods [11], the curve indicator random field [3], and the more recent work in [1].
2
Fields of Patterns (FoP)
Let G = [n] ? [m] be the grid of pixels in an n by m image. Let x = {x(i, j) | (i, j) ? G} be a
hidden binary image and y = {y(i, j) | (i, j) ? G} be a set of observations (such as a grayscale or
color image). Our goal is to estimate x from y.
We define p(x|y) using an energy function that is a sum of two terms,
1
exp(?E(x, y)) E(x, y) = EFoP (x) + Edata (x, y)
p(x|y) =
Z(y)
2
(1)
It is sometimes useful to think of EFoP (x) as a model for binary images and Edata (x, y) as a data
model even though technically there is no such distinction in a conditional model.
2.1
Singlescale FoP Model
The singlescale FoP model is one of the simplest energy models that can capture the basic properties
of contour maps or other images that contain thin objects. We use x[i, j] to denote the binary pattern
defined by x in the 3x3 window centered at pixel (i, j), treating values outside of the image as 0. A
singlescale FoP model is defined by the local patterns in x,
X
EFoP (x) =
V (x[i, j]).
(2)
(i,j)?G
Here V is a potential function assigning costs (or energies) to binary patterns. Note that there are 512
possible binary patterns in a 3x3 window. We can make the model invariant to rotations and mirror
symmetries by tying parameters together. The resulting model has 102 parameters (some patterns
have more symmetries than others) and can be learned from smaller datasets. We used invariant
models for all of the experiments reported in this paper.
2.2
Multiscale FoP Model
To capture non-local statistics we look at local patterns in a multiscale representation of x. For a
model with K scales let ?(x) = x0 , . . . , xK?1 be an image pyramid where x0 = x and xk+1 is
a coarsening of xk . Here xk is a binary image defined over a grid G k = [n/2k ] ? [m/2k ]. The
coarsening we use in practice is defined by a logical OR operation,
xk+1 (i, j) = xk (2i, 2j) ? xk (2i + 1, 2j) ? xk (2i, 2j + 1)k ? xk (2i + 1, 2j + 1)
(3)
This particular coarsening maps connected objects at one scale of resolution to connected objects at
the next scale, but other coarsenings may be appropriate in different applications.
A multiscale FoP model is defined by the local patterns in ?(x),
EFoP (x) =
K?1
X
X
V k (xk [i, j]).
(4)
k=0 (i,j)?G k
This model is parameterized by K potential functions V k ., one for each scale in the pyramid ?(x).
In many applications we expect the frequencies of a 3x3 pattern to be different at each scale. The
potential functions can encourage or discourage specific patterns to occur at specific scales.
Note that ?(x) is a deterministic function and the pyramid representation does not introduce new
random variables. The pyramid simply defines a convenient way to specify potential functions over
large regions of x. A single potential function in a multiscale model can depend on a large area of
x due to the coarsenings. For large enough K (proportional to log of the image size) the Markov
blanket of a pixel can be the whole image.
While the experiments in Section 5 use the conditional modeling approach specified by Equation
(1), we can also use EFoP to define priors over binary images. Samples from these priors illustrate the information that is captured by a FoP model, specially the added benefit of the multiscale
representation. Figure 2 shows samples from FoP priors trained on contour maps of natural images.
The empirical studies in [14] suggest that low-order Markov models can not capture the empirical
length distribution of contours in natural images. A multiscale FoP model can control the size
distribution of objects much better than a low-order MRF. After coarsening the diameter of an object
goes down by a factor of approximately two, and eventually the object is mapped to a single pixel.
The scale at which this happens can be captured by a 3x3 pattern with an ?on? pixel surrounded by
?off? pixels (this assumes there are no other objects nearby). Since the cost of a pattern depends on
the scale at which it appears we can assign a cost to an object that is based loosely upon its size.
2.3
Data Model
Let y be an input image and ?(y) be an image pyramid computed from y. Our data models are
defined by sums over pixels in the two pyramids ?(x) and ?(y). In our experiments y is a graylevel
3
(a)
(b)
(c)
Figure 2: (a) Examples of training images T extracted from the BSD. (b) Samples from a singlescale
FoP prior trained on T . (c) Samples from a multiscale FoP prior trained on T . The multiscale model
is better at capturing the lengths of contours and relationships between them.
image with values in {0, . . . , M ? 1}. The pyramid ?(y) is defined in analogy to ?(x) except that
we use a local average for coarsening instead of the logical OR,
y k+1 (i, j) = b(y k (2i, 2j) + y k (2i + 1, 2j) + y k (2i, 2j + 1) + y k (2i + 1, 2j + 1))/4c
(5)
The data model is parameterized by K vectors D0 , . . . , DK?1 ? RM
Edata (x, y) =
K?1
X
X
xk (i, j)Dk (y k (i, j))
(6)
k=0 (i,j)?G k
Here Dk (y k (i, j)) is an observation cost incurred when xk (i, j) = 1. There is no need to include an
observation cost when xk (i, j) = 0 because only energy differences affect the posterior p(x|y).
We note that it would be interesting to consider data models that capture complex relationships
between local patterns in ?(x) and ?(y). For example a local maximum in y k (i, j) might give
evidence for xk (i, j) = 1, or a particular 3x3 pattern in xk [i, j].
2.4
Log-Linear Representation
The energy function E(x, y) of a FoP model can be expressed by a dot product between a vector of
model parameters w and a feature vector ?(x, y). The vector ?(x, y) has one block for each scale.
In the k-th block we have: (1) 512 (or 102 for invariant models) entries counting the number of
times each 3x3 pattern occurs in xk ; and (2) M entries counting the number of times each possible
value for y(i, j) occurs where xk (i, j) = 1. The vector w specifies the cost for each pattern in each
scale (V k ) and the parameters of the data model (Dk ). We then have that E(x, y) = w ? ?(x, y).
This log-linear form is useful for learning the model parameters as described in Section 4.
3
Inference with a Band Sampler
In inference we have a set of observations y and want to estimate x. We use MCMC methods [13]
to draw samples from p(x|y) and estimate the posterior marginal probabilities p(x(i, j) = 1|y).
Sampling is also used for learning model parameters as described in Section 4.
In a block Gibbs sampler we repeatedly update x by picking a block of pixels B and sampling new
values for xB from p(xB |y, xB ). If the blocks are selected appropriately this defines a Markov chain
with stationary distribution p(x|y).
We can implement a block Gibbs sampler for a multiscale FoP model by keeping track of the image
pyramid ?(x) as we update x. To sample from p(xB |y, xB ) we consider each possible configuration
4
for xB . We can efficiently update ?(x) to reflect a possible configuration for xB and evaluate the
terms in E(x, y) that depend on xB . This takes O(K|B|) time for each configuration for xB . This
in turn leads to an O(K|B|2|B| ) time algorithm for sampling from p(xB |y, xB? ). The running time
can be reduced to O(K2|B| ) using Gray codes to iterate over configurations for xB .
Here we define a band sampler that updates all pixels in a horizontal or vertical band of x in a single
step. Consider an n by m image x and let B be a horizontal band of pixels with h rows. Since
|B| = mh a straightforward implementation of block sampling for B is completely impractical.
However, for an Ising model we can generate samples from p(xB |y, xB ) in O(m22h ) time using
the forward-backward algorithm for Markov models. We simply treat each column of B as a single
variable with 2h possible states. A similar idea can be used for FoP models.
Let S be a state space where a state specifies a joint configuration of binary values for the pixels in
a column of B. Note that |S| = 2h . Let z1 , . . . , zm be a representation of xB in terms of the state
of each column. For a singlescale FoP model the distribution p(z1 , . . . , zn |y, xB? ) is a 2nd-order
Markov model. This allows for efficient sampling using forward weights computed via dynamic
programming. Such an algorithm takes O(m23h ) time to generate a sample from p(xB |y, xB ),
which is efficient for moderate values of h.
In a multiscale FoP model the 3x3 patterns in the upper levels of ?(x) depend on many columns of
B. This means p(z1 , . . . , zn |xB? ) is no longer 2nd-order. Therefore instead of sampling xB directly
we use a Metropolis-Hastings approach. Let p be a multiscale FoP model we would like to sample
from. Let q be a singlescale FoP model that approximates p. Let x be the current state of the
Markov chain and x0 be a proposal generated by the singlescale band sampler for q. We accept x0
with probability min(1, ((p(x0 |y)q(x|y))/(p(x|y)q(x0 |y)))). Efficient computation of acceptance
probabilities can be done using the pyramid representations of x and y. For each proposal we update
?(x) to ?(x0 ) and compute the difference in energy due to the change under both p and q.
One problem with the Metropolis-Hastings approach is that if proposals are rejected very often the
resulting Markov chain mixes slowly. We can avoid this problem by noting that most of the work
required to generate a sample from the proposal distribution involves computing forward weights
that can be re-used to generate other samples. Each step of our band sampler for a multiscale FoP
model picks a band B (horizontal or vertical) and generates many proposals for xB , accepting each
one with the appropriate acceptance probability. As long as one of the proposals is accepted the
work done in computing forward weights is not wasted.
4
Learning
We can learn models using maximum-likelihood and stochastic gradient descent. This is similar to
what was done in [24, 15, 20]. But in our case we have a conditional model so we maximize the
conditional likelihood of the training examples.
Let T = {(x1 , yi ), . . . , (xN , yN )} be a training set with N examples. We define the training objective using the negative log-likelihood of the data plus a regularization term. The regularization
ensures no pattern is too costly. This helps the Markov chains used during learning and inference to
mix reasonably fast. Let L(xi , yi ) = ? log p(xi |yi ). The training objective is given by
N
X
?
O(w) = ||w||2 +
L(xi , yi ).
(7)
2
i=1
This objective is convex and
?O(w) = ?w +
N
X
?(xi , yi ) ? Ep(x|yi ) [?(x, yi )].
(8)
i=1
Here Ep(x|yi ) [?(x, yi )] is the expectation of ?(x, yi ) under the posterior p(x|yi ) defined by the
current model parameters w. A stochastic approximation to the gradient ?O(w) can be obtained
by sampling x0i from p(x|yi ). Let ? be a learning rate. In each stochastic gradient descent step we
sample x0i from p(x|yi ) and update w as follows
N
X
w := w ? ?(?w +
?(xi , yi ) ? ?(x0i , yi )).
(9)
i=1
5
To sample the x0i we run N Markov chains, one for each training example, using the band sampler
from Section 3. After each model update we advance each Markov chain for a small number of steps
using the latest model parameters to obtain new samples x0i .
5
Applications
To evaluate the ability of FoP to adapt to different problems we consider two different applications.
In both cases we estimate hidden binary images x from grayscale input images y. We used ground
truth binary images obtained from standard datasets and synthetic observations. For the experiments
described here we generate y by sampling a value y(i, j) for each pixel independently from a normal
distribution with standard deviation ?y and mean ?0 or ?1 , depending on x(i, j),
y(i, j) ? N (?x(i,j) , ?y2 ).
(10)
We have also done experiments with more complex data models but the results we obtained were
similar to the results described here.
5.1
Contour Detection
The BSD [12, 2] contains images of natural scenes and manual segmentations of the most salient
objects in those images. We used one manual segmentation for each image in the BSD500. From
each image we generated a contour map x indicating the location of boundaryes between segments
in the image. To generate the observations y we used ?0 = 150, ?1 = 100 and ?y = 40 in Equation
(10). Our training and test sets each have 200 examples. We first trained a 1-scale FoP model. We
then trained a 4-level FoP model using the 1-level model as a proposal distribution for the band
sampler (see Section 3). Training each model took 2 days on a 20-core machine. During training
and testing we used the band sampler with h = 3 rows. Inference involves estimating posterior
marginal probabilities for each pixel by sampling from p(x|y). Inference on each image took 20
minutes on an 8-core machine.
For comparison we implemented a baseline technique using linear filters. Following [10] we used the
second derivative of an elongated Gaussian filter together with its Hilbert transform. The filters had
an elongation factor of 4 and we experimented with different values for the base standard deviation
?b of the Gaussian. The sum of squared responses of both filters defines an oriented energy map. We
evaluated the filters at 16 orientations and took the maximum response at each pixel. We performed
non-maximum suppression along the dominant orientations to obtain a thin contour map.
Figure 3 illustrates our results on 3 examples from the test set. Results on more examples are available in the supplemental material. For the FoP models we show the posterior marginal probabilities
p(x(i, j) = 1|y). The darkness of a pixel is proportional to the marginal probability. The FoP
models do a good job suppressing noise and localizing the contours. The multiscale FoP model in
particular gives fairly clean results despite the highly noisy inputs. The baseline results at lower ?b
values suffer from significant noise, detecting many spurious edges. The baseline at higher ?b values
suppresses noise at the expense of having poor localization and missing high-curvature boundaries.
For a quantitative evaluation we compute precision-recall curves for the different models by thresholding the estimated contour maps at different values. Figure 4 shows the precision-recall curves.
The average precision (AP) was found by calculating the area under the precision-recall curves. The
1-level FoP model AP was 0.73. The 4-level FoP model AP was 0.78. The best baseline AP was
0.18 obtained with ?b = 1. We have also done experiments using lower observation noise levels ?y .
With low observation noise the 1-level and 4-level FoP results become similar and baseline results
improve significantly approaching the FoP results.
5.2
Binary Segmentation
For this experiment we obtained binary images from the Swedish Leaf Dataset [18]. We focused on
the class of Rowan leaves because they have complex shapes. Each image defines a segmentation
mask x. To generate the observations y we used ?0 = 150, ?1 = 100 and ?y = 100 in Equation
(10). We used a higher ?y compared to the previous experiment because the 2D nature of masks
makes it possible to recover them under higher noise. We used 50 examples for training and 25
6
Contour map x
Observation y
Baseline ?b = 1
Baseline ?b = 4
FoP 1
FoP 4
Figure 3: Contour detection results. Top-to-bottom: Hidden contour map x, input image y, output
of oriented filter baseline with ?b = 1 and ?b = 4, output of 1-level and 4-level FoP model.
examples for testing. We trained FoP models with the same procedure and parameters used for the
contour detection experiment. For a baseline, we used graph-cuts [5, 4] to perform MAP inference
with an Ising model. We set the data term using our knowledge of the observation model and picked
the pairwise discontinuity cost minimizing the per-pixel error rate in the test set.
Figure 5 illustrates the results of the different methods. Results on other images are available in the
supplemental material. The precision-recall curves are in Figure 4. Graph-cuts yields a precisionrecall point, with precision 0.893 and recall 0.916. The 1-level FoP model has a higher precision of
0.915 at the same recall. The 4-level FoP model raises the precision to 0.929 at the same recall. The
7
(a) Contour detection
(b) Binary segmentation
Figure 4: (a) Precision-recall curves for the contour detection experiment. (b) Precision-recall curves
for the segmentation experiment (the graph-cuts baseline yields a single precision-recall point).
Mask x
Observation y
Graph-cuts
FoP 1
FoP 4
Figure 5: Binary segmentation examples. The 4-level FoP model does a better job recovering pixels
near the object boundary and the stem of the leaves.
differences in precision are small because they are due to pixels near the object boundary but those
are the hardest pixels to get right. There are clear differences that can be seen by visual inspection.
6
Conclusion
We described a general framework for defining high-order image models. The idea involves modeling local properties in a multiscale representation of an image. This leads to a natural lowdimensional parameterization for high-order models that exploits standard pyramid representations
of images. Our experiments demonstrate the approach yields good results on two applications that
require very different image priors, illustrating the broad applicability of our models. An interesting
direction for future work is to consider FoP models for non-binary images.
Acknowledgements
We would like to thank Alexandra Shapiro for helpful discussions and initial experiments related to
this project. This material is based upon work supported by the National Science Foundation under
Grant No. 1161282.
8
References
[1] S. Alpert, M. Galun, B. Nadler, and R. Basri. Detecting faint curved edges in noisy images. In ECCV,
2010.
[2] Pablo Arbelaez, Michael Maire, Charless Fowlkes, and Jitendra Malik. Contour detection and hierarchical
image segmentation. PAMI, 33(5):898?916, May 2011.
[3] J. August and S.W. Zucker. Sketches with curvature: The curve indicator random field and markov
processes. PAMI, 25(4):387?400, April 2003.
[4] Y. Boykov and V. Kolmogorov. An experimental comparison of min-cut/max-flow algorithms for energy
minimization in vision. PAMI, 26(9):1124?1137, Sep 2004.
[5] Y. Boykov, O. Veksler, and R.Zabih. Efficient approximate energy minimization via graph cuts. PAMI,
20(12):1222?1239, Nov 2001.
[6] X. Descombes, J.F. Mangin, E. Pechersky, and M. Sigelle. Fine structures preserving markov model for
image processing. In SCIA, 1995.
[7] S.M. Eslami, N. Heess, and J. Winn. The shape boltzmann machine: a strong model of object shape.
CVPR, 2012.
[8] P. F. Felzenszwalb and D. P. Huttenlocher. Efficient belief propagation for early vision. IJCV, 70(1),
2006.
[9] P. F. Felzenszwalb and J. Schwartz. Hierarchical matching of deformable shapes. CVPR, 2007.
[10] T. Leung and J. Malik. Contour continuity in region-based image segmentation. ECCV, pages 544?559,
1998.
[11] Shyjan Mahamud, Lance R. Williams, Karvel K. Thornber, and Kanglin Xu. Segmentation of multiple
salient closed contours from real images. PAMI, 25(4):433?444, 2003.
[12] David R. Martin, Charless C. Fowlkes, and Jitendra Malik. Learning to detect natural image boundaries
using local brightness, color, and texture cues. PAMI, 26(5):530?549, 2004.
[13] R. Neal. Probabilistic inference using markov chain monte carlo methods. Technical Report CRG-TR93-1, Computer Science, University of Toronto, 1993.
[14] Xiaofeng Ren, Charless Fowlkes, and Jitendra Malik. Learning probabilistic models for contour completion in natural images. IJCV, 77(1-3):47?63, 2008.
[15] Stefan Roth and Michael J. Black. Fields of experts. IJCV, 82(2):205?229, 2009.
[16] A. Shashua and S. Ullman. Structural saliency: The detection of globally salient structures using a locally
connected network. In ICCV, pages 321?327, 1988.
[17] A. Shekhovtsov, P. Kohli, and C. Rother. Curvature prior for mrf-based segmentation and shape inpainting. In DAGM, 2012.
[18] O. J. O. S?oderkvist. Computer vision classification of leaves from swedish trees. Master?s thesis,
Link?oping University, September 2001.
[19] Rustem Takhanov and Vladimir Kolmogorov. Inference algorithms for pattern-based crfs on sequence
data. In ICML, 2013.
[20] T. Tieleman. Training restricted boltzmann machines using approximations to the likelihood gradient. In
ICML, 2008.
[21] Lance R. Williams and David W. Jacobs. Local parallel computation of stochastic completion fields.
Neural Computation, 9(4):859?881, 1997.
[22] Lance R. Williams and David W. Jacobs. Stochastic completion fields: A neural model of illusory contour
shape and salience. Neural Computation, 9(4):837?858, 1997.
[23] A.S. Willsky. Multiresolution markov models for signal and image processing. Proceedings of the IEEE,
90(8):1396?1458, 2002.
[24] S. C. Zhu, Y. N. Wu, and D.B. Mumford. Filters, random fields and maximum entropy (FRAME):
Towards a unified theory for texture modeling. IJCV, 27(2):1?20, 1998.
9
| 5283 |@word kohli:1 illustrating:1 nd:2 jacob:2 pick:1 brightness:1 inpainting:1 initial:1 configuration:5 contains:1 selecting:1 suppressing:1 rowan:1 current:2 assigning:1 john:2 realistic:1 shape:7 treating:1 update:8 stationary:1 cue:1 selected:1 leaf:4 parameterization:1 inspection:1 xk:17 core:2 accepting:1 coarse:2 detecting:4 location:1 toronto:1 along:1 become:1 ijcv:4 introduce:1 pairwise:1 x0:7 mask:4 expected:1 globally:1 quad:2 window:2 increasing:1 project:1 estimating:2 underlying:1 what:1 tying:1 suppresses:1 supplemental:2 unified:1 impractical:1 berkeley:1 quantitative:1 rustem:1 descombes:1 scaled:1 rm:1 k2:1 control:1 mangin:1 grant:1 schwartz:1 yn:1 local:18 treat:1 despite:1 eslami:1 approximately:1 ap:4 might:1 plus:1 pami:6 black:1 scia:1 testing:2 practice:2 block:9 implement:1 x3:15 precisionrecall:1 procedure:2 maire:1 area:4 empirical:2 significantly:1 convenient:1 matching:1 pre:1 suggest:1 get:1 darkness:1 map:13 deterministic:2 elongated:1 crfs:2 missing:1 go:1 straightforward:1 latest:1 independently:1 convex:2 focused:1 resolution:6 williams:3 assigns:1 graylevel:1 hierarchy:1 programming:2 cut:6 ising:2 huttenlocher:1 coarsened:1 observed:3 bottom:2 ep:2 capture:7 region:2 ensures:1 connected:3 dynamic:2 trained:6 depend:5 solving:1 segment:1 raise:1 technically:1 upon:2 localization:1 completely:1 sep:1 mh:1 joint:1 kolmogorov:2 train:1 fast:1 describe:3 effective:1 monte:1 neighborhood:3 outside:1 widely:1 cvpr:2 ability:1 statistic:1 think:1 highlighted:1 noisy:3 transform:1 sequence:1 took:3 lowdimensional:1 product:1 zm:1 neighboring:1 combining:1 multiresolution:2 deformable:1 object:13 help:1 illustrate:1 develop:1 completion:4 depending:1 x0i:5 job:2 strong:1 implemented:1 recovering:1 involves:11 blanket:1 concentrate:1 direction:1 closely:1 filter:11 stochastic:6 centered:1 material:3 require:2 assign:1 bi1:1 crg:1 around:2 considered:1 ground:1 normal:1 exp:1 nadler:1 early:1 estimation:1 minimization:2 stefan:1 gaussian:2 avoid:1 coarsen:1 likelihood:5 suppression:1 baseline:10 detect:2 helpful:1 inference:13 leung:1 dagm:1 accept:1 hidden:5 spurious:1 going:1 pixel:23 classification:1 orientation:2 fairly:1 singlescale:7 marginal:4 field:12 having:1 sampling:11 elongation:1 broad:1 look:1 hardest:1 icml:2 thin:3 future:1 others:1 report:1 oriented:2 national:1 subsampled:1 detection:11 acceptance:2 highly:1 evaluation:1 xb:21 chain:7 edge:2 encourage:3 tree:4 loosely:1 circle:1 re:1 column:4 modeling:5 localizing:1 restoration:2 zn:2 cost:10 applicability:1 deviation:2 entry:2 veksler:1 oping:1 too:1 motivating:1 reported:1 providence:2 synthetic:1 probabilistic:2 off:1 picking:1 michael:2 together:2 squared:1 reflect:2 thesis:1 slowly:1 expert:2 derivative:1 ullman:1 potential:5 jitendra:3 depends:2 performed:1 picked:1 closed:1 shashua:1 recover:1 parallel:1 who:1 efficiently:1 yield:3 saliency:1 shekhovtsov:1 ren:1 carlo:1 foe:2 history:1 manual:2 energy:11 frequency:5 dataset:2 logical:2 recall:10 color:2 knowledge:1 illusory:1 segmentation:15 hilbert:1 back:1 appears:3 higher:5 day:1 response:3 specify:1 april:1 swedish:2 evaluated:2 though:1 done:5 rejected:1 sketch:1 hastings:3 horizontal:4 expressive:1 multiscale:22 propagation:1 continuity:1 defines:4 artifact:1 gray:1 alexandra:1 brown:4 contain:1 y2:1 regularization:2 fop:43 spatially:1 neal:1 during:2 criterion:1 leftmost:1 demonstrate:1 image:72 recently:1 charles:3 boykov:2 common:1 rotation:1 approximates:1 significant:1 gibbs:2 grid:2 had:1 dot:1 zucker:1 longer:1 base:1 dominant:1 curvature:4 posterior:5 recent:2 moderate:1 binary:27 yi:15 captured:3 seen:1 preserving:1 maximize:1 signal:2 multiple:1 mix:2 d0:1 stem:1 technical:1 match:2 adapt:1 long:2 involving:1 basic:1 mrf:2 vision:5 expectation:1 sometimes:1 pyramid:14 thornber:1 proposal:7 want:1 fine:1 winn:1 appropriately:1 extra:1 nary:1 specially:1 flow:1 coarsening:7 structural:1 near:2 counting:2 noting:1 enough:1 variety:4 affect:1 iterate:1 approaching:1 idea:2 stereo:1 suffer:2 repeatedly:1 deep:1 heess:1 useful:2 clear:2 involve:1 band:11 locally:1 concentrated:1 zabih:1 simplest:1 diameter:1 reduced:1 specifies:2 generate:7 shapiro:1 estimated:1 track:1 per:1 alpert:1 salient:4 clean:1 backward:2 wasted:1 graph:5 sum:3 run:1 coarsenings:2 parameterized:2 master:1 wu:1 patch:1 draw:1 mahamud:1 capturing:1 occur:1 ri:2 scene:1 edata:3 nearby:1 generates:1 min:2 coarsest:1 relatively:1 martin:1 bsd:3 poor:1 smaller:1 metropolis:3 making:1 happens:1 restricted:2 invariant:3 iccv:1 equation:3 visualization:1 turn:1 eventually:1 available:2 operation:1 observe:1 hierarchical:2 spectral:1 appropriate:2 fowlkes:3 original:3 top:2 assumes:1 include:2 running:1 calculating:1 exploit:1 objective:3 malik:4 added:1 occurs:2 mumford:1 costly:1 september:1 gradient:4 thank:1 mapped:1 arbelaez:1 link:1 willsky:1 rother:1 length:2 code:1 relationship:2 minimizing:1 vladimir:1 expense:1 negative:1 design:1 implementation:1 boltzmann:3 perform:1 upper:1 vertical:3 observation:12 markov:17 datasets:2 descent:2 curved:1 defining:3 frame:2 august:1 pablo:1 david:3 pair:1 required:1 specified:1 z1:3 coherent:1 distinction:1 learned:1 discontinuity:1 pattern:41 max:1 lance:3 belief:1 roth:1 natural:7 indicator:2 zhu:1 improve:1 prior:15 acknowledgement:1 expect:1 interesting:2 proportional:2 analogy:1 foundation:1 incurred:1 thresholding:1 surrounded:1 row:2 eccv:2 supported:1 keeping:1 salience:1 felzenszwalb:3 benefit:1 curve:9 boundary:4 xn:1 avoids:1 contour:29 forward:5 approximate:1 nov:1 basri:1 xi:5 grayscale:2 learn:1 reasonably:1 nature:1 symmetry:2 complex:3 discourage:1 whole:1 noise:6 galun:1 repeated:2 x1:1 xu:1 precision:12 down:1 minute:1 xiaofeng:1 specific:2 oberlin:2 dk:4 experimented:1 faint:1 evidence:1 mirror:1 texture:2 illustrates:2 pff:1 entropy:1 simply:2 visual:1 expressed:1 pedro:1 corresponds:1 truth:1 tieleman:1 extracted:1 conditional:4 goal:1 towards:1 change:1 except:1 reducing:1 sampler:9 accepted:1 experimental:1 indicating:1 support:1 maximumlikelihood:1 evaluate:3 mcmc:4 |
4,731 | 5,284 | Weakly-supervised Discovery of
Visual Pattern Configurations
Hyun Oh Song
Yong Jae Lee*
Stefanie Jegelka
*
University of California, Berkeley
Trevor Darrell
University of California, Davis
Abstract
The prominence of weakly labeled data gives rise to a growing demand for object detection methods that can cope with minimal supervision. We propose an
approach that automatically identifies discriminative configurations of visual patterns that are characteristic of a given object class. We formulate the problem as a
constrained submodular optimization problem and demonstrate the benefits of the
discovered configurations in remedying mislocalizations and finding informative
positive and negative training examples. Together, these lead to state-of-the-art
weakly-supervised detection results on the challenging PASCAL VOC dataset.
1
Introduction
The growing amount of sparsely and noisily labeled image data demands robust detection methods
that can cope with a minimal amount of supervision. A prominent example of this scenario is the
abundant availability of labels at the image level (i.e., whether a certain object is present or absent
in the image); detailed annotations of the exact location of the object are tedious and expensive and,
consequently, scarce. Learning methods that can handle image-level labels circumvent the need
for such detailed annotations and therefore have the potential to effectively use the vast textually
annotated visual data available on the Web. Moreover, if the detailed annotations happen to be noisy
or erroneous, such weakly supervised methods can even be more robust than fully supervised ones.
Motivated by these developments, recent work has explored learning methods that decreasingly
rely on strong supervision. Early ideas for weakly supervised detection [11, 32] paved the way
by successfully learning part-based object models, albeit on simple object-centric datasets (e.g.,
Caltech-101). Since then, a number of approaches [21, 26, 29] have aimed at learning models from
more realistic and challenging data sets that feature large intra-category appearance variations and
background clutter. These approaches typically generate multiple candidate regions and retain the
ones that occur most frequently in the positively-labeled images. However, due to intra-category
variations and deformations, the identified (single) patches often correspond to only a part of the
object, such as a human face instead of the entire body. Such mislocalizations are a frequent problem
for weakly supervised detection methods.
Mislocalization and too large or too small bounding boxes are problematic in two respects. First,
detection is commonly phrased as multiple instance learning (MIL) and solved by non-convex optimization methods that alternatingly guess the location of the objects as positive examples (since
the true location is unknown) and train a detector based on those guesses. This procedure is heavily
affected by the initial localizations. Second, the detector is often trained in stages; in each stage one
adds informative ?hard? negative examples to the training data. If we are not given accurate true
object localizations in the training data, these hard examples must be derived from the detections
inferred in earlier rounds. The higher the accuracy of the initial localizations, the more informative
is the augmented training data ? and this is key to the accuracy of the final learned model.
In this work, we address the issue of mislocalizations by identifying characteristic, discriminative
configurations of multiple patches (rather than a single one). This part-based approach is motivated
1
by the observation that automatically discovered single ?discriminative? patches often correspond
to object parts. In addition, while background patches (e.g., of water or sky) can also occur throughout the positive images, they will re-occur in arbitrary rather than ?typical? configurations. We
develop an effective method that takes as input a set of images with labels of the form ?the object is
present/absent?, and automatically identifies characteristic part configurations of the given object.
To identify such configurations, we use two main criteria. First, useful patches are discriminative,
i.e., they occur in many positively-labeled images, and rarely in the negatively labeled ones. To identify such patches, we use a discriminative covering formulation similar to [29]. Second, the patches
should represent different parts, i.e., they may be close but should not overlap too much. In covering
formulations, one may rule out overlaps by saying that for two overlapping regions, one ?covers?
the other, i.e., they are treated as identical and picking one is as good as picking both. But identity is
a transitive relation, and the density of possible regions in detection would imply that all regions are
identical, strongly discouraging the selection of more than one part per image. Partial covers face
the problem of scale invariance. Hence, we instead formulate an independence constraint. This second criterion ensures that we select regions that may be close but are non-redundant and sufficiently
non-overlapping. We show that this constrained selection problem corresponds to maximizing a
submodular function subject to a matroid intersection constraint, which leads to approximation algorithms with theoretical worst-case bounds. Given candidate parts identified by these two criteria,
we effectively find frequently co-occurring configurations that take into account relative position,
scale, and viewpoint.
We demonstrate multiple benefits of the discovered configurations. First, we observe that configurations of patches can produce more accurate spatial coverage of the full object, especially when the
most discriminative pattern corresponds to an object part. Second, any overlapping region between
co-occurring visual patterns is likely to cover a part (but not the full) of the object of interest. Thus,
they can be used to generate mis-localized positives as informative hard negatives for training (see
white boxes in Figure 3), which can further reduce localization errors at test time.
In short, our main contribution is a weakly-supervised object detection method that automatically
discovers frequent configurations of discriminative visual patterns to train robust object detectors.
In our experiments on the challenging PASCAL VOC dataset, we find the inclusion of our discriminative, automatically detected configurations to outperform all existing state-of-the-art methods.
2
Related work
Weakly-supervised object detection. Object detectors have commonly been trained in a fullysupervised manner, using tight bounding box annotations that cover the object of interest (e.g., [10]).
To reduce laborious bounding box annotation costs, recent weakly-supervised approaches [3, 4, 11,
21, 26, 29, 32] use image-level object-presence labels with no information on object location.
Early efforts [11, 32] focused on simple datasets that have a single prominent object in each image
(e.g., Caltech-101). More recent approaches [21, 26, 29] work with the more challenging PASCAL
dataset that contains multiple objects in each image and large intra-category appearance variations.
Of these, Song et al. [29] achieve state-of-the-art results by finding discriminative image patches
that occur frequently in the positive images but rarely in the negative images, using deep Convolutional Neural Network (CNN) features [17] and a submodular cover formulation. We build on their
approach to identify discriminative patches. But, contrary to [29] which assumes patches to contain
entire objects, we assume patches to contain either full objects or merely object parts, and automatically piece together those patches to produce better full-object estimates. To this end, we change
the covering formulation and identify patches that are both representative and explicitly mutually
different. This leads to more robust object estimates and further allows our system to intelligently
select ?hard negatives? (mislocalized objects), both of which improve detection performance.
Visual data mining. Existing approaches discover high-level object categories [14, 7, 28], mid-level
patches [5, 16, 24], or low-level foreground features [18] by grouping similar visual patterns (i.e.,
images, patches, or contours) according to their texture, color, shape, etc. Recent methods [5, 16]
use weakly-supervised labels to discover discriminative visual patterns. We use related ideas, but
formulate the problem as a submodular optimization over matroids, which leads to approximation
algorithms with theoretical worst-case guarantees. Covering formulations have also been used in
2
[1, 2], but after running a trained object detector. An alternative discriminative approach is to use
spectral methods [34].
Modeling co-occurring visual patterns. It is known that modeling the spatial and geometric relationship between co-occurring visual patterns (objects or object-parts) often improves visual recognition performance [8, 18, 10, 11, 19, 23, 27, 24, 32, 33]. Co-occurring patterns are usually represented as doublets [24], higher-order constellations [11, 32] or star-shaped models [10]. Among
these, our work is most inspired by [11, 32], which learn part-based models with weak supervision. We use more informative deep CNN features and a different formulation, and show results on
more difficult datasets. Our work is also related to [19], which discovers high-level object compositions (?visual phrases? [8]), but with ground-truth bounding box annotations. In contrast, we aim to
discover part compositions to represent full objects and do so with less supervision.
3
Approach
Our goal is to find a discriminative set of patches that co-occur in the same configuration in many
positively-labeled images. We address this goal in two steps. First, we find a set of patches that are
discriminative; i.e., they occur frequently in positive images and rarely in negative images. Second,
we efficiently find co-occurring configurations of pairs of such patches. Our approach easily extends
beyond pairs; for simplicity and to retain configurations that occur frequently enough, we here
restrict ourselves to pairs.
Discriminative candidate patches. For identifying discriminative patches, we begin with a construction similar to that of Song et al. [29]. Let P be the set of positively-labeled images. Each
image I contains candidate boxes {bI,1 , . . . , bI,m } found via selective search [30]. For each bI,i , we
find its closest matching neighbor bI 0 ,j in each other image I 0 (regardless of the image label). The
K closest of those neighbors form the neighborhood N (bI,i ); the remaining ones are discarded.
Discriminative patches have neighborhoods mainly within images in P, i.e., if B(P) is the set of all
patches from images in P, then |N (b) ? B(P)| ? K. To identify a small, diverse and representative
set of such patches, like [29], we construct a bipartite graph G = (U, V, E), where both U and V
contain copies of B(P). Each patch b ? V is connected to the copy of its nearest neighbors in U (i.e.,
N (b) ? B(P)). These will be K or fewer, depending on whether the K nearest neighbors of b occur
in B(P) or in negatively-labeled images. The most representative patches maximize the covering
function
F (S) = |?(S)|,
(1)
where ?(S) = {u ? U | (b, u) ? E for some b ? S} ? U is the neighborhood of S ? V in the
bipartite graph. Figure 1 shows a cartoon illustration. The function F is monotone and submodular,
and the C maximizing elements (for a given C) can be selected greedily [20].
However, if we aim to find part configurations, we must select multiple, jointly informative patches
per image. Patches selected to merely maximize coverage can still be redundant, since the most
frequently occurring ones are often highly overlapping. A straightforward modification would be
to treat highly overlapping patches as identical. This identification would still admit a submodular
cover model as in Equation (1). But, in our case, the candidate patches are very densely packed in
the image, and, by transitivity, we would have to make all of them identical. In consequence, this
would completely rule out the selection of more than one patch in an image and thereby prohibit the
discovery of any co-occurring configurations.
Instead, we directly constrain our selection such that no two patches b, b0 ? V can be picked whose
neighborhoods overlap by more than a fraction ?. By overlap, we mean that the patches in the
neighborhoods of b, b0 overlap significantly (they need not be identical). This notion of diversity is
reminiscent of NMS and similar to that in [5], but we here phrase and analyze it as a constrained
submodular optimization problem. Our constraint can be expressed in terms of a different graph
GC = (V, EC ) with nodes V. In GC , there is an edge between b and b0 if their neighborhoods overlap
prohibitively, as illustrated in Figure 1. Our family of feasible solutions is
M = {S ? V | ? b, b0 ? S there is no edge (b, b0 ) ? EC }.
(2)
In other words, M is the family of all independent sets in GC . We aim to maximize
maxS?V F (S) s.t. S ? M.
3
(3)
V
U
Figure 1: Left: bipartite graph G that defines the utility function F and identifies discriminative
patches; right: graph GC that defines the diversifying independence constraints M. We may pick
C1 (yellow) and C3 (green) together, but not C2 (red) with any of those.
This problem is NP-hard. We solve it approximately via the following greedy algorithm. Begin with
S 0 = ?, and, in iteration t, add b ? argmaxb?V\S |?(b) \ ?(S t?1 )|. As we add b, we delete all of
b?s neighbors in GC from V. We continue until V = ?. If the neighborhoods of any b, b0 are disjoint
but contain overlapping elements (?(b) ? ?(b0 ) = ? but there exist u ? ?(b) and u0 ? ?(b0 ) that
overlap), then this algorithm amounts to the following simplified scheme: we first sort all b ? V in
non-increasing order by their degree ?(b), i.e., their number of neighbors in B(P), and visit them in
this order. We always add the currently highest b in the list to S, then delete it from the list, and with
it all its immediate (overlapping) neighbors in GC . The following lemma states an approximation
factor for the greedy algorithm, where ? is the maximum degree of any node in GC .
Lemma 1. The solution Sg returned by the greedy algorithm is a 1/(? + 2) approximation for
1
Problem (2): F (Sg ) ? ?+2
F (S ? ). If ?(b) ? ?(b0 ) = ? for all b, b0 ? V, then the worst-case
approximation factor is 1/(? + 1).
The proof relies on phrasing M as an intersection of matroids.
Definition 1 (Matroid). A matroid (V, Ik ) consists of a ground set V and a family Ik ? 2V of
?independent sets? that satisfy three axioms: (1) ? ? Ik ; (2) downward closedness: if S ? Ik then
T ? Ik for all T ? S; and (3) the exchange property: if S, T ? Ik and |S| < |T |, then there is an
element v ? T \ S such that S ? {v} ? Ik .
Proof. (Lemma 1) We will argue that Problem (2) is the problem of maximizing a monotone submodular function subject to the constraint that the solution lies in the intersection of ? + 1 matroids.
With this insight, the approximation factor of the greedy algorithm for submodular F follows from
[12] and that for non-intersecting ?(b) from [15], since in the latter case the problem is that of
finding a maximum weight vector in the intersection of ? + 1 matroids.
It remains to argue that M is an intersection of matroids. Our matroids will be partition matroids
(over the ground set V) whose independent sets are of the form Ik = {S | |S ? e| ? 1, for all e ?
Ek }. To define those, we partition the edges in GC into disjoint sets Ek , i.e., no two edges in Ek
share a common node. The Ek can be found by an edge coloring ? one Ek and Ik for each color k.
By Vizing?s theorem [31], we need at most ?+1 colors. The matroid Ik demands that for each edge
e ? Ek , we may only select one of its adjacent nodes. All matroids together say that for any edge
e ? E, we may only select one of the adjacent nodes, and that is the constraint in Equation (2), i.e.
T?+1
M = k=1 Ik . We do not ever need to explicitly compute Ek and Ik ; all we need to do is check
membership in the intersection, and this is equivalent to checking whether a set S is an independent
set in GC , which is achieved implicitly via the deletions in the algorithm.
From the constrained greedy algorithm, we obtain a set S ? V of discriminative patches. Together
with its neighborhood ?(b), each patch b ? V forms a representative cluster. Figure 2 shows some
example patches derived from the labels ?aeroplane? and ?motorbike?. The discovered patches
intuitively look like ?parts? of the objects, and are frequent but sufficiently different.
Finding frequent configurations. The next step is to find frequent configurations of co-occurring
clusters, e.g., the head patch of a person on top of the torso patch, or a bicycle with visible wheels.
4
Figure 2: Examples of discovered patch ?clusters? for aeroplane, motorbike, and cat. The discovered
patches intuitively look like object parts, and are frequent but sufficiently different.
A ?configuration? consists of patches from two clusters Ci , Cj , their relative location, and their
viewpoint and scale. In practice, we give preference to pairs that by themselves are very relevant
and maximize a weighted combination of co-occurrence count and coverage max{?(Ci ), ?(Cj )}.
All possible configurations of all pairs of patches amount to too many to explicitly write down and
count. Instead, we follow an efficient procedure for finding frequent configurations. Our approach
is inspired by [19], but does not require any supervision. We first find configurations that occur in at
least two images. To do so, we consider each pair of images I1 , I2 that have at least two co-occurring
clusters. For each correspondence of cluster patches across the images, we find a corresponding
transform operation (translation, scale, viewpoint change). This results in a point in a 4D transform
space, for each cluster correspondence. We quantize this space into B bins. Our candidate configurations will be pairs of cluster correspondences ((bI1 ,1 , bI2 ,1 ), (bI1 ,2 , bI2 ,2 )) ? (Ci ?Ci )?(Cj ?Cj )
that fall in the same bin, i.e., share the same transform and have the same relative location. Between
a given pair of images, there can be multiple such pairs of correspondences. We keep track of those
via a multi-graph GP = (P, EP ) that has a node for each image I ? P. For each correspondence
((bI1 ,1 , bI2 ,1 ), (bI1 ,2 , bI2 ,2 )), we draw an edge (I1 , I2 ) and label it by the clusters Ci , Cj and the
common relative position. As a result, there can be multiple edges (I1 , Ij ) in GP with different edge
labels.
The most frequently occurring configuration can now be read out by finding the largest connected
component in GP induced by retaining only edges with the same label. We use the largest component(s) as the characteristic configurations for a given image label (object class). If the component
is very small, then there is not enough information to determine co-occurrences, and we simply use
the most frequent single cluster. The final single ?correct? localization will be the smallest bounding
box that contains the full configuration.
Discovering mislocalized hard negatives. Discovering frequent configurations can not only lead
to better localization estimates of the full object, but they can also be used to generate mislocalized
estimates as ?hard negatives? when training the object detector. We exploit this idea as follows.
Let b1 , b2 be a discovered configuration within a given image. These patches typically constitute
co-occurring parts or a part and the full object. Our foreground estimate is the smallest box that
includes both b1 and b2 . Hence, any region within the foreground estimate that does not overlap
simultaneously with both b1 and b2 will capture only a fragment of the foreground object. We extract
the four largest such rectangular regions (see white boxes in Figure 3) as hard negative examples.
Specifically, we parameterize any rectangular region with [xl , xr , y t , y b ], i.e., its x-left, x-right,
y-top, and y-bottom coordinate values. Let the bounding box of bi (i = 1, 2) be [xli , xri , yit , yib ],
the foreground estimate be [xlf , xrf , yft , yfb ], and let xl = max(xl1 , xl2 ), xr = min(xr1 , xr2 ), y t =
max(y1t , y2t ), y b = min(y1b , y2b ). We generate four hard negatives: [xlf , xl , yfb , yft ], [xr , xrf , yfb , yft ],
[xlf , xrf , yft , y t ], [xlf , xrf , y b , yfb ]. If either b1 or b2 is very small in size relative to the foreground, the
resulting hard negatives can have high overlap with the foreground, which will introduce undesirable
noise (false negatives) when training the detector. Thus, we shrink any hard negative that overlaps
with the foreground estimate by more than 50%, until its overlap is 50% (we adjust the boundary
that does not coincide with any of the foreground estimation boundaries).
5
Figure 3: Automatically discovered foreground estimation box (magenta), hard negative (white),
and the patch (yellow) that induced the hard negative. Note that we are only showing the largest one
out of (up to) four hard negatives per image.
Note that simply taking arbitrary rectangular regions that overlap with the foreground estimation box
by some threshold will not always generate useful hard negatives (as we show in the experiments).
If the overlap threshold is too low, the selected regions will be uninformative, and if the overlap
threshold is too high, the selected regions will cover too much of the foreground. Our approach
selects informative hard negatives more robustly by ruling out the overlapping region between the
configuration patches, which is very likely be part of the foreground object but not the full object.
Mining positives and training the detector. While the discovered configurations typically lead
to better foreground localization, their absolute count can be relatively low compared to the total
number of positive images. This is due to inaccuracies in the initial patch discovery stage: for a
frequent configuration to be discovered, both of its patches must be found accurately. Thus, we also
mine additional positives from the set of remaining positive images P 0 that did not produce any of
the discovered configurations.
To do so, we train an initial object detector, using the foreground estimates derived from our discovered configurations as positive examples, and the corresponding discovered hard negative regions as
negatives. In addition, we mine negative examples in negative images as in [10]. We run the detector
on all selective search regions in P 0 and retain the region in each image with the highest detection
score as an additional positive training example. Our final detector is trained on this augmented
training data, and iteratively improved by latent SVM (LSVM) updates (see [10, 29] for details).
4
Experiments
In this section, we analyze: (1) detection performance of the models trained with the discovered
configurations, and (2) impact of the discovered hard negatives on detection performance.
Implementation details. We employ a recent region based detection framework [13, 29] and use the
same fc7 features from the CNN model [6] on region proposals [30] throughout the experiments. For
discriminative patch discovery, we use K = |P|/2, ? = K/20. For correspondence detection, we
discretize the 4D transform space of {x: relative horizontal shift, y: relative vertical shift, s: relative
scale, p: relative aspect ratio} with ?x = 30 px, ?y = 30 px, ?s = 1 px/px, ?p = 1 px/px.
We chose this binning scheme by examining a few qualitative examples so that scale and aspect
ratio agreement between the two paired instances are more strict, while their translation agreement
is more loose, in order to handle deformable objects. More details regarding the transform space
binning can be found in [22].
Discovered configurations. Figure 5 shows the discovered configurations (solid green and yellow
boxes) and foreground estimates (dashed magenta boxes) that have high degree in graph GP for all
20 classes in the PASCAL dataset. Our method consistently finds meaningful combinations such
as a wheel and body of bicycles, face and torso of people, locomotive basement and upper body
parts of trains/buses, and window and body frame of cars. Some failures include cases where the
algorithm latches onto different objects co-occurring in consistent configurations such as the lamp
and sofa combination (right column, second row from the bottom in Figure 5).
Weakly-supervised object detection. Following the evaluation protocol of the PASCAL VOC
dataset, we report detection results on the PASCAL test set using detection average precision. For a
direct comparison with the state-of-the-art weakly-supervised object detection method [29], we do
not use the extra instance level annotations such as pose, difficult, truncated and restrict the supervision to the image-level object presence annotations. Table 1 compares our detection results against
two baseline methods [25, 29] on the full dataset. Our method improves detection performance on
15 of the 20 classes. It is worth noting that our method yields significant improvement on the person
6
aero bike bird boat btl bus car cat chr cow tble dog horse mbk pson plnt shp sofa train tv mAP
[25] 13.4 44.0 3.1 3.1 0.0 31.2 43.9 7.1 0.1 9.3 9.9 1.5 29.4 38.3 4.6 0.1 0.4 3.8 34.2 0.0 13.9
[29] 27.6 41.9 19.7 9.1 10.4 35.8 39.1 33.6 0.6 20.9 10.0 27.7 29.4 39.2 9.1 19.3 20.5 17.1 35.6 7.1 22.7
ours1 31.9 47.0 21.9 8.7 4.9 34.4 41.8 25.6 0.3 19.5 14.2 23.0 27.8 38.7 21.2 17.6 26.9 12.8 40.1 9.2 23.4
ours2 36.3 47.6 23.3 12.3 11.1 36.0 46.6 25.4 0.7 23.5 12.5 23.5 27.9 40.9 14.8 19.2 24.2 17.1 37.7 11.6 24.6
Table 1: Detection average precision (%) on full PASCAL VOC 2007 test set. ours1 : before latent
updates. ours2 : after latent updates
w/o hard negatives
neighboring hard negatives
discovered hard negatives
ours + SVM
22.5
22.2
23.4
ours + LSVM
23.7
23.9
24.6
Table 2: Effect of our hard negative examples on full PASCAL VOC 2007 test set.
class, which is arguably the most important category in the PASCAL dataset. Figure 4 shows some
example high scoring detections on the test set. Our method produces more complete detections
since it is trained on better localized instances of the object-of-interest.
Figure 4: Example detections on test set. Green: our method, red: [29]
Impact of discovered hard negatives. To analyze the effect of our discovered hard negatives, we
compare to two baselines: (1) not adding any negative examples from positives images, and (2)
adding image regions around the foreground estimate, as conventionally implemented in fully supervised object detection algorithms [9, 13]. For the latter, we use the criterion from [13], where
all image regions in positive images with overlap score (intersection over union with respect to any
foreground region) less than 0.3 are used as ?neighboring? negative image regions on positive images. Table 2 shows the effect of our hard negative examples on detection mean average precision for
all classes (mAP). We also added neighboring negative examples to [29], but this decreases its mAP
from 20.3% to 20.2% (before latent updates) and from 22.7% to 21.8% (after latent updates). These
experiments show that adding neighboring negative regions does not lead to noticeable improvement over not adding any negative regions from positive images, while adding our automatically
discovered hard negative regions improves detection performance more substantially.
Conclusion. We developed a weakly-supervised object detection method that discovers frequent
configurations of discriminative visual patterns. We showed that the discovered configurations provide more accurate spatial coverage of the full object and provide a way to generate useful hard
negatives. Together, these lead to state-of-the-art weakly-supervised detection results on the challenging PASCAL VOC dataset.
Acknowledgement. This work was supported in part by DARPA?s MSEE and SMISC programs, by NSF awards IIS-1427425, IIS-1212798, IIS-1116411, and by
support from Toyota.
References
[1] O. Barinova, V. Lempitsky, and P. Kohli. On detection of multiple object instances using hough transforms. IEEE TPAMI, 2012.
[2] Y. Chen, H. Shioi, C. Fuentes-Montesinos, L. Koh, S. Wich, and A. Krause. Active detection via adaptive
submodularity. In ICML, 2014.
7
Figure 5: Example configurations that have high degree in graph GP . The solid green and yellow boxes show the discovered discriminative visual parts, and the dashed magenta box shows the
bounding box that tightly fits their configuration.
8
[3] T. Deselaers, B. Alexe, and V. Ferrari. Localizing objects while learning their appearance. In ECCV,
2010.
[4] T. Deselaers, B. Alexe, and V. Ferrari. Weakly supervised localization and learning with generic knowledge. IJCV, 2012.
[5] C. Doersch, S. Singh, A. Gupta, J. Sivic, and A. A. Efros. What Makes Paris Look like Paris? In
SIGGRAPH, 2012.
[6] J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng, and T. Darrell. DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition. arXiv e-prints, 2013.
[7] A. Faktor and M. Irani. Clustering by Composition Unsupervised Discovery of Image Categories. In
ECCV, 2012.
[8] A. Farhadi and A. Sadeghi. Recognition Using Visual Phrases. In CVPR, 2011.
[9] P. Felzenszwalb, D. McAllester, and D. Ramanan. A Discriminatively Trained, Multiscale, Deformable
Part Model. In CVPR, 2008.
[10] P. Felzenszwalb, R. Girshick, D. McAllester, and D. Ramanan. Object Detection with Discriminatively
Trained Part Based Models. TPAMI, 32(9), 2010.
[11] R. Fergus, P. Perona, and A. Zisserman. Object Class Recognition by Unsupervised Scale-Invariant
Learning. In CVPR, 2003.
[12] M. Fisher, G. Nemhauser, and L. Wolsey. An analysis of approximations for maximizing submodular set
functions - II. Math. Prog. Study, 8:73?87, 1978.
[13] R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich feature hierarchies for accurate object detection
and semantic segmentation. arXiv e-prints, 2013.
[14] K. Grauman and T. Darrell. Unsupervised learning of categories from sets of partially matching image
features. In CVPR, 2006.
[15] T. Jenkyns. The efficacy of the ?greedy? algorithm. In Proc. of 7th South Eastern Conference on Combinatorics, Graph Theory and Computing, pages 341?350, 1976.
[16] M. Juneja, A. Vedaldi, C. V. Jawahar, and A. Zisserman. Blocks that Shout: Distinctive Parts for Scene
Classification. In CVPR, 2013.
[17] A. Krizhevsky and I. S. G. Hinton. ImageNet Classification with Deep Convolutional Neural Networks.
In NIPS, 2012.
[18] Y. J. Lee and K. Grauman. Foreground Focus: Unsupervised Learning From Partially Matching Images.
IJCV, 85, 2009.
[19] C. Li, D. Parikh, and T. Chen. Automatic Discovery of Groups of Objects for Scene Understanding. In
CVPR, 2012.
[20] G. Nemhauser, L. Wolsey, and M. Fisher. An analysis of approximations for maximizing submodular set
functions?I. Mathematical Programming, 14(1):265?294, 1978.
[21] M. Pandey and S. Lazebnik. Scene recognition and weakly supervised object localization with deformable
part-based models. In ICCV, 2011.
[22] D. Parikh, C. L. Zitnick, and T. Chen. From Appearance to Context-Based Recognition: Dense Labeling
in Small Images. In CVPR, 2008.
[23] T. Quack, V. Ferrari, B. Leibe, and L. V. Gool. Efficient Mining of Frequent and Distinctive Feature
Configurations. In ICCV, 2007.
[24] S. Singh, A. Gupta, and A. A. Efros. Unsupervised Discovery of Mid-level Discriminative Patches. In
ECCV, 2012.
[25] P. Siva and T. Xiang. Weakly supervised object detector learning with model drift detection. In ICCV,
2011.
[26] P. Siva, C. Russell, and T. Xiang. In defence of negative mining for annotating weakly labelled data. In
ECCV, 2012.
[27] J. Sivic and A. Zisserman. Video Data Mining Using Configurations of Viewpoint Invariant Regions. In
CVPR, 2004.
[28] J. Sivic, B. Russell, A. Efros, A. Zisserman, and W. Freeman. Discovering object categories in image
collections. In ICCV, 2005.
[29] H. O. Song, R. Girshick, S. Jegelka, J. Mairal, Z. Harchaoui, and T. Darrell. On learning to localize
objects with minimal supervision. In ICML, 2014.
[30] J. Uijlings, K. van de Sande, T. Gevers, and A. Smeulders. Selective search for object recognition. In
IJCV, 2013.
[31] V. Vizing. On an estimate of the chromatic class of a p-graph. Diskret. Analiz., 3:25?30, 1964.
[32] M. Weber, M. Welling, and P. Perona. Unsupervised Learning of Models for Recognition. In ECCV,
2000.
[33] Y. Zhang and T. Chen. Efficient Kernels for Identifying Unbounded-order Spatial Features. In CVPR,
2009.
[34] J. Zou, D. Hsu, D. Parkes, and R. Adams. Contrastive learning using spectral methods. In NIPS, 2013.
9
| 5284 |@word kohli:1 cnn:3 tedious:1 prominence:1 contrastive:1 pick:1 locomotive:1 thereby:1 solid:2 initial:4 configuration:44 contains:3 fragment:1 score:2 efficacy:1 ours:2 existing:2 activation:1 must:3 reminiscent:1 realistic:1 happen:1 informative:7 partition:2 shape:1 visible:1 update:5 greedy:6 fewer:1 guess:2 selected:4 discovering:3 lamp:1 short:1 parkes:1 math:1 node:6 location:6 preference:1 zhang:2 unbounded:1 mathematical:1 c2:1 direct:1 ik:12 qualitative:1 consists:2 ijcv:3 introduce:1 manner:1 themselves:1 frequently:7 growing:2 multi:1 inspired:2 voc:6 freeman:1 automatically:8 window:1 farhadi:1 increasing:1 begin:2 discover:3 moreover:1 bike:1 what:1 msee:1 substantially:1 developed:1 finding:6 guarantee:1 berkeley:1 sky:1 prohibitively:1 grauman:2 ramanan:2 arguably:1 positive:16 before:2 treat:1 consequence:1 approximately:1 chose:1 bird:1 challenging:5 co:14 bi:6 practice:1 union:1 block:1 xr:3 procedure:2 axiom:1 significantly:1 vedaldi:1 matching:3 word:1 onto:1 close:2 selection:4 wheel:2 undesirable:1 context:1 equivalent:1 map:3 maximizing:5 straightforward:1 regardless:1 wich:1 convex:1 focused:1 formulate:3 rectangular:3 simplicity:1 identifying:3 rule:2 insight:1 oh:1 handle:2 notion:1 variation:3 coordinate:1 ferrari:3 analiz:1 construction:1 hierarchy:1 heavily:1 exact:1 programming:1 agreement:2 element:3 expensive:1 recognition:8 sparsely:1 labeled:8 binning:2 ep:1 bottom:2 aero:1 solved:1 capture:1 worst:3 parameterize:1 region:26 ensures:1 connected:2 decrease:1 highest:2 russell:2 mine:2 trained:8 weakly:18 tight:1 singh:2 negatively:2 localization:9 bipartite:3 basement:1 completely:1 distinctive:2 easily:1 darpa:1 siggraph:1 represented:1 cat:2 train:5 effective:1 detected:1 paved:1 horse:1 labeling:1 neighborhood:8 whose:2 y1t:1 solve:1 cvpr:9 say:1 annotating:1 gp:5 jointly:1 noisy:1 transform:5 final:3 tpami:2 intelligently:1 propose:1 frequent:12 neighboring:4 relevant:1 achieve:1 deformable:3 cluster:10 darrell:5 produce:4 adam:1 object:66 depending:1 develop:1 pose:1 ij:1 nearest:2 noticeable:1 b0:10 strong:1 coverage:4 implemented:1 submodularity:1 annotated:1 correct:1 human:1 mcallester:2 bin:2 exchange:1 require:1 montesinos:1 bi1:4 sufficiently:3 around:1 ground:3 bicycle:2 alexe:2 efros:3 early:2 smallest:2 argmaxb:1 estimation:3 proc:1 sofa:2 label:11 currently:1 jawahar:1 largest:4 successfully:1 weighted:1 hoffman:1 always:2 defence:1 aim:3 rather:2 chromatic:1 mil:1 deselaers:2 derived:3 focus:1 improvement:2 consistently:1 check:1 mainly:1 contrast:1 greedily:1 baseline:2 membership:1 typically:3 entire:2 perona:2 relation:1 selective:3 i1:3 selects:1 issue:1 among:1 classification:2 pascal:10 retaining:1 development:1 constrained:4 art:5 spatial:4 tzeng:1 construct:1 shaped:1 cartoon:1 identical:5 look:3 icml:2 unsupervised:6 foreground:19 np:1 report:1 employ:1 few:1 simultaneously:1 densely:1 tightly:1 ourselves:1 detection:36 interest:3 mining:5 highly:2 intra:3 evaluation:1 laborious:1 adjust:1 accurate:4 edge:11 partial:1 hough:1 abundant:1 re:1 mbk:1 deformation:1 girshick:3 theoretical:2 minimal:3 delete:2 smisc:1 instance:5 column:1 earlier:1 modeling:2 cover:7 localizing:1 phrase:3 cost:1 krizhevsky:1 xrf:4 examining:1 too:7 closedness:1 person:2 density:1 retain:3 lee:2 picking:2 together:6 intersecting:1 nm:1 admit:1 ek:7 li:1 account:1 potential:1 de:1 diversity:1 star:1 b2:4 availability:1 includes:1 satisfy:1 combinatorics:1 explicitly:3 piece:1 picked:1 analyze:3 red:2 sort:1 annotation:8 gevers:1 jia:1 contribution:1 smeulders:1 accuracy:2 convolutional:3 characteristic:4 efficiently:1 correspond:2 identify:5 yield:1 yellow:4 weak:1 identification:1 xli:1 accurately:1 worth:1 shout:1 alternatingly:1 detector:12 trevor:1 definition:1 shp:1 failure:1 against:1 proof:2 mi:1 hsu:1 dataset:8 color:3 car:2 improves:3 torso:2 cj:5 knowledge:1 segmentation:1 siva:2 coloring:1 centric:1 higher:2 supervised:18 follow:1 zisserman:4 improved:1 formulation:6 box:17 strongly:1 shrink:1 stage:3 until:2 horizontal:1 web:1 multiscale:1 overlapping:8 defines:2 effect:3 contain:4 true:2 hence:2 read:1 irani:1 iteratively:1 i2:2 illustrated:1 white:3 semantic:1 round:1 adjacent:2 transitivity:1 latch:1 davis:1 covering:5 prohibit:1 shioi:1 criterion:4 prominent:2 complete:1 demonstrate:2 image:54 lazebnik:1 weber:1 discovers:3 parikh:2 common:2 diversifying:1 significant:1 composition:3 doersch:1 automatic:1 xlf:4 inclusion:1 submodular:11 phrasing:1 supervision:8 etc:1 add:4 fc7:1 closest:2 recent:5 showed:1 noisily:1 scenario:1 certain:1 sande:1 continue:1 caltech:2 scoring:1 mislocalizations:3 yft:4 additional:2 determine:1 maximize:4 redundant:2 dashed:2 u0:1 ii:4 multiple:9 full:13 harchaoui:1 doublet:1 visit:1 award:1 paired:1 impact:2 arxiv:2 iteration:1 represent:2 kernel:1 achieved:1 c1:1 proposal:1 background:2 addition:2 uninformative:1 krause:1 extra:1 strict:1 south:1 subject:2 induced:2 contrary:1 presence:2 noting:1 enough:2 independence:2 remedying:1 matroid:4 fit:1 identified:2 restrict:2 cow:1 reduce:2 idea:3 regarding:1 faktor:1 absent:2 shift:2 whether:3 motivated:2 utility:1 aeroplane:2 effort:1 song:4 returned:1 constitute:1 deep:4 useful:3 detailed:3 aimed:1 xr2:1 amount:4 clutter:1 transforms:1 mid:2 category:8 generate:6 outperform:1 exist:1 problematic:1 nsf:1 xr1:1 disjoint:2 per:3 track:1 diverse:1 write:1 affected:1 group:1 key:1 four:3 threshold:3 yit:1 localize:1 btl:1 vast:1 graph:10 merely:2 monotone:2 fraction:1 run:1 extends:1 throughout:2 saying:1 family:3 ruling:1 prog:1 patch:52 draw:1 lsvm:2 bound:1 correspondence:6 occur:10 constraint:6 constrain:1 scene:3 phrased:1 yong:1 aspect:2 min:2 relatively:1 px:6 tv:1 according:1 combination:3 across:1 modification:1 intuitively:2 invariant:2 iccv:4 koh:1 equation:2 mutually:1 remains:1 bus:2 count:3 loose:1 end:1 available:1 operation:1 observe:1 leibe:1 spectral:2 generic:2 occurrence:2 robustly:1 alternative:1 motorbike:2 assumes:1 running:1 remaining:2 top:2 include:1 clustering:1 exploit:1 especially:1 build:1 malik:1 added:1 print:2 nemhauser:2 y2t:1 argue:2 water:1 relationship:1 illustration:1 ratio:2 difficult:2 xri:1 negative:38 rise:1 implementation:1 packed:1 unknown:1 fuentes:1 discretize:1 vertical:1 observation:1 upper:1 datasets:3 discarded:1 hyun:1 truncated:1 immediate:1 hinton:1 ever:1 head:1 frame:1 discovered:23 gc:9 arbitrary:2 drift:1 inferred:1 pair:9 dog:1 paris:2 c3:1 imagenet:1 sivic:3 california:2 learned:1 deletion:1 decreasingly:1 inaccuracy:1 nip:2 address:2 beyond:1 usually:1 pattern:11 juneja:1 program:1 max:4 green:4 video:1 gool:1 overlap:15 bi2:4 treated:1 rely:1 circumvent:1 boat:1 scarce:1 scheme:2 improve:1 sadeghi:1 imply:1 identifies:3 conventionally:1 stefanie:1 transitive:1 extract:1 geometric:1 discovery:7 discouraging:1 sg:2 checking:1 acknowledgement:1 relative:9 understanding:1 xiang:2 fully:2 discriminatively:2 wolsey:2 localized:2 degree:4 jegelka:2 consistent:1 viewpoint:4 share:2 translation:2 row:1 eccv:5 supported:1 copy:2 eastern:1 neighbor:7 fall:1 face:3 taking:1 felzenszwalb:2 matroids:8 absolute:1 benefit:2 van:1 boundary:2 contour:1 rich:1 commonly:2 xl2:1 coincide:1 simplified:1 adaptive:1 collection:1 ec:2 cope:2 welling:1 implicitly:1 keep:1 active:1 mairal:1 b1:4 discriminative:23 fergus:1 search:3 latent:5 pandey:1 table:4 learn:1 robust:4 quantize:1 zou:1 uijlings:1 protocol:1 zitnick:1 did:1 main:2 dense:1 bounding:7 y1b:1 noise:1 jae:1 positively:4 body:4 augmented:2 textually:1 representative:4 precision:3 position:2 yib:1 xl:3 candidate:6 lie:1 toyota:1 donahue:2 theorem:1 down:1 erroneous:1 magenta:3 barinova:1 showing:1 constellation:1 explored:1 list:2 svm:2 gupta:2 grouping:1 albeit:1 false:1 effectively:2 adding:5 ci:5 decaf:1 texture:1 downward:1 occurring:13 demand:3 chen:4 intersection:7 simply:2 appearance:4 likely:2 visual:16 vinyals:1 expressed:1 partially:2 xl1:1 corresponds:2 truth:1 relies:1 lempitsky:1 identity:1 goal:2 consequently:1 labelled:1 fisher:2 feasible:1 hard:27 change:2 typical:1 specifically:1 lemma:3 total:1 invariance:1 meaningful:1 rarely:3 select:5 chr:1 people:1 support:1 latter:2 |
4,732 | 5,285 | Encoding High Dimensional Local Features by Sparse
Coding Based Fisher Vectors
Lingqiao Liu1 , Chunhua Shen1,2 , Lei Wang3 , Anton van den Hengel1,2 , Chao Wang3
1
School of Computer Science, University of Adelaide, Australia
2
ARC Centre of Excellence for Robotic Vision
3
School of Computer Science and Software Engineering, University of Wollongong, Australia
Abstract
Deriving from the gradient vector of a generative model of local features, Fisher
vector coding (FVC) has been identified as an effective coding method for image classification. Most, if not all, FVC implementations employ the Gaussian
mixture model (GMM) to characterize the generation process of local features.
This choice has shown to be sufficient for traditional low dimensional local features, e.g., SIFT; and typically, good performance can be achieved with only a few
hundred Gaussian distributions. However, the same number of Gaussians is insufficient to model the feature space spanned by higher dimensional local features,
which have become popular recently. In order to improve the modeling capacity
for high dimensional features, it turns out to be inefficient and computationally
impractical to simply increase the number of Gaussians.
In this paper, we propose a model in which each local feature is drawn from a
Gaussian distribution whose mean vector is sampled from a subspace. With certain approximation, this model can be converted to a sparse coding procedure and
the learning/inference problems can be readily solved by standard sparse coding
methods. By calculating the gradient vector of the proposed model, we derive
a new fisher vector encoding strategy, termed Sparse Coding based Fisher Vector Coding (SCFVC). Moreover, we adopt the recently developed Deep Convolutional Neural Network (CNN) descriptor as a high dimensional local feature
and implement image classification with the proposed SCFVC. Our experimental evaluations demonstrate that our method not only significantly outperforms
the traditional GMM based Fisher vector encoding but also achieves the state-ofthe-art performance in generic object recognition, indoor scene, and fine-grained
image classification problems.
1
Introduction
Fisher vector coding is a coding method derived from the Fisher kernel [1] which was originally proposed to compare two samples induced by a generative model. Since its introduction to computer
vision [2], many improvements and variants have been proposed. For example, in [3] the normalization of Fisher vectors is identified as an essential step to achieve good performance; in [4] the spatial
information of local features is incorporated; in [5] the model parameters are learned through a endto-end supervised training algorithm and in [6] multiple layers of Fisher vector coding modules are
stacked into a deep architecture. With these extensions, Fisher vector coding has been established
as the state-of-the-art image classification approach.
Almost all of these methods share one common component: they all employ Gaussian mixture
model (GMM) as the generative model for local features. This choice has been proved effective in
modeling standard local features such as SIFT, which are often of low dimension. Usually, using a
1
mixture of a few hundred Gaussians has been sufficient to guarantee good performance. Generally
speaking, the distribution of local features can only be well captured by a Gaussian distribution
within a local region due to the variety of local feature appearances and thus the number of Gaussian
mixtures needed is essentially determined by the volume of the feature space of local features.
Recently, the choice of local features has gone beyond the traditional local patch descriptors such as
SIFT or SURF [7] and higher dimensional local features such as the activation of a pre-trained deep
neural-network [8] or pooled coding vectors from a local region [9, 10] have demonstrated promising performance. The higher dimensionality and rich visual content captured by those features make
the volume of their feature space much larger than that of traditional local features. Consequently, a
much larger number of Gaussian mixtures will be needed in order to model the feature space accurately. However, this would lead to the explosion of the resulted image representation dimensionality
and thus is usually computationally impractical.
To alleviate this difficulty, here we propose an alternative solution. We model the generation process
of local features as randomly drawing features from a Gaussian distribution whose mean vector is
randomly drawn from a subspace. With certain approximation, we convert this model to a sparse
coding model and leverage an off-the-shelf solver to solve the learning and inference problems.
With further derivation, this model leads to a new Fisher vector coding algorithm called Sparse
Coding based Fisher Vector Coding (SCFVC). Moreover, we adopt the recently developed Deep
Convolutional Neural Network to generate regional local features and apply the proposed SCFVC
to these local features to build an image classification system.
To demonstrate its effectiveness in encoding the high dimensional local feature, we conduct a series
of experiments on generic object, indoor scene and fine-grained image classification datasets, it is
shown that our method not only significantly outperforms the traditional GMM based Fisher vector
coding in encoding high dimensional local features but also achieves state-of-the-art performance in
these image classification problems.
2
2.1
Fisher vector coding
General formulation
Given two samples generated from a generative model, their similarity can be evaluated by using a
Fisher kernel [1]. The sample can take any form, including a vector or a vector set, as long as its generation process can be modeled. For a Fisher vector based image classification approach, the sample
is a set of local features extracted from an image which we denote it as X = {x1 , x2 , ? ? ? , xT }.
Assuming each xi is modeled by a p.d.f P (x|?) and is drawn i.i.d, in Fisher kernel a sample X can
be described by the gradient vector over the model parameter ?
X
GX
?? log P (xi |?).
(1)
? = ?? log P (X|?) =
i
T ?1 X
GX
G? ,
? F
The Fisher kernel is then defined as K(X, Y) =
where F is the information matrix
X XT
and is defined as F = E[G? G? ]. In practice, the role of the information matrix is less significant
and is often omitted for computational simplicity [3]. As a result, two samples can be directly
compared by the linear kernel of their corresponding gradient vectors which are often called Fisher
vectors. From a bag-of-features model perspective, the evaluation of the Fisher kernel for two
images can be seen as first calculating the gradient or Fisher vector of each local feature and then
performing sum-pooling. In this sense, the Fisher vector calculation for each local feature can be
seen as a coding method and we call it Fisher vector coding in this paper.
2.2
GMM based Fisher vector coding and its limitation
To implement the Fisher vector coding framework introduced above, one needs to specify the distribution P (x|?). In the literature, most, if not all, works choose GMM to model the generation
process of x, which can be described as follows:
? Draw a Gaussian model N (?k , ?k ) from the prior distribution P (k), k = 1, 2, ? ? ? , m .
? Draw a local feature x from N (?k , ?k ).
2
Generally speaking, the distribution of x resembles a Gaussian distribution only within a local region
of feature space. Thus, for a GMM, each of Gaussian essentially models a small partition of the
feature space and many of them are needed to depict the whole feature space. Consequently, the
number of mixtures needed will be determined by the volume of the feature space. For the commonly
used low dimensional local features, such as SIFT, it has been shown that it is sufficient to set the
number of mixtures to few hundreds. However, for higher dimensional local features this number
may be insufficient. This is because the volume of feature space usually increases quickly with the
feature dimensionality. Consequently, the same number of mixtures will result in a coarser partition
resolution and imprecise modeling.
To increase the partition resolution for higher dimensional feature space, one straightforward solution is to increase the number of Gaussians. However, it turns out that the partition resolution
increases slowly (compared to our method which will be introduced in the next section) with the
number of mixtures. In other words, much larger number of Gaussians will be needed and this will
result in a Fisher vector whose dimensionality is too high to be handled in practice.
3
3.1
Our method
Infinite number of Gaussians mixture
Our solution to this issue is to go beyond a fixed number of Gaussian distributions and use an
infinite number of them. More specifically, we assume that a local feature is drawn from a Gaussian
distribution with a randomly generated mean vector. The mean vector is a point on a subspace
spanned by a set of bases (which can be complete or over-complete) and is indexed by a latent
coding vector u. The detailed generation process is as follows:
? Draw a coding vector u from a zero mean Laplacian distribution P (u) =
1
2?
exp(? |u|
? ).
? Draw a local feature x from the Gaussian distribution N (Bu, ?),
where the Laplace prior for P (u) ensures the sparsity of resulting Fisher vector which can be helpful
for coding. Essentially, the above process resembles a sparse coding model. To show this relationship, let?s first write the marginal distribution of x:
Z
Z
P (x) =
P (x, u|B)du =
P (x|u, B)P (u)du.
(2)
u
u
The above formulation involves an integral operator which makes the likelihood evaluation difficult.
To simplify the calculation, we use the point-wise maximum within the integral term to approximate
the likelihood, that is,
P (x) ? P (x|u? , B)P (u? ).
u? = argmax P (x|u, B)P (u)
(3)
u
2
2
By assumming that ? = diag(?12 , ? ? ? , ?m
) and setting ?12 = ? ? ? = ?m
= ? 2 as a constant. The
logarithm of P (x) is written as
log(P (x|B)) = min
u
1
kx ? Buk22 + ?kuk1 ,
?2
(4)
which is exactly the objective value of a sparse coding problem. This relationship suggests that we
can learn the model parameter B and infer the latent variable u by using the off-the-shelf sparse
coding solvers.
One question for the above method is that compared to simply increasing the number of models
in traditional GMM, how much improvement is achieved by increasing the partition resolution. To
answer this question, we designed an experiment to compare these two schemes. In our experiment,
the partition resolution is roughly measured by the average distance (denoted as d ) between a feature
and its closest mean vector in the GMM or the above model. The larger d is, the lower the partition
resolution is. The comparison is shown in Figure 1. In Figure 1 (a), we increase the dimensionality
3
3
2.55
2.8
2.5
2.6
2.45
2.4
d
d
of local features 1 and for each dimension we calculate d in a GMM model with 100 mixtures. As
seen, d increases quickly with the feature dimensionality. In Figure 1 (b), we try to reduce d by
introducing more mixture distributions in GMM model. However, as can be seen, d drops slowly
with the increase in the number of Gaussians. In contrast, with the proposed method, we can achieve
much lower d by using only 100 bases. This result illustrates the advantage of our method.
2.2
2.4
2.35
GMM
GMM with 100 mixtures
2
1.8
100
Proposed model (with 100 bases)
2.3
200
300
400
500
600
700
800
900
1000
Dimensionality of regional local features
2.25
100
200
300
400
500
600
700
800
900
1000
Number of Gaussian mixtures
(a)
(b)
Figure 1: Comparison of two ways to increase the partition resolution. (a) For GMM, d (the average distance between a local feature and its closest mean vector) increases with the local feature
dimensionality. Here the GMM is fixed at 100 Gaussians. (b) d is reduced in two ways (1) simply
increasing the number of Gaussian distributions in the mixture. (2) using the proposed generation
process. As can be seen, the latter achieves much lower d even with a small number of bases.
3.2
Sparse coding based Fisher vector coding
Once the generative model of local features is established, we can readily derive the corresponding
Fisher coding vector by differentiating its log likelihood, that is,
? 12 kx ? Bu? k22 + ?ku? k1
? log(P (x|B))
= ?
?B
?B
u? = argmax P (x|u, B)P (u).
C(x) =
(5)
u
Note that the differentiation involves u? which implicitly interacts with B. To calculate this term,
we notice that the sparse coding problem can be reformulated as a general quadratic programming
problem by defining u+ and u? as the positive and negative parts of u, that is, the sparse coding
problem can be rewritten as
min
u+ ,u?
kx ? B(u+ ? u? )k22 + ?1T (u+ + u? )
s.t. u+ ? 0
0
+
u? ? 0
(6)
? T
By further defining u = (u , u ) , log(P (x|B)) can be expressed in the following general form,
1 T
T
log(P (x|B)) = L(B) = max
u0 v(B) ? u0 P(B)u0 ,
u0
2
(7)
where P(B) and v(B) are a matrix term and a vector term depending on B respectively. The
derivative of L(B) has been studied in [11]. According to the Lemma 2 in [11], we can differentiate
L(B) with respect to B as if u0 did not depend on B. In other words, we can firstly calculate u0 or
(x|B))
equivalently u? by solving the sparse coding problem and then obtain the Fisher vector ? log(P
?B
as
? ?12 kx ? Bu? k22 + ?ku? k1
= (x ? Bu? )u? T .
?B
(8)
1
This is achieved by performing PCA on a 4096-dimensional CNN regional descriptor. For more details
about the feature we used, please refer to Section 3.4
4
Table 1: Comparison of results on Pascal VOC 2007. The lower part of this table lists some results
reported in the literature. We only report the mean average precision over 20 classes. The average
precision for each class is listed in Table 2.
Methods
mean average precision
Comments
SCFVC (proposed)
GMMFVC
76.9%
73.8%
single scale, no augmented data
single scale, no augmented data
CNNaug-SVM [8]
CNN-SVM [8]
NUS [13]
GHM [14]
AGS [15]
77.2%
73.9%
70.5%
64.7%
71.1%
with augmented data, use CNN for whole image
no augmented data.use CNN for whole image
-
Note that the Fisher vector expressed in Eq. (8) has an interesting form: it is simply the outer product
between the sparse coding vector u? and the reconstruction residual term (x ? Bu? ). In traditional
sparse coding, only the kth dimension of a coding vector uk is used to indicate the relationship
between a local feature x and the kth basis. Here in the sparse coding based Fisher vector, the
coding value uk multiplied by the reconstruction residual is used to capture their relationship.
3.3
Pooling and normalization
From the i.i.d assumption in Eq. (1), the Fisher vector of the whole image is 2
X ? log(P (xi |B))
X
? log(P (I|B))
=
=
(xi ? Bu?i )u?i > .
?B
?B
xi ?I
(9)
xi ?I
This is equivalent to performing the sum-pooling for the extracted Fisher coding vectors. However,
it has been observed [3, 12] that the image signature obtained using sum-pooling tends to overemphasize the information from the background [3] or bursting visual words [12]. It is important to
apply normalization when sum-pooling is used. In this paper, we apply intra-normalization [12] to
normalize
the pooled Fisher vectors. More specifically, we apply l2 normalization to the subvectors
P
(x
? Bu?i )u?i,k ?k, where k indicates the kth dimension of the sparse coding u?i . Besides
i
xi ?I
intra-normalization, we also utilize the power normalization as suggested in [3].
3.4
Deep CNN based regional local features
Recently, the middle-layer activation of a pre-trained deep CNN has been demonstrated to be a
powerful image descriptor [8, 16]. In this paper, we employ this descriptor to generate a number
of local features for an image. More specifically, an input image is first resized to 512?512 pixels
and regions with the size of 227?227 pixels are cropped with the stride 8 pixels. These regions
are subsequently feed into the deep CNN and the activation of the sixth layer is extracted as local
features for these regions. In our implementation, we use the Caffe [17] package which provides a
deep CNN pre-trained on ILSVRC2012 dataset and its 6-th layer is a 4096-dimensional vector. This
strategy has demonstrated better performance than directly using deep CNN features for the whole
image recently [16].
Once regional local features are extracted, we encoded them using the proposed SCFVC method and
generate an image level representation by sum-pooling and normalization. Certainly, our method is
open to the choice of other high-dimensional local features. The reason for choosing deep CNN
features in this paper is that by doing so we can demonstrate state-of-the-art image classification
performance.
4
Experimental results
We conduct experimental evaluation of the proposed sparse coding based Fisher vector coding
(SCFVC) on three large datasets: Pascal VOC 2007, MIT indoor scene-67 and Caltech-UCSD Birds2
the vectorized form of
? log(P (I|B))
?B
is used as the image representation.
5
Table 2: Comparison of results on Pascal VOC 2007 for each of 20 classes. Besides the proposed
SCFVC and the GMMFVC baseline, the performance obtained by directly using CNN as global
feature is also compared.
SCFVC
GMMFVC
CNN-SVM
SCFVC
GMMFVC
CNN-SVM
aero
bike
bird
boat
bottle
bus
car
cat
chair
cow
89.5
87.1
88.5
84.1
80.6
81.0
83.7
80.3
83.5
83.7
79.7
82.0
43.9
42.8
42.0
76.7
72.2
72.5
87.8
87.4
85.3
82.5
76.1
81.6
60.6
58.6
59.9
69.6
64.0
58.5
table
dog
horse
mbike
person
plant
sheep
sofa
train
TV
72.0
66.9
66.5
77.1
75.1
77.8
88.7
84.9
81.8
82.1
81.2
78.8
94.4
93.1
90.2
56.8
53.1
54.8
71.4
70.8
71.1
67.7
66.2
62.6
90.9
87.9
87.2
75.0
71.3
71.8
Table 3: Comparison of results on MIT-67. The lower part of this table lists some results reported
in the literature.
Methods
Classification Accuracy
Comments
SCFVC (proposed)
GMMFVC
68.2%
64.3%
with single scale
with single scale
MOP-CNN [16]
VLAD level2 [16]
CNN-SVM [8]
FV+Bag of parts [19]
DPM [20]
68.9%
65.5%
58.4%
63.2%
37.6%
with three scales
with single best scale
use CNN for whole image
-
200-2011. These are commonly used evaluation benchmarks for generic object classification, scene
classification and fine-grained image classification respectively. The focus of these experiments is
to examine that whether the proposed SCFVC outperforms the traditional Fisher vector coding in
encoding high dimensional local features.
4.1
Experiment setup
In our experiments, the activations of the sixth layer of a pre-trained deep CNN are used as regional
local features. PCA is applied to further reduce the regional local features from 4096 dimensions to
2000 dimensions. The number of Gaussian distributions and the codebook size for sparse coding is
set to 100 throughout our experiments unless otherwise mentioned.
For the sparse coding, we use the algorithm in [18] to learn the codebook and perform the coding
vector inference. For all experiments, linear SVM is used as the classifier.
4.2
Main results
Pascal-07 Pascal VOC 2007 contains 9963 images with 20 object categories which form 20 binary
(object vs. non-object) classification tasks. The use of deep CNN features has demonstrated the
state-of-the-art performance [8] on this dataset. In contrast to [8], here we use the deep CNN features as local features to model a set of image regions rather than as a global feature to model the
whole image. The results of the proposed SCFVC and traditional Fisher vector coding, denoted as
GMMFVC, are shown in Table 1 and Table 2. As can be seen from Table 1, the proposed SCFVC
leads to superior performance over the traditional GMMFVC and outperforms GMMFVC by 3%.
By cross-referencing Table 2, it is clear that the proposed SCFVC outperforms GMMFVC in all
of 20 categories. Also, we notice that the GMMFVC is merely comparable to the performance of
directly using deep CNN as global features, namely, CNN-SVM in Table 1. Since both the proposed
SCFVC and GMMFVC adopt deep CNN features as local features, this observation suggests that
the advantage of using deep CNN features as local features can only be clearly demonstrated when
the appropriate coding method, i.e. the proposed SCFVC is employed. Note that to further boost the
6
Table 4: Comparison of results on Birds-200 2011. The lower part of this table lists some results
reported in the literature.
Methods
Classification Accuracy
Comments
SCFVC (proposed)
GMMFVC
66.4%
61.7%
with single scale
with single scale
CNNaug-SVM [8]
CNN-SVM [8]
DPD+CNN+LogReg [21]
DPD [22]
61.8%
53.3%
65.0%
51.0%
with augmented data, use CNN for the whole image
no augmented data, use CNN as global features
use part information
-
69
Classification Accuracy %
68
67
66
65
64
63
SCFV
GMMFV
62
61
0
200
400
600
800
1000
1200
1400
1600
1800
2000
Dimensionality of regional local features
Figure 2: The performance comparison of classification accuracy vs. local feature dimensionality
for the proposed SCFVC and GMMFVC on MIT-67.
performance, one can adopt some additional approaches like introducing augmented data or combining multiple scales. Some of the methods compared in Table 1 have employed these approaches
and we have commented this fact as so inform readers that whether these methods are directly comparable to the proposed SCFVC. We do not pursue these approaches in this paper since the focus of
our experiment is to compare the proposed SCFVC against traditional GMMFVC.
MIT-67 MIT-67 contains 6700 images with 67 indoor scene categories. This dataset is quite challenging because the differences between some categories are very subtle. The comparison of classification results are shown in Table 3. Again, we observe that the proposed SCFVC significantly outperforms traditional GMMFVC. To the best of our knowledge, the best performance on this dataset
is achieved in [16] by concatenating the features extracted from three different scales. The proposed
method achieves the same performance only using a single scale. We also tried to concatenate the
image representation generated from the proposed SCFVC with the global deep CNN feature. The
resulted performance can be as high as 70% which is by far the best performance achieved on this
dataset.
Birds-200-2011 Birds-200-2011 contains 11788 with 200 different birds species, which is a commonly used benchmark for fine-grained image classification. The experimental results on this
dataset are shown in Table 4. The advantage of SCFVC over GMMFVC is more pronounced on
this dataset: SCFVC outperforms GMMFVC by over 4%. We also notice two interesting observations: (1) GMMFVC even achieves comparable performance to the method of using the global
deep CNN feature with augmented data, namely, CNNaug-SVM in Table 4. (2) Although we do
not use any parts information (of birds), our method outperforms the result using parts information
(DPD+CNN+LogReg in Table 4). These two observations suggest that using deep CNN features
as local features is better for fine-grained problems and the proposed method can further boost its
advantage.
7
Table 5: Comparison of results on MIT-67 with three different settings: (1) 100-basis codebook with
1000 dimensional local features, denoted as SCFV-100-1000D (2) 400 Gaussian mixtures with 300
dimensional local features, denoted as GMMFV-400-300D (3) 1000 Gaussian mixtures with 100
dimensional local features denoted as GMMFV-1000-100D. They have the same/similar total image
representation dimensionality.
SCFV-100-1000D
68.1%
4.3
GMMFV-400-300D
64.0%
GMMFV-1000-100D
60.8%
Discussion
In the above experiments, the dimensionality of local features is fixed to 2000. But how about the
performance comparison between the proposed SCFV and traditional GMMFV on lower dimensional features? To investigate this issue, we vary the dimensionality of the deep CNN features from
100 to 2000 and compare the performance of the two Fisher vector coding methods on MIT-67. The
results are shown in Figure 2. As can be seen, for lower dimensionality (like 100), the two methods
achieve comparable performance and in general both methods benefit from using higher dimensional
features. However, for traditional GMMFVC, the performance gain obtained from increasing feature dimensionality is lower than that obtained by the proposed SCFVC. For example, from 100 to
1000 dimensions, the traditional GMMFVC only obtains 4% performance improvement while our
SCFVC achieves 7% performance gain. This validates our argument that the proposed SCFVC is
especially suited for encoding high dimensional local features.
Since GMMFVC works well for lower dimensional features, how about reducing the higher dimensional local features to lower dimensions and use more Gaussian mixtures? Will it be able to achieve
comparable performance to our SCFVC which uses higher dimensional local features but a smaller
number of bases? To investigate this issue, we also evaluate the classification performance on MIT67 using 400 Gaussian mixtures with 300-dimension local features and 1000 Gaussian mixtures with
100-dimension local features. Thus the total dimensionality of these two image representations will
be similar to that of our SCFVC which uses 100 bases and 1000-dimension local features. The comparison is shown in Table 5. As can be seen, the performance of these two settings are much inferior
to the proposed one. This suggests that some discriminative information may have already been lost
after the PCA dimensionality reduction and the discriminative power can not be re-boosted by simply introducing more Gaussian distributions. This verifies the necessity of using high dimensional
local features and justifies the value of the proposed method.
In general, the inference step in sparse coding can be slower than the membership assignment in
GMM model. However, the computational efficiency can be greatly improved by using an approximated sparse coding algorithm such as learned FISTA [23] or orthogonal matching pursuit [10].
Also, the proposed method can be easily generalized to several similar coding models, such as local
linear coding [24]. In that case, the computational efficiency is almost identical (or even faster if
approximated k-nearest neighbor algorithms are used) to the traditional GMMFVC.
5
Conclusion
In this work, we study the use of Fisher vector coding to encode high-dimensional local features.
Our main discovery is that traditional GMM based Fisher vector coding is not particular well suited
to modeling high-dimensional local features. As an alternative, we proposed to use a generation
process which allows the mean vector of a Gaussian distribution to be chosen from a point in a
subspace. This model leads to a new Fisher vector coding method which is based on sparse coding
model. Combining with the activation of the middle layer of a pre-trained CNN as high-dimensional
local features, we build an image classification system and experimentally demonstrate that the
proposed coding method is superior to the traditional GMM in encoding high-dimensional local
features and can achieve state-of-the-art performance in three image classification problems.
Acknowledgements This work was in part supported by Australian Research Council grants
FT120100969, LP120200485, and the Data to Decisions Cooperative Research Centre. Correspondence should be addressed to C. Shen (email: [email protected]).
8
References
[1] T. Jaakkola and D. Haussler, ?Exploiting generative models in discriminative classifiers,? in Proc. Adv.
Neural Inf. Process. Syst., 1998, pp. 487?493.
[2] F. Perronnin and C. R. Dance, ?Fisher kernels on visual vocabularies for image categorization,? in Proc.
IEEE Conf. Comp. Vis. Patt. Recogn., 2007.
[3] F. Perronnin, J. S?anchez, and T. Mensink, ?Improving the Fisher kernel for large-scale image classification,? in Proc. Eur. Conf. Comp. Vis., 2010.
[4] J. Krapac, J. J. Verbeek, and F. Jurie, ?Modeling spatial layout with fisher vectors for image categorization.? in Proc. IEEE Int. Conf. Comp. Vis., 2011, pp. 1487?1494.
[5] K. Simonyan, A. Vedaldi, and A. Zisserman, ?Deep fisher networks for large-scale image classification,?
in Proc. Adv. Neural Inf. Process. Syst., 2013.
[6] V. Sydorov, M. Sakurada, and C. H. Lampert, ?Deep fisher kernels?end to end learning of the Fisher
kernel GMM parameters,? in Proc. IEEE Conf. Comp. Vis. Patt. Recogn., 2014.
[7] H. Bay, A. Ess, T. Tuytelaars, and L. J. V. Gool, ?Speeded-up robust features (SURF),? Computer Vision
& Image Understanding, vol. 110, no. 3, pp. 346?359, 2008.
[8] A. S. Razavian, H. Azizpour, J. Sullivan, and S. Carlsson, ?CNN features off-the-shelf: an astounding
baseline for recognition,? 2014, http://arxiv.org/abs/1403.6382.
[9] S. Yan, X. Xu, D. Xu, S. Lin, and X. Li, ?Beyond spatial pyramids: A new feature extraction framework
with dense spatial sampling for image classification,? in Proc. Eur. Conf. Comp. Vis., 2012, pp. 473?487.
[10] L. Bo, X. Ren, and D. Fox, ?Hierarchical matching pursuit for image classification: Architecture and fast
algorithms,? in Proc. Adv. Neural Inf. Process. Syst., 2011, pp. 2115?2123.
[11] O. Chapelle, V. Vapnik, O. Bousquet, and S. Mukherjee, ?Choosing multiple parameters for support vector
machines,? Machine Learning, vol. 46, no. 1?3, pp. 131?159, 2002.
[12] R. Arandjelovi?c and A. Zisserman, ?All about VLAD,? in Proc. IEEE Int. Conf. Comp. Vis., 2013.
[13] Z. Song, Q. Chen, Z. Huang, Y. Hua, and S. Yan, ?Contextualizing object detection and classification.? in
Proc. IEEE Conf. Comp. Vis. Patt. Recogn., 2011.
[14] Q. Chen, Z. Song, Y. Hua, Z. Huang, and S. Yan, ?Hierarchical matching with side information for image
classification.? in Proc. IEEE Conf. Comp. Vis. Patt. Recogn., 2012, pp. 3426?3433.
[15] J. Dong, W. Xia, Q. Chen, J. Feng, Z. Huang, and S. Yan, ?Subcategory-aware object classification.? in
Proc. IEEE Conf. Comp. Vis. Patt. Recogn., 2013, pp. 827?834.
[16] Y. Gong, L. Wang, R. Guo, and S. Lazebnik, ?Multi-scale orderless pooling of deep convolutional activation features,? in Proc. Eur. Conf. Comp. Vis., 2014.
[17] Y. Jia, ?Caffe,? 2014, https://github.com/BVLC/caffe.
[18] H. Lee, A. Battle, R. Raina, and A. Y. Ng, ?Efficient sparse coding algorithms,? in Proc. Adv. Neural Inf.
Process. Syst., 2007, pp. 801?808.
[19] C. Doersch, A. Gupta, and A. A. Efros, ?Mid-level visual element discovery as discriminative mode
seeking,? in Proc. Adv. Neural Inf. Process. Syst., 2013.
[20] M. Pandey and S. Lazebnik, ?Scene recognition and weakly supervised object localization with deformable part-based models,? in Proc. IEEE Int. Conf. Comp. Vis., 2011, pp. 1307?1314.
[21] J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng, and T. Darrell, ?DeCAF: A deep convolutional activation feature for generic visual recognition,? in Proc. Int. Conf. Mach. Learn., 2013.
[22] N. Zhang, R. Farrell, F. Iandola, and T. Darrell, ?Deformable part descriptors for fine-grained recognition
and attribute prediction,? in Proc. IEEE Int. Conf. Comp. Vis., December 2013.
[23] K. Gregor and Y. LeCun, ?Learning fast approximations of sparse coding,? in Proc. Int. Conf. Mach.
Learn., 2010, pp. 399?406.
[24] J. Wang, J. Yang, K. Yu, F. Lv, T. Huang, and Y. Gong, ?Locality-constrained linear coding for image
classification,? in Proc. IEEE Conf. Comp. Vis. Patt. Recogn., 2010.
9
| 5285 |@word cnn:35 middle:2 open:1 tried:1 reduction:1 necessity:1 series:1 contains:3 outperforms:8 com:2 activation:7 gmail:1 written:1 readily:2 concatenate:1 partition:8 designed:1 drop:1 depict:1 v:2 generative:6 fvc:2 es:1 provides:1 codebook:3 gx:2 firstly:1 org:1 zhang:2 become:1 excellence:1 roughly:1 examine:1 multi:1 voc:4 solver:2 increasing:4 subvectors:1 moreover:2 bike:1 pursue:1 developed:2 ag:1 differentiation:1 impractical:2 guarantee:1 exactly:1 classifier:2 uk:2 grant:1 positive:1 engineering:1 local:69 kuk1:1 tends:1 encoding:8 mach:2 bird:6 resembles:2 studied:1 bursting:1 suggests:3 challenging:1 gone:1 speeded:1 jurie:1 lecun:1 practice:2 lost:1 implement:2 sullivan:1 procedure:1 yan:4 significantly:3 vedaldi:1 matching:3 imprecise:1 pre:5 word:3 suggest:1 operator:1 equivalent:1 demonstrated:5 straightforward:1 go:1 layout:1 resolution:7 shen:1 simplicity:1 haussler:1 deriving:1 spanned:2 laplace:1 programming:1 us:2 element:1 recognition:5 approximated:2 mukherjee:1 coarser:1 cooperative:1 observed:1 role:1 module:1 aero:1 solved:1 capture:1 wang:2 calculate:3 region:7 ensures:1 ilsvrc2012:1 adv:5 mentioned:1 signature:1 trained:5 depend:1 solving:1 weakly:1 localization:1 efficiency:2 basis:2 logreg:2 easily:1 cat:1 recogn:6 derivation:1 stacked:1 train:1 fast:2 effective:2 horse:1 choosing:2 caffe:3 whose:3 encoded:1 larger:4 solve:1 quite:1 drawing:1 wang3:2 otherwise:1 simonyan:1 tuytelaars:1 validates:1 differentiate:1 advantage:4 propose:2 reconstruction:2 product:1 combining:2 achieve:5 deformable:2 pronounced:1 normalize:1 exploiting:1 darrell:2 categorization:2 object:9 derive:2 depending:1 gong:2 measured:1 nearest:1 school:2 eq:2 involves:2 indicate:1 australian:1 attribute:1 subsequently:1 australia:2 alleviate:1 extension:1 exp:1 mop:1 efros:1 achieves:6 adopt:4 vary:1 omitted:1 proc:20 sofa:1 bag:2 council:1 hoffman:1 mit:7 clearly:1 gaussian:24 rather:1 shelf:3 resized:1 boosted:1 jaakkola:1 azizpour:1 encode:1 derived:1 focus:2 improvement:3 likelihood:3 indicates:1 greatly:1 contrast:2 baseline:2 sense:1 helpful:1 inference:4 perronnin:2 membership:1 typically:1 pixel:3 issue:3 classification:30 pascal:5 denoted:5 art:6 spatial:4 tzeng:1 constrained:1 marginal:1 once:2 aware:1 extraction:1 ng:1 sampling:1 identical:1 yu:1 report:1 simplify:1 employ:3 few:3 randomly:3 resulted:2 argmax:2 astounding:1 ab:1 detection:1 investigate:2 intra:2 evaluation:5 certainly:1 sheep:1 mixture:20 contextualizing:1 integral:2 explosion:1 orthogonal:1 unless:1 shen1:1 conduct:2 indexed:1 logarithm:1 fox:1 re:1 mbike:1 modeling:5 assignment:1 introducing:3 hundred:3 too:1 characterize:1 reported:3 arandjelovi:1 answer:1 eur:3 person:1 bu:7 lee:1 off:3 dong:1 quickly:2 again:1 choose:1 slowly:2 huang:4 overemphasize:1 conf:15 inefficient:1 derivative:1 li:1 syst:5 converted:1 stride:1 pooled:2 coding:62 int:6 farrell:1 vi:13 try:1 razavian:1 doing:1 liu1:1 level2:1 jia:2 accuracy:4 convolutional:4 descriptor:6 ofthe:1 anton:1 accurately:1 ren:1 comp:13 inform:1 email:1 sixth:2 against:1 pp:11 sampled:1 gain:2 proved:1 dataset:7 popular:1 vlad:2 knowledge:1 car:1 dimensionality:17 subtle:1 feed:1 higher:8 originally:1 supervised:2 specify:1 improved:1 zisserman:2 formulation:2 evaluated:1 mensink:1 mode:1 lei:1 k22:3 please:1 inferior:1 generalized:1 complete:2 demonstrate:4 image:44 wise:1 lazebnik:2 recently:6 common:1 superior:2 volume:4 significant:1 refer:1 doersch:1 centre:2 chapelle:1 similarity:1 base:6 closest:2 perspective:1 inf:5 chunhua:1 termed:1 certain:2 binary:1 caltech:1 captured:2 seen:8 additional:1 employed:2 u0:6 multiple:3 infer:1 faster:1 calculation:2 cross:1 long:1 lin:1 laplacian:1 verbeek:1 variant:1 prediction:1 vision:3 essentially:3 dpd:3 arxiv:1 kernel:10 normalization:8 pyramid:1 achieved:5 background:1 cropped:1 fine:6 addressed:1 regional:8 comment:3 induced:1 pooling:7 dpm:1 december:1 effectiveness:1 call:1 leverage:1 yang:1 variety:1 architecture:2 identified:2 cow:1 reduce:2 whether:2 lingqiao:1 handled:1 pca:3 song:2 reformulated:1 speaking:2 deep:24 generally:2 detailed:1 listed:1 clear:1 mid:1 bvlc:1 category:4 reduced:1 generate:3 http:2 notice:3 patt:6 write:1 vol:2 commented:1 drawn:4 gmm:19 utilize:1 merely:1 convert:1 sum:5 package:1 powerful:1 almost:2 throughout:1 reader:1 patch:1 draw:4 decision:1 comparable:5 layer:6 correspondence:1 quadratic:1 scene:6 software:1 x2:1 bousquet:1 argument:1 min:2 chair:1 performing:3 tv:1 according:1 battle:1 smaller:1 den:1 referencing:1 computationally:2 bus:1 turn:2 needed:5 end:3 pursuit:2 gaussians:8 rewritten:1 multiplied:1 apply:4 observe:1 hierarchical:2 generic:4 appropriate:1 alternative:2 slower:1 calculating:2 k1:2 build:2 especially:1 gregor:1 feng:1 seeking:1 objective:1 question:2 already:1 strategy:2 traditional:18 interacts:1 gradient:5 kth:3 subspace:4 distance:2 capacity:1 outer:1 reason:1 assuming:1 besides:2 modeled:2 relationship:4 insufficient:2 equivalently:1 difficult:1 setup:1 negative:1 implementation:2 perform:1 subcategory:1 observation:3 anchez:1 datasets:2 arc:1 benchmark:2 defining:2 incorporated:1 ucsd:1 introduced:2 bottle:1 dog:1 namely:2 fv:1 learned:2 established:2 boost:2 nu:1 beyond:3 suggested:1 able:1 usually:3 indoor:4 sparsity:1 including:1 max:1 endto:1 power:2 gool:1 difficulty:1 boat:1 residual:2 raina:1 scheme:1 improve:1 github:1 chao:1 prior:2 literature:4 l2:1 discovery:2 acknowledgement:1 understanding:1 carlsson:1 plant:1 generation:7 limitation:1 interesting:2 lv:1 sufficient:3 vectorized:1 share:1 supported:1 side:1 neighbor:1 differentiating:1 sparse:25 orderless:1 van:1 benefit:1 xia:1 dimension:11 vocabulary:1 rich:1 commonly:3 far:1 approximate:1 obtains:1 implicitly:1 global:6 robotic:1 xi:7 discriminative:4 pandey:1 latent:2 bay:1 table:21 promising:1 learn:4 ku:2 robust:1 improving:1 du:2 diag:1 surf:2 did:1 main:2 dense:1 whole:8 lampert:1 verifies:1 x1:1 augmented:8 xu:2 precision:3 concatenating:1 grained:6 donahue:1 xt:2 sift:4 list:3 svm:10 gupta:1 essential:1 vapnik:1 decaf:1 illustrates:1 justifies:1 kx:4 chen:3 suited:2 locality:1 simply:5 appearance:1 visual:5 vinyals:1 expressed:2 iandola:1 bo:1 hua:2 extracted:5 consequently:3 fisher:48 content:1 experimentally:1 fista:1 determined:2 infinite:2 specifically:3 reducing:1 lemma:1 called:2 specie:1 total:2 experimental:4 support:1 guo:1 latter:1 adelaide:1 evaluate:1 dance:1 |
4,733 | 5,286 | Self-Adaptable Templates for Feature Coding
Xavier Boix1,2? Gemma Roig1,2? Salomon Diether1 Luc Van Gool1
1
Computer Vision Laboratory, ETH Zurich, Switzerland
2
LCSL, Massachusetts Institute of Technology & Istituto Italiano di Tecnologia, Cambridge, MA
{xboix,gemmar}@mit.edu
{boxavier,gemmar,sdiether,vangool}@vision.ee.ethz.ch
Abstract
Hierarchical feed-forward networks have been successfully applied in object
recognition. At each level of the hierarchy, features are extracted and encoded,
followed by a pooling step. Within this processing pipeline, the common trend is
to learn the feature coding templates, often referred as codebook entries, filters, or
over-complete basis. Recently, an approach that apparently does not use templates
has been shown to obtain very promising results. This is the second-order pooling
(O2P) [1, 2, 3, 4, 5]. In this paper, we analyze O2P as a coding-pooling scheme.
We find that at testing phase, O2P automatically adapts the feature coding templates to the input features, rather than using templates learned during the training
phase. From this finding, we are able to bring common concepts of coding-pooling
schemes to O2P, such as feature quantization. This allows for significant accuracy
improvements of O2P in standard benchmarks of image classification, namely
Caltech101 and VOC07.
1
Introduction
Many object recognition schemes, inspired from biological vision, are based on feed-forward hierarchical architectures, e.g. [6, 7, 8]. In each level in the hierarchy, the algorithms can be usually
divided into the steps of feature coding and spatial pooling. The feature coding extracts similarities
between the set of input features and a set of templates (the so called filters, over-complete basis or
codebook), and then, the similarity responses are transformed using some non-linearities. Finally,
the spatial pooling extracts one single vector from the set of transformed responses. The specific architecture of the network (e.g. how many layers), and the specific algorithms for the coding-pooling
at each layer are usually set for a recognition task and dataset, cf. [9].
Second-order Pooling (O2P) is an alternative algorithm to the aforementioned coding-pooling
scheme. O2P has been introduced in medical imaging to analyze magnetic resonance images [1, 2],
and lately, O2P achieved state-of-the-art in some of the traditional computer vision tasks [3, 4, 5, 10].
A surprising fact of O2P is that it is formulated without feature coding templates [5]. This is in contrast to the common coding-pooling schemes, in which the templates are learned during a training
phase, and at testing phase, the templates remain fixed to the learned values.
Motivated by the intriguing properties of O2P, in this paper we try to re-formulate O2P as a codingpooling scheme. In doing so, we find that O2P actually computes similarities to feature coding
templates as the rest of the coding-pooling schemes. Yet, what remains uncommon of O2P, is that
the templates are ?recomputed? for each specific input, rather than being fixed to learned values. In
O2P, the templates are self-adapted to the input, and hence, they do not require learning.
From our formulation, we are able to bring common concepts of coding-pooling schemes to O2P,
such as feature quantization. This allows us to achieve significant improvements of the accuracy
?
Both first authors contributed equally.
1
of O2P for image classification. We report experiments on two challenging benchmarks for image
classification, namely Caltech101 [11], and VOC07 [12].
2
Preliminaries
In this Section, we introduce O2P as well as several coding-pooling schemes, and identify some
common terminology in the literature. This will serve as a basis for the new formulation of O2P,
that we introduce in the following section.
The algorithms that we analyze in this section are usually part of a layer of a hierarchical network
for object recognition. The input to these algorithms is a set of feature vectors that come from the
output of the previous layer, or from the raw image. Let {xi }N be the set of input feature vectors
to the algorithm, which is the set of N feature vectors, xi ? RM , indexed by i ? {1, . . . , N }.
The output of the algorithm is a single vector, which we denote as y, and it may have a different
dimensionality than the input vectors.
In the following subsections, we present the algorithms and terminology of template-based methods,
and then, we introduce the formulation of O2P that appears in the literature that apparently does not
use templates.
2.1
Coding-Pooling based on Evaluating Similarities to Templates
Template-based methods are build upon similarities between the input vectors and a set of templates.
Depending on the terminology of each algorithm, the templates may be denoted as filters, codebook,
or over-complete basis. From now on, we will refer to all of them as templates. We denote the set
of templates as {bk ? RM }P . In this paper, bk and the input feature vectors xi have the same
dimensionality, M . The set of templates is fixed to learned values during the training phase. There
are many possible learning algorithms, but analyzing them is not necessary here.
The algorithms that are interesting for our purposes, start by computing a similarity measure between
the input feature vectors {xi }N and the templates {bk }P . Let ?(xi , bk ) be the similarity function,
which depends on each algorithm. We define ? i as the vector that contains the similarities of xi to
the set of templates {bk }, and ? ? RM ?P the matrix whose columns are the vectors ? i , i.e.
?ki = ?(xi , bk ).
(1)
Once ? is computed, the algorithms that we analyze apply some non-linear transformation to ?, and
then, the resulting responses are merged together, with the so called pooling operation. The pooling
consists on generating one single response value for each template. We denote as gk (?) the function
that includes both the non-linear transformation and the pooling operation, where gk : RM ?P ? R.
We include both operations in the same function, but in the literature it is usually presented as two
separate steps. Finally, the output vector y is built using {gk (?)}P , {bk }P and {xi }N , depending
on the algorithm. It is also quite common to concatenate the outputs of neighboring regions to
generate the final output of the layer.
We now show how the presented terminology is applied to some methods based on evaluating similarities to templates, namely assignment-based methods and Fisher Vector. In the sequel, these
algorithms will be a basis to reformulate O2P.
Assignment-based Methods The popular Bag-of-Words and some of its variants fall into this
category, e.g. [13, 14, 15]. These methods consist on assigning each input vector xi to a set of
templates (the so called vector quantization), and then, building a histogram of the assignments,
which corresponds to the average pooling operation.
We now present them using our terminology. After computing the similarities to the templates, ?
(usually based on `2 distance), gk (?) computes both the vector quantization and the pooling. Let
s be the number of templates to which each input vector is assigned, and let ? 0i be the resulting
assignment vector of xi (i.e. ? 0i is the result of applying vector quantisation on xi ). ? 0i has s entries
set to 1 and the rest to 0, that indicate the assignment. Finally, gkP
(?) also computes the pooling for
the assignments corresponding to the template k, i.e. gk (?) = N1 i<N ? 0ki . The final output vector
is the concatenation of the resulting pooling of the different templates, y = (g1 (?), . . . , gP (?)).
2
Fisher Vectors It uses the first and second order statistics of the similarities between the features
and the templates [16]. Fisher Vector builds two vectors for each template bk , which are
1 X
1 X
(1)
(2)
?k =
?ki (bk ? xi ) ?k =
?ki (bk ? xi )2 ? Ck ,
(2)
Ak
Bk
i<N
i<N
1
1
exp ? (xi ? bk )t Dk (xi ? bk ) .
(3)
where ?ki =
Zk
2
Ak , Bk , Ck are learned constants, Zk a normalization factor and Dk is a learned constant matrix of
the model. Note that in Eq. (3), ?ki is a similarity between the feature vector xi and the template bk .
(1)
(2)
(1)
(2)
The final output vector is y = (?1 , ?1 . . . , ?P , ?P ). For further details we refer the reader
to [16].
We use our terminology to do a very simple re-write of the terms. We define gk (?) and bF
k (we use
the super-index F to indicate that are from Fisher vectors, and different from bk ) as
1
(1)
(2)
(1)
(2)
(? , ?k ).
(4)
gk (?) = k(?k , ?k )k2 , bF
k =
gk (?) k
We can see the templates of Fisher vectors, bF
k , are obtained from computing some transformations
to the original learned template bk , which involve the input set of features {xi }. gk (?) is the norm
(1)
(2)
of (?k , ?k ), which gives an idea of the importance of each template in {xi }, similarly to gk (?)
in assignment-based methods. Note that bF
k and gk (?) are related to only one fixed template, bk .
F
The final output vector becomes y = (g1 (?)bF
1 , . . . , gP (?)bP ).
2.2
Second-Order Pooling
Second-order Pooling (O2P) was introduced in medical imaging to describe the voxels produced in
diffusion tensor imaging [1], and to process tensor fields [2, 17]. O2P starts by building a correlation
matrix from the set of feature (column) vectors {xi ? RM }N , i.e.
1 X
xi xti ,
(5)
K=
N
i<N
where xti
M ?M
is the transpose vector of xi , and K ? R
is a square matrix. K is a symmetric positive
definite (SPD) matrix, and contains second-order statistics of {xi }. The set of SPD matrices form
a Riemannian manifold, and hence, the conventional operations in the Euclidean space can not be
used. Several metrics have been proposed for SPD matrices, and the most celebrated is the LogEuclidean metric [17]. Such metric consists of mapping the SPD matrices to the tangent space by
using the logarithm of the matrix, log(K). In the tangent space, the standard Euclidean metrics can
be used.
The logarithm of an SPD matrix can be computed in practice by applying the logarithm individually
to each of the eigenvalues of K [18]. Thus, the final output vector for O2P can be written as
!
X
t
y = vec (log(K)) = vec
log(?k )ek ek ,
(6)
k<M
where ek are the eigenvectors of K, and ?k the corresponding eigenvalues. The vec(?) operator
vectorizes log(K).
In Eq. (6), apparently, there are no similarities to a set of templates. The absence of templates makes
O2P look quite different from template-based methods. Recently, O2P achieved state-of-the-art
results in some computer vision tasks, e.g. in object detection [3], semantic segmentation [5, 10],
and for patch description [4]. Both reasons, motivates us to further analyze O2P in relation to
template-based methods.
3
Self-Adaptability of the Templates
In this section, we introduce a formulation that relates O2P and template-based methods. The new
formulation is based on comparing two final representation vectors, rather than defining how the
3
final vector y is built. We denote hyr , ys i as the inner product between yr and ys , which are the
final representation vectors from two sets of input feature vectors, {xri }N and {xsi }N , respectively,
where we use the superscripts r and s to indicate the respective representation for each set. It will
become clear during this section why we analyze hyr , ys i instead of y.
We divide the analysis in three subsections. In subsection 3.1, we re-write the formulation of the
template-based methods of Section 2 with the inner product hyr , ys i. In subsection 3.2, we do the
same for O2P, and this unveils that O2P is also based on evaluating similarities to templates. In
subsection 3.3, we analyze the characteristics of the templates in O2P, which have the particularity
that are self-adapted to the input.
3.1
Re-Formulation of Template-Based Methods
We re-write a generic formulation for the template-based methods described in Section 2 with the
inner product between two final output vectors. The algorithms of Section 2 can be expressed as
XX
hyr , ys i =
gk (? r )gq (? s )S(brk , bsq ),
(7)
k<P q<P
where ?ki = ?(xi , bk ),
and S(u, v) is a similarity function between the templates that depends on each algorithm. Recall
that gk (?) is a function that includes the non-linearities and the pooling of the similarities between
the input feature vectors and the the templates. To see how Eq. (7) arises naturally from the algorithms of Section 2, we now analyze them in terms of this formulation.
Assignment-Based Methods
as
The inner product between two final output vectors can be written
hyr , ys i =(g1 (? r ), . . . , gP (? r ))t (g1s (? s ), . . . , gPs (? s )) =
X
XX
=
gk (? r )gk (? s ) =
gk (? r )gq (? s )I(brk = bsq ),
(8)
k<P q<P
k<P
where the last step introduces an outer summation, and the indicator function I(?) eliminates the
unnecessary cross terms. Comparing this last equation to Eq. (7), we can identify that S(brk , bsq ) is
the indicator function (returns 1 when brk = bsq , and 0 otherwise).
Fisher Vectors The inner product between two final Fisher Vectors is
r
rF t
s
sF
s
sF
hyr , ys i =(g1 (? r )brF
1 , . . . , gP (? )bP ) (g1 (? )b1 , . . . , gP (? )bP )
XX
sF
=
gk (? r )gq (? s )I(brk = bsq )hbrF
k , bq i.
(9)
k<P q<P
The indicator function appears for the same reason as in Assignment-Based Methods. The final
sF
templates for each set of input vectors, brF
k , bk , respectively, are compared with each other with
rF t sF
rF
sF
t sF
the similarity (bk ) bq . Thus, S(bk , bq ) in Eq. (7) is equal to I(brk = bsq )(brF
k ) bq .
3.2
O2P as Coding-Pooling based on Template Similarities
We now re-formulate O2P, in the same way as we did for template-based methods in the previous
subsection. This will allow relating O2P to template-based methods, and show that O2P also uses
similarities to templates.
We re-write the definition of O2P in Eq. (6) with hyr , ys i. Using the property vec(A)t vec(B) =
tr(At B), where tr(?) is the trace function of a matrix, hyr , ys i becomes (in the supplementary
material we do the full derivation)
hyr , ys i = hvec (log(Kr )) , vec (log(Ks ))i =
X X
=
log(?rk ) log(?sq )herk , esq i2 ,
(10)
k<M q<M
where ek etk is a square matrix, and the eigenvectors, {erk }M and {esk }M , are compared all against
each other with herk , esq i2 . Going back to the generic formulation of template-based methods in
4
Method
Assignment-based
Fisher Vectors
O2P
S(brk , bsq )
?ki = ?(xi , bk ) templates
I(brk = bsq )
hxi , bk i
fixed
r
s
sF
sF
I(bk = bq )hbk , bP i
Eq. (3)
fixed/adapted
hbrk , bsq i2
hxi , bk i2
self-adapted
gk (?)
P 0
1
i ?ki
N
(1)
(2)
k(?k , P
?k )k2
log N1 i ?ki
Table 1: Summary Table of the elements of our formulation for Assignment-based methods, Fisher
Vectors and O2P.
Eq. (7), we can see that the similarity function between the templates, S(erk , esq ), can be identified in
O2P as herk , esq i2 . Also, note that in O2P the sums go over M , which is the number of eigenvectors,
and in Eq. (7), go over P , which is the number of templates. Finally, gk (?) in Eq. (7) corresponds
to log(?k ) in O2P.
At this point, we have expressed O2P in a similar way as template-based methods. Yet, we still have
to find the similarity between the input feature vectors and the templates. For that purpose, we use
the definition of eigenvalues and eigenvectors, i.e. ?k ek = Kek , and also that tr(ek etk ) = 1 (the
t
eigenvectors are orthonormal). Then,
P wet can derive the following equivalence: ?k = ?k tr(ek ek ) =
1
t
tr(Kek ek ). Replacing K by N i xi xi , we find that the eigenvalues, ?k , can be written using the
similarity between the input vectors, xi , and the eigenvectors, ek :
1 X
1 X
tr((xi xti )(ek etk )) =
hxi , ek i2 .
(11)
?k =
N i
N i
Finally, we can integrate all the above derivations in Eq. (10), and we obtain that
X X
hyr , ys i =
gk (? r )gq (? s )herk , esq i2 ,
(12)
k<M q<M
where gk (?) = log(?k ) = log
1 X
?ki
N
i<N
2
and ?ki = ?(xi , ek ) = hxi , ek i .
!
,
(13)
(14)
We can see by analyzing Eq. (12) that this equation takes the same form as the general equation
of template-based methods in Eq. (7). Note that the eigenvectors take the same role as the set of
templates, i.e. bk = ek and P = M . Also, observe that S(brk , bsq ) is the square of the inner product
between eigenvectors, ?(xi , bk ) is the square of the inner product between the input vectors and the
eigenvectors, and the pooling operation is the logarithm of the average of the similarities. In Table 1
we summarize the corresponding elements of all the described methods.
3.3
Self-Adaptative Templates
We define self-adaptative templates as templates that only depend on the input set of feature vectors, and are not fixed to predefined values. This is the case in O2P, because the templates in O2P
correspond to the eigenvectors computed from the set of input feature vectors. The templates in
O2P are not fixed to values learned during the training phase. Interestingly, the final templates in
F
Fisher Vectors, bF
k , are also partially self-adapted to the input vectors. Note that bk are obtained by
modifying the fixed learned templates, bk , with the input feature vectors.
Finally, note that in O2P the number of templates is equal to the dimensionality of the input feature
vectors. Thus, in O2P the number of templates can not be increased without changing the input
vectors? length, M . This begs the following question: do M templates allow for sufficient generalization for object recognition for any set of input vectors? We analyze this question in the next
section.
4
Application: Quantization for O2P
We observe in the experiments section that the performance of O2P degrades when the number of
vectors in the set of input features increases. It is reasonable that M templates are not sufficient
when the number of different vectors in {xi }N increases, specially when they are very different
5
Algorithm 1: Sparse Quantization in O2P
Input: {xi }N , k
Output: y
foreach i = {1, . . . , N } do
? i ? Set k highest values of xi to its vector entry: xi , and the rest to 0
x
end
P
?ix
? ti
K = N1 i x
y = vec(log(K))
from each other. We now introduce an algorithm to increase the robustness of O2P to the variability
of the input vectors.
We quantize the input feature vectors, {xi }, before computing O2P. Quantization may discard details, and hence, reduce the variability among vectors. In the experiments section it is reported
that this allows preventing the degradation of performance in object recognition, when the number
of input feature vectors increases. The quantization algorithm that we use is sparse quantization
(SQ) [15, 19], because SQ does not change the dimensionality of the feature vector. Also, SQ is fast
to compute, and does not increase the computational cost of O2P.
Sparse Quantization for O2P For the quantization of {xi } we use SQ, which is a quantization
to the set of k-sparse vectors. Let Rqk be the set of k-sparse vectors, i.e. {s ? Rq : ksk0 ? k}.
Also, we define Bqk = {0, 1}qk = {s ? {0, 1}q : ksk0 = k}, which is the set of binary vectors
with k elements set to one and (q ? k) set to zero. The cardinality of |Bqk | is equal to kq . The
quantization of a vector v ? Rq into a codebook {ci } is a mapping of v to the closest element in
? ? = arg minv? ?{ci } k?
? ? is the quantized vector v. In the case of SQ, the
{ci }, i.e. v
v ? vk2 , where v
codebook {ci } contains the set of k-sparse vectors. These may be any of the previously introduced
types: Rqk , Bqk . An important advantage of SQ over a general quantization is that it can be computed
much more efficiently. The naive way to compute a general quantization is to evaluate the nearest
neighbor of v in {ci }, which may be costly to compute for large codebooks and high-dimensional
v. In contrast, SQ can be computed by selecting the k higher values of the set {vi }, i.e. for SQ into
Rqk , v?i = vi if i is one of the k-highest entries of vector v, and 0 otherwise. For SQ into Bqk , the
dimension indexed by the k-highest are set to 1 instead of vi , and 0 otherwise. (We refer the reader
to [15, 19] for a more detailed explanation on SQ).
In Algorithm 1 we depict the implementation of SQ in O2P, which highlights its simplicity. The
computational cost of SQ is negligible compared to the cost of computing O2P. We use the set of
k-sparse vectors in RM
k for SQ, which worked best in practice, as shown in the following.
5
Experiments
In this section, we analyze O2P in image classification from dense sampled SIFT descriptors. This
setup is common in image classification, and it allows direct comparison to previous works on O2P.
We report results on the Caltech101 [11] and VOC07 [12] datasets, using the standard evaluation
benchmarks, which are the mean average precision accuracy across all classes.
5.1
Implementation Details
We use the standard pipeline for image classification. We never use flipped or blurred images to
extend the training set.
Pipeline. For Caltech101, the image is re-sized to take a maximum height and width of 300
pixels, which is the standard resizing protocol for this dataset. For VOC07 the size of the images
remains the same as the original. We extract SIFT [8] from patches on a regular grid, at different
scales. In Caltech 101, we extract them at every 8 pixels and at the scales of 16, 32 and 48 pixels
diameter. In VOC07, SIFT is sampled at each 4 pixels and at the scales of 12, 24 and 36 pixels
diameter. O2P is computed using the SIFT descriptors as input, and using spatial pyramids. In
6
Caltech101, we generate the pooling regions dividing the image in 4 ? 4, 2 ? 2 and 1 ? 1 regions,
and in VOC07 in 3 ? 1, 2 ? 2 and 1 ? 1 regions. To generate the final descriptor for the whole
image, we concatenate the descriptors for each pooled region. We apply the power normalization to
the final feature dimensions, sign(x)|x|3/4 , that was shown to work well in practice [5]. Finally, we
use a linear one-versus-rest SVM classifier for each class with the parameter C of the SVM set to
1000. We use the LIBLINEAR library for the SVM[20].
Other Feature Codings. As a sanity check of our results, we replace O2P with the Bag-ofWords [13] baseline, without changing any of the parameters. In Caltech101, we replace the average
pooling of Bag-of-Words by max-pooling (without normalization) as it performs better. The codebook is learned by randomly picking a set of patches as codebook entries, which was shown to work
well for the encodings we are evaluating [14]. We use a codebook of 8192 entries, since with more
entries the performance does not increase significantly, but the computational cost does.
5.2
Results on Caltech101
We use 3 random splits of 30 images per class for training and the rest for testing. In Fig. 1a, results
are shown for different spatial pyramid configurations, as well as different levels of quantization.
Note that SQ with k = 128 is not introducing any quantization, as SIFT features are 128 dimensional
vectors. Note that using SQ increases the performance more than 5% compared to when not using
SQ (k = 128), when using only the first level of the pyramid. For the other levels of the pyramid,
there is less improvement with SQ. This is in accordance with the observation that in smaller regions
there are less SIFT vectors, the variability is smaller, and the limited amount of templates is able to
better capture the meaningful information than in bigger regions. We can also see that for small k
of SQ, the performance degrades due to the introduction of too much quantization.
We also run experiments with Bag-of-Words with max-pooling (74.8%), and O2P without SQ
(76.52%), and both of them are surpassed by O2P with SQ (78.63%). In [5], O2P accuracy is
reported to be 79.2% with SIFT descriptor (we do not compare to their version of enriched SIFT,
since all our experiments are with normal SIFT). We inspected the code of [5], and we found that
the difference of accuracy mainly comes from using a more drastic resizing of the image, that takes
a maximum of 100 pixels of width and height (usually in the literature it is 300 pixels). Note that resizing is another way of discarding information, and hence, O2P may benefit from that. We confirm
this by resizing the image back to 300 pixels in [5]?s code, and the accuracy is 77.1%, similar to the
one that we report without SQ in our code. The accuracy is not exactly the same due to differences
in the SIFT parameters in [5]. Also, we tested SQ in [5]?s code with the resizing to a maximum of
100 pixels, and the accuracy increased to 79.45%, which is higher than reported in [5], and close to
state-of-the-art results using SIFT descriptors (80.3%) [21].
5.3
Results on VOC07
In Fig. 1b, we run the same experiment as in Caltech101. Note that the impact of SQ is even more
evident than in Caltech101. In Table 2 we report the per-class accuracy, in addition to the mean
average precision reported in Fig. 1b. We follow the evaluation procedure as described in [12].
With the full pyramid, when we use SQ the accuracy increases from 18.81% to 50.97%. In contrast to Caltech101, O2P with SQ performance is similar to our implementation of Bag-of-Words
(51.14%). Thus, under adverse conditions for O2P, i.e. images with high variability such as in
VOC07 and with a high number of input vectors, we can use SQ and obtain huge improvements of
the O2P?s accuracy. The best reported results [22] in VOC07 are around 10% better than O2P with
SQ, yet we obtain more than 30% improvement from the baseline.
6
Conclusions
We found that O2P can be posed as a coding-pooling scheme based on evaluating similarities to templates. The templates of O2P self-adapt to the input, while the rest of the analyzed methods do not.
In practice, our formulation was used to improve the performance of O2P in image classification.
We are currently analyzing self-adaptative templates in deep hierarchical networks.
7
1 pyr.
1+2 pyr.
1+2+3 pyr.
1+2+3 pyr. w/o SQ
Caltech 101
PASCAL VOC 2007
50.97%
78.63%
Mean average precision
Mean accuracy
0.8
76.52%
0.75
75.55%
0.7
0.65
65.14%
0.6
0.55
SQ selected in val. set
5
20
40
60
80
100
128
0.5
49.09%
0.4
41.20%
0.3
0.2
0.1
18.81%
5
20
40
60
80
100
Sparse Quantization
Sparse Quantization
(a)
(b)
128
35
14
33
27
4
74
28
69
61
12
56
17
56
43
8
19
7
18
20
6
44
9
41
28
7
41
21
37
30
12
78
55
74
66
49
35
7
36
33
5
Average
50
26
47
40
22
TV/Monitor
52
14
47
38
10
Sofa
Potted Plant
Sheep
69
40
68
58
29
Train
Motorbike
Person
51
19
50
37
8
Horse
23
6
20
15
6
63
18
62
58
9
Cow
Dinning Table
Dog
45
12
41
32
11
Chair
Bottle
Bus
53
9
50
41
7
Car
Bird
Boat
72
34
71
66
21
Cat
Bicycle
3 Pyr. O2P + SQ
3 Pyr. O2P w/o SQ
2 Pyr. O2P + SQ
1 Pyr. O2P + SQ
1 Pyr. O2P w/o SQ
Aeroplane
Figure 1: Results for different numbers of non-zero entries of SQ. Note that SQ at k = 128 is not
introducing any quantization, since SIFT features are 128 dimensional vectors. (a) Caltech 101
(using 30 images per class for training), (b) VOC07.
50
10
51
37
7
67
16
66
56
9
45
12
44
36
9
50.97
18.81
49.09
41.20
12.53
Table 2: PASCAL VOC 2007 classification results. The average score provides the per-class average. We report results for O2P, with and without SQ, with the first plus second plus third levels of
pyramids (3 Pyr.), O2P with SQ with the first plus second levels of pyramids (2 Pyr.), and O2P with
and without SQ only with the first level of pyramids (1 Pyr.).
Acknowledgments: We thank the ERC for support from AdG VarCity.
References
[1] D. Le Bihan, J.-F. Mangin, C. Poupon, C. A. Clark, S. Pappata, N. Molko, and H. Chabriat,
?Diffusion tensor imaging: concepts and applications,? Journal of magnetic resonance imaging, 2001.
[2] J. Weickert and H. Hagen, Visualization and Processing of Tensor Fields. Springer, 2006.
[3] O. Tuzel, F. Porikli, and P. Meer, ?Region covariance: A fast descriptor for detection and
classification,? in ECCV, 2006.
[4] P. Li and Q. Wang, ?Local log-euclidean covariance matrix (L2 ECM) for image representation
and its applications,? in ECCV, 2012.
[5] J. Carreira, R. Caseiro, J. Batista, and C. Sminchisescu, ?Semantic segmentation with secondorder pooling,? in ECCV, 2012.
[6] K. Fukushima, ?Neocognitron: A self-organizing neural network model for a mechanism of
pattern recognition unaffected by shift in position,? Biological cybernetics, 1980.
[7] M. Riesenhuber and T. Poggio, ?Hierarchical models of object recognition in cortex,? Nature
neuroscience, 1999.
[8] D. G. Lowe, ?Distinctive image features from scale-invariant keypoints,? IJCV, 2004.
[9] J. Bergstra, D. Yamins, and D. Cox, ?Making a science of model search: Hyperparameter
optimization in hundreds of dimensions for vision architectures,? in ICML, 2013.
8
[10] R. Girshick, J. Donahue, T. Darrell, and J. Malik, ?Rich feature hierarchies for accurate object
detection and semantic segmentation,? in CVPR, 2014.
[11] L. Fei-Fei, R. Fergus, and P. Perona, ?One-shot learning of object categories,? TPAMI, 2006.
[12] M. Everingham, L. Van Gool, C. Williams, J. Winn, and A. Zisserman, ?The PASCAL visual
object classes (VOC) challenge,? IJCV, 2010.
[13] G. Csurka, C. R. Dance, L. Fan, J. Willamowski, and C. Bray, ?Visual categorization with bags
of keypoints,? in Workshop on Statistical Learning in Computer Vision, ECCV, 2004.
[14] A. Coates and A. Ng, ?The importance of encoding versus training with sparse coding and
vector quantization,? in ICML, 2011.
[15] X. Boix, G. Roig, and L. Van Gool, ?Nested sparse quantization for efficient feature coding,?
in ECCV, 2012.
[16] J. Sanchez, F. Perronnin, T. Mensink, and J. Verbeek, ?Image classification with the fisher
vector: Theory and practice,? IJCV, 2013.
[17] V. Arsigny, P. Fillard, X. Pennec, and N. Ayache, ?Geometric means in a novel vector space
structure on symmetric positive-definite matrices,? Journal on matrix analysis and applications, 2007.
[18] R. Bhatia, Positive definite matrices. Princeton University Press, 2009.
[19] X. Boix, M. Gygli, G. Roig, and L. Van Gool, ?Sparse quantization for patch description,? in
CVPR, 2013.
[20] R. E. Fan, K. W. Chang, C. J. Hsieh, X. R. Wang, and C. J. Lin, ?LIBLINEAR: A library for
large linear classification,? JMLR, 2008.
[21] O. Duchenne, A. Joulin, and J. Ponce, ?A graph-matching kernel for object categorization,? in
ICCV, 2011.
[22] X. Zhou, K. Yu, T. Zhang, and T. S. Huang, ?Image classification using super-vector coding of
local image descriptors,? in ECCV, 2010.
9
| 5286 |@word cox:1 version:1 norm:1 everingham:1 bf:6 covariance:2 hsieh:1 tr:6 shot:1 liblinear:2 celebrated:1 contains:3 configuration:1 selecting:1 score:1 batista:1 interestingly:1 comparing:2 surprising:1 yet:3 intriguing:1 assigning:1 written:3 concatenate:2 depict:1 selected:1 yr:1 potted:1 quantized:1 codebook:8 provides:1 zhang:1 height:2 direct:1 become:1 consists:2 ijcv:3 introduce:5 inspired:1 voc:3 automatically:1 xti:3 cardinality:1 becomes:2 xx:3 linearity:2 what:1 erk:2 finding:1 transformation:3 porikli:1 every:1 ti:1 exactly:1 rm:6 k2:1 classifier:1 mangin:1 medical:2 positive:3 before:1 negligible:1 accordance:1 local:2 brf:3 encoding:2 ak:2 analyzing:3 plus:3 bird:1 k:1 equivalence:1 salomon:1 challenging:1 limited:1 acknowledgment:1 testing:3 practice:5 minv:1 definite:3 sq:40 procedure:1 tuzel:1 eth:1 significantly:1 matching:1 word:4 regular:1 close:1 operator:1 applying:2 conventional:1 gool1:1 go:2 williams:1 formulate:2 simplicity:1 orthonormal:1 meer:1 hvec:1 roig:2 gygli:1 hierarchy:3 inspected:1 gps:1 us:2 secondorder:1 trend:1 element:4 recognition:8 hagen:1 role:1 wang:2 capture:1 region:8 brk:9 highest:3 rq:2 unveils:1 depend:1 esk:1 serve:1 upon:1 distinctive:1 basis:5 cat:1 derivation:2 train:1 fast:2 describe:1 horse:1 bhatia:1 sanity:1 whose:1 encoded:1 quite:2 supplementary:1 posed:1 cvpr:2 particularity:1 otherwise:3 tested:1 resizing:5 statistic:2 g1:5 gp:5 final:15 superscript:1 advantage:1 eigenvalue:4 tpami:1 product:7 gq:4 neighboring:1 organizing:1 achieve:1 adapts:1 description:2 gkp:1 g1s:1 gemma:1 darrell:1 generating:1 categorization:2 object:11 depending:2 derive:1 nearest:1 eq:13 dividing:1 come:2 indicate:3 switzerland:1 merged:1 filter:3 modifying:1 material:1 require:1 generalization:1 preliminary:1 biological:2 summation:1 around:1 normal:1 exp:1 mapping:2 bicycle:1 purpose:2 sofa:1 bag:6 wet:1 currently:1 individually:1 successfully:1 bqk:4 mit:1 super:2 rather:3 ck:2 zhou:1 ecm:1 vectorizes:1 ponce:1 improvement:5 check:1 mainly:1 contrast:3 baseline:2 vk2:1 perronnin:1 voc07:10 perona:1 relation:1 transformed:2 going:1 pixel:9 arg:1 classification:12 aforementioned:1 among:1 denoted:1 pascal:3 resonance:2 spatial:4 art:3 field:2 once:1 equal:3 never:1 ng:1 flipped:1 look:1 icml:2 yu:1 report:5 randomly:1 phase:6 n1:3 fukushima:1 detection:3 huge:1 evaluation:2 sheep:1 uncommon:1 introduces:1 analyzed:1 predefined:1 accurate:1 necessary:1 istituto:1 poggio:1 respective:1 bq:5 indexed:2 euclidean:3 logarithm:4 divide:1 re:8 girshick:1 increased:2 column:2 assignment:11 cost:4 introducing:2 entry:8 kq:1 hundred:1 o2p:78 too:1 reported:5 person:1 caseiro:1 sequel:1 picking:1 together:1 huang:1 ek:15 return:1 li:1 bergstra:1 coding:22 pooled:1 includes:2 blurred:1 depends:2 vi:3 csurka:1 try:1 lowe:1 apparently:3 analyze:10 doing:1 start:2 square:4 accuracy:12 descriptor:8 kek:2 characteristic:1 qk:1 correspond:1 identify:2 efficiently:1 raw:1 produced:1 cybernetics:1 unaffected:1 definition:2 against:1 lcsl:1 naturally:1 pyr:12 di:1 riemannian:1 sampled:2 dataset:2 popular:1 massachusetts:1 adaptative:3 recall:1 subsection:6 car:1 dimensionality:4 segmentation:3 adaptability:1 actually:1 back:2 adaptable:1 appears:2 feed:2 higher:2 follow:1 response:4 zisserman:1 formulation:12 mensink:1 gemmar:2 correlation:1 bihan:1 replacing:1 building:2 concept:3 xavier:1 hence:4 assigned:1 symmetric:2 laboratory:1 semantic:3 i2:7 during:5 self:11 width:2 neocognitron:1 evident:1 complete:3 performs:1 bring:2 image:24 novel:1 recently:2 common:7 foreach:1 extend:1 relating:1 fillard:1 significant:2 refer:3 cambridge:1 vec:7 grid:1 similarly:1 erc:1 hxi:4 similarity:24 cortex:1 quantisation:1 closest:1 discard:1 pennec:1 binary:1 caltech:3 relates:1 full:2 keypoints:2 adapt:1 cross:1 lin:1 divided:1 equally:1 y:11 bigger:1 impact:1 verbeek:1 variant:1 xsi:1 vision:7 metric:4 surpassed:1 histogram:1 normalization:3 kernel:1 pyramid:8 achieved:2 addition:1 winn:1 rest:6 eliminates:1 specially:1 pooling:32 sanchez:1 ee:1 split:1 spd:5 architecture:3 identified:1 cow:1 inner:7 idea:1 reduce:1 codebooks:1 shift:1 motivated:1 aeroplane:1 etk:3 deep:1 clear:1 involve:1 eigenvectors:10 detailed:1 amount:1 category:2 diameter:2 generate:3 coates:1 sign:1 neuroscience:1 per:4 write:4 hyperparameter:1 recomputed:1 terminology:6 monitor:1 changing:2 vangool:1 diffusion:2 imaging:5 graph:1 sum:1 run:2 reader:2 reasonable:1 patch:4 layer:5 ki:12 followed:1 weickert:1 fan:2 adapted:5 bray:1 worked:1 fei:2 bp:4 chair:1 tv:1 remain:1 across:1 smaller:2 making:1 invariant:1 iccv:1 pipeline:3 equation:3 zurich:1 remains:2 previously:1 bus:1 visualization:1 mechanism:1 yamins:1 drastic:1 italiano:1 end:1 operation:6 apply:2 observe:2 hierarchical:5 generic:2 magnetic:2 alternative:1 robustness:1 motorbike:1 original:2 hbk:1 cf:1 include:1 build:2 tensor:4 malik:1 question:2 degrades:2 costly:1 ofwords:1 traditional:1 distance:1 separate:1 thank:1 concatenation:1 outer:1 manifold:1 reason:2 length:1 code:4 index:1 reformulate:1 setup:1 gk:21 xri:1 trace:1 implementation:3 motivates:1 contributed:1 observation:1 datasets:1 benchmark:3 riesenhuber:1 defining:1 variability:4 introduced:3 bk:30 namely:3 dog:1 bottle:1 learned:11 able:3 usually:6 pattern:1 summarize:1 challenge:1 built:2 rf:3 max:2 explanation:1 gool:3 power:1 indicator:3 boat:1 ksk0:2 scheme:10 improve:1 technology:1 library:2 lately:1 extract:4 naive:1 literature:4 voxels:1 tangent:2 val:1 l2:1 geometric:1 plant:1 highlight:1 interesting:1 versus:2 clark:1 integrate:1 sufficient:2 begs:1 eccv:6 caltech101:10 summary:1 last:2 transpose:1 allow:2 institute:1 fall:1 template:78 neighbor:1 sparse:12 van:4 benefit:1 dimension:3 evaluating:5 rich:1 computes:3 preventing:1 forward:2 author:1 confirm:1 b1:1 unnecessary:1 willamowski:1 xi:36 fergus:1 search:1 ayache:1 why:1 table:6 rqk:3 promising:1 learn:1 zk:2 nature:1 sminchisescu:1 quantize:1 protocol:1 did:1 joulin:1 dense:1 whole:1 enriched:1 fig:3 referred:1 boix:2 adg:1 precision:3 position:1 sf:9 jmlr:1 third:1 ix:1 donahue:1 rk:1 specific:3 discarding:1 sift:12 dk:2 svm:3 consist:1 workshop:1 quantization:24 importance:2 kr:1 ci:5 visual:2 duchenne:1 expressed:2 partially:1 chang:1 springer:1 ch:1 corresponds:2 nested:1 extracted:1 ma:1 sized:1 formulated:1 luc:1 fisher:11 absence:1 change:1 replace:2 adverse:1 tecnologia:1 carreira:1 degradation:1 called:3 bsq:10 meaningful:1 support:1 arises:1 ethz:1 evaluate:1 princeton:1 dance:1 |
4,734 | 5,287 | Learning From Weakly Supervised Data by The
Expectation Loss SVM (e-SVM) algorithm
Junhua Mao
Department of Statistics
University of California, Los Angeles
[email protected]
Jun Zhu
Department of Statistics
University of California, Los Angeles
[email protected]
Alan Yuille
Department of Statistics
University of California, Los Angeles
[email protected]
Abstract
In many situations we have some measurement of confidence on ?positiveness?
for a binary label. The ?positiveness? is a continuous value whose range is a
bounded interval. It quantifies the affiliation of each training data to the positive
class. We propose a novel learning algorithm called expectation loss SVM (eSVM) that is devoted to the problems where only the ?positiveness? instead of a
binary label of each training sample is available. Our e-SVM algorithm can also
be readily extended to learn segment classifiers under weak supervision where the
exact positiveness value of each training example is unobserved. In experiments,
we show that the e-SVM algorithm can effectively address the segment proposal
classification task under both strong supervision (e.g. the pixel-level annotations
are available) and the weak supervision (e.g. only bounding-box annotations are
available), and outperforms the alternative approaches. Besides, we further validate this method on two major tasks of computer vision: semantic segmentation
and object detection. Our method achieves the state-of-the-art object detection
performance on PASCAL VOC 2007 dataset.
1
Introduction
Recent work in computer vision relies heavily on manually labeled datasets to achieve satisfactory
performance. However, the detailed hand-labelling of datasets is expensive and impractical for large
datasets such as ImageNet [6]. It is better to have learning algorithms that can work with data that
has only been weakly labelled, for example by putting a bounding box around an object instead of
segmenting it or parsing it into parts.
In this paper we present a learning algorithm called expectation loss SVM (e-SVM). It requires
a method that can generate a set of proposals for the true label (e.g., the exact silhouette of the
object). But this set of proposals may be very large, each proposal may be only partially correct
(the correctness can be quantified by a continues value between 0 and 1 called ?positiveness?), and
several proposals may be required to obtain the correct label. In the training stage, our algorithm
can deal with the strong supervised case where the positiveness of the proposals are observed, and
can easily extend to the weakly supervised case by treating the positiveness as latent variables. In
the testing stage, it will predict the label for each proposal and provide a confidence score.
There are some alternative approaches for this problem, such as Support Vector Classification (SVC)
and Support Vector Regression (SVR). For the SVC algorithm, because this is not a standard binary
1
Annotations
Segment Proposals
IoU Ratios
Test images
0.79
0.02
...
0
Train
e-SVM
...
latent
latent
...
latent
Classifiers
Confidence of
class ?dog?: 3.49
...
0.25
-2.76
Figure 1: The illustration of our algorithm. In the training process, the e-SVM model can handle
two types of annotations: pixel level (strong supervision) and bounding box (weak supervision)
annotations. For pixel level annotations, we set the positiveness of the proposal as IoU overlap
ratios with the groundtruth and train classifiers using basic e-SVM. For bounding box annotations,
we treat the positiveness as latent variables and use latent e-SVM to train classifiers. In the testing
process, the e-SVM will provide each segment proposal a class label and a confidence score. (Best
viewed in color)
classification problem, one might need to binarize the positiveness using ad-hoc heuristics to determine a threshold, which degrades its performance [18]. To address this problem, previous works
usually used SVR [4, 18] to train the class confidence prediction models in segmentic segmentation. However, it is also not a standard regression problem since the value of positiveness belongs
to a bounded interval [0, 1]. We compare our e-SVM to these two related methods in the segment proposal confidence prediction problem. The positiveness of each segment proposal is set as
the Intersection over Union (IoU) overlap ratio between the proposal and the pixel level instance
groundtruth. We test our algorithm under two types of scenarios with different annotations: the
pixel level annotations (positiveness is observed) and the bounding box annotations (positiveness
is unobserved). Experiments show that our model outperforms SVC and SVR in both scenarios.
Figure 1 illustrates the framework of our algorithm.
We further validate our approach on two fundamental computer vision tasks: (i) semantic segmentation, and (ii) object detection. Firstly, we consider semantic segmentation. There has recently been
impressive progress at this task using rich appearance cues. Segments are extracted from images
[1, 3, 4, 12], appearance cues are computed for each segment [5, 21, 25], and classifiers are trained
using groundtruth pixel labeling [18]. Methods of this type are almost always among the winners
of the PASCAL VOC segmentation challenge [5]. But all these methods rely on datasets which
have been hand-labelled at the pixel level. For this application we generate the segment proposals
using CPMC segments [4]. The positiveness of each proposal is set as the Intersection over Union
(IoU) overlap ratio. We show that appearance cues learnt by e-SVM, using either the bounding
box annotations or pixel level annotations, are more effective than those learnt with SVC and SVR
on PASCAL VOC 2011 [9] segmentation dataset. Our algorithm is also flexible enough to utilize
additional bounding box annotations to further improve the results.
Secondly, we address object detection by exploiting the effectiveness of segmentation cues and coupling them to existing object detection methods. For this application, the data is only weakly labeled
because the groundtruth for object detection is typically specified by bounding boxes (e.g. PASCAL
VOC [8, 9] and Imagenet [6]), which means that pixel level groundtruth is not available. We use
either CPMC or super-pixels as methods for producing segment proposals. IoU is again used to represent the positiveness of the proposals. We test our approach on the PASCAL dataset using, as our
base detector, the Regions with CNN features (RCNN) [14] (currently state of the art on PASCAL
and outperforms previous works by a large margin). This method first used selective search method
[24] to extract candidate bounding boxes. For each candidate bounding box, it extracts features by
deep networks [16] learned on Imagenet dataset and fine-tuned on PASCAL. We couple our appearance cues to this system by simple concatenating our spatial confidence map features based on the
trained e-SVM classifiers and the deep learning features, and then train a linear SVM. We show that
this simple approach yields an average improvement of 1.5 percent on per-class average precision
(AP).
We note that our approach is general. It can use any segment proposal detectors, any image features,
and any classifiers. When applied to object detection it could use any base detector, and we could
couple the appearance cues with the base detector in many different ways (we choose the simplest).
2
In addition, it can handle other classification problems where only the ?positiveness? of the samples
instead of binary labels are available.
2
Related work on weakly supervised learning and weighted SVMs
We have introduced some of the most relevant works published recently for semantic segmentation
or object detection. In this section, we will briefly review related work of weakly supervised learning methods for segment classification, and discuss the connection to instance weighted SVM in
literature.
The problem settings for most previous works generally assumed that they only get a set of accompanying words of an image or a set of image level labeling, which is different from the problem
settings in this paper. Multiple Instance Learning (MIL) [7, 2] was adopted to solve these problems
[20, 22]. MIL handles cases where at least one positive instance is present in a positive bag and
only the labels of a set of bags are available. Vezhnevets et.al. [26] proposed a Multi-Image Model
(MIM) to solve this problem and showed that MIL in [22] is a special case of MIM. Later, [26]
developed MIM to a generalized MIM and used it as their segmentation model. Recently, Liu et.al.
[19] presented a weakly-supervised dual clustering approach to handle this task.
Our weakly supervised problem setting is in the middle between these settings and the strong supervision case (i.e. the full pixel level annotations are available). It is also very important and useful
because bounding box annotations of large-scale image dataset are already available (e.g. Imagenet
[6]) while the pixel level annotations of large datasets are still hard to obtain. This weakly supervised problem cannot be solved by MIL. We cannot assume that at least one ?completely? positive
instance (i.e. a CPMC segment proposals) is present in a positive bag (i.e. a groundtruth instance)
since most of the proposals will contain both foreground pixels and background pixels. We will
show how our e-SVM and its latent extension address this problem in the next sections.
In machine learning literature, the weighted SVM (WSVM) methods [23, 27, ?] also use an instancedependent weight on the cost of each example, and can improve the robustness of model estimation
[23], alleviate the effect of outliers [27], leverage privileged information [17] or deal with unbalanced classification problems. The difference between our e-SVM and WSVMs mainly lies in that
it weights labels instead of data points, which leads to each example contributing both to the costs
of positive and negative labels. Although the loss function of e-SVM model is different from those
of WSVMs, it can be effortlessly solved by any standard SVM solver (e.g., LibLinear [10]) like
those used in WSVMs. This is an advantage because it does not require a specific solver for the
implementation of our e-SVM.
3
The expectation loss SVM model
In this section, we will first describe the basic formulation of our expectation loss SVM model
(e-SVM) in section 3.1 when the positiveness of each segment proposal is observed. Then, in section 3.2, a latent e-SVM model is introduced to handle the weak supervision situation where the
positiveness of each segment proposal is unobserved.
3.1
The basic e-SVM model
We are given a set of training images D. Using some segmentation method (we adopt CPMC [4]
in this work), we can generate a set of foreground segment proposals {S1 , S2 , . . . , SN } from these
images. For each segment Si , we extract feature xi , xi ? Rd .
Suppose the pixelwise annotations are available for all the groundtruth instances in D. For each
object class, we can calculate the IoU ratio ui (ui ? [0, 1]) between each segment Si and the
groundtruth instances labeling, and set the positiveness of Si as ui (although positiveness can be
some functions of IoU ratio, for simplicity, we just set it as IoU and use ui to represent the positiveness in the following paragraphs). Because many foreground segments overlap partially with
the groundtruth instances (i.e. 0 < ui < 1), it is not a standard binary classification problem for
training. Of course, we can define a threshold ?b and treat all the segments whose ui ? ?b as positive
examples and the segments whose ui < ?b as negative examples. In this way, this problem is transferred to a Support Vector Classification (SVC) problem. But it needs some heuristics to determine
?b and its performance is only partially satisfactory [18].
3
To address this issue, we proposed our expectation loss SVM model as an extension of the classical
SVC models. In this model, we treat the label Yi of each segment as an unobserved random variable.
Yi ? {?1, +1}. Given xi , we assume that Yi follows a Bernoulli distribution. The probability of
Yi = 1 given xi (i.e. the success probability of the Bernoulli distribution) is denoted as ?i . We
assume that ?i is a function of the positiveness ui , i.e. ?i = g(ui ). In the experiment, we simply set
?i = ui .
Similar to the traditional linear SVC problem, we adopt a linear function as the prediction function:
F (xi ) = wT xi + b. For simplicity, we denote [w b] as w, [xi 1] as xi and F (xi ) = wT xi in the
remaining part of the paper. The loss function of our e-SVM is the expectation over the random
variables Yi :
N
1
1 X
L(w) =?w ? wT w +
EY [max(0, 1 ? Yi wT xi )]
2
N i=1 i
N
1
1 X +
[l ? Pr(Yi = +1|xi ) + li? ? Pr(Yi = ?1|xi )]
=?w ? wT w +
2
N i=1 i
(1)
N
1
1 X +
=?w ? wT w +
{l ? g(ui ) + li? ? [1 ? g(ui )]}
2
N i=1 i
where li+ = max(0, 1 ? wT xi ) and li? = max(0, 1 + wT xi ).
Given the pixelwise groundtruth annotations, g(ui ) is known. From Equation 1, we can see that it
is equivalent to ?weight? each sample with a function of its positiveness. The standard linear SVM
solver is used to solve this model with loss function of L(w). In the experiments, we show that
the performance of our e-SVM is much better than SVC and slightly better than Support Vector
Regression (SVR) in the segment classification task.
3.2
The latent e-SVM model
One of the advantage of our e-SVM model is that we can easily extend it to the situation where only
bounding box annotations are available (this type of labeling is of most interest in the paper). Under
this weakly supervised setting, we cannot obtain the exact value of the positiveness (IoU) ui for each
segment. Instead, ui will be treated as a latent variable which will be determined by minimizing the
following loss function:
N
1 X +
1
L(w, u) = ?w ? wT w +
{l ? g(ui ) + li? ? [1 ? g(ui )]} + ?R ? R(u)
2
N i=1 i
(2)
where u denotes {ui }i=1,...,N . R(u) is a regularization term for u. We can see that the loss function
in Equation 1 is a special case of that in Equation 2 by setting u as constant and ?R equal to 0.
When u is fixed, L(w, u) is a standard linear SVM loss, which is convex with respect to w. When
w is fixed, L(w, u) is also a convex function if R(u) is a convex function with respect to u. The IoU
between a segment Si and groundtruth bounding boxes, denoted as ubb
i , can serve as an initialization
for ui . We can iteratively fix u and w, and solve the two convex optimization problems until it
converges. The pseudo-code for the optimization algorithm is shown in Algorithm 1.
Algorithm 1 The optimization for training latent e-SVM
Initialization:
1: u(cur) ? ubb ;
Process:
2: repeat
3:
w(new) ? arg minw L(w, u(cur) );
4:
u(new) ? arg minu L(w(new) , u);
5:
u(cur) ? u(new) ;
6: until Converge
4
If we do not add any regularization term on u (i.e. set ?R = 0), u will become 0 or 1 in the
optimization step in line 4 of algorithm 1 because the loss function becomes a linear function with
respect to u when w is fixed. It turns to be similar to a latent SVM and can lead the algorithm to
stuck in the local minimal as shown in the experiments. The regularization term will prevent this
situation under the assumption that the true value of u should be around ubb .
There are a lot of different designs of the regularization term R(u). In practice, we use the following
one based on the cross entropy between two Bernoulli distributions with success probability ubb
i and
ui respectively.
R(u) = ?
N
1 X bb
[u ? log(ui ) + (1 ? ubb
i ) ? log(1 ? ui )]
N i=1 i
N
1 X
=?
DKL [Bern(ubb
i )||Bern(ui )] + C
N i=1
(3)
where C is a constant value with respect to u. DKL (.) represents the KL distance between two
Bernoulli distributions. This regularization term is a convex function with respect to u and achieves
its minimal when u = ubb . It is a strong regularization term since its value increases very fast when
u 6= ubb .
4
4.1
Visual Tasks
Semantic segmentation
We can easily apply our e-SVM model to the semantic segmentation task with the framework proposed by Carreira et al. [5]. Firstly, CPMC segment proposals [4] are generated and the secondorder pooling features [5] are extracted from each segment. Then we train the segment classifiers
using either e-SVM or latent e-SVM according to whether the groundtruth pixel-level annotations
are available. In the testing stage, the CPMC segments are sorted based on their confidence scores
output by the trained classifiers. The top ones will be selected to produce the predicted semantic
label map.
4.2
Object detection
For the task of object detection, we can only acquire bounding-box annotations instead of pixel-level
labeling. Therefore, it is natural to apply our latent e-SVM in this task to provide complementary
information for the current object detection system.
In the state-of-the-art object detection systems [11, 13, 24, 14], the window candidates of foreground
object are extracted from images and the confidence scores are predicted on them. Window candidates are extracted either by sliding window approaches (used in e.g. the deformable part-based
model [11, 13]) or most recently, the Selective Search method [24] (used in e.g. the Region Convolutional Neural Networks [14]). This method lowers down the number of window candidates
compared to the traditional sliding window approach.
Original Image
Confidence Map
e-SVM
classifiers
Pooling in
each bins
Mapping
segment
confidence
to pixels
(a)
Features
(c)
(b)
Figure 2: The illustration of our spatial confidence map features for window candidates based on
e-SVM. The confidence scores of the segments are mapped to pixels to generate a pixel-level confidence map. We will divide a window candidate into m ? m spatial bins and pool the confidence
scores of the pixels in each bin. It leads to a m ? m dimensional feature.
5
It is not easy to directly incorporate confidence scores of the segments into these object detection
systems based on window candidates. The difficulty lies in two aspects. First, only some of the
segments are totally inside a window candidate or totally outside the window candidate. It might be
hard to calculate the contribution of the confidence score of a segment that only partially overlaps
with a window candidate. Second, the window candidates (even the groundtruth bounding boxes)
will contain some of the background regions. Some regions (e.g. the regions near the boundary of
the window candidates) will have higher probability to be the background region than the regions in
the center. Treating them equally will harm the accuracy of the whole detection system.
In order to solve these issues, we propose a new spatial confidence map feature. Given an image and
a set of window candidates, we first calculate the confidence scores of all the segments in the image
using the learned e-SVM models. The confidence score for a segment S is denoted as CfdScore(S).
For each pixel, the confidence score is set as the maximum confidence score of all the segments
that contain this pixel. CfdScore(p) = max?S,p?S CfdScore(S). In this way, we can handle the
difficulty of partial overlapping between segments and candidate windows. For the second difficulty,
we divide each candidate window into M = m ? m spatial bins and pool the confidence scores of
the pixels in each bin. Because the classifiers are trained with the one-vs-all scheme, our spatial
confidence map feature is class-specific. It leads to a (M ? K)-dimensional feature for each candidate window, where K refers to the total number of object classes. After that, we encode it by
additive kernels approximation mapping [25] and obtain the final feature representation of candidate
windows. The feature generating process is illustrated in Figure 2. In the testing stage, we can
concatenate this segment feature with the features from other object detection systems.
5
Experiments
In this section, we first evaluate the performance of e-SVM method on segment proposal classification, by using two new evaluation criterions for this task. After that, we apply our method to two
essential tasks in computer vision: semantic segmentation and object detection. For semantic segmentation task, we test the proposed eSVM and latent eSVM on two different scenarios (i.e., with
pixel-level groundtruth label annotation and with only bounding-box object annotation) respectively.
For object detection task, we combine our confidence map feature with the state-of-the-art object detection system, and show our method can obtain non-trivial improvement on detection performance.
5.1
Performance evaluation on e-SVM
We use PASCAL VOC 2011 [9] segmentation dataset in this experiment. It is a subset of the whole
PASCAL 2011 datasets with 1112 images in the training set and 1111 images in the validation set,
with 20 foreground object classes in total. We use the official training set and validation set for
training and testing respectively. Similar to [5], we extract 150 CPMC [4] segment proposals for
each image and compute the second-order pooling features on each segment. Besides, we use the
same sequential pasting scheme [5] as the inference algorithm in testing.
5.1.1
Evaluation criteria
In literature [5], the supervised learning framework of segment-based prediction model either regressed the overlapping value or converted it to a binary classification problem via a threshold value, and evaluate the performance by certain task-specific criterion (i.e., the pixel-wise accuracy used
for semantic segmentation). In this paper, we adopt a direct performance evaluation criteria for the
segment-wise target class prediction task, which is consistent with the learning problem itself and
not biased to particular tasks. Unfortunately, we have not found any work on this sort of direct
performance evaluation, and thus introduce two new evaluation criteria for this purpose. We first
briefly describe them as follows:
Threshold Average Precision Curve (TAPC) Although the ground-truth target value (i.e., the
overlap rate of segment and bounding box) is a real value in the range of [0, 1], we can transform
original prediction problem to a series of binary problems, each of which is conducted by thresholding the original groundtruth overlap rate. Thus, we calculate the Precison-Recall Curve as well
as AP on each of binary classification problem, and compute the mean AP w.r.t. different threshold
values as a performance measurement for the segment-based class confidence prediction problem.
6
e-SVM
SVR
SVC-0.0
SVC-0.2
SVC-0.4
SVC-0.6
SVC-0.8
TAPC
36.69
35.23
22.48
33.96
35.62
32.57
26.73
e-SVM
SVR
SVC-0
SVC-0.2
SVC-0.4
SVC-0.6
SVC-0.8
NDCG
0.8750
0.8652
0.8153
0.8672
0.8656
0.8485
0.8244
TAPC
TAPC
38
36
34
32
30
28
26
24
22
20
33.00
0.8700
29.50
0.8500
26.00
0.8300
22.50
0.8100
19.00
0.7900
15.50
0.7700
0.7500
12.00
e-SVM
SVR
L-eSVM
SVC-0.0 SVC-0.2 SVC-0.4 SVC-0.6 SVC-0.8
SVR
SVC-0.0 SVC-0.2 SVC-0.4 SVC-0.6 SVC-0.8
NDCG
NDCG
0.8800
0.8700
0.8700
0.8500
0.8600
0.8300
0.8500
0.8400
0.8100
0.8300
0.7900
0.8200
0.7700
0.8100
0.7500
0.8000
e-SVM
SVR
L-eSVM
SVC-0 SVC-0.2 SVC-0.4 SVC-0.6 SVC-0.8
(a) Using pixel level annotations
SVR
SVC-0 SVC-0.2 SVC-0.4 SVC-0.6 SVC-0.8
(b) Using bounding box annotations
Figure 3: Performance evaluation and comparison to SVC and SVR
Normalized Discounted Cumulative Gain (NDCG) [15] Considering that a higher confidence
value is expected to be predicted for the segment with higher overlap rate, we think this prediction
problem can be treated as a ranking problem, and thus we use the Normalized Discounted Cumulative Gain (NDCG), which is common performance measurement for ranking problem, as another
kind of performance evaluation criterion in this paper.
5.1.2
Comparisons to SVC and SVR
Based on the TAPC and NDCG introduced above, we evaluate the performance of our e-SVM model
on PASCAL VOC 2011 segmentation dataset, and compare the results to two common methods (i.e.
SVC and SVR) in literature. Note that we test the SVC?s performance with a variety of binary classification problems, each of which are trained by using different threshold values (e.g., 0, 0.2, 0.4,
0.6 and 0.8 as shown in figure 3). In figure 3 (a) and (b), we show the experimental results w.r.t. the
model/classifier trained with clean pixel-wise object class labels and weakly-labelled bounding-box
annotation, respectively. For both cases, we can see that our method obtains consistently superior
performance than SVC model for all different threshold values. Besides, we can see that the TAPC
and NDCG of our method are higher than those of SVR, which is a popular regression model for
continuously valued target variable based on the max-margin principle.
5.2
Results of semantic segmentation
For the semantic segmentation task, we test our e-SVM model with PASCAL VOC 2011 segmtation
dataset using training set for training and validation set for testing. We evaluate the performance
under two different data annotation settings, i.e., training with pixel-wise semantic class label maps
and object bounding-box annotations. The accuracy w.r.t. these two settings are 36.8% and 27.7%
respectively, which are comparable to the results of the state-of-the-art segment confidence prediction model (i.e., SVR) [5] used in semantic segmentation task.
5.3
Results of object detection
As mentioned in Section 4.2, one of the natural applications of our e-SVM method is the object
detection task. Most recently, Girshick et.al [14] presented a Regions with CNN features method
(RCNN) using the Convolutional Neural Network pre-trained on the ImageNet Dataset [6] and finetuned on the PASCAL VOC datasets. They achieved a significantly improvement over the previous
state-of-the-art algorithms (e.g. Deformable Part-based Model (DPM) [11])and push the detection
7
RCNN
Ours
Gain
RCNN (bb)
Ours (bb)
Gain
RCNN
Ours
Gain
RCNN (bb)
Ours (bb)
Gain (bb)
plane
64.1
63.7
-0.4
68.1
70.4
2.3
table
45.8
47.8
2.0
54.5
56.4
1.9
bike
69.2
70.2
1.0
72.8
74.2
1.4
dog
55.8
57.9
2.1
61.2
62.9
1.8
bird
50.4
51.9
1.5
56.8
59.1
2.3
horse
61.0
61.2
0.3
69.1
69.3
0.2
boat
41.2
42.5
1.3
43.0
44.7
1.6
motor.
66.8
67.5
0.8
68.6
69.9
1.4
bottle
33.2
33.4
0.2
36.8
38.0
1.2
person
53.9
54.9
1.0
58.7
59.6
0.9
bus
62.8
63.2
0.4
66.3
67.2
1.0
plant
30.9
34.5
3.7
33.4
35.6
2.2
car
70.5
71.3
0.8
74.2
74.6
0.3
sheep
53.3
55.8
2.5
62.9
64.6
1.7
cat
61.8
62.0
0.2
67.6
69.0
1.3
sofa
49.2
51.0
1.8
51.1
53.2
2.1
chair
32.4
34.7
2.3
34.4
36.7
2.3
train
56.9
58.4
1.6
62.5
64.3
1.8
cow
58.4
58.7
0.2
63.5
64.3
0.8
tv
64.1
65.0
0.9
64.8
65.5
0.7
Average
54.1
55.3
1.2
58.5
60.0
1.5
Table 1: Detection results on PASCAL 2007. ?bb? means the result after applying bounding box
regression. Gain means the improved AP of our system compared to RCNN under the same settings
(both with bounding box or without). The better results in the comparisons are bold.
performance into a very high level (The average AP is 58.5 with boundary regularization on PASCAL VOC 2007).
A question arises: can we further improve their performance? The answer is yes. In our method,
we first learn the latent e-SVM models based on the object bounding-box annotation, and calculate
the spatial confidence map features as in section 4.2. Then we simply concatenate them with RCNN
the features to train object classifiers on candidate windows. We use PASCAL VOC 2007 dataset
in this experiment. As shown in table 1, our method can improve the average AP by 1.2 before
applying bounding boxes regression. For some categories that the original RCNN does not perform
well, such as potted plant, the gain of AP is up to 3.65. After applying bounding box regression for
both RCNN and our algorithm, the gain of performance is 1.5 on average.
In the experiment, we set m = 5 and adopt average pooling on the pixel level confidence scores
within each spatial bin. We also modified the bounding box regularization method used in [14] by
augmenting the fifth layer features with additive kernels approximation methods [25]. It will lead to
a slightly improved performance.
In summary, we achieved an average AP of 60.0, which is 1.5 higher than the best known results
(the original RCNN with bounding box regression) of this dataset. Please note that we only use the
annotations on PASCAL VOC 2007 to train the e-SVM classifiers and have not considered context.
The results are expected to be further improved if the data in ImageNet is used.
6
Conclusion
We present a novel learning algorithm call e-SVM that can well handle the situation in which the
labels of training data are continuous values whose range is a bounded interval. It can be applied
to segment proposal classification task and can be easily extended to learn segment classifiers under
weak supervision (e.g. only bounding box annotations are available). We apply this method on
two major tasks of computer vision (i.e., semantic segmentation and object detection), and obtain
the state-of-the-art object detection performance on PASCAL VOC 2007 dataset. We believe that,
with the ever growing size of datesets, it is increasingly important to learn segment classifiers under
weak supervision to reduce the amount of labeling required. In future work, we will consider using
the bounding box annotation from large datasets, such as ImageNet, to further improve semantic
segmentation performance on PASCAL VOC.
Acknowledgements. We gratefully acknowledge funding support from the National Science Foundation (NSF) with award CCF-1317376, and from the National Institute of Health NIH Grant
5R01EY022247-03. We also thank the NVIDIA Corporation for providing GPUs in our experiments.
8
References
[1] R. Achanta, A. Shaji, K. Smith, A. Lucchi, P. Fua, and S. Susstrunk. SLIC superpixels compared to
state-of-the-art superpixel methods. TPAMI, 34(11):2274?2282, 2012.
[2] S. Andrews, I. Tsochantaridis, and T. Hofmann. Support vector machines for multiple-instance learning.
In Advances in Neural Information Processing Systems 15, pages 561?568. MIT Press, 2003.
[3] P. Arbelaez, B. Hariharan, C. Gu, S. Gupta, and J. Malik. Semantic segmentation using regions and parts.
In CVPR, 2012.
[4] J. Carreira and C. Sminchisescu. Cpmc: Automatic object segmentation using constrained parametric
min-cuts. TPAMI, 34(7):1312?1328, 2012.
[5] J. a. Carreira, R. Caseiro, J. Batista, and C. Sminchisescu. Semantic segmentation with second-order
pooling. In ECCV, pages 430?443, 2012.
[6] J. Deng, A. Berg, , J. Winn, and A. Zisserman. The PASCAL Visual Object Classes Challenge 2010
(VOC2010) Results. http://www.image-net.org/challenges/LSVRC/2012/index.
[7] T. G. Dietterich, R. H. Lathrop, and T. Lozano-P?erez. Solving the multiple instance problem with axisparallel rectangles. Artif. Intell., 89(1-2):31?71, Jan. 1997.
[8] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman.
The
PASCAL Visual Object Classes Challenge 2007 (VOC2007) Results.
http://www.pascalnetwork.org/challenges/VOC/voc2007/workshop/index.html.
[9] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman.
The
PASCAL Visual Object Classes Challenge 2011 (VOC2011) Results.
http://www.pascalnetwork.org/challenges/VOC/voc2011/workshop/index.html.
[10] R.-E. Fan, K.-W. Chang, C.-J. Hsieh, X.-R. Wang, and C.-J. Lin. LIBLINEAR: A library for large linear
classification. JMLR, 9:1871?1874, 2008.
[11] P. F. Felzenszwalb, R. B. Girshick, D. McAllester, and D. Ramanan. Object detection with discriminatively trained part-based models. TPAMI, 32(9):1627?1645, 2010.
[12] P. F. Felzenszwalb and D. P. Huttenlocher. Efficient graph-based image segmentation. IJCV, 59(2):167?
181, Sept. 2004.
[13] S. Fidler, R. Mottaghi, A. L. Yuille, and R. Urtasun. Bottom-up segmentation for top-down detection. In
CVPR, pages 3294?3301, 2013.
[14] R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich feature hierarchies for accurate object detection
and semantic segmentation. In CVPR, 2014.
[15] K. J?arvelin and J. Kek?al?ainen. Cumulated gain-based evaluation of ir techniques. TOIS, 20(4):422?446,
2002.
[16] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural
networks. In NIPS, pages 1106?1114, 2012.
[17] M. Lapin, M. Hein, and B. Schiele. Learning using privileged information: Svm+ and weighted svm.
Neural Networks, 53:95?108, 2014.
[18] F. Li, J. Carreira, and C. Sminchisescu. Object recognition as ranking holistic figure-ground hypotheses.
In CVPR, pages 1712?1719, 2010.
[19] Y. Liu, J. Liu, Z. Li, J. Tang, and H. Lu. Weakly-supervised dual clustering for image semantic segmentation. In Computer Vision and Pattern Recognition (CVPR), 2013 IEEE Conference on, pages 2075?2082.
IEEE, 2013.
[20] A. M?uller and S. Behnke. Multi-instance methods for partially supervised image segmentation. In PSL,
pages 110?119, 2012.
[21] X. Ren, L. Bo, and D. Fox. Rgb-(d) scene labeling: Features and algorithms. In CVPR, June 2012.
[22] J. Shotton, M. Johnson, and R. Cipolla. Semantic texton forests for image categorization and segmentation. In CVPR, pages 1?8, 2008.
[23] J. Suykens, J. D. Brabanter, L. Lukas, and J. Vandewalle. Weighted least squares support vector machines:
robustness and sparse approximation. NEUROCOMPUTING, 48:85?105, 2002.
[24] J. Uijlings, K. van de Sande, T. Gevers, and A. Smeulders. Selective search for object recognition. IJCV,
104(2):154?171, 2013.
[25] A. Vedaldi and A. Zisserman. Efficient additive kernels via explicit feature maps. TPAMI, 34(3):480?492,
2012.
[26] A. Vezhnevets, V. Ferrari, and J. M. Buhmann. Weakly supervised structured output learning for semantic
segmentation. In CVPR, pages 845?852, 2012.
[27] X. Yang, Q. Song, and A. Cao. Weighted support vector machine for data classification. In IJCNN, 2005.
9
| 5287 |@word cnn:2 middle:1 briefly:2 everingham:2 rgb:1 hsieh:1 liblinear:2 liu:3 series:1 score:14 tuned:1 ours:4 batista:1 outperforms:3 existing:1 current:1 si:4 parsing:1 readily:1 concatenate:2 additive:3 hofmann:1 motor:1 treating:2 ainen:1 v:1 cue:6 selected:1 plane:1 smith:1 potted:1 firstly:2 org:3 direct:2 become:1 precison:1 ijcv:2 combine:1 paragraph:1 inside:1 introduce:1 expected:2 growing:1 multi:2 voc:15 discounted:2 window:19 solver:3 totally:2 mjhustc:1 becomes:1 considering:1 bounded:3 bike:1 kind:1 developed:1 unobserved:4 corporation:1 impractical:1 pasting:1 pseudo:1 classifier:16 ramanan:1 grant:1 producing:1 segmenting:1 positive:7 before:1 local:1 treat:3 ap:8 ndcg:7 might:2 bird:1 initialization:2 achanta:1 quantified:1 range:3 testing:7 union:2 practice:1 jan:1 significantly:1 vedaldi:1 confidence:30 word:1 refers:1 pre:1 svr:16 get:1 cannot:3 tsochantaridis:1 context:1 applying:3 www:3 equivalent:1 map:11 center:1 williams:2 convex:5 simplicity:2 handle:7 ferrari:1 target:3 suppose:1 heavily:1 hierarchy:1 exact:3 hypothesis:1 secondorder:1 superpixel:1 expensive:1 finetuned:1 recognition:3 continues:1 cut:1 labeled:2 huttenlocher:1 observed:3 bottom:1 solved:2 wang:1 calculate:5 region:9 mentioned:1 ui:23 schiele:1 trained:8 weakly:13 solving:1 segment:52 arvelin:1 yuille:3 serve:1 completely:1 gu:1 easily:4 voc2010:1 cat:1 train:9 fast:1 effective:1 describe:2 axisparallel:1 labeling:7 horse:1 outside:1 whose:4 heuristic:2 solve:5 valued:1 cvpr:8 statistic:3 think:1 transform:1 itself:1 final:1 hoc:1 advantage:2 tpami:4 brabanter:1 net:1 propose:2 relevant:1 cao:1 holistic:1 achieve:1 deformable:2 validate:2 los:3 exploiting:1 sutskever:1 darrell:1 produce:1 generating:1 categorization:1 converges:1 object:40 coupling:1 andrew:1 augmenting:1 stat:1 progress:1 strong:5 predicted:3 iou:10 correct:2 mcallester:1 bin:6 require:1 fix:1 alleviate:1 secondly:1 extension:2 accompanying:1 effortlessly:1 around:2 considered:1 ground:2 minu:1 mapping:2 predict:1 major:2 achieves:2 adopt:4 purpose:1 estimation:1 sofa:1 bag:3 label:16 currently:1 correctness:1 weighted:6 uller:1 mit:1 always:1 super:1 modified:1 mil:4 encode:1 susstrunk:1 june:1 improvement:3 consistently:1 bernoulli:4 mainly:1 superpixels:1 inference:1 typically:1 selective:3 pixel:29 arg:2 issue:2 classification:17 flexible:1 pascal:21 among:1 dual:2 denoted:3 html:2 art:8 spatial:8 special:2 constrained:1 equal:1 manually:1 represents:1 foreground:5 future:1 national:2 intell:1 neurocomputing:1 detection:28 interest:1 evaluation:9 sheep:1 devoted:1 accurate:1 partial:1 minw:1 fox:1 divide:2 hein:1 girshick:3 minimal:2 instance:12 cost:2 subset:1 krizhevsky:1 vandewalle:1 conducted:1 johnson:1 pixelwise:2 answer:1 learnt:2 person:1 caseiro:1 fundamental:1 pool:2 continuously:1 lucchi:1 again:1 choose:1 li:7 converted:1 de:1 bold:1 ranking:3 ad:1 later:1 lot:1 sort:1 annotation:32 gevers:1 contribution:1 smeulders:1 square:1 ir:1 hariharan:1 accuracy:3 convolutional:3 kek:1 yield:1 yes:1 weak:6 lu:1 ren:1 published:1 detector:4 couple:2 cur:3 gain:10 dataset:12 popular:1 recall:1 color:1 car:1 segmentation:32 higher:5 supervised:13 zisserman:4 improved:3 fua:1 formulation:1 box:29 just:1 stage:4 until:2 hand:2 overlapping:2 believe:1 artif:1 effect:1 dietterich:1 contain:3 true:2 normalized:2 ccf:1 lozano:1 regularization:8 fidler:1 pascalnetwork:2 iteratively:1 satisfactory:2 semantic:22 illustrated:1 deal:2 please:1 criterion:6 generalized:1 percent:1 image:21 wise:4 novel:2 svc:43 recently:5 funding:1 common:2 superior:1 nih:1 vezhnevets:2 winner:1 extend:2 measurement:3 positiveness:25 rd:1 automatic:1 erez:1 gratefully:1 supervision:9 impressive:1 base:3 add:1 recent:1 showed:1 belongs:1 scenario:3 certain:1 nvidia:1 sande:1 affiliation:1 binary:9 success:2 yi:8 mottaghi:1 additional:1 ey:1 deng:1 determine:2 converge:1 ii:1 sliding:2 multiple:3 full:1 alan:1 cross:1 lin:1 equally:1 award:1 dkl:2 privileged:2 prediction:9 regression:8 basic:3 vision:6 expectation:7 represent:2 kernel:3 texton:1 achieved:2 suykens:1 proposal:27 addition:1 background:3 fine:1 interval:3 winn:3 biased:1 pooling:5 dpm:1 effectiveness:1 call:1 near:1 leverage:1 yang:1 shotton:1 enough:1 easy:1 variety:1 behnke:1 cow:1 reduce:1 angeles:3 psl:1 whether:1 song:1 deep:3 generally:1 useful:1 detailed:1 amount:1 svms:1 category:1 simplest:1 generate:4 http:3 nsf:1 per:1 slic:1 putting:1 threshold:7 prevent:1 clean:1 utilize:1 rectangle:1 graph:1 cpmc:8 almost:1 groundtruth:15 comparable:1 layer:1 fan:1 ijcnn:1 scene:1 regressed:1 ucla:3 aspect:1 chair:1 min:1 shaji:1 gpus:1 transferred:1 department:3 tv:1 according:1 structured:1 wsvm:1 slightly:2 increasingly:1 s1:1 outlier:1 pr:2 equation:3 bus:1 discus:1 turn:1 adopted:1 available:12 apply:4 alternative:2 robustness:2 original:5 denotes:1 clustering:2 remaining:1 top:2 tois:1 classical:1 malik:2 already:1 question:1 degrades:1 mim:4 parametric:1 traditional:2 distance:1 thank:1 mapped:1 arbelaez:1 evaluate:4 binarize:1 trivial:1 urtasun:1 besides:3 code:1 index:3 illustration:2 ratio:6 minimizing:1 acquire:1 providing:1 unfortunately:1 negative:2 implementation:1 design:1 perform:1 datasets:8 acknowledge:1 situation:5 extended:2 ever:1 hinton:1 introduced:3 dog:2 required:2 specified:1 kl:1 connection:1 imagenet:8 bottle:1 california:3 learned:2 nip:1 address:5 usually:1 pattern:1 challenge:7 max:5 gool:2 overlap:8 treated:2 rely:1 natural:2 difficulty:3 buhmann:1 boat:1 zhu:1 scheme:2 improve:5 voc2007:2 library:1 jun:1 extract:4 health:1 sept:1 sn:1 review:1 literature:4 acknowledgement:1 contributing:1 loss:13 plant:2 discriminatively:1 validation:3 rcnn:11 foundation:1 consistent:1 thresholding:1 principle:1 eccv:1 course:1 summary:1 repeat:1 bern:2 institute:1 felzenszwalb:2 lukas:1 fifth:1 sparse:1 van:3 boundary:2 curve:2 cumulative:2 rich:2 stuck:1 bb:7 obtains:1 silhouette:1 harm:1 assumed:1 xi:15 continuous:2 latent:16 search:3 quantifies:1 table:3 learn:4 forest:1 sminchisescu:3 uijlings:1 official:1 voc2011:2 junhua:1 bounding:29 s2:1 whole:2 complementary:1 precision:2 mao:1 explicit:1 concatenating:1 candidate:19 lie:2 jmlr:1 donahue:1 tang:1 down:2 specific:3 svm:58 gupta:1 essential:1 workshop:2 sequential:1 effectively:1 cumulated:1 labelling:1 illustrates:1 push:1 margin:2 entropy:1 intersection:2 simply:2 appearance:5 visual:4 partially:5 bo:1 chang:1 cipolla:1 truth:1 relies:1 extracted:4 viewed:1 sorted:1 labelled:3 hard:2 carreira:4 determined:1 lsvrc:1 wt:9 called:3 total:2 lathrop:1 experimental:1 berg:1 support:8 arises:1 unbalanced:1 incorporate:1 lapin:1 |
4,735 | 5,288 | Predicting Useful Neighborhoods
for Lazy Local Learning
Kristen Grauman
University of Texas at Austin
[email protected]
Aron Yu
University of Texas at Austin
[email protected]
Abstract
Lazy local learning methods train a classifier ?on the fly? at test time, using only
a subset of the training instances that are most relevant to the novel test example.
The goal is to tailor the classifier to the properties of the data surrounding the test
example. Existing methods assume that the instances most useful for building the
local model are strictly those closest to the test example. However, this fails to
account for the fact that the success of the resulting classifier depends on the full
distribution of selected training instances. Rather than simply gathering the test
example?s nearest neighbors, we propose to predict the subset of training data that
is jointly relevant to training its local model. We develop an approach to discover
patterns between queries and their ?good? neighborhoods using large-scale multilabel classification with compressed sensing. Given a novel test point, we estimate
both the composition and size of the training subset likely to yield an accurate local
model. We demonstrate the approach on image classification tasks on SUN and
aPascal and show its advantages over traditional global and local approaches.
1
Introduction
Many domains today?vision, speech, biology, and others?are flush with data. Data availability, combined with recent large-scale annotation efforts and crowdsourcing developments, have
yielded labeled datasets of unprecedented size. Though a boon for learning approaches, large labeled datasets also present new challenges. Beyond the obvious scalability concerns, the diversity
of the data can make it difficult to learn a single global model that will generalize well. For example,
a standard binary dog classifier forced to simultaneously account for the visual variations among
hundreds of dog breeds may be ?diluted? to the point it falls short in detecting new dog instances.
Furthermore, with training points distributed unevenly across the feature space, the model capacity
required in any given region of the space will vary. As a result, if we train a single high capacity
learning algorithm, it may succeed near parts of the decision boundary that are densely populated
with training examples, yet fail in poorly sampled areas of the feature space.
Local learning methods offer a promising direction to address these challenges. Local learning is an
instance of ?lazy learning?, where one defers processing of the training data until test time. Rather
than estimate a single global model from all training data, local learning methods instead focus on a
subset of the data most relevant to the particular test instance. This helps learn fine-grained models
tailored to the new input, and makes it possible to adjust the capacity of the learning algorithm to
the local properties of the data [5]. Local methods include classic nearest neighbor classification as
well as various novel formulations that use only nearby points to either train a model [2, 3, 5, 13, 29]
or learn a feature transformation [8, 9, 15, 25] that caters to the novel input.
A key technical question in local learning is how to determine which training instances are relevant
to a test instance. All existing methods rely on an important core assumption: that the instances most
useful for building a local model are those that are nearest to the test example. This assumption is
well-motivated by the factors discussed above, in terms of data density and intra-class variation.
1
Furthermore, identifying training examples solely based on proximity has the appeal of permitting
specialized similarity functions (whether learned or engineered for the problem domain), which can
be valuable for good results, especially in structured input spaces.
On the other hand, there is a problem with this core assumption. By treating the individual nearness
of training points as a metric of their utility for local training, existing methods fail to model how
those training points will actually be employed. Namely, the relative success of a locally trained
model is a function of the entire set or distribution of the selected data points?not simply the
individual pointwise nearness of each one against the query. In other words, the ideal target subset
consists of a set of instances that together yield a good predictive model for the test instance.
Based on this observation, we propose to learn the properties of a ?good neighborhood? for local
training. Given a test instance, the goal is to predict which subset of the training data should be
enlisted to train a local model on the fly. The desired prediction task is non-trivial: with a large
labeled dataset, the power set of candidates is enormous, and we can observe relatively few training
instances for which the most effective neighborhood is known. We show that the problem can be
cast in terms of large-scale multi-label classification, where we learn a mapping from an individual
instance to an indicator vector over the entire training set that specifies which instances are jointly
useful to the query. Our approach maintains an inherent bias towards neighborhoods that are local,
yet makes it possible to discover subsets that (i) deviate from a strict nearest-neighbor ranking and
(ii) vary in size.
The proposed technique is a general framework to enhance local learning. We demonstrate its impact
on image classification tasks for computer vision, and show its substantial advantages over existing
local learning strategies. Our results illustrate the value in estimating the size and composition of
discriminative neighborhoods, rather than relying on proximity alone.
2
Related Work
Local learning algorithms Lazy local learning methods are most relevant to our work. Existing methods primarily vary in how they exploit the labeled instances nearest to a test point. One
strategy is to identify a fixed number of neighbors most similar to the test point, then train a model
with only those examples (e.g., a neural network [5], SVM [29], ranking function [3, 13], or linear
regression [2]). Alternatively, the nearest training points can be used to learn a transformation of
the feature space (e.g., Linear Discriminant Analysis); after projecting the data into the new space,
the model is better tailored to the query?s neighborhood properties [8, 9, 15, 25]. In local selection
methods, strictly the subset of nearby data is used, whereas in locally weighted methods, all training
points are used but weighted according to their distance [2]. All prior methods select the local neighborhood based on proximity, and they typically fix its size. In contrast, our idea is to predict the set
of training instances that will produce an effective discriminative model for a given test instance.
Metric learning The question ?what is relevant to a test point?? also brings to mind the metric
learning problem. Metric learning methods optimize the parameters of a distance function so as to
best satisfy known (dis)similarity constraints between training data [4]. Most relevant to our work
are those that learn local metrics; rather than learn a single global parameterization, the metric varies
in different regions of the feature space. For example, to improve nearest neighbor classification,
in [11] a set of feature weights is learned for each individual training example, while in [26, 28]
separate metrics are trained for clusters discovered in the training data.
Such methods are valuable when the data is multi-modal and thus ill-suited by a single global metric.
Furthermore, one could plug a learned metric into the basic local learning framework. However,
we stress that learning what a good neighbor looks like (metric learning?s goal) is distinct from
learning what a good neighborhood looks like (our goal). Whereas a metric can be trained with
pairwise constraints indicating what should be near or far, jointly predicting the instances that ought
to compose a neighborhood requires a distinct form of learning, which we tackle in this work.
Hierarchical classification For large multi-class problems, hierarchical classification approaches
offer a different way to exploit ?locality? among the training data. The idea is to assemble a tree of
decision points, where at each node only a subset of labels are considered (e.g., [6, 12, 21]). Such
methods are valuable for reducing computational complexity at test time, and broadly speaking they
share the motivation of focusing on finer-grained learning tasks to improve accuracy. However,
2
otherwise the work is quite distant from our problem. Hierarchical methods precompute groups of
labels to isolate in classification tasks, and apply the same classifiers to all test instances; lazy local
learning predicts at test time what set of training instances are relevant for each novel test instance.
Weighting training instances Our problem can be seen as deciding which training instances to
?trust? most. Various scenarios call for associating weights with training instances such that some
influence the learned parameters more than others. For example, weighted instances can reflect label
confidences [27], help cope with imbalanced training sets [24], or resist the influence of outliers [20].
However, unlike our setting, the weights are given at training time and they are used to create a
single global model. Methods to estimate the weights per example arise in domain adaptation,
where one aims to give more weight to source domain samples distributed most like those in the
target domain [14, 17, 18]. These are non-local, offline approaches, whereas we predict useful
neighborhoods in an online, query-dependent manner. Rather than close the mismatch between a
source and target domain, we aim to find a subset of training data amenable to a local model.
Active learning Active learning [23] aims to identify informative unlabeled training instances,
with the goal of minimizing labeling effort when training a single (global) classifier. In contrast, our
goal is to ignore those labeled training points that are irrelevant to a particular novel instance.
3
Approach
We propose to predict the set of training instances which, for a given test example, are likely to
compose an effective neighborhood for local classifier learning. We use the word ?neighborhood?
to refer to such a subset of training data?though we stress that the optimal subset need not consist
of strictly rank-ordered nearest neighbor points.
Our approach has three main phases: (i) an offline stage where we generate positive training neighborhoods (Sec. 3.1), (ii) an offline stage where we learn a mapping from individual examples to their
useful neighborhoods (Sec. 3.2), and (iii) an online phase where we apply the learned model to infer
a novel example?s neighborhood, train a local classifier, and predict the test label (Sec. 3.3).
3.1
Generating training neighborhoods
Let T = {(x1 , c1 ), . . . , (xM , cM )} denote the set of M category-labeled training examples. Each
xi ? <d is a vector in some d-dimensional feature space, and each ci ? {1, . . . , C} is its target
category label. Given these examples, we first aim to generate a set of training neighborhoods, N =
{(xn1 , yn1 ), . . . , (xnN , ynN )}. Each training neighborhood (xni , yni ) consists of an individual
instance xni paired with a set of training instance indices capturing its target ?neighbors?, the latter
being represented as a M -dimensional indicator vector yni . If yni (j) = 1, this means xj appears
in the target neighborhood for xni . Otherwise, yni (j) = 0. Note that the dimensionality of this
target indicator vector is M , the number of total available training examples. We will generate N
such pairs, where typically N M .
As discussed above, there are very good motivations for incorporating nearby points for local learning. Indeed, we do not intend to eschew the ?locality? aspect of local learning. Rather, we start from
the premise that points near to a query are likely relevant?but relevance is not necessarily preserved
purely by their rank order, nor must the best local set be within a fixed radius of the query (or have a
fixed set size). Instead, we aim to generalize the locality concept to jointly estimate the members of
a neighborhood such that taken together they are equipped to train an accurate query-specific model.
With these goals in mind, we devise an empirical approach to generate the pairs (xni , yni ) ? N .
The main idea is to sample a series of candidate neighborhoods for each instance xni , evaluate their
relative success at predicting the training instance?s label, and record the best candidate.
Specifically, for instance xni , we first compute its proximity to the M ? 1 other training images in
the feature space. (We simply apply Euclidean distance, but a task-specific kernel or learned metric
could also be used here.) Then, for each of a series of possible neighborhood sizes {k1 , . . . , kK },
we sample a neighborhood of size k from among all training images, subject to two requirements:
(i) points nearer to xni are more likely to be chosen, and (ii) the category label composition within
the neighborhood set is balanced. In particular, for each possible category label 1, . . . , C we sample
k
C training instances without replacement, where the weight associated with an instance is inversely
3
related to its (normalized) distance to xni . We repeat the sampling S times for each value of k,
yielding K ? S candidates per instance xni .
Next, for each of these candidates, we learn a local model. Throughout we employ linear support
vector machine (SVM) classifiers, both due to their training efficiency and because lower capacity
models are suited to the sparse, local datasets under consideration; however, kernelized/non-linear
models are also possible.1 Note that any number of the K ? S sampled neighborhoods may yield a
classifier that correctly predicts xni ?s category label cni . Thus, to determine which among the successful classifiers is best, we rank them by their prediction confidences. Let pks (xni ) = P (cni |xni )
be the posterior estimated by the s-th candidate classifier for neighborhood size k, as computed via
Platt scaling using the neighborhood points. To automatically select the best k for instance xni , we
average these posteriors across all samples per k value, then take the one with the highest probabilPS
ity: k ? = arg max S1 s=1 pks (xni ). The averaging step aims to smooth the estimated probability
k
using the samples for that value of k, each of which favors near points but varies in its composition.
Finally, we obtain a single neighborhood pair (xni , yni ), where yni is the indicator vector for the
?
neighborhood sampled with size k ? having the highest posterior pks .
In general we can expect higher values of S and denser samplings of k to provide best results, though
at a correspondingly higher computational cost during this offline training procedure.
3.2
Learning to predict neighborhoods with compressed sensing
With the training instance-neighborhood pairs in hand, we next aim to learn a function capturing
their relationships. This function must estimate the proper neighborhood for novel test instances.
We are faced with a non-trivial learning task. The most straightforward approach might be to learn
a binary decision function for each xi ? T , trained with all xnj for which ynj (i) = 1 as positives.
However, this approach has several problems. First, it would require training M binary classifiers,
and in the applications of interest M ?the number of all available category-labeled examples?may
be very large, easily reaching the millions. Second, it would fail to represent the dependencies
between the instances appearing in a single training neighborhood, which ought to be informative
for our task. Finally, it is unclear how to properly gather negative instances for such a naive solution.
Instead, we pose the learning task as a large-scale multi-label classification problem. In multi-label
classification, a single data point may have multiple labels. Typical examples include image and web
page tagging [16, 19] or recommending advertiser bid phrases [1]. In our case, rather than predict
which labels to associate with a novel example, we want to predict which training instances belong
in its neighborhood. This is exactly what is encoded by the target indicator vectors defined above,
yni . Furthermore, we want to exploit the fact that, compared to the number of all labeled training
images, the most useful local neighborhoods will contain relatively few examples.
Therefore, we adopt a large-scale multi-label classification approach based on compressed sensing [19] into our framework. With it, we can leverage sparsity in the high-dimensional target neighborhood space to efficiently learn a prediction function that jointly estimates all useful neighbors.
First, for each of the N training neighborhoods, we project its M -dimensional neighborhood vector yni to a lower-dimensional space using a random transformation: zni = ? yni , where ? is a
D ? M random matrix, and D denotes the compressed indicators? dimensionality. Then, we learn
regression functions to map the original features to these projected values zn1 , . . . , znN as targets.
That is, we obtain a series of D M regression functions f1 , . . . , fD minimizing the loss in the
compressed indicator vector space. Given a novel instance xq , those same regression functions are
applied to map to the reduced space, [f1 (xq ), . . . , fD (xq )]. Finally, we predict the complete indicator vector by recovering the M -dimensional vector using a standard reconstruction algorithm from
the compressed sensing literature.
We employ the Bayesian multi-label compressed sensing framework of [19], since it unifies the
regression and sparse recovery stages, yielding accurate results for a compact set of latent variables.
Due to compressed sensing guarantees, an M -dimensional indicator vector with l nonzero entries
can be recovered efficiently using D = O(l log Ml ) [16].
1
In our experiments the datasets have binary labels (C = 2); in the case of C > 2 the local model must be
multi-class, e.g., a one-versus-rest SVM.
4
3.3
Inferring the neighborhood for a novel example
All processing so far is performed offline. At test time, we are given a novel example xq , and must
predict its category label. We first predict its neighborhood using the compressed sensing approach
overviewed in the previous section, obtaining the M -dimensional vector y?q . The entries of this
vector are real-valued, and correspond to our relative confidence that each category-labeled instance
xi ? T belongs in xq ?s neighborhood.
Past multi-label classification work focuses its evaluation on the precision of (a fixed number of) the
top few most confident predictions and the raw reconstruction error [16, 19], and does not handle
the important issue of how to truncate the values to produce hard binary decisions. In contrast, our
setting demands that we extract both the neighborhood size estimate as well as the neighborhood
composition from the estimated real-valued indicator vector.
To this end, we perform steps paralleling the training procedure defined in Sec. 3.1, as follows. First,
we use the sorted confidence values in y?q to generate a series of candidate neighborhoods of sizes
varying from k1 to kK , each time ensuring balance among the category labels. That is, for each
k, we take the Ck most confident training instances per label. Recall that all M training instances
referenced by y?q have a known category label among 1, . . . , C. Analogous to before, we then apply
each of the K candidate predicted neighborhoods in turn to train a local classifier. Of those, we
return the category label prediction from the classifier with the most confident decision value.
Note that this process automatically selects the neighborhood size k to apply for the novel input. In
contrast, existing local learning approaches typically manually define this parameter and fix it for
all test examples [5, 8, 13, 15, 29]. Our results show that approach is sub-optimal; not only does the
most useful neighborhood deviate from the strict ranked list of neighbors, it also varies in size.
We previously explored an alternative approach for inference, where we directly used the confidences in y?q as weights in an importance-weighted SVM. That is, for each query, we trained a model
with all M data points, but modulated their influence according to the soft indicator vector y?q , such
that less confident points incurred lower slack penalties. However, we found that approach inferior,
likely due to the difficulty in validating the slack scale factor for all training instances (problematic
in the local learning setting) as well as the highly imbalanced datasets we tackle in the experiments.
3.4
Discussion
While local learning methods strive to improve accuracy over standard global models, their lazy use
of training data makes them more expensive to apply. This is true of any local approach that needs to
compute distances to neighbors and train a fresh classifier online for each new test example. In our
case, using Matlab, the run-time for processing a single novel test point can vary from 30 seconds to
30 minutes. It is dominated by the compressed sensing reconstruction step, which takes about 80%
of the computation time and is highly dependent on the complexity of the trained model. One could
improve performance by using approximate nearest neighbor methods to sort T , or pre-computing
a set of representative local models. We leave these implementation improvements as future work.
The offline stages of our algorithm (Secs. 3.1 and 3.2) require about 5 hours for datasets with M =
14, 000, N = 2, 000, d = 6, 300, and D = 2, 000. The run-time is dominated by the SVM
evaluation of K ? S candidate training neighborhoods on the N images, which could be performed
in parallel. The compressed sensing formulation is quite valuable for efficiency here; if we were to
instead naively train M independent classifiers, the offline run-time would be on the order of days.
We found that building category-label balance into the training and inference algorithms was crucial
for good results when dealing with highly imbalanced datasets. Earlier versions of our method
that ignored label balance would often predict neighborhoods with only the same label as the query.
Local methods typically handle this by reverting to a nearest neighbor decision. However, as we will
see below, this can be inferior to explicitly learning to identify a local and balanced neighborhood,
which can be used to build a more sophisticated classifier (like an SVM).
Finally, while our training procedure designates a single neighborhood as the prediction target, it is
determined by a necessarily limited sample of candidates (Sec. 3.1). Our confidence ranking step
accounts for the differences between those candidates that ultimately make the same label prediction.
Nonetheless, the non-exhaustive training samples mean that slight variations on the target vectors
5
may be equally good in practice. This suggests future extensions to explicitly represent ?missing?
entries in the indicator vector during training or employ some form of active learning.
4
Experiments
We validate our approach on an array of binary image classification tasks on public datasets.
Datasets We consider two challenging datasets with visual attribute classification tasks. The SUN
Attributes dataset [22] (SUN) contains 14,340 scene images labeled with binary attributes of various types (e.g., materials, functions, lighting). We use all images and randomly select 8 attribute
categories. We use the 6,300-dimensional HOG 2 ? 2 features provided by the authors, since they
perform best for this dataset [22]. The aPascal training dataset [10] contains 6,440 object images
labeled with attributes describing the objects? shapes, materials, and parts. We use all images and
randomly select 6 attribute categories. We use the base features from [10], which include color,
texture, edges, and HOG. We reduce their dimensionality to 200 using PCA. For both datasets, we
treat each attribute as a separate binary classification task (C = 2).
Implementation Details For each attribute, we compose a test set of 100 randomly chosen images
(balanced between positives and negatives), and use all other images for T . This makes M =
14, 240 for SUN and M = 6, 340 for aPascal. We use N = 2, 000 training neighborhoods for
both, and set D = {2000, 1000} for SUN and aPascal, roughly 15% of their original label indicator
lengths. Generally higher values of D yield better accuracy (less compression), but for a greater
expense. We fix the number of samples S = 100, and consider neighborhood sizes from k1 = 50
and kK = 500, in increments of 10 to 50.
Baselines and Setup We compare to the following methods: (1) Global: for each test image, we
apply the same global classifier trained with all M training images; (2) Local: for each test image,
we apply a classifier trained with only its nearest neighbors, as measured with Euclidean distance on
the image features. This baseline considers a series of k values, like our method, and independently
selects the best k per test point according to the confidence of the resulting local classifiers (see
Sec. 3.3). (3) Local+ML: same as Local, except the Euclidean distance is replaced with a learned
metric. We apply the ITML metric learning algorithm [7] using the authors? public code.
Global represents the default classification approach, and lets us gauge to what extent the classification task requires local models at all (e.g., how multi-modal the dataset is). The two Local baselines
represent the standard local learning approach [3, 5, 13, 15, 25, 29], in which proximal data points
are used to train a model per test case, as discussed in Sec. 2. By using proximity instead of y?q to
define neighborhoods, they isolate the impact of our compressed sensing approach.
All results reported for our method and the Local baselines use the automatically selected k value
per test image (cf. Sec. 3.3), unless otherwise noted. Each local method independently selects its
best k value. All methods use the exact same image features and train linear SVMs, with the cost
parameter cross-validated based on the Global baseline. To ensure the baselines do not suffer from
the imbalanced data, we show results for the baselines using both balanced (B) and unbalanced (U)
training sets. For the balanced case, for Global we randomly downsample the negatives and average
results over 10 such runs, and for Local we gather the nearest k2 neighbors from each class.
SUN Results The SUN attributes are quite challenging classification tasks. Images within the
same attribute exhibit wide visual variety. For example, the attribute ?eating? (see Fig. 1, top right) is
positive for any images where annotators could envision eating occurring, spanning from an restaurant scene, to home a kitchen, to a person eating, to a banquet table close-up. Furthermore, the
attribute may occupy only a portion of the image (e.g., ?metal? might occupy any subset of the
pixels). It is exactly this variety that we expect local learning may handle well.
Table 1 shows the results on SUN. Our method outperforms all baselines for all attributes. Global
benefits from a balanced training set (B), but still underperforms our method (by 6 points on average). We attribute this to the high intra-class variability of the dataset. Most notably, conventional
Local learning performs very poorly?whether or not we enforce balance. (Recall that the test sets
are always balanced, so chance is 0.50.) Adding metric learning to local (Local+ML) improves
things only marginally, likely because the attributes are not consistently localized in the image. We
also implemented a local metric learning baseline that clusters the training points then learns a met6
Global
B
U
0.80 0.60
0.73 0.55
0.69 0.59
0.77 0.56
0.64 0.57
0.70 0.54
0.78 0.77
0.60 0.67
Attribute
hiking
eating
exercise
farming
metal
still water
clouds
sunny
Local
B
U
0.51 0.56
0.50 0.50
0.50 0.53
0.51 0.54
0.50 0.50
0.51 0.53
0.70 0.74
0.65 0.67
Local+ML
B
U
0.55 0.65
0.50 0.51
0.50 0.53
0.52 0.57
0.50 0.51
0.51 0.52
0.74 0.75
0.62 0.60
Ours
Local
0.85
0.78
0.74
0.83
0.67
0.76
0.80
0.73
0.53
0.50
0.50
0.51
0.50
0.50
0.65
0.59
Local+ML
k = 400
0.53
0.50
0.50
0.51
0.50
0.50
0.74
0.57
Ours
0.89
0.79
0.75
0.81
0.67
0.71
0.79
0.72
Ours
Fix-k*
0.89
0.82
0.77
0.88
0.70
0.81
0.84
0.78
Table 1: Accuracy (% of correctly labeled images) for the SUN dataset. B and U refers to balanced and
unbalanced training data, respectively. All local results to left of double line use k values automatically selected
per method and per test instance; all those to the right use a fixed k for all queries. See text for details.
Local
?clouds?
Local
Ours
Ours
Local
?sunny?
Local
Ours
Local
?farming?
Ours
?exercise?
Ours
Local
?eating?
Ours
?hiking?
Figure 1: Example neighborhoods using visual similarity alone (Local) and compressed sensing inference
(Ours) on SUN. For each attribute, we show a positive test image and its top 5 neighbors. Best viewed on pdf.
ric per cluster, similar to [26, 28], then proceeds as Local+ML. Its results are similar to those of
Local+ML (see Supp. file).
The results left of the double bar correspond to auto-selected k values per query, which averaged
k = 106 with a standard deviation of 24 for our method; see Supp. file for per attribute statistics. The
rightmost columns of Table 1 show results when we fix k for all the local methods for all queries,
as is standard practice.2 Here too, our gain over Local is sizeable, assuring that Local is not at any
disadvantage due to our k auto-selection procedure.
The rightmost column, Fix-k*, shows our results had we been able to choose the optimal fixed k
(applied uniformly to all queries). Note this requires peeking at the test labels, and is something
of an upper bound. It is useful, however, to isolate the quality of our neighborhood membership
confidence estimates from the issue of automatically selecting the neighborhood size. We see there
is room for improvement on the latter.
Our method is more expensive at test time than the Local baseline due to the compressed sensing
reconstruction step (see Sec. 3.4). In an attempt to equalize that factor, we also ran an experiment
where the Local method was allowed to check more candidate k values than our method. Specifically, it could generate as many (proximity-based) candidate neighborhoods at test time as would fit
in the run-time required by our approach, where k ranges from 20 up to 6,000 in increments of 10.
Preliminary tests, however, showed that this gave no accuracy improvement to the baseline. This
indicates our method?s higher computational overhead is warranted.
Despite its potential to handle intra-class variations, the Local baseline fails on SUN because the
neighbors that look most similar are often negative, leading to near-chance accuracy. Even when
we balance its local neighborhood by label, the positives it retrieves can be quite distant (e.g., see
?exercise? in Fig. 1). Our approach, on the other hand, combines locality with what it learned about
2
We chose k = 400 based on the range where the Local baseline had best results.
7
Attribute
wing
wheel
plastic
cloth
furry
shiny
Global
B
U
0.69 0.76
0.84 0.86
0.67 0.71
0.74 0.72
0.80 0.80
0.72 0.77
Local
B
U
0.58 0.67
0.61 0.71
0.50 0.60
0.70 0.67
0.58 0.75
0.56 0.67
Local+ML
B
U
0.59 0.67
0.62 0.69
0.50 0.54
0.72 0.68
0.60 0.71
0.57 0.64
Ours
0.71
0.78
0.64
0.72
0.81
0.72
Local
0.50
0.54
0.50
0.69
0.54
0.52
Local+ML
k = 400
0.53
0.63
0.50
0.65
0.63
0.55
Ours
0.66
0.74
0.54
0.64
0.72
0.62
Ours
Fix-k*
0.78
0.81
0.67
0.77
0.82
0.73
Table 2: Accuracy (% of correctly labeled images) for the aPascal dataset, formatted as in Table 1
useful neighbor combinations, attaining much better results. Altogether, our gains over both Local
and Local+ML?20 points on average?support our central claim that learning what makes a good
neighbor is not equivalent to learning what makes a good neighborhood.
Figure 1 shows example test images and the top 5 images in the neighborhoods produced by both
Local and our approach. We stress that while Local?s neighbors are ranked based on visual similarity,
our method?s ?neighborhood? uses visual similarity only to guide its sampling during training, then
directly predicts which instances are useful. Thus, purer visual similarity in the retrieved examples
is not necessarily optimal. We see that the most confident neighborhood members predicted by
our method are more often positives. Relying solely on visual similarity, Local can retrieve less
informative instances (e.g., see ?farming?) that share global appearance but do not assist in capturing
the class distribution. The attributes where the Local baseline is most successful, ?sunny? and
?cloudy?, seem to differ from the rest in that (i) they exhibit more consistent global image properties,
and (ii) they have many more positives in the dataset (e.g., 2,416 positives for ?sunny? vs. only 281
for ?farming?). In fact, this scenario is exactly where one would expect traditional visual ranking for
local learning to be adequate. Our method does well not only in such cases, but also where image
nearness is not a good proxy for relevance to classifier construction.
aPascal Results Table 2 shows the results on the aPascal dataset. Again we see a clear and consistent advantage of our approach compared to the conventional Local baselines, with an average
accuracy gain of 10 points across all the Local variants. The addition of metric learning again
provides a slight boost over local, but is inferior to our method, again showing the importance of
learning good neighborhoods. On average, the auto-selected k values for this dataset were 144 with
a standard deviation of 20 for our method; see Supp. file for per attribute statistics.
That said, on this dataset Global has a slight advantage over our method, by 2.7 points on average.
We attribute Global?s success on this dataset to two factors: the images have better spatial alignment
(they are cropped to the boundaries of the object, as opposed to displaying a whole scene as in
SUN), and each attribute exhibits lower visual diversity (they stem from just 20 object classes, as
opposed to 707 scene classes in SUN). See Supp. file. For this data, training with all examples
is most effective. While this dataset yields a negative result for local learning on the whole, it is
nonetheless a positive result for the proposed form of local learning, since we steadily outperform
the standard Local baseline. Furthermore, in principle, our approach could match the accuracy of
the Global method if we let kK = M during training; in that case our method could learn that for
certain queries, it is best to use all examples. This is a flexibility not offered by traditional local
methods. However, due to run-time considerations, at the time of writing we have not yet verified
this in practice.
5
Conclusions
We proposed a new form of lazy local learning that predicts at test time what training data is relevant
for the classification task. Rather than rely solely on feature space proximity, our key insight is to
learn to predict a useful neighborhood. Our results on two challenging image datasets show our
method?s advantages, particularly when categories are multi-modal and/or its similar instances are
difficult to match based on global feature distances alone. In future work, we plan to explore ways
to exploit active learning during training neighborhood generation to reduce its costs. We will also
pursue extensions to allow incremental additions to the labeled data without complete retraining.
Acknowledgements We thank Ashish Kapoor for helpful discussions. This research is supported
in part by NSF IIS-1065390.
8
References
[1] R. Agrawal, A. Gupta, Y. Prabhu, and M. Varma. Multi-label learning with millions of labels: Recommending advertiser bid phrases for web pages. In WWW, 2013.
[2] C. Atkeson, A. Moore, and S. Schaal. Locally weighted learning. AI Review, 1997.
[3] S. Banerjee, A. Dubey, J. Machchhar, and S. Chakrabarti. Efficient and accurate local learning for ranking.
In SIGIR Wkshp, 2009.
[4] A. Bellet, A. Habrard, and M. Sebban. A survey on metric learning for feature vectors and structured
data. CoRR, abs/1306.6709, 2013.
[5] L. Bottou and V. Vapnik. Local learning algorithms. Neural Comp, 1992.
[6] L. Cai and T. Hofmann. Hierarchical document categorization with support vector machines. In CIKM,
2004.
[7] J. Davis, B. Kulis, P. Jain, S. Sra, and I. Dhillon. Information-Theoretic Metric Learning. In ICML, 2007.
[8] C. Domeniconi and D. Gunopulos. Adaptive nearest neighbor classification using support vector machines. In NIPS, 2001.
[9] K. Duh and K. Kirchhoff. Learning to rank with partially-labeled data. In SIGIR, 2008.
[10] A. Farhadi, I. Endres, D. Hoiem, and D. Forsyth. Describing objects by their attributes. In CVPR, 2009.
[11] A. Frome, Y. Singer, and J. Malik. Image retrieval and classification using local distance functions. In
NIPS, 2006.
[12] T. Gao and D. Koller. Discriminative learning of relaxed hierarchy for large-scale visual recognition. In
ICCV, 2011.
[13] X. Geng, T. Liu, T. Qin, A. Arnold, H. Li, and H. Shum. Query dependent ranking using k-nearest
neighbor. In SIGIR, 2008.
[14] B. Gong, K. Grauman, and F. Sha. Connecting the dots with landmarks: Discriminatively learning
domain-invariant features for unsupervised domain adaptation. In ICML, 2013.
[15] T. Hastie and R. Tibshirani. Discriminant adaptive nearest neighbor classification. PAMI, 1996.
[16] D. Hsu, S. Kakade, J. Langford, and T. Zhang. Multi-label prediction via compressed sensing. In NIPS,
2009.
[17] J. Huang, A. Smola, A. Gretton, K. Borgwardt, and B. Scholkopf. Correcting sample selection bias by
unlabeled data. In NIPS, 2007.
[18] J. Jiang and C. Zhai. Instance weighting for domain adapation in NLP. In ACL, 2007.
[19] A. Kapoor and P. Jaina nd R. Viswanathan. Multilabel classification using Bayesian compressed sensing.
In NIPS, 2012.
[20] M. Lapin, M. Hein, and B. Schiele. Learning using privileged information: SVM+ and weighted SVM.
Neural Networks, 53, 2014.
[21] M. Marszalek and C. Schmid. Constructing category hierarchies for visual recognition. In ECCV, 2008.
[22] G. Patterson and J. Hays. Sun attribute database: Discovering, annotating, and recognizing scene attributes. In CVPR, 2012.
[23] B. Settles. Active Learning Literature Survey. Computer Sciences Technical Report 1648, University of
Wisconsin?Madison, 2009.
[24] K. Veropoulos, C. Campbell, and N. Cristianini. Controlling the sensitivity of support vector machines.
In IJCAI, 1999.
[25] P. Vincent and Y. Bengio. K-local hyperplane and convex distance nearest neighbor algorithms. In NIPS,
2001.
[26] K. Weinberger and L. Saul. Distance metric learning for large margin nearest neighbor classification.
JMLR, 2009.
[27] X. Wu and R. Srihari. Incorporating prior knowledge with weighted margin support vector machines. In
KDD, 2004.
[28] L. Yang, R. Jin, R. Sukthankar, and Y. Liu. An efficent algorithm for local distance metric learning. In
AAAI, 2006.
[29] H. Zhang, A. Berg, M. Maire, and J. Malik. SVM-KNN: Discriminative nearest neighbor classification
for visual category recognition. In CVPR, 2006.
9
| 5288 |@word kulis:1 version:1 compression:1 retraining:1 nd:1 liu:2 series:5 contains:2 selecting:1 hoiem:1 shum:1 ours:13 document:1 envision:1 rightmost:2 yni:10 existing:6 xnj:1 recovered:1 past:1 outperforms:1 yet:3 must:4 distant:2 informative:3 kdd:1 shape:1 hofmann:1 treating:1 v:1 alone:3 selected:6 discovering:1 parameterization:1 short:1 core:2 record:1 nearness:3 detecting:1 provides:1 node:1 zhang:2 shiny:1 chakrabarti:1 scholkopf:1 consists:2 compose:3 overhead:1 combine:1 manner:1 pairwise:1 notably:1 tagging:1 indeed:1 roughly:1 nor:1 multi:13 relying:2 automatically:5 equipped:1 farhadi:1 project:1 discover:2 estimating:1 provided:1 what:11 cm:1 pursue:1 transformation:3 ought:2 guarantee:1 tackle:2 exactly:3 grauman:3 classifier:22 k2:1 platt:1 positive:10 before:1 local:106 referenced:1 treat:1 gunopulos:1 despite:1 jiang:1 solely:3 marszalek:1 pami:1 might:2 chose:1 acl:1 suggests:1 challenging:3 limited:1 range:2 averaged:1 practice:3 procedure:4 maire:1 area:1 empirical:1 word:2 confidence:8 pre:1 refers:1 close:2 selection:3 unlabeled:2 wheel:1 influence:3 writing:1 sukthankar:1 optimize:1 conventional:2 map:2 equivalent:1 missing:1 www:1 straightforward:1 independently:2 convex:1 sigir:3 survey:2 identifying:1 recovery:1 correcting:1 insight:1 array:1 varma:1 retrieve:1 ity:1 classic:1 handle:4 variation:4 increment:2 analogous:1 target:12 today:1 construction:1 assuring:1 paralleling:1 exact:1 hierarchy:2 us:1 controlling:1 associate:1 expensive:2 particularly:1 recognition:3 predicts:4 labeled:15 database:1 veropoulos:1 cloud:2 fly:2 region:2 sun:14 highest:2 valuable:4 ran:1 substantial:1 balanced:8 equalize:1 complexity:2 schiele:1 cristianini:1 ultimately:1 multilabel:2 trained:8 predictive:1 purely:1 patterson:1 efficiency:2 easily:1 kirchhoff:1 various:3 represented:1 retrieves:1 surrounding:1 train:12 forced:1 distinct:2 effective:4 jain:1 query:16 labeling:1 neighborhood:67 exhaustive:1 quite:4 encoded:1 valued:2 denser:1 cvpr:3 defers:1 compressed:16 otherwise:3 favor:1 statistic:2 annotating:1 knn:1 breed:1 jointly:5 enlisted:1 peeking:1 online:3 advantage:5 agrawal:1 unprecedented:1 cai:1 propose:3 reconstruction:4 adaptation:2 qin:1 relevant:10 kapoor:2 poorly:2 flexibility:1 validate:1 scalability:1 ijcai:1 cluster:3 requirement:1 double:2 produce:2 generating:1 incremental:1 leave:1 categorization:1 diluted:1 help:2 illustrate:1 develop:1 object:5 gong:1 pose:1 measured:1 nearest:18 recovering:1 c:1 predicted:2 implemented:1 frome:1 differ:1 direction:1 radius:1 attribute:26 engineered:1 settle:1 public:2 material:2 require:2 premise:1 fix:7 f1:2 kristen:1 preliminary:1 strictly:3 extension:2 proximity:7 considered:1 deciding:1 mapping:2 predict:14 claim:1 vary:4 adopt:1 label:33 utexas:2 create:1 gauge:1 weighted:7 always:1 aim:7 rather:8 reaching:1 ck:1 varying:1 eating:5 validated:1 focus:2 schaal:1 properly:1 improvement:3 rank:4 consistently:1 indicates:1 check:1 contrast:4 baseline:16 helpful:1 inference:3 dependent:3 downsample:1 membership:1 cloth:1 entire:2 typically:4 kernelized:1 koller:1 zn1:1 selects:3 pixel:1 issue:2 arg:1 classification:26 among:6 ill:1 development:1 plan:1 spatial:1 having:1 sampling:3 manually:1 biology:1 represents:1 yu:2 look:3 icml:2 unsupervised:1 geng:1 future:3 others:2 report:1 inherent:1 few:3 primarily:1 employ:3 randomly:4 simultaneously:1 densely:1 individual:6 replaced:1 phase:2 kitchen:1 replacement:1 attempt:1 ab:1 interest:1 fd:2 highly:3 intra:3 evaluation:2 adjust:1 alignment:1 yielding:2 amenable:1 accurate:4 xni:15 edge:1 unless:1 tree:1 euclidean:3 desired:1 ynj:1 hein:1 instance:53 column:2 soft:1 earlier:1 disadvantage:1 phrase:2 cost:3 deviation:2 subset:13 entry:3 habrard:1 hundred:1 recognizing:1 successful:2 too:1 itml:1 reported:1 dependency:1 varies:3 proximal:1 endres:1 combined:1 confident:5 person:1 density:1 borgwardt:1 sensitivity:1 enhance:1 together:2 ashish:1 connecting:1 again:3 reflect:1 central:1 aaai:1 opposed:2 choose:1 huang:1 farming:4 strive:1 leading:1 return:1 wing:1 li:1 supp:4 account:3 potential:1 diversity:2 attaining:1 sizeable:1 sec:10 availability:1 forsyth:1 satisfy:1 explicitly:2 ranking:6 depends:1 aron:2 performed:2 portion:1 start:1 sort:1 maintains:1 parallel:1 annotation:1 accuracy:9 efficiently:2 yield:5 identify:3 correspond:2 cni:2 generalize:2 bayesian:2 vincent:1 unifies:1 raw:1 plastic:1 produced:1 marginally:1 lighting:1 comp:1 finer:1 against:1 nonetheless:2 steadily:1 obvious:1 associated:1 xn1:1 sampled:3 gain:3 dataset:14 hsu:1 recall:2 color:1 knowledge:1 dimensionality:3 improves:1 sophisticated:1 actually:1 campbell:1 focusing:1 appears:1 higher:4 day:1 modal:3 formulation:2 though:3 furthermore:6 just:1 stage:4 smola:1 until:1 langford:1 hand:3 web:2 trust:1 banerjee:1 brings:1 quality:1 building:3 concept:1 normalized:1 contain:1 true:1 yn1:1 nonzero:1 furry:1 moore:1 dhillon:1 during:5 inferior:3 davis:1 noted:1 pdf:1 stress:3 complete:2 demonstrate:2 theoretic:1 performs:1 image:34 consideration:2 novel:14 specialized:1 znn:1 sebban:1 million:2 discussed:3 belong:1 slight:3 refer:1 composition:5 ai:1 populated:1 had:2 dot:1 similarity:7 base:1 something:1 closest:1 imbalanced:4 recent:1 posterior:3 showed:1 retrieved:1 irrelevant:1 belongs:1 scenario:2 certain:1 hay:1 binary:8 success:4 formatted:1 devise:1 seen:1 greater:1 relaxed:1 employed:1 determine:2 advertiser:2 ii:5 full:1 multiple:1 gretton:1 infer:1 stem:1 smooth:1 technical:2 match:2 plug:1 offer:2 cross:1 retrieval:1 efficent:1 wkshp:1 permitting:1 equally:1 paired:1 privileged:1 impact:2 prediction:8 ensuring:1 regression:5 basic:1 variant:1 vision:2 metric:21 kernel:1 tailored:2 represent:3 underperforms:1 c1:1 preserved:1 whereas:3 want:2 fine:1 addition:2 cropped:1 unevenly:1 source:2 crucial:1 rest:2 unlike:1 strict:2 file:4 isolate:3 subject:1 validating:1 thing:1 member:2 seem:1 call:1 near:5 leverage:1 ideal:1 yang:1 iii:1 bengio:1 bid:2 xj:1 variety:2 restaurant:1 fit:1 gave:1 associating:1 hastie:1 reduce:2 idea:3 texas:2 whether:2 motivated:1 pca:1 utility:1 assist:1 effort:2 penalty:1 suffer:1 speech:1 speaking:1 matlab:1 adequate:1 ignored:1 useful:13 generally:1 clear:1 dubey:1 locally:3 svms:1 category:17 reduced:1 generate:6 specifies:1 occupy:2 outperform:1 pks:3 problematic:1 nsf:1 estimated:3 cikm:1 per:13 correctly:3 tibshirani:1 broadly:1 group:1 key:2 enormous:1 verified:1 ynn:1 run:6 tailor:1 throughout:1 wu:1 home:1 decision:6 ric:1 scaling:1 capturing:3 bound:1 adapation:1 assemble:1 yielded:1 hiking:2 constraint:2 scene:5 nearby:3 dominated:2 aspect:1 cloudy:1 relatively:2 flush:1 structured:2 according:3 viswanathan:1 truncate:1 precompute:1 sunny:4 combination:1 across:3 bellet:1 kakade:1 s1:1 projecting:1 outlier:1 iccv:1 gathering:1 invariant:1 taken:1 previously:1 turn:1 slack:2 fail:3 describing:2 reverting:1 mind:2 singer:1 end:1 available:2 apply:9 observe:1 hierarchical:4 enforce:1 appearing:1 alternative:1 weinberger:1 altogether:1 original:2 denotes:1 top:4 include:3 cf:1 ensure:1 nlp:1 madison:1 exploit:4 k1:3 especially:1 build:1 malik:2 intend:1 question:2 strategy:2 sha:1 traditional:3 unclear:1 exhibit:3 said:1 distance:12 separate:2 thank:1 duh:1 capacity:4 landmark:1 evaluate:1 considers:1 discriminant:2 trivial:2 extent:1 fresh:1 spanning:1 water:1 prabhu:1 length:1 code:1 pointwise:1 index:1 kk:4 relationship:1 minimizing:2 balance:5 zhai:1 difficult:2 setup:1 hog:2 expense:1 negative:5 implementation:2 proper:1 zni:1 perform:2 upper:1 observation:1 datasets:12 jin:1 variability:1 discovered:1 dog:3 required:2 namely:1 cast:1 pair:4 resist:1 learned:8 boost:1 hour:1 nip:6 nearer:1 address:1 beyond:1 bar:1 proceeds:1 below:1 pattern:1 mismatch:1 xm:1 able:1 sparsity:1 challenge:2 eschew:1 max:1 power:1 apascal:7 ranked:2 rely:2 difficulty:1 predicting:3 indicator:13 improve:4 inversely:1 naive:1 extract:1 auto:3 schmid:1 xq:5 deviate:2 prior:2 faced:1 literature:2 text:1 acknowledgement:1 review:1 relative:3 wisconsin:1 loss:1 expect:3 discriminatively:1 generation:1 versus:1 localized:1 annotator:1 incurred:1 boon:1 gather:2 metal:2 consistent:2 proxy:1 offered:1 displaying:1 principle:1 share:2 austin:2 eccv:1 repeat:1 supported:1 dis:1 offline:7 bias:2 guide:1 allow:1 arnold:1 neighbor:26 fall:1 wide:1 correspondingly:1 saul:1 sparse:2 distributed:2 benefit:1 boundary:2 default:1 author:2 adaptive:2 projected:1 atkeson:1 far:2 cope:1 approximate:1 compact:1 ignore:1 dealing:1 ml:10 global:22 active:5 recommending:2 discriminative:4 xi:3 alternatively:1 latent:1 designates:1 table:7 promising:1 learn:16 sra:1 obtaining:1 warranted:1 bottou:1 necessarily:3 constructing:1 domain:9 main:2 motivation:2 whole:2 arise:1 allowed:1 caters:1 x1:1 fig:2 representative:1 precision:1 fails:2 inferring:1 sub:1 exercise:3 candidate:13 jmlr:1 weighting:2 learns:1 grained:2 minute:1 specific:2 showing:1 sensing:14 appeal:1 list:1 svm:9 explored:1 gupta:1 concern:1 consist:1 incorporating:2 naively:1 vapnik:1 adding:1 corr:1 importance:2 ci:1 texture:1 occurring:1 demand:1 margin:2 suited:2 locality:4 simply:3 likely:6 appearance:1 explore:1 gao:1 visual:13 srihari:1 lazy:7 ordered:1 partially:1 xnn:1 chance:2 succeed:1 goal:7 sorted:1 viewed:1 towards:1 room:1 hard:1 specifically:2 typical:1 reducing:1 determined:1 averaging:1 except:1 uniformly:1 hyperplane:1 total:1 domeniconi:1 indicating:1 select:4 berg:1 support:6 overviewed:1 latter:2 modulated:1 unbalanced:2 relevance:2 lapin:1 crowdsourcing:1 |
4,736 | 5,289 | A Unified Semantic Embedding:
Relating Taxonomies and Attributes
Sung Ju Hwang?
Disney Research
Pittsburgh, PA
[email protected]
Leonid Sigal
Disney Research
Pittsburgh, PA
[email protected]
Abstract
We propose a method that learns a discriminative yet semantic space for object
categorization, where we also embed auxiliary semantic entities such as supercategories and attributes. Contrary to prior work, which only utilized them as side information, we explicitly embed these semantic entities into the same space where
we embed categories, which enables us to represent a category as their linear combination. By exploiting such a unified model for semantics, we enforce each category to be generated as a supercategory + a sparse combination of attributes, with
an additional exclusive regularization to learn discriminative composition. The
proposed reconstructive regularization guides the discriminative learning process
to learn a model with better generalization. This model also generates compact semantic description of each category, which enhances interoperability and enables
humans to analyze what has been learned.
1
Introduction
Object categorization is a challenging problem that requires drawing boundaries between groups of
objects in a seemingly continuous space. Semantic approaches have gained a lot of attention recently
as object categorization became more focused on large-scale and fine-grained recognition tasks and
datasets. Attributes [1, 2, 3, 4] and semantic taxonomies [5, 6, 7, 8] are two popular semantic sources
which impose certain relations between the category models, including a more recently introduced
analogies [9] that induce even higher-order relations between them. While many techniques have
been introduced to utilize each of the individual semantic sources for object categorization, no unified model has been proposed to relate them.
We propose a unified semantic model where we can learn to place categories, supercategories, and
attributes as points (or vectors) in a hypothetical common semantic space, and taxonomies provide
specific topological relationships between these semantic entities. Further, we propose a discriminative learning framework, based on dictionary learning and large margin embedding, to learn each
of these semantic entities to be well separated and pseudo-orthogonal, such that we can use them to
improve visual recognition tasks such as category or attribute recognition.
However, having semantic entities embedded into a common space is not enough to utilize the
vast number of relations that exist between the semantic entities. Thus, we impose a graph-based
regularization between the semantic embeddings, such that each semantic embedding is regularized
by sparse combination of auxiliary semantic embeddings. This additional requirement imposed on
the discriminative learning model would guide the learning such that we obtain not just the optimal
model for class discrimination, but to learn a semantically plausible model which has a potential to
be more robust and human-interpretable; we call this model Unified Semantic Embedding (USE).
?
Now at Ulsan National Institute of Science and Technology in Ulsan, South Korea
1
Figure 1: Concept: We regularize each category
to be represented by its supercategory + a sparse
combination of attributes, where the regularization
parameters are learned. The resulting embedding
model improves the generalization ability by the
specific relations between the semantic entities, and
also is able to compactly represent a novel category
in this manner. For example, given a novel category
tiger, our model can describe it as a striped feline.
The observation we make to draw the relation between the categories and attributes, is that a category
can be represented as the sum of its supercategory + the category-specific modifier, which in many
cases can be represented by a combination of attributes. Further, we want the representation to be
compact. Instead of describing a dalmatian as a domestic animal with a lean body, four legs, a
long tail, and spots, it is more efficient to say it is a spotted dog (Figure 1). It is also more exact
since the higher-level category dog contains all general properties of different dog breeds, including
indescribable dog-specific properties, such as the shape of the head, and its posture.
This exemplifies how a human would describe an object, to efficiently communicate and understand
the concept. Such decomposition of a category into attributes+supercategory can hold for categories
at any level. For example, supercategory feline can be described as a stalking carnivore.
With the addition of this new generative objective, our goal is to learn a discriminative model that
can be compactly represented as a combination of semantic entities, which helps learn a model that is
semantically more reasonable. We want to balance between these two discriminative and generative
objectives when learning a model for each object category. For object categories that have scarce
training examples, we can put more weight on the generative part of the model.
Contributions: Our contributions are threefold: (1) We show a multitask learning formulation for
object categorization that learns a unified semantic space for supercategories and attributes, while
drawing relations between them. (2) We propose a novel sparse-coding based regularization that
enforces the object category representation to be reconstructed as the sum of a supercategory and a
sparse combination of attributes. (3) We show from the experiments that the generative learning with
the sparse-coding based regularization helps improve object categorization performance, especially
in the one or few-shot learning case, by generating semantically plausible predictions.
2
Related Work
Semantic methods for object recognition. For many years, vision researchers have sought to
exploit external semantic knowledge about the object to incorporate semantics into learning of the
model. Taxonomies, or class hierarchies were the first to be explored by vision researchers [5, 6], and
were mostly used to efficiently rule out irrelevant category hypotheses leveraging class hierarchical
structure [8, 10]. Attributes are visual or semantic properties of an object that are common across
multiple categories, mostly regarded as describable mid-level representations. They have been used
to directly infer categories [1, 2], or as additional supervision to aid the main categorization problem
in the multitask learning framework [3]. While many methods have been proposed to leverage either
of these two popular types of semantic knowledge, little work has been done to relate the two, which
our paper aims to address.
Discriminative embedding for object categorization. Since the conventional kernel-based multiclass SVM does not scale due to its memory and computational requirements for today?s large-scale
classification tasks, embedding-based methods have gained recent popularity. Embedding-based
methods perform classification on a low dimensional shared space optimized for class discrimination. Most methods learn two linear projections, for data instances and class labels, to a common
lower-dimensional space optimized by ranking loss. Bengio et al. [10] solves the problem using
stochastic gradient, and also provides a way to learn a tree structure which enables one to efficiently
predict the class label at the test time. Mensink et al. [11] eliminated the need of class embedding by
replacing them with the class mean, which enabled generalization to new classes at near zero cost.
There are also efforts in incorporating semantic information into the learned embedding space.
Weinberger et al. [7] used the taxonomies to preserve the inter-class similarities in the learned space,
2
in terms of distance. Akata et al. [4] used attributes and taxonomy information as labels, replacing
the conventional unit-vector based class representation with more structured labels to improve on
zero-shot performance. One most recent work in this direction is DEVISE [12], which learns embeddings that maximize the ranking loss, as an additional layer on top of the deep network for both
images and labels. However, these models impose structure only on the output space, and structure
on the learned space is not explicitly enforced, which is our goal.
Recently, Hwang et al. [9] introduced one such model, which regularizes the category quadruplets,
that form an analogy, to form a parallelogram. Our goal is similar, but we explore a more general
compositional relationship, which we learn without any manual supervision.
Multitask learning. Our work can be viewed as a multitask learning method, since we relate
each model for different semantic entities by learning both the joint semantic space and enforcing
geometric constraints between them. Perhaps the most similar work is [13], where the parameter
of each model is regularized while fixing the parameter for its parent-level models. We use similar
strategy but instead of enforcing sharing between the models, we simply learn each model to be
close to its approximation obtained using higher-level (more abstract) concepts.
Sparse coding. Our method to approximate each category embedding as a sum of its direct supercategory plus a sparse combination of attributes, is similar to the objective of sparse coding. One
work that is specifically relevant to ours is Mairal et al. [14], where the learning objective is to reduce both the classification and reconstruction error, given class labels. In our model, however, the
dictionary atoms are also discriminatively learned with supervision, and are assembled to be a semantically meaningful combination of a supercategory + attributes, while [14] learns the dictionary
atoms in an unsupervised way.
3
Approach
We now explain our unified semantic embedding model, which learns a discriminative common
low-dimensional space to embed both the images and semantic concepts including object categories,
while enforcing relationships between them using semantic reconstruction.
Suppose that we have a d-dimensional image descriptor and m-dimensional vector describing labels
associated with the instances, including category labels at different semantic granularities and attributes. Our goal then is to embed both images and the labels onto a single unified semantic space,
where the images are associated with their corresponding semantic labels.
To formally state the problem, given a training set D that has N labeled examples, i.e. D =
d
{xi , yi }N
i=1 , where xi ? R denotes image descriptors and yi ? {1, . . . , m} are their labels associated with m unique concepts, we want to embed each xi as zi , and each label yi as uyi in the
de -dimensional space, such that the similarity between zi and uyi , S(zi , uyi ) is maximized.
One way to solve the above problem is to use regression, using S(zi , uyi ) = ?kzi ? uyi k22 . That is,
we estimate the data embedding zi as zi = W xi , and minimize their distances to the correct label
embeddings uyi ? Rm where the dimension for yi is set to 1 and every other dimension is set to 0:
m X
N
X
min
kW xi ? uyi k22 + ?kW k2F .
(1)
W
c=1 i=1
The above ridge regression will project each instance close to its correct embedding. However, it
does not guarantee that the resulting embeddings are well separated. Therefore, most embedding
methods for categorization add in discriminative constraints which ensure that the projected instances have higher similarity to their own category embedding than to others. One way to enforce
this is to use large-margin constraints on distance: kW xi ?uyi k22 +1 ? kW xi ?uc k22 +?ic , yi 6= c
which can be translated into to the following discriminative loss:
X
LC (W , U , xi , yi ) =
[1 + kW xi ? uyi k22 ? kW xi ? uc k22 ]+ , ?c 6= yi ,
(2)
c
where U is the columwise concatenation of each label embedding vector, such that uj denotes jth
column of U . After replacing the generative loss in the ridge regression formula with the discriminative loss, we get the following discriminative learning problem:
N
X
(3)
min
LC (W , U , xi , yi ) + ?kW k2F + ?kU k2F , yi ? {1, . . . , m},
W ,U
i
3
where ? regularizes W and U from shooting to infinity. This is one of the most common objective
used for learning discriminative category embeddings for multi-class classification [10, 7], while
ranking loss-based [15] models have been also explored for LC . Bilinear model on a single variable
W has been also used in Akata et al. [4], which uses structured labels (attributes) as uyi .
3.1
Embedding auxiliary semantic entities.
Now we describe how we embed the supercategories and attributes onto the learned shared space.
Supercategories. While our objective is to better categorize entry level categories, categories in
general can appear at different semantic granularities. For example, a zebra could be both an equus,
and an odd-toed ungulate. To learn the embeddings for the supercategories, we map each data
instance to be closer to its correct supercategory embedding than to its siblings: kW xi ?us k22 +1 ?
kW xi ? uc k22 + ?sc , ?s ? Pyi and c ? Ss where Pyi denotes the set of superclasses at all levels
for class s, and Ss is the set of its siblings. The constraints can be translated into the following loss
term:
X X
[1 + kW xi ? us k22 ? kW xi ? uc k22 ]+ .
(4)
LS (W , U , xi , yi ) =
s?Pyi c?Ss
Attributes. Attributes can be considered normalized basis vectors for the semantic space, whose
combination represents a category. Basically, we want to maximize the correlation between the
projected instance that possess the attribute, and its correct attribute embedding, as follows:
LA (W , U , xi , yi ) = 1 ?
X
(W xi )T yia ua , kua k2 ? 1, yia ? {0, 1}, ?a ? Ayi ,
(5)
a
where Ac is the set of all attributes for class c and ua is an embedding vector for an attribute a.
3.2
Relationship between the categories, supercategories, and attributes
Simply summing up all previously defined loss functions while adding {us } and {ua } as additional columns of U will result in a multi-task formulation that implicitly associate the semantic
entities, through the shared data embedding W . However, we want to further utilize the relationships between the semantic entities, to explicitly impose structural regularization on the semantic
embeddings U . One simple and intuitive relation is that an object class can be represented as the
combination of its parent level category plus a sparse combination of attributes, which translates into
the following constraint:
uc = up + U A ?c , c ? Cp , k?c k0 ?1 , ?c 0, ?c ? {1, . . . , C},
(6)
A
where U is the aggregation of all attribute embeddings {ua }, Cp is the set of children classes for
class p, ?1 is the sparsity parameter, and C is the number of categories. We require ? to be nonnegative, since it makes more sense and more efficient to describe an object with attributes that it
might have, rather than describing it by attributes that it might not have.
We rewrite Eq. 7 into a regularization term as follows, replacing the `0 -norm constraints with `1 norm regularizations for tractable optimization:
C
X
R(U , B) =
kuc ? up ? U A ?c k22 + ?2 k?c + ?o k22 ,
c
(7)
c ? Cp , o ? Pc ? Sc , 0 ?c ?1 , ?c ? {1, . . . , C},
where B is the matrix whose jth column vector ?j is the reconstruction weight for class j, Sc is the
set of all sibling classes for class c, and ?2 is the parameters to enforce exclusivity.
The exclusive regularization term is used to prevent the semantic reconstruction ?c for class c from
fitting to the same attributes fitted by its parents and siblings. This is because attributes common
across parent and child, and between siblings, are less discriminative. This regularization is especially useful for discrimination between siblings, which belong to the same superclass and only
differ by the category-specific modifier. By generating unique semantic decomposition for each
class, we can better discriminate between any two categories using a semantic combination of discriminatively learned auxiliary entities.
4
With the sparsity regularization enforced by ?1 , the simple sum of the two weights will prevent the
two (super)categories from having high weight for a single attribute, which will let each category
embedding to fit to exclusive attribute set. This, in fact, is the exclusive lasso regularizer introduced
in [16], except for the nonnegativity constraint on ?c , which makes the problem easier to solve.
3.3
Unified semantic embeddings for object categorization
After augmenting the categorization objective in Eq. 3 with the superclass and attributes loss and the
sparse-coding based regularization in Eq. 7, we obtain the following multitask learning formulation
that jointly learns all the semantic entities along with the sparse-coding based regularization:
min
N
X
W ,U ,B
kwj k22
LC (W , U , xi , yi ) + ?1 (LS (W , U , xi , yi ) + LA (W , U , xi , yi )) + ?2 R(U , B);
i=1
?
?, kuk k22
(8)
? ?, 0 ?c ?1 ?j ? {1, . . . , d}, ?k ? {1, . . . , m}, ?c ? {1, . . . , C + S},
where S is the number of supercategories, wj is W ?s jth column, and ?1 and ?2 are parameters to
balance between the main and auxiliary tasks, and discriminative and generative objective.
Eq. 8 could be also used for knowledge transfer when learning a model for a novel set of categories,
by replacing U A in R(U , B) with U S , learned on class set S to transfer the knowledge from.
3.4
Numerical optimization
Eq. 8 is not jointly convex in all variables, and has both discriminative and generative terms. This
problem is similar to the problem in [14], where the objective is to learn the dictionary, sparse
coefficients, and classifier parameters together, and can be optimized using a similar alternating
optimization, while each subproblem differs. We first describe how we optimize for each variable.
Learning of W and U . The optimization of both embedding models are similar, except for the
reconstructive regularization on U . and the main bottleneck lies in the minimization of the O(N m)
large-margin losses. Since the losses are non-differentiable, we solve the problems using stochastic
subgradient method. Specifically, we implement the proximal gradient algorithm in [17], handling
the `-2 norm constraints with proximal operators.
Learning B. This is similar to the sparse coding problem, but simpler. We use projected gradient
t+ 1
method, where at each iteration t, we project the solution of the objective ?c 2 for category c to `-1
norm ball and nonnegative orthant, to obtain ?ct that satisfies the constraints.
Alternating optimization. We decompose Eq. 8 to two convex problems: 1) Optimization of the
data embedding W and approximation parameter B (Since the two variable do not have direct link
between them) , and 2) Optimization of the category embedding U . We alternate the process of
optimizing each of the convex problems while fixing the remaining variables, until the convergence
criterion 1 is met, or the maximum number of iteration is reached.
Run-time complexity. Training: Optimization of W and U using proximal stochastic gradient [17], have time complexities of O(de d(k + 1)) and O(de (dk + C)) respectively. Both terms are
dominated by the gradient computation for k(k N ) sampled constraints, that is O(de dk). Outer
loop for alternation converges within 5-10 iterations depending on . Test: Test time complexity is
exactly the same as in LME, which is O(de (C + d)).
4
Results
We validate our method for multiclass categorization performance on two different datasets generated from a public image collection, and also test for knowledge transfer on few-shot learning.
4.1
Datasets
We use Animals with Attributes dataset [1], which consists of 30, 475 images of 50 animal classes,
with 85 class-level attributes 2 . We use the Wordnet hierarchy to generate supercategories. Since
kW t+1 ? W t k2 + kU t+1 ? U t k2 + kB t+1 ? B t k2 <
Attributes are defined on color (black, orange), texture (stripes, spots), parts (longneck, hooves), and other
high-level behavioral properties (slow, hibernate, domestic) of the animals
1
2
5
there is no fixed training/test split, we use {30,30,30} random split for training/validation/test. We
generate the following two datasets using the provided features. 1) AWA-PCA: We compose a 300dimensional feature vectors by performing PCA on each of 6 types of features provided, including
SIFT, rgSIFT, SURF, HoG, LSS, and CQ to have 50 dimensions per each feature type, and concatenating them. 2) AWA-DeCAF: For the second dataset, we use the provided 4096-D DeCAF features
[18] obtained from the layer just before the output layer of a deep convolutional neural network.
4.2
Baselines
We compare our proposed method against multiple existing embedding-based categorization approaches, that either do not use any semantic information, or use semantic information but do not
explicitly embed semantic entities. For non-semantics baselines, we use the following: 1)Ridge
Regression: A linear regression with `-2 norm (Eq. 1). 2) NCM: Nearest mean classifier from [11],
which uses the class mean as category embeddings (uc = x?c ). We use the code provided by the
authors3 . 3) LME: A base large-margin embedding (Eq. 3) solved using alternating optimization.
For implicit semantic baselines, we consider two different methods. 4) LMTE: Our implementation
of the Weinberger et al. [7], which enforces the semantic similarity between class embeddings as
distance constraints [7], where U is regularized to preserve the pairwise class similarities from a
given taxonomy. 5-7) ALE, HLE, AHLE: Our implementation of the attribute label embedding in
Akata et al. [4], which encodes the semantic information by representing each class with structured
labels that indicate the class? association with superclasses and attributes. We implement variants
that use attributes (ALE), leaf level + superclass labels (HLE), and both (AHLE) labels.
For our models, we implement multiple variants to analyze the impact of each semantic entity and
the proposed regularization. 1) LME-MTL-S: The multitask semantic embedding model learned
with supercategories. 2) LME-MTL-A: The multitask embedding model learned with attributes. 3)
USE-No Reg.: The unified semantic embedding model learned using both attributes and supercategories, without semantic regularization. 4) USE-Reg: USE with the sparse coding regularization.
For parameters, the projection dimension de = 50 for all our models. 4 For other parameters, we
find the optimal value by cross-validation on the validation set. We set ?1 = 1 that balances the main
and auxiliary task equally, and search for ?2 for discriminative/generative tradeoff, in the range of
{0.01, 0.1, 0.2 . . . , 1, 10}, and set `-2 norm regularization parameter ? = 1. For sparsity parameter
?1 , we set it to select on average several (3 or 4) attributes per class, and for disjoint parameter ?2 ,
we use 10?1 , without tuning for performance.
No
semantics
Implicit
semantics
Explicit
semantics
USE
Method
Ridge Regression
NCM [11]
LME
LMTE [7]
ALE [4]
HLE [4]
AHLE [4]
LME-MTL-S
LME-MTL-A
USE-No Reg.
USE-Reg.
1
19.31 ? 1.15
18.93 ? 1.71
19.87 ? 1.56
20.76 ? 1.64
15.72 ? 1.14
17.09 ? 1.09
16.65 ? 0.47
20.77 ? 1.41
20.65 ? 0.83
21.07 ? 1.53
21.64 ? 1.02
Flat hit @ k (%)
2
28.34 ? 1.53
29.75 ? 0.92
30.47 ? 1.56
30.71 ? 1.35
25.63 ? 1.44
27.52 ? 1.20
26.55 ? 0.77
32.09 ? 1.67
31.51 ? 0.72
31.59 ? 1.57
32.69 ? 0.83
5
44.17 ? 2.33
47.33 ? 1.60
48.07 ? 1.06
47.76 ? 2.25
43.42 ? 1.67
45.49 ? 0.61
43.05 ? 1.22
50.94 ? 1.21
49.40 ? 0.62
50.11 ? 1.51
52.04 ? 1.02
Hierarchical precision @ k (%)
2
5
28.95 ? 0.54
39.39 ? 0.17
30.81 ? 0.53
43.43 ? 0.53
30.98 ? 0.62
42.63 ? 0.56
31.05 ? 0.71
43.13 ? 0.29
29.26 ? 0.50
43.71 ? 0.34
30.51 ? 0.48
44.76 ? 0.20
29.49 ? 0.89
43.41 ? 0.65
33.71 ? 0.94
45.73 ? 0.71
31.69 ? 0.49
43.47 ? 0.23
33.67 ? 0.55
45.41 ? 0.43
33.37 ? 0.74
47.17 ? 0.91
Table 1: Multiclass classification performance on AWA-PCA dataset (300-D PCA features).
4.3
Multiclass categorization
We first evaluate the suggested multitask learning framework for categorization performance. We
report the average classification performance and standard error over 5 random training/test splits
in Table 1 and 2, using both flat hit@k, which is the accuracy for the top-k predictions made, and
hierarchical precision@k from [12], which is a precision the given label is correct at k, at all levels.
Non-semantic baselines, ridge regression and NCM, were outperformed by our most basic LME
model. For implicit semantic baselines, ALE-variants underperformed even the ridge regression
3
4
http://staff.science.uva.nl/?tmensink/code.php
Except for ALE variants where de =m, the number of semantic entities.
6
No
semantics
Implicit
semantics
Explicit
semantics
USE
Method
Ridge Regression
NCM [11]
LME
LMTE [7]
ALE [4]
HLE [4]
AHLE [4]
LME-MTL-S
LME-MTL-A
USE-No Reg.
USE-Reg.
1
38.39 ? 1.48
43.49 ? 1.23
44.76 ? 1.77
38.92 ? 1.12
36.40 ? 1.03
33.56 ? 1.64
38.01 ? 1.69
45.03 ? 1.32
45.55 ? 1.71
45.93 ? 1.76
46.42 ? 1.33
Flat hit @ k (%)
2
48.61 ? 1.29
57.45 ? 0.91
58.08 ? 2.05
49.97 ? 1.16
50.43 ? 1.92
45.93 ? 2.56
52.07 ? 1.19
57.73 ? 1.75
58.60 ? 1.76
59.37 ? 1.32
59.54 ? 0.73
5
62.12 ? 1.20
75.48 ? 0.58
75.11 ? 1.48
63.35 ? 1.38
70.25 ? 1.97
64.66 ? 1.77
71.53 ? 1.41
74.43 ? 1.26
74.97 ? 1.15
74.97 ? 1.15
76.62 ? 1.45
Hierarchical precision @ k (%)
2
5
38.51 ? 0.61
41.73 ? 0.54
45.25 ? 0.52
50.32 ? 0.47
44.84 ? 0.98
49.87 ? 0.39
38.67 ? 0.46
41.72 ? 0.45
42.52 ? 1.17
52.46 ? 0.37
46.11 ? 2.65
56.79 ? 2.05
44.43 ? 0.66
54.39 ? 0.55
46.05 ? 0.89
51.08 ? 0.36
44.23 ? 0.95
48.52 ? 0.29
47.13 ? 0.62
51.04 ? 0.46
47.39 ? 0.82
53.35 ? 0.30
Table 2: Multiclass classification performance on AWA-DeCAF dataset (4096-D DeCAF features).
baseline with regard to the top-1 classification accuracy 5 , while they improve upon the top-2 recognition accuracy and hierarchical precision. This shows that hard-encoding structures in the label
space do not necessarily improve the discrimination performance, while it helps to learn a more
semantic space. LMTE makes substantial improvement on 300-D features, but not on DeCAF features.
Explicit embedding of semantic entities using our method improved both the top-1 accuracy and
the hierarchical precision, with USE variants achieving the best performance in both. Specifically,
adding superclass embeddings as auxiliary entities improves the hierarchical precision, while using
attributes improves the flat top-k classification accuracy. USE-Reg, especially, made substantial
improvements on flat hit and hierarchical precision @ 5, which shows the proposed regularization?s
effectiveness in learning a semantic space that also discriminates well.
Category
Otter
Skunk
Deer
Moose
Equine
Primate
Ground-truth attributes
Supercategory + learned attributes
An animal that swims, fish, water, new world, small, flippers,
furry, black, brown, tail, . . .
A musteline mammal that is quadrapedal, flippers, furry,
ocean
An animal that is smelly, black, stripes, white, tail, furry,
ground, quadrapedal, new world, walks, . . .
A musteline mammal that has stripes
An animal that is brown, fast, horns, grazer, forest,
quadrapedal, vegetation, timid, hooves, walks, . . .
A deer that has spots, nestspot, longneck, yellow, hooves
An animal that has horns, brown, big, quadrapedal, new
world, vegetation, grazer, hooves, strong, ground,. . .
A deer that is arctic, stripes, black
N/A
N/A
An odd-toed ungulate, that is lean and active
An animal, that has hands and bipedal
Table 3: Semantic description generated using ground truth attributes labels and learned semantic decomposition of each categorys. For ground truth labels, we show top-10 ranked by their human-ranked relevance. For
our method, we rank the attributes by their learned weights. Incorrect attributes are colored in red.
4.3.1
Qualitative analysis
Besides learning a space that is both discriminative and generalizes well, our method?s main advantage, over existing methods, is its ability to generate compact, semantic descriptions for each
category it has learned. This is a great caveat, since in most models, including the state-of-the
art deep convolutional networks, humans cannot understand what has been learned; by generating
human-understandable explanation, our model can communicate with the human, allowing understanding of the rationale behind the categorization decisions, and to possibly allow feedback for
correction.
To show the effectiveness of using supercategory+attributes in the description, we report the learned
reconstruction for our model, compared against the description generated by its ground-truth attributes in Table 3. The results show that our method generates compact description of each category, focusing on its discriminative attributes. For example, our method select attributes such as
flippers for otter, and stripes for skunk, instead of attributes common and nondescriminative such as
tail. Note that some attributes that are ranked less relevant by humans were selected for their discriminativity, e.g., yellow for dear and black for moose, both of which human annotators regarded
5
We did extensive parameter search for the ALE variants.
7
placental
ungulate
horns
hooves
longneck
carnivore
pads
stalker
paws
aquatic
plankton
ocean
swims
rodent
plankton
g.ape
feline
plankton
yellow
orange
canine
longneck
whale
longleg
plains
fields
primate
plankton
hands
bipedal
even?toed
hunter
dog
strainteeth
toughskin
longneck
bear
musteln procyonid
pinnpd
dolphin
plankton strainteethlongneck
walks
planktonbaleen
longneck longneck toughskin
ground
longneck
strainteethplankton strainteeth
stalker
sheperd
cat
strainteeth
toughskin
hairless
big cat
toughskin
longneck
big
ruminant
plankton
meatteeth
hunter
bovid
hooves
horns
grazer
odd?toed ungulate
plankton
meatteeth
hunter
domestic
plankton
longneck
toughskin
equine
hunter
meatteeth
small
deer
muscle
bovine
small
meatteeth
s
es
irl
ha
s in
s
rn ou hsk ds
ho ulb ug pa e
b to ic iv
:
t ct
w h
co bus low es y a
: : s dom ilit s
g e
ox lo
h
a
v
ffa fis an oo
ck
e
bu p: le n h
s
gn
ee e: lea
im
on
sh lop ts e e l
sw
h
te po tiv ng
is
an : s ac ora ite h f
er : in w wh et
de se ello els inte ush
oo y nn tra b s
m e: tu : s eck lain
f
f
k
h
ra a us n p
et
gi we am long wn
te
o
g: ot s br s at
pi op ripe in usk me ws
k t
pp st s e er la
hi a: ugh ngl talk ds c
br to ju s a
ze e: s: tive tic p ts
rs ro ac es po
ho oce in om w s ly
in e: d llo el
rh es at: ye sm
am c ng us
Si ian tro lbo
rs : s bu h
Pe cat ert fis
er
b es ts e eth alk
bo : d spo ang atte y st y
r e ra ra
n :
o
lio ard up ic m g g tic g ss
op ro st le es rle
le r: g me long om hai
d g
e do :
tig e: erd eak gle
lli h w lon
co hep a: ts
u o
.s h p e
G ua : s scl h t w
h n u fis as lo
f
s
hi ia
C at tic m ge ts on
lm rc an po kt
da : a s or es s lan
p l
f
ol rn ip s ta
w ho str sk as
u
x: n: : t co r
fo oo da up aze
s
cc an gro gr e tic
ra t p ks ck oov arc s
n
e h ly
ai
an us n
l
gi r: t ong low me unt
te : l s s s o
ot sel orn ean ig m
ea h oc b
s
w k: r: ave
im
un ea c
sw
sk r b ar: l
c
la be tai ive esti n
a
po ly : ct
m
le
k
a
z
iz ac in do at th
gr pb le: ive me tee
s
m a ct th ck
nd
hu wh : a ttee r bu ks ha
l
n
ue hi a e a cts
bl lp me raz ll w se
do e: s g ma in
c. hal usk s s skin
w : t in h
r
k. us nta oug ous aze
r u t lb r c
g i
al o :
w : m ee l bu ns est
al nz da or m
se pa ipe : h do
im b ey up eth l
ch lla: nk gro tte da e
ri mo st ea ipe tiv
go er ore s m r b nac
id : f ld lke s i
sp se fie sta lk s
a
ou rce ds w ipe
m fie pa hes str
t: l: tc ds ss
ra irre pa pa irle
s ha
u r:
h
sq ste wim lys h bus
m : s f us s
ha er ton n b usk
av nk to n t
be pla ank kto els
t: pl an n
ba it: pl tun
:
bb nt on
ra ha nkt
ep la
el e: p
ol
m
Figure 2: Learned discriminative attributes association on the AWA-PCA dataset. Incorrect attributes are
colored in red.
AWA?DeCAF
70
40
60
35
30
No transfer
AHLE
USE
USE?Reg.
25
20
15
0
2
4
6
8
Number of training examples
10
Accuracy (%)
Accuracy (%)
AWA?PCA
45
Class
Humpback
whale
Leopard
50
40
No transfer
AHLE
USE
USE?Reg.
30
20
0
2
4
6
8
Number of training examples
10
Learned decomposition
A baleen whale, with plankton, flippers, blue, skimmer, arctic
A big cat that is orange, claws, black
An even-toed ungulate, that is gray,
Hippopotamus
bulbous, water, smelly, hands
A primate, that is mountains, strong,
Chimpanzee
stalker, black
A domestic cat, that is arctic, nestspot,
Persian Cat
fish, bush
Figure 3: Few-shot experiment result on the AWA dataset, and generated semantic decompositions.
as brown. Further, our method selects discriminative attributes for each supercategory, while there
is no provided attribute label for supercategories.
Figure 2 shows the discriminative attributes disjointly selected at each node on the class hierarchy.
We observe that coarser grained categories fit to attributes that are common throughout all its children (e.g. pads, stalker and paws for carnivore), while the finer grained categories fit to attributes
that help for finer-grained distinctions (e.g. orange for tiger, spots for leopard, and desert for lion).
4.4 One-shot/Few-shot learning
Our method is expected to be especially useful for few-shot learning, by generating a richer description than existing methods, that approximate the new input category using either trained categories
or attributes. For this experiment, we divide the 50 categories into predefined 40/10 training/test
split, and compare with the knowledge transfer using AHLE. We assume that no attribute label is
provided for test set. For AHLE, and USE, we regularize the learning of W with W S learned
on training class set S by adding ?2 kW ? W S k22 , to LME (Eq. 3). For USE-Reg we use the
reconstructive regularizer to regularize the model to generate semantic decomposition using U S .
Figure 3 shows the result, and the learned semantic decomposition of each novel category. While all
methods make improvements over the no-transfer baseline, USE-Reg achieves the most improvement, improving two-shot result on AWA-DeCAF from 38.93% to 49.87%, where USE comes in
second with 44.87%. Most learned reconstructions look reasonable, and fit to discriminative traits
that help to discriminate between the test classes, which in this case are colors; orange for leopard,
gray for hippopotamus, blue for humpback whale, and arctic (white) for Persian cat.
5
Conclusion
We propose a unified semantic space model that learns a discriminative space for object categorization, with the help of auxiliary semantic entities such as supercategories and attributes. The auxiliary
entities aid object categorization both indirectly, by sharing a common data embedding, and directly,
by a sparse-coding based regularizer that enforces the category to be generated by its supercategory
+ a sparse combination of attributes. Our USE model improves both the flat-hit accuracy and hierarchical precision on the AWA dataset, and also generates semantically meaningful decomposition
of categories, that provides human-interpretable rationale.
8
References
[1] Christoph Lampert, Hannes Nickisch, and Stefan Harmeling. Learning to Detect Unseen Object Classes by Between-Class Attribute Transfer. In IEEE Conference on Computer Vision
and Pattern Recognition (CVPR), 2009.
[2] Ali Farhadi, Ian Endres, Derek Hoiem, and David Forsyth. Describing Objects by their Attributes. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2009.
[3] Sung Ju Hwang, Fei Sha, and Kristen Grauman. Sharing features between objects and their
attributes. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages
1761?1768, 2011.
[4] Zeynep Akata, Florent Perronnin, Zaid Harchaoui, and Cordelia Schmid. Label-Embedding for
Attribute-Based Classification. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 819?826, 2013.
[5] Marcin Marszalek and Cordelia Schmid. Constructing category hierarchies for visual recognition. In European Conference on Computer Vision (ECCV), 2008.
[6] Gregory Griffin and Pietro Perona. Learning and using taxonomies for fast visual categorization. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1?8,
2008.
[7] Kilian Q. Weinberger and Olivier Chapelle. Large margin taxonomy embedding for document
categorization. In Neural Information Processing Systems (NIPS), pages 1737?1744, 2009.
[8] Tianshi Gao and Daphne Koller. Discriminative learning of relaxed hierarchy for large-scale
visual recognition. International Conference on Computer Vision (ICCV), pages 2072?2079,
2011.
[9] Sung Ju Hwang, Kristen Grauman, and Fei Sha. Analogy-preserving semantic embedding for
visual object categorization. In International Conference on Machine Learning (ICML), pages
639?647, 2013.
[10] Samy Bengio, Jason Weston, and David Grangier. Label Embedding Trees for Large MultiClass Task. In Neural Information Processing Systems (NIPS), 2010.
[11] Thomas Mensink, Jakov Verbeek, Florent Perronnin, and Gabriela Csurka. Distance-based
image classification: Generalizing to new classes at near zero cost. IEEE Transactions on
Pattern Analysis and Machine Intelligence (TPAMI), 35(11), 2013.
[12] Andrea Frome, Greg Corrado, Jon Shlens, Samy Bengio, Jeffrey Dean, Marc?Aurelio Ranzato,
and Tomas Mikolov. Devise: A deep visual-semantic embedding model. In Neural Information
Processing Systems (NIPS), 2013.
[13] Alon Zweig and Daphna Weinshall. Hierarchical regularization cascade for joint learning. In
International Conference on Machine Learning (ICML), volume 28, pages 37?45, 2013.
[14] Julien Mairal, Francis Bach, Jean Ponce, Guillermo Sapiro, and Andrew Zisserman. Supervised dictionary learning. In Neural Information Processing Systems (NIPS), pages 1033?
1040, 2008.
[15] Jason Weston, Samy Bengio, and Nicolas Usunier. Wsabie: Scaling up to large vocabulary
image annotation. In International Joint Conferences on Artificial Intelligence (IJCAI), 2011.
[16] Yang Zhou, Rong Jin, and Steven C. H. Hoi. Exclusive lasso for multi-task feature selection.
Journal of Machine Learning Research, 9:988?995, 2010.
[17] John Duchi and Yoram Singer. Efficient online and batch learning using forward backward
splitting. Journal of Machine Learning Research, 10, 2009.
[18] Jeff Donahue, Yangqing Jia, Oriol Vinyals, Judy Hoffman, Ning Zhang, Eric Tzeng, and Trevor
Darrell. DeCAF: A deep convolutional activation feature for generic visual recognition. In
International Conference on Machine Learning (ICML), 2014.
9
| 5289 |@word multitask:8 norm:6 nd:1 hu:1 r:2 llo:1 decomposition:8 mammal:2 shot:8 ld:1 contains:1 paw:2 hoiem:1 ours:1 document:1 existing:3 com:2 nt:1 si:1 yet:1 activation:1 john:1 numerical:1 shape:1 enables:3 musteline:2 flipper:4 zaid:1 interpretable:2 discrimination:4 generative:8 leaf:1 selected:2 intelligence:2 dear:1 colored:2 caveat:1 provides:2 node:1 simpler:1 lbo:1 daphne:1 zhang:1 rc:1 along:1 direct:2 uyi:10 shooting:1 consists:1 incorrect:2 qualitative:1 fitting:1 compose:1 behavioral:1 manner:1 pairwise:1 inter:1 expected:1 ra:6 andrea:1 os:1 multi:3 ol:2 eck:1 little:1 str:2 farhadi:1 ua:5 domestic:4 project:2 provided:6 what:2 tic:4 mountain:1 weinshall:1 disneyresearch:2 unified:11 sung:3 guarantee:1 pseudo:1 esti:1 every:1 hypothetical:1 sapiro:1 lys:1 exactly:1 ro:2 rm:1 k2:4 classifier:2 hit:5 unit:1 ly:3 grauman:2 appear:1 before:1 bilinear:1 encoding:1 id:1 marszalek:1 might:2 plus:2 black:7 meatteeth:4 k:2 nz:1 challenging:1 christoph:1 co:3 lop:1 range:1 unique:2 horn:4 enforces:3 pla:1 harmeling:1 implement:3 differs:1 spot:4 sq:1 lla:1 eth:2 cascade:1 projection:2 induce:1 supercategory:13 get:1 onto:2 close:2 equine:2 operator:1 hle:4 put:1 cannot:1 selection:1 optimize:1 conventional:2 imposed:1 map:1 disjointly:1 dean:1 go:1 attention:1 l:3 convex:3 focused:1 pyi:3 feline:3 tomas:1 splitting:1 rule:1 regarded:2 shlens:1 regularize:3 enabled:1 embedding:39 ert:1 hierarchy:5 today:1 suppose:1 exact:1 olivier:1 us:2 samy:3 hypothesis:1 pa:7 associate:1 recognition:13 ze:1 utilized:1 stripe:5 rgsift:1 lean:2 coarser:1 labeled:1 exclusivity:1 ep:1 steven:1 subproblem:1 solved:1 wj:1 kilian:1 ranzato:1 ilit:1 substantial:2 discriminates:1 ungulate:5 complexity:3 ong:1 dom:1 trained:1 unt:1 rewrite:1 ali:1 upon:1 eric:1 basis:1 compactly:2 translated:2 po:4 joint:3 k0:1 represented:5 cat:7 talk:1 regularizer:3 separated:2 fast:2 describe:5 reconstructive:3 artificial:1 sc:3 deer:4 whose:2 richer:1 jean:1 plausible:2 cvpr:5 solve:3 say:1 drawing:2 s:5 ive:2 ability:2 gro:2 gi:2 unseen:1 breed:1 jointly:2 ip:1 seemingly:1 lke:1 online:1 advantage:1 differentiable:1 tpami:1 propose:5 reconstruction:6 ste:1 tu:1 relevant:2 loop:1 description:7 intuitive:1 validate:1 hibernate:1 exploiting:1 parent:4 convergence:1 requirement:2 dolphin:1 ijcai:1 darrell:1 generating:4 categorization:21 converges:1 object:25 help:6 depending:1 alon:1 andrew:1 augmenting:1 fixing:2 ac:4 hoof:6 op:2 nearest:1 odd:3 eq:9 strong:2 solves:1 auxiliary:9 ard:1 frome:1 indicate:1 come:1 met:1 differ:1 direction:1 ning:1 correct:5 attribute:71 stochastic:3 kb:1 human:10 public:1 orn:1 hoi:1 require:1 generalization:3 decompose:1 kristen:2 im:3 leopard:3 pl:2 correction:1 hold:1 gabriela:1 rong:1 considered:1 ic:3 ground:7 great:1 predict:1 mo:1 lm:1 dictionary:5 sought:1 achieves:1 interoperability:1 outperformed:1 label:27 wim:1 hoffman:1 minimization:1 stefan:1 aim:1 super:1 rather:1 ck:3 zhou:1 hippopotamus:2 sel:1 exemplifies:1 tte:1 lon:1 ponce:1 improvement:4 rank:1 ave:1 baseline:7 detect:1 sense:1 hsk:1 am:2 el:4 humpback:2 nn:1 perronnin:2 fis:3 pad:2 w:1 relation:7 perona:1 koller:1 marcin:1 selects:1 semantics:9 bovine:1 classification:11 oce:1 animal:9 art:1 orange:5 uc:6 tzeng:1 field:1 having:2 ng:2 eliminated:1 atom:2 whale:4 kw:13 represents:1 arctic:4 unsupervised:1 k2f:3 look:1 icml:3 jon:1 others:1 report:2 few:5 sta:1 preserve:2 national:1 individual:1 jeffrey:1 bipedal:2 sh:1 nl:1 pc:1 behind:1 baleen:1 predefined:1 kt:1 closer:1 korea:1 orthogonal:1 tree:2 iv:1 divide:1 walk:3 fitted:1 instance:6 column:4 gn:1 hep:1 ar:1 cost:2 entry:1 gr:2 gregory:1 proximal:3 nickisch:1 endres:1 ju:4 st:4 international:5 bu:4 together:1 possibly:1 external:1 yia:2 potential:1 de:8 coding:9 coefficient:1 tra:1 forsyth:1 explicitly:4 ranking:3 csurka:1 tiv:2 lot:1 ayi:1 placental:1 analyze:2 jason:2 reached:1 red:2 aggregation:1 francis:1 annotation:1 irre:1 jia:1 contribution:2 minimize:1 om:2 php:1 greg:1 became:1 descriptor:2 convolutional:3 efficiently:3 maximized:1 accuracy:8 yellow:3 hunter:4 basically:1 usk:3 lli:1 researcher:2 cc:1 finer:2 tianshi:1 explain:1 fo:1 manual:1 sharing:3 trevor:1 against:2 pp:1 derek:1 associated:3 sampled:1 dataset:7 popular:2 wh:2 knowledge:6 color:2 improves:4 ou:2 akata:4 ea:3 focusing:1 higher:4 ta:1 oov:1 mtl:6 supervised:1 zisserman:1 improved:1 erd:1 hannes:1 formulation:3 done:1 mensink:2 ox:1 just:2 implicit:4 correlation:1 until:1 hand:3 d:4 replacing:5 irl:1 perhaps:1 gray:2 hwang:5 aquatic:1 hal:1 irle:1 nac:1 k22:15 concept:5 normalized:1 brown:4 ye:1 regularization:21 alternating:3 furry:3 semantic:72 white:2 ll:1 quadruplet:1 ue:1 oc:1 criterion:1 ridge:7 duchi:1 cp:3 spo:1 image:10 tro:1 novel:5 recently:3 common:10 ug:1 volume:1 tail:4 belong:1 association:2 relating:1 vegetation:2 he:1 trait:1 composition:1 zebra:1 longneck:10 ai:1 tuning:1 grangier:1 chapelle:1 supervision:3 similarity:5 add:1 base:1 own:1 recent:2 optimizing:1 irrelevant:1 certain:1 tee:1 alternation:1 yi:14 devise:2 muscle:1 ous:1 preserving:1 additional:5 relaxed:1 impose:4 staff:1 quadrapedal:4 ey:1 maximize:2 corrado:1 ale:7 multiple:3 persian:2 harchaoui:1 infer:1 fie:2 cross:1 long:3 zweig:1 bach:1 equally:1 spotted:1 impact:1 prediction:2 variant:6 regression:9 basic:1 verbeek:1 vision:9 iteration:3 represent:2 kernel:1 lea:1 addition:1 want:5 fine:1 ore:1 ank:1 source:2 stalker:4 ot:2 posse:1 south:1 ape:1 chimpanzee:1 contrary:1 leveraging:1 effectiveness:2 call:1 structural:1 near:2 yang:1 leverage:1 granularity:2 bengio:4 enough:1 embeddings:13 split:4 ee:2 wn:1 fit:4 zi:6 nta:1 lasso:2 florent:2 reduce:1 raz:1 multiclass:6 sibling:6 translates:1 tradeoff:1 br:2 bottleneck:1 pca:6 alk:1 effort:1 swim:2 ipe:3 strainteeth:3 compositional:1 deep:5 useful:2 se:4 awa:10 gle:1 mid:1 ang:1 category:55 generate:4 http:1 exist:1 fish:2 lsigal:1 disjoint:1 popularity:1 per:2 blue:2 threefold:1 iz:1 group:1 four:1 lan:1 pb:1 achieving:1 yangqing:1 prevent:2 kuk:1 utilize:3 backward:1 vast:1 graph:1 subgradient:1 pietro:1 lain:1 sum:4 year:1 enforced:2 run:1 communicate:2 place:1 throughout:1 reasonable:2 parallelogram:1 draw:1 decision:1 griffin:1 scaling:1 authors3:1 modifier:2 layer:3 ct:5 hi:3 topological:1 nonnegative:2 constraint:11 infinity:1 fei:2 striped:1 ri:1 flat:6 encodes:1 dominated:1 generates:3 oo:3 min:3 performing:1 claw:1 mikolov:1 cordelia:2 structured:3 alternate:1 combination:14 supercategories:13 ball:1 across:2 wsabie:1 ush:1 lp:1 describable:1 primate:3 leg:1 iccv:1 previously:1 bus:2 describing:4 tai:1 singer:1 ge:1 tractable:1 scl:1 usunier:1 generalizes:1 observe:1 hierarchical:10 enforce:3 indirectly:1 generic:1 ocean:2 batch:1 weinberger:3 ho:3 thomas:1 top:7 denotes:3 ensure:1 remaining:1 sw:2 rce:1 daphna:1 procyonid:1 exploit:1 yoram:1 tig:1 especially:4 uj:1 bl:1 objective:10 skin:1 posture:1 strategy:1 ulb:1 exclusive:5 sha:2 hai:1 enhances:1 gradient:5 distance:5 link:1 entity:21 concatenation:1 outer:1 lio:1 timid:1 me:5 zeynep:1 water:2 enforcing:3 code:2 besides:1 relationship:5 cq:1 balance:3 inte:1 mostly:2 hairless:1 taxonomy:9 relate:3 hog:1 ba:1 implementation:2 understandable:1 perform:1 allowing:1 plankton:9 canine:1 observation:1 av:1 datasets:4 sm:1 arc:1 jin:1 orthant:1 t:5 regularizes:2 head:1 disney:2 rn:2 lb:1 introduced:4 tive:1 dog:5 david:2 extensive:1 optimized:3 dalmatian:1 learned:23 distinction:1 nip:4 assembled:1 address:1 able:1 suggested:1 lion:1 pattern:6 sparsity:3 toughskin:5 tun:1 including:6 memory:1 explanation:1 ia:1 ranked:3 regularized:3 nkt:1 scarce:1 representing:1 improve:5 technology:1 julien:1 lk:1 schmid:2 prior:1 geometric:1 understanding:1 embedded:1 loss:11 discriminatively:2 moose:2 rationale:2 bear:1 analogy:3 annotator:1 validation:3 bulbous:1 ripe:1 sigal:1 pi:1 lo:2 eccv:1 guillermo:1 carnivore:3 jth:3 side:1 guide:2 understand:2 allow:1 institute:1 smelly:2 sparse:17 regard:1 boundary:1 dimension:4 feedback:1 world:3 plain:1 vocabulary:1 forward:1 collection:1 made:2 projected:3 ig:1 kzi:1 eak:1 bb:1 reconstructed:1 approximate:2 compact:4 transaction:1 implicitly:1 skunk:2 otter:2 active:1 mairal:2 summing:1 pittsburgh:2 discriminative:26 xi:21 continuous:1 search:2 nestspot:2 un:1 sk:2 table:5 underperformed:1 learn:14 ku:2 robust:1 nicolas:1 transfer:8 ean:1 forest:1 improving:1 necessarily:1 european:1 constructing:1 marc:1 da:4 uva:1 surf:1 did:1 main:5 sp:1 rh:1 big:4 aurelio:1 lampert:1 child:3 body:1 slow:1 aid:2 lc:4 precision:9 n:1 judy:1 nonnegativity:1 explicit:3 concatenating:1 lie:1 pe:1 learns:7 grained:4 ian:2 donahue:1 formula:1 embed:8 ffa:1 specific:5 sift:1 lme:12 grazer:3 er:5 explored:2 dk:2 svm:1 ton:1 incorporating:1 adding:3 gained:2 decaf:8 texture:1 te:3 margin:5 nk:2 easier:1 rodent:1 tc:1 generalizing:1 simply:2 explore:1 gao:1 visual:8 ncm:4 vinyals:1 bo:1 kwj:1 ch:1 truth:4 satisfies:1 ma:1 weston:2 superclass:6 goal:4 viewed:1 ite:1 ello:1 jeff:1 shared:3 leonid:1 tiger:2 hard:1 rle:1 specifically:3 except:3 semantically:5 kuc:1 wordnet:1 toed:5 discriminate:2 e:7 la:5 est:1 meaningful:2 desert:1 formally:1 select:2 categorize:1 relevance:1 bush:1 oriol:1 incorporate:1 evaluate:1 reg:11 handling:1 |
4,737 | 529 | Neural Computing with Small Weights
Kai-Yeung Siu
Dept. of Electrical & Computer Engineering
University of California, Irvine
Irvine, CA 92717
J ehoshua Bruck
IBM Research Division
Almaden Research Center
San Jose, CA 95120-6099
Abstract
An important issue in neural computation is the dynamic range of weights
in the neural networks. Many experimental results on learning indicate
that the weights in the networks can grow prohibitively large with the
size of the inputs. Here we address this issue by studying the tradeoffs
between the depth and the size of weights in polynomial-size networks
of linear threshold elements (LTEs). We show that there is an efficient
way of simulating a network of LTEs with large weights by a network
of LTEs with small weights. In particular, we prove that every depth-d,
polynomial-size network of LTEs with exponentially large integer weights
can be simulated by a depth-(2d + 1), polynomial-size network of LTEs
with polynomially bounded integer weights. To prove these results, we
use tools from harmonic analysis of Boolean functions. Our technique is
quite general, it provides insights to some other problems. For example,
we are able to improve the best known results on the depth of a network
of linear threshold elements that computes the COM PARISO N, SUM
and PRO DU CT of two n-bits numbers, and the MAX 1M U M and the
SORTING of n n-bit numbers.
1
Introduction
The motivation for this work comes from the area of neural networks, where a
linear threshold element is the basic processing element. Many experimental results
on learning have indicated that the magnitudes of the coefficients in the threshold
elements grow very fast with the size of the inputs and therefore limit the practical
use of the network. One natural question to ask is the following: How limited
944
Neural Computing with Small Weights
is the computational power of the network if we restrict ourselves to threshold
elements with only "small" growth in the coefficients? We answer this question by
showing that we can trade-off an exponential growth with a polynomial growth in
the magnitudes of coefficients by increasing the depth of the network by a factor of
almost two and a polynomial growth in the size.
Linear Threshold Functions: A linear threshold function f(X) is a Boolean
function such that
f(X)
= sgn(F(X? = {_II
where
if F(X) > 0
if F(X) < 0
n
F(X) =
2:=
Wi . Xi
+ Wo
i=l
Throughout this paper, a Boolean function will be defined as f : {I, _I}n --+
{I, -I}; namely, 0 and 1 are represented by 1 and -1, respectively. Without loss
of generality, we can assume F(X):/; 0 for all X E {I,-I}n. The coefficients Wi
are commonly referred to as the weights of the threshold function. We denote the
class of all linear threshold functions by LT1 ?
---
LT1 functions: In this paper, we shall study a subclass of LT1 which we denote
by
is characterized by
1 . Each function f(X) = sgn(L:~=l Wi' Xi + wo) in
IT
IT1
the property that the weights Wi are integers and bounded by a polynomial in n,
i.e. IWil ~ n C for some constant c > O.
Threshold Circuits: A threshold circuit [5, 10] is a Boolean network in which
every gate computes an LT1 function. The size of a threshold circuit is the number
of LT1 elements in the circuit. Let LTk denote the class of threshold circuits of
depth k with the size bounded by a polynomial in the number of inputs. We define
LTk similarly except that we allow each gate in LTk to compute an LTI function.
-
---
---
Although the definition of (LTd linear threshold function allows the weights to be
real numbers, it is known [12] that we can replace each of the real weights by integers
of O( n log n) bits, where n is the number of input Boolean variables. So in the rest
of the paper, we shall assume without loss of generality that all weights are integers.
However, this still allows the magnitudes of the weights to increase exponentially
fast with the size of the inputs. It is natural to ask if this is necessary. In other
words, is there a linear threshold function that must require exponentially large
weights? Since there are 2n(n~) linear threshold functions in n variables [8, 14, 15],
there exists at least one which requires O(n 2 ) bits to specify the weights. By the
pigeonhole principle, at least one weight of such a function must need O(n) bits,
and thus is exponentially large in magnitude. i.e.
-
L TI ~ LT1
The above result was proved in [9] using a different method by explicitly constructing
an LT1 function and proving that it is not in LT1 . In the following section, we
shall show that the COMPARISON function (to be defined later) also requires
exponentially large weights. We will refer to this function later on in the proof of
-
945
946
Siu and Bruck
our main results. Main Results: The fact that we can simulate a linear threshold
function with exponentially large weights in a 'constant' number oflayers of elements
with 'small' weights follows from the results in [3] and [11]. Their results showed
that the sum of n n-bit numbers is computable in a constant number of layers of
'counting' gates, which in turn can be simulated by a constant number of layers of
threshold elements with 'small' weights. However, it was not explicitly stated how
many layers are needed in each step of their construction and direct application of
their results would yield a constant such as 13. In this paper, we shall reduce the
constant to 3 by giving a more 'depth'-efficient algorithm and by using harmonic
analysis of Boolean functions [1,2,6]. We then generalize this result to higher depth
circuits and show how to simulate a threshold circuit of depth-d and exponentially
large weights in a depth-(2d + 1) threshold circuit of 'small' weights, i.e. LTd ~
fr2d+l.
As another application of harmonic analysis, we also show that the
COM P ARISON and ADDITION of two n-bit numbers is computable with only
two layers of elements with 'small' weights, while it was only known to be computable in 3 layers [5]. We also indicate how our 'depth'-efficient algorithm can be
applied to show that the product of two n-bit numbers can be computed in LT4 .
In addition, we show that the MAXIMUM and SORTING ofn n-bit numbers
can be computed in fr3 and LT4 , respectively.
--
2
Main Results
Definition: Let X = (Xl, ... , Xn), Y = (YI, ... , Yn) E {I, _l}n. We consider X
and Y as two n-bit numbers representing E?=l Xi? 2' and E?=l Yi . 2i , respectively.
The COMPARISON function is defined as
C(X, Y) = 1 iff X
In other words,
~ Y
n
C(X, Y) =
sgn{L:: 2i(Xi - yd + I}
i=l
Lemma 1
COMPARISON
-
tt LTI
On the other hand, using harmonic analysis [2], we can show the following:
Lemma 2
COMPARISON E
m
Spectral representation of Boolean functions: Recently, harmonic analysis
has been found to be a powerful tool in studying the computational complexity of
Boolean functions [1, 2, 7]. The idea is that every Boolean function f : {I, _1}n -+
{I, -I} can be represented as a polynomial over the field of rational numbers as
follows:
f(X)
aa xa
= L
aE{O,l}n
Neural Computing with Small Weights
h
were
al
X a = x al
1 x2
an ?
.?. Xn
Such representation is unique and the coefficients of the polynomial,
{a, l}n}, are called the spectral coefficients of f.
{aal Q E
We shall define the Ll spectral norm of f to be
IIfll
=
~ laal?
ae{O,I}n
The proof of Lemma 2 is based on the spectral techniques developed in [2]. Using
probabilistic arguments, it was proved in [2] that if a Boolean function has Ll spec.tral norm which is polynomially bounded, then the function is computable in LT2 ?
We observe (together with Noga Alon) that the techniques in [2] can be generalized
to show that any Boolean function with polynomially bounded Ll spectral norm
can even be closely approximated by a sparse polynomial. This observation is crucial
when we extend our result from a single element to networks of elements with large
weights.
Lemma 3 Let f(X) : {I, _l}n --+ {I, -I} such that IIfll
for any k > 0, there exists a sparse polynomial
F(X)
= N1 2:'.:: wa Xa
~
n C for some c. Then
such that
aes
IF(X) - f(X)1 ~ n- k ,
where Wa and N are integers, S c {O, l}n, the size of S, Wa and N are all bounded
by a polynomial in n. Hence, f(X) E
2?
LT
As a consequence of this result, Lemma 2 follows since it can be shown that
COM PARISON has a polynQmially bounded Ll spectral norm.
Now we are ready to state our main results. Although most linear threshold functions require exponentially large weights, we can always simulate them by 3 layers
of
elements.
in
Theorem 1
-
LTI ~ LT3
The result stated in Theorem 1 implies that a depth-d threshold circuit with exponentially large weights can be simulated by a depth-3d threshold circuit with
polynomially large weights. Using the result of Lemma 3, we can actually obtain a
more depth-efficient simulation.
Theorem 2
As another consequence of Lemma 3, we have the following :
947
948
Siu and Bruck
Corollary 1 Let /1 (X), ... , fm(X) be functions with polynomially bounded Ll spectral norms, and g(/1 (X), ... , fm(X? be an fi\ function with fi(X) 's as inputs,
I.e.
m
g(/1(X), ... , fm(X?
= sgn(2: Wdi(X) + wo)
i=l
Then 9 can be expressed as a sign of a sparse polynomial in X with polynomially
many number of monomial terms xcr 's and polynomially bounded integer coefficients. Hence 9 E LT2.
---
If all LTI functions have polynomially bounded Ll spectral norms, then it would
follow that LTI C iT 2 ? However, even the simple MAJORITY function does not
have a polynomially bounded Ll spectral norm. We shall prove this fact via the
following theorem. (As in Lemma 3, by a sparse polynomial we mean a polynomial
with only polynomially many monomial terms xcr's).
Theorem 3 The
iT l
function MAJORITY:
n
sgn(2: X i)
i=l
cannot be approximated by a sparse polynomial with an error o( n -1).
Other applications of the harmonic analysis techniques and the results of Lemma 3
yields the following theorems:
Theorem 4
Let x, y be two n-bit numbers. Then
ADDITION(x, y) E
m
---
Theorem 5 The product of two n-bit integers can be computed in LT4 ?
---
Theorem 6 The MAX I MU M of n n-bit numbers can be computed in LT3.
Theorem 7 The SORTING ofn n-bit numbers can be computed in
3
IT4 .
Concluding Remarks
Our main result indicates that for networks of linear threshold elements, we can
trade-off arbitrary real weights with polynomially bounded integer weights, at the
expense of a polynomial increase in the size and a factor of almost two in the depth of
the network. The proofs of the results in this paper can be found in [13]. We would
like to mention that our results have recently been improved by Goldmann, Hastad
and Razborov [4]. They showed that any polynomial-size depth-d network oflinear
threshold elements with arbitrary weights can be simulated by a polynomial-size
depth-( d + 1) network with "small" (polynomially bounded integer) weights. While
our construction can be made explicit, only the existence of the simulation result is
proved in [4]; it is left as an open problem in [4] if there is an explicit construction
of their results.
Neural Computing with Small Weights
Acknowledgements
This work was done while Kai-Yeung Siu was a research student associate at IBM
Almaden Research Center and was supported in part by the Joint Services Program
at Stanford University (US Army, US Navy, US Air Force) under Contract DAAL0388-C-0011, and the Department of the Navy (NAVELEX), NASA Headquarters,
Center for Aeronautics and Space Information Sciences under Grant NAGW-419S6.
References
[1] J. Bruck. Harmonic Analysis of Polynomial Threshold Functions. SIAM Journal on Discrete Mathematics, May 1990.
[2] J. Bruck and R. Smolensky. Polynomial Threshold Functions, ACo Functions
and Spectral Norms. Technical Report RJ 7140, IBM Research, November
1989. Appeared in IEEE Symp. on Found. of Compo Sci. October, 1990.
[3] A. K. Chandra, L. Stockmeyer, and U. Vishkin. Constant depth reducibility.
Siam J. Comput ., 13:423-439, 1984.
[4] M. Goldmann, J. Hastad, and A. Razborov Majority Gates vS. General
Weighted Threshold Gates. Unpublished Manuscript.
[5] A. HajnaI, W . Maass, P. PudIak, M. Szegedy, and G. Turan. Threshold circuits
of bounded depth . IEEE Symp. Found. Compo Sci., 28:99-110, 1987.
[6] R. J. Lechner. Harmonic analysis of switching functions. In A. Mukhopadhyay,
editor, Recent Development in Switching Theory. Academic Press, 1971.
[7] N. LiniaI, Y. Mansour, and N. Nisan. Constant Depth Circuits, Fourier Transforms, and Learnability. Proc. 30th IEEE Symp. Found. Compo Sci., 1989.
[8] S. Muroga and 1. Toda. Lower Bound of the Number of Threshold Functions.
IEEE Trans. on Electronic Computers, EC 15, 1966.
[9] J. Myhill and W. H. Kautz. On the Size of Weights Required for Linear-Input
Switching Functions. IRE Trans. on Electronic Computers, EC 10, 1961.
[10] I. Parberry and G. Schnitger. Parallel Computation with Threshold Functions
. Journal of Computer and System Sciences, 36(3):278-302, 1988.
[11] N. Pippenger. The complexity of computations by networks. IBM J. Res.
Develop. , 31(2), March 1987.
[12] P. Raghavan. Learning in Threshold Networks: A Computation Model and
Applications. Technical Report RC 13859, IBM Research, July 1988.
[13] K.-Y. Siu and J. Bruck. On the Power of Threshold Circuits with Small
Weights. SIAM J. Discrete Math., 4(3):423-435, August 1991.
[14] D. R. Smith. Bounds on the Number of Threshold Functions. IEEE Trans. on
Electronic Computers, EC 15, 1966.
[15] S. Yajima and T. Ibaraki. A Lower Bound on the Number of Threshold Functions. IEEE Trans. on Electronic Computers, EC 14, 1965.
949
| 529 |@word toda:1 indicate:2 polynomial:21 norm:8 come:1 implies:1 hence:2 question:2 open:1 closely:1 maass:1 simulation:2 sgn:5 raghavan:1 ll:7 mention:1 require:2 simulated:4 sci:3 it1:1 majority:3 generalized:1 yajima:1 tt:1 aes:1 pro:1 com:3 wdi:1 schnitger:1 harmonic:8 recently:2 must:2 fi:2 aco:1 iwil:1 october:1 expense:1 stated:2 exponentially:9 extend:1 v:1 proc:1 spec:1 arison:1 refer:1 observation:1 smith:1 november:1 tool:2 compo:3 weighted:1 similarly:1 mathematics:1 ire:1 provides:1 math:1 always:1 mansour:1 arbitrary:2 august:1 rc:1 aeronautics:1 direct:1 corollary:1 prove:3 showed:2 recent:1 namely:1 unpublished:1 symp:3 required:1 indicates:1 california:1 yi:2 trans:4 address:1 able:1 smolensky:1 appeared:1 increasing:1 program:1 july:1 ii:1 bounded:14 max:2 circuit:13 issue:2 headquarters:1 rj:1 almaden:2 power:2 iifll:2 development:1 technical:2 characterized:1 academic:1 developed:1 turan:1 natural:2 field:1 force:1 bruck:6 representing:1 improve:1 every:3 subclass:1 ti:1 growth:4 muroga:1 basic:1 ae:2 prohibitively:1 report:2 chandra:1 yeung:2 tral:1 grant:1 aal:1 yn:1 acknowledgement:1 reducibility:1 addition:3 service:1 engineering:1 loss:2 limit:1 consequence:2 switching:3 lt3:2 ourselves:1 grow:2 n1:1 crucial:1 noga:1 rest:1 yd:1 integer:10 principle:1 editor:1 limited:1 counting:1 range:1 ibm:5 practical:1 unique:1 ibaraki:1 supported:1 ltk:3 restrict:1 necessary:1 fm:3 reduce:1 idea:1 monomial:2 tradeoff:1 computable:4 allow:1 area:1 re:1 sparse:5 depth:20 xn:2 ltd:2 word:2 wo:3 computes:2 boolean:11 commonly:1 made:1 hastad:2 cannot:1 san:1 ec:4 remark:1 polynomially:12 center:3 siu:5 transforms:1 learnability:1 answer:1 xi:4 parberry:1 insight:1 sign:1 siam:3 s6:1 probabilistic:1 proving:1 off:2 contract:1 discrete:2 razborov:2 together:1 shall:6 ca:2 construction:3 du:1 threshold:35 constructing:1 associate:1 element:15 lti:5 approximated:2 main:5 motivation:1 mukhopadhyay:1 szegedy:1 sum:2 pigeonhole:1 jose:1 powerful:1 electrical:1 student:1 referred:1 coefficient:7 almost:2 throughout:1 lt1:8 explicitly:2 electronic:4 nisan:1 oflinear:1 trade:2 later:2 explicit:2 exponential:1 xl:1 comput:1 bit:14 mu:1 complexity:2 layer:6 ct:1 parallel:1 kautz:1 bound:3 dynamic:1 theorem:10 ofn:2 air:1 showing:1 division:1 x2:1 yield:2 exists:2 joint:1 generalize:1 fourier:1 myhill:1 represented:2 simulate:3 argument:1 concluding:1 magnitude:4 fast:2 department:1 sorting:3 march:1 lt:1 ehoshua:1 navy:2 quite:1 definition:2 kai:2 stanford:1 wi:4 ready:1 lt4:3 army:1 lechner:1 expressed:1 vishkin:1 proof:3 lt2:2 parison:1 irvine:2 rational:1 proved:3 aa:1 ask:2 turn:1 oflayers:1 needed:1 product:2 pippenger:1 actually:1 nasa:1 manuscript:1 replace:1 higher:1 studying:2 follow:1 iff:1 stockmeyer:1 specify:1 improved:1 goldmann:2 observe:1 done:1 spectral:10 except:1 generality:2 simulating:1 xa:2 lemma:9 called:1 experimental:2 hand:1 gate:5 existence:1 alon:1 develop:1 indicated:1 dept:1 giving:1 |
4,738 | 5,290 | Zero-Shot Recognition with Unreliable Attributes
Kristen Grauman
University of Texas at Austin
Austin, TX 78701
[email protected]
Dinesh Jayaraman
University of Texas at Austin
Austin, TX 78701
[email protected]
Abstract
In principle, zero-shot learning makes it possible to train a recognition model
simply by specifying the category?s attributes. For example, with classifiers for
generic attributes like striped and four-legged, one can construct a classifier for
the zebra category by enumerating which properties it possesses?even without
providing zebra training images. In practice, however, the standard zero-shot
paradigm suffers because attribute predictions in novel images are hard to get
right. We propose a novel random forest approach to train zero-shot models that
explicitly accounts for the unreliability of attribute predictions. By leveraging
statistics about each attribute?s error tendencies, our method obtains more robust
discriminative models for the unseen classes. We further devise extensions to handle the few-shot scenario and unreliable attribute descriptions. On three datasets,
we demonstrate the benefit for visual category learning with zero or few training
examples, a critical domain for rare categories or categories defined on the fly.
1
Introduction
Visual recognition research has achieved major successes in recent years using large datasets and
discriminative learning algorithms. The typical scenario assumes a multi-class task where one has
ample labeled training images for each class (object, scene, etc.) of interest. However, many realworld settings do not meet these assumptions. Rather than fix the system to a closed set of thoroughly
trained object detectors, one would like to acquire models for new categories with minimal effort
and training examples. Doing so is essential not only to cope with the ?long-tailed? distribution of
objects in the world, but also to support applications where new categories emerge dynamically?for
example, when a scientist defines a new phenomenon of interest to be detected in her visual data.
Zero-shot learning offers a compelling solution. In zero-shot learning, a novel class is trained via
description?not labeled training examples [10, 18, 8]. In general, this requires the learner to have
access to some mid-level semantic representation, such that a human teacher can define a novel
unseen class by specifying a configuration of those semantic properties. In visual recognition, the
semantic properties are attributes shared among categories, like black, has ears, or rugged. Supposing the system can predict the presence of any such attribute in novel images, then adding a new
category model amounts to defining its attribute ?signature? [8, 3, 18, 24, 19]. For example, even
without labeling any images of zebras, one could build a zebra classifier by instructing the system
that zebras are striped, black and white, etc. Interestingly, computational models for attribute-based
recognition are supported by the cognitive science literature, where researchers explore how humans
conceive of objects as bundles of attributes [25, 17, 5].
So, in principle, if we could perfectly predict attribute presence1 , zero-shot learning would offer
an elegant solution to generating novel classifiers on the fly. The problem, however, is that we
can?t assume perfect attribute predictions. Visual attributes are in practice quite difficult to learn
1
and have an attribute vocabulary rich enough to form distinct signatures for each category of interest
1
accurately?often even more so than object categories themselves. This is because many attributes
are correlated with one another (given only images of furry brown bears, how do we learn furry and
brown separately? [6]), and abstract linguistic properties can have very diverse visual instantiations
(compare a bumpy road to a bumpy rash). Thus, attribute-based zero-shot recognition remains in the
?proof of concept? realm, in practice falling short of alternate transfer methods [23].
We propose an approach to train zero-shot models that explicitly accounts for the unreliability of
attribute predictions. Whereas existing methods take attribute predictions at face value, our method
during training acknowledges the known biases of the mid-level attribute models. Specifically,
we develop a random forest algorithm that, given attribute signatures for each category, exploits
the attribute classifiers? receiver operating characteristics to select discriminative and predictable
decision nodes. We further generalize the idea to account for unreliable class-attribute associations.
Finally, we extend the solution to the ?few-shot? setting, where a small number of category-labeled
images are also available for training.
We demonstrate the idea on three large datasets of object and scene categories, and show its clear
advantages over status quo models. Our results suggest the valuable role attributes can play for
low-cost visual category learning, in spite of the inherent difficulty in learning them reliably.
2
Related Work
Most existing zero-shot models take a two-stage classification approach: given a novel image, first its
attributes are predicted, then its class label is predicted as a function of those attributes. For example,
in [3, 18, 30], each unseen object class is described by a binary indicator vector (?signature?) over its
attributes; a new image is mapped to the unseen class with the signature most similar to its attribute
predictions. The probabilistic Direct Attribute Prediction (DAP) method [8] takes a similar form, but
adds priors for the classes and attributes and computes a MAP prediction of the unseen class label.
A topic model variant is explored in [31]. The DAP model has gained traction and is often used in
other work [23, 19, 29]. In all of the above methods, as in ours, training an unseen class amounts to
specifying its attribute signature. In contrast to our approach, none of the existing methods account
for attribute unreliability when learning an unseen category. As we will see in the results, this has a
dramatic impact on generalization.
We stress that attribute unreliability is distinct from attribute strength. The former (our focus) pertains to how reliable the mid-level classifier is, whereas the latter pertains to how strongly an image
exhibits an attribute (e.g., as modeled by relative [19] or probabilistic [8] attributes). PAC bounds
on the tolerable error for mid-level classifiers are given in [18], but that work does not propose a
solution to mitigate the influence of their uncertainty.
While the above two-stage attribute-based formulation is most common, an alternative zero-shot
strategy is to exploit external knowledge about class relationships to adapt classifiers to an unseen
class. For example, an unseen object?s classifier can be estimated by combining the nearest existing classifiers (trained with images) in the ImageNet hierarchy [23, 14], or by combining classifiers
based on label co-occurrences [13]. In a similar spirit, label embeddings [1] or feature embeddings [4] can exploit semantic information for zero-shot predictions. Unlike these models, we focus
on defining new categories through language-based description (with attributes). This has the advantage of giving a human supervisor direct control on the unseen class?s definition, even if its
attribute signature is unlike that observed in any existing trained model.
Acknowledging that attribute classifiers are often unreliable, recent work abandons purely semantic
attributes in favor of discovering mid-level features that are both detectable and discriminative for
a set of class labels [11, 22, 26, 15, 30, 27, 1]. However, there is no guarantee that the discovered
features will align with semantic properties, particularly ?nameable? ones. This typically makes
them inapplicable to zero-shot learning, since a human supervisor can no longer define the unseen
class with concise semantic terms. Nonetheless, one can attempt to assign semantics post-hoc (e.g.,
[30]). We demonstrate that our method can benefit zero-shot learning with such discovered (pseudo)attributes as well.
Our idea for handling unreliable attributes in random forests is related to fractional tuples for handling missing values in decision trees [21]. In that approach, points with missing values are distributed down the tree in proportion to the observed values in all other data. Similar concepts are
explored in [28] to handle features represented as discrete distributions and in [16] to propagate
2
instances with soft node memberships. Our approach also entails propagating training instances in
proportion to uncertainty. However, our zero-shot scenario is distinct, and, accordingly, the training
and testing domains differ in important ways. At training time, rather than build a decision tree from
labeled data points, we construct each tree using the unseen classes? attribute signatures. Then, at
test time, the inputs are attribute classifier predictions. Furthermore, we show how to propagate both
signatures and data points through the tree simultaneously, which makes it possible to account for
inter-dependencies among the input dimensions and also enables a few-shot extension.
3
Approach
Given a vocabulary of M visual attributes, each unseen class k is described in terms of its attribute
signature Ak , which is an M -dimensional vector where Ak (i) gives the association of attribute
i with class k.2 Typically the association values would be binary?meaning that the attribute is
always present/absent in the class?but they may also be real-valued when such fine-grained data
is available. We model each unseen class with a single signature (e.g., whales are big and gray).
However, it is straightforward to handle the case where a class has a multi-modal definition (e.g.,
whales are big and gray OR whales are big and black), by learning a zero-shot model per ?mode?.
Whether the attribute vocabulary is hand-designed [8, 3, 19, 29, 23] or discovered [30, 11, 22], our
approach assumes it is expressive enough to discriminate between the categories.
Suppose there are K unseen classes of interest, for which we have no training images. Our zero-shot
method takes as input the K attribute signatures and a dataset of images labeled with attributes, and
produces a classifier for each unseen class as output. At test time, the goal is to predict which unseen
class appears in a novel image.
In the following, we first describe the initial stage of building the attribute classifiers (Sec. 3.1).
Then we introduce a zero-shot random forest trained with attribute signatures (Sec. 3.2). Next we
explain how to augment that training procedure to account for attribute unreliability (Sec. 3.2.2) and
signature uncertainty (Sec. 3.2.3). Finally, we present an extension to few-shot learning (Sec. 3.3).
3.1
Learning the attribute vocabulary
As in any attribute-based zero-shot method [3, 8, 18, 23, 19, 7, 29], we first must train classifiers to
predict the presence or absence of each of the M attributes in novel images. Importantly, the images
used to train the attribute classifiers may come from a variety of objects/scenes and need not contain
any instances of the unseen categories. The fact that attributes are shared across category boundaries
is precisely what allows zero-shot learning.
We train one SVM per attribute, using a training set of images xi (represented with standard descriptors) with binary M -dimensional label vectors yi , where yi (m) = 1 indicates that attribute m
is present in xi . Let a
?m (x) denote the Platt probability score from the m-th such SVM applied to
test input x.
3.2
Zero-shot random forests
Next we introduce our key contribution: a random forest model for zero-shot learning.
3.2.1
Basic formulation: Signature random forest
First we define a basic random forest training algorithm for the zero-shot setting. The main idea is
to train an ensemble of decision trees using attribute signatures?not image descriptors or vectors
of attribute predictions. In the zero-shot setting, this is all the training information available. Later,
at test time, we will have an image in hand, and we will apply the trained random forest to estimate
its class posteriors.
Recall that the k-th unseen class is defined by its attribute signature Ak ? <M . We treat each such
signature as the lone positive ?exemplar? for its class, and discriminatively train random forests to
distinguish all the signatures, A1 , . . . , AK . We take a one-versus-all approach, training one forest
for each unseen class. So, when training class k, the K ? 1 other class signatures are the negatives.
2
We use ?class? and ?category? to refer to an object or scene, e.g., zebra or beach, and ?attribute? to refer
to a property, e.g., striped or sunny. ?Unseen? means we have no training images for that class.
3
For each class, we build an ensemble of decision trees in a breadth-first manner. Each tree is learned
by recursively splitting the signatures into subsets at each node, starting at the root. Let In denote
an indicator vector of length K that records which signatures appear at node n. For the root node,
all K signatures are present, so we have In = [1, . . . , 1]. Following the typical random forest
protocol [2], the training instances are recursively split according to a randomized test; it compares
one dimension of the signature against a threshold t, then propagates each one to the left child l
or right child r depending on the outcome, yielding indicator vectors Il and Ir . Specifically, if
In (k) = 1, then if Ak (m) > t, we have Ir (k) = 1. Otherwise, Ir (k) = 0. Further, Il = In ? Ir .
Thus, during training we must choose two things at each node: the query attribute m and the threshold t, represented jointly as the split (m, t). We sample a limited number of (m, t) combinations3
and choose the one that maximizes the expected information gain IGbasic :
`
?
IGbasic (m, t) = H(pIn ) ? P (Ai (m) > t|In (i) = 1) H(pIl ) + P (Ai (m) ? t|In (i) = 1) H(pIr ) (1)
?
?
kIl k1
kIr k1
= H(pIn ) ?
(2)
H(pIl ) +
H(pIr ) ,
kIn k1
kIn k1
P
where H(p) = ? i p(i) log2 p(i) is the entropy of a distribution p. The 1-norm on an indicator
vector I sums up the occurrences I(k) of each signature, which for now are binary, I(k) ? {0, 1}.
Since we are training a zero-shot forest to discriminate class k from the rest, the distribution over
class labels at node n is a length-2 vector:
P
In (k)
i6=k In (i)
p In =
,
.
(3)
kIn k1
kIn k1
We grow each tree in the forest to a fixed, maximum depth, terminating a branch prematurely if less
than 5% of training samples have reached a node on it. We learn J = 100 trees per forest.
Given a novel test image xtest , we compute its predicted attribute signature a
?(xtest ) =
[?
a1 (xtest ), . . . , a
?M (xtest )] by applying the attribute SVMs. Then, to predict the posterior for
class k, we use a
?(xtest ) to traverse to a leaf node in each tree of k?s forest. Let Pkj (`) denote
the fraction of positive training instances at a leaf node ` in tree j of the forest for class k. Then
P
P (k|?
a(xtest )) = J1 j Pkj (`), the average of the posteriors across the ensemble.
If we somehow had perfect attribute classifiers, this basic zero-shot random forest (in fact, one such
tree alone) would be sufficient. Next, we show how to adapt the training procedure defined so far to
account for their unreliability.
3.2.2
Accounting for attribute prediction unreliability
While our training ?exemplars? are the true attribute signatures for each unseen class, the test images will have only approximate estimates of the attributes they contain. We therefore augment the
zero-shot random forest to account for this unreliability during training. The main idea is to generalize the recursive splitting procedure above such that a given signature can pursue multiple paths
down the tree. Critically, those paths will be determined by the false positive/true positive rates of
the individual attribute predictors. In this way, we expand each idealized training signature into a
distribution in the predicted attribute space. Essentially, this preemptively builds in the appropriate
?cushion? of expected errors when choosing discriminative splits.
Implementing this idea requires two primary extensions to the formulation in Sec. 3.2.1: (i) we
inject attribute validation data and its associated attribute classification error statistics into the tree
formation process, and (ii) we redefine the information gain to account for the partial propagation
of training signatures. We explain each of these components in turn next.
First, in addition to signatures, at each node we maintain a set of validation data in order to gauge
the error tendencies of each attribute classifier. For the experiments in this paper (Sec 4), our method
reserves some attribute classifier training data for this purpose. Denote this set of attribute-labeled
images as DV . During random forest training, this data is recursively propagated down the tree
following each split once it is chosen. Let DV (n) ? DV denote the set of validation data inherited
at node n. At the root, DV (n) = DV .
3
With binary Ai (m), all 0 < t < 1 are equivalent in Sec 3.2.1. Selecting t becomes important in Sec 3.2.2.
4
With validation data thus injected, we can estimate the test-time receiver operating characteristic
(ROC)4 for an attribute classifier at any node in the tree. For example, the estimated false positive
rate at node n for attribute m at threshold t is FP(n, m, t) = Pn (?
am (x) > t | y(m) = 0), which is
the fraction of examples in DV (n) for which the attribute m is absent, but the SVM predicts it to be
present at threshold t. Here, y(m) denotes the m-th attribute?s label for image x.
For any node n, let In0 be a real-valued indicator vector, such that In0 (k) ? [0, 1] records the fractional
occurrence of the training signature for class k at node n. At the root node, In0 (k) = 1, ?k. For a
split (m, t) at node n, a signature Ak splits into the right and left child nodes according to its ROC
for attribute m at the operating point specified by t. In particular, we have:
Ir0 (k) = In0 (k)Pn (?
am (x) > t | y(m) = Ak (m)), and Il0 (k) = In0 (k)Pn (?
am (x) ? t | y(m) = Ak (m)),
(4)
where x ? DV (n) . When Ak (m) = 1, the probability terms are TP(n, m, t) and FN(n, m, t)
respectively; when Ak (m) = 0, they are FP(n, m, t) and TN(n, m, t). In this way, we channel all
predicted negatives to the left child node. In contrast, a naive random forest (RF) trained on signatures assumes ideal attribute classifiers and channels all ground truth negatives?i.e., true negatives
and false positives?through the left node.
To illustrate the meaning of this fractional propagation, consider a class ?elephant? known to have
the attribute ?gray?. If the ?gray? attribute classifier fires only on 60% of the ?gray? samples in the
validation set, i.e., TP=0.6, then only 0.6 fraction of the ?elephant? signature is passed on to the
positive (i.e., right) node. This process repeats through more levels until fractions of the single ?elephant? signature have reached all leaf nodes. Thus, a single class signature emulates the estimated
statistics of a full training set of class-labeled instances with attribute predictions.
We stress two things about the validation data propagation. First, the data in DV is labeled by
attributes only; it has no unseen class labels and never features in the information gain computation.
Its only role is to estimate the ROC values. Second, the recursive sub-selection of the validation data
is important to capture the dependency of TP/FP rates at higher level splits. For example, if we were
to select split (m, t) at the root, then the fractional signatures pushed to the left child must all have
A(m) < t, meaning that for a candidate split (m, s) at the left child, where s > t, the correct TP and
FP rates are both 0. This is accounted for when we use DV (n) to compute the ROC, but would not
have been, had we just used DV . Thus, our formulation properly accounts for dependencies between
attributes when selecting discriminative thresholds, an issue not addressed by existing methods for
missing [21] or probabilistically distributed features [28].
Next, we redefine the information gain. When building a zero-shot tree conscious of attribute unreliability, we choose the split maximizing the expected information gain according to the fractionally
propagated signatures (compare to Eqn. (2)):
0
kIl k1
kIr0 k1
IGzero (m, t) = H(pIn0 ) ?
H(pIl0 ) + 0 H(pIr0 ) .
(5)
kIn0 k1
kIn k1
The distribution pIz0 , z ? {l, r} is computed as in Eqn. (3). For full pseudocode and a schematic
illustration of our method, please see supp.
The discriminative splits under this criterion will be those that not only distinguish the unseen classes
but also persevere (at test time) as a strong signal in spite of the attribute classifiers? error tendencies. This means the trees will prefer both reliable attributes that are discriminative among the
classes, as well as less reliable attributes coupled with intelligently selected operating points that
remain distinctive. Furthermore, they will omit splits that, though highly discriminative in terms of
idealized signatures, were found to be ?unlearnable? among the validation data. For example, in
the extreme case, if an attribute classifier cannot distinguish positives and negatives, meaning that
TPR=FPR, then the signatures of all classes are equally likely to propagate to the left or right, i.e.,
Ir0 (k)/In0 (k) = Ir0 (j)/In0 (j) and Il0 (k)/In0 (k) = Il0 (j)/In0 (j) for all k, j, which yields an information gain of 0 in Eqn. (5) (see supp). Thus, our method, while explicitly making the best of imperfect
attribute classification, inherently prefers more learnable attributes.
4
The ROC captures the true positive (TP) vs. false positive (FP) rates (equivalently the true negative (TN)
and false negative (FN) rates) as a function of a decision value threshold.
5
The proposed approach produces unseen category classifiers with zero category-labeled images.
The attribute-labeled validation data is important to our solution?s robustness. If that data perfectly
represented the true attribute errors on images from the unseen classes (which we cannot access,
of course, because images from those classes appear only at test time), then our training procedure
would be equivalent to building a random forest on the test samples? attribute classifier outputs.
3.2.3
Accounting for class signature uncertainty
Beyond attribute classifier unreliability, our framework can also deal with another source of zeroshot uncertainty: instances of a class often deviate from class-level attribute signatures. To tackle
this, we redefine the soft indicators Ir0 and Il0 in Eqn. 4, appending a term to account for annotation
noise. Please see supp. for details.
3.3
Extending to few-shot random forests
Our approach also admits a natural extension to few-shot training. Extensions of zero-shot models
to the few-shot setting have been attempted before [31, 26, 14, 1]. In this case, we are given not only
attribute signatures, but also a dataset DT consisting of a small number of images with their class
labels. We essentially use the signatures A1 , . . . , AK as a prior for selecting good tree splits that
also satisfy the traditional training examples. The information gain on the signatures is as defined in
Sec. 3.2.2, while the information gain on the training images, for which we can compute classifier
outputs, uses the standard measure defined in Sec. 3.2.1. Using some notation shortcuts, for few-shot
training we recursively select the split that maximizes the combined information gain:
IGf ew (m, t) = ? IGzero (m, t){A1 , . . . , AK } + (1 ? ?) IGbasic (m, t){DT },
(6)
where ? controls the role of the signature-based prior. Intuitively, we can expect lower values of ? to
suffice as the size of DT increases, since with more training examples we can more precisely learn
the class?s appearance. This few-shot extension can be interpreted as a new way to learn random
forests with descriptive priors.
4
Experiments
Datasets and setup We use three datasets: (1) Animals with Attributes (AwA) [8] (M = 85
attributes, K = 10 unseen classes, 30,475 total images), (2) aPascal/aYahoo objects (aPY) [3] (M =
65, K = 12, 15,339 images) (3) SUN scene attributes (SUN) [20] (M = 102, K = 10, 14,340
images). These datasets capture a wide array of categories (animals, indoor and outdoor scenes,
household objects, etc.) and attributes (parts, affordances, habitats, shapes, materials, etc.). The
attribute-labeled images originate from 40, 20, and 707 ?seen? classes in each dataset, respectively;
we use the class labels solely to map to attribute annotations. We use the unseen class splits specified
in [9] for AwA and aPY, and randomly select the 10 unseen classes for SUN (see supp.). For all three,
we use the features provided with the datasets, which include color histograms, SIFT, PHOG, and
others (see [9, 3, 20] for details).
Following [8], we train attribute SVMs with combined ?2 -kernels, one kernel per feature channel,
and set C = 10. Our method reserves 20% of the attribute-labeled images as ROC validation data,
then pools it with the remaining 80% to train the final attribute classifiers. We stress that our method
and all baselines have access to exactly the same amount of attribute-labeled data.
We report results as mean and standard error measured over 20 random trials. Based on crossvalidation, we use tree depths of (AwA-9, aPY-6, SUN-8), and generate (#m, #t) tests per node
(AwA-(10,7), aPY-(8,2), SUN-(4,5)). When too few validation points (< 10 positives or negatives)
reach a node n, we revert to computing statistics over the full validation set DV rather than DV (n).
Baselines In addition to several state-of-the-art published results and ablated variants of our
method, we also compare to two baselines: (1) SIGNATURE RF: random forests trained on classattribute signatures as described in Sec. 3.2.1, without an attribute uncertainty model, and (2) DAP:
Direct Attribute Prediction [8, 9], which is a leading attribute-based zero-shot object recognition
method widely used in the literature [8, 3, 18, 30, 8, 23, 19, 29].5
5
We use the authors? code: http://attributes.kyb.tuebingen.mpg.de/
6
Uniform noise levels
Attribute?specific noise levels
100
100
ours
signature?RF
DAP
80
accuracy(%)
accuracy(%)
80
60
40
20
0
60
40
ours
signature?RF
DAP
20
0
0.5
1
1.5
2
2.5
3
3.5
0
4
0
noise level ?
0.5
1
1.5
2
2.5
3
3.5
4
mean noise level ?
Figure 1: Zero-shot accuracy on AwA as a function of attribute uncertainty, in controlled noise
scenarios.
Method/Dataset
DAP
SIGNATURE - RF
OURS W / O ROC PROP, SIG UNCERTAINTY
OURS W / O SIG UNCERTAINTY
OURS
OURS + TRUE ROC
AwA
40.50
36.65 ? 0.16
39.97 ? 0.09
41.88 ? 0.08
43.01 ? 0.07
54.22 ? 0.03
aPY
18.12
12.70 ? 0.38
24.25 ? 0.18
24.79 ? 0.11
26.02 ? 0.05
33.54 ? 0.07
SUN
52.50
13.20 ? 0.34
47.46 ? 0.29
56.18 ? 0.27
56.18 ? 0.27
66.65 ? 0.31
Table 1: Zero-shot learning accuracy on all three datasets. Accuracy is percentage of correct category predictions on unseen class images, ? standard error.
4.1
Zero-shot object and scene recognition
Controlled noise experiments Our approach is designed to overcome the unreliability of attribute
classifiers. To glean insight into how it works, we first test it with controlled noise in the test images?
attribute predictions. We start with hypothetical perfect attribute classifier scores a
?m (x) = Ak (m)
for x in class k, then progressively add noise to represent increasing errors in the predictions. We
examine two scenarios: (1) where all attribute classifiers are equally noisy, and (2) where the average
noise level varies per attribute. See supp. for details on the noise model.
Figure 1 shows the results using AwA. By definition, all methods are perfectly accurate with zero
noise. Once the attributes are unreliable (i.e., noise > 0), however, our approach is consistently
better. Furthermore, our gains are notably larger in the second scenario where noise levels vary
per attribute (right plot), illustrating how our approach properly favors more learnable attributes
as discussed in Sec. 3.2.2. In contrast, SIGNATURE - RF is liable to break down with even minor
imperfections in attribute prediction. These results affirm that our method benefits from both (1)
estimating and accounting for classifier noisiness and (2) avoiding uninformative attribute classifiers.
Real unreliable attributes experiments Next we present the key zero-shot results for our method
applied to three challenging datasets using over 250 real attribute classifiers. Table 1 shows the
results. Our method significantly outperforms the existing DAP method [9]. This is an important
result: DAP is today the most commonly used model for zero-shot object recognition, whether using
this exact DAP formulation [8, 23, 19, 29] or very similar non-probabilistic variants [3, 30]. Note that
our approach beats DAP despite the fact we use only 80% of the attribute-labelled images to train
attribute classifiers. This indicates that modeling how good/bad the attribute classifiers are is even
more important than having better attribute classifiers. Furthermore, this demonstrates that modeling
only the confidence of an attribute?s presence in a test image (which DAP does) is inadequate; our
idea to characterize their error tendencies during training is valuable.
Our substantial improvements over SIGNATURE - RF also confirm it is imperative to model attribute
classifier unreliability. Our gains over DAP are especially large on SUN and aPY, which have fewer
positive training samples per attribute, leading to less reliable attribute classifiers?exactly where
our method is needed most. On AwA too, we outperform DAP on 7 out of 10 categories, with
largest gains on ?giant panda?(10.2%),?whale seal?(9.4%) and ?persian cat?(7.4%), classes that are
very different from the train classes. Further, if we repeat the experiment on AwA reducing to 500
randomly chosen images for attribute training, our overall accuracy gain over DAP widens to 8 points
(28.0 ? 0.9 vs. 20.42).
7
58
50 shot (our prior)
100 shot (our prior)
200 shot (our prior)
50 shot (baseline)
100 shot (baseline)
200 shot (baseline)
56
accuracy(%)
54
52
50
Method
Lampert et al. [8]
Yu and Aloimonos [31]
Rohrbach et al. [24]
Kankuekul et al. [7]
Yu et al. [30]
OURS (named attributes)
OURS (discovered attributes)
48
46
44
42
40
38
36
0
0.2
0.4
0.6
0.8
1
1.2
lambda
Accuracy
40.5
40.0
35.7
32.7
48.3
43.0 ? 0.07
48.7 ? 0.09
(b) Zero-shot vs. state of the art
(a) Few-shot. Stars denote selected ?.
Figure 2: (a) Few-shot results. (b) Zero-shot results on AwA compared to the state of the art.
Table 1 also helps isolate the impact of two components of our method: the model of signature
uncertainty (see OURS W / O SIG UNCERTAINTY), and the recursive propagation of validation data
(see OURS W / O ROC PROP, SIG UNCERTAINTY). For the latter, we further compute TPR/FPRs
globally on the full validation dataset DV rather than for node-specific subsets DV (n). We see both
aspects contribute to our full method?s best performance (see OURS). Finally, OURS + TRUE ROC
provides an ?upper bound? on the accuracy achievable with our method for these datasets; this is the
result attainable were we to use the unseen class images as validation data DV . This also points to
an interesting direction for future work: to better model expected error rates on images with unseen
attribute combinations. Our initial attempts in this regard included focusing validation data on seen
class images with signatures most like those of the unseen classes, but the impact was negligible.
Figure 2b compares our method against all published results on AwA, using both named and discovered attributes. When using standard AwA named attributes, our method comfortably outperforms
all prior methods. Further, when we use the discovered attributes from [30], it performs comparably
to their attribute decoding method, achieving the state-of-the-art on AwA. This result was obtained
using a generalization of our method to handle the continuous attribute strength signatures of [30].
4.2
Few-shot object and scene recognition
Finally, we demonstrate our few-shot extension. Figure 2a shows the results, as a function of both the
amount of labeled training images and the prior-weighting parameter ? (cf. Sec 3.3).6 When ? = 0,
we rely solely on the training images DT ; when ? = 1, we rely solely on the attribute signatures i.e.,
zero-shot learning. As a baseline, we compare to a method that uses solely the few training images
to learn the unseen classes (dotted lines). We see the clear advantage of our attribute signature prior
for few-shot random forest training. Furthermore, we see that, as expected, the optimal ? shifts
towards 0 as more samples are added. Still, even with 200 training images in DT , the prior plays
a role (e.g., the best ? = 0.3 on blue curve). The star per curve indicates the ? value our method
selects automatically with cross-validation.
5
Conclusion
We introduced a zero-shot training approach that models unreliable attributes?both due to classifier
predictions and uncertainty in their association with unseen classes. Our results on three challenging
datasets indicate the method?s promise, and suggest that the elegance of zero-shot learning need
not be abandoned in spite of the fact that visual attributes remain very difficult to predict reliably.
Further, our idea is applicable to other uses of semantic mid-level concepts for higher tasks e.g.,
poselets for action recognition [12], discriminative mid-level patches for location recognition [27]
etc., and in domains outside computer vision. In future work, we plan to develop extensions to
accommodate inter-attribute correlations in the random forest tests and multi-label random forests
to improve scalability for many unseen classes.
Acknowledgements: We thank Christoph Lampert and Felix Yu for helpful discussions and sharing
their code. This research is supported in part by NSF IIS-1065390 and ONR ATL.
6
These are for AwA; see supp. for similar results on the other two datasets.
8
References
[1] Z. Akata, F. Perronnin, Z. Harchaoui, and C. Schmid. Label-Embedding for Attribute-Based Classification. In CVPR, 2013.
[2] L. Breiman. Random forests. Machine learning, 45(1):5?32, 2001.
[3] A. Farhadi, I. Endres, D. Hoiem, and D. Forsyth. Describing objects by their attributes. In CVPR, 2009.
[4] A. Frome, G. Corrado, J. Shlens, S. Bengio, J. Dean, M. Ranzato, and T. Mikolov. Devise: A deep
visual-semantic embedding model. In NIPS, 2013.
[5] P. G?ardenfors. Conceptual Spaces: The Geometry of Thought, volume 106. 2000.
[6] D. Jayaraman, F. Sha, and K. Grauman. Decorrelating semantic visual attributes by resisting the urge to
share. In CVPR, 2014.
[7] P. Kankuekul, A. Kawewong, S. Tangruamsub, and O. Hasegawa. Online incremental attribute-based
zero-shot learning. In CVPR, 2012.
[8] C Lampert, H Nickisch, and S Harmeling. Learning to Detect Unseen Object Classes by Between-class
Attribute Transfer. In CVPR, 2009.
[9] Christoph H Lampert, Hannes Nickisch, and Stefan Harmeling. Attribute-based classification for zeroshot visual object categorization. TPAMI, 2014.
[10] H. Larochelle, D. Erhan, and Y. Bengio. Zero-data learning of new tasks. In AAAI, 2008.
[11] D. Mahajan, S. Sellamanicka, and V. Nair. A joint learning framework for attribute models and object
descriptions. In ICCV, 2011.
[12] S. Maji, L. Bourdev, and J. Malik. Action recognition from a distributed representation of pose and
appearance. In CVPR, 2011.
[13] T. Mensink, E. Gavves, and C. Snoek. COSTA: Co-occurrence statistics for zero-shot classification. In
CVPR, 2014.
[14] T. Mensink and J. Verbeek. Metric learning for large scale image classification: Generalizing to new
classes at near-zero cost. In ECCV, 2012.
[15] R. Mittelman, H. Lee, B. Kuipers, and S. Savarese. Weakly Supervised Learning of Mid-Level Features
with Beta-Bernoulli Process Restricted Boltzmann Machines. In CVPR, 2013.
[16] C. Olaru and L. Wehenkel. A complete fuzzy decision tree technique. Fuzzy Sets and Systems,
138(2):221?254, Sept 2003.
[17] D. Osherson, E. Smith, T. Myers, E. Shafir, and M. Stob. Extrapolating human probability judgment.
Theory and Decision, 36:103?129, 1994.
[18] M. Palatucci, D. Pomerleau, G. Hinton, and T. Mitchell. Zero-shot Learning with Semantic Output Codes.
In NIPS, 2009.
[19] D. Parikh and K. Grauman. Relative attributes. In ICCV, 2011.
[20] G Patterson and J Hays. SUN Attribute Database: Discovering, Annotating, and Recognizing Scene
Attributes. In CVPR, 2012.
[21] J. Quinlan. Induction of decision trees. Machine learning, pages 81?106, 1986.
[22] M. Rastegari, A. Farhadi, and D. Forsyth. Attribute discovery via predictable discriminative binary codes.
In ECCV, 2012.
[23] M. Rohrbach, M. Stark, and B. Schiele. Evaluating knowledge transfer and zero-shot learning in a largescale setting. In CVPR, 2011.
[24] M. Rohrbach, M. Stark, G. Szarvas, I. Gurevych, and B. Schiele. What helps where and why? semantic
relatedness for knowledge transfer. In CVPR, 2010.
[25] E. Rosch and B. Lloyd. Cognition and categorization. 1978.
[26] V. Sharmanska, N. Quadrianto, and C. Lampert. Augmented attribute representations. In ECCV, 2012.
[27] S. Singh, A. Gupta, and A. Efros. Unsupervised discovery of mid-level discriminative patches. In ECCV,
2012.
[28] S. Tsang, B. Kao, K. Yip, W.-S. Ho, and S. Lee. Decision Trees for Uncertain Data. IEEE Transactions
on Knowledge and Data Engineering, 23(1):64?78, January 2011.
[29] N. Turakhia and D. Parikh. Attribute dominance: what pops out? In ICCV, 2013.
[30] F. Yu, L. Cao, R. Feris, J. Smith, and S.-F. Chang. Designing Category-Level Attributes for Discriminative
Visual Recognition. In CVPR, 2013.
[31] X. Yu and Y. Aloimonos. Attribute-based transfer learning for object categorization with zero/one training
example. In ECCV, 2010.
9
| 5290 |@word trial:1 illustrating:1 achievable:1 proportion:2 norm:1 seal:1 propagate:3 accounting:3 xtest:6 attainable:1 concise:1 dramatic:1 accommodate:1 shot:64 recursively:4 initial:2 configuration:1 quo:1 score:2 selecting:3 hoiem:1 ours:13 interestingly:1 outperforms:2 existing:7 must:3 fn:2 gavves:1 j1:1 shape:1 enables:1 kyb:1 designed:2 plot:1 progressively:1 extrapolating:1 v:3 alone:1 preemptively:1 discovering:2 leaf:3 selected:2 phog:1 accordingly:1 fewer:1 fpr:1 smith:2 short:1 feris:1 record:2 provides:1 node:26 contribute:1 traverse:1 location:1 direct:3 beta:1 redefine:3 manner:1 introduce:2 jayaraman:2 inter:2 snoek:1 notably:1 expected:5 themselves:1 mpg:1 examine:1 multi:3 affordances:1 globally:1 automatically:1 kuiper:1 farhadi:2 increasing:1 becomes:1 provided:1 estimating:1 notation:1 suffice:1 maximizes:2 what:3 interpreted:1 pursue:1 fuzzy:2 lone:1 rugged:1 giant:1 guarantee:1 pseudo:1 mitigate:1 hypothetical:1 tackle:1 exactly:2 grauman:4 classifier:42 demonstrates:1 platt:1 control:2 shafir:1 omit:1 unreliability:12 appear:2 positive:12 before:1 scientist:1 negligible:1 treat:1 felix:1 engineering:1 despite:1 ak:13 meet:1 path:2 solely:4 black:3 dynamically:1 specifying:3 challenging:2 christoph:2 co:2 limited:1 harmeling:2 testing:1 practice:3 recursive:3 procedure:4 urge:1 significantly:1 thought:1 confidence:1 road:1 spite:3 suggest:2 get:1 cannot:2 selection:1 influence:1 applying:1 equivalent:2 map:2 dean:1 missing:3 maximizing:1 straightforward:1 starting:1 splitting:2 insight:1 array:1 importantly:1 shlens:1 embedding:2 handle:4 atl:1 hierarchy:1 play:2 suppose:1 today:1 exact:1 us:3 designing:1 sig:4 recognition:14 particularly:1 predicts:1 labeled:14 database:1 observed:2 role:4 fly:2 capture:3 tsang:1 sun:8 ranzato:1 valuable:2 substantial:1 predictable:2 schiele:2 legged:1 signature:59 trained:8 terminating:1 weakly:1 singh:1 purely:1 inapplicable:1 distinctive:1 patterson:1 learner:1 joint:1 osherson:1 represented:4 tx:2 cat:1 maji:1 train:12 revert:1 distinct:3 describe:1 detected:1 query:1 labeling:1 formation:1 outcome:1 choosing:1 outside:1 quite:1 widely:1 valued:2 larger:1 cvpr:12 otherwise:1 elephant:3 annotating:1 favor:2 statistic:5 unseen:37 jointly:1 noisy:1 abandon:1 final:1 online:1 hoc:1 advantage:3 descriptive:1 tpami:1 intelligently:1 myers:1 propose:3 cao:1 combining:2 description:4 kao:1 scalability:1 crossvalidation:1 extending:1 produce:2 generating:1 perfect:3 categorization:3 incremental:1 object:21 help:2 depending:1 develop:2 illustrate:1 pose:1 propagating:1 measured:1 bourdev:1 exemplar:2 nearest:1 minor:1 strong:1 c:2 predicted:5 come:1 indicate:1 poselets:1 differ:1 direction:1 frome:1 larochelle:1 correct:2 attribute:170 human:5 olaru:1 material:1 implementing:1 pkj:2 assign:1 fix:1 generalization:2 kristen:1 extension:9 ground:1 cognition:1 predict:6 reserve:2 kin0:1 major:1 vary:1 efros:1 purpose:1 applicable:1 label:13 utexas:2 largest:1 gauge:1 stefan:1 imperfection:1 always:1 rather:4 pn:3 breiman:1 probabilistically:1 linguistic:1 focus:2 noisiness:1 properly:2 consistently:1 improvement:1 indicates:3 bernoulli:1 contrast:3 baseline:7 am:3 detect:1 helpful:1 perronnin:1 membership:1 typically:2 her:1 expand:1 selects:1 semantics:1 issue:1 among:4 classification:7 overall:1 augment:2 animal:2 art:4 plan:1 yip:1 construct:2 once:2 never:1 beach:1 having:1 whale:4 yu:5 unsupervised:1 future:2 report:1 others:1 inherent:1 few:17 randomly:2 simultaneously:1 individual:1 geometry:1 consisting:1 fire:1 maintain:1 attempt:2 interest:4 highly:1 extreme:1 yielding:1 bundle:1 accurate:1 partial:1 tree:24 savarese:1 ablated:1 minimal:1 uncertain:1 instance:7 soft:2 compelling:1 modeling:2 mittelman:1 tp:5 cost:2 subset:2 rare:1 imperative:1 predictor:1 uniform:1 recognizing:1 supervisor:2 inadequate:1 too:2 characterize:1 dependency:3 teacher:1 varies:1 endres:1 nickisch:2 rash:1 thoroughly:1 combined:2 randomized:1 probabilistic:3 lee:2 decoding:1 pool:1 aaai:1 ear:1 bumpy:2 choose:3 lambda:1 cognitive:1 external:1 inject:1 leading:2 stark:2 supp:6 account:11 de:1 star:2 sec:14 lloyd:1 forsyth:2 satisfy:1 explicitly:3 idealized:2 later:1 root:5 break:1 closed:1 doing:1 reached:2 start:1 panda:1 pil:2 inherited:1 annotation:2 contribution:1 il:2 ir:4 accuracy:9 conceive:1 characteristic:2 acknowledging:1 descriptor:2 ensemble:3 emulates:1 yield:1 judgment:1 generalize:2 accurately:1 critically:1 comparably:1 none:1 liable:1 researcher:1 published:2 detector:1 explain:2 reach:1 suffers:1 sharing:1 definition:3 against:2 nonetheless:1 elegance:1 proof:1 associated:1 resisting:1 propagated:2 gain:13 costa:1 dataset:5 mitchell:1 recall:1 realm:1 knowledge:4 fractional:4 color:1 akata:1 appears:1 focusing:1 higher:2 dt:5 supervised:1 modal:1 hannes:1 formulation:5 decorrelating:1 though:1 strongly:1 affirm:1 furthermore:5 just:1 stage:3 cushion:1 mensink:2 until:1 correlation:1 hand:2 eqn:4 expressive:1 propagation:4 somehow:1 defines:1 mode:1 gray:5 building:3 brown:2 concept:3 contain:2 true:8 former:1 furry:2 semantic:12 dinesh:1 white:1 deal:1 mahajan:1 during:5 please:2 criterion:1 stress:3 dap:14 demonstrate:4 complete:1 tn:2 performs:1 image:47 meaning:4 novel:10 parikh:2 common:1 pseudocode:1 kil:2 volume:1 association:4 extend:1 tpr:2 discussed:1 comfortably:1 zeroshot:2 refer:2 zebra:6 ai:3 i6:1 language:1 had:2 access:3 entail:1 longer:1 operating:4 etc:5 add:2 align:1 posterior:3 recent:2 scenario:6 hay:1 binary:6 success:1 onr:1 yi:2 devise:2 seen:2 gurevych:1 paradigm:1 corrado:1 signal:1 ii:2 branch:1 multiple:1 full:5 persian:1 harchaoui:1 adapt:2 offer:2 long:1 cross:1 post:1 equally:2 a1:4 controlled:3 impact:3 prediction:19 variant:3 basic:3 schematic:1 verbeek:1 essentially:2 vision:1 metric:1 histogram:1 kernel:2 represent:1 ir0:4 palatucci:1 achieved:1 szarvas:1 whereas:2 addition:2 separately:1 fine:1 addressed:1 uninformative:1 grow:1 source:1 rest:1 unlike:2 posse:1 isolate:1 supposing:1 elegant:1 thing:2 ample:1 leveraging:1 spirit:1 near:1 presence:3 ideal:1 split:15 enough:2 embeddings:2 bengio:2 variety:1 perfectly:3 imperfect:1 idea:8 enumerating:1 texas:2 absent:2 shift:1 whether:2 pir:2 passed:1 effort:1 prefers:1 action:2 deep:1 clear:2 awa:14 amount:4 traction:1 mid:9 conscious:1 svms:2 category:27 generate:1 http:1 outperform:1 percentage:1 nsf:1 dotted:1 estimated:3 kankuekul:2 per:9 glean:1 blue:1 diverse:1 discrete:1 promise:1 dominance:1 key:2 four:1 fractionally:1 threshold:6 falling:1 achieving:1 breadth:1 fraction:4 year:1 sum:1 realworld:1 uncertainty:13 injected:1 nameable:1 named:3 patch:2 decision:10 prefer:1 pushed:1 bound:2 distinguish:3 strength:2 precisely:2 striped:3 scene:9 aspect:1 mikolov:1 according:3 alternate:1 combination:1 sunny:1 across:2 remain:2 making:1 dv:15 intuitively:1 handling:2 iccv:3 restricted:1 remains:1 pin:2 detectable:1 turn:1 describing:1 needed:1 available:3 apply:1 generic:1 appropriate:1 tolerable:1 occurrence:4 appending:1 alternative:1 robustness:1 ho:1 abandoned:1 assumes:3 denotes:1 include:1 remaining:1 cf:1 wehenkel:1 log2:1 quinlan:1 in0:9 household:1 widens:1 exploit:3 giving:1 k1:10 build:4 especially:1 malik:1 added:1 rosch:1 strategy:1 primary:1 sha:1 traditional:1 exhibit:1 thank:1 mapped:1 topic:1 originate:1 tuebingen:1 induction:1 length:2 code:4 modeled:1 relationship:1 illustration:1 providing:1 acquire:1 equivalently:1 difficult:2 setup:1 hasegawa:1 negative:8 kir:1 reliably:2 pomerleau:1 boltzmann:1 upper:1 datasets:12 beat:1 january:1 defining:2 hinton:1 prematurely:1 discovered:6 sharmanska:1 introduced:1 specified:2 imagenet:1 instructing:1 learned:1 pop:1 nip:2 aloimonos:2 beyond:1 indoor:1 fp:5 rf:7 reliable:4 critical:1 apascal:1 difficulty:1 natural:1 rely:2 indicator:6 largescale:1 improve:1 habitat:1 acknowledges:1 naive:1 coupled:1 schmid:1 sept:1 deviate:1 prior:11 literature:2 acknowledgement:1 discovery:2 relative:2 expect:1 bear:1 discriminatively:1 interesting:1 versus:1 validation:17 sufficient:1 propagates:1 principle:2 share:1 austin:4 eccv:5 course:1 accounted:1 supported:2 repeat:2 bias:1 wide:1 face:1 emerge:1 benefit:3 regard:1 curve:2 distributed:3 boundary:1 dimension:2 vocabulary:4 world:1 depth:2 rich:1 computes:1 overcome:1 author:1 commonly:1 evaluating:1 far:1 erhan:1 cope:1 transaction:1 approximate:1 obtains:1 relatedness:1 status:1 unreliable:8 confirm:1 instantiation:1 receiver:2 conceptual:1 tuples:1 discriminative:13 xi:2 continuous:1 tailed:1 why:1 table:3 learn:6 transfer:5 robust:1 channel:3 inherently:1 rastegari:1 correlated:1 forest:29 domain:3 protocol:1 main:2 big:3 noise:14 lampert:5 quadrianto:1 child:6 augmented:1 roc:10 sub:1 candidate:1 outdoor:1 weighting:1 grained:1 kin:5 down:4 bad:1 specific:2 pac:1 sift:1 learnable:2 explored:2 svm:3 admits:1 gupta:1 essential:1 false:5 adding:1 gained:1 entropy:1 generalizing:1 simply:1 explore:1 likely:1 appearance:2 rohrbach:3 visual:13 il0:4 chang:1 truth:1 prop:2 nair:1 goal:1 towards:1 labelled:1 shared:2 absence:1 shortcut:1 hard:1 included:1 typical:2 specifically:2 determined:1 reducing:1 total:1 discriminate:2 tendency:4 attempted:1 ew:1 select:4 support:1 latter:2 pertains:2 phenomenon:1 avoiding:1 unlearnable:1 |
4,739 | 5,291 | Articulated Pose Estimation by a Graphical Model
with Image Dependent Pairwise Relations
Alan Yuille
University of California, Los Angeles
Los Angeles, CA 90024
[email protected]
Xianjie Chen
University of California, Los Angeles
Los Angeles, CA 90024
[email protected]
Abstract
We present a method for estimating articulated human pose from a single static
image based on a graphical model with novel pairwise relations that make adaptive use of local image measurements. More precisely, we specify a graphical
model for human pose which exploits the fact the local image measurements can
be used both to detect parts (or joints) and also to predict the spatial relationships
between them (Image Dependent Pairwise Relations). These spatial relationships
are represented by a mixture model. We use Deep Convolutional Neural Networks
(DCNNs) to learn conditional probabilities for the presence of parts and their spatial relationships within image patches. Hence our model combines the representational flexibility of graphical models with the efficiency and statistical power of
DCNNs. Our method significantly outperforms the state of the art methods on the
LSP and FLIC datasets and also performs very well on the Buffy dataset without
any training.
1
Introduction
Articulated pose estimation is one of the fundamental challenges in computer vision. Progress in
this area can immediately be applied to important vision tasks such as human tracking [2], action
recognition [25] and video analysis.
Most work on pose estimation has been based on graphical model [8, 6, 27, 1, 10, 2, 4]. The graph
nodes represent the body parts (or joints), and the edges model the pairwise relationships between
the parts. The score function, or energy, of the model contains unary terms at each node which
capture the local appearance cues of the part, and pairwise terms defined at the edges which capture
the local contextual relations between the parts. Recently, DeepPose [23] advocates modeling pose
in a holistic manner and captures the full context of all body parts in a Deep Convolutional Neural
Network (DCNN) [12] based regressor.
In this paper, we present a graphical model with image dependent pairwise relations (IDPRs). As
illustrated in Figure 1, we can reliably predict the relative positions of a part?s neighbors (as well as
the presence of the part itself) by only observing the local image patch around it. So in our model
the local image patches give input to both the unary and pairwise terms. This gives stronger pairwise
terms because data independent relations are typically either too loose to be helpful or too strict to
model highly variable poses.
Our approach requires us to have a method that can extract information about pairwise part relations,
as well as part presence, from local image patches. We require this method to be efficient and to
share features between different parts and part relationships. To do this, we train a DCNN to output
1
Lower Arm:
Upper Arm:
Elbow:
Wrist:
Figure 1: Motivation. The local image measurements around a part, e.g., in an image patch, can reliably
predict the relative positions of all its neighbors (as well as detect the part). Center Panel: The local image
patch centered at the elbow can reliably predict the relative positions of the shoulder and wrist, and the local
image patch centered at the wrist can reliably predict the relative position of the elbow. Left & Right Panels: We
define different types of pairwise spatial relationships (i.e., a mixture model) for each pair of neighboring parts.
The Left Panel shows typical spatial relationships the elbow can have with its neighbors, i.e., the shoulder and
wrist. The Right Panel shows typical spatial relationships the wrist can have with its neighbor, i.e., the elbow.
estimates for the part presence and spatial relationships which are used in our unary and pairwise
terms of our score function. The weight parameters of different terms in the model are trained using
Structured Supported Vector Machine (S-SVM) [24]. In summary, our model combines the representational flexibility of graphical models, including the ability to represent spatial relationships,
with the data driven power of DCNNs.
We perform experiments on two standard pose estimation benchmarks: LSP dataset [10] and FLIC
dataset [20]. Our method outperforms the state of the art methods by a significant margin on both
datasets. We also do cross-dataset evaluation on Buffy dataset [7] (without training on this dataset)
and obtain strong results which shows the ability of our model to generalize.
2
The Model
The Graphical Model and its Variables: We represent human pose by a graphical model G =
(V, E) where the nodes V specify the positions of the parts (or joints) and the edges E indicates
which parts are spatially related. For simplicity, we impose that the graph structure forms a K?node
tree, where K = |V|. The positions of the parts are denoted by l, where li = (x, y) specifies the
pixel location of part i, for i ? {1, . . . , K}. For each edge in the graph (i, j) ? E, we specify a
discrete set of spatial relationships indexed by tij , which corresponds to a mixture of different spatial
relationships (see Figure 1). We denote the set of spatial relationships by t = {tij , tji |(i, j) ? E}.
The image is written as I. We will define a score function F (l, t|t) as follows as a sum of unary and
pairwise terms.
Unary Terms: The unary terms give local evidence for part i ? V to lie at location li and is based
on the local image patch I(li ). They are of form:
U (li |I) = wi ?(i|I(li ); ?),
(1)
where ?(.|.; ?) is the (scalar-valued) appearance term with ? as its parameters (specified in the next
section), and wi is a scalar weight parameter.
Image Dependent Pairwise Relational (IDPR) Terms: These IDPR terms capture our intuition
that neighboring parts (i, j) ? E can roughly predict their relative spatial positions using only local
information (see Figure 1). In our model, the relative positions of parts i and j are discretized
into several types tij ? {1, . . . , Tij } (i.e., a mixture of different relationships) with corresponding
t
mean relative positions rijij plus small deformations which are modeled by the standard quadratic
2
deformation term. More formally, the pairwise relational score of each edge (i, j) ? E is given by:
R(li , lj , tij , tji |I) =
+
t
t
hwijij , ?(lj ? li ? rijij )i + wij ?(tij |I(li ); ?)
,
t
t
hwjiji , ?(li ? lj ? rjiji )i + wji ?(tji |I(lj ); ?)
(2)
where ?(?l = [?x, ?y]) = [?x ?x2 ?y ?y 2 ]| are the standard quadratic deformation features,
?(.|.; ?) is the Image Dependent Pairwise Relational (IDPR) term with ? as its parameters (specified
t
t
in the next section), and wijij , wij , wjiji , wji are the weight parameters. The notation h., .i specifies
dot product and boldface indicates a vector.
The Full Score: The full score F (l, t|I) is a function of the part locations l, the pairwise relation
types t, and the input image I. It is expressed as the sum of the unary and pairwise terms:
X
X
F (l, t|I) =
U (li |I) +
R(li , lj , tij , tji |I) + w0 ,
(3)
i?V
(i,j)?E
where w0 is a scalar weight on constant 1 (i.e., the bias term).
t
t
The model consists of three sets of parameters: the mean relative positions r = {rijij , rjiji |(i, j) ? E}
of different pairwise relation types; the parameters ? of the appearance terms and IDPR terms; and
t
t
the weight parameters w (i.e., wi , wijij , wij , wjiji , wji and w0 ). See Section 4 for the learning of
these parameters.
2.1
Image Dependent Terms and DCNNs
The appearance terms and IDPR terms depend on the image patches. In other words, a local image
patch I(li ) not only gives evidence for the presence of a part i, but also about the relationship tij
between it and its neighbors j ? N(i), where j ? N(i) if, and only if, (i, j) ? E. This requires
us to learn distribution for the state variables i, tij conditioned on the image patches I(li ). In order
to specify this distribution we must define the state space more precisely, because the number of
pairwise spatial relationships varies for different parts with different numbers of neighbors (see
Figure 1), and we need also consider the possibility that the patch does not contain a part.
We define c to be the random variable which denotes which part is present c = i for i ? {1, ..., K}
or c = 0 if no part is present (i.e., the background). We define mcN(c) to be the random variable that
determines the spatial relation types of c and takes values in McN(c) . If c = i has one neighbor j
(e.g., the wrist), then MiN(i) = {1, . . . , Tij }. If c = i has two neighbors j and k (e.g., the elbow),
then MiN(i) = {1, . . . , Tij } ? {1, . . . , Tik }. If c = 0, then we define M0N(0) = {0}.
The full space is represented as:
PK
S = ?K
c=0 {c} ? McN(c)
(4)
The size of the space is |S| = c=0 |McN(c) |. Each element in this space corresponds to a part with
all the types of its pairwise relationships, or the background.
We use DCNN [12] to learn the conditional probability distribution p(c, mcN(c) |I(li ); ?). DCNN is
suitable for this task because it is very efficient and enables us to share features. See section 4 for
more details.
We specify the appearance terms ?(.|.; ?) and IDPR terms ?(.|.; ?) in terms of p(c, mcN(c) |I(li ); ?)
by marginalization:
?(i|I(li ); ?) = log(p(c = i|I(li ); ?))
?(tij |I(li ); ?) = log(p(mij = tij |c = i, I(li ); ?))
2.2
(5)
(6)
Relationship to other models
We now briefly discuss how our method relates to standard models.
Pictorial Structure: We recover pictorial structure models [6] by only allowing one relationship
type (i.e., Tij = 1). In this case, our IDPR term conveys no information. Our model reduces to
3
standard unary and (image independent) pairwise terms. The only slight difference is that we use
DCNN to learn the unary terms instead of using HOG filters.
Mixtures-of-parts: [27] describes a model with a mixture of templates for each part, where each
template is called a ?type? of the part. The ?type? of each part is defined by its relative position with
respect to its parent. This can be obtained by restricting each part in our model to only predict the
relative position of its parent (i.e., Tij = 1, if j is not parent of i). In this case, each part is associated
with only one informative IDPR term, which can be merged with the appearance term of each part
to define different ?types? of part in [27]. Also this method does not use DCNNs.
Conditional Random Fields (CRFs): Our model is also related to the conditional random field
literature on data-dependent priors [18, 13, 15, 19]. The data-dependent priors and unary terms
are typically modeled separately in the CRFs. In this paper, we efficiently model all the image
dependent terms (i.e. unary terms and IDPR terms) together in a single DCNN by exploiting the fact
the local image measurements are reliable for predicting both the presence of a part and the pairwise
relationships of a part with its neighbors.
3
Inference
To detect the optimal configuration for each person, we search for the configurations of the locations
l and types t that maximize the score function: (l? , t? ) = arg maxl,t F (l, t|I). Since our relational
graph is a tree, this can be done efficiently via dynamic programming.
Let K(i) be the set of children of part i in the graph (K(i) = ?, if part i is a leaf), and Si (li |I) be
maximum score of the subtree rooted at part i with part i located at li . The maximum score of each
subtree can be computed as follow:
X
Si (li |I) = U (li |I) +
max (R(li , lk , tik , tki |I) + Sk (lk |I))
(7)
k?K(i)
lk ,tik ,tki
Using Equation 7, we can recursively compute the overall best score of the model, and the optimal
configuration of locations and types can be recovered by the standard backward pass of dynamic
programming.
Computation: Since our pairwise term is a quadratic function of locations, li and lj , the max
operation over lk in Equation 7 can be accelerated by using the generalized distance transforms [6].
The resulting approach is very efficient, taking O(T 2 LK) time once the image dependent terms are
computed, where T is the number of relation types, L is the total number of locations, and K is the
total number of parts in the model. This analysis assumes that all the pairwise spatial relationships
have the same number of types, i.e., Tij = Tji = T, ?(i, j) ? E.
The computation of the image dependent terms is also efficient. They are computed over all the
locations by a single DCNN. Applying DCNN in a sliding fashion is inherently efficient, since the
computations common to overlapping regions are naturally shared [22].
4
Learning
Now we consider the problem of learning the model parameters from images with labeled part
locations, which is the data available in most of the human pose datasets [17, 7, 10, 20]. We derive
type labels tij from part location annotations and adopt a supervised approach to learn the model.
Our model consists of three sets of parameters: the mean relative positions r of different pairwise
relation types; the parameters ? of the image dependent terms; and the weight parameters w. They
are learnt separately by the K-means algorithm for r, DCNN for ?, and S-SVM for w.
Mean Relative Positions and Type Labels: Given the labeled positive images {(In , ln )}N
n=1 , let
dij be the relative position from part i to its neighbor j. We cluster the relative positions over the
training set {dnij }N
n=1 to get Tij clusters (in the experiments Tij = 11 for all pairwise relations).
Each cluster corresponds to a set of instances of part i that share similar spatial relationship with
its neighbor part j. Thus we define each cluster as a pairwise relation type tij from part i to j in
t
our model, and use the center of each cluster as the mean relative position rijij associated with each
4
type. In this way, the mean relative positions of different pairwise relation types are learnt, and the
type label tnij for each training instance is derived based on its cluster index. We use K-means in our
experiments by setting K = Tij to do the clustering.
Parameters of Image Dependent Terms: After deriving type labels, each local image patch I(ln )
centered at an annotated part location is labeled with category label cn ? {1, . . . , K}, that indicates which part is present, and also the type labels mncn N(cn ) that indicate its relation types with
all its neighbors. In this way, we get a set of labelled patches {I(ln ), cn , mncn N(cn ) }KN
n=1 from positive images (each positive image provides K part patches), and also a set of background patches
{I(ln ), 0, 0} sampled from negative images.
Given the labelled part patches and background patches, we train a multi-class DCNN classifier by
standard stochastic gradient descent using softmax loss. The DCNN consists of five convolutional
layers, 2 max-pooling layers and three fully-connected layers with a final |S| dimensions softmax
output, which is defined as our conditional probability distribution, i.e., p(c, mcN(c) |I(li ); ?). The
architecture of our network is summarized in Figure 2.
Weight Parameters: Each pose in the positive image is now labeled with annotated part locations
and derived type labels: (In , ln , tn ). We use S-SVM to learn the weight parameters w. The structure
prediction problem is simplified by using 0 ? 1 loss, that is all the training examples either have all
dimensions of its labels correct or all dimensions of its labels wrong. We denote the former ones
as pos examples, and the later ones as neg examples. Since the full score function (Equation 3) is
linear in the weight parameters w, we write the optimization function as:
X
1
min hw, wi + C
max(0, 1 ? yn hw, ?(In , ln , tn )i),
(8)
w 2
n
where yn ? {1, ?1}, and ?(In , ln , tn ) is a sparse feature vector representing the n-th example
and is the concatenation of the image dependent terms (calculated from the learnt DCNN), spatial
deformation features, and constant 1. Here yn = 1 if n ? pos, and yn = ?1 if n ? neg.
5
Experiment
This section introduces the datasets, clarifies the evaluation metrics, describes our experimental
setup, presents comparative evaluation results and gives diagnostic experiments.
5.1
Datasets and Evaluation Metrics
We perform our experiments on two publicly available human pose estimation benchmarks: (i)
the ?Leeds Sports Poses? (LSP) dataset [10], that contains 1000 training and 1000 testing images
from sport activities with annotated full-body human poses; (ii) the ?Frames Labeled In Cinema?
(FLIC) dataset [20] that contains 3987 training and 1016 testing images from Hollywood movies
with annotated upper-body human poses. We follow previous work and use the observer-centric
annotations on both benchmarks. To train our models, we also use the negative training images from
the INRIAPerson dataset [3] (These images do not contain people).
We use the most popular evaluation metrics to allow comparison with previous work. Percentage
of Correct Parts (PCP) is the standard evaluation metric on several benchmarks including the LSP
dataset. However, as discussed in [27], there are several alternative interpretations of PCP that can
lead to severely different results. In our paper, we use the stricter version unless otherwise stated,
that is we evaluate only a single highest-scoring estimation result for one test image and a body part
is considered as correct if both of its segment endpoints (joints) lie within 50% of the length of the
ground-truth annotated endpoints (Each test image on the LSP dataset contains only one annotated
person). We refer to this version of PCP as strict PCP.
On the FLIC dataset, we use both strict PCP and the evaluation metric specified with it [20]: Percentage of Detected Joints (PDJ). PDJ measures the performance using a curve of the percentage
of correctly localized joints by varying localization precision threshold. The localization precision
threshold is normalized by the scale (defined as distance between left shoulder and right hip) of each
ground truth pose to make it scale invariant. There are multiple people in the FLIC images, so each
5
7
5
3
3
3
dense
+
dropout
9
128
128
x3
6
54
4096
x5
3
4
3
dense
9x
9
8
128
dense
+
dropout
conv
9x
9
9
x1
128
3
conv
9x
9x
18
36
32
3
3
conv
3
conv
+
norm
+
pool
3
5
7
conv
+
norm
+
pool
OR
4096
|S|
Figure 2: Architectures of our DCNNs. The size of input patch is 36 ? 36 pixels on the LSP dataset, and
54 ? 54 pixels on the FLIC dataset. The DCNNs consist of five convolutional layers, 2 max-pooling layers
and three fully-connected (dense) layers with a final |S| dimensions output. We use dropout, local response
normalization (norm) and overlapping pooling (pool) described in [12].
ground truth person is also annotated with a torso detection box. During evaluation, we return a
single highest-scoring estimation result for each ground truth person by restricting our neck part to
be localized inside a window defined by the provided torso box.
5.2
Implementation detail
Data Augmentation: Our DCNN has millions of parameters, while only several thousand of training images are available. In order to reduce overfitting, we augment the training data by rotating
the positive training images through 360? . These images are also horizontally flipped to double the
training images. This increases the number of training examples of body parts with different spatial
relationships with its neighbors (See the elbows along the diagonal of the Left Panel in Figure 1).
We hold out random positive images as a validation set for the DCNN training. Also the weight
parameters w are trained on this held out set to reduce overfitting to training data.
Note that our DCNN is trained using local part patches and background patches instead of the whole
images. This naturally increases the number of training examples by a factor of K (the number of
parts). Although the number of dimensions of the DCNN final output also increases linearly with the
number of parts, the number of parameters only slightly increase in the last fully-connected layer.
This is because most of the parameters are shared between different parts, and thus we can benefit
from larger K by having more training examples per parameter. In our experiments, we increase K
by adding the midway points between annotated parts, which results in 26 parts on the LSP dataset
and 18 parts on the FLIC dataset. Covering a person by more parts also reduces the distance between
neighboring parts, which is good for modeling foreshortening [27].
Graph Structure: We define a full-body graph structure for the LSP dataset, and a upper-body
graph structure for the FLIC dataset respectively. The graph connects the annotated parts and their
midways points to form a tree (See the skeletons in Figure 5 for clarification).
Settings: We use the same number of types for all pairs of neighbors for simplicity. We set it as
11 on all datasets (i.e., Tij = Tji = 11, ?(i, j) ? E), which results in 11 spatial relation types
for the parts with one neighbor (e.g., the wrist), 112 spatial relation types for the parts with two
neighbors (e.g., the elbow), and so forth (recall Figure 1). The patch size of each part is set as
36 ? 36 pixels on the LSP dataset, and 54 ? 54 pixels on the FLIC dataset, as the FLIC images are
of higher resolution. We use similar DCNN architectures on both datasets, which differ in the first
layer because of different input patch sizes. Figure 2 visualizes the architectures we used. We use
the Caffe [9] implementation of DCNN in our experiments.
5.3
Benchmark results
We show strict PCP results on the LSP dataset in Table 1, and on the FLIC dataset in Table 2. We
also show PDJ results on the FLIC dataset in Figure 3. As is shown, our method outperforms state
of the art methods by a significant margin on both datasets (see the captions for detailed analysis).
Figure 5 shows some estimation examples on the LSP and FLIC datasets.
6
Method
Ours
Torso
92.7
Head
87.8
U.arms
69.2
L.arms
55.4
U.legs
82.9
L.legs
77.0
Mean
75.0
Pishchulin et al. [16]
Ouyang et al. [14]
DeepPose* [23]
Pishchulin et al. [15]
Eichner&Ferrari [4]
Yang&Ramanan [26]
88.7
85.8
87.5
86.2
84.1
85.6
83.1
78.1
80.1
77.1
61.5
63.3
56
54.2
56.5
52.5
44.9
46.6
38
33.9
37.4
35.9
78.8
76.5
77
75.7
74.3
69.5
73.4
72.2
71
68.0
69.3
65.6
69.2
68.6
62.9
64.3
60.8
Table 1: Comparison of strict PCP results on the LSP dataset. Our method improves on all parts by a significant
margin, and outperforms the best previously published result [16] by 5.8% on average. Note that DeepPose uses
Person-Centric annotations and is trained with an extra 10,000 images.
MODEC[20]
U.arms
97.0
L.arms
86.8
84.4
52.1
Mean
91.9
68.3
Table 2: Comparison of strict PCP results on the
FLIC dataset. Our method significantly outperforms
MODEC [20].
5.4
Wrists
0.8
Percentage of Detected Joints (PDJ)
Method
Ours
Percentage of Detected Joints (PDJ)
Elbows
1
0.9
MODEC: 75.5%
DeepPose: 91.0%
Ours: 94.9%
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0
0.05
0.1
0.15
Normalized Precision Threshold
0.2
1
0.9
0.8
MODEC: 57.9%
DeepPose: 80.9%
Ours: 92.0%
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0
0.05
0.1
0.15
Normalized Precision Threshold
0.2
Figure 3: Comparison of PDJ curves of elbows and
wrists on the FLIC dataset. The legend shows the
PDJ numbers at the threshold of 0.2.
Diagnostic Experiments
We perform diagnostic experiments to show the cross-dataset generalization ability of our model,
and better understand the influence of each term in our model.
Cross-dataset Generalization: We directly apply the trained model on the FLIC dataset to the
official test set of Buffy dataset [7] (i.e., no training on the Buffy dataset), which also contains
upper-body human poses. The Buffy test set includes a subset of people whose upper-body can be
detected. We get the newest detection windows from [5], and compare our results to previously
published work on this subset.
Most previous work was evaluated with the official evaluation toolkit of Buffy, which uses a less
strict PCP implementation1 . We refer to this version of PCP as Buffy PCP and report it along with the
strict PCP in Table 3. We also show the PDJ curves in Figure 4. As is shown by both criterions, our
method significantly outperforms the state of the arts, which shows the good generalization ability
of our method. Also note that both DeepPose [23] and our method are trained on the FLIC dataset.
Compared with Figure 3, the margin between our method and DeepPose significantly increases in
Figure 4, which implies that our model generalizes better to the Buffy dataset.
L.arms
89.0
84.1
Mean
92.9
89.3
97.8
94.3
95.3
93.2
93.2
68.6
57.5
63.0
60.6
60.3
83.2
75.9
79.2
76.9
76.8
Elbows
Table 3: Cross-dataset PCP results on Buffy test subset. The PCP numbers are Buffy PCP unless otherwise stated. Note that our method is trained on the
FLIC dataset.
Wrists
1
0.9
0.8
Percentage of Detected Joints (PDJ)
Yang[27]
Yang[27] strict
Sapp[21]
FLPM[11]
Eichner[5]
U.arms
96.8
94.5
Percentage of Detected Joints (PDJ)
Method
Ours*
Ours* strict
Yang: 80.4%
MODEC: 77.0%
DeepPose*: 83.4%
Ours*: 93.2%
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0
0.05
0.1
0.15
Normalized Precision Threshold
0.2
1
0.9
0.8
Yang: 57.4%
MODEC: 58.8%
DeepPose*: 64.6%
Ours*: 89.4%
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0
0.05
0.1
0.15
Normalized Precision Threshold
0.2
Figure 4: Cross-dataset PDJ curves on Buffy test
subset. The legend shows the PDJ numbers at the
threshold of 0.2. Note that both our method and
DeepPose [23] are trained on the FLIC dataset.
1
A part is considered correctly localized if the average distance between its endpoints (joints) and groundtruth is less than 50% of the length of the ground-truth annotated endpoints.
7
Method
Unary-Only
No-IDPRs
Torso
56.3
87.4
Head
66.4
74.8
U.arms
28.9
60.7
L.arms
15.5
43.0
U.legs
50.8
73.2
L.legs
45.9
65.1
Mean
40.5
64.6
Full Model
92.7
87.8
69.2
55.4
82.9
77.0
75.0
Table 4: Diagnostic term analysis strict PCP results on the LSP dataset. The unary term alone is still not
powerful enough to get good results, even though it?s trained by a DCNN classifier. No-IDPRs method, whose
pairwise terms are not dependent on the image (see Terms Analysis in Section 5.4), can get comparable performance with the state-of-the-art, and adding IDPR terms significantly boost our final performance to 75.0%.
Terms Analysis: We design two experiments to better understand the influence of each term in
our model. In the first experiment, we use only the unary terms and thus all the parts are localized
independently. In the second experiment, we replace the IDPR terms with image independent priors
(i.e., in Equation 2, wij ?(tij |I(li ); ?) and wji ?(tji |I(lj ); ?) are replaced with scalar prior terms
t
t
bijij and bjiji respectively), and retrain the weight parameters along with the new prior terms. In this
case, our pairwise relational terms do not depend on the image, but instead is a mixture of Gaussian
deformations with image independent biases. We refer to the first experiment as Unary-Only and the
second one as No-IDPRs, short for No IDPR terms. The experiments are done on the LSP dataset
using identical appearance terms for fair comparison. We show strict PCP results in Table 4. As is
shown, all terms in our model significantly improve the performance (see the caption for detail).
6
Conclusion
We have presented a graphical model for human pose which exploits the fact the local image measurements can be used both to detect parts (or joints) and also to predict the spatial relationships
between them (Image Dependent Pairwise Relations). These spatial relationships are represented
by a mixture model over types of spatial relationships. We use DCNNs to learn conditional probabilities for the presence of parts and their spatial relationships within image patches. Hence our
model combines the representational flexibility of graphical models with the efficiency and statistical power of DCNNs. Our method outperforms the state of the art methods on the LSP and FLIC
datasets and also performs very well on the Buffy dataset without any training.
Figure 5: Results on the LSP and FLIC datasets. We show the part localization results along with the graph
skeleton we used in the model. The last row shows some failure cases, which are typically due to large foreshortening, occlusions and distractions from clothing or overlapping people.
7
Acknowledgements
This research has been supported by grants ONR MURI N000014-10-1-0933, ONR N00014-12-10883 and ARO 62250-CS. The GPUs used in this research were generously donated by the NVIDIA
Corporation.
8
References
[1] X. Chen, R. Mottaghi, X. Liu, S. Fidler, R. Urtasun, and A. Yuille. Detect what you can: Detecting and
representing objects using holistic models and body parts. In Computer Vision and Pattern Recognition
(CVPR), 2014.
[2] N.-G. Cho, A. L. Yuille, and S.-W. Lee. Adaptive occlusion state estimation for human pose tracking
under self-occlusions. Pattern Recognition, 2013.
[3] N. Dalal and B. Triggs. Histograms of oriented gradients for human detection. In Computer Vision and
Pattern Recognition (CVPR), 2005.
[4] M. Eichner and V. Ferrari. Appearance sharing for collective human pose estimation. In Asian Conference
on Computer Vision (ACCV), 2012.
[5] M. Eichner, M. Marin-Jimenez, A. Zisserman, and V. Ferrari. 2d articulated human pose estimation and
retrieval in (almost) unconstrained still images. International Journal of Computer Vision (IJCV), 2012.
[6] P. F. Felzenszwalb and D. P. Huttenlocher. Pictorial structures for object recognition. International
Journal of Computer Vision (IJCV), 2005.
[7] V. Ferrari, M. Marin-Jimenez, and A. Zisserman. Progressive search space reduction for human pose
estimation. In Computer Vision and Pattern Recognition (CVPR), 2008.
[8] M. A. Fischler and R. A. Elschlager. The representation and matching of pictorial structures. IEEE
Transactions on Computers, 1973.
[9] Y. Jia. Caffe: An open source convolutional architecture for fast feature embedding. http://caffe.
berkeleyvision.org/, 2013.
[10] S. Johnson and M. Everingham. Clustered pose and nonlinear appearance models for human pose estimation. In British Machine Vision Conference (BMVC), 2010.
[11] L. Karlinsky and S. Ullman. Using linking features in learning non-parametric part models. In European
Conference on Computer Vision (ECCV), 2012.
[12] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural
networks. In Neural Information Processing Systems (NIPS), 2012.
[13] J. Lafferty, A. McCallum, and F. C. Pereira. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In International Conference on Machine Learning (ICML), 2001.
[14] W. Ouyang, X. Chu, and X. Wang. Multi-source deep learning for human pose estimation. In Computer
Vision and Pattern Recognition (CVPR), 2014.
[15] L. Pishchulin, M. Andriluka, P. Gehler, and B. Schiele. Poselet conditioned pictorial structures. In
Computer Vision and Pattern Recognition (CVPR), 2013.
[16] L. Pishchulin, M. Andriluka, P. Gehler, and B. Schiele. Strong appearance and expressive spatial models
for human pose estimation. In International Conference on Computer Vision (ICCV), 2013.
[17] D. Ramanan. Learning to parse images of articulated bodies. In Neural Information Processing Systems
(NIPS), 2006.
[18] C. Rother, V. Kolmogorov, and A. Blake. Grabcut: Interactive foreground extraction using iterated graph
cuts. In ACM Transactions on Graphics (TOG), 2004.
[19] B. Sapp, C. Jordan, and B. Taskar. Adaptive pose priors for pictorial structures. In Computer Vision and
Pattern Recognition (CVPR), 2010.
[20] B. Sapp and B. Taskar. Modec: Multimodal decomposable models for human pose estimation. In Computer Vision and Pattern Recognition (CVPR), 2013.
[21] B. Sapp, A. Toshev, and B. Taskar. Cascaded models for articulated pose estimation. In European
Conference on Computer Vision (ECCV), 2010.
[22] P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. LeCun. Overfeat: Integrated recognition, localization and detection using convolutional networks. In International Conference on Learning
Representations (ICLR), 2014.
[23] A. Toshev and C. Szegedy. Deeppose: Human pose estimation via deep neural networks. In Computer
Vision and Pattern Recognition (CVPR), 2014.
[24] I. Tsochantaridis, T. Hofmann, T. Joachims, and Y. Altun. Support vector machine learning for interdependent and structured output spaces. In International Conference on Machine Learning (ICML), 2004.
[25] C. Wang, Y. Wang, and A. L. Yuille. An approach to pose-based action recognition. In Computer Vision
and Pattern Recognition (CVPR), 2013.
[26] Y. Yang and D. Ramanan. Articulated pose estimation with flexible mixtures-of-parts. In Computer Vision
and Pattern Recognition (CVPR), 2011.
[27] Y. Yang and D. Ramanan. Articulated human detection with flexible mixtures of parts. IEEE Transactions
on Pattern Analysis and Machine Intelligence (TPAMI), 2013.
9
| 5291 |@word version:3 briefly:1 dalal:1 stronger:1 norm:3 everingham:1 triggs:1 open:1 recursively:1 reduction:1 configuration:3 contains:5 score:11 liu:1 jimenez:2 ours:8 outperforms:7 recovered:1 contextual:1 si:2 chu:1 written:1 must:1 pcp:17 informative:1 midway:2 hofmann:1 enables:1 newest:1 cue:1 leaf:1 alone:1 intelligence:1 mccallum:1 short:1 provides:1 detecting:1 node:4 location:12 org:1 zhang:1 five:2 along:4 consists:3 ijcv:2 combine:3 advocate:1 inside:1 manner:1 pairwise:31 roughly:1 multi:2 discretized:1 window:2 elbow:11 conv:5 provided:1 estimating:1 notation:1 panel:5 what:1 ouyang:2 corporation:1 donated:1 interactive:1 stricter:1 eichner:4 classifier:2 wrong:1 ramanan:4 grant:1 yn:4 segmenting:1 positive:6 local:19 tki:2 severely:1 marin:2 deeppose:11 plus:1 lecun:1 wrist:10 testing:2 cxj:1 x3:1 area:1 significantly:6 matching:1 word:1 altun:1 get:5 tsochantaridis:1 context:1 applying:1 influence:2 center:2 crfs:2 independently:1 resolution:1 simplicity:2 decomposable:1 immediately:1 deriving:1 embedding:1 ferrari:4 caption:2 programming:2 us:2 element:1 recognition:15 located:1 cut:1 muri:1 labeled:5 huttenlocher:1 gehler:2 taskar:3 wang:3 capture:4 thousand:1 region:1 connected:3 highest:2 intuition:1 pdj:12 skeleton:2 fischler:1 schiele:2 m0n:1 dynamic:2 trained:9 depend:2 segment:1 yuille:5 localization:4 tog:1 efficiency:2 karlinsky:1 po:2 joint:12 multimodal:1 represented:3 kolmogorov:1 articulated:8 train:3 fast:1 detected:6 labeling:1 caffe:3 whose:2 larger:1 valued:1 cvpr:10 otherwise:2 ability:4 itself:1 final:4 sequence:1 tpami:1 aro:1 product:1 neighboring:3 holistic:2 flexibility:3 representational:3 forth:1 los:4 exploiting:1 parent:3 cluster:6 double:1 sutskever:1 comparative:1 object:2 derive:1 pose:32 stat:1 progress:1 strong:2 c:1 indicate:1 implies:1 differ:1 merged:1 annotated:10 tji:7 filter:1 stochastic:1 correct:3 centered:3 human:21 require:1 generalization:3 clustered:1 clothing:1 hold:1 around:2 considered:2 ground:5 blake:1 predict:8 adopt:1 estimation:19 tik:3 label:9 hollywood:1 generously:1 dcnn:19 gaussian:1 varying:1 derived:2 joachim:1 indicates:3 detect:5 helpful:1 inference:1 dependent:16 unary:15 typically:3 lj:7 integrated:1 relation:19 wij:4 pixel:5 dcnns:9 arg:1 overall:1 classification:1 denoted:1 augment:1 flexible:2 overfeat:1 spatial:25 art:6 softmax:2 andriluka:2 field:3 once:1 having:1 extraction:1 identical:1 flipped:1 progressive:1 icml:2 foreground:1 report:1 oriented:1 foreshortening:2 asian:1 pictorial:6 replaced:1 connects:1 occlusion:3 detection:5 possibility:1 highly:1 evaluation:9 introduces:1 mixture:10 held:1 edge:5 unless:2 tree:3 indexed:1 rotating:1 deformation:5 instance:2 hip:1 modeling:2 subset:4 krizhevsky:1 dij:1 johnson:1 too:2 graphic:1 kn:1 varies:1 learnt:3 cho:1 person:6 fundamental:1 international:6 lee:1 probabilistic:1 regressor:1 pool:3 together:1 augmentation:1 return:1 ullman:1 li:27 szegedy:1 summarized:1 includes:1 later:1 observer:1 observing:1 recover:1 annotation:3 jia:1 implementation1:1 publicly:1 convolutional:7 efficiently:2 clarifies:1 generalize:1 iterated:1 published:2 visualizes:1 sharing:1 failure:1 energy:1 conveys:1 naturally:2 associated:2 static:1 sampled:1 dataset:40 cinema:1 popular:1 recall:1 improves:1 torso:4 sapp:4 centric:2 higher:1 supervised:1 follow:2 specify:5 response:1 zisserman:2 bmvc:1 done:2 box:2 evaluated:1 though:1 parse:1 expressive:1 nonlinear:1 overlapping:3 wijij:2 contain:2 normalized:5 former:1 hence:2 fidler:1 spatially:1 illustrated:1 x5:1 during:1 self:1 rooted:1 covering:1 berkeleyvision:1 criterion:1 generalized:1 tn:3 performs:2 image:62 novel:1 recently:1 lsp:16 common:1 endpoint:4 million:1 discussed:1 slight:1 interpretation:1 linking:1 measurement:5 significant:3 refer:3 unconstrained:1 dot:1 toolkit:1 driven:1 n00014:1 nvidia:1 poselet:1 onr:2 mottaghi:1 wji:4 neg:2 scoring:2 impose:1 grabcut:1 maximize:1 ii:1 relates:1 full:8 sliding:1 multiple:1 reduces:2 alan:1 cross:5 retrieval:1 prediction:1 vision:19 metric:5 histogram:1 represent:3 normalization:1 background:5 separately:2 source:2 extra:1 strict:12 pooling:3 legend:2 lafferty:1 jordan:1 presence:7 yang:7 enough:1 marginalization:1 architecture:5 reduce:2 cn:4 angeles:4 action:2 deep:5 tij:23 detailed:1 transforms:1 category:1 http:1 specifies:2 mcn:7 percentage:7 diagnostic:4 correctly:2 per:1 discrete:1 write:1 threshold:8 backward:1 graph:11 sum:2 powerful:1 you:1 almost:1 groundtruth:1 patch:24 comparable:1 dropout:3 layer:8 quadratic:3 activity:1 precisely:2 x2:1 ucla:2 toshev:2 min:3 gpus:1 structured:2 describes:2 slightly:1 wi:4 leg:4 invariant:1 iccv:1 ln:7 equation:4 previously:2 discus:1 loose:1 available:3 operation:1 generalizes:1 apply:1 alternative:1 eigen:1 denotes:1 assumes:1 clustering:1 graphical:11 exploit:2 parametric:1 diagonal:1 gradient:2 iclr:1 distance:4 concatenation:1 w0:3 urtasun:1 boldface:1 rother:1 length:2 modeled:2 relationship:27 index:1 modec:7 sermanet:1 setup:1 hog:1 negative:2 stated:2 implementation:2 reliably:4 design:1 collective:1 perform:3 allowing:1 upper:5 datasets:11 benchmark:5 descent:1 accv:1 relational:5 shoulder:3 head:2 hinton:1 frame:1 pair:2 specified:3 imagenet:1 california:2 boost:1 nip:2 pattern:12 challenge:1 including:2 reliable:1 video:1 max:5 power:3 suitable:1 leeds:1 predicting:1 cascaded:1 arm:10 representing:2 pishchulin:4 movie:1 improve:1 mathieu:1 lk:5 extract:1 prior:6 literature:1 acknowledgement:1 interdependent:1 relative:16 loss:2 fully:3 localized:4 validation:1 flic:21 elschlager:1 share:3 row:1 eccv:2 summary:1 supported:2 last:2 bias:2 allow:1 understand:2 neighbor:16 template:2 taking:1 felzenszwalb:1 sparse:1 benefit:1 curve:4 maxl:1 dimension:5 calculated:1 adaptive:3 simplified:1 transaction:3 overfitting:2 fergus:1 search:2 sk:1 table:8 learn:7 mncn:2 ca:2 inherently:1 european:2 official:2 pk:1 dense:4 linearly:1 motivation:1 whole:1 child:1 fair:1 body:12 x1:1 retrain:1 fashion:1 precision:6 position:18 pereira:1 lie:2 hw:2 british:1 svm:3 evidence:2 consist:1 restricting:2 adding:2 subtree:2 conditioned:2 margin:4 chen:2 appearance:10 horizontally:1 expressed:1 tracking:2 sport:2 scalar:4 mij:1 corresponds:3 truth:5 determines:1 acm:1 conditional:7 buffy:12 labelled:2 shared:2 replace:1 typical:2 called:1 total:2 pas:1 neck:1 experimental:1 clarification:1 formally:1 distraction:1 people:4 support:1 accelerated:1 evaluate:1 |
4,740 | 5,292 | Exact Post Model Selection Inference for Marginal
Screening
Jason D. Lee
Computational and Mathematical Engineering
Stanford University
Stanford, CA 94305
[email protected]
Jonathan E. Taylor
Department of Statistics
Stanford University
Stanford, CA 94305
[email protected]
Abstract
We develop a framework for post model selection inference, via marginal screening, in linear regression. At the core of this framework is a result that characterizes the exact distribution of linear functions of the response y, conditional on
the model being selected (?condition on selection" framework). This allows us
to construct valid confidence intervals and hypothesis tests for regression coefficients that account for the selection procedure. In contrast to recent work in
high-dimensional statistics, our results are exact (non-asymptotic) and require no
eigenvalue-like assumptions on the design matrix X. Furthermore, the computational cost of marginal regression, constructing confidence intervals and hypothesis testing is negligible compared to the cost of linear regression, thus making
our methods particularly suitable for extremely large datasets. Although we focus
on marginal screening to illustrate the applicability of the condition on selection
framework, this framework is much more broadly applicable. We show how to
apply the proposed framework to several other selection procedures including orthogonal matching pursuit and marginal screening+Lasso.
1
Introduction
Consider the model
yi = ?(xi ) + i , i ? N (0, ? 2 I),
(1)
p
where ?(x) is an arbitrary function, and xi ? R . Our goal is to perform inference on
(X T X)?1 X T ?, which is the best linear predictor of ?. In the classical setting of n > p , the
least squares estimator ?? = (X T X)?1 X T y is a commonly used estimator for (X T X)?1 X T ?.
Under the linear model assumption ? = X? 0 , the exact distribution of ?? is
?? ? N (? 0 , ? 2 (X T X)?1 ).
(2)
Using the normal distribution, we can test the hypothesis H0 : ?j0 = 0 and form confidence intervals
for ?j0 using the z-test.
However in the high-dimensional p > n setting, the least squares estimator is an underdetermined
problem, and the predominant approach is to perform variable selection or model selection [4].
There are many approaches to variable selection including AIC/BIC, greedy algorithms such as
forward stepwise regression, orthogonal matching pursuit, and regularization methods such as the
Lasso. The focus of this paper will be on the model selection procedure known as marginal screening, which selects the k most correlated features xj with the response y.
Marginal screening is the simplest and most commonly used of the variable selection procedures
[9, 21, 16]. Marginal screening requires only O(np) computation and is several orders of magnitude
1
faster than regularization methods such as the Lasso; it is extremely suitable for extremely large
datasets where the Lasso may be computationally intractable to apply. Furthermore, the selection
properties are comparable to the Lasso [8].
Since marginal screening utilizes the response variable y, the confidence intervals and statistical
tests based on the distribution in (2) are not valid; confidence intervals with nominal 1 ? ? coverage
may no longer cover at the advertised level:
Pr ?j0 ? C1?? (x) < 1 ? ?.
Several authors have previously noted this problem including recent work in [13, 14, 15, 2]. A major
line of work [13, 14, 15] has described the difficulty of inference post model selection: the distribution of post model selection estimates is complicated and cannot be approximated in a uniform
sense by their asymptotic counterparts.
In this paper, we describe how to form exact confidence intervals for linear regression coefficients
post model selection. We assume the model (1), and operate under the fixed design matrix X
setting. The linear regression coefficients constrained to a subset of variables S is linear in ?,
eTj (XST XS )?1 XST ? = ? T ? for some ?. We derive the conditional distribution of ? T y for any
vector ?, so we are able to form confidence intervals for regression coefficients.
In Section 2 we discuss related work on high-dimensional statistical inference, and Section 3 introduces the marginal screening algorithm and shows how z intervals may fail to have the correct
coverage properties. Section 4 and 5 show how to represent the marginal screening selection event
as constraints on y, and construct pivotal quantities for the truncated Gaussian. Section 6 uses these
tools to develop valid confidence intervals, and Section 7 evaluates the methodology on two real
datasets.
Although the focus of this paper is on marginal screening, the ?condition on selection" framework,
first proposed for the Lasso in [12], is much more general; we use marginal screening as a simple and
clean illustration of the applicability of this framework. In Section 8, we discuss several extensions
including how to apply the framework to other variable/model selection procedures and to nonlinear
regression problems. Section 8 covers 1) marginal screening+Lasso, a screen and clean procedure
that first uses marginal screening and cleans with the Lasso, and orthogonal matching pursuit (OMP).
2
Related Work
Most of the theoretical work on high-dimensional linear models focuses on consistency. Such results
establish, under restrictive assumptions on X, the Lasso ?? is close to the unknown ? 0 [19] and
selects the correct model [26, 23, 11]. We refer to the reader to [4] for a comprehensive discussion
about the theoretical properties of the Lasso.
There is also recent work on obtaining confidence intervals and significance testing for penalized Mestimators such as the Lasso. One class of methods uses sample splitting or subsampling to obtain
confidence intervals and p-values [24, 18]. In the post model selection literature, the recent work of
[2] proposed the POSI approach, a correction to the usual t-test confidence intervals by controlling
the familywise error rate for all parameters in any possible submodel. The POSI methodology is
extremely computationally intensive and currently only applicable for p ? 30.
A separate line of work establishes the asymptotic normality of a corrected estimator obtained by
? z,
?inverting? the KKT conditions [22, 25, 10]. The corrected estimator ?b has the form ?b = ?? + ???
? is an approximate inverse to the Gram matrix
where z? is a subgradient of the penalty at ?? and ?
X T X. The two main drawbacks to this approach are 1) the confidence intervals are valid only when
the M-estimator is consistent, and thus require restricted eigenvalue conditions on X, 2) obtaining
? and 3) the method is specific to regularized
? is usually much more expensive than obtaining ?,
?
estimators, and does not extend to marginal screening, forward stepwise, and other variable selection
methods.
Most closely related to our work is the ?condition on selection" framework laid out in [12] for the
Lasso. Our work extends this methodology to other variable selection methods such as marginal
screening, marginal screening followed by the Lasso (marginal screening+Lasso) and orthogonal
matching pursuit. The primary contribution of this work is the observation that many model selection
2
methods, including marginal screening and Lasso, lead to ?selection events" that can be represented
as a set of constraints on the response variable y. By conditioning on the selection event, we can
characterize the exact distribution of ? T y. This paper focuses on marginal screening, since it is
the simplest of variable selection methods, and thus the applicability of the ?condition on selection
event" framework is most transparent. However, this framework is not limited to marginal screening
and can be applied to a wide a class of model selection procedures including greedy algorithms such
as orthogonal matching pursuit. We discuss some of these possible extensions in Section 8, but leave
a thorough investigation to future work.
A remarkable aspect of our work is that we only assume X is in general position, and the test is exact,
meaning the distributional results are true even under finite samples. By extension, we do not make
any assumptions on n and p, which is unusual in high-dimensional statistics [4]. Furthermore, the
computational requirements of our test are negligible compared to computing the linear regression
coefficients.
3
Marginal Screening
Let X ? Rn?p be the design matrix, y ? Rn the response variable, and assume the model
yi = ?(xi ) + i , i ? N (0, ? 2 I). We will assume that X is in general position and has unit norm
columns. The algorithm estimates ?? via Algorithm 1. The marginal screening algorithm chooses
Algorithm 1 Marginal screening algorithm
1:
2:
3:
4:
Input: Design matrix X, response y, and model size k.
Compute |X T y|.
Let S? be the index of the k largest entries of |X T y|.
Compute ??S? = (XST? XS? )?1 XST? y
the k variables with highest absolute dot product with y, and then fits a linear model over those k
variables. We will assume k ? min(n, p). For any fixed subset of variables S, the distribution of
??S = (XST XS )?1 XST y is
??S ? N (XST XS )?1 XST ?, ? 2 (XST XS )?1
(3)
?
We will use the notation ?j?S
:= (?S? )j , where j is indexing a variable in the set S. The z-test
intervals for a regression coefficient are
C(?, j, S) := ??j?S ? ?z1??/2 (XST XS )jj , ??j?S + ?z1??/2 (XST XS )jj
(4)
?
and each interval has 1 ? ? coverage, meaning Pr ?j?S
? C(?, j, S) = 1 ? ?. However if S? is
chosen using a model selection procedure that depends on y, the distributional
result (3)no longer
?
? < 1 ? ?.
holds and the z-test intervals will not cover at the 1 ? ? level, and Pr ?j?S? ? C(?, j, S)
3.1
Failure of z-test confidence intervals
We will illustrate empirically that the z-test intervals do not cover at 1 ? ? when S? is chosen by
marginal screening in Algorithm 1. For this experiment we generated X from a standard normal
with n = 20 and p = 200. The signal vector is 2 sparse with ?10 , ?20 = SNR, y = X? 0 + , and
? N (0, 1). The confidence intervals were constructed for the k = 2 variables selected by the
marginal screening algorithm. The z-test intervals were constructed via (4) with ? = .1, and the
adjusted intervals were constructed using Algorithm 2. The results are described in Figure 1.
4
Representing the selection event
Since Equation (3) does not hold for a selected S? when the selection procedure depends on y, the
z-test intervals are not valid. Our strategy will be to understand the conditional distribution of y
3
Coverage Proportion
1
0.9
0.8
0.7
Adjusted
Z test
0.6
0.5
0.4
?1
0
log10 SNR
1
Figure 1: Plots of the coverage proportion across a range of SNR (log-scale). We see that the
coverage proportion of the z intervals can be far below the nominal level of 1 ? ? = .9, even at
SNR =5. The adjusted intervals always have coverage proportion .9. Each point represents 500
independent trials.
and contrasts (linear functions of y) ? T y, then construct inference conditional on the selection event
? We will use E(y)
?
E.
to represent a random variable, and E to represent an element of the range of
?
?
E(y). In the case of marginal screening, the selection event E(y)
corresponds to the set of selected
?
variables S and signs s:
n
o
?
E(y)
= y : sign(xTi y)xTi y > ?xTj y for all i ? S? and j ? S?c
n
o
= y : s?i xTi y > ?xTj y and s?i xTi y ? 0 for all i ? S? and j ? S?c
n
o
? s?)y ? b(S,
? s?)
= y : A(S,
(5)
? s?) and vector b(S,
? s?)1 . We will use the selection event E
? and the selected
for some matrix A(S,
?
variables/signs pair (S, s?) interchangeably since they are in bijection.
F
The space Rn is partitioned by the selection events, Rn =
(S,s) {y : A(S, s)y ?
2
b(S,
s)}
.
The
vector
y
can
be
decomposed
with
respect
to
the
partition
as follows y =
P
y
1
(A(S,
s)y
?
b(S,
s)).
S,s
Theorem 4.1. The distribution of y conditional on the selection event is a constrained Gaussian,
d
?
y|{E(y)
= E} = z {A(S, s)z ? b}, z ? N (?, ? 2 I).
Proof. The event
E is in bijection with a pair (S, s), and y is unconditionally Gaussian. Thus the
conditional y {A(S, s)y ? b(S, s)} is a Gaussian constrained to the set {A(S, s)y ? b(S, s)}.
5
Truncated Gaussian test
This section summarizes the recent tools developed in [12] for testing contrasts3 ? T y of a constrained Gaussian y. The results are stated without proof and the proofs can be found in [12]. The
primary result is a one-dimensional pivotal quantity for ? T ?. This pivot relies on characterizing the
distribution of ? T y as a truncated normal. The key step to deriving this pivot is the following lemma:
Lemma 5.1. The conditioning set can be rewritten in terms of ? T y as follows:
{Ay ? b} = {V ? (y) ? ? T y ? V + (y), V 0 (y) ? 0}
1
b can be taken to be 0 for marginal screening, but this extra generality is needed for other model selection
methods.
2
It is also possible to use a coarser partition, where each element of the partition only corresponds to a
subset of variables S. See [12] for details.
3
A contrast of y is a linear function of the form ? T y.
4
where
?=
A??
? T ??
(6)
bj ? (Ay)j + ?j ? T y
j: ?j <0
?j
(7)
bj ? (Ay)j + ?j ? T y
.
j: ?j >0
?j
(8)
V ? = V ? (y) = max
V + = V + (y) = min
V 0 = V 0 (y) = min bj ? (Ay)j
(9)
j: ?j =0
Moreover, (V + , V ? , V 0 ) are independent of ? T y.
[a,b]
Theorem 5.2. Let ?(x) denote the CDF of a N (0, 1) random variable, and let F?,?2 denote the
CDF of T N (?, ?, a, b), i.e.:
?((x ? ?)/?) ? ?((a ? ?)/?)
[a,b]
F?,?2 (x) =
.
(10)
?((b ? ?)/?) ? ?((a ? ?)/?)
[V ? ,V + ]
Then F?T ?, ?T ?? (? T y) is a pivotal quantity, conditional on {Ay ? b}:
[V ? ,V + ]
F?T ?, ?T ?? (? T y) {Ay ? b} ? Unif(0, 1)
where V
?
(11)
+
and V are defined in (7) and (8).
300
1
250
0.8
200
0.6
150
0.4
100
0.2
50
0
0
0.2
0.4
0.6
0.8
0
1
empirical cdf
Unif(0,1) cdf
0
0.2
0.4
0.6
0.8
1
[V ? ,V + ]
Figure 2: Histogram and qq plot of F?T ?, ?T ?? (? T y) where y is a constrained Gaussian. The
distribution is very close to Unif(0, 1), which is in agreement with Theorem 5.2.
6
Inference for marginal screening
In this section, we apply the theory summarized in Sections 4 and 5 to marginal screening. In
particular, we will construct confidence intervals for the selected variables.
To summarize the developments so far, recall that our model (1) says that y ? N (?, ? 2 I).
?
The distribution of interest is y|{E(y)
= E}, and by Theorem 4.1, this is equivalent to
y|{A(S, s)z ? b(S, s)}, where y ? N (?, ? 2 I). By applying Theorem 5.2, we obtain the pivotal
quantity
[V ? ,V + ]
?
F?T ?, ?2 ||?||2 (? T y) {E(y)
= E} ? Unif(0, 1)
(12)
2
for any ?, where V ? and V + are defined in (7) and (8).
In this section, we describe how to form confidence intervals for the components of ?S?? =
(XST? XS? )?1 XST? ?. The best linear predictor of ? that uses only the selected variables is ?S?? , and
?? ? = (X T X ? )?1 X T y is an unbiased estimate of ? ? . If we choose
S
?
S
S
?
S
?j =
?
S
T
?1 T
((XS? XS? ) XS? ej )T ,
5
(13)
?
th
then ?jT ? = ?j?
? , so the above framework provides a method for inference about the j variable in
S
?
the model S.
6.1
Confidence intervals for selected variables
?
Next, we discuss how to obtain confidence intervals for ?j?
The standard way
?.
S
to
obtain an interval is to invert a pivotal quantity [5].
In other words, since
[V ? ,V + ]
? = E} = ? one can define a (1 ? ?) (conditional)
Pr ?2 ? F? ? , ?2 ||?j ||2 (?jT y) ? 1 ? ?2 {E
?
j?S
?
confidence interval for ?j,
? as
E
n
?
?o
[V ? ,V + ]
.
x : ? Fx, ?2 ||?j ||2 (?jT y) ? 1 ?
2
2
(14)
In fact, F is monotone decreasing in x, so to find its endpoints, one need only solve for the root of a
smooth one-dimensional function. The monotonicity is a consequence of the fact that the truncated
Gaussian distribution is a natural exponential family and hence has monotone likelihood ratio in ?
[17].
We now formalize the above observations in the following result, an immediate consequence of
Theorem 5.2.
? s?)) and U? = U? (?j , (S,
? s?))
Corollary 6.1. Let ?j be defined as in (13), and let L? = L? (?j , (S,
be the (unique) values satisfying
?
?
[V ? ,V + ]
[V ? ,V + ]
FL? , ?2 ||?j ||2 (?jT y) = 1 ?
FU? , ?2 ||?j ||2 (?jT y) =
(15)
2
2
?
?
Then [L? , U? ] is a (1 ? ?) confidence interval for ?j?
? , conditional on E:
S
?
?
(16)
P ?j?
? ? [L? , U? ] {E = E} = 1 ? ?.
S
?
?
Proof. The confidence region of ?j?
? is the set of ?j such that the test of H0 : ?j?S
? accepts at the
S
[V ? ,V + ]
1 ? ? level. The function Fx, ?2 ||?j ||2 (?jT y) is monotone in x, so solving for L? and U? identify the
most extreme values where H0 is still accepted. This gives a 1 ? ? confidence interval.
Next, we establish the unconditional coverage of the constructed confidence intervals and the false
coverage rate (FCR) control [1].
?
Corollary 6.2. For each j ? S,
?
j
j
Pr ?j?
(17)
? ? [L? , U? ] = 1 ? ?.
S
j j
Furthermore, the FCR of the intervals [L? , U? ] j?E? is ?.
Proof. By (16), the conditional coverage of the confidence intervals are 1 ? ?. The coverage holds
?
for every element of the partition {E(y)
= E}, so
X
? = E} Pr(E
? = E)
Pr ? ? ? ? [L? , U? ] {E
Pr ? ? ? ? [Lj , U j ] =
j?S
?
?
j?S
E
X
? = E) = 1 ? ?.
=
(1 ? ?) Pr(E
E
Remark 6.3. We would like to emphasize that the previous Corollary shows that the constructed
?
confidence intervals are unconditionally valid. The conditioning on the selection event E(y)
=E
was only for mathematical convenience to work out the exact pivot. Unlike standard z-test intervals,
?
the coverage target, ?j?
? , and the interval [L? , U? ] are random. In a typical confidence interval
S
only the interval is random; however in the post-selection inference setting, the selected model is
random, so both the interval and the target are necessarily random [2].
We summarize the algorithm for selecting and constructing confidence intervals below.
6
Algorithm 2 Confidence intervals for selected variables
1: Input: Design matrix X, response y, model size k.
2: Use Algorithm 1 to select a subset of variables S? and signs s? = sign(X T? y).
? s?) and b = b(S,
? s?) using (5). Let ?j = (X T )? ej .
3: Let A = A(S,
?
S
S
4: Solve for Lj? and U?j using Equation (15) where V ? and V + are computed via (7) and (8) using
the A, b, and ?j previously defined.
?
5: Output: Return the intervals [Lj? , U?j ] for j ? S.
7
Experiments
In Figure 1, we have already seen that the confidence intervals constructed using Algorithm 2 have
exactly 1 ? ? coverage proportion. In this section, we perform two experiments on real data where
the linear model does not hold, the noise is not Gaussian, and the noise variance is unknown.
7.1
Diabetes dataset
The diabetes dataset contains n = 442 diabetes patients measured on p = 10 baseline variables [6].
The baseline variables are age, sex, body mass index, average blood pressure, and six blood serum
measurements, and the response y is a quantitative measure of disease progression measured one
yk
year after the baseline. Since the noise variance ? 2 is unknown, we estimate it by ? 2 = ky??
n?p ,
Coverage Proportion
1
0.8
0.6
Z?test
Adjusted
Nominal
0.4
0.2
0.6
0.8
1??
1
Figure 3: Plot of 1 ? ? vs the coverage proportion for diabetes dataset. The nominal curve is the
line y = x. The coverage proportion of the adjusted intervals agree with the nominal coverage level,
but the z-test coverage proportion is strictly below the nominal level. The adjusted intervals perform
well, despite the noise being non-Gaussian, and ? 2 unknown.
where y? = X ?? and ?? = (X T X)?1 X T y. For each trial we generated new responses y?i = X ?? + ?,
and ? is bootstrapped from the residuals ri = yi ? y?i . We used marginal screening to select k = 2
variables, and then fit linear regression on the selected variables. The adjusted confidence intervals
were constructed using Algorithm 2 with the estimated ? 2 . The nominal coverage level is varied
across 1 ? ? ? {.5, .6, .7, .8, .9, .95, .99}. From Figure 3, we observe that the adjusted intervals
always cover at the nominal level, whereas the z-test is always below. The experiment was repeated
2000 times.
7.2
Riboflavin dataset
Our second data example is a high-throughput genomic dataset about riboflavin (vitamin B2) production rate [3]. There are p = 4088 variables which measure the log expression level of different
genes, a single real-valued response y which measures the logarithm of the riboflavin production
rate, and n = 71 samples. We first estimate ? 2 by using cross-validation [20], and apply marginal
screening with k = 30, as chosen in [3]. We then use Algorithm 2 to identify genes significant at
7
? = 10%. The genes identified as significant were YCKE_at, YOAB_at, and YURQ_at. After
using Bonferroni to control for FWER, we found YOAB_at remained significant.
8
Extensions
The purpose of this section is to illustrate the broad applicability of the condition on selection framework. For expository purposes, we focused the paper on marginal screening where the framework
is particularly easy to understand. In the rest of this section, we show how to apply the framework
to marginal screening+Lasso, and orthogonal matching pursuit. This is a non-exhaustive list of
selection procedures where the condition on selection framework is applicable, but we hope this incomplete list emphasizes the ease of constructing tests and confidence intervals post-model selection
via conditioning.
8.1
Marginal screening + Lasso
The marginal screening+Lasso procedure was introduced in [7] as a variable selection method for
k
the ultra-high dimensional setting of p = O(en ). Fan et al. [7] recommend applying the marginal
screening algorithm with k = n ? 1, followed by the Lasso on the selected variables. This is a
two-stage procedure, so to properly account for the selection we must encode the selection event
of marginal screening followed by Lasso. This can be done by representing the two stage selection
as a single event. Let (S?m , s?m ) be the variables and signs selected by marginal screening, and the
(S?L , z?L ) be the variables and signs selected by Lasso [12]. In Proposition 2.2 of [12], it is shown
how to encode the Lasso selection event (S?L , z?L ) as a set of constraints {AL y ? bL } 4 , and in
Section 4 we showed how to encode the marginal screening selection event (S?m , s?m ) as a set of
constraints {Am y ? bm }. Thus the selection event of marginal screening+Lasso can be encoded
as {AL y ? bL , Am y ? bm }. Using these constraints, the hypothesis test and confidence intervals
described in Algorithm 2 are valid for marginal screening+Lasso.
8.2
Orthogonal Matching Pursuit
Orthogonal matching pursuit (OMP) is a commonly used variable selection method. At each iteration, OMP selects the variable most correlated with the residual r, and then recomputes the residual
using the residual of least squares using the selected variables. Similar to Section 4, we can represent
the OMP selection event as a set of linear constraints on y.
?
E(y)
= y : sign(xTpi ri )xTpi ri > ?xTj ri , for all j 6= pi and all i ? [k]
= {y : s?i xTpi (I ? XS?i?1 XS??
i?1
s?i xTpi (I ? XS?i?1 XS??
i?1
)y > ?xTj (I ? XS?i?1 XS??
)y and
i?1
)y > 0, for all j 6= pi , and all i ? [k] }
The selection event encodes that OMP selected a certain variable and the sign of the correlation of
that variable with the residual, at steps 1 to k. The primary difference between the OMP selection
event and the marginal screening selection event is that the OMP event also describes the order at
which the variables were chosen.
9
Conclusion
Due to the increasing size of datasets, marginal screening has become an important method for
fast variable selection. However, the standard hypothesis tests and confidence intervals used in
linear regression are invalid after using marginal screening to select important variables. We have
described a method to form confidence intervals after marginal screening. The condition on selection
framework is not restricted to marginal screening, and also applies to OMP and marginal screening
+ Lasso. The supplementary material also discusses the framework applied to non-negative least
squares.
4
The Lasso selection event is with respect to the Lasso optimization problem after marginal screening.
8
References
[1] Yoav Benjamini and Daniel Yekutieli. False discovery rate?adjusted multiple confidence intervals for
selected parameters. Journal of the American Statistical Association, 100(469):71?81, 2005.
[2] Richard Berk, Lawrence Brown, Andreas Buja, Kai Zhang, and Linda Zhao. Valid post-selection inference. Annals of Statistics, 41(2):802?837, 2013.
[3] Peter B?hlmann, Markus Kalisch, and Lukas Meier. High-dimensional statistics with a view toward
applications in biology. Statistics, 1, 2014.
[4] Peter Lukas B?hlmann and Sara A van de Geer. Statistics for High-dimensional Data. Springer, 2011.
[5] George Casella and Roger L Berger. Statistical inference, volume 70. Duxbury Press Belmont, CA, 1990.
[6] Bradley Efron, Trevor Hastie, Iain Johnstone, and Robert Tibshirani. Least angle regression. The Annals
of statistics, 32(2):407?499, 2004.
[7] Jianqing Fan and Jinchi Lv. Sure independence screening for ultrahigh dimensional feature space. Journal
of the Royal Statistical Society: Series B (Statistical Methodology), 70(5):849?911, 2008.
[8] Christopher R Genovese, Jiashun Jin, Larry Wasserman, and Zhigang Yao. A comparison of the lasso
and marginal regression. The Journal of Machine Learning Research, 98888:2107?2143, 2012.
[9] Isabelle Guyon and Andr? Elisseeff. An introduction to variable and feature selection. The Journal of
Machine Learning Research, 3:1157?1182, 2003.
[10] Adel Javanmard and Andrea Montanari. Confidence intervals and hypothesis testing for high-dimensional
regression. arXiv preprint arXiv:1306.3171, 2013.
[11] Jason Lee, Yuekai Sun, and Jonathan E Taylor. On model selection consistency of penalized m-estimators:
a geometric theory. In Advances in Neural Information Processing Systems, pages 342?350, 2013.
[12] Jason D Lee, Dennis L Sun, Yuekai Sun, and Jonathan E Taylor. Exact inference after model selection
via the lasso. arXiv preprint arXiv:1311.6238, 2013.
[13] Hannes Leeb and Benedikt M P?tscher. The finite-sample distribution of post-model-selection estimators
and uniform versus nonuniform approximations. Econometric Theory, 19(1):100?142, 2003.
[14] Hannes Leeb and Benedikt M P?tscher. Model selection and inference: Facts and fiction. Econometric
Theory, 21(1):21?59, 2005.
[15] Hannes Leeb and Benedikt M P?tscher. Can one estimate the conditional distribution of post-modelselection estimators? The Annals of Statistics, pages 2554?2591, 2006.
[16] Jeff Leek.
Prediction:
the lasso vs just using the top 10 predictors.
http://simplystatistics.tumblr.com/post/18132467723/
prediction-the-lasso-vs-just-using-the-top-10.
[17] Erich L. Lehmann and Joseph P. Romano. Testing Statistical Hypotheses. Springer, 3 edition, 2005.
[18] Nicolai Meinshausen, Lukas Meier, and Peter B?hlmann. P-values for high-dimensional regression. Journal of the American Statistical Association, 104(488), 2009.
[19] Sahand N Negahban, Pradeep Ravikumar, Martin J Wainwright, and Bin Yu. A unified framework
for high-dimensional analysis of m-estimators with decomposable regularizers. Statistical Science,
27(4):538?557, 2012.
[20] Stephen Reid, Robert Tibshirani, and Jerome Friedman. A study of error variance estimation in lasso
regression. arXiv preprint arXiv:1311.5274, 2013.
[21] Virginia Goss Tusher, Robert Tibshirani, and Gilbert Chu. Significance analysis of microarrays applied
to the ionizing radiation response. Proceedings of the National Academy of Sciences, 98(9):5116?5121,
2001.
[22] Sara van de Geer, Peter B?hlmann, and Ya?acov Ritov. On asymptotically optimal confidence regions and
tests for high-dimensional models. arXiv preprint arXiv:1303.0518, 2013.
[23] M.J. Wainwright. Sharp thresholds for high-dimensional and noisy sparsity recovery using `1 -constrained
quadratic programming (lasso). 55(5):2183?2202, 2009.
[24] Larry Wasserman and Kathryn Roeder.
37(5A):2178, 2009.
High dimensional variable selection.
Annals of statistics,
[25] Cun-Hui Zhang and S Zhang. Confidence intervals for low-dimensional parameters with highdimensional data. arXiv preprint arXiv:1110.2563, 2011.
[26] P. Zhao and B. Yu. On model selection consistency of lasso. 7:2541?2563, 2006.
9
| 5292 |@word trial:2 norm:1 proportion:9 sex:1 unif:4 elisseeff:1 pressure:1 contains:1 series:1 selecting:1 daniel:1 bootstrapped:1 bradley:1 com:1 nicolai:1 chu:1 must:1 belmont:1 partition:4 plot:3 v:3 greedy:2 selected:17 tscher:3 core:1 jiashun:1 provides:1 bijection:2 zhang:3 mathematical:2 constructed:7 become:1 javanmard:1 andrea:1 decomposed:1 decreasing:1 xti:4 increasing:1 notation:1 moreover:1 mass:1 linda:1 developed:1 unified:1 thorough:1 every:1 quantitative:1 exactly:1 control:2 unit:1 kalisch:1 reid:1 negligible:2 engineering:1 consequence:2 despite:1 meinshausen:1 sara:2 ease:1 limited:1 range:2 unique:1 testing:5 procedure:12 j0:3 empirical:1 matching:8 confidence:38 word:1 cannot:1 close:2 selection:66 convenience:1 applying:2 gilbert:1 equivalent:1 serum:1 go:1 focused:1 decomposable:1 splitting:1 recovery:1 wasserman:2 estimator:11 iain:1 submodel:1 deriving:1 fx:2 qq:1 annals:4 controlling:1 nominal:8 target:2 exact:9 programming:1 us:4 kathryn:1 hypothesis:7 agreement:1 diabetes:4 element:3 approximated:1 particularly:2 expensive:1 satisfying:1 distributional:2 coarser:1 preprint:5 modelselection:1 region:2 sun:3 highest:1 yk:1 disease:1 solving:1 represented:1 recomputes:1 describe:2 fast:1 h0:3 exhaustive:1 encoded:1 stanford:6 solve:2 valued:1 say:1 supplementary:1 kai:1 statistic:10 benedikt:3 noisy:1 eigenvalue:2 product:1 academy:1 ky:1 etj:1 requirement:1 leave:1 illustrate:3 develop:2 derive:1 radiation:1 measured:2 coverage:19 drawback:1 correct:2 closely:1 larry:2 material:1 bin:1 require:2 transparent:1 investigation:1 ultra:1 proposition:1 underdetermined:1 adjusted:9 extension:4 strictly:1 correction:1 hold:4 normal:3 lawrence:1 bj:3 major:1 purpose:2 estimation:1 applicable:3 currently:1 largest:1 establishes:1 tool:2 hope:1 genomic:1 gaussian:10 always:3 ej:2 corollary:3 encode:3 focus:5 properly:1 likelihood:1 contrast:3 baseline:3 sense:1 am:2 inference:13 roeder:1 familywise:1 lj:3 selects:3 development:1 constrained:6 marginal:51 construct:4 biology:1 represents:1 broad:1 yu:2 throughput:1 genovese:1 future:1 np:1 recommend:1 richard:1 national:1 comprehensive:1 xtj:4 friedman:1 screening:50 interest:1 predominant:1 introduces:1 extreme:1 pradeep:1 unconditional:1 regularizers:1 fu:1 orthogonal:8 incomplete:1 taylor:4 logarithm:1 theoretical:2 column:1 cover:5 yoav:1 hlmann:4 cost:2 applicability:4 subset:4 entry:1 snr:4 predictor:3 uniform:2 virginia:1 characterize:1 chooses:1 negahban:1 lee:3 yao:1 posi:2 choose:1 american:2 zhao:2 return:1 account:2 de:2 summarized:1 b2:1 coefficient:6 depends:2 root:1 jason:3 view:1 characterizes:1 complicated:1 contribution:1 square:4 variance:3 identify:2 emphasizes:1 casella:1 trevor:1 evaluates:1 failure:1 proof:5 dataset:5 recall:1 efron:1 formalize:1 methodology:4 response:11 hannes:3 ritov:1 done:1 generality:1 furthermore:4 roger:1 stage:2 just:2 correlation:1 jerome:1 dennis:1 christopher:1 nonlinear:1 brown:1 true:1 unbiased:1 counterpart:1 regularization:2 hence:1 yekutieli:1 interchangeably:1 bonferroni:1 noted:1 fwer:1 ay:6 meaning:2 empirically:1 conditioning:4 endpoint:1 volume:1 extend:1 jdl17:1 association:2 refer:1 measurement:1 significant:3 isabelle:1 consistency:3 erich:1 benjamini:1 dot:1 longer:2 recent:5 showed:1 certain:1 jianqing:1 yi:3 seen:1 george:1 omp:8 signal:1 stephen:1 multiple:1 yuekai:2 smooth:1 faster:1 vitamin:1 cross:1 post:12 ravikumar:1 prediction:2 regression:18 patient:1 arxiv:10 histogram:1 represent:4 iteration:1 invert:1 c1:1 whereas:1 interval:56 xst:13 extra:1 operate:1 unlike:1 rest:1 sure:1 easy:1 xj:1 bic:1 fit:2 independence:1 hastie:1 lasso:34 identified:1 andreas:1 microarrays:1 intensive:1 pivot:3 six:1 expression:1 sahand:1 adel:1 penalty:1 peter:4 jj:2 remark:1 romano:1 simplest:2 http:1 andr:1 fiction:1 sign:9 estimated:1 tibshirani:3 broadly:1 key:1 threshold:1 blood:2 clean:3 econometric:2 asymptotically:1 subgradient:1 monotone:3 year:1 inverse:1 angle:1 lehmann:1 extends:1 laid:1 reader:1 family:1 guyon:1 utilizes:1 summarizes:1 comparable:1 fl:1 followed:3 aic:1 fan:2 quadratic:1 constraint:6 ri:4 encodes:1 markus:1 mestimators:1 aspect:1 extremely:4 min:3 martin:1 department:1 expository:1 across:2 describes:1 partitioned:1 joseph:1 cun:1 making:1 restricted:2 pr:9 indexing:1 advertised:1 taken:1 computationally:2 equation:2 agree:1 previously:2 discus:5 fail:1 leek:1 needed:1 unusual:1 pursuit:8 rewritten:1 apply:6 progression:1 observe:1 duxbury:1 tumblr:1 top:2 subsampling:1 log10:1 restrictive:1 establish:2 classical:1 society:1 bl:2 already:1 quantity:5 strategy:1 primary:3 usual:1 separate:1 toward:1 index:2 berger:1 illustration:1 ratio:1 robert:3 stated:1 negative:1 design:5 unknown:4 perform:4 observation:2 datasets:4 finite:2 jin:1 truncated:4 immediate:1 rn:4 varied:1 nonuniform:1 arbitrary:1 sharp:1 buja:1 introduced:1 inverting:1 pair:2 meier:2 z1:2 accepts:1 acov:1 able:1 usually:1 below:4 jinchi:1 sparsity:1 summarize:2 including:6 max:1 royal:1 wainwright:2 suitable:2 event:23 difficulty:1 natural:1 regularized:1 residual:5 normality:1 representing:2 unconditionally:2 literature:1 discovery:1 geometric:1 asymptotic:3 ultrahigh:1 versus:1 remarkable:1 lv:1 age:1 validation:1 consistent:1 pi:2 production:2 penalized:2 understand:2 johnstone:1 wide:1 characterizing:1 lukas:3 absolute:1 sparse:1 van:2 curve:1 gram:1 valid:8 forward:2 commonly:3 author:1 bm:2 far:2 approximate:1 emphasize:1 gene:3 monotonicity:1 kkt:1 xi:3 ca:3 obtaining:3 necessarily:1 constructing:3 significance:2 main:1 montanari:1 noise:4 edition:1 repeated:1 pivotal:5 body:1 en:1 screen:1 position:2 exponential:1 theorem:6 remained:1 specific:1 jt:6 list:2 x:17 intractable:1 stepwise:2 false:2 hui:1 magnitude:1 applies:1 springer:2 corresponds:2 relies:1 cdf:4 conditional:11 goal:1 invalid:1 jeff:1 typical:1 corrected:2 lemma:2 berk:1 geer:2 accepted:1 leeb:3 ya:1 select:3 highdimensional:1 jonathan:4 correlated:2 |
4,741 | 5,293 | On Iterative Hard Thresholding Methods for
High-dimensional M-Estimation
Prateek Jain?
Ambuj Tewari?
Purushottam Kar?
Microsoft Research, INDIA
?
University of Michigan, Ann Arbor, USA
{prajain,t-purkar}@microsoft.com, [email protected]
?
Abstract
The use of M-estimators in generalized linear regression models in high dimensional settings requires risk minimization with hard L0 constraints. Of the known
methods, the class of projected gradient descent (also known as iterative hard
thresholding (IHT)) methods is known to offer the fastest and most scalable solutions. However, the current state-of-the-art is only able to analyze these methods
in extremely restrictive settings which do not hold in high dimensional statistical models. In this work we bridge this gap by providing the first analysis for
IHT-style methods in the high dimensional statistical setting. Our bounds are tight
and match known minimax lower bounds. Our results rely on a general analysis
framework that enables us to analyze several popular hard thresholding style algorithms (such as HTP, CoSaMP, SP) in the high dimensional regression setting.
Finally, we extend our analysis to the problem of low-rank matrix recovery.
1
Introduction
Modern statistical estimation is routinely faced with real world problems where the number of parameters p handily outnumbers the number of observations n. In general, consistent estimation of
parameters is not possible in such a situation. Consequently, a rich line of work has focused on
models that satisfy special structural assumptions such as sparsity or low-rank structure. Under
these assumptions, several works (for example, see [1, 2, 3, 4, 5]) have established that consistent
estimation is information theoretically possible in the ?n p? regime as well.
The question of efficient estimation, however, is faced with feasibility issues since consistent estimation routines often end-up solving NP-hard problems. Examples include sparse regression which
requires loss minimization with sparsity constraints and low-rank regression which requires dealing
with rank constraints which are not efficiently solvable in general [6].
Interestingly, recent works have demonstrated that these hardness results can be avoided by assuming
certain natural conditions over the loss function being minimized such as restricted strong convexity
(RSC) and restricted strong smoothness (RSS). The estimation routines proposed in these works
typically make use of convex relaxations [5] or greedy methods [7, 8, 9] which do not suffer from
infeasibility issues.
Despite this, certain limitations have precluded widespread use of these techniques. Convex
relaxation-based methods typically suffer from slow rates as they solve non-smooth optimization
problems apart from being hard to analyze in terms of global guarantees. Greedy methods, on the
other hand, are slow in situations with non-negligible sparsity or relatively high rank, owing to their
incremental approach of adding/removing individual support elements.
Instead, the methods of choice for practical applications are actually projected gradient (PGD) methods, also referred to as iterative hard thresholding (IHT) methods. These methods directly project
1
the gradient descent update onto the underlying (non-convex) feasible set. This projection can be
performed efficiently for several interesting structures such as sparsity and low rank. However, traditional PGD analyses for convex problems viz. [10] do not apply to these techniques due to the
non-convex structure of the problem.
An exception to this is the recent work [11] that demonstrates that PGD with non-convex regularization can offer consistent estimates for certain high-dimensional problems. However, the work in [11]
is only able to analyze penalties such as SCAD, MCP and capped L1 . Moreover, their framework
cannot handle commonly used penalties such as L0 or low-rank constraints.
Insufficiency of RIP based Guarantees for M-estimation. As noted above, PGD/IHT-style methods have been very popular in literature for sparse recovery and several algorithms including Iterative
Hard Thresholding (IHT) [12] or GraDeS [13], Hard Thresholding Pursuit (HTP) [14], CoSaMP
[15], Subspace Pursuit (SP) [16], and OMPR(`) [17] have been proposed. However, the analysis
of these algorithms has traditionally been restricted to settings that satisfy the Restricted Isometry
property (RIP) or incoherence property. As the discussion below demonstrates, this renders these
analyses inaccessible to high-dimensional statistical estimation problems.
All existing results analyzing these methods require the condition number of the loss function, restricted to sparse vectors, to be smaller than a universal constant. The best known such constant is
due to the work of [17] that requires a bound on the RIP constant ?2k ? 0.5 (or equivalently a bound
1+?2k
1??2k ? 3 on the condition number). In contrast, real-life high dimensional statistical settings,
wherein pairs of variables can be arbitrarily correlated, routinely require estimation methods to performunder arbitrarily
large condition numbers. In particular if two variates have a covariance matrix
1
1?
like
, then the restricted condition number (on a support set of size just 2) of the sam1?
1
ple matrix cannot be brought down below 1/ even with infinitely many samples. In particular when
< 1/6, none of the existing results for hard thresholding methods offer any guarantees. Moreover,
most of these analyses consider only the least squares objective. Although recent attempts have
been made to extend this to general differentiable objectives [18, 19], the results continue to require
that the restricted condition number be less than a universal constant and remain unsatisfactory in a
statistical setting.
Overview of Results. Our main contribution in this work is an analysis of PGD/IHT-style methods
in statistical settings. Our bounds are tight, achieve known minmax lower bounds [20], and hold
for arbitrary differentiable, possibly even non-convex functions. Our results hold even when the
underlying condition number is arbitrarily large and only require the function to satisfy RSC/RSS
conditions. In particular, this reveals that these iterative methods are indeed applicable to statistical
settings, a result that escaped all previous works.
Our first result shows that the PGD/IHT methods achieve global convergence if used with a relaxed
projection step. More formally, if the optimal parameter is s? -sparse and the problem satisfies
RSC and RSS constraints ? and L respectively (see Section 2), then PGD methods offer global
convergence so long as they employ projection to an s-sparse set where s ? 4(L/?)2 s? . This
gives convergence rates that are identical to those of convex relaxation and greedy methods for the
Gaussian sparse linear model. We then move to a family of efficient ?fully corrective? methods and
show as before, that for arbitrary functions satisfying the RSC/RSS properties, these methods offer
global convergence.
Next, we show that these results allow PGD-style methods to offer global convergence in a variety
of statistical estimation problems such as sparse linear regression and low rank matrix regression.
Our results effortlessly extend to the noisy setting as a corollary and give bounds similar to those of
[21] that relies on solving an L1 regularized problem.
Our proofs are able to exploit that even though hard-thresholding is not the prox-operator for any
convex prox function, it still provides strong contraction when projection is performed onto sets of
sparsity s s? . This crucial observation allows us to provide the first unified analysis for hard
thresholding based gradient descent algorithms. Our empirical results confirm our predictions with
respect to the recovery properties of IHT-style algorithms on badly-conditioned sparse recovery
problems, as well as demonstrate that these methods can be orders of magnitudes faster than their
L1 and greedy counterparts.
2
Organization. Section 2 sets the notation and the problem statement. Section 3 introduces the
PGD/IHT algorithm that we study and proves that the method guarantees recovery assuming the
RSC/RSS property. We also generalize our guarantees to the problem of low-rank matrix regression.
Section 4 then provides crisp sample complexity bounds and statistical guarantees for the PGD/IHT
estimators. Section 5 extends our analysis to a broad family of compressive sensing algorithms that
include the so-called fully-corrective hard thresholding methods and provide similar results for them
as well. We present some empirical results in Section 6 and conclude in Section 7.
2
Problem Setup and Notations
High-dimensional Sparse Estimation. Given data points X = [X1 , . . . , Xn ]T , where Xi ? Rp ,
and the target Y = [Y1 , . . . , Yn ]T , where Yi ? R, the goal is to compute an s? -sparse ? ? s.t.,
? ? = arg
min
?,k?k0 ?s?
f (?).
(1)
P
Typically, f can be thought of as an empirical risk function i.e. f (?) = n1 i `(hXi , ?i, Yi ) for some
loss function ` (see examples in Section 4). However, for our analysis of PGD and other algorithms,
we need not assume any other property of f other than differentiability and the following two RSC
and RSS properties.
Definition 1 (RSC Property). A differentiable function f : Rp ? R is said to satisfy restricted
strong convexity (RSC) at sparsity level s = s1 + s2 with strong convexity constraint ?s if the
following holds for all ?1 , ?2 s.t. k?1 k0 ? s1 and k?2 k0 ? s2 :
?s
f (?1 ) ? f (?2 ) ? h?1 ? ?2 , ?? f (?2 )i +
k?1 ? ?2 k22 .
2
Definition 2 (RSS Property). A differentiable function f : Rp ? R is said to satisfy restricted
strong smoothness (RSS) at sparsity level s = s1 + s2 with strong convexity constraint Ls if the
following holds for all ?1 , ?2 s.t. k?1 k0 ? s1 and k?2 k0 ? s2 :
f (?1 ) ? f (?2 ) ? h?1 ? ?2 , ?? f (?2 )i +
Ls
k?1 ? ?2 k22 .
2
Low-rank Matrix Regression. Low-rank matrix regression is similar to sparse estimation as presented above except that each data point is now a matrix i.e. Xi ? Rp1 ?p2 , the goal being to estimate
a low-rank matrix W ? Rp1 ?p2 that minimizes the empirical loss function on the given data.
W ? = arg
min
W,rank(W )?r
f (W ).
(2)
For this problem the RSC and RSS properties for f are defined similarly as in Definition 1, 2 except
that the L0 norm is replaced by the rank function.
3
Iterative Hard-thresholding Method
In this section we study the popular projected gradient descent (a.k.a iterative hard thresholding)
method for the case of the feasible set being the set of sparse vectors (see Algorithm 1 for pseudocode). The projection operator Ps (z), can be implemented efficiently in this case by projecting
z onto the set of s-sparse vectors by selecting the s largest elements (in magnitude) of z. The standard projection property implies that kPs (z) ? zk22 ? k? 0 ? zk22 for all k? 0 k0 ? s. However, it
turns out that we can prove a significantly stronger property of hard thresholding for the case when
k? 0 k0 ? s? and s? s. This property is key to analysing IHT and is formalized below.
Lemma 1. For any index set I, any z ? RI , let ? = Ps (z). Then for any ? ? ? RI such that
k? ? k0 ? s? , we have
|I| ? s ?
k? ? zk22 .
k? ? zk22 ?
|I| ? s?
See Appendix A for a detailed proof.
Our analysis combines the above observation with the RSC/RSS properties of f to provide geometric
convergence rates for the IHT procedure below.
3
Algorithm 1 Iterative Hard-thresholding
1: Input: Function f with gradient oracle, sparsity level s, step-size ?
2: ? 1 = 0, t = 1
3: while not converged do
4:
? t+1 = Ps (? t ? ??? f (? t )), t = t + 1
5: end while
6: Output: ? t
Theorem 1. Let f have RSC and RSS parameters given by L2s+s? (f ) = L and ?2s+s? (f ) = ?
2 ?
2
. Also let ? ? =
s and ? = 3L
respectively. Let Algorithm 1 be invoked with f , s ? 32 L
?
0
f (? )
arg min?,k?k0 ?s? f (?). Then, the ? -th iterate of Algorithm 1, for ? = O( L
? ? log( )) satisfies:
f (? ? ) ? f (? ? ) ? .
Proof. (Sketch) Let S t = supp(? t ), S ? = supp(? ? ), S t+1 = supp(? t+1 ) and I t = S ? ?S t ?S t+1 .
Using the RSS property and the fact that supp(? t ) ? I t and supp(? t+1 ) ? I t , we have:
L
f (? t+1 ) ? f (? t ) ? h? t+1 ? ? t , g t i + k? t+1 ? ? t k22 ,
2
L t+1
2
1
= k?I t ? ?It t +
? g t t k2 ?
kg t t k2 ,
2
3L I 2 2L I 2
?1 L
|I t | ? s
1
1
? ? t
? k?I?t ? ?It t + ? gIt t k22 ?
(kgIt t \(S t ?S ? ) k22 + kgSt t ?S ? k22 ),
?
2 |I | ? s
L
2L
(3)
where ?1 follows from an application of Lemma 1 with I = I t and the Pythagoras theorem. The
above equation has three critical terms. The first term can be bounded using the RSS condition.
1
kgSt t ?S ? k22 bounds the third term
Using f (? t ) ? f (? ? ) ? hgSt t ?S ? , ? t ? ? ? i ? ?2 k? t ? ? ? k22 ? 2?
in (3). The second term is more interesting as in general elements of gSt ? can be arbitrarily small.
However, elements of gIt t \(S t ?S ? ) should be at least as large as gSt ? \S t+1 as they are selected by
hard-thresholding. Combining this insight with bounds for gSt ? \S t+1 and with (3), we obtain the
theorem. See Appendix A for a detailed proof.
3.1
Low-rank Matrix Regression
We now generalize our previous analysis to a projected gradient descent (PGD) method for low-rank
matrix regression. Formally, we study the following problem:
min f (W ), s.t., rank(W ) ? s.
(4)
W
The hard-thresholding projection step for low-rank matrices can be solved using SVD i.e.
P Ms (W ) = Us ?s VsT ,
where W = U ?V is the singular value decomposition of W . Us , Vs are the top-s singular vectors
(left and right, respectively) of W and ?s is the diagonal matrix of the top-s singular values of W .
To proceed, we first note a property of the above projection similar to Lemma 1.
Lemma 2. Let W ? Rp1 ?p2 be a rank-|I t | matrix and let p1 ? p2 . Then for any rank-s? matrix
W ? ? Rp1 ?p2 we have
|I t | ? s
kP Ms (W ) ? W k2F ? t
kW ? ? W k2F .
(5)
|I | ? s?
T
Proof. Let W = U ?V T be the singular value decomposition of W . Now, kP Ms (W ) ? W k2F =
P|I t |
2
2
i=s+1 ?i = kPs (diag(?)) ? diag(?)k2 , where ?1 ? ? ? ? ? ?|I t | ? 0 are the singular values of
W . Using Lemma 1, we get:
|I t | ? s
|I t | ? s
?
2
kP Ms (W ) ? W k2F ? t
k?
?
diag(?)k
?
kW ? ? W k2F ,
(6)
2
|I | ? s?
|I t | ? s?
P
where the last step uses the von Neumann?s trace inequality (T r(A ? B) ? i ?i (A)?i (B)).
4
The following result for low-rank matrix regression immediately follows from Lemma 4.
Theorem 2. Let f have RSC and RSS parameters given by L2s+s? (f ) = L and ?2s+s? (f ) = ?.
Replace the projection operator Ps in Algorithm 1 with its matrix counterpart P Ms as defined in (5).
2 ?
2
s , ? = 3L
. Also let W ? = arg minW,rank(W )?s? f (W ).
Suppose we invoke it with f, s ? 32 L
?
0
f (W )
Then the ? -th iterate of Algorithm 1, for ? = O( L
) satisfies:
? ? log(
f (W ? ) ? f (W ? ) ? .
Proof. A proof progression similar to that of Theorem 1 suffices. The only changes that need to be
made are: firstly Lemma 2 has to be invoked in place of Lemma 1. Secondly, in place of considering vectors restricted to a subset of coordinates viz. ?S , gIt , we would need to consider matrices
restricted to subspaces i.e. WS = US UST W where US is a set of singular vectors spanning the
range-space of S.
4
High Dimensional Statistical Estimation
This section elaborates on how the results of the previous section can be used to give guarantees for
IHT-style techniques in a variety of statistical estimation problems. We will first present a generic
convergence result and then specialize it to various settings. Suppose we have a sample of data
points Z1:n and a loss function L(?; Z1:n ) that depends on a parameter ? and the sample. Then we
can show the following result. (See Appendix B for a proof.)
Theorem 3. Let ?? be any s? -sparse vector. Suppose L(?; Z1:n ) is differentiable and satisfies RSC and RSS at sparsity level s + s? with parameters ?s+s? and Ls+s? respectively, for
2
L2s+s?
s ? 32 ?2s+s
s? . Let ? ? be the ? -th iterate of Algorithm 1 for ? chosen as in Theorem 1
?
and ? be the function value error incurred by Algorithm 1. Then we have
s
?
? Z1:n )k?
2
2 s + s? k?L(?;
?
?
k? ? ? k2 ?
+
.
?s+s?
?s+s?
Note that the result does not require the loss function to be convex. This fact will be crucially used
later. We now apply the above result to several statistical estimation scenarios.
? Xi i + ?i where
Sparse Linear Regression. Here Zi = (Xi , Yi ) ? Rp ? R and Yi = h?,
2
?i ? N (0, ? ) is label noise. The empirical loss is the usual least squares loss i.e. L(?; Z1:n ) =
1
2
n kY ? X?k2 . Suppose X1:n are drawn i.i.d. from a sub-Gaussian distribution with covariance
? with ?jj ? 1 for all j. Then [22, Lemma 6] immediately implies that RSC and RSS at
p
sparsity level k hold, with probability at least 1 ? e?c0 n , with ?k = 12 ?min (?) ? c1 k log
and
n
p
?
Lk = 2?max (?) + c1 k log
(c
,
c
are
universal
constants).
So
we
can
set
k
=
2s
+
s
and
if
0
1
n
n > 4c1 k log p/?min (?) then we have ?k ? 41 ?min (?) and Lk ? 2.25?max (?) which means that
2 ?
Lk /9?k ? ?(?) := ?max (?)/?min (?). Thus it is enough to choose
q s = 2592?(?) s and ap? Z1:n )k? = kX T ?/nk? ? 2? log p with probability at least
ply Theorem 3. Note that k?L(?;
n
1?c2 p?c3 (c2 , c3 are universal constants). Putting everything together, we have the following bound
with high probability:
r
r
?(?)
s? log p
?
k?? ? ? k2 ? 145
?
+2
,
?min (?)
n
?min (?)
where is the function value error incurred by Algorithm 1.
Noisy and Missing Data. We now look at cases with feature noise as well. More specifically,
? i ?s that are corrupted versions of Xi ?s. Two models of noise are
assume that we only have access to X
? i = Xi +Wi where Wi ? N (0, ?W ), and b) (missing
popular in literature [21]: a) (additive noise) X
?
data) X is an R?{?}-valued matrix obtained by independently, with probability ? ? [0, 1), replacing
each entry in X with ?. For the case of additive noise (missing data can be handled similarly),
? i , Yi ) and L(?; Z1:n ) = 1 ? T ??
? ? ?? T ? where ?
?=X
? T X/n
? ? ?W and ?? = X
? T Y /n are
Zi = (X
2
5
Algorithm 2 Two-stage Hard-thresholding
1: Input: function f with gradient oracle, sparsity level s, sparsity expansion level `
2: ? 1 = 0, t = 1
3: while not converged do
4:
g t = ?? f (? t ), S t = supp(? t )
5:
Z t = S t ? (largest ` elements of |gSt t |)
6:
? t = arg min?,supp(?)?Z t f (?)
// fully corrective step
t
t
e
7:
? = Ps (? )
8:
? t+1 = arg min?,supp(?)?supp(?et ) f (?), t = t + 1
// fully corrective step
9: end while
10: Output: ? t
unbiased estimators of ? and ?T ?? respectively. [21, Appendix A, Lemma 1] implies that RSC, RSS
at sparsity level k hold, with failure probability exponentially small in n, with ?k = 12 ?min (?) ?
(k?k2 +k?W k2 )2
op
op
, 1) log p.
k? (p)/n and Lk = 1.5?max (?) + k? (p)/n for ? (p) = c0 ?min (?) max(
2
(?)
?min
Thus for k = 2s + s? and n ? 4k? (p)/?min (?) we have Lk /?k ? 7?(?). Note that L(?; Z1:n )
is non-convex but we can still apply Theorem 3 with s = 1568?(?)2 s? because RSC, RSS hold.
? Z1:n )k? ?
Using thephigh probability upper bound (see [21, Appendix A, Lemma 2]) k?L(?;
?
c1 ?
? k?k2 log p/n gives us the following
r
r
?
?(?)
?
? 2 s log p + 2
k?? ? ? k2 ? c2
?
? k?k
?min (?)
n
?min (?)
q
where ?
? = k?W k2op + k?k2op (k?W kop + ?) and is the function value error in Algorithm 1.
5
Fully-corrective Methods
In this section, we study a variety of ?fully-corrective? methods. These methods keep the optimization objective fully minimized over the support of the current iterate. To this end, we first prove a
fundamental theorem for fully-corrective methods that formalizes the intuition that for such methods, a large function value should imply a large gradient at any sparse ? as well. This result is similar
to Lemma 1 of [17] but holds under RSC/RSS conditions (rather than the RIP condition as in [17]),
as well as for the general loss functions. See Appendix C for a detailed proof.
Lemma 3. Consider a function f with RSC parameter given by L2s+s? (f ) = L and RSS parameter
given by ?2s+s? (f ) = ?. Let ? ? = arg min?,k?k0 ?s? f (?) with S ? = supp(? ? ). Let S t ? [p] be
any subset of co-ordinates s.t. |S t | ? s. Let ? t = arg min?,supp(?)?S t f (?). Then, we have:
2?(f (? t ) ? f (? ? )) ? kgSt t ?S ? k22 ? ?2 k?St t \S ? k22
Two-stage Methods. We will, for now, concentrate on a family of two-stage fully corrective methods that contains popular compressive sensing algorithms like CoSaMP and Subspace Pursuit (see
Algorithm 2 for pseudocode). These algorithms have thus far been analyzed only under RIP conditions for the least squares objective. Using our analysis framework developed in the previous
sections, we present a generic RSC/RSS-based analysis for general two-stage methods for arbitrary
loss functions. Our analysis shall use the following key observation that the the hard thresholding
step in two stage methods does not increase the objective function a lot.
We defer the analysis of partial hard thresholding methods to a later version of the paper. This family
includes the OMPR(`) method [17], which is known to provide the best known RIP guarantees in
the compressive sensing setting. Using our proof techniques, we can show that this method offers
geometric convergence rates in the statistical setting as well.
Lemma 4. Let Zt ? [n] and |Zt | ? q. Let ? t = arg min?,supp(?)?Zt f (?) and ?bt = Pq (? t ).
Then, the following holds:
`
L
? (f (? t ) ? f (? ? )).
f (?bt ) ? f (? t ) ? ?
? s + ` ? s?
6
0
0.1
0.2
0.3
Noise level (sigma)
0.4
(a)
100
50
0
0.5
1
1.5
2
Dimensionality (p)
2.5
2
10
0
HTP
GraDeS
L1
FoBa
10
?2
10
?3
10
0
100
4
x 10
(b)
200 300 400
Sparsity (s*)
500
Support Recovery Error
20
150
10
HTP
GraDeS
L1
FoBa
Runtime (sec)
40
Runtime (sec)
Support Recovery Error
HTP
GraDeS
L1
FoBa
60
0
4
200
80
40
CoSaMP
HTP
GraDeS
30
20
10
0
80
100 120 140 160
Projected Sparsity (s)
(c)
(d)
Figure 1: A comparison of hard thresholding techniques (HTP) and projected gradient methods
(GraDeS) with L1 and greedy methods (FoBa) on sparse noisy linear regression tasks. 1(a) gives
the number of undiscovered elements from supp(? ? ) as label noise levels are increased. 1(b) shows
the variation in running times with increasing dimensionality p. 1(c) gives the variation in running
times (in logscale) when the true sparsity level s? is increased keeping p fixed. HTP and GraDeS are
clearly much more scalable than L1 and FoBa. 1(d) shows the recovery properties of different IHT
methods under large condition number (? = 50) setting as the size of projected set is increased.
Proof. Let v t = ?? f (? t ). Then, using the RSS property we get:
?2 L
L
|`|
?1 L
f (?bt ) ? f (? t ) ? h?bt ? ? t , v t i + k?bt ? ? t k22 = k?bt ? ? t k22 ?
kw ? ? t k22 ,
2
2
2 |s + ` ? s? |
(7)
t
= 0 and by
where w is any vector such that wZt = 0 and kwk0 ? s? . ?1 follows by observing vZ
t
t
?
b
noting that supp(? ) ? Zt . ?2 follows by Lemma 1 and the fact that kwk0 ? s . Now, using the
RSC property and the fact that ?? f (? t ) = 0, we have:
?
kw ? ? t k22 ? f (? t ) ? f (w) ? f (? t ) ? f (? ? ).
(8)
2
The result now follows by combining (7) and (8).
Theorem 4. Let f have RSC and RSS parameters given by ?2s+s? (f ) = ? and L2s+` (f ) =
2
L2 ?
?
?
L resp. Call Algorithm 2 with f , ` ? s? and s ? 4 L
?2 ` + s ? ` ? 4 ?2 s . Also let ? =
0
f (? )
arg min?,k?k0 ?s? f (?). Then, the ? -th iterate of Algorithm 2, for ? = O( L
? ? log( ) satisfies:
f (? ? ) ? f (? ? ) ? .
See Appendix C for a detailed proof.
6
Experiments
We conducted simulations on high dimensional sparse linear regression problems to verify our predictions. Our experiments demonstrate that hard thresholding and projected gradient techniques can
not only offer recovery in stochastic setting, but offer much more scalable routines for the same.
Data: Our problem setting is identical to the one described in the previous section. We fixed a
parameter vector ?? by choosing s? random coordinates and setting them randomly to ?1 values.
? Xi i + ?i where
Data samples were generated as Zi = (Xi , Yi ) where Xi ? N (0, Ip ) and Yi = h?,
2
?
?i ? N (0, ? ). We studied the effect of varying dimensionality p, sparsity s , sample size n and
label noise level ? on the recovery properties of the various algorithms as well as their run times.
We chose baseline values of p = 20000, s? = 100, ? = 0.1, n = fo ? s? log p where fo is the
oversampling factor with default value fo = 2. Keeping all other quantities fixed, we varied one of
the quantities and generated independent data samples for the experiments.
Algorithms: We studied a variety of hard-thresholding style algorithms including HTP [14],
GraDeS [13] (or IHT [12]), CoSaMP [15], OMPR [17] and SP [16]. We compared them with a
standard implementation of the L1 projected scaled sub-gradient technique [23] for the lasso problem and a greedy method FoBa [24] for the same.
7
Evaluation Metrics: For the baseline noise level ? = 0.1, we found that all the algorithms were
able to recover the support set within an error of 2%. Consequently, our focus shifted to running
times for these experiments. In the experiments where noise levels were varied, we recorded, for
each method, the number of undiscovered support set elements.
Results: Figure1 describes the results of our experiments in graphical form. For sake of clarity
we included only HTP, GraDeS, L1 and FoBa results in these graphs. Graphs for the other algorithms CoSaMP, SP and OMPR can be seen in the supplementary material. The graphs indicate that
whereas hard thresholding techniques are equally effective as L1 and greedy techniques for recovery in noisy settings, as indicated by Figure1(a), the former can be much more efficient and scalable
than the latter. For instance, as Figure1(b), for the base level of p = 20000, HTP was 150? faster
than the L1 method. For higher values of p, the runtime gap widened to more than 350?. We also
note that in both these cases, HTP actually offered exact support recovery whereas L1 was unable to
recover 2 and 4 support elements respectively.
Although FoBa was faster than L1 on Figure1(b) experiments, it was still slower than HTP by 50?
and 90? for p = 20000 and 25000 respectively. Moreover, due to its greedy and incremental
nature, FoBa was found to suffer badly in settings with larger true sparsity levels. As Figure 1(c)
indicates, for even moderate sparsity levels of s? = 300 and 500, FoBa is 60 ? 75? slower than
HTP. As mentioned before, the reason for this slowdown is the greedy approach followed by FoBa:
whereas HTP took less than 5 iterations to converge for these two problems, FoBa spend 300 and
500 iterations respectively. GraDeS was found to offer much lesser run times in comparison being
slower than HTP by 30 ? 40? for larger values of p and 2 ? 5? slower for larger values of s? .
Experiments on badly conditioned problems. We also ran experiments to verify the performance
of IHT algorithms in high condition number setting. Values of p, s? and ? were kept at baseline
? we selected s? /2 random coordinates from
levels. After selecting the optimal parameter vector ?,
?
its support and s /2 random coordinates outside its support and constructed a covariance matrix
with heavy correlations between these chosen coordinates. The condition number of the resulting
matrix was close to 50. Samples were drawn from this distribution and the recovery properties of
the different IHT-style algorithms was observed as the projected sparsity levels s were increased.
Our results (see Figure 1(d)) corroborate our theoretical observation that these algorithms show
a remarkable improvement in recovery properties for ill-conditioned problems with an enlarged
projection size.
7
Discussion and Conclusions
In our work we studied iterative hard thresholding algorithms and showed that these techniques
can offer global convergence guarantees for arbitrary, possibly non-convex, differentiable objective
functions, which nevertheless satisfy Restricted Strong Convexity/Smoothness (RSC/RSM) conditions. Our results apply to a large family of algorithms that includes existing algorithms such as
IHT, GraDeS, CoSaMP, SP and OMPR. Previously the analyses of these algorithms required stringent RIP conditions that did not allow the (restricted) condition number to be larger than universal
constants specific to these algorithms.
Our basic insight was to relax this stringent requirement by running these iterative algorithms with
an enlarged support size. We showed that guarantees for high-dimensional M-estimation follow
seamlessly from our results by invoking results on RSC/RSM conditions that have already been
established in the literature for a variety of statistical settings. Our theoretical results put hard
thresholding methods on par with those based on convex relaxation or greedy algorithms. Our
experimental results demonstrate that hard thresholding methods outperform convex relaxation and
greedy methods in terms of running time, sometime by orders of magnitude, all the while offering
competitive or better recovery properties.
Our results apply to sparsity and low rank structure, arguably two of the most commonly used
structures in high dimensional statistical learning problems. In future work, it would be interesting
to generalize our algorithms and their analyses to more general structures. A unified analysis for
general structures will probably create interesting connections with existing unified frameworks
such as those based on decomposability [5] and atomic norms [25].
8
References
[1] Peter B?uhlmann and Sara Van De Geer. Statistics for high-dimensional data: methods, theory and applications. Springer, 2011.
[2] Sahand Negahban, Martin J Wainwright, et al. Estimation of (near) low-rank matrices with noise and
high-dimensional scaling. The Annals of Statistics, 39(2):1069?1097, 2011.
[3] Garvesh Raskutti, Martin J Wainwright, and Bin Yu. Minimax rates of estimation for high-dimensional
linear regression over `q -balls. Information Theory, IEEE Transactions on, 57(10):6976?6994, 2011.
[4] Angelika Rohde and Alexandre B Tsybakov. Estimation of high-dimensional low-rank matrices. The
Annals of Statistics, 39(2):887?930, 2011.
[5] Sahand N Negahban, Pradeep Ravikumar, Martin J Wainwright, Bin Yu, et al. A unified framework
for high-dimensional analysis of M-estimators with decomposable regularizers. Statistical Science,
27(4):538?557, 2012.
[6] Balas Kausik Natarajan. Sparse approximate solutions to linear systems. SIAM Journal on Computing,
24(2):227?234, 1995.
[7] Ji Liu, Ryohei Fujimaki, and Jieping Ye. Forward-backward greedy algorithms for general convex smooth
functions over a cardinality constraint. In Proceedings of The 31st International Conference on Machine
Learning, pages 503?511, 2014.
[8] Ali Jalali, Christopher C Johnson, and Pradeep D Ravikumar. On learning discrete graphical models using
greedy methods. In NIPS, pages 1935?1943, 2011.
[9] Shai Shalev-Shwartz, Nathan Srebro, and Tong Zhang. Trading accuracy for sparsity in optimization
problems with sparsity constraints. SIAM Journal on Optimization, 20(6):2807?2832, 2010.
[10] Yurii Nesterov. Introductory lectures on convex optimization: A basic course, volume 87 of Applied
Optimization. Springer, 2004.
[11] P. Loh and M. J. Wainwright. Regularized M-estimators with nonconvexity: Statistical and algorithmic
theory for local optima, 2013. arXiv:1305.2436 [math.ST].
[12] Thomas Blumensath and Mike E. Davies. Iterative hard thresholding for compressed sensing. Applied
and Computational Harmonic Analysis, 27(3):265 ? 274, 2009.
[13] Rahul Garg and Rohit Khandekar. Gradient descent with sparsification: an iterative algorithm for sparse
recovery with restricted isometry property. In ICML, 2009.
[14] Simon Foucart. Hard thresholding pursuit: an algorithm for compressive sensing. SIAM J. on Num. Anal.,
49(6):2543?2563, 2011.
[15] Deanna Needell and Joel A. Tropp. CoSaMP: Iterative Signal Recovery from Incomplete and Inaccurate
Samples. Appl. Comput. Harmon. Anal., 26:301?321, 2008.
[16] Wei Dai and Olgica Milenkovic. Subspace pursuit for compressive sensing signal reconstruction. IEEE
Trans. Inf. Theory, 55(5):22302249, 2009.
[17] Prateek Jain, Ambuj Tewari, and Inderjit S. Dhillon. Orthogonal matching pursuit with replacement. In
Annual Conference on Neural Information Processing Systems, 2011.
[18] Sohail Bahmani, Bhiksha Raj, and Petros T Boufounos. Greedy sparsity-constrained optimization. The
Journal of Machine Learning Research, 14(1):807?841, 2013.
[19] Xiaotong Yuan, Ping Li, and Tong Zhang. Gradient hard thresholding pursuit for sparsity-constrained
optimization. In Proceedings of The 31st International Conference on Machine Learning, 2014.
[20] Yuchen Zhang, Martin J. Wainwright, and Michael I. Jordan. Lower bounds on the performance of
polynomial-time algorithms for sparse linear regression. arXiv:1402.1918, 2014.
[21] P. Loh and M. J. Wainwright. High-dimension regression with noisy and missing data: Provable guarantees with non-convexity. Annals of Statistics, 40(3):1637?1664, 2012.
[22] Alekh Agarwal, Sahand N. Negahban, and Martin J. Wainwright. Fast global convergence of gradient
methods for high-dimensional statistical recovery. Annals of Statistics, 40(5):2452?2482, 2012.
[23] Mark Schmidt. Graphical Model Structure Learning with L1-Regularization. PhD thesis, University of
British Columbia, 2010.
[24] Tong Zhang. Adaptive Forward-Backward Greedy Algorithm for Learning Sparse Representations. IEEE
Trans. Inf. Theory, 57:4689?4708, 2011.
[25] Venkat Chandrasekaran, Benjamin Recht, Pablo A Parrilo, and Alan S Willsky. The convex geometry of
linear inverse problems. Foundations of Computational Mathematics, 12(6):805?849, 2012.
9
| 5293 |@word milenkovic:1 version:2 polynomial:1 norm:2 stronger:1 c0:2 r:23 crucially:1 git:3 covariance:3 contraction:1 decomposition:2 simulation:1 invoking:1 bahmani:1 minmax:1 contains:1 liu:1 selecting:2 offering:1 interestingly:1 undiscovered:2 existing:4 current:2 com:1 ust:1 additive:2 enables:1 update:1 v:1 greedy:15 selected:2 rp1:4 num:1 provides:2 math:1 firstly:1 zhang:4 c2:3 constructed:1 ryohei:1 yuan:1 prove:2 specialize:1 blumensath:1 combine:1 introductory:1 theoretically:1 indeed:1 hardness:1 p1:1 grade:11 considering:1 increasing:1 kwk0:2 project:1 cardinality:1 underlying:2 moreover:3 notation:2 bounded:1 prateek:2 kg:1 minimizes:1 developed:1 compressive:5 unified:4 sparsification:1 guarantee:11 tewaria:1 formalizes:1 rohde:1 runtime:3 demonstrates:2 k2:10 scaled:1 yn:1 arguably:1 before:2 negligible:1 local:1 insufficiency:1 despite:1 analyzing:1 incoherence:1 foba:12 ap:1 chose:1 garg:1 studied:3 sara:1 appl:1 co:1 fastest:1 range:1 practical:1 atomic:1 procedure:1 universal:5 empirical:5 thought:1 significantly:1 projection:10 davy:1 matching:1 get:2 onto:3 cannot:2 close:1 operator:3 put:1 risk:2 crisp:1 demonstrated:1 missing:4 jieping:1 l:3 convex:17 focused:1 independently:1 formalized:1 recovery:18 immediately:2 decomposable:1 needell:1 estimator:5 insight:2 handle:1 traditionally:1 coordinate:5 variation:2 resp:1 target:1 suppose:4 annals:4 rip:7 exact:1 us:1 element:8 satisfying:1 natarajan:1 observed:1 mike:1 solved:1 ran:1 mentioned:1 intuition:1 benjamin:1 convexity:6 inaccessible:1 complexity:1 nesterov:1 angelika:1 tight:2 solving:2 ali:1 k0:11 routinely:2 various:2 corrective:8 jain:2 fast:1 effective:1 kp:5 choosing:1 outside:1 shalev:1 supplementary:1 solve:1 valued:1 larger:4 spend:1 elaborates:1 relax:1 compressed:1 statistic:5 noisy:5 ip:1 differentiable:6 took:1 reconstruction:1 combining:2 achieve:2 ky:1 convergence:10 cosamp:8 p:5 neumann:1 requirement:1 optimum:1 incremental:2 op:2 p2:5 strong:8 implemented:1 implies:3 indicate:1 trading:1 concentrate:1 owing:1 stochastic:1 stringent:2 kgit:1 material:1 everything:1 bin:2 require:5 suffices:1 secondly:1 hold:10 effortlessly:1 algorithmic:1 estimation:20 sometime:1 applicable:1 label:3 uhlmann:1 bridge:1 largest:2 vz:1 create:1 minimization:2 sohail:1 brought:1 htp:16 clearly:1 gaussian:2 rather:1 varying:1 corollary:1 l0:3 focus:1 viz:2 improvement:1 unsatisfactory:1 rank:25 indicates:1 seamlessly:1 contrast:1 baseline:3 zk22:4 inaccurate:1 typically:3 bt:6 w:1 issue:2 arg:10 ill:1 art:1 special:1 constrained:2 identical:2 kw:4 broad:1 look:1 k2f:5 yu:2 icml:1 future:1 minimized:2 np:1 employ:1 modern:1 randomly:1 individual:1 replaced:1 geometry:1 replacement:1 microsoft:2 n1:1 attempt:1 organization:1 evaluation:1 fujimaki:1 joel:1 introduces:1 analyzed:1 pradeep:2 regularizers:1 l2s:5 partial:1 minw:1 orthogonal:1 harmon:1 incomplete:1 yuchen:1 theoretical:2 rsc:23 increased:4 instance:1 corroborate:1 subset:2 entry:1 decomposability:1 conducted:1 johnson:1 corrupted:1 st:4 recht:1 fundamental:1 negahban:3 siam:3 international:2 invoke:1 michael:1 together:1 von:1 thesis:1 recorded:1 choose:1 possibly:2 style:9 li:1 supp:14 prox:2 de:1 parrilo:1 sec:2 includes:2 satisfy:6 depends:1 performed:2 vst:1 later:2 lot:1 analyze:4 observing:1 competitive:1 recover:2 shai:1 defer:1 simon:1 contribution:1 square:3 accuracy:1 efficiently:3 generalize:3 none:1 converged:2 ping:1 fo:3 iht:18 definition:3 failure:1 proof:12 petros:1 popular:5 dimensionality:3 routine:3 actually:2 alexandre:1 higher:1 figure1:4 follow:1 wherein:1 rahul:1 wei:1 though:1 just:1 stage:5 correlation:1 hand:1 sketch:1 tropp:1 replacing:1 christopher:1 widespread:1 indicated:1 usa:1 effect:1 ye:1 k22:14 unbiased:1 true:2 counterpart:2 verify:2 regularization:2 former:1 dhillon:1 noted:1 m:5 generalized:1 demonstrate:3 l1:15 rsm:2 harmonic:1 invoked:2 garvesh:1 pseudocode:2 raskutti:1 ji:1 overview:1 exponentially:1 volume:1 extend:3 ompr:5 smoothness:3 mathematics:1 similarly:2 pq:1 hxi:1 access:1 alekh:1 base:1 isometry:2 recent:3 purushottam:1 showed:2 raj:1 moderate:1 apart:1 inf:2 scenario:1 certain:3 kar:1 inequality:1 arbitrarily:4 continue:1 life:1 yi:7 seen:1 dai:1 relaxed:1 converge:1 signal:2 smooth:2 alan:1 match:1 faster:3 offer:11 long:1 escaped:1 equally:1 ravikumar:2 feasibility:1 prediction:2 scalable:4 regression:18 basic:2 metric:1 arxiv:2 iteration:2 agarwal:1 c1:4 whereas:3 singular:6 crucial:1 probably:1 jordan:1 call:1 structural:1 near:1 noting:1 enough:1 variety:5 iterate:5 variate:1 zi:3 lasso:1 lesser:1 handled:1 sahand:3 penalty:2 suffer:3 render:1 peter:1 loh:2 proceed:1 jj:1 tewari:2 detailed:4 tsybakov:1 differentiability:1 handily:1 outperform:1 oversampling:1 shifted:1 discrete:1 shall:1 key:2 putting:1 nevertheless:1 drawn:2 clarity:1 kept:1 backward:2 nonconvexity:1 graph:3 relaxation:5 run:2 inverse:1 extends:1 family:5 place:2 chandrasekaran:1 appendix:7 scaling:1 bound:13 followed:1 oracle:2 badly:3 annual:1 constraint:9 ri:2 sake:1 nathan:1 extremely:1 min:22 xiaotong:1 gst:4 bhiksha:1 relatively:1 martin:5 scad:1 ball:1 smaller:1 remain:1 describes:1 wi:2 s1:4 projecting:1 restricted:14 mcp:1 equation:1 previously:1 turn:1 prajain:1 end:4 umich:1 yurii:1 pursuit:7 apply:5 progression:1 generic:2 schmidt:1 slower:4 rp:4 thomas:1 top:2 running:5 include:2 graphical:3 logscale:1 exploit:1 restrictive:1 prof:1 objective:6 move:1 question:1 quantity:2 already:1 usual:1 traditional:1 diagonal:1 said:2 jalali:1 gradient:15 subspace:4 unable:1 spanning:1 reason:1 khandekar:1 provable:1 assuming:2 willsky:1 index:1 providing:1 equivalently:1 setup:1 statement:1 trace:1 sigma:1 implementation:1 anal:2 zt:4 perform:1 upper:1 observation:5 descent:6 situation:2 y1:1 varied:2 arbitrary:4 ordinate:1 pablo:1 pair:1 widened:1 required:1 c3:2 z1:9 connection:1 established:2 nip:1 trans:2 capped:1 able:4 precluded:1 k2op:2 below:4 deanna:1 regime:1 sparsity:25 ambuj:2 including:2 max:5 wainwright:7 critical:1 natural:1 rely:1 regularized:2 solvable:1 minimax:2 imply:1 lk:5 columbia:1 faced:2 literature:3 geometric:2 l2:1 rohit:1 loss:11 fully:9 par:1 lecture:1 interesting:4 limitation:1 srebro:1 remarkable:1 foundation:1 incurred:2 offered:1 consistent:4 olgica:1 thresholding:29 heavy:1 course:1 last:1 keeping:2 slowdown:1 allow:2 india:1 sparse:22 van:1 default:1 xn:1 world:1 dimension:1 rich:1 forward:2 commonly:2 made:2 projected:10 avoided:1 wzt:1 ple:1 far:1 adaptive:1 transaction:1 approximate:1 keep:1 dealing:1 confirm:1 global:7 reveals:1 conclude:1 xi:9 shwartz:1 iterative:13 infeasibility:1 nature:1 expansion:1 diag:3 sp:5 did:1 main:1 s2:4 noise:11 x1:2 enlarged:2 referred:1 venkat:1 slow:2 tong:3 sub:2 comput:1 ply:1 third:1 removing:1 down:1 theorem:11 kop:1 specific:1 british:1 sensing:6 foucart:1 balas:1 adding:1 phd:1 magnitude:3 conditioned:3 kx:1 nk:1 gap:2 michigan:1 infinitely:1 pythagoras:1 inderjit:1 springer:2 satisfies:5 relies:1 goal:2 ann:1 consequently:2 replace:1 feasible:2 hard:32 analysing:1 change:1 specifically:1 pgd:12 except:2 included:1 lemma:15 boufounos:1 called:1 geer:1 arbor:1 svd:1 experimental:1 outnumbers:1 exception:1 formally:2 support:12 mark:1 latter:1 correlated:1 |
4,742 | 5,294 | A Safe Screening Rule for Sparse Logistic Regression
Jiayu Zhou
Arizona State University
Tempe, AZ 85287
[email protected]
Jie Wang
Arizona State University
Tempe, AZ 85287
[email protected]
Jun Liu
SAS Institute Inc.
Cary, NC 27513
[email protected]
Peter Wonka
Arizona State University
Tempe, AZ 85287
[email protected]
Jieping Ye
Arizona State University
Tempe, AZ 85287
[email protected]
Abstract
The `1 -regularized logistic regression (or sparse logistic regression) is a widely
used method for simultaneous classification and feature selection. Although many
recent efforts have been devoted to its efficient implementation, its application to
high dimensional data still poses significant challenges. In this paper, we present a
fast and effective sparse logistic regression screening rule (Slores) to identify the
?0? components in the solution vector, which may lead to a substantial reduction
in the number of features to be entered to the optimization. An appealing feature
of Slores is that the data set needs to be scanned only once to run the screening and
its computational cost is negligible compared to that of solving the sparse logistic
regression problem. Moreover, Slores is independent of solvers for sparse logistic regression, thus Slores can be integrated with any existing solver to improve
the efficiency. We have evaluated Slores using high-dimensional data sets from
different applications. Experiments demonstrate that Slores outperforms the existing state-of-the-art screening rules and the efficiency of solving sparse logistic
regression can be improved by one magnitude.
1
Introduction
Logistic regression (LR) is a popular and well established classification method that has been widely
used in many domains such as machine learning [4, 7], text mining [3, 8], image processing [9, 15],
bioinformatics [1, 13, 19, 27, 28], medical and social sciences [2, 17] etc. When the number of
feature variables is large compared to the number of training samples, logistic regression is prone
to over-fitting. To reduce over-fitting, regularization has been shown to be a promising approach.
Typical examples include `2 and `1 regularization. Although `1 regularized LR is more challenging
to solve compared to `2 regularized LR, it has received much attention in the last few years and
the interest in it is growing [20, 25, 28] due to the increasing prevalence of high-dimensional data.
The most appealing property of `1 regularized LR is the sparsity of the resulting models, which is
equivalent to feature selection.
In the past few years, many algorithms have been proposed to efficiently solve the `1 regularized
LR [5, 12, 11, 18]. However, for large-scale problems, solving the `1 regularized LR with higher
accuracy remains challenging. One promising solution is by ?screening?, that is, we first identify
the ?inactive? features, which have 0 coefficients in the solution and then discard them from the
optimization. This would result in a reduced feature matrix and substantial savings in computational
cost and memory size. In [6], El Ghaoui et al. proposed novel screening rules, called ?SAFE?,
to accelerate the optimization for a class of `1 regularized problems, including LASSO [21], `1
1
regularized LR and `1 regularized support vector machines. Inspired by SAFE, Tibshirani et al.
[22] proposed ?strong rules? for a large class of `1 regularized problems, including LASSO, elastic
net, `1 regularized LR and more general convex problems. In [26], Xiang et al. proposed ?DOME?
rules to further improve SAFE rules for LASSO based on the observation that SAFE rules can be
understood as a special case of the general ?sphere test?. Although both strong rules and the sphere
tests are more effective in discarding features than SAFE for solving LASSO, it is worthwhile to
mention that strong rules may mistakenly discard features that have non-zero coefficients in the
solution and the sphere tests are not easy to be generalized to handle the `1 regularized LR. To the
best of our knowledge, the SAFE rule is the only screening test for the `1 regularized LR that is
?safe?, that is, it only discards features that are guaranteed to be absent from the resulting models.
In this paper, we develop novel screening rules, called
?Slores?, for the `1 regularized LR. The proposed screening tests detect inactive features by estimating an upper
bound of the inner product between each feature vector
and the ?dual optimal solution? of the `1 regularized LR, which is unknown. The more accurate the estimation
is, the more inactive features can be detected. An accurate estimation of such an upper bound turns out to be
quite challenging. Indeed most of the key ideas/insights
behind existing ?safe? screening rules for LASSO heavily rely on the least square loss, which are not applicable
for the `1 regularized LR case due to the presence of the
logistic loss. To this end, we propose a novel framework Figure 1: Comparison of Slores, strong
to accurately estimate an upper bound. Our key techni- rule and SAFE on the prostate cancer
cal contribution is to formulate the estimation of an upper data set.
bound of the inner product as a constrained convex optimization problem and show that it admits
a closed form solution. Therefore, the estimation of the inner product can be computed efficiently.
Our extensive experiments have shown that Slores discards far more features than SAFE yet requires
much less computational efforts. In contrast with strong rules, Slores is ?safe?, i.e., it never discards
features which have non-zero coefficients in the solution.
To illustrate the effectiveness of Slores, we compare Slores, strong rule and SAFE on a data set of
prostate cancer along a sequence of 86 parameters equally spaced on the ?/?max scale from 0.1 to
0.95, where ? is the parameter for the `1 penalty and ?max is the smallest tuning parameter [10] such
that the solution of the `1 regularized LR is 0 [please refer to Eq. (1)]. The data matrix contains 132
patients with 15154 features. To measure the performance of different screening rules, we compute
the rejection ratio which is the ratio between the number of features discarded by screening rules
and the number of features with 0 coefficients in the solution. Therefore, the larger the rejection
ratio is, the more effective the screening rule is. The results are shown in Fig. 1. We can see that
Slores discards far more features than SAFE especially when ?/?max is large while the strong rule
is not applicable when ?/?max ? 0.5. We present more results and discussions to demonstrate the
effectiveness of Slores in Section 6. For proofs of the lemmas, corollaries, and theorems, please
refer to the long version of this paper [24].
2
Basics and Motivations
In this section, we briefly review the basics of the `1 regularized LR and then motivate the general
screening rules via the KKT conditions. Suppose we are given a set of training samples {xi }m
i=1
and the associate labels b ? <m , where xi ? <p and bi ? {1, ?1} for all i ? {1, . . . , m}. The `1
regularized logistic regression is:
m
1 X
? i i ? bi c)) + ?k?k1 ,
min
log(1 + exp(?h?, x
?,c m
i=1
(LRP? )
? i = bi xi , and ? > 0 is the
where ? ? <p and c ? < are the model parameters to be estimated, x
? i and the j th
tuning parameter. We denote by X ? <m?p the data matrix with the ith row being x
j
? .
column being x
2
Let C = {? ? <m : ?i ? (0, 1), i = 1, . . . , m} and f (y) = y log(y) + (1 ? y) log(1 ? y) for
y ? (0, 1). The dual problem of (LRP? ) [24] is given by
(
)
m
1 X
? T ?k? ? m?, h?, bi = 0, ? ? C .
min g(?) =
f (?i ) : kX
(LRD? )
?
m i=1
To simplify notations, we denote the feasible set of problem (LRD? ) as F? , and let (??? , c?? ) and
??? be the optimal solutions of problems (LRP? ) and (LRD? ) respectively. In [10], the authors have
shown that for some special choice of the tuning parameter ?, both of (LRP? ) and (LRD? ) have
closed form solutions. In fact, let P = {i : bi = 1}, N = {i : bi = ?1}, and m+ and m? be the
cardinalities of P and N respectively. We define
? T ??
?max = 1 kX
k? ,
(1)
?max
m
where
(
[??? max ]i
=
m?
m ,
m+
m ,
if i ? P,
if i ? N ,
i = 1, . . . , m.
(2)
([?]i denotes the ith component of a vector.) Then, it is known [10] that ??? = 0 and ??? = ??? max
whenever ? ? ?max . When ? ? (0, ?max ], it is known that (LRD? ) has a unique optimal solution
[24]. We can now write the KKT conditions of problems (LRP? ) and (LRD? ) as
?
if [??? ]j > 0,
?m?,
?
j
? i ? ?m?,
h?? , x
(3)
if [??? ]j < 0, j = 1, . . . , p.
?
[?m?, m?], if [??? ]j = 0.
In view of Eq. (3), we can see that
? j i| < m? ? [??? ]j = 0.
|h??? , x
(R1)
?j i
|h??? , x
?j
In other words, if
< m?, then the KKT conditions imply that the coefficient of x in the
solution ??? is 0 and thus the j th feature can be safely removed from the optimization of (LRP? ).
However, for the general case in which ? < ?max , (R1) is not applicable since it assumes the
knowledge of ??? . Although it is unknown, we can still estimate a region A? which contains ??? . As
? j i| < m?, we can also conclude that [??? ]j = 0 by (R1). In other words,
a result, if max??A? |h?, x
(R1) can be relaxed as
? j ) := max |h?, x
? j i| < m? ? [??? ]j = 0.
T (??? , x
(R10 )
??A?
In this paper, (R10 ) serves as the foundation for constructing our screening rules, Slores. From
? j ) are more aggressive in discarding
(R10 ), it is easy to see that screening rules with smaller T (??? , x
? ?j
features. To give a tight estimation of T (?? , x ), we need to restrict the region A? which includes
?j )
??? as small as possible. In Section 3, we show that the estimation of the upper bound T (??? , x
can be obtained via solving a convex optimization problem. We show in Section 4 that the convex
optimization problem admits a closed form solution and derive Slores in Section 5 based on (R10 ).
3
Estimating the Upper Bound via Solving a Convex Optimization Problem
? j ) of |h??? , x
? j i|. In
In this section, we present a novel framework to estimate an upper bound T (??? , x
the subsequent development, we assume a parameter ?0 and the corresponding dual optimal ??? 0 are
given. In our Slores rule to be presented in Section 5, we set ?0 and ??? 0 to be ?max and ??? max given
? j ) as a constrained convex optimization
in Eqs. (1) and (2). We formulate the estimation of T (??? , x
problem in this section, which will be shown to admit a closed form solution in Section 4.
?i
1
1
1
4
For the dual function g(?), it follows that [?g(?)]i = m
log( 1??
), [?2 g(?)]i,i = m
?i (1??i ) ? m .
i
4
Since ?2 g(?) is a diagonal matrix, it follows that ?2 g(?) m
I, where I is the identity matrix.
4
Thus, g(?) is strongly convex with modulus ? = m [16]. Rigorously, we have the following lemma.
Lemma 1. Let ? > 0 and ?1 , ?2 ? F? , then
a).
g(?2 ) ? g(?1 ) ? h?g(?1 ), ?2 ? ?1 i +
2
m k?2
? ?1 k22 .
(4)
b). If ?1 6= ?2 , the inequality in (4) becomes a strict inequality, i.e., ??? becomes ?>?.
3
Given ? ? (0, ?0 ], it is easy to see that both of ??? and ??? 0 belong to F?0 . Therefore, Lemma 1 can
be a useful tool to bound ??? with the knowledge of ??? 0 . In fact, we have the following theorem.
Theorem 2. Let ?max ? ?0 > ? > 0, then the following holds:
i
mh ? ?
(5)
a).
k??? ? ??? 0 k22 ?
g ?0 ??0 ? g(??? 0 ) + 1 ? ??0 h?g(??? 0 ), ??? 0 i
2
b). If ??? 6= ??? 0 , the inequality in (5) becomes a strict inequality, i.e., ??? becomes ?<?.
Theorem 2 implies that ??? is inside a ball centred at ??? 0 with radius
r h
r=
m
2
g
? ?
?0 ??0
? g(??? 0 ) + (1 ?
i
?
?
?
?0 )h?g(??0 ), ??0 i
.
(6)
Recall that to make our screening rules more aggressive in discarding features, we need to get a tight
? j ) of |h??? , x
? j i| [please see (R10 )]. Thus, it is desirable to further restrict the
upper bound T (??? , x
possible region A? of ??? . Clearly, we can see that
h??? , bi = 0
(7)
???
?j i
h??? 0 , x
since
is feasible for problem (LRD? ). On the other hand, we call the set I?0 = {j :
|m?0 |, j = 1, . . . , p} the ?active set? of ??? 0 . We have the following lemma for the active set.
=
? j i| =
Lemma 3. Given the optimal solution ??? of problem (LRD? ), the active set I? = {j : |h??? , x
m?, j = 1, . . . , p} is not empty if ? ? (0, ?max ].
Since ?0 ? (0, ?max ], we can see that I?0 is not empty by Lemma 3. We pick j0 ? I?0 and set
? ? = sign(h??? 0 , x
? j0 i)?
x
xj0 .
(8)
It follows that h?
x? , ??? 0 i = m?0 . Due to the feasibility of ??? for problem (LRD? ), ??? satisfies
? ? i ? m?.
h??? , x
(9)
As a result, Theorem 2, Eq. (7) and (9) imply that ??? is contained in the following set:
? ? i ? m?}.
A??0 := {? : k? ? ??? 0 k22 ? r2 , h?, bi = 0, h?, x
? j i| ? max??A?? |h?, x
? j i|. Therefore, (R10 ) implies that if
Since ??? ? A??0 , we can see that |h??? , x
0
? j ; ??? 0 )
T (??? , x
? j i|
:= max |h?, x
??A?
?
(UBP)
0
? j can be discarded from the optimization
is smaller than m?, we can conclude that [??? ]j = 0 and x
? j ) with T (??? , x
? j ; ??? 0 ) and A??0
of (LRP? ). Notice that, we replace the notations A? and T (??? , x
?
? j ; ??? 0 ), (R10 )
to emphasize their dependence on ??0 . Clearly, as long as we can solve for T (??? , x
would be an applicable screening rule to discard features which have 0 coefficients in ??? . We give a
closed form solution of problem (UBP) in the next section.
4
Solving the Convex Optimization Problem (UBP)
In this section, we show how to solve the convex optimization problem (UBP) based on the standard
Lagrangian multiplier method. We first transform problem (UBP) into a pair of convex minimization
problem (UBP0 ) via Eq. (11) and then show that the strong duality holds for (UBP0 ) in Lemma 6. The
strong duality guarantees the applicability of the Lagrangian multiplier method. We then give the
closed form solution of (UBP0 ) in Theorem 8. After we solve problem (UBP0 ), it is straightforward
to compute the solution of problem (UBP) via Eq. (11).
Before we solve (UBP) for the general case, it is worthwhile to mention a special case in which
xj ,bi
? j ? h?kbk
P?
xj = x
2 b = 0. Clearly, P is the projection operator which projects a vector onto the
2
orthogonal complement of the space spanned by b. In fact, we have the following theorem.
Theorem 4. Let ?max ? ?0 > ? > 0, and assume ??? 0 is known. For j ? {1, . . . , p}, if P?
xj = 0,
? ?j ?
then T (?? , x ; ??0 ) = 0.
4
Because of (R10 ), we immediately have the following corollary.
Corollary 5. Let ? ? (0, ?max ) and j ? {1, . . . , p}. If P?
xj = 0, then [??? ]j = 0.
For the general case in which P?
xj 6= 0, let
? j ; ??? 0 ) := max h?, +?
? j ; ??? 0 ) := max h?, ??
T+ (??? , x
xj i, T? (??? , x
xj i.
??A?
?
??A?
?
0
(10)
0
Clearly, we have
? j ; ??? 0 ) = max{T+ (??? , x
? j ; ??? 0 ), T? (??? , x
? j ; ??? 0 )}.
T (??? , x
(11)
Therefore, we can solve problem (UBP) by solving the two sub-problems in (10).
Let ? ? {+1, ?1}. Then problems in (10) can be written uniformly as
? j ; ??? 0 ) = max h?, ??
xj i.
T? (??? , x
??A?
?
(UBPs )
0
To make use of the standard Lagrangian multiplier method, we transform problem (UBPs ) to the
following minimization problem:
? j ; ??? 0 ) = min h?, ???
?T? (??? , x
xj i
??A?
?
(UBP0 )
0
j
by noting that max??A?? h?, ??
x i = ? min??A?? h?, ???
xj i.
0
0
Lemma 6. Let ?max ? ?0 > ? > 0 and assume ??? 0 is known. The strong duality holds for problem
(UBP0 ). Moreover, problem (UBP0 ) admits an optimal solution in A??0 .
Because the strong duality holds for problem (UBP0 ) by Lemma 6, the Lagrangian multiplier method
is applicable for (UBP0 ). In general, we need to first solve the dual problem and then recover the
? ? are defined by
optimal solution of the primal problem via KKT conditions. Recall that r and x
Eq. (6) and (8) respectively. Lemma 7 derives the dual problems of (UBP0 ) for different cases.
Lemma 7. Let ?max ? ?0 > ? > 0 and assume ??? 0 is known. For j ? {1, . . . , p} and P?
xj 6= 0,
? = ???
let x
xj . Denote
n
o
x,P?
x? i
U1 = {(u1 , u2 ) : u1 > 0, u2 ? 0} and U2 = (u1 , u2 ) : u1 = 0, u2 = ? hP?
.
kP?
x? k2
2
a). If
hP?
x,P?
x? i
kP?
xk2 kP?
x ? k2
max
(u1 ,u2 )?U1
? (?1, 1], the dual problem of (UBP0 ) is equivalent to:
? i ? 21 u1 r2 .
g?(u1 , u2 ) = ? 2u1 1 kP?
x + u2 P?
x? k22 + u2 m(?0 ? ?) + h??? 0 , x
(UBD0 )
Moreover, g?(u1 , u2 ) attains its maximum in U1 .
b). If
hP?
x,P?
x? i
kP?
xk2 kP?
x ? k2
= ?1, the dual problem of (UBP0 ) is equivalent to:
(
max
(u1 ,u2 )?U1 ?U2
g?(u1 , u2 ) =
g?(u1 , u2 ),
kP?
xk2
? kP?
x? k2 m?,
if (u1 , u2 ) ? U1 ,
if (u1 , u2 ) ? U2 .
(UBD00 )
We can now solve problem (UBP0 ) in the following theorem.
Theorem 8. Let ?max ? ?0 > ? > 0, d =
? = ???
and P?
xj 6= 0, let x
xj .
a). If
hP?
x,P?
x? i
kP?
xk2 kP?
x ? k2
m(?0 ??)
rkP?
x? k2
and assume ??? 0 is known. For j ? {1, . . . , p}
? d, then
? j ; ??? 0 ) = rkP?
? i;
T? (??? , x
xk2 ? h??? 0 , x
5
(12)
b). If
hP?
x,P?
x? i
kP?
xk2 kP?
x ? k2
< d, then
? j ; ??? 0 ) = rkP?
? i,
T? (??? , x
x + u?2 P?
x? k2 ? u?2 m(?0 ? ?) ? h??? 0 , x
(13)
where
u?2 =
?
?a1 + ?
,
2a2
? 4
kP?
x k2 (1 ?
?
a2 =
d2 ),
a1 = 2hP?
x, P?
x ikP?
x? k22 (1 ? d2 ),
? 2
a0 = hP?
x, P?
x i ? d2 kP?
xk22 kP?
x? k22 ,
? = a21 ? 4a2 a0 = 4d2 (1 ? d2 )kP?
x? k42 (kP?
xk22 kP?
x? k22 ? hP?
x, P?
x? i2 ).
(14)
Notice that, although the dual problems of (UBP0 ) in Lemma 7 are different, the resulting upper
? j ; ??? 0 ) can be given by Theorem 8 in a uniform way. The tricky part is how to deal
bound T? (??? , x
with the extremal cases in which
5
hP?
x,P?
x? i
kP?
xk2 kP?
x ? k2
? {?1, +1}.
The proposed Slores Rule for `1 Regularized Logistic Regression
Using (R10 ), we are now ready to construct the screening rules for the `1 Regularized Logistic
Regression. By Corollary 5, we can see that the orthogonality between the j th feature and the
? j from the resulting model. For the general case in which
response vector b implies the absence of x
? j ; ??? 0 ) = max{T+ (??? , x
? j ; ??? 0 ), T? (??? , x
? j ; ??? 0 )} < m?,
P?
xj 6= 0, (R10 ) implies that if T (??? , x
then the j th feature can be discarded from the optimization of (LRP? ). Notice that, letting ? = ?1,
? j ; ??? 0 ) have been solved by Theorem 8. Rigorously, we have the
? j ; ??? 0 ) and T? (??? , x
T+ (??? , x
following theorem.
Theorem 9 (Slores). Let ?0 > ? > 0 and assume ??? 0 is known.
1. If ? ? ?max , then ??? = 0;
2. If ?max ? ?0 > ? > 0 and either of the following holds:
(a) P?
xj = 0,
? j ; ??? 0 ) : ? = ?1} < m?,
(b) max{T? (??? , x
?
then [?? ]j = 0.
Based on Theorem 9, we construct the Slores rule as summarized below in Algorithm 1.
Notice that, the output R of Slores is the indices
?
of the features that need to be entered to the Algorithm 1 R = Slores(X, b, ?, ?0 , ??0 )
optimization. As a result, suppose the output
Initialize R := {1, . . . , p};
if ? ? ?max then
of Algorithm 1 is R = {j1 , . . . , jk }, we can
set R = ?;
substitute the full matrix X in problem (LRP? )
else
j1
j
k
? ) and
with the sub-matrix XR = (?
x ,...,x
for j = 1 to p do
just solve for [??? ]R and c?? .
j
if P?
x = 0 then
remove j from R;
? j ; ??? 0 ) : ? = ?1} < m?
else if max{T? (??? , x
then
remove j from R;
end if
end for
end if
Return: R
On the other hand, Algorithm 1 implies that
Slores needs five inputs. Since X and b come
with the data and ? is chosen by the user, we only need to specify ??? 0 and ?0 . In other words,
we need to provide Slores with a dual optimal solution of problem (LRD? ) for an arbitrary parameter. A natural choice is by setting
?0 = ?max and ??? 0 = ??? max given by Eq. (1)
and Eq. (2) in closed form.
6
Experiments
We evaluate our screening rules using the newgroup data set [10] and Yahoo web pages data sets
[23]. The newgroup data set is cultured from the data by Koh et al. [10]. The Yahoo data sets include 11 top-level categories, each of which is further divided into a set of subcategories. In
6
our experiment we construct five balanced binary classification datasets from the topics of Computers, Education, Health, Recreation, and Science. For each topic, we choose samples from
one subcategory as the positive class and randomly sample an equal number of samples from the
rest of subcategories as the negative class. The statistics of the data sets are given in Table 1.
Table 1: Statistics of the test data sets.
We compare the performance of Slores and the
Data
set
m
p
no. nonzeros
strong rule which achieves state-of-the-art pernewsgroup
11269 61188
1467345
formance for `1 regularized LR. We do not inComputers
216
25259
23181
clude SAFE because it is less effective in disEducation
254
20782
28287
carding features than strong rules and requires
Health
228
18430
40145
much higher computational time [22]. Fig. 1
Recreation
370
25095
49986
has shown the performance of Slores, strong
Science
222
24002
37227
rule and SAFE. We compare the efficiency of
the three screening rules on the same prostate
cancer data set in Table 2. All of the screen- Table 2: Running time (in seconds) of Slores,
ing rules are tested along a sequence of 86 pa- strong rule, SAFE and the solver.
rameter values equally spaced on the ?/?max
Slores Strong Rule
SAFE
Solver
scale from 0.1 to 0.95. We repeat the procedure
0.37
0.33
1128.65
10.56
100 times and during each time we undersample 80% of the data. We report the total running time of the three screening rules over the 86 values
of ?/?max in Table 2. For reference, we also report the total running time of the solver1 . We observe
that the running time of Slores and strong rule is negligible compared to that of the solver. However,
SAFE takes much longer time even than the solver.
In Section 6.1, we evaluate the performance of Slores and strong rule. Recall that we use the rejection ratio, i.e., the ratio between the number of features discarded by the screening rules and the
number of features with 0 coefficients in the solution, to measure the performance of screening rules.
Note that, because no features with non-zero coefficients in the solution would be mistakenly discarded by Slores, its rejection ratio is no larger than one. We then compare the efficiency of Slores
and strong rule in Section 6.2.
The experiment settings are as follows. For each data set, we undersample 80% of the date and
run Slores and strong rules along a sequence of 86 parameter values equally spaced on the ?/?max
scale from 0.1 to 0.95. We repeat the procedure 100 times and report the average performance and
running time at each of the 86 values of ?/?max . Slores, strong rules and SAFE are all implemented
in Matlab. All of the experiments are carried out on a Intel(R) (i7-2600) 3.4Ghz processor.
6.1
Comparison of Performance
In this experiment, we evaluate the performance of the Slores and the strong rule via the rejection
ratio. Fig. 2 shows the rejection ratio of Slores and strong rule on six real data sets. When ?/?max >
0.5, we can see that both Slores and strong rule are able to identify almost 100% of the inactive
features, i.e., features with 0 coefficients in the solution vector. However, when ?/?max ? 0.5,
strong rule can not detect the inactive features. In contrast, we observe that Slores exhibits much
stronger capability in discarding inactive features for small ?, even when ?/?max is close to 0.1.
Taking the data point at which ?/?max = 0.1 for example, Slores discards about 99% inactive
features for the newsgroup data set. For the other data sets, more than 80% inactive features are
identified by Slores. Thus, in terms of rejection ratio, Slores significantly outperforms the strong
rule. Moreover, the discarded features by Slores are guaranteed to have 0 coefficients in the solution.
But strong rule may mistakenly discard features which have non-zero coefficients in the solution.
6.2
Comparison of Efficiency
We compare efficiency of Slores and the strong rule in this experiment. The data sets for evaluating
the rules are the same as Section 6.1. The running time of the screening rules reported in Fig. 3
includes the computational cost of the rules themselves and that of the solver after screening. We
plot the running time of the screening rules against that of the solver without screening. As indicated
by Fig. 2, when ?/?max > 0.5, Slores and strong rule discards almost 100% of the inactive features.
1
In this paper, the ground truth is computed by SLEP [14].
7
(a) newsgroup
(b) Computers
(c) Education
(d) Health
(e) Recreation
(f) Science
Figure 2: Comparison of the performance of Slores and strong rules on six real data sets.
(a) newsgroup
(b) Computers
(c) Education
(d) Health
(e) Recreation
(f) Science
Figure 3: Comparison of the efficiency of Slores and strong rule on six real data sets.
As a result, the size of the feature matrix involved in the optimization of problem (LRP? ) is greatly
reduced. From Fig. 3, we can observe that the efficiency is improved by about one magnitude on
average compared to that of the solver without screening. However, when ?/?max < 0.5, strong
rule can not identify any inactive features and thus the running time is almost the same as that of the
solver without screening. In contrast, Slores is still able to identify more than 80% of the inactive
features for the data sets cultured from the Yahoo web pages data sets and thus the efficiency is
improved by roughly 5 times. For the newgroup data set, about 99% inactive features are identified
by Slores which leads to about 10 times savings in running time. These results demonstrate the
power of the proposed Slores rule in improving the efficiency of solving the `1 regularized LR.
7
Conclusions
In this paper, we propose novel screening rules to effectively discard features for `1 regularized
LR. Extensive numerical experiments on real data demonstrate that Slores outperforms the existing
state-of-the-art screening rules. We plan to extend the framework of Slores to more general sparse
formulations, including convex ones, like group Lasso, fused Lasso, `1 regularized SVM, and nonconvex ones, like `p regularized problems where 0 < p < 1.
8
References
[1] M. Asgary, S. Jahandideh, P. Abdolmaleki, and A. Kazemnejad. Analysis and identification of ?-turn
types using multinomial logistic regression and artificial neural network. Bioinformatics, 23(23):3125?
3130, 2007.
[2] C. Boyd, M. Tolson, and W. Copes. Evaluating trauma care: The TRISS method, trauma score and the
injury severity score. Journal of Trauma, 27:370?378, 1987.
[3] J. R. Brzezinski and G. J. Knafl. Logistic regression modeling for context-based classification. In DEXA
Workshop, pages 755?759, 1999.
[4] K. Chaudhuri and C. Monteleoni. Privacy-preserving logistic regression. In NIPS, 2008.
[5] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani. Least angle regression. Ann. Statist., 32:407?499,
2004.
[6] L. El Ghaoui, V. Viallon, and T. Rabbani. Safe feature elimination for the lasso and sparse supervised
learning problems. arXiv:1009.4219v2.
[7] J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: a statistical view of boosting. The
Annals of Statistics, 38(2), 2000.
[8] A. Genkin, D. Lewis, and D. Madigan. Large-scale bayesian logistic regression for text categorization.
Technometrics, 49:291?304(14), 2007.
[9] S. Gould, J. Rodgers, D. Cohen, G. Elidan, and D. Koller. Multi-class segmentation with relative location
prior. International Journal of Computer Vision, 80(3):300?316, 2008.
[10] K. Koh, S. J. Kim, and S. Boyd. An interior-point method for large scale `1 -regularized logistic regression.
J. Mach. Learn. Res., 8:1519?1555, 2007.
[11] B. Krishnapuram, L. Carin, M. Figueiredo, and A. Hartemink. Sparse multinomial logistic regression:
Fast algorithms and generalization bounds. IEEE Trans. Pattern Anal. Mach. Intell., 27:957?968, 2005.
[12] S. Lee, H. Lee, P. Abbeel, and A. Ng. Efficient l1 regularized logistic regression. In In AAAI-06, 2006.
[13] J. Liao and K. Chin. Logistic regression for disease classification using microarray data: model selection
in a large p and small n case. Bioinformatics, 23(15):1945?1951, 2007.
[14] J. Liu, S. Ji, and J. Ye. SLEP: Sparse Learning with Efficient Projections. Arizona State University, 2009.
[15] S. Martins, L. Sousa, and J. Martins. Additive logistic regression applied to retina modelling. In ICIP (3),
pages 309?312. IEEE, 2007.
[16] Y. Nesterov. Introductory Lectures on Convex Optimization: A Basic Course. Springer, 2004.
[17] S. Palei and S. Das. Logistic regression model for prediction of roof fall risks in bord and pillar workings
in coal mines: An approach. Safety Science, 47:88?96, 2009.
[18] M. Park and T. Hastie. `1 regularized path algorithm for generalized linear models. J. R. Statist. Soc. B,
69:659?677, 2007.
[19] M. Sartor, G. Leikauf, and M. Medvedovic. LRpath: a logistic regression approach for identifying enriched biological groups in gene expression data. Bioinformatics, 25(2):211?217, 2009.
[20] D. Sun, T. Erp, P. Thompson, C. Bearden, M. Daley, L. Kushan, M. Hardt, K. Nuechterlein, A. Toga, and
T. Cannon. Elucidating a magnetic resonance imaging-based neuroanatomic biomarker for psychosis:
classification analysis using probabilistic brain atlas and machine learning algorithms. Biological Psychiatry, 66:1055?1?60, 2009.
[21] R. Tibshirani. Regression shringkage and selection via the lasso. J. R. Statist. Soc. B, 58:267?288, 1996.
[22] R. Tibshirani, J. Bien, J. Friedman, T. Hastie, N. Simon, J. Taylor, and R. Tibshirani. Strong rules for
discarding predictors in lasso-type problems. J. R. Statist. Soc. B, 74:245?266, 2012.
[23] N. Ueda and K. Saito. Parametric mixture models for multi-labeled text. Advances in neural information
processing systems, 15:721?728, 2002.
[24] J. Wang, J. Zhou, J. Liu, P. Wonka, and J. Ye. A safe screening rule for sparse logistic regression.
arXiv:1307.4145v2, 2013.
[25] T. T. Wu, Y. F. Chen, T. Hastie, E. Sobel, and K. Lange. Genome-wide association analysis by lasso
penalized logistic regression. Bioinformatics, 25:714?721, 2009.
[26] Z. J. Xiang and P. J. Ramadge. Fast lasso screening tests based on correlations. In IEEE ICASSP, 2012.
[27] J. Zhu and T. Hastie. Kernel logistic regression and the import vector machine. In T. G. Dietterich,
S. Becker, and Z. Ghahramani, editors, NIPS, pages 1081?1088. MIT Press, 2001.
[28] J. Zhu and T. Hastie. Classification of gene microarrays by penalized logistic regression. Biostatistics,
5:427?443, 2004.
9
| 5294 |@word briefly:1 version:1 stronger:1 pillar:1 solver1:1 d2:5 pick:1 mention:2 reduction:1 liu:4 contains:2 score:2 outperforms:3 existing:4 past:1 com:1 yet:1 written:1 import:1 numerical:1 subsequent:1 j1:2 additive:2 remove:2 plot:1 atlas:1 asu:4 ith:2 lr:18 boosting:1 location:1 five:2 rabbani:1 along:3 fitting:2 introductory:1 inside:1 privacy:1 coal:1 indeed:1 roughly:1 themselves:1 growing:1 multi:2 brain:1 inspired:1 solver:10 increasing:1 cardinality:1 becomes:4 estimating:2 moreover:4 notation:2 project:1 biostatistics:1 guarantee:1 safely:1 k2:10 tricky:1 medical:1 before:1 negligible:2 understood:1 positive:1 safety:1 mach:2 tempe:4 slores:51 path:1 challenging:3 ramadge:1 bi:9 unique:1 prevalence:1 xr:1 procedure:2 j0:2 saito:1 significantly:1 projection:2 boyd:2 word:3 madigan:1 krishnapuram:1 get:1 onto:1 close:1 selection:4 cal:1 operator:1 interior:1 context:1 risk:1 equivalent:3 lagrangian:4 jieping:2 straightforward:1 attention:1 convex:12 thompson:1 formulate:2 identifying:1 immediately:1 rule:65 insight:1 spanned:1 handle:1 annals:1 cultured:2 suppose:2 heavily:1 user:1 associate:1 pa:1 jk:1 labeled:1 wang:3 solved:1 region:3 sun:1 removed:1 substantial:2 balanced:1 disease:1 nesterov:1 rigorously:2 mine:1 motivate:1 solving:9 tight:2 efficiency:10 accelerate:1 mh:1 icassp:1 fast:3 effective:4 kp:20 detected:1 artificial:1 quite:1 widely:2 solve:10 larger:2 statistic:3 transform:2 sequence:3 net:1 propose:2 product:3 entered:2 date:1 chaudhuri:1 shringkage:1 az:4 empty:2 r1:4 categorization:1 illustrate:1 develop:1 derive:1 pose:1 received:1 sa:2 eq:9 soc:3 implemented:1 strong:33 implies:5 come:1 safe:22 radius:1 elimination:1 education:3 abbeel:1 generalization:1 biological:2 hold:5 ground:1 exp:1 achieves:1 smallest:1 xk2:7 a2:3 estimation:7 applicable:5 label:1 extremal:1 cary:1 tool:1 lrp:10 minimization:2 mit:1 clearly:4 zhou:3 cannon:1 corollary:4 modelling:1 biomarker:1 greatly:1 contrast:3 psychiatry:1 attains:1 kim:1 detect:2 el:2 integrated:1 a0:2 koller:1 classification:7 dual:10 yahoo:3 development:1 plan:1 art:3 special:3 constrained:2 initialize:1 resonance:1 equal:1 once:1 saving:2 never:1 construct:3 ng:1 park:1 carin:1 lrd:10 report:3 prostate:3 simplify:1 few:2 retina:1 randomly:1 genkin:1 intell:1 roof:1 friedman:2 technometrics:1 screening:34 interest:1 mining:1 elucidating:1 recreation:4 mixture:1 behind:1 primal:1 devoted:1 sobel:1 accurate:2 orthogonal:1 taylor:1 re:1 column:1 modeling:1 injury:1 cost:3 applicability:1 uniform:1 predictor:1 slep:2 reported:1 international:1 lee:2 probabilistic:1 fused:1 aaai:1 choose:1 admit:1 return:1 aggressive:2 centred:1 summarized:1 includes:2 coefficient:11 inc:1 toga:1 view:2 closed:7 jiayu:2 recover:1 capability:1 simon:1 contribution:1 square:1 accuracy:1 formance:1 efficiently:2 spaced:3 identify:5 identification:1 bayesian:1 accurately:1 processor:1 simultaneous:1 monteleoni:1 whenever:1 against:1 involved:1 proof:1 hardt:1 popular:1 recall:3 knowledge:3 efron:1 segmentation:1 higher:2 supervised:1 response:1 improved:3 specify:1 formulation:1 evaluated:1 strongly:1 just:1 correlation:1 hand:2 working:1 mistakenly:3 web:2 logistic:29 indicated:1 modulus:1 dietterich:1 ye:4 k22:7 multiplier:4 regularization:2 i2:1 deal:1 during:1 please:3 generalized:2 chin:1 demonstrate:4 l1:1 image:1 novel:5 multinomial:2 ji:1 cohen:1 belong:1 extend:1 rodgers:1 association:1 significant:1 refer:2 tuning:3 hp:9 longer:1 etc:1 recent:1 discard:11 nonconvex:1 inequality:4 binary:1 preserving:1 relaxed:1 care:1 xj0:1 elidan:1 full:1 desirable:1 nonzeros:1 ing:1 sphere:3 long:2 divided:1 rameter:1 equally:3 a1:2 feasibility:1 prediction:1 regression:30 basic:3 liao:1 patient:1 vision:1 arxiv:2 kernel:1 else:2 microarray:1 rest:1 strict:2 effectiveness:2 call:1 presence:1 noting:1 easy:3 xj:16 hastie:7 lasso:12 restrict:2 identified:2 reduce:1 inner:3 idea:1 lange:1 microarrays:1 absent:1 i7:1 inactive:12 six:3 expression:1 becker:1 effort:2 penalty:1 peter:2 trauma:3 matlab:1 jie:2 useful:1 statist:4 category:1 medvedovic:1 reduced:2 notice:4 sign:1 estimated:1 tibshirani:6 write:1 group:2 key:2 erp:1 r10:10 viallon:1 imaging:1 year:2 run:2 angle:1 almost:3 ueda:1 wu:1 bound:11 guaranteed:2 arizona:5 scanned:1 orthogonality:1 u1:19 min:4 martin:2 gould:1 ball:1 smaller:2 appealing:2 kbk:1 ghaoui:2 koh:2 xk22:2 remains:1 turn:2 letting:1 end:4 serf:1 observe:3 worthwhile:2 v2:2 magnetic:1 sousa:1 substitute:1 denotes:1 assumes:1 include:2 top:1 running:9 k1:1 especially:1 ghahramani:1 parametric:1 dependence:1 diagonal:1 exhibit:1 topic:2 index:1 psychosis:1 ratio:9 nc:1 wonka:3 negative:1 implementation:1 anal:1 unknown:2 subcategory:1 upper:9 observation:1 datasets:1 discarded:6 severity:1 arbitrary:1 complement:1 pair:1 extensive:2 icip:1 established:1 nip:2 trans:1 able:2 below:1 pattern:1 sparsity:1 challenge:1 bien:1 including:3 memory:1 max:50 power:1 natural:1 rely:1 regularized:29 zhu:2 improve:2 imply:2 ready:1 carried:1 jun:2 health:4 text:3 review:1 prior:1 xiang:2 relative:1 loss:2 subcategories:2 lecture:1 foundation:1 editor:1 row:1 prone:1 cancer:3 course:1 penalized:2 repeat:2 last:1 figueiredo:1 institute:1 johnstone:1 techni:1 taking:1 fall:1 wide:1 sparse:11 ghz:1 evaluating:2 genome:1 author:1 dome:1 far:2 social:1 cope:1 emphasize:1 gene:2 kkt:4 active:3 conclude:2 xi:3 table:5 promising:2 learn:1 elastic:1 improving:1 rkp:3 constructing:1 domain:1 da:1 motivation:1 enriched:1 fig:6 intel:1 screen:1 k42:1 sub:2 a21:1 daley:1 theorem:15 discarding:5 undersample:2 r2:2 admits:3 svm:1 derives:1 workshop:1 effectively:1 magnitude:2 kx:2 chen:1 rejection:7 contained:1 hartemink:1 u2:17 springer:1 truth:1 satisfies:1 lewis:1 identity:1 ann:1 replace:1 absence:1 feasible:2 typical:1 uniformly:1 lemma:13 called:2 total:2 duality:4 newsgroup:3 support:1 ustc:1 bioinformatics:5 evaluate:3 tested:1 |
4,743 | 5,295 | Generalized Dantzig Selector:
Application to the k-support norm
Soumyadeep Chatterjee?
Sheng Chen?
Arindam Banerjee
Dept. of Computer Science & Engg.
University of Minnesota, Twin Cities
{chatter,shengc,banerjee}@cs.umn.edu
Abstract
We propose a Generalized Dantzig Selector (GDS) for linear models, in which any
norm encoding the parameter structure can be leveraged for estimation. We investigate both computational and statistical aspects of the GDS. Based on conjugate
proximal operator, a flexible inexact ADMM framework is designed for solving
GDS. Thereafter, non-asymptotic high-probability bounds are established on the
estimation error, which rely on Gaussian widths of the unit norm ball and the error
set. Further, we consider a non-trivial example of the GDS using k-support norm.
We derive an efficient method to compute the proximal operator for k-support
norm since existing methods are inapplicable in this setting. For statistical analysis, we provide upper bounds for the Gaussian widths needed in the GDS analysis,
yielding the first statistical recovery guarantee for estimation with the k-support
norm. The experimental results confirm our theoretical analysis.
1 Introduction
The Dantzig Selector (DS) [3, 5] provides an alternative to regularized regression approaches such as
Lasso [19, 22] for sparse estimation. While DS does not consider a regularized maximum likelihood
approach, [3] has established clear similarities between the estimates from DS and Lasso. While
norm regularized regression approaches have been generalized to more general norms [14, 2], the
literature on DS has primarily focused on the sparse L1 norm case, with a few notable exceptions
which have considered extensions to sparse group-structured norms [11].
In this paper, we consider linear models of the form y = X? ? + w, where y ? Rn is a set of
observations, X ? Rn?p is a design matrix with i.i.d. standard Gaussian entries, and w ? Rn
is i.i.d. standard Gaussian noise. For any given norm R(?), the parameter ? ? is assumed to be
structured in terms of having a low value of R(? ? ). For this setting, we propose the following
Generalized Dantzig Selector (GDS) for parameter estimation:
!
"
?? = argmin R(?) s.t. R? XT (y ? X?) ? ?p ,
(1)
??Rp
where R? (?) is the dual norm of R(?), and ?p is a suitable constant. If R(?) is the L1 norm, (1)
reduces to standard DS [5]. A key novel aspect of GDS is that the constraint is in terms of the dual
norm R? (?) of the original structure inducing norm R(?). It is instructive to contrast GDS with
the recently proposed atomic norm based estimation framework [6] which, unlike GDS, considers
constraints based on the L2 norm of the error ?y ? X??2 .
In this paper, we consider both computational and statistical aspects of the GDS. For the L1 -norm
Dantzig selector, [5] proposed a primal-dual interior point method since the optimization is a linear
program. DASSO and its generalization proposed in [10, 9] focused on homotopy methods, which
?
Both authors contributed equally.
1
provide a piecewise linear solution path through a sequential simplex-like algorithm. However, none
of the algorithms above can be immediately extended to our general formulation. In recent work,
the Alternating Direction Method of Multipliers (ADMM) has been applied to the L1 -norm Dantzig
selection problem [12, 21], and the linearized version in [21] proved to be efficient. Motivated
by such results for DS, we propose a general inexact ADMM [20] framework for GDS where the
primal update steps, interestingly, turn out respectively to be proximal updates involving R(?) and
its convex conjugate, the indicator of R? (x) ? ?p . As a result, by Moreau decomposition, it suffices
to develop efficient proximal update for either R(?) or its conjugate. On the statistical side, we
establish non-asymptotic high-probability bounds on the estimation error ??? ? ? ? ?2 . Interestingly,
the bound depends on the Gaussian width of the unit norm ball of R(?) as well as the Gaussian width
of intersection of error cone and unit sphere [6, 16].
As a non-trivial example of the GDS framework, we consider estimation using the recently proposed
k-support norm [1, 13]. We show that proximal operators for k-support norm can be efficiently
computed in O(p log p + log k log(p ? k)), and hence the estimation can be done efficiently. Note
that existing work [1, 13] on k-support norm has focused on the proximal operator for the square of
the k-support norm, which is not directly applicable in our setting. On the statistical side, we provide
upper bounds for the Gaussian widths of the unit norm ball and the error cone as needed in the GDS
framework, yielding the first statistical recovery guarantee for estimation with the k-support norm.
The rest of the paper is organized as follows: We establish general optimization and statistical
recovery results for GDS for any norm in Section 2. In Section 3, we present efficient algorithms
and estimation error bounds for the k-support norm. We present experimental results in Section 4
and conclude in Section 5. All technical analysis and proofs can be found in [7].
2 General Optimization and Statistical Recovery Guarantees
The problem in (1) is a convex program, and a suitable choice of ?p ensures that the feasible set
is not empty. We start the section with an inexact ADMM framework for solving problems of the
form (1), and then present bounds on the estimation error establishing statistical consistency of GDS.
2.1 General Optimization Framework using Inexact ADMM
For convenience, we temporarily drop the subscript p of ?p . We let A = XT X, b = XT y, and
define the set C? = {v : R? (v) ? ?}. The optimization problem is equivalent to
min R(?)
?,v
s.t. b ? A? = v, v ? C? .
(2)
Due to the nonsmoothness of both R and R? , solving (2) can be quite challenging and a generally
applicable algorithm is Alternating Direction Method of Multipliers (ADMM) [4]. The augmented
Lagrangian function for (2) is given as
?
LR (?, v, z) = R(?) + ?z, A? + v ? b? + ||A? + v ? b||22 ,
(3)
2
where z is the Lagrange multiplier and ? controls the penalty introduced by the quadratic term. The
iterative updates of the variables (?, v, z) in standard ADMM are given by
? k+1 ? argmin LR (?, vk , zk ) ,
(4)
vk+1 ? argmin LR (? k+1 , v, zk ) ,
(5)
zk+1 ? zk + ?(A? k+1 + vk+1 ? b) .
(6)
?
v?C?
Note that update (4) amounts to a norm regularized least squares problem for ?, which can be
computationally expensive. Thus we use an inexact update for ? instead, which can alleviate the
computational cost and lead to a quite simple algorithm. Inspired by [21, 20], we consider a simpler
subproblem for the ?-update which minimizes
%
? $%
%A? k + vk ? b%2 +
L#kR (?, vk , zk ) = R(?) + ?zk , A? + vk ? b? +
2
2
(7)
%2 (
' ?%
&
k
T
k
k
k
2 ? ? ? , A (A? + v ? b) + %? ? ? %2 ,
2
2
Algorithm 1 ADMM for Generalized Dantzig Selector
Input: A = XT X, b = XT y, ?, ?
Output: Optimal ?? of (1)
1: Initialize (?, v, z)
2: while not converged do
!
3:
? k+1 ? prox 2R ? k ? ?2 AT (A? k + vk ? b +
??
!
k"
4:
vk+1 ? prox C b ? A? k+1 ? z?
?
5:
zk+1 ? zk + ?(A? k+1 + vk+1 ? b)
6: end while
"
zk
? )
where ? is a user-defined parameter. L#kR (?, vk , zk ) can be viewed as an approximation of
LR (?, vk , zk ) with the quadratic term linearized at ? k . Then the update (4) is replaced by
? k+1 ? argmin L#kR (?, vk , zk )
?
)
*
! k 2 T
zk "%
2R(?) 1 %
%2
%
k
k
+ %? ? ? ? A (A? + v ? b + ) % .
= argmin
??
2
?
?
2
?
(8)
Similarly the update of v in (5) can be recast as
1%
zk %2
vk+1 ? argmin LR (? k+1 , v, zk ) = argmin %v ? (b ? A? k+1 ? )%2 .
?
v?C?
v?C? 2
(9)
In fact, the updates of both ? and v are to compute certain proximal operators [15]. In general, the
proximal operator proxh (?) of a closed proper convex function h : Rp ? R ? {+?} is defined as
+1
,
?w ? x?22 + h(w) .
proxh (x) = argmin
2
w?Rp
Hence it is easy to see that (8) and (9) correspond to prox 2R (?) and prox
??
C? (?) is the indicator function of set C? given by
)
0
if x ? C?
(x)
=
.
C?
+?
if otherwise
C?
(?), respectively, where
In Algorithm 1, we provide our general ADMM for the GDS. For the ADMM to work, we need two
subroutines that can efficiently compute the proximal operators for the functions in Line 3 and 4
respectively. The simplicity of the proposed approach stems from the fact that we in fact need only
one subroutine, for any one of the functions, since the functions are conjugates of each other.
Proposition 1 Given ? > 0 and a norm R(?), the two functions, f (x) = ?R(x) and g(x) =
are convex conjugate to each other, thus giving the following identity,
x = proxf (x) + proxg (x) .
C? (x)
(10)
Proof: The Proposition 1 simply follows from the definition of convex conjugate and dual norm,
and (10) is just Moreau decomposition provided in [15].
The decomposition enables conversion of the two types of proximal operator to each other at negligible cost (i.e., vector subtraction). Thus we have the flexibility in Algorithm 1 to focus on the
proximal operator that is efficiently computable, and the other can be simply obtained through (10).
Remark on convergence: Note that Algorithm 1 is a special case of inexact Bregman ADMM proposed in [20], which matches the case of linearizing quadratic penalty term by using B??? (?, ?k ) =
?
1
2
2 ?? ? ?k ?2 as Bregman divergence. In order to converge, the algorithm requires 2 to be larger than
T
the spectral radius of A A, and the convergence rate is O(1/T ) according to Theorem 2 in [20].
3
2.2 Statistical Recovery for Generalized Dantzig Selector
Our goal is to provide non-asymptotic bounds on ??? ? ? ? ?2 between the true parameter ? ? and
? = ?? ? ? ? . For any set ? ? Rp , we
the minimizer ?? of (1). Let the error vector be defined as ?
would measure the size of this set using its Gaussian width [17, 6], which is defined as ?(?) =
Eg [supz?? ?g, z?] , where g is a vector of i.i.d. standard Gaussian entries. We also consider the
? defined as
error cone TR (? ? ), generated by the set of possible error vectors ? and containing ?,
TR (? ? ) := cone {? ? Rp : R(? ? + ?) ? R(? ? )} .
(11)
Note that this set contains a restricted set of directions and does not in general span the entire space
of Rp . With these definitions, we obtain our main result.
Theorem 1 Suppose that both design matrix X and noise w consists of i.i.d. Gaussian entries with
zero mean variance 1 and X has normalized columns, i.e. ?X(j) ?2 = 1, j = 1, . . . , p. If we solve
the problem (1) with
.
?p ? cE R? (XT w) ,
(12)
where c > 1 is a constant, then, with probability at least (1 ? ?1 exp(??2 n)), we have
/
4 R(? ? )?p
?
?
,
?? ? ? ?2 ?
(?n ? ?(TR (? ? ) ? Sp?1 ))
(13)
where ?(TR (? ? ) ? Sp?1 ) is the Gaussian width of the intersection of the error cone TR (? ? ) and the
unit spherical shell S?p?1 , and ?n is the expected length of a length n i.i.d. standard Gaussian vector
n
with ?n+1
< ?n < n, and ?1 , ?2 > 0 are constants.
Remark: The choice of ?p is also intimately connected to the notion of Gaussian width. Note that
for X with unit length columns, XT w = z is an i.i.d. standard Gaussian vector. Therefore the right
hand side of (12) can be written as
0
1
- ? T .
E R (X w) = E
sup ?u, z? = ? ({u : R(u) ? 1}) ,
(14)
u: R(u)?1
which is the Gaussian width of the unit ball of the norm R(?).
Example: L1 -norm Dantzig Selector When R(?) is chosen to be L1 norm, the dual norm is the
L? norm, and (1) is reduced to the standard DS, given by
?? = argmin ???1
??Rp
s.t. ?XT (y ? X?)?? ? ? .
We know that prox????1 (?) is given by the elementwise soft-thresholding operation
.
prox????1 (x) i = sign(xi ) ? max(0, |xi | ? ?) .
(15)
(16)
Based on Proposition 1, the ADMM updates in Algorithm 1 can be instantiated as
!
2
zk "
? k+1 ? prox 2???1 ? k ? AT (A? k + vk ? u + ) ,
??
?
?
k
!
z
zk "
vk+1 ? (u ? A? k+1 ? ) ? prox????1 u ? A? k+1 ?
,
?
?
zk+1 ? zk + ?(A? k+1 + vk+1 ? u) ,
where the update of v leverages the decomposition (10). Similar updates were used in [21] for
L1 -norm Dantzig selector.
For statistical recovery, we assume that ? ? is s-sparse, i.e., contains s non-zero entries, and that
?? ? ?2 = 1, so that ?? ? ?1 ? s. It was shown in [6] that!the" Gaussian width of the -set (TL1 (? ?.) ?
Sp?1 ) is upper bounded as ?(TL1 (? ? )?Sp?1 )2 ? 2s log ps + 54 s. Also note that E R? (XT w) =
4
E[?XT w?? ] ? log p, since XT w is a vector of i.i.d. standard Gaussian entries [5]. Therefore, if
we solve (15) with ?p = 2 log p, then
5
34
/
? ? log p
32??
s
log
p
1
?
2
??? ? ? ?2 ? $
(17)
( =O
! "
n
?n ? 2s log ps + 54 s
with high probability, which agrees with known results for DS [3, 5].
3 Dantzig Selection with k-support norm
We first introduce some notations. Given any ? ? Rp , let |?| denote its absolute-valued counterpart
and ? ? denote the permutation of ? with its elements arranged in decreasing order. In previous
work [1, 13], the k-support norm has been defined as
?
?
? 9
?
9
???sp
=
min
?v
?
:
supp(v
)
?
I,
v
=
?
,
(18)
I 2
I
I
k
? (k)
?
(k)
I?G
I?G
where G (k) denotes the set of subsets of {1, . . . , p} of cardinality at most k. The unit ball of this
norm is the set Ck = conv {? ? Rp : ???0 ? k, ???2 ? 1} . The dual norm of the k-support norm
is given by
3 k
51
,
+
9 ?2 2
sp?
|?|i
.
(19)
???k = max ??G ?2 : G ? G (k) =
i=1
Note that k = 1 gives the L1 norm and its dual norm is L? norm. The k-support norm was
proposed in order to overcome some of the empirical shortcomings of the elastic net [23] and the
(group)-sparse regularizers. It was shown in [1] to behave similarly as the?elastic net in the sense
that the unit norm ball of the k-support norm is within a constant factor of 2 of the unit elastic net
ball. Although multiple papers have reported good empirical performance of the k-support norm on
selecting correlated features, where L1 regularization fails, there exists no statistical analysis of the
k-support norm. Besides, current computational methods consider square of k-support norm in their
formulation, which might fail to work out in certain cases.
In the rest of this section, we focus on GDS with R(?) = ???sp
k given as
?? = argmin ???sp
k
??Rp
?
?XT (y ? X?)?sp
? ?p .
k
s.t.
(20)
For the indicator function C? (?) of the dual norm, we present a fast algorithm for computing its
proximal operator by exploiting the structure of its solution, which can be directly plugged in Algorithm 1 to solve (20). Further, we prove statistical recovery bounds for k-support norm Dantzig
selection, which hold even for a high-dimensional scenario, where n < p.
3.1 Computation of Proximal Operator
?
In order to solve (20), either prox????sp
(?) or prox C (?) for ? ? ?sp
should be efficiently comk
k
?
putable. Existing methods [1, 13] are inapplicable to our scenario since they compute the proximal
operator for squared k-support norm, from which prox C (?) cannot be directly obtained. In Theo?
rem 2, we show that prox C (?) can be efficiently computed, and thus Algorithm 1 is applicable.
?
?
?
? ?, then w? = prox C (x) = x. If ?x?sp
> ?,
Theorem 2 Given ? > 0 and x ? Rp , if ?x?sp
k
k
?
=r
=
s
?
? 2
define Asr = i=s+1 |x|i , Bs = i=1 (|x|i ) , in which 0 ? s < k and k ? r ? p, and construct
the nonlinear equation of ?,
>
?2
1+?
2
(k ? s)Asr
? ?2 (1 + ?)2 + Bs = 0 .
(21)
r ? s + (k ? s)?
5
Let ?sr be given by
)
nonnegative root of (21) if s > 0 and the root exists
0
otherwise
.
(22)
Then the proximal operator w? = proxIC (x) is given by
?
?
?
1
?
if 1 ? i ? s?
?
1+? ? r? |x|i
?
? 2 ?s2 ?B
?
s?
if s? < i ? r? and ?s? r? = 0
k?s?
|w? |?i =
As ? r ?
?
?
if s? < i ? r? and ?s? r? > 0
?
r ? ?s? +(k?s? )?s? r?
?
?
?
|x|i
if r? < i ? p
,
(23)
?sr =
where the indices s? and r? with computed |w? |? satisfy the following two inequalities:
|w? |?s? > |w? |?k ,
|x|?r? +1
|w? |?k
(24)
|x|?r?
?
<
.
(25)
There might be multiple pairs of (s, r) satisfying the inequalities (24)-(25), and we choose the pair
with the smallest ?|x|? ? |w|? ?2 . Finally, w? is obtained by sign-changing and reordering |w? |? to
conform to x.
Remark: The nonlinear equation (21) is quartic, for which we can use general formula to get all the
roots [18]. In addition, if it exists, the nonnegative root is unique, as shown in the proof [7].
Theorem 2 indicates that computing prox C (?) requires sorting of entries in |x| and a two?
dimensional search of s? and r? . Hence the total time complexity is O(p log p + k(p ? k)). However, a more careful observation can particularly reduce the search complexity from O(k(p ? k)) to
O(log k log(p ? k)), which is motivated by Theorem 3.
Theorem 3 In search of (s? , r? ) defined in Theorem 2, there can be only one r? for a given candidate
s? of s? , such that the inequality (25) is satisfied. Moreover if such r? exists, then for any r < r?, the
? ?k violates the first part of (25), and for r > r?, |w|
? ?k violates the second part of (25).
associated |w|
On the other hand, based on the r?, we have following assertion of s? ,
?
? > s? if r? does not exist
? ?k satisfies (24) .
? s? if r? exists and corresponding |w|
(26)
s?
?
? ?k violates (24)
< s? if r? exists but corresponding |w|
Based on Theorem 3, the accelerated search procedure for finding (s? , r? ) is to execute a twodimensional binary search, and Algorithm 2 gives the details. Therefore the total time complexity
becomes O(p log p + log k log(p ? k)). Compared with previous proximal operators for squared
k-support norm, this complexity is better than that in [1], and roughly the same as the one in [13].
3.2 Statistical Recovery Guarantees for k-support norm
The analysis of the generalized Dantzig Selector for k-support norm consists of addressing two key
challenges. First, note that Theorem 1 requires an appropriate choice of ?p . Second, one needs
to compute the Gaussian width of the subset of the error set TR (? ? ) ? Sp?1 . For the k-support
norm, we can get upper bounds to both of these quantities. We start by defining some notation. Let
G ? ? G (k) be the set of groups intersecting with the support of ? ? , and let S be the union of groups
in G ? , such that s = |S|. Then, we have the following bounds which are used for choosing ?p , and
bounding the Gaussian width.
Theorem 4 For the k-support norm Generalized Dantzig Selection problem (20), we obtain
B2
A4
$ ep (
.
E R? (XT w) ? k
+1
2 log
k
A4
( ? B2 C s D
$
CsD
+2 + k ?
+s.
2k log p ? k ?
?(TR (? ? ) ? Sp?1 )2 ?
k
k
6
(27)
(28)
Algorithm 2 Algorithm for computing prox
?
C?
(?) of ? ? ?sp
k
Input: x, k, ?
Output: w? = prox C (x)
?
sp?
1: if ?x?k ? ? then
2:
w? := x
3: else
4:
l := 0, u := k ? 1, and sort |x| to get |x|?
5:
while l ? u do
? based on (23)
6:
s? := ?(l + u)/2?, and binary search for r? that satisfies (25) and compute w
7:
if r? does not exist then
8:
l := s? + 1
9:
else if r? exists and (24) is satisfied then
? l := s? + 1
10:
w? := w,
11:
else if r? exists but (24) is not satisfied then
12:
u := s? ? 1
13:
end if
14:
end while
15: end if
Our analysis technique for these bounds follows [16]. Similar results were obtained in [8] in the
context of calculating norms of Gaussian vectors, and our bounds are of the same order as those
(2
$2
! "
+
1
2 log ep
, and under the assumptions of Theorem 1, we
of [8]. Choosing ?p = 2k
k
obtain the following result on the error bound for the minimizer of (20).
Corollary 1 Suppose that we obtain i.i.d. Gaussian measurements X, and the noise w is i.i.d.
N (0, 1). If we solve (20) with ?p chosen as above. Then, with high probability, we obtain
$2
?2
/
! " ? (
! " ? ?
+ k
8?? ??sp
2k log ep
sk log kp + sk
k
k
? .
?
??? ? ? ? ?2 ?
= O?
(29)
(?n ? ?(TR (? ? ) ? Sp?1 ))
n
Remark The error bound provides a natural interpretation for the two special cases of the k-support
norm, viz. k = 1 and k = p. First, for kA= 1 the k-support
norm is exactly the same as the L1 norm,
B
2
s log p
and the error bound obtained will be O
, the same as known results of DS, and shown in
n
Section 2.2. Second, for k = p, the k-support norm is equal to the L2 norm, and the error cone
"
!/ (11)
sp
is then simply a half space (there is no structural constraint) and the error bound scales as O
n .
4 Experimental Results
On the optimization side, we focus on the efficiency of different proximal operators related to ksupport norm. On the statistical side, we concentrate on the behavior and performance of GDS with
k-support norm. All experiments are implemented in MATLAB.
4.1 Efficiency of Proximal Operator
We tested four proximal operators related to k-support norm, which are normal prox C (?) in The?
1
orem 2 and the accelerated one in Theorem 3, prox 2?
2 (?) in [1], and prox ? ???2 (?) in [13].
(???sp
k )
2
?
The dimension p of vector varied from 1000 to 10000, and the ratio p/k = {200, 100, 50, 20}. As
illustrated in Figure 1, the speedup of accelerated prox C (?) is considerable compared with the
?
1
normal one and prox 2?
2 (?). It is also slightly better than the prox ? ???2 (?).
(???sp
k )
2
?
4.2 Statistical Recovery on Synthetic Data
Data generation We fixed p = 600, and ? ? = (10, . . . , 10, 10, . . . , 10, 10, . . . , 10, 0, 0, . . . , 0)
I JK L I JK L I JK L I JK L
10
10
10
570
throughout the experiment, in which nonzero entries were divided equally into three groups. The
design matrix X were generated from a normal distribution such that the entries in the same group
7
p / k = 200
p / k = 100
?2
?1
?2
?2
?3
?3
?4
?4
?4
10000
5000
p
10000
0
?1
?3
5000
p
1
0
log(time)
?1
p / k = 20
1
0
log(time)
0
log(time)
p / k = 50
1
log(time)
1
?1
?2
?3
5000
p
10000
?4
5000
p
10000
Figure 1: Efficiency of proximal operators. Diamond: proxIC (?) in Theorem 2, Square: prox
?
1 (???sp )2
k
2?
(?)
in [1], Downward-pointing triangle: prox ? ???2 (?) in [13], Upward-pointing triangle: accelerated proxIC (?)
?
2
?
in Theorem 3. For each (p, k), 200 vectors are randomly generated for testing. Time is measured in seconds.
have the same mean sampled from N (0, 1). X was normalized afterwards. The response vector y
was given by y = X? ? + 0.01 ? N (0, 1). The number of samples n is specified later.
ROC curves with different k We fixed n = 400 to obtain the ROC plot for k = {1, 10, 50} as
shown in Figure 2(a). ?p ranged from 10?2 to 103 . Small k gets better ROC curve.
60
k=1
k = 10
k = 50
?2 error : ??? ? ?? ?2
0.8
0.7
Mean of error
40
0.6
TPR
2.5
k=1
k = 10
k = 50
50
?2 error : ??? ? ?? ?2
1
0.9
0.5
1.5
30
0.4
20
0.3
0.2
2
1
0.5
10
0.1
0
0
0.2
0.4
FPR
0.6
(a) ROC curves
0.8
1
0
0
50
100
150
n
200
(b) L2 error vs. n
250
300
0
0
5
10
15
20
25
30
35
40
k
(c) L2 error vs. k
Figure 2: (a) The true positive rate reaches 1 quite early for k = 1, 10. When k = 50, the ROC gets worse
due to the strong smoothing effect introduced by large k. (b) For each k, the L2 error is large when the sample
is inadequate. As n increases, the error decreases dramatically for k = 1, 10 and becomes stable afterwards,
while the decrease is not that significant for k = 50 and the error remains relatively large. (c) Both mean and
standard deviation of L2 error are decreasing as k increases until it exceeds the number of nonzero entries in
?? , and then the error goes up for larger k.
L2 error vs. n We investigated how the L2 error ??? ? ? ? ?2 of Dantzig selector changes as the
number of samples increases, where k = {1, 10, 50} and n = {30, 60, 90, . . . , 300}. k = 1, 10 give
small errors when n is sufficiently large.
L2 error vs. k We also looked at the L2 error with different k. We again fixed n = 400 and
varied k from 1 to 39. For each k, we repeated the experiment 100 times, and obtained the mean and
standard deviation plot in Figure 2(c). The result shows that the k-support-norm GDS with suitable
k outperforms the L1 -norm DS (i.e. k = 1) when correlated variables present in data.
5 Conclusions
In this paper, we introduced the GDS, which generalizes the standard L1 -norm Dantzig selector to
estimation with any norm, such that structural information encoded in the norm can be efficiently
exploited. A flexible framework based on inexact ADMM is proposed for solving the GDS, which
only requires one of conjugate proximal operators to be efficiently solved. Further, we provide a
unified statistical analysis framework for the GDS, which utilizes Gaussian widths of certain restricted sets for proving consistency. In the non-trivial example of k-support norm, we showed that
the proximal operators used in the inexact ADMM can be computed more efficiently compared to
previously proposed variants. Our statistical analysis for the k-support norm provides the first result
of consistency of this structured norm. Further, experimental results provided sound support to the
theoretical development in the paper.
Acknowledgements
The research was supported by NSF grants IIS-1447566, IIS-1422557, CCF-1451986, CNS1314560, IIS-0953274, IIS-1029711, and by NASA grant NNX12AQ39A.
8
References
[1] Andreas Argyriou, Rina Foygel, and Nathan Srebro. Sparse prediction with the k-support
norm. In NIPS, pages 1466?1474, 2012.
[2] Arindam Banerjee, Sheng Chen, Farideh Fazayeli, and Vidyashankar Sivakumar. Estimation
with norm regularization. In NIPS, 2014.
[3] Peter J Bickel, Ya?acov Ritov, and Alexandre B Tsybakov. Simultaneous analysis of lasso and
dantzig selector. The Annals of Statistics, 37(4):1705?1732, 2009.
[4] Stephen P. Boyd, Neal Parikh, Eric Chu, Borja Peleato, and Jonathan Eckstein. Distributed
optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends in Machine Learning, 3(1):1?122, 2011.
[5] Emmanuel Candes and Terence Tao. The Dantzig selector: Statistical estimation when p is
much larger than n. The Annals of Statistics, 35(6):2313?2351, December 2007.
[6] Venkat Chandrasekaran, Benjamin Recht, Pablo A Parrilo, and Alan S Willsky. The convex
geometry of linear inverse problems. Foundations of Computational Mathematics, 12(6):805?
849, 2012.
[7] Soumyadeep Chatterjee, Sheng Chen, and Arindam Banerjee. Generalized dantzig selector:
Application to the k-support norm. ArXiv e-prints, June 2014.
[8] Yehoram Gordon, Alexander E. Litvak, Shahar Mendelson, and Alain Pajor. Gaussian averages
of interpolated bodies and applications to approximate reconstruction. Journal of Approximation Theory, 149(1):59?73, 2007.
[9] Gareth M. James and Peter Radchenko. A generalized Dantzig selector with shrinkage tuning.
Biometrika, 96(2):323?337, 2009.
[10] Gareth M. James, Peter Radchenko, and Jinchi Lv. DASSO: connections between the Dantzig
selector and lasso. Journal of the Royal Statistical Society Series B, 71(1):127?142, 2009.
[11] Han Liu, Jian Zhang, Xiaoye Jiang, and Jun Liu. The group Dantzig selector. In AISTATS,
pages 461?468, 2010.
[12] Zhaosong Lu, Ting Kei Pong, and Yong Zhang. An alternating direction method for finding
dantzig selectors. Computational Statistics & Data Analysis, 56(12):4037 ? 4046, 2012.
[13] Andrew M. McDonald, Massimiliano Pontil, and Dimitris Stamos. New perspectives on ksupport and cluster norms. ArXiv e-prints, March 2014.
[14] Sahand N Negahban, Pradeep Ravikumar, Martin J Wainwright, Bin Yu, et al. A unified framework for high-dimensional analysis of m-estimators with decomposable regularizers. Statistical Science, 27(4):538?557, 2012.
[15] Neal Parikh and Stephen Boyd. Proximal algorithms. Foundations and Trends in Optimization,
1(3):127?239, 2014.
[16] Nikhil S Rao, Ben Recht, and Robert D Nowak. Universal measurement bounds for structured
sparse signal recovery. In AISTATS, pages 942?950, 2012.
[17] Mark Rudelson and Roman Vershynin. On sparse reconstruction from Fourier and Gaussian
measurements. Communications on Pure and Applied Mathematics, 61(8):1025?1045, 2008.
[18] Ian Stewart. Galois Theory, Third Edition. Chapman Hall/CRC Mathematics Series. Taylor &
Francis, 2003.
[19] Robert Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal
Statistical Society. Series B (Methodological), pages 267?288, 1996.
[20] Huahua Wang and Arindam Banerjee. Bregman alternating direction method of multipliers. In
NIPS, 2014.
[21] Xiangfeng Wang and Xiaoming Yuan. The linearized alternating direction method of multipliers for Dantzig selector. SIAM Journal on Scientific Computing, 34(5), 2012.
[22] Peng Zhao and Bin Yu. On model selection consistency of lasso. The Journal of Machine
Learning Research, 7:2541?2563, 2006.
[23] Hui Zou and Trevor Hastie. Regularization and variable selection via the elastic net. Journal
of the Royal Statistical Society: Series B (Statistical Methodology), 67(2):301?320, 2005.
9
| 5295 |@word version:1 norm:80 linearized:3 decomposition:4 tr:8 liu:2 contains:2 series:4 selecting:1 interestingly:2 outperforms:1 existing:3 current:1 ka:1 chu:1 written:1 engg:1 enables:1 designed:1 drop:1 update:13 plot:2 v:4 half:1 fpr:1 lr:5 provides:3 simpler:1 zhang:2 yuan:1 consists:2 prove:1 introduce:1 peng:1 proxic:3 expected:1 behavior:1 roughly:1 inspired:1 rem:1 spherical:1 decreasing:2 cardinality:1 pajor:1 conv:1 provided:2 becomes:2 bounded:1 notation:2 moreover:1 argmin:10 minimizes:1 unified:2 finding:2 guarantee:4 exactly:1 biometrika:1 control:1 unit:10 grant:2 positive:1 negligible:1 encoding:1 jiang:1 establishing:1 subscript:1 path:1 sivakumar:1 might:2 dantzig:24 challenging:1 unique:1 testing:1 atomic:1 union:1 procedure:1 litvak:1 pontil:1 universal:1 empirical:2 boyd:2 get:5 convenience:1 interior:1 selection:7 operator:20 cannot:1 twodimensional:1 context:1 xiangfeng:1 equivalent:1 lagrangian:1 go:1 convex:6 focused:3 simplicity:1 recovery:10 immediately:1 decomposable:1 pure:1 estimator:1 supz:1 proving:1 notion:1 annals:2 suppose:2 user:1 element:1 trend:2 expensive:1 satisfying:1 particularly:1 jk:4 ep:3 subproblem:1 solved:1 wang:2 ensures:1 connected:1 rina:1 decrease:2 benjamin:1 pong:1 complexity:4 solving:4 inapplicable:2 efficiency:3 eric:1 triangle:2 instantiated:1 fast:1 shortcoming:1 massimiliano:1 kp:1 choosing:2 quite:3 encoded:1 larger:3 solve:5 valued:1 nikhil:1 tested:1 otherwise:2 statistic:3 net:4 propose:3 reconstruction:2 flexibility:1 inducing:1 exploiting:1 convergence:2 empty:1 p:2 cluster:1 ben:1 derive:1 develop:1 andrew:1 measured:1 strong:1 implemented:1 c:1 direction:7 concentrate:1 radius:1 violates:3 bin:2 crc:1 suffices:1 generalization:1 homotopy:1 alleviate:1 proposition:3 extension:1 hold:1 sufficiently:1 considered:1 hall:1 normal:3 exp:1 proxg:1 pointing:2 bickel:1 early:1 smallest:1 estimation:15 applicable:3 radchenko:2 agrees:1 city:1 gaussian:24 ksupport:2 ck:1 shrinkage:2 corollary:1 focus:3 viz:1 june:1 vk:16 methodological:1 likelihood:1 indicates:1 contrast:1 sense:1 entire:1 stamos:1 subroutine:2 tao:1 upward:1 dual:8 flexible:2 development:1 smoothing:1 special:2 initialize:1 equal:1 construct:1 asr:2 having:1 chapman:1 yu:2 tl1:2 simplex:1 piecewise:1 gordon:1 primarily:1 few:1 roman:1 randomly:1 divergence:1 replaced:1 geometry:1 investigate:1 zhaosong:1 umn:1 fazayeli:1 pradeep:1 yielding:2 primal:2 regularizers:2 bregman:3 nowak:1 plugged:1 taylor:1 theoretical:2 column:2 soft:1 rao:1 assertion:1 stewart:1 farideh:1 cost:2 addressing:1 entry:9 subset:2 deviation:2 inadequate:1 reported:1 shengc:1 proximal:23 gd:22 synthetic:1 recht:2 vershynin:1 negahban:1 siam:1 terence:1 intersecting:1 squared:2 again:1 satisfied:3 containing:1 leveraged:1 choose:1 worse:1 zhao:1 supp:1 prox:24 parrilo:1 twin:1 b2:2 satisfy:1 notable:1 depends:1 later:1 root:4 closed:1 sup:1 francis:1 start:2 sort:1 candes:1 square:4 variance:1 efficiently:9 correspond:1 none:1 lu:1 converged:1 simultaneous:1 reach:1 trevor:1 definition:2 inexact:8 james:2 proof:3 associated:1 sampled:1 proved:1 organized:1 nasa:1 proxh:2 alexandre:1 methodology:1 response:1 formulation:2 done:1 arranged:1 execute:1 ritov:1 just:1 until:1 d:10 sheng:3 hand:2 nonlinear:2 banerjee:5 nonsmoothness:1 scientific:1 effect:1 normalized:2 multiplier:6 true:2 counterpart:1 ranged:1 hence:3 regularization:3 ccf:1 alternating:6 nonzero:2 neal:2 illustrated:1 eg:1 width:13 proxf:1 linearizing:1 generalized:10 comk:1 mcdonald:1 l1:12 arindam:4 novel:1 recently:2 parikh:2 interpretation:1 tpr:1 elementwise:1 measurement:3 significant:1 tuning:1 consistency:4 mathematics:3 similarly:2 minnesota:1 stable:1 han:1 similarity:1 vidyashankar:1 recent:1 showed:1 quartic:1 perspective:1 scenario:2 certain:3 inequality:3 binary:2 shahar:1 exploited:1 subtraction:1 converge:1 signal:1 ii:4 stephen:2 multiple:2 afterwards:2 sound:1 reduces:1 stem:1 exceeds:1 technical:1 match:1 borja:1 alan:1 sphere:1 divided:1 equally:2 ravikumar:1 prediction:1 involving:1 regression:3 variant:1 arxiv:2 addition:1 else:3 jian:1 rest:2 unlike:1 sr:2 december:1 structural:2 leverage:1 easy:1 hastie:1 lasso:6 reduce:1 andreas:1 computable:1 motivated:2 sahand:1 penalty:2 peter:3 remark:4 matlab:1 dramatically:1 generally:1 clear:1 amount:1 tsybakov:1 reduced:1 exist:2 nsf:1 sign:2 tibshirani:1 conform:1 group:7 thereafter:1 key:2 four:1 changing:1 ce:1 cone:6 inverse:1 throughout:1 chandrasekaran:1 utilizes:1 bound:18 quadratic:3 nonnegative:2 constraint:3 yong:1 interpolated:1 aspect:3 nathan:1 fourier:1 min:2 span:1 relatively:1 martin:1 xiaoming:1 speedup:1 structured:4 according:1 yehoram:1 ball:7 march:1 conjugate:7 slightly:1 intimately:1 b:2 restricted:2 computationally:1 equation:2 remains:1 previously:1 turn:1 nnx12aq39a:1 fail:1 foygel:1 needed:2 know:1 end:4 generalizes:1 operation:1 spectral:1 appropriate:1 alternative:1 rp:11 original:1 denotes:1 rudelson:1 a4:2 calculating:1 giving:1 ting:1 emmanuel:1 establish:2 society:3 print:2 quantity:1 looked:1 considers:1 trivial:3 willsky:1 length:3 besides:1 index:1 ratio:1 robert:2 design:3 proper:1 contributed:1 diamond:1 upper:4 conversion:1 observation:2 behave:1 defining:1 extended:1 communication:1 rn:3 varied:2 peleato:1 introduced:3 pablo:1 pair:2 eckstein:1 specified:1 connection:1 huahua:1 established:2 nip:3 acov:1 dimitris:1 jinchi:1 challenge:1 orem:1 program:2 recast:1 max:2 royal:3 dasso:2 wainwright:1 suitable:3 natural:1 rely:1 regularized:4 indicator:3 jun:1 soumyadeep:2 literature:1 l2:10 acknowledgement:1 asymptotic:3 reordering:1 permutation:1 generation:1 srebro:1 lv:1 foundation:3 thresholding:1 supported:1 theo:1 alain:1 side:5 absolute:1 sparse:8 moreau:2 distributed:1 overcome:1 dimension:1 curve:3 author:1 kei:1 approximate:1 selector:20 confirm:1 assumed:1 conclude:1 xi:2 search:6 iterative:1 sk:2 zk:19 elastic:4 investigated:1 zou:1 sp:23 aistats:2 main:1 csd:1 s2:1 noise:3 bounding:1 edition:1 repeated:1 body:1 augmented:1 venkat:1 roc:5 fails:1 candidate:1 third:1 ian:1 theorem:14 formula:1 xt:13 exists:8 mendelson:1 sequential:1 kr:3 hui:1 downward:1 chatterjee:2 chen:3 sorting:1 intersection:2 simply:3 lagrange:1 temporarily:1 minimizer:2 satisfies:2 gareth:2 shell:1 viewed:1 identity:1 goal:1 careful:1 admm:14 feasible:1 considerable:1 change:1 total:2 experimental:4 ya:1 exception:1 support:37 mark:1 jonathan:1 alexander:1 accelerated:4 dept:1 argyriou:1 instructive:1 correlated:2 |
4,744 | 5,296 | Parallel Feature Selection inspired by Group Testing
Yingbo Zhou?
Utkarsh Porwal?
CSE Department
SUNY at Buffalo
{yingbozh, utkarshp}@buffalo.edu
Ce Zhang
CS Department
University of Wisconsin-Madison
[email protected]
Hung Ngo
CSE Department
SUNY at Buffalo
[email protected]
Christopher R?e
CS Department
Stanford University
[email protected]
XuanLong Nguyen
EECS Department
University of Michigan
[email protected]
Venu Govindaraju
CSE Department
SUNY at Buffalo
[email protected]
Abstract
This paper presents a parallel feature selection method for classi?cation that scales
up to very high dimensions and large data sizes. Our original method is inspired
by group testing theory, under which the feature selection procedure consists of a
collection of randomized tests to be performed in parallel. Each test corresponds
to a subset of features, for which a scoring function may be applied to measure
the relevance of the features in a classi?cation task. We develop a general theory providing suf?cient conditions under which true features are guaranteed to
be correctly identi?ed. Superior performance of our method is demonstrated on
a challenging relation extraction task from a very large data set that have both
redundant features and sample size in the order of millions. We present comprehensive comparisons with state-of-the-art feature selection methods on a range of
data sets, for which our method exhibits competitive performance in terms of running time and accuracy. Moreover, it also yields substantial speedup when used
as a pre-processing step for most other existing methods.
1
Introduction
Feature selection (FS) is a fundamental and classic problem in machine learning [10, 4, 12]. In
classi?cation, FS is the following problem: Given a universe U of possible features, identify a
subset of features F ? U such that using the features in F one can build a model to best predict
the target class. The set F not only in?uences the model?s accuracy, its computational cost, but also
the ability of an analyst to understand the resulting model. In applications, such as gene selection
from micro-array data [10, 4], text categorization [3], and ?nance [22], U may contain hundreds of
thousands of features from which one wants to select only a small handful for F .
While the overall goal is to have an FS method that is both computationally ef?cient and statistically
sound, natural formulations of the FS problem are known to be NP-hard [2]. For large scale data,
scalability is a crucial criterion, because FS often serves not as an end but a means to other sophisticated subsequent learning. In reality, practitioners often resort to heuristic methods, which can
broadly be categorized into three types: wrapper, embedded, and ?lter [10, 4, 12]. In the wrapper
method, a classi?er is used as a black-box to test on any subset of features. In ?lter methods no
classi?er is used; instead, features are selected based on generic statistical properties of the (labeled)
?
* denotes equal contribution
1
data such as mutual information and entropy. Embedded methods have built in mechanisms for FS
as an integral part of the classi?er training. Devising a mathematically rigorous framework to explain and justify FS heuristics is an emerging research area. Recently Brown et al. [4] considered
common FS heuristics using a formulation based on conditional likelihood maximization.
The primary contribution of this paper is a new framework for parallelizable feature selection, which
is inspired by the theory of group testing. By exploiting parallelism in our test design we obtain a
FS method that is easily scalable to millions of features and samples or more, while preserving
useful statistical properties in terms of classi?cation accuracy, stability and robustness. Recall that
group testing is a combinatorial search paradigm [7] in which one wants to identify a small subset of
?positive items? from a large universe of possible items. In the original application, items are blood
samples of WWII draftees and an item is positive if it is infected with syphilis. Testing individual
blood sample is very expensive; the group testing approach is to distribute samples into pools in
a smart way. If a pool is tested negative, then all samples in the pool are negative. On the other
hand, if a pool is tested positive then at least one sample in the pool is positive. We can think of
the FS problem in the group testing framework: there is a presumably small, unknown subset F of
relevant features in a large universe of N features. Both FS and group testing algorithms perform
the same basic operation: apply a ?test? to a subset T of the underlying universe; this test produces
a score, s(T ), that is designed to measure the quality of the features T (or return positive/negative
in the group testing case). From the collection of test scores the relevant features are supposed to
be identi?ed. Most existing FS algorithms can be thought of as sequential instantiations in this
framework1 : we select the set T to test based on the scores of previous tests. For example, let X =
(X1 , . . . , XN ) be a collection of features (variables) and Y be the class label. In the joint mutual
information (JMI) method [25], the feature set T is grown sequentially by adding one feature at each
iteration. The?
next feature?s score, s(Xk ), is de?ned relative to the set of features already selected in
T : s(Xk ) = Xj ?T I(Xk , Xj ; Y ). As each such scoring operation takes a non-negligible amount
of time, a sequential method may take a long time to complete.
A key insight is that group testing needs not be done sequentially. With a good pooling design, all
the tests can be performed in parallel in which we determine the pooling design without knowing
any pool?s test outcome. From the vector of test outcomes, one can identify exactly the collection
of positive blood samples. Parallel group testing, commonly called non-adaptive group testing
(NAGT) is a natural paradigm and has found numerous applications in many areas of mathematics,
computer Science, and biology [18]. It is natural to wonder whether a ?parallel? FS scheme can be
designed for machine learning in the same way NAGT was possible: all feature sets T are speci?ed
in advance, without knowing the scores of any other tests, and from the ?nal collection of scores the
features are identi?ed. This paper initiates a mathematical investigation of this possibility.
At a high level, our parallel feature selection (PFS) scheme has three inter-related components: (1)
the test design indicates the collection of subsets of features to be tested, (2) the scoring function
s : 2[N ] ? R that assigns a score to each test, and (3) the feature identi?cation algorithm that
identi?es the ?nal selected feature set from the test scores. The design space is thus very large. Every
combination of the three components leads to a new PFS scheme.2 We argue that PFS schemes are
preferred over sequential FS for two reasons:
1. scalability, the tests in a PFS schem can be performed in parallel, and thus the scheme can
be scaled to large datasets using standard parallel computing techniques, and
2. stability, errors in individual trials do not affect PFS methods as dramatically as sequential
methods. In fact, we will show in this paper that increasing the number of tests improves
the accuracy of our PFS scheme.
We propose and study one such PFS approach. We show that our approach has comparable (and
sometimes better) empirical quality compared to previous heuristic approaches while providing
sound statistical guarantees and substantially improved scalability.
Our technical contributions We propose a simple approach for the ?rst and the third components
of a PFS scheme. For the second component, we prove a suf?cient condition on the scoring function
under which the feature identi?cation algorithm we propose is guaranteed to identify exactly the set
1
A notable exception is the MIM method, which is easily parallelizable and can be regarded as a special
implementation of our framework
2
It is important to emphasize that this PFS framework is applicable to both ?lter and wrapper approaches.
In the wrapper approach, the score s(T ) might be the training error of some classi?er, for instance.
2
of original (true) features. In particular, we introduce a notion called C-separability, which roughly
indicates the strength of the scoring function in separating a relevant feature from an irrelevant
feature. We show that when s is C-separable and we can estimate s, we are able to guarantee exact
recovery of the right set of features with high probability. Moreover, when C > 0, the number of
tests can be asymptotically logarithmic in the number of features in U .
In theory, we provide suf?cient conditions (a Na??ve Bayes assumption) according to which one can
obtain separable scoring functions, including the KL divergence and mutual information (MI). In
practice, we demonstrate that MI is separable even when the suf?cient condition does not hold,
and moreover, on generated synthetic data sets, our method is shown recover exactly the relevant
features. We proceed to provide a comprehensive evaluation of our method on a range of real-world
data sets of both large and small sizes. It is the large scale data sets where our method exhibits
superior performance. In particular, for a huge relation extraction data set (TAC-KBP) that has
millions redundant features and samples, we outperform all existing methods in accuracy and time,
in addition to generating plausible features (in fact, many competing methods could not ?nish the
execution). For the more familiar NIPS 2013 FS Challenge data, our method is also competitive
(best or second-best) on the two largest data sets. Since our method hinges on the accuracy of score
functions, which is dif?cult achieve for small data, our performance is more modest in this regime
(staying in the middle of the pack in terms of classi?cation accuracy). Nonetheless, we show that our
method can be used as a preprocessing step for other FS methods to eliminate a large portion of the
feature space, thereby providing substantial computational speedups while retaining the accuracy of
those methods.
2
Parallel Feature Selection
The general setting Let N be the total number of input features. For each subset T ? [N ] :=
{1, . . . , N }, there is a score s(T ) normalized to be in [0, 1] that assesses the ?quality? of features in
T . We select a collection of t tests, each of which is a subset T ? [N ] such that from the scores
of all tests we can identify the unknown subset F of d relevant variables that are most important
to the classi?cation task. We encode the collection of t tests with a binary matrix A = (aij ) of
dimension t ? N , where aij = 1 iff feature j belongs to test i. Corresponding to each row i of A is
a ?test score? si = s({j | aij = 1}) ? [0, 1]. Specifying A is called test design, identifying F from
the score vector (si )i?[t] is the job of the feature identi?cation algorithm. The scheme is inherently
parallel because all the tests must be speci?ed in advance and executed in parallel; then the features
are selected from all the test outcomes.
Test design and feature identi?cation Our test design and feature identi?cation algorithms are
extremely simple. We construct the test matrix A randomly by putting a feature in the test with
probability p (to be chosen later). Then, from the test scores we rank the features and select d
top-ranked features. The ranking function is de?ned as follows. Given a t ? N test matrix A, let
aj denote its jth column. The dot-product ?aj , s? is the total score of all the tests that feature j
participates in. We de?ne ?(j) = ?aj , s? to be the rank of feature j with respect to the test matrix
A and the score function s.
The scoring function The crucial piece stiching together the entire scheme is the scoring function. The following theorem explains why the above test design and feature identi?cation strategy
make sense, as long as one can choose a scoring function s that satis?es a natural separability property. Intuitively, separable scoring functions require that adding more hidden features into a test set
increase its score.
De?nition 2.1 (Separable scoring function). Let C ? 0 be a real number. The score function
s : 2[N ] ? [0, 1] is said to be C-separable if the following property holds: for every f ? F and
f? ?
/ F , and for every T ? [N ] ? {f, f?}, we have s(T ? {f }) ? s(T ? {f?}) ? C.
In words, with a separable scoring function adding a relevant feature should be better than adding
an irrelevant feature to a given subset T of features. Due to space limination, the proofs of the
following theorem, propositions, and corollaries can be found in the supplementary materials. The
essence of the idea is that, when s can separate relevant features from irrelevant features, with high
probability a relevant feature will be ranked higher than an irrelevant feature. Hoeffding?s inequality
is then used to bound the number of tests.
3
Theorem 2.2. Let A be the random t ? N test matrix obtained by setting each entry to be 1 with
probability p ? [0, 1] and 0 with probability 1 ? p. If the scoring function s is C-separable, then the
expected rank of a feature in F is at least the expected rank of a feature not in F .
Furthermore, if C > 0, then for any ? ? (0, 1), with probability at least 1 ? ? every feature in F
has rank higher than every feature not in F , provided that the number of tests t satis?es
?
?
d(N ? d)
2
t? 2 2
log
.
(1)
C p (1 ? p)2
?
By setting p = 1/2 in the above theorem, we obtain the following. It is quite remarkable that,
assuming we can estimate the scores accurately, we only need about O(log N ) tests to identify F .
Corollary 2.3. Let C > 0 be a constant such that there is a C-separable scoring function s. Let
d = |F |, where F is the set of hidden features. Let ? ? (0, 1) be an arbitrary constant. Then, there
is a distribution of t ? N test matrices A with t = O(log(d(N ? d)/?)) such that, by selecting a
test matrix randomly from the distribution, the d top-ranked features are exactly the hidden features
with probability at least 1 ? ?.
Of course, in reality estimating the scores accurately is a very dif?cult problem, both statistically
and computationally, depending on what the scoring function is. We elaborate more on this point
below. But ?rst, we show that separable scoring functions exist, under certain assumption about the
underlying distribution.
Suf?cient conditions for separable scoring functions We demonstrate the existence of separable
scoring functions given some suf?cient conditions on the data. In practice, loss functions such as
classi?cation error and other surrogate losses may be used as scoring functions. For binary classi?cation, information-theoretic quantities such as Kullback-Leibler divergence, Hellinger distance and
the total variation ? all of which special cases of f -divergences [5, 1] ? may also be considered.
For multi-class classi?cation, mutual information (MI) is a popular choice.
The data pairs (X, Y ) are assumed to be iid samples from a joint distribution P (X, Y ). The following result shows that under the so-called ?naive Bayes? condition, i.e., all components of random
vector X are conditionally independent given label variable Y , the Kullback-Leibler distance is a
separable scoring function in a binary classi?cation setting:
Proposition 2.4. Consider the binary classi?cation setting, i.e., Y ? {0, 1} and assume that the
naive Bayes condition holds. De?ne score function to be the Kullback-Leibler divergence:
s(T ) := KL(P (X T |Y = 0)||P (X T |Y = 1)).
Then s is a separable scoring function. Moreover, s is C-separable, where C := minf ?F s(f ).
Proposition 2.5. Consider the multi-class classi?cation setting, and assume that the naive Bayes
condition holds. Moreover, for any pair f ? F and f? ?
/ F , the following holds for any T ?
[N ] ? {f, f?}
I(Xf ; Y ) ? I(Xf ; X T ) ? I(Xf?; Y ) ? I(Xf?; X T ).
Then, the MI function s(T ) := I(XT ; Y ) is a separable scoring function.
We note the naturalness of the condition so required, as quantity I(Xf ; Y ) ? I(Xf ; XT ) may be
viewed as the relevance of feature f with respect to the label Y , subtracted by the redundancy with
other existing features T . If we assume further that X f? is independent of both X T and the label Y ,
and there is a positive constant C such that I(Xf ; Y ) ? I(Xf ; XT ) ? C for any f ? F , then s(T )
is obviously a C-separable scoring function. It should be noted that the naive Bayes conditions are
suf?cient, but not necessary for a scoring function to be C-separable.
Separable scoring functions for ?lters and wrappers. In practice, information-based scoring
functions need to be estimated from the data. Consistent estimators of scoring functions such as KL
divergence (more generally f -divergences) and MI are available (e.g., [20]). This provides the theoretical support for applying our test technique to ?lter methods: when the number of training data is
suf?ciently large, a consistent estimate of a separable scoring function must also be a separable scoring function. On the other hand, a wrapper method uses a classi?cation algorithm?s performance as
a scoring function for testing. Therefore, the choice of the underlying (surrogate) loss function plays
a critical role. The following result provides the existence of loss functions which induce separable
scoring functions for the wrapper method:
4
T
Proposition 2.6. Consider the binary classi?cation setting, and let P0T :=
? P (X T |Y = 0), P1 :=
P (X T |Y = 1). Assume that an f -divergence of the form: s(T ) = ?(dP0T /dP1T )dP1T is a
separable scoring function for some convex function ? : R+ ? R. Then there exists a surrogate
loss function l : R ? R ? R+ under which the minimum l-risk: Rl (T ) := inf g E [l(Y, g(X T ))] is
also a separable scoring function. Here the in?mum is taken over all measurable classi?er functions
g acting on feature input X T , E denotes expectation with respect to the joint distribution of X T
and Y .
This result follows from Theorem 1 of [19], who established a precise correspondence between f divergences de?ned by convex ? and equivalent classes of surrogate losses l. As a consequence,
if the Hellinger distance between P0T and P1T is separable, then the wrapper method using the
Adaboost classi?er corresponds to a separable scoring function. Similarly, a separable KullbackLeibler divergence implies that of a logistic regression based wrapper; while a separable variational
distance implies that of a SVM based wrapper.
3
Experimental results
3.1 Synthetic experiments
In this section, we synthetically illustrate that separable scoring functions exist and our PFS framework is sound beyond the Na??ve Bayes assumption (NBA). We ?rst show that MI is C-separable for
large C even when the NBA is violated. The NBA was only needed in Propositions 2.4 and 2.5 in
order for the proofs to go through. Then, we show that our framework recovers exactly the relevant
features for two common classes of input distributions.
We generate 1, 000 data points from two separated 2-D Gaussians with the same covariance matrix but different means, one centered
at (?2, ?2) and the other at (2, 2). We start
with the identity covariance matrix, and gradually change the off diagonal element to ?0.999,
representing highly correlated features. Then,
we add 1,000 dimensional zero mean Gaussian
noise with the same covariance matrix, where
the diagonal is 1 and the off-diagonal elements
increases from 0 gradually to 0.999. We then
calculate the MI between two features and the Figure 1: Illustration of MI as a separable scorclass label, and the two features are selected in ing function for the case of statistically dependent
three settings: 1) the two genuine dimensions; features. The top left point shows the scores for
2) one of the genuine feature and one from the the 1st setting; the middle points shows the scores
noisy dimensions; 3) two random pair from the for the 2nd setting; and the bottom points shows
noisy dimensions. The MI that we get from the scores for the 3rd setting.
these three conditions is shown in Figure 1. It is clear from this ?gure MI is a separable scoring
function, despite the fact that the NBA is violated.
We also synthetically evaluated our entire PFS idea, using two multinomials and two Gaussians to
generate two binary classi?cation task data. Our PFS scheme is able to capture exactly the relevant
features in most cases. Details are in the supplementary material section due to lack of space.
3.2 Real-world data experiment results
This section evaluates our approach in terms of accuracy, scalability, and robustness accross a range
of real-world data sets: small, medium, and large. We will show that our PFS scheme works very
well on medium and large data sets; because, as was shown in Section 3.1, with suf?cient data to
estimate test scores, we expect our method to work well in terms of accuracy. On the small datasets,
our approach is only competitive and does not dominate existing approaches, due to the lack of data
to estimate scores well. However, we show that we can still use our PFS scheme as a pre-processing
step to ?lter down the number of dimensions; this step reduces the dimensionality, helps speed up
existing FS methods from 3-5 times while keeps their accuracies.
3.2.1 The data sets and competing methods
Large: TAC-KBP is a large data set with the number of samples and dimensions in the millions3 ;
its domain is on relation extraction from natural language text. Medium: GISETTE and MADE3
http://nlp.cs.qc.cuny.edu/kbp/2010/
5
LON are two largest data sets from the NIPS 2003 feature selection challenge4 , with the number of
dimensions in the thousands. Small: Colon, Leukemia, Lymph, NCI9, and Lung are chosen from
the small Micro-array datasets [6], along with the UCI datasets5 . These sets typically have a few
hundreds to a few thousands variables, with only tens of data samples.
We compared our method with various baseline methods including mutual information
maximization[14] (MIM), maximum relevancy minimum redundancy[21] (MRMR), conditional
mutual information maximization[9] (CMIM), joint mutual information[25] (JMI), double input
symmetrical relevance[16] (DISR), conditional infomax feature extraction[15] (CIFE), interaction
capping[11] (ICAP), fast correlation based ?lter[26] (FCBF), local learning based feature selection
[23] (LOGO), and feature generating machine [24] (FGM).
3.2.2
Accuracy
?????????
??
?????
????
????
??
????
??
????
??????
????
?????????????
??????
????
????????
????????
????????
????????
????????
???????
???????
??????
??????????
????????????
?????????
?????
?????????
?????????
???????
?????????
????????????????????????????
Figure 2: Result from different methods on TAC-KBP dataset. (a) Precision/Recall of different
methods; (b) Top-5 keywords appearing in the Top-20 features selected by our method. Dotted lines
in (a) are FGM (or MIM) with our approach as pre-processing step.
Accuracy results on large data set. As shown in Figure 2(a), our method dominates both MIM
and FGM. Given the same precision, our method achieves 2-14? higher recall than FGM, and 1.22.4? higher recall than MIM. Other competitors do not ?nish execution in 12 hours. We compare the
top-features produced by our method and MIM, and ?nd that our method is able to extract features
that are strong indicators only when they are combined with other features, while MIM, which tests
features individually, ignores this type of combination. We then validate that the features selected
by our method makes intuitive sense. For each relation, we select the top-20 features and report the
keyword in these features.6 As shown in Figure 2(b), these top-features selected by our method are
good indicators of each relation. We also observe that using our approach as the pre-processing step
improves the quality of FGM signi?cantly. In Figure 2(a) (the broken lines), we run FGM (MIM)
on the top-10K features produced by our approach. We see that running FGM with pre-processing
achieves up to 10? higher recall given the same precision than running FGM on all 1M features.
Accuracy results on medium data sets Since the focus of the evaluation is to analyze the ef?cacy
of feature selection approaches, we employed the same strategy as Brown et al.[4] i.e. the ?nal
classi?cation is done using k-nearest neighbor classi?er with k ?xed to three, and applied Euclidean
distance7 .
We denote our method by Fk (and Wk ), where F denotes ?lter (and W denotes wrapper method).
k denotes the number of tests (i.e. let N be the dimension of data, then the total number of tests is
kN ). We bin each dimension of the data into ?ve equal distanced bins when the data is real valued,
otherwise the data is not processed8 . MI is used as the scoring function for ?lter method, and loglikelihood is used for scoring the wrapper method. The wrapper we used is logistic regression9 .
4
5
6
http://www.nipsfsc.ecs.soton.ac.uk/datasets/
http://archive.ics.uci.edu/ml/
Following the syntax used by Mintz et al. [17], if a feature has the form [?poss wif e ?prop of ], we report
the keyword as wife in Figure 2(b).
7
The classi?er for FGM is linear support vector machine (SVM), since it optimized for the SVM criteria.
8
For SVM based method, the real valued data is not processed, and all data is normalized to have unit length.
9
The logistic regressor used in wrapper is only to get the testing scores, the ?nal classi?cation scheme is
still k-NN.
6
(a)
(b)
Figure 3: Result from real world datasets: a) curve showing the ratio between the errors of various
methods applied on original data and on ?ltered data, where a large portion of the dimension is
?ltered out (value larger than one indicates performance improvement); b) the speed up we get by
applying our method as a pre-processing method on various methods across different datasets, the
?at dashed line indicates the location where the speed up is one.
For GISETTE we select up to 500 features and for MADELON we select up to 100 features. To get
the test results, we use the features according to the smallest validation error for each method, and
the results on test set are illustrated in table 4.
Table 1: Test set balanced error rate (%) from different methods on NIPS datasets
Datasets
Best
Perf.
2nd Best
Perf.
3rd Best
Perf.
Median
Perf.
Ours
(F3 )
Ours
(W3 )
Ours
(F10 )
Ours
(W10 )
GISETTE
MADELON
2.15
10.61
3.06
11.28
3.09
12.33
3.86
25.92
4.85
22.61
2.72
10.17
4.69
18.39
2.89
10.50
Accuracy results on the small data sets. As expected, due to the lack of data to estimate scores,
our accuracy performance is average for this data set. Numbers can be found in the supplementary
materials. However, as suggested by theorem A.3 (in supplementary materials), our method can also
be used as a preprocessing step for other feature selection method to eliminate a large portion of the
features. In this case, we use the ?lter methods to ?lter out e + 0.1 of the input features, where e is
the desired proportion of the features that one wants to reserve.
Using our method as preprocessing step achieves 3-5 times speedup as compare to the time spend
by original methods that take multiple passes through the datasets, and keeps or improves the performance in most of the cases (see ?gure 3 a and b). The actual running time can be found in
supplementary materials.
Scalability
???????????????
3.2.3
3600000
?
360000
?
?????? ??????
3600000
?
?????????????????
3600000
?
360000
?
360000
?
36000
?
36000
?
3600
?
3600
?
??????????????
???????????????
36000
?
???????????
3600
?
360
?
1
?
10
?
100
?
???????????
1000
?
360
?
10000
?
100000
?
1000000
?
??????????????
360
?
10000
?
100000
?
1000000
?
??????????????
Figure 4: Scalability Experiment of Our Approach
We validate that our method is able to run on large-scale data set ef?ciently, and the ability to take
advantage of parallelism is the key to its scalability.
7
Experiment Setup Given the TAC-KBP data set, we report the execution time by varying the
degree of parallelism, number of features, and number of examples. We ?rst produce a series of
data sets by sub-sampling the original data set with different number examples ({104 , 105 , 106 })
and number of features ({104 , 105 , 106 }). We also try different degree of parallelism by running
our approach using a single thread, 4-threads on a 4-core CPU, 32 threads on a single 8-CPU (4core/CPU) machine, and multiple machines available in the national Open Science Grid (OSG).
For each combination of number of features, number of examples, and degree of parallelism, we
estimate the throughput as the number of tests that we can run in 1 second, and estimate the total
running time accordingly. We also ran our largest data set (106 rows and 106 columns) on OSG and
report the actual run time.
Degree of Parallelism Figure 4(a) reports the (estimated) run time on the largest data set (106
rows and 106 columns) with different degree of parallelism. We ?rst observe that running our
approach requires non-trivial amount of computational resources?if we only use a single thread, we
need about 400 hours to ?nish our approach. However, the running time of our approach decreases
linearly with the number of cores that we used. If we run our approach on a single machine with 32
cores, it ?nishes in just 11 hours. This linear speed-up behavior allows our approach to scale to very
large data set?when we run our approach on the national Open Science Grid, we observed that our
approach is able to ?nish in 2.2 hours (0.7 hours for actual execution, and 1.5 hours for scheduling
overhead).
The Impact of Number of Features and Number of Examples Figure 4(b,c) report the run time
with different number of features and number of examples, respectively. In Figure 4(b), we ?x the
number of examples to be 105 , and vary the number of features, and in Figure 4(c), we ?x the number
of features to be 106 and vary the number of examples. We see that as the number of features or the
number of examples increase, our approach uses more time; however, the running time never grows
super-linearly. This behavior implies the potential of our approach to scale to even larger data sets.
3.2.4
Stability and robustness
Our method exhibits several robustness properties. In particular, the proof of Theorem 2.2 suggests
that as the number of tests are increased the performance also improves. Therefore, in this section
we empirically evaluate this observation. We picked four datasets: KRVSKP, Landset, Splice and
Waveform from the UCI datasets and both NIPS datasets.
(a)
(b)
(c)
(d)
Figure 5: Change of performance with respect of number of tests on several UCI datasets with (a)
?lter and (b) wrapper methods; and (c) GISETTE and (d) MADELON datasets.
The trend is pretty clear as can be observed from ?gure 5. The performance of both wrapper and
?lter methods improves as we increase the number of tests, which can be attributed to the increase of
robustness against inferior estimates for the test scores as the number of tests increases. In addition,
apart from MADELON dataset, the performance converges fast, normally around k = 10 ? 15.
Additional stability experiments can be found in the supplementary materials, where we evaluate
ours and other methods in terms of consistency index.
References
[1] S. M. Ali and S. D. Silvey. A general class of coef?cients of divergence of one distribution from another.
J. Royal Stat. Soc. Series B, 28:131?142, 1966.
[2] Edoardo Amaldi and Viggo Kann. On the approximability of minimizing nonzero variables or unsatis?ed
relations in linear systems, 1997.
8
[3] Ron Bekkerman, Ran El-Yaniv, Naftali Tishby, and Yoad Winter. Distributional word clusters vs. words
for text categorization. J. Mach. Learn. Res., 3:1183?1208, March 2003.
[4] Gavin Brown, Adam Pocock, Ming-Jie Zhao, and Mikel Luj?an. Conditional likelihood maximisation: A
unifying framework for information theoretic feature selection. JMLR, 13:27?66, 2012.
[5] I. Csisz?ar. Information-type measures of difference of probability distributions and indirect observation.
Studia Sci. Math. Hungar, 2:299?318, 1967.
[6] C. H. Q. Ding and H. Peng. Minimum redundancy feature selection from microarray gene expression
data. J. Bioinformatics and Computational Biology, pages 185?206, 2005.
[7] Ding-Zhu Du and Frank K. Hwang. Combinatorial group testing and its applications, volume 12 of Series
on Applied Mathematics. World Scienti?c Publishing Co. Inc., River Edge, NJ, second edition, 2000.
[8] Devdatt P. Dubhashi and Alessandro Panconesi. Concentration of measure for the analysis of randomized
algorithms. Cambridge University Press, Cambridge, 2009.
[9] Francois Fleuret and Isabelle Guyon. Fast binary feature selection with conditional mutual information.
Journal of Machine Learning Research, 5:1531?1555, 2004.
[10] Isabelle Guyon and Andr?e Elisseeff. An introduction to variable and feature selection. J. Mach. Learn.
Res., 3:1157?1182, March 2003.
[11] A. Jakulin and I. Bratko. Machine learning based on attribute interactions: Ph.D. dissertation. 2005.
[12] Ron Kohavi and George H. John. Wrappers for feature subset selection. Artif. Intell., 97(1-2):273?324,
December 1997.
[13] Ludmila I. Kuncheva. A stability index for feature selection. In Arti?cial Intelligence and Applications,
pages 421?427, 2007.
[14] David D. Lewis. Feature selection and feature extraction for text categorization. In In Proceedings of
Speech and Natural Language Workshop, pages 212?217. Morgan Kaufmann, 1992.
[15] Dahua Lin and Xiaoou Tang. Conditional infomax learning: An integrated framework for feature extraction and fusion. In ECCV (1), pages 68?82, 2006.
[16] P. E. Meyer and G. Bontempi. On the use of variable complementarity for feature selection in cancer
classi?cation. In Proceedings of EvoWorkshop, pages 91?102. Springer-Verlag, 2006.
[17] Mike Mintz, Steven Bills, Rion Snow, and Daniel Jurafsky. Distant supervision for relation extraction
without labeled data. In ACL/IJCNLP, pages 1003?1011, 2009.
[18] Hung Q. Ngo, Ely Porat, and Atri Rudra. Ef?ciently decodable compressed sensing by list-recoverable
codes and recursion. In Proceedings of STACS, volume 14, pages 230?241, 2012.
[19] X. Nguyen, M. J. Wainwright, and M. I. Jordan. On surrogate losses and f -divergences. Annals of
Statistics, 37(2):876?904, 2009.
[20] X. Nguyen, M. J. Wainwright, and M. I. Jordan. Estimating divergence functionals and the likelihood
ratio by by convex risk minimization. IEEE Trans. on Information Theory, 56(11):5847?5861, 2010.
[21] H. Peng, F. Long, and C. Ding. Feature selection based on mutual information: criteria of maxdependency, max-relevance, and min-redundancy. IEEE Transactions on PAMI, 27:1226?1238, 2005.
[22] Herv?e Stoppiglia, G?erard Dreyfus, R?emi Dubois, and Yacine Oussar. Ranking a random feature for
variable and feature selection. J. Mach. Learn. Res., 3:1399?1414, March 2003.
[23] Y. Sun, S. Todorovic, and S. Goodison. Local-learning-based feature selection for high-dimensional data
analysis. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 32(9):1610?1626, Sept 2010.
[24] Mingkui Tan, Li Wang, and Ivor W. Tsang. Learning sparse svm for feature selection on very high
dimensional datasets. In ICML, pages 1047?1054, 2010.
[25] Howard Hua Yang and John E. Moody. Data visualization and feature selection: New algorithms for
nongaussian data. In NIPS, pages 687?702, 1999.
[26] Lei Yu and Huan Liu. Ef?cient feature selection via analysis of relevance and redundancy. Journal of
Machine Learning Research, 5:1205?1224, 2004.
[27] Ce Zhang, Feng Niu, Christopher R?e, and Jude W. Shavlik. Big data versus the crowd: Looking for
relationships in all the right places. In ACL (1), pages 825?834, 2012.
9
| 5296 |@word trial:1 madelon:4 middle:2 proportion:1 bekkerman:1 nd:3 open:2 relevancy:1 covariance:3 elisseeff:1 arti:1 thereby:1 wrapper:17 liu:1 series:3 score:30 selecting:1 daniel:1 ours:5 existing:6 si:2 must:2 john:2 subsequent:1 distant:1 designed:2 v:1 intelligence:2 selected:8 devising:1 item:4 accordingly:1 cult:2 xk:3 core:4 dissertation:1 gure:3 provides:2 math:1 cse:3 location:1 ron:2 zhang:2 mathematical:1 along:1 consists:1 prove:1 overhead:1 hellinger:2 introduce:1 peng:2 inter:1 expected:3 behavior:2 p1:1 roughly:1 multi:2 inspired:3 ming:1 actual:3 cpu:3 accross:1 increasing:1 provided:1 estimating:2 moreover:5 underlying:3 gisette:4 medium:4 what:1 xed:1 substantially:1 emerging:1 nj:1 guarantee:2 cial:1 every:5 exactly:6 scaled:1 uk:1 unit:1 normally:1 positive:7 negligible:1 local:2 consequence:1 despite:1 mach:3 jakulin:1 niu:1 pami:1 black:1 might:1 logo:1 acl:2 specifying:1 challenging:1 suggests:1 co:1 dif:2 cacy:1 jurafsky:1 range:3 statistically:3 testing:15 rion:1 practice:3 maximisation:1 procedure:1 area:2 empirical:1 thought:1 pre:6 word:3 induce:1 get:4 selection:27 scheduling:1 risk:2 applying:2 www:1 measurable:1 equivalent:1 demonstrated:1 bill:1 go:1 convex:3 qc:1 chrismre:1 assigns:1 recovery:1 identifying:1 insight:1 estimator:1 array:2 regarded:1 dominate:1 classic:1 stability:5 notion:1 variation:1 hung:2 annals:1 target:1 play:1 tan:1 exact:1 us:2 complementarity:1 element:2 trend:1 expensive:1 distributional:1 labeled:2 stacs:1 bottom:1 role:1 observed:2 mike:1 ding:3 steven:1 capture:1 wang:1 thousand:3 calculate:1 tsang:1 sun:1 keyword:2 decrease:1 devdatt:1 ran:2 substantial:2 balanced:1 alessandro:1 broken:1 yingbo:1 smart:1 ali:1 easily:2 joint:4 po:1 indirect:1 xiaoou:1 various:3 grown:1 separated:1 fast:3 outcome:3 crowd:1 quite:1 heuristic:4 stanford:2 plausible:1 supplementary:6 valued:2 loglikelihood:1 otherwise:1 larger:2 compressed:1 ability:2 statistic:1 think:1 noisy:2 obviously:1 advantage:1 propose:3 interaction:2 product:1 relevant:10 uci:4 cients:1 iff:1 achieve:1 supposed:1 f10:1 intuitive:1 validate:2 csisz:1 scalability:7 exploiting:1 rst:5 double:1 yaniv:1 cluster:1 francois:1 produce:2 categorization:3 adam:1 generating:2 staying:1 converges:1 help:1 depending:1 develop:1 illustrate:1 stat:1 ac:1 ludmila:1 nearest:1 keywords:1 kuncheva:1 job:1 soc:1 strong:1 c:5 signi:1 implies:3 waveform:1 snow:1 attribute:1 centered:1 govind:1 material:6 bin:2 explains:1 require:1 mingkui:1 investigation:1 proposition:5 mathematically:1 ijcnlp:1 hold:5 around:1 considered:2 ic:1 gavin:1 presumably:1 predict:1 krvskp:1 reserve:1 unsatis:1 achieves:3 vary:2 smallest:1 applicable:1 combinatorial:2 label:5 pfs:14 individually:1 largest:4 minimization:1 gaussian:1 super:1 zhou:1 varying:1 corollary:2 encode:1 focus:1 lon:1 improvement:1 rank:5 likelihood:3 indicates:4 rigorous:1 baseline:1 sense:2 colon:1 dependent:1 el:1 nn:1 eliminate:2 entire:2 typically:1 integrated:1 hidden:3 relation:7 overall:1 uences:1 retaining:1 art:1 special:2 mutual:9 equal:2 construct:1 genuine:2 extraction:7 f3:1 sampling:1 never:1 biology:2 yu:1 icml:1 leukemia:1 minf:1 throughput:1 amaldi:1 np:1 report:6 micro:2 few:2 randomly:2 winter:1 decodable:1 ve:3 comprehensive:2 mintz:2 individual:2 divergence:12 familiar:1 national:2 intell:1 huge:1 satis:2 possibility:1 highly:1 evaluation:2 utkarsh:1 scienti:1 bontempi:1 silvey:1 integral:1 edge:1 rudra:1 necessary:1 huan:1 modest:1 euclidean:1 desired:1 re:3 theoretical:1 instance:1 column:3 increased:1 infected:1 ar:1 maximization:3 cost:1 subset:12 entry:1 hundred:2 wonder:1 tishby:1 kullbackleibler:1 kn:1 eec:1 synthetic:2 combined:1 st:1 fundamental:1 randomized:2 river:1 cantly:1 participates:1 off:2 pool:6 infomax:2 together:1 regressor:1 moody:1 nongaussian:1 na:2 choose:1 hoeffding:1 resort:1 zhao:1 return:1 li:1 distribute:1 potential:1 de:6 wk:1 inc:1 notable:1 ely:1 ranking:2 piece:1 performed:3 later:1 picked:1 try:1 analyze:1 portion:3 competitive:3 bayes:6 czhang:1 parallel:12 recover:1 start:1 lung:1 cife:1 contribution:3 ass:1 accuracy:16 kaufmann:1 who:1 yield:1 identify:6 accurately:2 produced:2 iid:1 cation:24 explain:1 parallelizable:2 coef:1 ed:6 evaluates:1 competitor:1 nonetheless:1 against:1 proof:3 mi:11 recovers:1 soton:1 attributed:1 dataset:2 govindaraju:1 popular:1 studia:1 recall:5 framework1:1 improves:5 dimensionality:1 sophisticated:1 yacine:1 higher:5 adaboost:1 kann:1 improved:1 formulation:2 done:2 box:1 evaluated:1 furthermore:1 just:1 correlation:1 hand:2 christopher:2 lack:3 logistic:3 quality:4 aj:3 hwang:1 grows:1 artif:1 lei:1 contain:1 true:2 brown:3 normalized:2 leibler:3 nonzero:1 illustrated:1 conditionally:1 inferior:1 essence:1 noted:1 naftali:1 criterion:3 syntax:1 complete:1 demonstrate:2 theoretic:2 variational:1 dreyfus:1 ef:5 recently:1 superior:2 common:2 multinomial:1 rl:1 empirically:1 volume:2 million:3 dahua:1 isabelle:2 cambridge:2 tac:4 rd:2 fk:1 mathematics:2 similarly:1 grid:2 consistency:1 language:2 dot:1 cuny:1 supervision:1 add:1 irrelevant:4 belongs:1 inf:1 apart:1 certain:1 verlag:1 inequality:1 binary:7 scoring:38 nition:1 preserving:1 minimum:3 additional:1 george:1 morgan:1 speci:2 employed:1 determine:1 paradigm:2 redundant:2 dashed:1 recoverable:1 multiple:2 sound:3 reduces:1 ing:1 technical:1 xf:8 distanced:1 long:3 lin:1 impact:1 scalable:1 basic:1 regression:1 luj:1 expectation:1 iteration:1 sometimes:1 jude:1 addition:2 want:3 median:1 microarray:1 crucial:2 kohavi:1 archive:1 pass:1 pooling:2 jmi:2 cmim:1 december:1 jordan:2 ngo:2 practitioner:1 ciently:3 yang:1 synthetically:2 wif:1 xj:2 affect:1 w3:1 competing:2 p0t:2 idea:2 knowing:2 nishes:1 panconesi:1 whether:1 thread:4 expression:1 herv:1 edoardo:1 f:17 speech:1 proceed:1 todorovic:1 jie:1 dramatically:1 useful:1 generally:1 xuanlong:2 clear:2 fleuret:1 amount:2 ten:1 ph:1 processed:1 generate:2 http:3 outperform:1 exist:2 andr:1 dotted:1 estimated:2 correctly:1 broadly:1 group:12 key:2 putting:1 redundancy:5 four:1 blood:3 suny:3 wisc:1 ce:2 nal:4 lter:12 asymptotically:1 wife:1 run:8 kbp:5 place:1 guyon:2 comparable:1 nba:4 spend:1 bound:1 guaranteed:2 correspondence:1 strength:1 handful:1 speed:4 emi:1 extremely:1 approximability:1 min:1 separable:32 ned:3 speedup:3 department:6 according:2 combination:3 march:3 across:1 separability:2 pocock:1 intuitively:1 gradually:2 taken:1 computationally:2 resource:1 visualization:1 mechanism:1 needed:1 initiate:1 serf:1 umich:1 end:1 naturalness:1 operation:2 available:2 gaussians:2 apply:1 observe:2 generic:1 nance:1 appearing:1 subtracted:1 robustness:5 existence:2 original:6 denotes:5 running:9 top:9 nlp:1 publishing:1 hinge:1 madison:1 unifying:1 build:1 dubhashi:1 feng:1 already:1 quantity:2 maxdependency:1 mim:8 primary:1 strategy:2 concentration:1 diagonal:3 surrogate:5 said:1 exhibit:3 distance:4 venu:1 separate:1 separating:1 fgm:9 sci:1 argue:1 trivial:1 reason:1 mikel:1 analyst:1 assuming:1 length:1 code:1 index:2 relationship:1 illustration:1 providing:3 ratio:2 minimizing:1 hungar:1 setup:1 executed:1 frank:1 negative:3 design:9 implementation:1 unknown:2 perform:1 observation:2 datasets:15 howard:1 buffalo:6 looking:1 precise:1 arbitrary:1 david:1 pair:3 required:1 kl:3 nipsfsc:1 optimized:1 lymph:1 identi:10 established:1 hour:6 nip:5 trans:1 able:5 beyond:1 suggested:1 parallelism:7 below:1 pattern:1 regime:1 challenge:1 built:1 including:2 royal:1 max:1 wainwright:2 critical:1 natural:6 ranked:3 indicator:2 bratko:1 recursion:1 zhu:1 representing:1 scheme:13 dubois:1 numerous:1 ne:2 ltered:2 perf:4 naive:4 extract:1 sept:1 text:4 relative:1 wisconsin:1 embedded:2 loss:7 expect:1 suf:9 versus:1 remarkable:1 validation:1 degree:5 lters:1 consistent:2 row:3 eccv:1 course:1 cancer:1 jth:1 aij:3 understand:1 shavlik:1 neighbor:1 sparse:1 curve:1 dimension:11 xn:1 world:5 ignores:1 collection:8 commonly:1 adaptive:1 preprocessing:3 nguyen:3 ec:1 transaction:2 functionals:1 emphasize:1 preferred:1 kullback:3 gene:2 keep:2 ml:1 sequentially:2 instantiation:1 symmetrical:1 assumed:1 search:1 porat:1 why:1 reality:2 table:2 learn:3 pack:1 pretty:1 inherently:1 du:1 domain:1 universe:4 linearly:2 big:1 noise:1 edition:1 categorized:1 x1:1 cient:10 elaborate:1 precision:3 sub:1 meyer:1 jmlr:1 third:1 capping:1 splice:1 tang:1 theorem:7 down:1 xt:3 showing:1 er:8 sensing:1 list:1 svm:5 dominates:1 fusion:1 exists:1 workshop:1 sequential:4 adding:4 execution:4 entropy:1 michigan:1 logarithmic:1 p1t:1 ivor:1 erard:1 atri:1 hua:1 springer:1 corresponds:2 lewis:1 prop:1 w10:1 conditional:6 goal:1 viewed:1 identity:1 porwal:1 mrmr:1 nish:4 hard:1 change:2 justify:1 acting:1 classi:26 called:4 total:5 e:3 experimental:1 exception:1 select:7 support:2 relevance:5 violated:2 bioinformatics:1 evaluate:2 mum:1 tested:3 correlated:1 |
4,745 | 5,297 | Spectral k-Support Norm Regularization
Andrew M. McDonald, Massimiliano Pontil, Dimitris Stamos
Department of Computer Science
University College London
{a.mcdonald,m.pontil,d.stamos}@cs.ucl.ac.uk
Abstract
The k-support norm has successfully been applied to sparse vector prediction
problems. We observe that it belongs to a wider class of norms, which we call the
box-norms. Within this framework we derive an efficient algorithm to compute
the proximity operator of the squared norm, improving upon the original method
for the k-support norm. We extend the norms from the vector to the matrix setting
and we introduce the spectral k-support norm. We study its properties and show
that it is closely related to the multitask learning cluster norm. We apply the norms
to real and synthetic matrix completion datasets. Our findings indicate that spectral k-support norm regularization gives state of the art performance, consistently
improving over trace norm regularization and the matrix elastic net.
1
Introduction
In recent years there has been a great deal of interest in the problem of learning a low rank matrix
from a set of linear measurements. A widely studied and successful instance of this problem arises
in the context of matrix completion or collaborative filtering, in which we want to recover a low
rank (or approximately low rank) matrix from a small sample of its entries, see e.g. [1, 2]. One
prominent method to solve this problem is trace norm regularization: we look for a matrix which
closely fits the observed entries and has a small trace norm (sum of singular values) [3, 4, 5]. Besides
collaborative filtering, this problem has important applications ranging from multitask learning, to
computer vision and natural language processing, to mention but a few.
In this paper, we propose new techniques to learn low rank matrices. These are inspired by the notion
of the k-support norm [6], which was recently studied in the context of sparse vector prediction and
shown to empirically outperform the Lasso [7] and Elastic Net [8] penalties. We note that this
norm can naturally be extended to the matrix setting and its characteristic properties relating to the
cardinality operator translate in a natural manner to matrices. Our approach is suggested by the
observation that the k-support norm belongs to a broader class of norms, which makes it apparent
that they can be extended to spectral matrix norms. Moreover, it provides a link between the spectral
k-support norm and the cluster norm, a regularizer introduced in the context of multitask learning
[9]. This result allows us to interpret the spectral k-support norm as a special case of the cluster
norm and furthermore adds a new perspective of the cluster norm as a perturbation of the former.
The main contributions of this paper are threefold. First, we show that the k-support norm can
be written as a parametrized infimum of quadratics, which we term the box-norms, and which are
symmetric gauge functions. This allows us to extend the norms to orthogonally invariant matrix
norms using a classical result by von Neumann [10]. Second, we show that the spectral box-norm
is essentially equivalent to the cluster norm, which in turn can be interpreted as a perturbation of
the spectral k-support norm, in the sense of the Moreau envelope [11]. Third, we use the infimum
framework to compute the box-norm and the proximity operator of the squared norm in O(d log d)
time. Apart from improving on the O(d(k + log d)) algorithm in [6], this method allows one to use
optimal first order optimization algorithms [12] with the cluster norm. Finally, we present numerical
1
experiments which indicate that the spectral k-support norm shows a significant improvement in
performance over regularization with the trace norm and the matrix elastic net, on four popular
matrix completion benchmarks.
The paper is organized as follows. In Section 2 we recall the k-support norm, and define the boxnorm. In Section 3 we study their properties, we introduce the corresponding spectral norms, and
we observe the connection to the cluster norm. In Section 4 we compute the norm and we derive
a fast method to compute the proximity operator. Finally, in Section 5 we report on our numerical
experiments. The supplementary material contains derivations of the results in the body of the paper.
2
Preliminaries
In this section, we recall the k-support norm and we introduce the box-norm and its dual. The ksupport norm k ? k(k) was introduced in [6] as the norm whose unit ball is the convex hull of the
set of vectors of cardinality at most k and `2 -norm no greater than one. The authors show that the
k-support norm can be written as the infimal convolution [11]
8
9
<X
=
X
kwk(k) = inf
kvg k2 : vg 2 Rd , supp(vg ) ? g,
v g = w , w 2 Rd ,
(1)
:
;
g2Gk
g2Gk
where Gk is the collection of all subsets of {1, . . . , d} containing at most k elements, and for any
v 2 Rd , the set supp(v) = {i : vi 6= 0} denotes the support of v. When used as a regularizer,
the norm encourages vectors w to be a sum of a limited number of vectors with small support. The
k-support norm is a special case of the group lasso with overlap [13], where the cardinality of the
support sets is at most k. Despite the complicated form of the primal norm, the dual norm has a
simple formulation, namely the `2 -norm of the k largest components of the vector
v
u k
uX
kuk?,(k) = t (|u|#i )2 , u 2 Rd ,
(2)
i=1
where |u|# is the vector obtained from u by reordering its components so that they are non-increasing
in absolute value [6]. The k-support norm includes the `1 -norm and `2 -norm as special cases. This
is clear from the dual norm since for k = 1 and k = d, it is equal to the `1 -norm and `2 -norm,
respectively. We note that while definition (1) involves a combinatorial number of variables, [6]
observed that the norm can be computed in O(d log d).
We now define the box-norm, and in the following section we will show that the k-support norm is
a special case of this family.
Pd
Definition 2.1. Let 0 ? a ? b and c 2 [ad, bd] and let ? = {? 2 Rd : a ? ?i ? b, i=1 ?i ? c}.
The box-norm is defined as
v
u
d
u
X
wi2
kwk? = t inf
, w 2 Rd .
(3)
?2? i=1 ?i
This formulation will be fundamental in deriving the proximity operator in Section 4.1. Note that
we may assume without loss of generality that b = 1, as by rescaling we obtain an equivalent norm,
however we do not explicitly fix b in the sequel.
s
d
P
Proposition 2.2. The norm (3) is well defined and the dual norm is kuk?,? = sup
?i u2i .
?2? i=1
The result holds true in the more general case that ? is a bounded convex subset of the strictly
positive orthant (for related results see [14, 15, 16, 17, 18, 19] and references therein). In this
paper we limit ourselves to the box constraints above. In particular we note that the constraints are
invariant with respect to permutation of the components of ?, and as we shall see this property is
key to extend the norm to matrices.
2
3
Properties of the Norms
In this section, we study the properties of the vector norms, and we extend the norms to the matrix
setting. We begin by deriving the dual box-norm.
Proposition 3.1. The dual box-norm is given by
q
kuk?,? = akuk22 + (b a)kuk2?,(k) + (b a)(? k)(|u|#k+1 )2 ,
(4)
where ? =
c da
b a
and k is the largest integer not exceeding ?.
We see from (4) that the dual norm decomposes into two `2 -norms plus a residual term, which
vanishes if ? = k, and for the rest of this paper we assume this holds, which loses little generality.
Note that setting a = 0, b = 1, and c = k 2 {1, . . . , d}, the dual box-norm (4) is the `2 -norm of the
largest k components of u, and we recover the dual k-support norm in equation (2). It follows that
the k-support norm is a box-norm with parameters a = 0, b = 1, c = k.
The following infimal convolution interpretation of the box-norm provides a link between the boxnorm and the k-support norm, and illustrates the effect of the parameters.
Proposition 3.2. If 0 < a ? b and c = (b a)k + da, for k 2 {1, . . . , d}, then
8
9
v
<X u
=
2
2
X
X
X
vg,i
vg,i
u
t
kwk? = inf
+
: v g 2 Rd ,
vg = w .
(5)
:
;
b
a
g2Gk
i2g
i2g
/
g2Gk
Notice that if b = 1, then as a tends to zero, we obtain the expression of the k-support norm (1),
recovering in particular the support constraints. If a is small and positive, the support constraints
are not imposed, however effectively most of the weight for each vg tends to be concentrated on
supp(g). Hence, Proposition 3.2 suggests that the box-norm regularizer will encourage vectors w
whose dominant components are a subset of a union of a small number of groups g 2 Gk .
The previous results have characterized the k-support norm as a special case of the box-norm. Conversely, the box-norm can be seen as a perturbation of the k-support norm with a quadratic term.
Proposition 3.3. Let k?k? be the box-norm on Rd with parameters 0 < a < b and c = k(b a)+da,
for k 2 {1, . . . , d}, then
?
1
1
2
kwk? = min
kw zk22 +
kzk2(k) .
(6)
d
a
b a
z2R
Consider the regularization problem minw2Rd kXw yk22 + kwk2? , with data X and response y.
Using Proposition 3.3 and setting w = u + z, we see that this problem is equivalent to
?
min kX(u + z) yk22 + kuk22 +
kzk2(k) .
a
b a
u,z2Rd
Furthermore, if (?
u, z?) solves this problem then w
?=u
? + z? solves problem (6). The solution w
? can
therefore be interpreted as the superposition of a vector which has small `2 norm, and a vector which
has small k-support norm, with the parameter a regulating these two components. Specifically, as
a tends to zero, in order to prevent the objective from blowing up, u
? must also tend to zero and we
recover k-support norm regularization. Similarly, as a tends to b, z? vanishes and we have a simple
ridge regression problem.
3.1
The Spectral k-Support Norm and the Spectral Box-Norm
We now turn our focus to the matrix norms. For this purpose, we recall that a norm k ? k on Rd?m is
called orthogonally invariant if kW k = kU W V k, for any orthogonal matrices U 2 Rd?d and
V 2 Rm?m . A classical result by von Neumann [10] establishes that a norm is orthogonally
invariant if and only if it is of the form kW k = g( (W )), where (W ) is the vector formed by
the singular values of W in nonincreasing order, and g is a symmetric gauge function, that is a norm
which is invariant under permutations and sign changes of the vector components.
3
Lemma 3.4. If ? is a convex bounded subset of the strictly positive orthant in Rd which is invariant
under permutations, then k ? k? is a symmetric gauge function.
In particular, this readily applies to both the k-support norm and box-norm. We can therefore extend
both norms to orthogonally invariant norms, which we term the spectral k-support norm and the
spectral box-norm respectively, and which we write (with some abuse of notation) as kW k(k) =
k (W )k(k) and kW k? = k (W )k? . We note that since the k-support norm subsumes the `1 and
`2 -norms for k = 1 and k = d respectively, the corresponding spectral k-support norms are equal
to the trace and Frobenius norms respectively. We first characterize the unit ball of the spectral
k-support norm.
Proposition 3.5. The unit ball of the spectral k-support norm is the convex hull of the set of matrices
of rank at most k and Frobenius norm no greater than one.
Referring to the unit ball characterization of the k-support norm, we note that the restriction on the
cardinality of the vectors whose convex hull defines the unit ball naturally extends to a restriction
on the rank operator in the matrix setting. Furthermore, as noted in [6], regularization using the
k-support norm encourages vectors to be sparse, but less so that the `1 -norm. In matrix problems, as
the extreme points of the unit ball have rank k, Proposition 3.5 suggests that the spectral k-support
norm for k > 1 should encourage matrices to have low rank, but less so than the trace norm.
3.2
Cluster Norm
We end this section by briefly discussing the cluster norm, which was introduced in [9] as a convex
relaxation of a multitask clustering problem. The norm is defined, for every W 2 Rd?m , as
r
kW kcl =
inf tr(S 1 W > W )
(7)
S2Sm
where Sm = {S 2 Rm?m , S ? 0 : aI
S
bI, tr S = c}, and 0 < a ? b. In [9] the authors
state that the cluster norm of W equals the box-norm of the vector formed by the singular values of
W where c = (b a)k + da. Here we provide a proof of this result. Denote by i (?) the eigenvalues
of a matrix which we write in nonincreasing order 1 (?)
???
2 (?)
d (?). Note that if ?i are
the eigenvalues of S then ?i = d i+1 (S 1 ). We have that
tr(S
1
W > W ) = tr(S
1
U ?2 U > )
m
X
1
d i+1 (S
) i (W > W ) =
i=1
d
X
2
i (W )
i=1
?i
where we have used the inequality [20, Sec. H.1.h] for S 1 , W > W ? 0. Since this inequality is
attained whenever S = U Diag(?)U , where U are the eigenvectors of W > W , we see that kW kcl =
k (W )k? , that is, the cluster norm coincides with the spectral box-norm. In particular, we see that
the spectral k-support norm is a special case of the cluster norm, where we let a tend to zero, b = 1
and c = k. Moreover, the methods to compute the norm and its proximity operator described in the
following section can directly be applied to the cluster norm.
As in the case of the vector norm (Proposition 3.3), the spectral box-norm or cluster norm can be
written as a perturbation of spectral k-support norm with a quadratic term.
Proposition 3.6. Let k ? k? be a matrix box-norm with parameters a, b, c and let k =
1
kW k2? = min kW
Z a
Zk2F +
1
b
a
c da
b a .
Then
kZk2(k) .
In other words, this result shows that the cluster norm can be seen as the Moreau envelope [11] of a
spectral k-support norm.
4
Computing the Norms and their Proximity Operator
In this section, we compute the norm and the proximity operator of the squared norm by explicitly
solving the optimization problem in (3). We begin with the vector norm.
4
Theorem 4.1. For every w 2 Rd it holds that
kwk2? =
1
1
1
kwQ k22 + kwI k21 + kwL k22 ,
b
p
a
where wQ = (|w|#1 , . . . , |w|#q ), wI = (|w|#q+1 , . . . , |w|#d ` ), wL = (|w|#d
` are the unique integers in {0, . . . , d} that satisfy q + ` ? d,
|wq |
b
p=c
qb
d `
1 X
|wq+1 |
|wi | >
,
p i=q+1
b
|wd ` |
a
(8)
#
`+1 , . . . , |w|d ),
d `
1 X
|wd `+1 |
|wi | >
,
p i=q+1
a
and q and
(9)
`a and we have defined |w0 | = 1 and |wd+1 | = 0.
Proof. (Sketch) We need to solve the optimization problem
inf
?
?X
d
i=1
d
X
wi2
: a ? ?i ? b,
?i ? c .
?i
i=1
(10)
We assume without loss of generality that the wi are ordered nonincreasing in absolute values, and
it follows that at the optimum the ?i are also ordered nonincreasing. We further assume that wi 6= 0
for all i and c ? db, so the sum constraint will be tight at the optimum. The Lagrangian is given by
!
d
d
X
wi2
1 X
L(?, ?) =
+ 2
?i c
?
?
i=1 i
i=1
Pd
where 1/?2 is a strictly positive multiplier to be chosen such that S(?) := i=1 ?i (?) = c. We
can then solve the original problem by minimizing the Lagrangian over the constraint ? 2 [a, b]d .
Due to the decoupling effect of the multiplier we can solve the simplified problem componentwise,
obtaining the solution
?i = ?i (?) = min(b, max(a, ?|wi |))
(11)
where S(?) = c. The minimizer has the form ? = (b, . . . , b, ?q+1 , . . . , ?d ` , a, . . . , a), where q, `
Pd `
are determined by the value of ?. From S(?) = c we get ? = p/( i=q+1 |wi |). The value
of the norm in (8) follows by substituting ? into the objective. Finally, by construction we have
?q b > ?q+1 and ?d ` > a ?d `+1 , which give rise to the conditions in (9).
Theorem 4.1 suggests two methods for computing the box-norm. First we find ? such that S(?) = c;
this value uniquely determines ? in (11), and the norm follows by substitution into (10). Alternatively, we identify q and ` that jointly satisfy (9) and we compute the norm using (8). Taking
advantage of the structure of ? in the former method leads to a computation time that is O(d log d).
Theorem 4.2. The computation of the box-norm can be completed in O(d log d) time.
The k-support norm is a special case of the box-norm, and as a direct corollary of Theorem 4.1 and
Theorem 4.2, we recover [6, Proposition 2.1].
4.1
Proximity Operator
Proximal gradient methods can be used to solve optimization problems of the form minw f (w) +
g(w), where f is a convex loss function with Lipschitz continuous gradient, > 0 is a regularization parameter, and g is a convex function for which the proximity operator can be computed
efficiently, see [12, 21, 22] and references therein. The proximity operator of g with parameter ? > 0
is defined as
?
1
prox?g (w) = argmin
kx wk2 + ?g(x) : x 2 Rd .
2
We now use the infimum formulation of the box-norm to derive the proximity operator of the squared
norm.
5
Algorithm 1 Computation of x = prox
2
k?k2?
Require: parameters a, b, c, .
n
od
2d
b+
1. Sort points ?i i=1 = a+
|wj | , |wj |
2.
3.
4.
5.
j=1
(w).
such that ?i ? ?i+1 ;
Identify points ?i and ?i+1 such that S(?i ) ? c and S(?i+1 ) c by binary search;
Find ?? between ?i and ?i+1 such that S(?? ) = c by linear interpolation;
Compute ?i (?? ) for i = 1, . . . , d;
wi
Return xi = ??ii+
for i = 1, . . . , d.
Theorem 4.3. The proximity operator of the square of the box-norm at point w 2 Rd with parameter
?d w d
?1 w 1
2 is given by prox k?k2 (w) = ( ?1 + , . . . , ?d + ), where
?
2
?i = ?i (?) = min(b, max(a, ?|wi |
))
(12)
Pd
and ? is chosen such that S(?) := i=1 ?i (?) = c. Furthermore, the computation of the proximity
operator can be completed in O(d log d) time.
The proof follows a similar reasoning to the proof of Theorem 4.1. Algorithm 1 illustrates the
computation of the proximity operator for the squared box-norm in O(d log d) time. This includes
the k-support as a special case, where we let a tend to zero, and set b = 1 and c = k, which
improves upon the complexity of the O(d(k + log d)) computation provided in [6], and we illustrate
the improvement empirically in Table 1.
4.2
Proximity Operator for Orthogonally Invariant Norms
The computational considerations outlined above can be naturally extended to the matrix setting by
using von Neumann?s trace inequality (see, e.g. [23]). Here we comment on the computation of the
proximity operator, which is important for our numerical experiments in the following section. The
proximity operator of an orthogonally invariant norm k ? k = g( (?)) is given by
proxk?k (W ) = U diag(proxg ( (W )))V > , W 2 Rm?d ,
where U and V are the matrices formed by the left and right singular vectors of W (see e.g. [24,
Prop 3.1]). Using this result we can employ proximal gradient methods to solve matrix regularization
problems using the squared spectral k-support norm and spectral box-norm.
5
Numerical Experiments
In this section, we report on the statistical performance of the spectral regularizers in matrix completion experiments. We also offer an interpretation of the role of the parameters in the box-norm
and we empirically verify the improved performance of the proximity operator computation (see
Table 1). We compare the trace norm (tr) [25], matrix elastic net (en) [26], spectral k-support (ks)
and the spectral box-norm (box). The Frobenius norm, which is equal to the spectral k-support
norm for k = d, performed considerably worse than the trace norm and we omit the results here.
We report test error and standard deviation, matrix rank (r) and optimal parameter values for k and
a, which were determined by validation, as were the regularization parameters. When comparing
performance, we used a t-test to determine statistical significance at a level of p < 0.001. For the
optimization we used an accelerated proximal gradient method (FISTA), see e.g. [12, 21, 22], with
the percentage change in objective as convergence criterion, with a tolerance of 10 5 for the simulated datasets and 10 3 for the real datasets. As is typical with spectral regularizers we found that
the spectrum of the learned matrix exhibited a rapid decay to zero. In order to explicitly impose a
low rank on the solution we included a final step where we hard-threshold the singular values of the
final matrix below a level determined by validation. We report on both sets of results below.
5.1
Simulated Data
Matrix Completion. We applied the norms to matrix completion on noisy observations of low rank
matrices. Each m ? m matrix was generated as W = AB > + E, where A, B 2 Rm?r , r ? m, and
6
Table 1: Comparison of proximity operator algorithms for the k-support norm (time in s), k = 0.05d.
Algorithm 1 is the method in [6], Algorithm 2 is our method.
d
Alg. 1
Alg. 2
1,000
2,000
4,000
8,000
16,000
32,000
0.0443
0.0011
0.1567
0.0016
0.5907
0.0026
2.3065
0.0046
9.0080
0.0101
35.6199
0.0181
0.03
5
k value
a value
4
0.02
0.01
3
2
0
2
4
6
SNR
8
1
10
Figure 1: Impact of signal to noise on a.
2
4
6
8
true rank
10
Figure 2: Impact of matrix rank on k.
the entries of A, B and E are i.i.d. standard Gaussian. We set m = 100, r 2 {5, 10} and we sampled
uniformly a percentage ? 2 {10%, 20%, 30%} of the entries for training, and used a fixed 10% for
validation. The error was measured as ktrue predictedk2 /ktruek2 [5] and averaged over 100 trials.
The results are summarized in Table 2. In the thresholding case, all methods recovered the rank of
the true noiseless matrix. The spectral box-norm generated the lowest test errors in all regimes, with
the spectral k-support a close second, in particular in the thresholding case. This suggests that the
non zero parameter a in the spectral box-norm counteracted the noise to some extent.
Role of Parameters. In the same setting we investigated the role of the parameters in the boxnorm. As previously discussed, parameter b can be set to 1 without loss of generality. Figure 1
shows the optimal value of a chosen by validation for varying signal to noise ratios (SNR), keeping
k fixed. We see that for greater noise levels (smaller SNR), the optimal value for a increases. While
for a > 0, the recovered solutions are not sparse, as we show below this can still lead to improved
performance in experiments, in particular in the presence of noise. Figure 2 shows the optimal value
of k chosen by validation for matrices with increasing rank, keeping a fixed. We notice that as the
rank of the matrix increases, the optimal k value increases, which is expected since it is an upper
bound on the sum of the singular values.
Table 2: Matrix completion on simulated data sets, without (left) and with (right) thresholding.
dataset
norm
test error
r
k
rank 5 tr
?=10% en
ks
box
0.8184 (0.03)
0.8164 (0.03)
0.8036 (0.03)
0.7805 (0.03)
rank 5 tr
?=20% en
ks
box
0.4085 (0.03) 23
0.4081 (0.03) 23
0.4031 (0.03) 21 3.1
0.3898 (0.03) 100 1.3
rank 10 tr
?=20% en
ks
box
0.6356 (0.03)
0.6359 (0.03)
0.6284 (0.03)
0.6243 (0.03)
rank 10 tr
?=30% en
ks
box
dataset
a
20
20
16 3.6
87 2.9 1.7e-2
norm
test error
r
k
a
rank 5 tr
?=10% en
ks
box
0.7799 (0.04)
0.7794 (0.04)
0.7728 (0.04)
0.7649 (0.04)
5
5
5 4.23
5 3.63 8.1e-3
9e-3
rank 5 tr
?=20% en
ks
box
0.3449 (0.02)
0.3445 (0.02)
0.3381 (0.02)
0.3380 (0.02)
5
5
5 2.97
5 3.28 1.9e-3
27
27
24 4.4
89 1.8
9e-3
rank 10 tr
?=20% en
ks
box
0.6084 (0.03)
0.6074 (0.03)
0.6000 (0.03)
0.6000 (0.03)
10
10
10 5.02
10 5.22 1.9e-3
0.3642 (0.02) 36
0.3638 (0.002 36
0.3579 (0.02) 33 5.0
0.3486 (0.02) 100 2.5
9e-3
rank 10 tr
?=30% en
ks
box
0.3086 (0.02)
0.3082 (0.02)
0.3025 (0.02)
0.3025 (0.02)
10
10
10 5.13
10 5.16
7
3e-4
Table 3: Matrix completion on real data sets, without (left) and with (right) thresholding.
dataset
norm test error
r
k
dataset
a
norm test error
r
k
a
MovieLens tr
100k
en
? = 50%
ks
box
0.2034
0.2034
0.2031
0.2035
87
87
102 1.00
943 1.00 1e-5
MovieLens tr
100k
en
? = 50%
ks
box
0.2017 13
0.2017 13
0.1990 9 1.87
0.1989 10 2.00 1e-5
MovieLens tr
1M
en
? = 50%
ks
box
0.1821 325
0.1821 319
0.1820 317 1.00
0.1817 3576 1.09 3e-5
MovieLens tr
1M
en
? = 50%
ks
box
0.1790
0.1789
0.1782
0.1777
17
17
17 1.80
19 2.00 1e-6
Jester 1
20 per line
tr
en
ks
box
0.1787
0.1787
0.1764
0.1766
98
98
84 5.00
100 4.00 1e-6
Jester 1
20 per line
tr
en
ks
box
0.1752
0.1752
0.1739
0.1726
11
11
11 6.38
11 6.40 2e-5
Jester 3
8 per line
tr
en
ks
box
0.1988
0.1988
0.1970
0.1973
49
49
46 3.70
100 5.91 1e-3
Jester 3
8 per line
tr
en
ks
box
0.1959
0.1959
0.1942
0.1940
3
3
3 2.13
3 4.00 8e-4
5.2
Real Data
Matrix Completion (MovieLens and Jester). In this section we report on matrix completion on
real data sets. We observe a percentage of the (user, rating) entries of a matrix and the task is to predict the unobserved ratings, with the assumption that the true matrix has low rank. The datasets we
considered were MovieLens 100k and MovieLens 1M (http://grouplens.org/datasets/movielens/),
which consist of user ratings of movies, and Jester 1 and Jester 3 (http://goldberg.berkeley.edu/jesterdata/), which consist of users and ratings of jokes (Jester 2 showed essentially identical performance
to Jester 1). Following [4], for MovieLens we uniformly sampled ? = 50% of the available entries
for each user for training, and for Jester 1 and Jester 3 we sampled 20, respectively 8, ratings per
user, and we used 10% for validation. The error was measured as normalized mean absolute error,
ktrue predictedk2
#observations/(rmax rmin ) , where rmin and rmax are lower and upper bounds for the ratings [4]. The
results are outlined in Table 3. In the thresholding case, the spectral box and k-support norms had
the best performance. In the absence of thresholding, the spectral k-support showed slightly better
performance. Comparing to the synthetic data sets, this suggests that in the absence of noise the
parameter a did not provide any benefit. We note that in the absence of thresholding our results for
the trace norm on MovieLens 100k agreed with those in [3].
6
Conclusion
We showed that the k-support norm belongs to the family of box-norms and noted that these can
be naturally extended from the vector to the matrix setting. We also provided a connection between
the k-support norm and the cluster norm, which essentially coincides with the spectral box-norm.
We further observed that the cluster norm is a perturbation of the spectral k-support norm, and we
were able to compute the norm and its proximity operator. Our experiments indicate that the spectral
box-norm and k-support norm consistently outperform the trace norm and the matrix elastic net on
various matrix completion problems. With a single parameter to validate, compared to two for the
spectral box-norm, our results suggest that the spectral k-support norm is a powerful alternative to
the trace norm and the elastic net, which has the same number of parameters. In future work, we
would like to study the application of the norms to clustering problems in multitask learning [9],
in particular the impact of centering. It would also be valuable to derive statistical inequalities and
Rademacher complexities for these norms.
Acknowledgements
We would like to thank Andreas Maurer, Charles Micchelli and especially Andreas Argyriou for
useful discussions. Part of this work was supported by EPSRC Grant EP/H027203/1.
8
References
[1] N. Srebro, J. D. M. Rennie, and T. S. Jaakkola. Maximum-margin matrix factorization. Advances in
Neural Information Processing Systems 17, 2005.
[2] J. Abernethy, F. Bach, T. Evgeniou, and J.-P. Vert. A new approach to collaborative filtering: Operator
estimation with spectral regularization. Journal of Machine Learning Research, Vol. 10:803?826, 2009.
[3] M Jaggi and M. Sulovsky. A simple algorithm for nuclear norm regularized problems. Proceedings of the
27th International Conference on Machine Learning, 2010.
[4] K.-C. Toh and S. Yun. An accelerated proximal gradient algorithm for nuclear norm regularized least
squares problems. SIAM Journal on Imaging Sciences, 4:573?596, 2011.
[5] R. Mazumder, T. Hastie, and R. Tibshirani. Spectral regularization algorithms for learning large incomplete matrices. Journal of Machine Learning Research, 11:2287?2322, 2010.
[6] A. Argyriou, R. Foygel, and N. Srebro. Sparse prediction with the k-support norm. In Advances in Neural
Information Processing Systems 25, pages 1466?1474, 2012.
[7] R. Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society,
Vol. 58:267?288, 1996.
[8] H. Zou and T. Hastie. Regularization and variable selection via the elastic net. Journal of the Royal
Statistical Society, Series B, 67(2):301?320, 2005.
[9] L. Jacob, F. Bach, and J.-P. Vert. Clustered multi-task learning: a convex formulation. Advances in Neural
Information Processing Systems (NIPS 21), 2009.
[10] J. Von Neumann. Some matrix-inequalities and metrization of matric-space. Tomsk. Univ. Rev. Vol I,
1937.
[11] R. T. Rockafellar. Convex Analysis. Princeton University Press, 1970.
[12] Y. Nesterov. Gradient methods for minimizing composite objective function. Center for Operations
Research and Econometrics, 76, 2007.
[13] L. Jacob, G. Obozinski, and J.-P. Vert. Group lasso with overlap and graph lasso. Proc of the 26th Int.
Conf. on Machine Learning, 2009.
[14] Y. Grandvalet. Least absolute shrinkage is equivalent to quadratic penalization. In ICANN 98, pages
201?206. Springer London, 1998.
[15] C. A. Micchelli and M. Pontil. Learning the kernel function via regularization. Journal of Machine
Learning Research, 6:1099?1125, 2005.
[16] M. Szafranski, Y. Grandvalet, and P. Morizet-Mahoudeaux. Hierarchical penalization. In Advances in
Neural Information Processing Systems 21, 2007.
[17] C. A. Micchelli, J. M. Morales, and M. Pontil. Regularizers for structured sparsity. Advances in Comp.
Mathematics, 38:455?489, 2013.
[18] A. Maurer and M. Pontil. Structured sparsity and generalization. The Journal of Machine Learning
Research, 13:671?690, 2012.
[19] G. Obozinski and F. Bach. Convex relaxation for combinatorial penalties. CoRR, 2012.
[20] A. W. Marshall and I. Olkin. Inequalities: Theory of Majorization and its Applications. Academic Press,
1979.
[21] P. L. Combettes and J.-C. Pesquet. Proximal splitting methods in signal processing. In Fixed-Point
Algorithms for Inv Prob. Springer, 2011.
[22] A. Beck and M. Teboulle. A fast iterative shrinkage-thresholding algorithm for linear inverse problems.
SIAM J. Imaging Sciences, 2(1):183?202, 2009.
[23] A. S. Lewis. The convex analysis of unitarily invariant matrix functions. Journal of Convex Analysis,
2:173?183, 1995.
[24] A. Argyriou, C. A. Micchelli, M. Pontil, L. Shen, and Y. Xu. Efficient first order methods for linear
composite regularizers. CoRR, abs/1104.1436, 2011.
[25] J.-F. Cai, E. J. Candes, and Z. Shen. A singular value thresholding algorithm for matrix completion. SIAM
Journal on Optimization, 20(4):1956?1982, 2008.
[26] H. Li, N. Chen, and L. Li. Error analysis for matrix elastic-net regularization algorithms. IEEE Transactions on Neural Networks and Learning Systems, 23-5:737?748, 2012.
[27] W. Rudin. Functional Analysis. McGraw Hill, 1991.
[28] D. P. Bertsekas, A. Nedic, and A. E. Ozdaglar. Convex Analysis and Optimization. Athena Scientific,
2003.
[29] R. A. Horn and C. R. Johnson. Topics in Matrix Analysis. Cambridge University Press, 1991.
9
| 5297 |@word multitask:5 trial:1 briefly:1 norm:185 jacob:2 mention:1 tr:21 substitution:1 contains:1 series:1 recovered:2 wd:3 od:1 comparing:2 mahoudeaux:1 toh:1 olkin:1 written:3 readily:1 bd:1 must:1 numerical:4 rudin:1 provides:2 characterization:1 org:1 u2i:1 direct:1 manner:1 introduce:3 expected:1 rapid:1 multi:1 inspired:1 little:1 cardinality:4 increasing:2 begin:2 provided:2 moreover:2 bounded:2 notation:1 lowest:1 argmin:1 interpreted:2 rmax:2 finding:1 unobserved:1 berkeley:1 every:2 k2:4 rm:4 uk:1 unit:6 grant:1 omit:1 ozdaglar:1 bertsekas:1 positive:4 tends:4 limit:1 despite:1 interpolation:1 approximately:1 abuse:1 plus:1 therein:2 studied:2 wk2:1 k:17 suggests:5 conversely:1 limited:1 factorization:1 bi:1 averaged:1 unique:1 horn:1 union:1 pontil:6 vert:3 composite:2 word:1 suggest:1 get:1 close:1 selection:2 operator:23 context:3 restriction:2 equivalent:4 imposed:1 lagrangian:2 center:1 szafranski:1 convex:14 shen:2 splitting:1 deriving:2 nuclear:2 notion:1 construction:1 user:5 goldberg:1 element:1 sulovsky:1 econometrics:1 observed:3 role:3 epsrc:1 ep:1 wj:2 valuable:1 pd:4 vanishes:2 complexity:2 nesterov:1 solving:1 tight:1 upon:2 various:1 regularizer:3 derivation:1 univ:1 massimiliano:1 fast:2 kcl:2 london:2 abernethy:1 apparent:1 whose:3 widely:1 solve:6 supplementary:1 rennie:1 jointly:1 noisy:1 final:2 advantage:1 eigenvalue:2 metrization:1 net:8 cai:1 ucl:1 propose:1 translate:1 frobenius:3 validate:1 convergence:1 cluster:17 optimum:2 neumann:4 rademacher:1 wider:1 derive:4 andrew:1 ac:1 completion:12 illustrate:1 measured:2 solves:2 recovering:1 c:1 involves:1 indicate:3 i2g:2 closely:2 hull:3 material:1 require:1 fix:1 clustered:1 generalization:1 preliminary:1 proposition:11 strictly:3 hold:3 proximity:20 considered:1 great:1 proxg:1 predict:1 substituting:1 purpose:1 estimation:1 proc:1 combinatorial:2 superposition:1 grouplens:1 largest:3 wl:1 gauge:3 successfully:1 establishes:1 gaussian:1 ksupport:1 shrinkage:3 varying:1 broader:1 jaakkola:1 corollary:1 focus:1 improvement:2 consistently:2 rank:25 sense:1 zk22:1 stamos:2 kwq:1 dual:9 jester:11 art:1 special:8 equal:4 evgeniou:1 identical:1 kw:9 look:1 future:1 report:5 few:1 employ:1 beck:1 ourselves:1 ab:2 interest:1 regulating:1 extreme:1 primal:1 regularizers:4 nonincreasing:4 encourage:2 minw:1 orthogonal:1 incomplete:1 maurer:2 instance:1 teboulle:1 marshall:1 deviation:1 entry:6 subset:4 snr:3 successful:1 johnson:1 characterize:1 proximal:5 synthetic:2 considerably:1 referring:1 fundamental:1 international:1 siam:3 sequel:1 squared:6 von:4 containing:1 worse:1 conf:1 rescaling:1 return:1 li:2 supp:3 prox:3 kwl:1 sec:1 subsumes:1 includes:2 summarized:1 rockafellar:1 int:1 satisfy:2 explicitly:3 kzk2:3 vi:1 ad:1 performed:1 kwk:4 sup:1 recover:4 sort:1 complicated:1 candes:1 collaborative:3 contribution:1 formed:3 square:2 majorization:1 characteristic:1 efficiently:1 identify:2 comp:1 whenever:1 definition:2 centering:1 naturally:4 proof:4 sampled:3 dataset:4 popular:1 recall:3 improves:1 organized:1 agreed:1 blowing:1 attained:1 response:1 improved:2 formulation:4 box:57 generality:4 furthermore:4 sketch:1 defines:1 infimum:3 scientific:1 unitarily:1 effect:2 k22:2 normalized:1 true:4 multiplier:2 verify:1 former:2 regularization:16 hence:1 symmetric:3 deal:1 encourages:2 uniquely:1 noted:2 coincides:2 criterion:1 prominent:1 yun:1 hill:1 ridge:1 mcdonald:2 reasoning:1 ranging:1 consideration:1 ktrue:2 recently:1 charles:1 functional:1 empirically:3 extend:5 interpretation:2 discussed:1 relating:1 interpret:1 kwk2:2 measurement:1 significant:1 counteracted:1 cambridge:1 ai:1 rd:15 outlined:2 mathematics:1 similarly:1 language:1 had:1 add:1 dominant:1 jaggi:1 recent:1 showed:3 perspective:1 inf:5 belongs:3 apart:1 inequality:6 binary:1 discussing:1 seen:2 greater:3 impose:1 determine:1 signal:3 ii:1 characterized:1 academic:1 offer:1 bach:3 impact:3 prediction:3 regression:2 vision:1 essentially:3 noiseless:1 kernel:1 want:1 singular:7 kvg:1 envelope:2 rest:1 exhibited:1 kwi:1 comment:1 tend:3 db:1 call:1 integer:2 presence:1 yk22:2 fit:1 hastie:2 lasso:5 pesquet:1 andreas:2 expression:1 penalty:2 useful:1 clear:1 eigenvectors:1 concentrated:1 http:2 outperform:2 percentage:3 notice:2 sign:1 per:5 tibshirani:2 write:2 threefold:1 shall:1 vol:3 group:3 key:1 four:1 threshold:1 prevent:1 kuk:3 imaging:2 graph:1 relaxation:2 year:1 sum:4 prob:1 inverse:1 powerful:1 extends:1 family:2 infimal:2 bound:2 quadratic:4 constraint:6 rmin:2 min:5 qb:1 department:1 structured:2 ball:6 smaller:1 slightly:1 wi:9 rev:1 invariant:10 equation:1 previously:1 foygel:1 turn:2 end:1 available:1 operation:1 apply:1 observe:3 hierarchical:1 spectral:42 alternative:1 original:2 denotes:1 clustering:2 completed:2 matric:1 especially:1 classical:2 society:2 micchelli:4 objective:4 joke:1 gradient:6 link:2 thank:1 simulated:3 parametrized:1 athena:1 w0:1 topic:1 extent:1 besides:1 ratio:1 minimizing:2 gk:2 trace:12 rise:1 zk2f:1 upper:2 observation:3 convolution:2 datasets:5 sm:1 benchmark:1 orthant:2 extended:4 perturbation:5 inv:1 kxw:1 rating:6 introduced:3 namely:1 connection:2 componentwise:1 learned:1 nip:1 able:1 suggested:1 below:3 dimitris:1 wi2:3 regime:1 sparsity:2 max:2 royal:2 overlap:2 natural:2 regularized:2 residual:1 nedic:1 movie:1 orthogonally:6 acknowledgement:1 reordering:1 loss:4 permutation:3 filtering:3 srebro:2 vg:6 validation:6 penalization:2 thresholding:9 grandvalet:2 morale:1 proxk:1 supported:1 keeping:2 taking:1 absolute:4 sparse:5 moreau:2 tolerance:1 benefit:1 author:2 collection:1 simplified:1 transaction:1 mcgraw:1 xi:1 z2r:1 alternatively:1 spectrum:1 continuous:1 search:1 iterative:1 decomposes:1 table:7 learn:1 ku:1 elastic:8 decoupling:1 obtaining:1 mazumder:1 improving:3 alg:2 investigated:1 zou:1 da:5 diag:2 did:1 significance:1 main:1 icann:1 noise:6 morizet:1 body:1 xu:1 en:17 combettes:1 kuk22:1 exceeding:1 third:1 theorem:7 kuk2:1 k21:1 decay:1 consist:2 effectively:1 corr:2 illustrates:2 kx:2 margin:1 chen:1 ordered:2 ux:1 applies:1 springer:2 loses:1 minimizer:1 determines:1 lewis:1 prop:1 obozinski:2 lipschitz:1 absence:3 change:2 fista:1 included:1 specifically:1 determined:3 typical:1 hard:1 uniformly:2 movielens:10 lemma:1 called:1 college:1 wq:3 support:59 arises:1 accelerated:2 princeton:1 argyriou:3 |
4,746 | 5,298 | Beta-Negative Binomial Process and Exchangeable
Random Partitions for Mixed-Membership Modeling
Mingyuan Zhou
IROM Department, McCombs School of Business
The University of Texas at Austin, Austin, TX 78712, USA
[email protected]
Abstract
The beta-negative binomial process (BNBP), an integer-valued stochastic process,
is employed to partition a count vector into a latent random count matrix. As the
marginal probability distribution of the BNBP that governs the exchangeable random partitions of grouped data has not yet been developed, current inference for
the BNBP has to truncate the number of atoms of the beta process. This paper
introduces an exchangeable partition probability function to explicitly describe
how the BNBP clusters the data points of each group into a random number of
exchangeable partitions, which are shared across all the groups. A fully collapsed Gibbs sampler is developed for the BNBP, leading to a novel nonparametric
Bayesian topic model that is distinct from existing ones, with simple implementation, fast convergence, good mixing, and state-of-the-art predictive performance.
1
Introduction
For mixture modeling, there is a wide selection of nonparametric Bayesian priors, such as the Dirichlet process [1] and the more general family of normalized random measures with independent increments (NRMIs) [2, 3]. Although a draw from an NRMI usually consists of countably infinite atoms
that are impossible to instantiate in practice, one may transform the infinite-dimensional problem
into a finite one by marginalizing out the NRMI. For instance, it is well known that the marginalization of the Dirichlet process random probability measure under multinomial sampling leads to
the Chinese restaurant process [4, 5]. The general structure of the Chinese restaurant process is
broadened by [5] to the so called exchangeable partition probability function (EPPF) model, leading
to fully collapsed inference and providing a unified view of the characteristics of various nonparametric Bayesian mixture-modeling priors. Despite significant progress on EPPF models in the past
decade, their use in mixture modeling (clustering) is usually limited to a single set of data points.
Moving beyond mixture modeling of a single set, there has been significant recent interest in mixedmembership modeling, i.e., mixture modeling of grouped data x1 , . . . , xJ , where each group xj =
{xji }i=1,mj consists of mj data points that are exchangeable within the group. To cluster the mj
data points in each group into a random, potentially unbounded number of partitions, which are
exchangeable and shared across all the groups, is a much more challenging statistical problem.
While the hierarchical Dirichlet process (HDP) [6] is a popular choice, it is shown in [7] that a wide
variety of integer-valued stochastic processes, including the gamma-Poisson process [8, 9], betanegative binomial process (BNBP) [10, 11], and gamma-negative binomial process (GNBP), can all
be applied to mixed-membership modeling. However, none of these stochastic processes are able
to describe their marginal distributions that govern the exchangeable random partitions of grouped
data. Without these marginal distributions, the HDP exploits an alternative representation known as
the Chinese restaurant franchise [6] to derive collapsed inference, while fully collapsed inference is
available for neither the BNBP nor the GNBP.
1
The EPPF provides a unified treatment to mixture modeling, but there is hardly a unified treatment
to mixed-membership modeling. As the first step to fill that gap, this paper thoroughly investigates
the law of the BNBP that governs its exchangeable random partitions of grouped data. As directly
deriving the BNBP?s EPPF for mixed-membership modeling is difficult, we first randomize the
group sizes {mj }j and derive the joint distribution of {mj }j and their random partitions on a shared
list of exchangeable clusters; we then derive the marginal distribution of the group-size count vector
m = (m1 , . . . , mJ )T , and use Bayes? rule to further arrive at the BNBP?s EPPF that describes the
prior distribution of a latent column-exchangeable random count matrix, whose jth row sums to mj .
The general method to arrive at an EPPF for mixed-membership modeling using an integer-valued
stochastic process is an important contribution. We make several additional contributions: 1) We
derive a prediction rule for the BNBP to simulate exchangeable random partitions of grouped data
governed by its EPPF. 2) We construct a BNBP topic model, derive a fully collapsed Gibbs sampler that analytically marginalizes out not only the topics and topic weights, but also the infinitedimensional beta process, and provide closed-form update equations for model parameters. 3) The
straightforward to implement BNBP topic model sampling algorithm converges fast, mixes well,
and produces state-of-the-art predictive performance with a compact representation of the corpus.
1.1
Exchangeable Partition Probability Function
Let ?m = {A1 , . . . , Al } denote a random partition of the set [m] = {1, 2, . . . , m}, where there
are l partitions and each element i ? [m] belongs to one and only one set Ak from ?m . If P (?m =
{A1 , . . . , Al }|m) depends only on the number and sizes of the Ak ?s, regardless of their order, then
it is called an exchangeable partition probability function (EPPF) of ?m . An EPPF of ?m is an
EPPF of ? := (?1 , ?2 , . . .) if P (?m |n) = P (?m |m) does not depend on n, where P (?m |n)
denotes the marginal partition probability for [m] when it is known the sample size is n. Such a
constraint can also be expressed as an addition rule for the EPPF [5]. In this paper, the addition rule
is not required and the proposed EPPF is allowed to be dependent on the group sizes (or sample
size if the number of groups is one). Detailed discussions about sample size dependent EPPFs
can be found in [12]. We generalize the work of [12] to model the partition of a count vector into a
latent column-exchangeable random count matrix. A marginal sampler for ?-stable Poisson-Kigman
mixture models (but not mixed-membership models) is proposed in [13], encompassing a large class
of random probability measures and their corresponding EPPFs of ?. Note that the BNBP is not
within that class and both the BNBP?s EPPF and perdition rule are dependent on the group sizes.
1.2
Beta Process
The beta process B ? BP(c, B0 ) is a completely random measure defined on the product space
[0, 1] ? ?, with a concentration parameter c > 0 and a finite and continuous base measure B0 over
a complete separable metric space ? [14, 15] . We define the L?evy measure of the beta process as
?(dpd?) = p?1 (1 ? p)c?1 dpB0 (d?).
(1)
A draw from B ? BP(c, B0 ) can be represented as a countably infinite sum as B =
P
?
k=1 pk ??k , ?k ? g0 , where ?0 = B0 (?) is the mass parameter and g0 (d?) = B0 (d?)/?0
is the base distribution. The beta process is unique in that the beta distribution
is not infinitely
P
divisible, and its measure on a Borel set A ? ?, expressed as B(A) =
k:?k ?A pk , could be
larger than P
one and hence clearly not a beta random variable. In this paper we will work with
Q(A) = ? k:?k ?A ln(1 ? pk ), defined as a logbeta random variable, to analyze model properties
and derive closed-form Gibbs sampling update equations. We provide these details in the Appendix.
2
Exchangeable Cluster/Partition Probability Functions for the BNBP
The integer-valued beta-negative binomial process (BNBP) is defined as
Xj |B ? NBP(rj , B), B ? BP(c, B0 ),
(2)
where for the jth group rj is the negative binomialP
dispersion parameter and Xj |B ? NBP(rj , B)
is a negative binomial process such that Xj (A) = k:?k ?A njk , njk ? NB(rj , pk ) for each Borel
set A ? ?. The negative binomial distribution n ? NB(r, p) has probability mass function (PMF)
n
r
fN (n) = ?(n+r)
n!?(r) p (1 ? p) for n ? Z, where Z = {0, 1, . . .}. Our definition of the BNBP follows
2
those of [10, 7, 11], where for inference [10, 7] used finite truncation and [11] used slice sampling.
There are two recent papers [16, 17] that both marginalize out the beta process from the negative
binomial process, with the predictive structures of the BNBP described as the negative binomial
Indian buffet process (IBP) [16] and ?ice cream? buffet process [17], respectively. Both processes
are also related to the ?multi-scoop? IBP of [10], and they all generalize the binary-valued IBP [18].
Different from these two papers on infinite random count matrices, this paper focuses on generating
a latent column-exchangeable random count matrix, each of whose row sums to a fixed observed
integer. This paper generalizes the techniques developed in [17, 12] to define an EPPF for mixedmembership modeling and derive truncation-free fully collapsed inference.
The BNBP by nature is an integer-valued stochastic process as Xj (A) is a random count for each
Borel set A ? ?. As the negative binomial process is also a gamma-Poisson mixture process, we
can augment (2) as a beta-gamma-Poisson process as
Xj |?j ? PP(?j ), ?j |rj , B ? ?P[rj , B/(1 ? B)], B ? BP(c, B0 ),
where Xj |?j ? PP(?j ) is a Poisson process such thatPXj (A) ? Pois[?j (A)], and ?j |B ?
?P[rj , B/(1?B)] is a gamma process such that ?j (A) = k:?k ?A ?jk , ?jk ? Gamma[rj , pk /(1?
pk )], for each Borel set A ? ?. The mixed-membership-modeling potentials of the BNBP become
clear under this augmented representation. The Poisson process provides a bridge to link count modeling and mixture modeling [7], since Xj ? PP(?j ) can be equivalently generated by first drawing
a total random count mj := Xj (?) ? Pois[?j (?)] and then assigning this random count to disjoint
disjoint Borel sets of ? using a multinomial distribution.
2.1
Exchangeable Cluster Probability Function
In mixed-membership modeling, the size of each group is observed rather being random, thus although the BNBP?s augmented representation is instructive, it is still unclear how exactly the integervalued stochastic process leads to a prior distribution on exchangeable random partitions of grouped
data. The first step for us to arrive at such a prior distribution is to build a sample size dependent
model that treats the number of data points to be clustered (partitioned) in each group as random.
Below we will first derive an exchangeable cluster probability function (ECPF) governed by the
BNBP to describe the joint distribution of the random group sizes and their random partitions over a
random, potentially unbounded number of exchangeable clusters shared across all the groups. Later
we will show how to derive the EPPF from the ECPF using Bayes? rule.
Pmj
Lemma 1. Denote ?k (zji ) as a unit point mass at zji = k, njk = i=1
?k (zji ), and Xj (A) =
P
k:?k ?A njk as the number of data points in group j assigned to the atoms within the Borel set
A ? ?. The Xj ?s generated via the group size dependent model as
P?
?
zji ? k=1 ?jjk
(?) ?k , mj ? Pois(?j (?)),
?j ? ?P[rj , B/(1 ? B)], B ? BP(c, B0 )
(3)
is equivalent in distribution to the Xj ?s generated from a BNBP as in (2).
Proof. With B =
P?
pk ??k and ?j =
P?
k=1 ?jk ??k , the joint distribution of the cluster indices
Qmj
?jz
P? ji
z j = (zj1 , . . . , zjmj ) given ?j and mj can be expressed as p(z j |?j , mj ) = i=1
=
0 =1 ?jk0
k
Q? njk
1
P
mj
k=1 ?jk , which is not in a fully factorized form. As mj is linked to the total random
( ?
k=1 ?jk )
mass ?j (?) with a Poisson distribution, we have the joint likelihood of z j and mj given ?j as
Q? n
(4)
f (z j , mj |?j ) = f (z j |?j , mj )Pois(mj , ?j (?)) = m1j ! k=1 ?jkjk e??jk ,
k=1
which is fully factorized and hence amenable to marginalization. Since ?jk ? Gamma[rj , pk /(1 ?
pk )], we can marginalize ?jk out analytically as f (z j , mj |rj , B) = E?j [f (z j , mj |?j )], leading to
Q? ?(njk +rj ) njk
f (z j , mj |rj , B) = m1j ! k=1 ?(r
pk (1 ? pk )rj .
(5)
j)
Multiplying the above equation with a multinomial coefficient transforms the prior distribution
for the categorical random variables z j to the prior distribution for a random count vector as
Q?
m !
f (nj1 , . . . , nj? |rj , B) = Q? jnjk ! f (z j , mj |rj , B) = k=1 NB(njk ; rj , pk ). Thus in the prior,
k=1
3
for each group, the sample size dependent model in ( 3)
njk ? NB(rj , pk ) random number of
Pdraws
?
data points independently
at
each
atom.
With
X
:=
n
j
k=1 jk ??k , we have Xj |B ? NBP(rj , B)
P
such that Xj (A) = k:?k ?A njk , njk ? NB(rj , pk ).
The Lemma below provides a finite-dimensional distribution obtained by marginalizing out the
infinite-dimensional beta process from the BNBP. The proof is provided in the Appendix.
Lemma 2 (ECPF). The exchangeable cluster probability function (ECPF) of the BNBP, which describes the joint distribution of the random count vector m := (m1 , . . . , mJ )T and its exchangeable
random partitions z = (z11 , . . . , zJmJ ), can be expressed as
h
i
K
? J e??0 [?(c+r? )??(c)] QKJ
?(njk +rj )
?(n?k )?(c+r? ) QJ
QJ
f (z, m|r, ?0 , c) = 0
,
(6)
k=1
j=1
?(c+n?k +r? )
?(rj )
m !
j=1
j
where KJ is the number of observed points of discontinuity for which n?k > 0, r := (r1 , . . . , rJ )T ,
PJ
PJ
r? := j=1 rj , n?k := j=1 njk , and mj ? Z is the random size of group j.
2.2
Exchangeable Partition Probability Function and Prediction Rule
Having the ECPF does not directly lead to the EPPF for the BNBP, as an EPPF describes the distribution of the exchangeable random partitions of the data groups whose sizes are observed rather than
being random. To arrive at the EPPF, first we organize z into a random count matrix NJ ? ZJ?KJ ,
whose jth row represents the partition of the mj data points into the KJ shared exchangeable clusters and whose order of these KJ nonzero columns is chosen uniformly at random from one of the
KJ ! possible permutations, then we obtain a prior distribution on a BNBP random count matrix as
QJ
m !
f (NJ |r, ?0 , c) = K1J ! j=1 QKJ j
f (z, m|r, ?0 , c)
=
k=1 njk !
KJ ?? [?(c+r? )??(c)]
0
?0 e
KJ !
QKJ
?(n?k )?(c+r? )
k=1 ?(c+n?k +r? )
?(njk +rj )
j=1 njk !?(rj ) .
QJ
(7)
As described in detail in [17], although the matrix prior does not appear to be simple, direct calculation shows that this random count matrix has KJ ? Pois {?0 [?(c + r? ) ? ?(c)]} independent and
identically distributed (i.i.d.) columns that can be generated via
n:k ? DirMult(n?k , r1 , . . . , rJ ), n?k ? Digam(r? , c),
(8)
where n:k := (n1k , . . . , nJk )T is the count vector for the kth column (cluster),
the Dirichlet-multinomial (DirMult) distribution [19] has PMF DirMult(n:k |n?k , r) =
?(njk +rj )
?(r? ) QJ
QJn?k !
, and the digamma distribution [20] has PMF Digam(n|r, c) =
j=1
?(rj )
n ! ?(n?k +r? )
j=1
jk
?(r+n)?(c+r)
1
?(c+r)??(c) n?(c+n+r)?(r) ,
where n = 1, 2, . . .. Thus in the prior, the BNBP generates a Poisson
random number of clusters, the size of each cluster follows a digamma distribution, and each cluster
is further partitioned into the J groups using a Dirichlet-multinomial distribution [17].
With both the ECPF and random count matrix prior governed by the BNBP, we are ready to derive
both the EPPF and prediction rule, given in the following two Lemmas, with proofs in the Appendix.
P
Lemma 3 (EPPF). Let PK n:k =m denote a summation over all sets of count vectors with
k=1
PJ
PK
n
=
m,
where
m
=
:k
?
k=1
j=1 mj and n?k ? 1. The group-size dependent exchangeable
partition probability function (EPPF) governed by the BNBP can be expressed as
K
QKJ h ?(n?k )?(c+r? ) QJ ?(njk +rj ) i
?0 J
QJ
f (z|m, r, ?0 , c) = P
m?
K 0 =1
mj !
j=1
0
?0K
K0!
k=1
?(c+n?k +r? )
QK 0
P
PK 0
k0 =1
n:k0 =m
j=1
?(rj )
?(n?k0 )?(c+r? )
k0 =1 ?(c+n?k0 +r? )
?(njk0 +rj )
j=1 njk0 !?(rj )
QJ
,
(9)
which is a function of the cluster sizes {njk }k=1,KJ , regardless of the orders of the indices k?s.
Although the EPPF is fairly complicated, one may derive a simple prediction rule, as shown below,
to simulate exchangeable random partitions of grouped data governed by this EPPF.
Lemma 4 (Prediction Rule). With y ?ji representing the variable y that excludes the contribution
of xji , the prediction rule of the BNBP group size dependent model in (3) can be expressed as
?
?ji
?ji
?k
? n?ji
(n?ji
jk + rj ), for k = 1, . . . , KJ ;
c+n?k +r?
?ji
P (zji = k|z , m, r, ?0 , c) ?
(10)
? ?0
?ji
r
,
if
k
=
K
+
1.
j
J
c+r?
4
(a) ri = 1
25 18 1 3
Group
4
6
8
1
48
4
45
1
48
1
49
1
1
50
2
5 1 1
2
1
1
2 11
2
1
10 33 12
3
4
5
Partition
6
7
8
1
33 9 3 4
8 34
1
49
1
28 14
6 20 26 1
2
8
1 1
15 14 13
1
4 1
1 1 2
8 17 17 7
10
10 17 18 13
2
12
14
1
1
1
1
1
1
1
1 1
16 18 12
1
1
1
6 13 29 7
1 2
1 1
2
2 2 1 1 1
1
1
22 21 6
1
6
1 1
1
10 31 3
1
1
2
4
26 9 9 1 2
2 19 24 2
4 21 14 9
1 3
33 8 3 3 1
49
10
1 1
4 11 18 1 13 3
1
1
1
39 5 1 1 2
2
1
1
3 7
2 37 2
49
(c) ri = 100
(b) ri = 10
1
23
Group
27
Group
49
2
2
5
Partition
10
15
Partition
Figure 1:
Random draws from the EPPF that governs the BNBP?s exchangeable random partitions of
10 groups (rows), each of which has 50 data points. The parameters of the EPPF are set as c = 2,
?0 = ?(c+P 12rj )??(c) , and (a) rj = 1, (b) rj = 10, or (c) rj = 100 for all the 10 groups. The jth row
j
of each matrix, which sums to 50, represents the partition of the mj = 50 data points of the jth group over
a random number of exchangeable clusters, and the kth column of each matrix represents the kth nonempty
cluster in order of appearance in Gibbs sampling (the empty clusters are deleted).
2.3
Simulating Exchangeable Random Partitions of Grouped Data
While the EPPF (9) is not simple, the prediction rule (10) clearly shows that the probability of
selecting k is proportional to the product of two terms: one is related to the kth cluster?s overall
popularity across groups, and the other is only related to the kth cluster?s popularity at that group
and that group?s dispersion parameter; and the probability of creating a new cluster is related to ?0 ,
c, r? and rj . The BNBP?s exchangeable random partitions of the group-size count vector m, whose
prior distribution is governed by (9), can be easily simulated via Gibbs sampling using (10).
Running Gibbs sampling using (10) for 2500 iterations and displaying the last sample, we show in
Figure 1 (a)-(c) three distinct exchangeable random partitions of the same group-size count vector,
under three different parameter settings. It is clear that the dispersion parameters {rj }j play a
critical role in controlling how overdispersed the counts are: the smaller the {rj }j are, the more
overdispersed the counts in each row and those in each column are. This is unsurprising as in the
BNBP?s prior, we have njk ? NB(rj , pk ) and n:k ? DirMult(n?k , r1 , . . . , rJ ). Figure 1 suggests
that it is important to infer rj rather than setting them in a heuristic manner or fixing them.
3
Beta-Negative Binomial Process Topic Model
With the BNBP?s EPPF derived, it becomes evident that the integer-valued BNBP also governs a
prior distribution for exchangeable random partitions of grouped data. To demonstrate the use of
the BNBP, we apply it to topic modeling [21] of a document corpus, a special case of mixture
modeling of grouped data, where the words of the jth document xj1 , . . . , xjmj constitute a group
xj (mj words in document j), each word xji is an exchangeable group member indexed by vji in
a vocabulary with V unique terms. We choose the base distribution as a V dimensional Dirichlet
distribution as g0 (?) = Dir(?; ?, . . . , ?), and choose a multinomial distribution to link a word to a
topic (atom). We express the hierarchical construction of the BNBP topic model as
P?
?
xji ? Mult(?zji ), ?k ? Dir(?, . . . , ?), zji ? k=1 ?jjk
(?) ?k , mj ? Pois(?j (?)),
B
?j ? ?P rj , 1?B
, rj ? Gamma(a0 , 1/b0 ), B ? BP(c, B0 ), ?0 ? Gamma(e0 , 1/f0 ). (11)
Pmj
Let n
:=
i=1 ?v (xji )?k (zji ). Multiplying (4) and the data likelihood f (xj |z j , ?) =
QV vjk
Q?
nvjk
(?
)
, where ? = (?1 , . . . , ?? ), we have f (xj , z j , mj |?, ?j ) =
v=1Q k=1 vk
Q?
V
Q
? QV
v=1 nvjk !
k=1
v=1 Pois(nvjk ; ?vk ?jk ). Thus the BNBP topic model can also be conk=1
mj !
sidered as an infinite Poisson factor model [10], where the term-document
word count matrix
P?
(mvj )v=1:V, j=1:J is factorized under the Poisson likelihood as mvj =
k=1 nvjk , nvjk ?
Pois(?vk ?jk ), whose likelihood f ({nvjk }v,k |?, ?j ) is different from f (xj , z j , mj |?, ?j ) up to
a multinomial coefficient.
QJ
The full conditional likelihood f (x, z, m|?, ?) =
j=1 f (xj , z j , mj |?, ?j ) can be further
nQ
o Q? QJ ?njk e??jk
Q
?
V
nv?k
k=1
jk
Qj=1
expressed as f (x, z, m|?, ?) =
?
, where the
J
k=1
v=1 ?vk
m !
j=1
j
marginalization of ? from the first right-hand-side term is the product of Dirichlet-multinomial distributions and the second right-hand-side term leads to the ECPF. Thus we have a fully marginalized
5
QKJ h ?(V ?) QV ?(nv?k +?) i
likelihood as f (x, z, m|?0 , c, r) = f (z, m|?0 , c, r) k=1
. Directly
v=1
?(n?k +V ?)
?(?)
applying Bayes? rule to this fully marginalized likelihood, we construct a nonparametric Bayesian
fully collapsed Gibbs sampler for the BNBP topic model as
?
?ji
?
?k
n?ji
?ji
? ?+nvji?ji
?k
?
? (n?ji
?ji
jk + rj ), for k = 1, . . . , KJ ;
?ji
V
?+n
c+n
+r
?
?k
?k
P (zji = k|x, z , ?0 , m, c, r) ?
(12)
?
?ji
? 1 ? ?0 ? r ,
if
k
=
K
+
1.
j
J
V
c+r?
In the Appendix we include all the other closed-form Gibbs sampling update equations.
3.1
Comparison with Other Collapsed Gibbs Samplers
One may compare the collapsed Gibbs sampler of the BNBP topic model with that of latent Dirichlet
allocation (LDA) [22], which, in our notation, can be expressed as
P (zji = k|x, z ?ji , m, ?, K) ?
?+n?ji
v ?k
ji
V ?+n?ji
?k
? (n?ji
jk + ?), for k = 1, . . . , K,
(13)
where the number of topics K and the topic proportion Dirichlet smoothing parameter ? are both
tuning parameters. The BNBP topic model is a nonparametric Bayesian algorithm that removes the
need to tune these parameters. One may also compare the BNBP topic model with the HDP-LDA
[6, 23], whose direct assignment sampler in our notation can be expressed as
? ?+n?ji
vji ?k
?ji
?
+ ??
rk ), for k = 1, . . . , KJ?ji ;
?ji ? (njk
P (zji = k|x, z ?ji , m, ?, r? ) ? V ?+n?k
(14)
?1
?ji
r? ),
if k = KJ + 1;
V ? (??
e j ? DP(?, G),
e
where ? is the concentration parameter for the group-specific Dirichlet processes ?
e k ) and r?? = G(?\D
e
e
and r?k = G(?
J ) are the measures of the globally shared Dirichlet process G ?
e
DP(?0 , G0 ) over the observed points of discontinuity and absolutely continuous space, respectively.
Comparison between (14) and (12) shows that distinct from the HDP-LDA that combines a topic?s
global and local popularities in an additive manner as (n?ji
rk ), the BNBP topic model comjk + ??
bines them in a multiplicative manner as
as the product of n?ji
?k and
n?ji
jk +rj
c+n?ji
?k +r?
n?ji
?k
c+n?ji
?k +r?
? (n?ji
jk + rj ). This term can also be rewritten
, the latter of which represents how much the jth document
contributes to the overall popularity of the kth topic. Therefore, the BNBP and HDP-LDA have distinct mechanisms to automatically shrink the number of topics. Note that while the BNBP sampler
in (12) is fully collapsed, the direct assignment sampler of the HDP-LDA in (14) is only partially
e nor the concentration parameter ? are
collapsed as neither the globally shared Dirichlet process G
e (but still
marginalized out. To derive a collapsed sampler for the HDP-LDA that marginalizes out G
not ?), one has to use the Chinese restaurant franchise [6], which has cumbersome book-keeping as
each word is indirectly linked to its topic via a latent table index.
4
Example Results
We consider the JACM1 , PsyReview2 , and NIPS123 corpora, restricting the vocabulary to terms that
occur in five or more documents. The JACM corpus includes 536 documents, with V = 1539 unique
terms and 68,055 total word counts. The PsyReview corpus includes 1281 documents, with V =
2566 and 71,279 total word counts. The NIPS12 corpus includes 1740 documents, with V = 13, 649
and 2,301,375 total word counts. To evaluate the BNBP topic model4 and its performance relative to
that of the HDP-LDA, which are both nonparametric Bayesian algorithms, we randomly choose 50%
1
http://www.cs.princeton.edu/?blei/downloads/
http://psiexp.ss.uci.edu/research/programs data/toolbox.htm
3
http://www.cs.nyu.edu/?roweis/data.html
4
Matlab code available in http://mingyuanzhou.github.io/
2
6
Number of topics
(a) HDP?LDA, JACM
(b) BNBP Topic Model, JACM
1000
1000
100
100
10
10
0
500
1000
1500
0
Number of topics
(c) HDP?LDA, PsyReview
500
1000
1000
100
100
10
1500
10
0
500
1000
1500
0
(e) HDP?LDA, NIPS12
Number of topics
1000
(d) BNBP Topic Model, PsyReview
500
1000
1500
(f) BNBP Topic Model, NIPS12
1000
1000
100
100
10
10
0
500
1000
Gibbs sampling iteration
1500
0
500
1000
Gibbs sampling iteration
1500
Figure 2: The inferred number of topics KJ for the first 1500 Gibbs sampling iterations for the (a) HDP-LDA
and (b) BNBP topic model on JACM. (c)-(d) and (e)-(f) are analogous plots to (a)-(c) for the PsyReview and
NIPS12 corpora, respectively. From bottom to top in each plot, the red, blue, magenta, black, green, yellow,
and cyan curves correspond to the results for ? = 0.50, 0.25, 0.10, 0.05, 0.02, 0.01, and 0.005, respectively.
of the words in each document as training, and use the remaining ones to calculate per-word heldout
perplexity. We set the hyperparameters as a0 = b0 = e0 = f0 = 0.01. We consider 2500 Gibbs
sampling iterations and collect the last 1500 samples. In each iteration, we randomize the ordering
of the words. For each collected sample, we draw the topics (?k |?) ? Dir(? + n1?k , . . . , ? +
nJ?k ), and the topics weights (?jk |?) ? Gamma(njk + rj , pk ) for the BNBP and topic proportions
(? k |?) ? Dir(nj1 + ??
r1 , . . . , njKJ + ??
rKJ ) for the HDP, with which the per-word perplexity is
P P
(s) (s)
P P test
s
k ?vk ?jk
1
, where s ? {1, . . . , S} is the index
computed as exp ? mtest v j mvj ln P P
P
(s) (s)
??
s
v
k
?vk ?jk
of a collected MCMC sample, mtest
vj is the number of test words at term v in document j, and
P P
mtest = v j mtest
.
The
final
results
are averaged over five random training/testing partitions.
vj
The evaluation method is similar to those used in [24, 25, 26, 10]. Similar to [26, 10], we set the
topic Dirichlet smoothing parameter as ? = 0.01, 0.02, 0.05, 0.10, 0.25, or 0.50. To test how the
algorithms perform in more extreme settings, we also consider ? = 0.001, 0.002, and 0.005. All
algorithms are implemented with unoptimized Matlab code. On a 3.4 GHz CPU, the fully collapsed
Gibbs sampler of the BNBP topic model takes about 2.5 seconds per iteration on the NIPS12 corpus
when the inferred number of topics is around 180. The direct assignment sampler of the HDP-LDA
has comparable computational complexity when the inferred number of topics is similar. Note that
when the inferred number of topics KJ is large, the sparse computation technique for LDA [27, 28]
may also be used to considerably speed up the sampling algorithm of the BNBP topic model.
We first diagnose the convergence and mixing of the collapsed Gibbs samplers for the HDP-LDA
and BNBP topic model via the trace plots of their samples. The three plots in the left column of
Figures 2 show that the HDP-LDA travels relatively slowly to the target distributions of the number
of topics, often reaching them in more than 300 iterations, whereas the three plots in the right column
show that the BNBP topic model travels quickly to the target distributions, usually reaching them
in less than 100 iterations. Moreover, Figures 2 shows that the chains of the HDP-LDA are taking
in small steps and do not traverse their distributions quickly, whereas the chains of the BNBP topic
models mix very well locally and traverse their distributions relatively quickly.
A smaller topic Dirichlet smoothing parameter ? generally supports a larger number of topics, as
shown in the left column of Figure 3, and hence often leads to lower perplexities, as shown in
the middle column of Figure 3; however, an ? that is as small as 0.001 (not commonly used in
practice) may lead to more than a thousand topics and consequently overfit the corpus, which is
particularly evident for the HDP-LDA on both the JACM and PsyReview corpora. Similar trends
are also likely to be observed on the larger NIPS2012 corpus if we allow the values of ? to be
even smaller than 0.001. As shown in the middle column of Figure 3, for the same ?, the BNBP
topic model, usually representing the corpus with a smaller number of topics, often have higher
perplexities than those of the HDP-LDA, which is unsurprising as the BNBP topic model has a
multiplicative control mechanism to more strongly shrink the number of topics, whereas the HDP
has a softer additive shrinkage mechanism. Similar performance differences have also been observed
7
(a)
(b)
0
0.01
0.1
Topic Dirichlet parameter ?
(d)
0.5
Heldout perplexity
Number of topics K
2
10
0
0.01
0.1
Topic Dirichlet parameter ?
280
260
240
0.5
10
1000
10
BNBP Topic Model
HDP?LDA
0
0.5
900
1000
0.01
0.1
Topic Dirichlet parameter ?
1000
900
800
0.5
10
2200
2000
2000
1800
1600
1400
1200
0.01
0.1
Topic Dirichlet parameter ?
100
Number of topics K
1000 2000
(i)
2200
1000
100
Number of topics K
(f)
1100
(h)
2
0.01
0.1
Topic Dirichlet parameter ?
300
1200
(g)
Heldout perplexity
Number of topics K
0.01
0.1
Topic Dirichlet parameter ?
(e)
1100
800
0.5
10
10
260
1200
BNBP Topic Model
HDP?LDA
4
280
240
10
10
300
Heldout perplexity
4
320
Heldout perplexity
Heldout perplexity
Number of topics K
2
10
10
(c)
320
BNBP Topic Model
HDP?LDA
Heldout perplexity
4
10
0.5
1800
1600
1400
1200
1000
10
100
Number of topics K
1000
Figure 3: Comparison between the HDP-LDA and BNBP topic model with the topic Dirichlet smoothing parameter ? ? {0.001, 0.002, 0.005, 0.01, 0.02, 0.05, 0.10, 0.25, 0.50}. For the JACM corpus: (a) the posterior
mean of the inferred number of topics KJ and (b) per-word heldout perplexity, both as a function of ?, and (c)
per-word heldout perplexity as a function of the inferred number of topics KJ ; the topic Dirichlet smoothing
parameter ? and the number of topics KJ are displayed in the logarithmic scale. (d)-(f) Analogous plots to
(a)-(c) for the PsyReview corpus. (g)-(i) Analogous plots to (a)-(c) for the NIPS12 corpus, where the results of
? = 0.002 and 0.001 for the HDP-LDA are omitted.
in [7], where the HDP and BNBP are inferred under finite approximations with truncated block
Gibbs sampling. However, it does not necessarily mean that the HDP-LDA has better predictive
performance than the BNBP topic model. In fact, as shown in the right column of Figure 3, the
BNBP topic model?s perplexity tends to be lower than that of the HDP-LDA if their inferred number
of topics are comparable and the ? is not overly small, which implies that the BNBP topic model is
able to achieve the same predictive power as the HDP-LDA, but with a more compact representation
of the corpus under common experimental settings. While it is interesting to understand the ultimate
potentials of the HDP-LDA and BNBP topic model for out-of-sample prediction by setting the
? to be very small, a moderate ? that supports a moderate number of topics is usually preferred
in practice, for which the BNBP topic model could be a preferred choice over the HDP-LDA, as
our experimental results on three different corpora all suggest that the BNBP topic model could
achieve lower-perplexity using the same number of topics. To further understand why the BNBP
topic model and HDP-LDA have distinct characteristics, one may view them from a count-modeling
perspective [7] and examine how they differently control the relationship between the variances and
means of the latent topic usage count vectors {(n1k , . . . , nJk )}k .
We also find that the BNBP collapsed Gibbs sampler clearly outperforms the blocked Gibbs sampler
of [7] in terms of convergence speed, computational complexity and memory requirement. But a
blocked Gibbs sampler based on finite truncation [7] or adaptive truncation [11] does have a clear
advantage that it is easy to parallelize. The heuristics used to parallelize an HDP collapsed sampler
[24] may also be modified to parallelize the proposed BNBP collapsed sampler.
5
Conclusions
A group size dependent exchangeable partition probability function (EPPF) for mixed-membership
modeling is developed using the integer-valued beta-negative binomial process (BNBP). The exchangeable random partitions of grouped data, governed by the EPPF of the BNBP, are strongly influenced by the group-specific dispersion parameters. We construct a BNBP nonparametric Bayesian
topic model that is distinct from existing ones, intuitive to interpret, and straightforward to implement. The fully collapsed Gibbs sampler converges fast, mixes well, and has state-of-the-art predictive performance when a compact representation of the corpus is desired. The method to derive the
EPPF for the BNBP via a group size dependent model is unique, and it is of interest to further investigate whether this method can be generalized to derive new EPPFs for mixed-membership modeling
that could be introduced by other integer-valued stochastic processes, including the gamma-Poisson
and gamma-negative binomial processes.
8
References
[1] T. S. Ferguson. A Bayesian analysis of some nonparametric problems. Ann. Statist., 1973.
[2] E. Regazzini, A. Lijoi, and I. Pr?unster. Distributional results for means of normalized random
measures with independent increments. Annals of Statistics, 2003.
[3] A. Lijoi and I. Pr?unster. Models beyond the Dirichlet process. In N. L. Hjort, C. Holmes,
P. M?uller, and S. G. Walker, editors, Bayesian nonparametrics. Cambridge Univ. Press, 2010.
[4] D. Blackwell and J. MacQueen. Ferguson distributions via P?olya urn schemes. The Annals of
Statistics, 1973.
[5] J. Pitman. Combinatorial stochastic processes. Lecture Notes in Mathematics. SpringerVerlag, 2006.
[6] Y. W. Teh, M. I. Jordan, M. J. Beal, and D. M. Blei. Hierarchical Dirichlet processes. JASA,
2006.
[7] M. Zhou and L. Carin. Negative binomial process count and mixture modeling. To appear in
IEEE Trans. Pattern Anal. Mach. Intelligence, 2014.
[8] A. Y. Lo. Bayesian nonparametric statistical inference for Poisson point processes. Zeitschrift
fur, pages 55?66, 1982.
[9] M. K. Titsias. The infinite gamma-Poisson feature model. In NIPS, 2008.
[10] M. Zhou, L. Hannah, D. Dunson, and L. Carin. Beta-negative binomial process and Poisson
factor analysis. In AISTATS, 2012.
[11] T. Broderick, L. Mackey, J. Paisley, and M. I. Jordan. Combinatorial clustering and the beta
negative binomial process. To appear in IEEE Trans. Pattern Anal. Mach. Intelligence, 2014.
[12] M. Zhou and S. G. Walker. Sample size dependent species models. arXiv:1410.3155, 2014.
[13] M. Lomel??, S. Favaro, and Y. W. Teh. A marginal sampler for ?-Stable Poisson-Kingman
mixture models. arXiv preprint arXiv:1407.4211, 2014.
[14] N. L. Hjort. Nonparametric Bayes estimators based on beta processes in models for life history
data. Ann. Statist., 1990.
[15] R. Thibaux and M. I. Jordan. Hierarchical beta processes and the Indian buffet process. In
AISTATS, 2007.
[16] C. Heaukulani and D. M. Roy. The combinatorial structure of beta negative binomial processes.
arXiv:1401.0062, 2013.
[17] M. Zhou, O.-H. Madrid-Padilla, and J. G. Scott. Priors for random count matrices derived from
a family of negative binomial processes. arXiv:1404.3331v2, 2014.
[18] T. L. Griffiths and Z. Ghahramani. Infinite latent feature models and the Indian buffet process.
In NIPS, 2005.
[19] R. E. Madsen, D. Kauchak, and C. Elkan. Modeling word burstiness using the Dirichlet distribution. In ICML, 2005.
[20] M. Sibuya. Generalized hypergeometric, digamma and trigamma distributions. Annals of the
Institute of Statistical Mathematics, pages 373?390, 1979.
[21] D. Blei, A. Ng, and M. Jordan. Latent Dirichlet allocation. J. Mach. Learn. Res., 2003.
[22] T. L. Griffiths and M. Steyvers. Finding scientific topics. PNAS, 2004.
[23] C. Wang, J. Paisley, and D. M. Blei. Online variational inference for the hierarchical Dirichlet
process. In AISTATS, 2011.
[24] D. Newman, A. Asuncion, P. Smyth, and M. Welling. Distributed algorithms for topic models.
JMLR, 2009.
[25] H. M. Wallach, I. Murray, R. Salakhutdinov, and D. Mimno. Evaluation methods for topic
models. In ICML, 2009.
[26] J. Paisley, C. Wang, and D. Blei. The discrete infinite logistic normal distribution for mixedmembership modeling. In AISTATS, 2011.
[27] I. Porteous, D. Newman, A. Ihler, A. Asuncion, P. Smyth, and M. Welling. Fast collapsed
Gibbs sampling for latent Dirichlet allocation. In SIGKDD, 2008.
[28] D. Mimno, M. Hoffman, and D. Blei. Sparse stochastic inference for latent Dirichlet allocation.
In ICML, 2012.
9
| 5298 |@word m1j:2 middle:2 proportion:2 njk:25 selecting:1 document:11 past:1 existing:2 outperforms:1 current:1 yet:1 assigning:1 fn:1 additive:2 partition:38 remove:1 plot:7 update:3 mackey:1 intelligence:2 instantiate:1 nq:1 blei:6 provides:3 evy:1 traverse:2 five:2 unbounded:2 favaro:1 direct:4 beta:21 become:1 vjk:1 consists:2 combine:1 manner:3 xji:5 nor:2 examine:1 multi:1 olya:1 salakhutdinov:1 globally:2 automatically:1 cpu:1 becomes:1 provided:1 notation:2 moreover:1 mass:4 factorized:3 irom:1 nbp:3 heaukulani:1 developed:4 unified:3 finding:1 nj:4 exactly:1 control:2 exchangeable:37 broadened:1 unit:1 appear:3 organize:1 ice:1 local:1 treat:1 tends:1 io:1 zeitschrift:1 despite:1 mach:3 ak:2 parallelize:3 black:1 downloads:1 wallach:1 suggests:1 challenging:1 collect:1 limited:1 averaged:1 unique:4 testing:1 practice:3 block:1 implement:2 mult:1 word:17 griffith:2 suggest:1 marginalize:2 selection:1 nb:6 collapsed:19 impossible:1 applying:1 www:2 equivalent:1 straightforward:2 regardless:2 independently:1 mixedmembership:3 rule:13 holmes:1 estimator:1 fill:1 deriving:1 steyvers:1 eppf:30 increment:2 analogous:3 annals:3 controlling:1 play:1 construction:1 target:2 smyth:2 elkan:1 element:1 trend:1 roy:1 jk:22 particularly:1 distributional:1 observed:7 role:1 bottom:1 preprint:1 wang:2 calculate:1 thousand:1 ordering:1 burstiness:1 govern:1 complexity:2 broderick:1 mccombs:2 depend:1 predictive:6 titsias:1 completely:1 easily:1 joint:5 htm:1 k0:6 differently:1 various:1 tx:1 represented:1 univ:1 distinct:6 fast:4 describe:3 jk0:1 newman:2 whose:8 heuristic:2 larger:3 valued:9 drawing:1 s:1 statistic:2 transform:1 final:1 online:1 beal:1 advantage:1 product:4 uci:1 mixing:2 achieve:2 roweis:1 intuitive:1 convergence:3 cluster:21 nj1:2 r1:4 empty:1 produce:1 generating:1 requirement:1 franchise:2 converges:2 unster:2 derive:14 jjk:2 fixing:1 school:1 ibp:3 b0:11 progress:1 implemented:1 c:2 implies:1 nips12:6 lijoi:2 stochastic:9 softer:1 clustered:1 summation:1 around:1 normal:1 exp:1 omitted:1 travel:2 combinatorial:3 utexas:1 bridge:1 grouped:11 qv:3 hoffman:1 uller:1 clearly:3 modified:1 rather:3 reaching:2 zhou:6 nrmi:2 poi:8 shrinkage:1 derived:2 focus:1 vk:6 model4:1 fur:1 likelihood:7 digamma:3 sigkdd:1 inference:9 dependent:11 membership:10 ferguson:2 a0:2 unoptimized:1 overall:2 html:1 augment:1 art:3 special:1 fairly:1 smoothing:5 marginal:7 construct:3 having:1 ng:1 atom:5 sampling:15 represents:4 icml:3 carin:2 randomly:1 gamma:13 qkj:5 n1:1 interest:2 investigate:1 evaluation:2 introduces:1 mixture:12 extreme:1 chain:2 amenable:1 indexed:1 pmf:3 desired:1 regazzini:1 re:1 e0:2 instance:1 column:14 modeling:25 assignment:3 njk0:2 unsurprising:2 conk:1 thibaux:1 dir:4 considerably:1 thoroughly:1 quickly:3 choose:3 slowly:1 marginalizes:2 creating:1 book:1 leading:3 kingman:1 potential:2 psiexp:1 includes:3 coefficient:2 explicitly:1 depends:1 later:1 view:2 multiplicative:2 diagnose:1 closed:3 analyze:1 linked:2 red:1 bayes:4 complicated:1 asuncion:2 trigamma:1 contribution:3 n1k:2 qk:1 characteristic:2 variance:1 correspond:1 yellow:1 generalize:2 bayesian:10 none:1 multiplying:2 history:1 influenced:1 cumbersome:1 definition:1 pp:3 proof:3 ihler:1 treatment:2 popular:1 mingyuan:2 higher:1 nonparametrics:1 shrink:2 strongly:2 overfit:1 hand:2 logistic:1 lda:29 scientific:1 usage:1 usa:1 xj1:1 normalized:2 analytically:2 hence:3 assigned:1 overdispersed:2 nonzero:1 generalized:2 evident:2 complete:1 demonstrate:1 variational:1 novel:1 common:1 multinomial:8 ji:34 m1:2 interpret:1 significant:2 blocked:2 cambridge:1 gibbs:22 paisley:3 tuning:1 lomel:1 mathematics:2 moving:1 stable:2 f0:2 base:3 posterior:1 recent:2 madsen:1 perspective:1 belongs:1 moderate:2 perplexity:14 binary:1 life:1 additional:1 employed:1 full:1 mix:3 rj:50 infer:1 z11:1 pnas:1 calculation:1 a1:2 prediction:8 metric:1 poisson:15 dpd:1 iteration:9 arxiv:5 addition:2 whereas:3 walker:2 nv:2 cream:1 member:1 jordan:4 integer:9 hjort:2 identically:1 easy:1 divisible:1 variety:1 marginalization:3 restaurant:4 xj:20 qjn:1 texas:1 qj:11 whether:1 ultimate:1 hardly:1 constitute:1 matlab:2 generally:1 governs:4 detailed:1 clear:3 tune:1 transforms:1 nonparametric:10 locally:1 statist:2 http:4 zj:1 disjoint:2 popularity:4 per:5 overly:1 blue:1 discrete:1 express:1 group:41 deleted:1 neither:2 pj:3 excludes:1 sum:4 arrive:4 family:2 draw:4 appendix:4 investigates:1 comparable:2 cyan:1 occur:1 constraint:1 bp:6 ri:3 generates:1 simulate:2 speed:2 mtest:4 separable:1 urn:1 relatively:2 department:1 mvj:3 truncate:1 across:4 describes:3 smaller:4 partitioned:2 vji:2 pr:2 ln:2 equation:4 pmj:2 count:32 nonempty:1 mechanism:3 available:2 generalizes:1 rewritten:1 apply:1 hierarchical:5 v2:1 indirectly:1 simulating:1 alternative:1 buffet:4 binomial:18 dirichlet:29 clustering:2 denotes:1 running:1 include:1 top:1 marginalized:3 remaining:1 porteous:1 exploit:1 rkj:1 ghahramani:1 chinese:4 build:1 murray:1 padilla:1 g0:4 randomize:2 concentration:3 unclear:1 kth:6 dp:2 link:2 simulated:1 kauchak:1 topic:86 collected:2 hdp:33 code:2 index:4 relationship:1 providing:1 equivalently:1 difficult:1 dunson:1 bnbp:78 potentially:2 trace:1 negative:18 bine:1 implementation:1 anal:2 perform:1 teh:2 dispersion:4 macqueen:1 finite:6 displayed:1 truncated:1 inferred:8 introduced:1 required:1 toolbox:1 blackwell:1 hypergeometric:1 discontinuity:2 trans:2 nip:2 beyond:2 able:2 usually:5 below:3 pattern:2 scott:1 program:1 including:2 green:1 memory:1 power:1 critical:1 business:1 representing:2 scheme:1 github:1 ready:1 categorical:1 kj:18 prior:16 marginalizing:2 relative:1 law:1 fully:13 encompassing:1 permutation:1 heldout:9 mixed:10 interesting:1 lecture:1 proportional:1 allocation:4 jasa:1 displaying:1 editor:1 austin:2 row:6 lo:1 last:2 truncation:4 free:1 jth:7 keeping:1 side:2 allow:1 understand:2 institute:1 wide:2 taking:1 sparse:2 pitman:1 distributed:2 slice:1 curve:1 ghz:1 vocabulary:2 mimno:2 infinitedimensional:1 commonly:1 adaptive:1 welling:2 compact:3 countably:2 preferred:2 global:1 corpus:18 continuous:2 latent:11 decade:1 why:1 table:1 mj:33 nature:1 jz:1 learn:1 contributes:1 necessarily:1 vj:2 aistats:4 pk:19 hyperparameters:1 allowed:1 x1:1 augmented:2 madrid:1 borel:6 governed:7 jmlr:1 hannah:1 rk:2 magenta:1 specific:2 list:1 zj1:1 nyu:1 restricting:1 gap:1 zji:11 logarithmic:1 appearance:1 infinitely:1 jacm:6 likely:1 expressed:9 logbeta:1 partially:1 conditional:1 consequently:1 ann:2 shared:7 springerverlag:1 infinite:9 uniformly:1 sampler:20 lemma:6 called:2 total:5 specie:1 experimental:2 support:2 latter:1 indian:3 absolutely:1 evaluate:1 mcmc:1 princeton:1 instructive:1 |
4,747 | 5,299 | The Infinite Mixture of Infinite Gaussian Mixtures
Bartek Rajwa
Bindley Bioscience Center
Purdue University
W. Lafayette, IN 47907
[email protected]
Halid Z. Yerebakan
Department of
Computer and Information Science
IUPUI
Indianapolis, IN 46202
[email protected]
Murat Dundar
Department of
Computer and Information Science
IUPUI
Indianapolis, IN 46202
[email protected]
Abstract
Dirichlet process mixture of Gaussians (DPMG) has been used in the literature for
clustering and density estimation problems. However, many real-world data exhibit cluster distributions that cannot be captured by a single Gaussian. Modeling
such data sets by DPMG creates several extraneous clusters even when clusters are
relatively well-defined. Herein, we present the infinite mixture of infinite Gaussian mixtures (I2 GMM) for more flexible modeling of data sets with skewed and
multi-modal cluster distributions. Instead of using a single Gaussian for each cluster as in the standard DPMG model, the generative model of I2 GMM uses a single
DPMG for each cluster. The individual DPMGs are linked together through centering of their base distributions at the atoms of a higher level DP prior. Inference
is performed by a collapsed Gibbs sampler that also enables partial parallelization. Experimental results on several artificial and real-world data sets suggest
the proposed I2 GMM model can predict clusters more accurately than existing
variational Bayes and Gibbs sampler versions of DPMG.
1
Introduction
The traditional approach to fitting a Gaussian mixture model onto the data involves using the wellknown expectation-maximization algorithm to estimate component parameters [7]. The major limitation of this approach is the need to define the number of clusters in advance. Although there are
several ways to predict the number of clusters in a data set in an offline manner, these techniques
are in general suboptimal as they decouple the two interdependent tasks: predicting the number of
clusters and predicting model parameters.
Dirichlet process mixture of Gaussians (DPMG), also known as the infinite Gaussian mixture model
(IGMM), is a Gaussian mixture model (GMM) with a Dirichlet process (DP) prior defined over
mixture components [8]. Unlike traditional mixture modeling, DPMG predicts the number of clusters while simultaneously performing model inference. In the DPMG model the number of clusters
can arbitrarily grow to better accommodate data as needed. DPMG in general works well when
the clusters are well-defined with Gaussian-like distributions. When the distributions of clusters are
heavy-tailed, skewed, or multi-modal multiple mixture components per cluster may be needed for
more accurate modeling of cluster data. Since there is no dependency structure in DPMG to asso1
ciate mixture components with clusters, additional mixture components produced during inference
are all treated as independent clusters. This results in a suboptimal clustering of underlying data.
We propose the infinite mixture of IGMMs (I2 GMM) for more accurate clustering of data sets exhibiting skewed and multi-modal cluster distributions. The underlying generative model of I2 GMM
employs a different DPMG for each cluster data. A dependency structure is imposed across individual DPMGs through centering of their base distibutions at one of the atoms of the higher level DP.
This way individual cluster data are modeled by lower level DPs using one DPMG for each cluster
and atoms defining the base distributions of individual clusters and cluster proportions are modeled
by the higher level DP. Our model allows sharing of the covariance matrices across mixture components of the same DPMG. The data model, which is conjugate to the base distributions of both
higher and lower level DPs, makes obtaining closed form solutions of posterior predictive distributions possible. We use a collapsed Gibbs sampler scheme for inference. Each scan of the Gibbs
sampler involves two loops. One that iterates over individual data instances to sample component
indicator variables and another one that iterates over components to sample cluster indicator variables. Conditioned on the cluster indicator variables, component indicator variables can be sampled
in a parallel fashion, which significantly speeds up inference under certain circumstances.
2
Related Work
Dependent Dirichlet processes (DDP) have been studied in the literature for modeling collection
of distributions that vary in time, in spatial region, in covariate space, or in grouped data settings
(images, documents, biological samples). Previous work most related to the current work involves
studies that investigate DDP in grouped data settings.
Teh et al. uses a hierarchical DP (HDP) prior over the base distributions of individual DP models to
introduce a sharing mechanism that allows for sharing of atoms across multiple groups [15]. When
each group data is modeled by a different DPMG this allows for sharing of the same mean vector
and covariance matrix across multiple groups. Such a dependency may potentially be useful in a
multi-group setting. However, when all data are contained in a single group as in the current study
sharing the same mixture component across multiple cluster distributions leads to shared mixture
components being statistically unidentifiable.
The HDP-RE model by Kim & Smyth [10] and transformed DP by Sudderth et al. [14] relaxes the
exact sharing imposed by HDP to have a dependency structure between multiple groups that allow
for components to share perturbed copies of atoms. Although such a sharing mechanism may be
useful for modeling random variations in component parameters across multiple groups, it is not
very useful for clustering data sets with skewed and multi-modal distributions. Both HDP-RE and
transformed DP still model each group data by a single DPMG and suffer from the same drawbacks
as DPMG when clustering data sets with skewed and multi-modal distributions.
The nested Dirichlet Pricess (nDP) by Rodriguez et al. [13] is a DP whose base distribution is in
turn another DP. This model is introduced for modeling multi-group data sets where groups share
not just individual mixture components as in HDP but the entire mixture model defined by a DPMG.
nDP can be adapted to single group data sets with multiple clusters but with the restriction that
each DPMG is shared only once to ensure identifiability. Such a restriction practically eliminates
dependencies across DPMGs modeling different clusters and would not offer clustering property at
the group level.
Unlike existing work which creates dependencies across multiple DPMG through exact or perturbed
sharing of mixture components or through sharing of the entire mixture model, proposed I2 GMM
model associates each cluster with a distinct atom of the higher level DP through centering of the
base distribution of the corresponding DPMG at that atom. Thus, the higher level DP defines metaclusters whereas lower level DPs model actual cluster data. Mixture components associated with
the same DPMG have their own mean vectors but share the same covariance matrix. Apart from
preserving the conjugacy of the data model covariance sharing across mixture components of the
same DPMG allows for identification of clusters that differ in cluster shapes even when they are not
well separated by their means.
2
3
Dirichlet Process Mixture
Dirichlet process is a distribution over discrete distributions. It is parameterized by a concentration
parameter ? and a base distribution H denoted by DP (?H). Each probability mass in a sample
discrete distribution is called as atom. According to the stick-breaking construction of DP [9], each
sample from a DP can be considered as a collection of countably infinite number of atoms. In this
representation base distribution is a prior over the locations of the atoms and concentration parameter affects the distribution of the atom weights, i.e., stick lengths. Another popular characterization
of DP includes the Chinese restaurant process (CRP) [3] which we utilize during model inference.
Discrete nature of its samples makes DP suitable as a prior distribution over mixture weights in
mixture models. Although samples from DP are defined by an infinite dimensional discrete distribution, the posterior distribution conditioned on a finite data always uses finite number of mixture
components.
We denote each data instance by xi ? Rd where i ? {1, ..., n}, n is the total number of data
instances. For each instance, ?i indicates the set of parameters from which the instance is sampled.
For the Gaussian data model ?i = {?i , ?i } where ?i denotes the mean vector and ?i the covariance
matrix. The generative model of the Dirichlet Process Gaussian Mixture is given by (1).
xi ? p(xi |?i )
?i ? G
G ? DP (?H)
(1)
Owing to the discreteness of the distribution G, ?i ?s corresponding to different instances will not be
all distinct. It is this property of DP that offers clustering over ?i and in turn over data instances.
Choosing H from a family of distributions conjugate to the Gaussian distribution produces a closedform solution for the posterior predictive distribution of DPMG. The bivariate prior over the atoms
of G is defined in (2).
H
=
N IW (?0 , ?0 , ?0 , m) = N (?|?0 ,
?
) ? W ?1 (?|?0 , m)
?0
(2)
where ?0 is the prior mean and ?0 is a scaling constant that controls the deviation of the mean
vectors from the prior mean. The parameter ?0 is the scaling matrix and m is degrees of freedom.
The posterior predictive distribution for a Gaussian data model and NIW prior can be obtained
by integrating out ? and ? analytically. Integrating out ? and ? leaves us with the component
indicator variables ti for each instance xi as the only random variables in the state space. Using the
CRP representation of DP, ti ?s can be sampled as in (3).
?p(xi )
if k = K + 1
?i
p(ti = k|X, t ) ?
(3)
?i
? ?i
n?i
if k ? K
k p(xi |Ak , x
k )
? k ) denote the posterior predictive distributions for an empty and ocwhere p(xi ) and p(xi |Ak , x
cupied component, respectively, both of which are multivariate Student-t distributions. X and t
denote the sets of all data instances and their corresponding indicator variables, respectively. nk is
? k are the scatter matrix and sample mean
the number of data instances in component k. Ak and x
for component k, respectively. The superscript ?i notation indicates the exclusion of the effect of
instance i from the corresponding variable. Inference for DPMG can also be performed using the
stick-breaking representation of DP with the actual inference performed either by a Gibbs sampler
or through variational Bayes [5, 11].
4
The Infinite Mixture of Infinite Gaussian Mixture Models
When modeling data sets containing skewed and multi-modal clusters, DPMG tends to produce
multiple components for each cluster. Owing to the single-layer structure of DPMG, no direct associations among different components of the same cluster can be made. As a result of this limitation
all components are treated as independent clusters resulting in a situation where the number of clusters are overpredicted and the actual cluster data are split into multiple subclusters. A more flexible
model for clustering data sets with skewed and multi-modal clusters can be obtained using a two3
layer generative model as in (4).
xi
?i
Gj
Hj
?
?
?
=
(?j , ?j ) ?
G ?
H =
N (xi |?i , ?j )
Gj
DP (?Hj )
N (?j , ?j /?1 )
(4)
G
DP (?H)
N IW (?0 , ?0 , ?0 , m)
In this model, top layer DP generates cluster-specific parameters ?j and ?j according to the base
distribution H and concentration parameter ?. These parameters in turn define the base distributions
Hj of the bottom layer DPs. Since each Hj is representing a different cluster, Hj ?s can be considered
as meta-clusters from which mixture components of the corresponding cluster are generated. In this
model both the number of clusters and the number of mixture components within a cluster can
be potentially infinite hence the name I2 GMM. The top layer DP models the number of clusters,
their sizes, and the base distribution of the bottom layer DPs whereas each bottom layer DP models
the number of components in a cluster and their sizes. Allowing atom locations in the bottom
layer DPGMs to be different than their corresponding cluster atom provides the flexibility to model
clusters that cannot be effectively modeled by a single Gaussian. The scaling parameter ?1 adjusts
within cluster scattering of the component mean vectors whereas the scaling parameter ?0 adjusts
between cluster scattering of the cluster-specific mean vectors. Expressing both H and Hj ?s as
functions of ?j not only preserves the conjugacy of the model but also allows for sharing of the
same covariance matrix across mixture components of the same cluster.
Posterior inference for the proposed model in (4) can be performed by a collapsed Gibbs sampler
n
by iteratively sampling component indicator variables t = {ti }i=1 of data instances and cluster
K
indicator variables c = {ck }k=1 of mixture components. When sampling ti we restrict sampling
with components whose cluster indicator variables are equal to cti in addition to a new component.
The conditional distribution for sampling ti can be expressed by the following equation.
?p(xi )
if k = K + 1
p(ti = k|X, t?i , c) ?
(5)
?i
?i
?
n?i
p(x
|A
,
x
,
S
)
if
k : ck = cti
i
c
k
k
k
k
? ` , n` }`:c` =ck . When sampling component indicator variables, owing to the
where Sck = {A` , x
dependency among data instances, removing a data instance from a component not only affect the
parameters of the components it belongs to but also the corresponding cluster parameters. Technically speaking the parameters of both the component and corresponding cluster has to be updated
for exact inference. However, updating cluster parameters for every data instance removed will significantly slow down inference. For practical purposes we only update component parameters and
assume that removing a single data instance does not significantly change cluster parameters. The
conditional distribution for sampling ck can be expressed by the following equation.
Q
? Q
if j = J + 1
?k
i:ti =k p(xi )
p(ck = j|X, t, c ) ?
(6)
mj i:ti =k p(xi |Sj ) if j ? J
? ` , n` }`:c` =j , J is the number of clusters, and mj is the number of mixture
where Sj = {A` , x
components assigned to cluster j. Next, we discuss the derivation of the component-level posterior
? ?i
predictive distributions, i.e., p(xi |A?i
k ,x
k , Sck ), which can be obtained by evaluating the integral
in (7).
Z Z
?i
?
? ?i
p(xi |A?i
,
x
,
S
)
=
p(xi |?k , ?ck )p(?k , ?ck |A?i
(7)
c
k
k
k
k ,x
k , Sck )??k ??ck
To evaluate the integral in (7) we need the posterior distribution of the component parameters,
? ?i
namely p(?k , ?ck |A?i
k ,x
k , Sck ), which is proportional to
? ?i
p(?k , ?ck |A?i
k ,x
k , Sck ) ?
=
? ?i
p(?k , ?ck , A?i
k ,x
k |Sck )
?i
p(?
xk |?k , ?ck )p(A?i
k |?ck )p(?k |?ck , Sck )p(?k |Sck )
4
(8)
where
p(?
x?i
k |?k , ?ck ) =
p(A?i
=
k |?ck )
p(?k |?ck , Sck ) =
p(?ck |Sck )
=
?
?
=
?
?
=
?1
N ?k , (n?i
?ck
k )
W ?ck , n?i
k ?1
? ?
N (?,
? ?1 ?cP
)
k
P
?1
W
?
+
0
`:c` =ck A` , m +
`:c` =ck (n` ? 1)
P
n` ?1
? ` +?0 ?0
`:c` =ck (n` +?1 ) x
P
n` ?1
`:c` =ck (n` +?1 ) +?0
P
n ?
( `:c =c (n `+?1 ) +?0 )?1
`
k
1
`
P
n` ?1
`:c` =ck (n` +?1 ) +?0 +?1
? ?i
Once we substitute p(?k , ?ck |A?i
k ,x
k , Sck ) into (7) and evaluate the integral we obtain
?i
?i
? k , Sck ) in the form of a multivariate Student-t distribution.
p(xi |Ak , x
? v)
? ?i
? ?,
p(xi |Ak?i , x
k , Sck ) = stu ? t(?,
(9)
? and the degrees of freedom (v) are given below.
The location vector (?
?), the scale matrix (?),
Location vector:
?=
?
? ?i
?
n?i
??
k x
k +?
?i
nk + ?
?
(10)
Scale matrix:
?=
?
?0 +
P
`:c` =ck
A` + A?i
k +
n?i
?
k ?
(?
x?i
k
n?i
?
k +?
? x?i
? T
? ?)(?
k ? ?)
(?
?+n?i
k )v
(?
?+n?i
k +1)
(11)
Degrees of freedom:
v =m+
X
(n` ? 1) + n?i
k ?d+1
(12)
`:c` =ck
? ?i
The cluster-level posterior predictive distributions can be readily obtained from p(xi |A?i
k ,x
k , Sck )
? k , and nk from (10)-(12). Similarly, posterior predictive distribution for an empty
by dropping Ak , x
? k , and nk .
component/cluster can be obtained by dropping Sck from (10)-(12) in addition to Ak , x
Thanks to the two-layer structure of the proposed model, the inference for I2 GMM can be partially
parallelized. Conditioned on the cluster indicator variables, component indicator variables for data
instances in the same cluster can be sampled independent of the data instances in other clusters.
The amount of actual speed up that can be achieved by parallelization depends on multiple factors
including the number of clusters, cluster sizes, and how fast the other loop that iterates over cluster
indicator variables can be run.
5
Experiments
We evaluate the proposed I2 GMM model on five different data sets and compare its performance
against three different versions of DPMG in terms of clustering accuracy and run time.
5.1
Data Sets
Flower formed by Gaussians: We generated a flower-shaped two-dimensional artificial data set
using a different Gaussian mixture model for each of the four different parts (petals, stem, and two
leaves) of the flower. Each part is considered as a separate cluster. Although covariance matrices
are same for all Gaussian components within a mixture they do differ between mixtures to create
clusters of different shapes. Petals are formed by a mixture of nine Gaussians sharing a spherical
covariance. Stem is formed by a mixture of four Gaussians sharing a diagonal covariance. Each leaf
is formed by a mixture of two Gaussians sharing a full covariance. There are a total of seventeen
Gaussian components, four clusters, and 17,000 instances (1000 instances per component) in this
data set. Scatter plot of this data set is shown in Fig 1a.
Lymphoma: Lymphoma data set is one of the data sets used in the FlowCAP (Flow Cytometry Critical Assessment of Population Identification Methods) 2010 competition [1]. This data set consists
5
of thirty sub-data sets each generated from a lymph node biopsy sample of a patient using a flow
cytometer. Flow cytometry is a single-cell screening, analysis, and sorting technology that plays a
crucial role in research and clinical immunology, hematology, and oncology. The cellular phenotypes are defined in FC by combinations of morphological features (measured by elastic light scatter)
and abundances of surface and intracellular markers revealed by fluorescently labeled antibodies. In
the lymphoma data set each of the sub-data set contains thousands of instances with each instance
representing a cell by a five-dimensional feature vector. For each sub-data set cell populations are
manually gated by experts. Each sub-data has between two to four cell populations, i.e., clusters.
Owing to the intrinsic mechanical and optical limitations of a flow cytometer, distributions of cell
populations in the FC data end up being heavy-tailed or skewed, which makes their modeling by a
single Gaussian highly impractical [12]. Although clusters in this data set are relatively well-defined
accurate modeling of cell distributions is a challenge due to skewed nature of distributions.
Rare cell populations: This data set is a small subset of one of the data sets used in the FlowCAP
2012 competition [1]. The data set contains about 279,546 instances with each instance characterizing a white blood cell in a six-dimensional feature space. There are three clusters manually labeled
by experts. This is an interesting data set for two reasons. First, clusters are highly unbalanced in
terms of the number of instances belonging to each cluster. Two of the clusters, which are highly
significant for measuring immunological response of the patient, are extremely rare. The ratios of
the number of instances available from each of the two rare classes to the total number of instances
are 0.0004 and 0.0005, respectively. Second, the third cluster, which contains all cells not belonging to one of the two rare-cell populations, has a distribution that is both skewed and multi-modal
making it extremely challenging to recover its distribution as a single cluster.
Hyperspectral imagery: This data set is a flightline over a university campus. The hyperspectral
data provides image data in 126 spectral bands in the visible and infrared regions. A total of 21,518
pixels from eight different land cover types are manually labeled. Some of the land cover types
such as roof tops have multi-modal distributions. Cluster sizes are also relatively unbalanced with
pixels belonging to roof tops constituting about one half of the labeled pixels. To reduce run time the
dimensionality is reduced by projecting the original data onto its first thirty principal components.
The data with reduced dimensionality is used in all experiments.
Letter recognition: This is a benchmark data set available through the UCI machine learning repository [4]. There are twenty six well-balanced clusters (one for each letter) in this data set.
2
1
0
?1
?2
?3
?4
?5
?6
?7
?3
?2
?1
0
1
2
3
(b) I2 GMM
(a) True Clusters
4
4
2
2
0
0
?2
?2
?4
?4
?6
?6
?8
?8
2
1
0
?1
?2
?3
?4
?5
?6
?7
?10
?4
?3
?2
?1
0
(c) VB
1
2
3
4
?10
?4
?3
?2
?1
0
1
(d) KD-VB
2
3
4
?3
?2
?1
0
1
2
3
(e) ColGibbs
Figure 1: Clusters predicted by I2 GMM, VB, KD-VB, and ColGibbs on the flower data set. Black
contours in the first figure indicate distributions of individual Gaussian components forming the
flower. Each color refers to a different cluster. Points denote data instances.
6
Table 1: Micro and macro F1 scores produced by I2 GMM, VB, KD-VB, and ColGibbs on the five
data sets. For each data set the first line includes micro F1 scores and the second line macro F1
scores. Numbers in parenthesis indicate standard deviations across ten repetitions. Results for the
lyphoma data set are the average of results from thirty sub-data sets.
Data set
I2 GMM
I2 GMMp
VB
KD-VB
ColGibbs
Flower
0.975 (0.032) 0.991 (0.003) 0.640 (0.087)
0.584
0.525 (0.010)
0.982 (0.015) 0.990 (0.002) 0.643 (0.059)
0.639
0.611 (0.009)
Lymphoma
0.920 (0.016) 0.922 (0.020) 0.454 (0.056)
0.819
0.634 (0.034)
0.847 (0.021) 0.847 (0.022) 0.509 (0.044)
0.762
0.656 (0.029)
Rare classes
0.487 (0.031) 0.493 (0.020) 0.182 (0.015)
0.353
0.234 (0.059)
0.756 (0.012) 0.756 (0.010) 0.441 (0.032)
0.472
0.638 (0.023)
Hyperspectral
0.624 (0.017) 0.626 (0.021) 0.433 (0.031)
0.554
0.427 (0.024)
0.667 (0.018) 0.661 (0.012) 0.580 (0.034)
0.380
0.596 (0.020)
Letter Recognition 0.459 (0.015) 0.467 (0.017) 0.420 (0.015)
0.267
0.398 (0.018)
0.460 (0.015) 0.467 (0.017) 0.420 (0.015)
0.267
0.399 (0.018)
5.2
Benchmark Models and Evaluation Metric
We compare the performance of the proposed I2 GMM model with three different versions of DPMG.
These include the collapsed Gibbs sampler version (ColGibbs) discussed in Section 3, the variational
Bayes version (VB) introduced in [5], and the KD-tree based accelerated variational Bayes version
(KD-VB) introduced in [11]. For I2 GMM and ColGibbs we used our own implementations develR
oped in C++. For VB and KD-VB we used existing MATLAB
(Natick,
MA) implementations 1 .
In order to see the effect of parallelization over execution times we ran the proposed technique in
two modes: parallelized (I2 GMMp) and unparallelized (I2 GMM).
All data sets are scaled to have unit variance for each feature. The ColGibbs model has five free
parameters (?, ?0 , m, ?0 , ?0 ), I2 GMM model has two more parameters (?1 , ?) than ColGibbs. We
use vague priors with ? and ? by fixing their value to one. We set m to the minimum feasible value,
which is d+2, to achieve maximum degrees of freedom in the shape of the covariance matrices. The
prior mean ?0 is set to the mean of the entire data. The scale matrix ?0 is set to I/s, where I is the
identity matrix. This leaves the scaling constant s of ?0 , ?0 , and ?1 as the three free parameters. We
use s = 150/(d(logd)), ?0 = 0.05, and ?1 = 0.5 in experiments with all five data sets described
above.
Micro and macro F1 scores are used as performance measures for comparing clustering accuracy of
these four techniques. As one-to-many matchings are expected between true and predicted clusters,
the F1 score for a true cluster is computed as the maximum of the F1 scores for all predicted clusters.
The Gibbs sampler for ColGibbs and I2 GMM are run for 1500 sweeps. The first 1000 samples are
ignored as burn-in and eleven samples drawn with fifty sweeps apart are saved for final evaluation.
We used an approach similar to the one proposed in [6] for matching cluster labels across different
samples. The mode of cluster labels computed across ten samples are assigned as the final cluster
label for each data instance. ColGibbs and I2 GMM use stochastic sampling whereas VB use a
random initialization stage. Thus, these three techniques may produce results that vary from one
run to other on the same data set. Therefore we repeat each experiment ten times and report average
results of ten repetitions for these three techniques.
5.3
Results and Discussion
Micro and macro F1 produced by the four techniques on all five data sets are reported in Table 1. On
the flower data set I2 GMM achieves almost perfect micro and macro F1 scores and correctly predicts
the true number of clusters. The other three techniques produce several extraneous clusters which
lead to poor F1 scores. Clusters predicted by each of the four techniques are shown in Fig. 1. As
expected ColGibbs identify distributions of individual Gaussian components as clusters as opposed
to the actual clusters formed by mixtures of Gaussians. The piece-wise linear cluster boundaries
1
https://sites.google.com/site/kenichikurihara/academic-software/
variational-dirichlet-process-gaussian-mixture-model
7
Table 2: Execution times for I2 GMM, I2 GMMp, VB, KD-VB, and ColGibbs in seconds on the
five data sets. Numbers in parenthesis indicate standard deviations across ten repetitions. For the
lymphoma data set results reported are average run-time per sub-data set.
Data set
I2 GMM
I2 GMMp
VB
KD-VB
ColGibbs
Flower
54 (2)
41 (4)
1 (0.2)
7
59 (1)
Lymphoma
119 (4)
85 (4)
51 (10)
3
63 (3)
Rare classes
9,738 (349) 5,034 (220) 2171 (569)
16
7,250 (182)
Hyperspectral
5,385 (109) 3,456 (174) 582 (156)
2
7,455 (221)
Letter Recognition
1545 (63)
953 (26)
122 (22)
12
2,785 (123)
obtained by VB and KD-VB, splitting original clusters into multiple subclusters, can be explained
by simplistic model assumptions and approximations that characterize variational Bayes algorithms.
On the lymphoma data set the proposed I2 GMM model achieves an average micro and macro F1
scores of 0.920 and 0.848, respectively. These values are not only significantly higher than corresponding F1 scores produced by the other three techniques but also on par with the best performing
techniques in the FlowCAP 2010 competition [2]. Results for thirty individual sub-data sets in the
lymphoma data set are available in the supplementary document. A similar trend is also observed
with the other three real-world data sets as I2 GMM achieves the best F1 score among the four techniques. Between I2 GMM and ColGibbs, I2 GMM consistently generates less number of clusters
across all data sets as expected. Overall, among the three different versions of DPMG that differ in
the inference algorithm used, there is no clear consensus across five data sets as to which version
predicts clusters more accurately. However, the proposed I2 GMM model which extends DPMG to
skewed and multi-modal clusters, clearly stands out as the most accurate model on all five data sets.
Run time results included in Table 2 favors variational Bayes techniques over the Gibbs samplerbased ones as expected. Despite longer run times, significantly higher F1 scores achieved on data
sets with diverse characteristics suggest that I2 GMM can be preferred over DPMG for more accurate
clustering. Results also suggest that I2 GMM can benefit from parallelization. The actual amount of
improvement in execution time depend on data characteristics as well as how fast the unparallelized
loop can be run. The largest gain by parallelization is obtained on the rare classes data set which
offered almost two-fold increase by parallelization on an eight-core workstation.
6
Conclusions
We introduced I2 GMM for more effective clustering of multivariate data sets containing skewed
and multi-modal clusters. The proposed model extends DPMG to introduce dependencies between
components and clusters by a two-layer generative model. Unlike standard DPMG where each
cluster is modeled by a single Gaussian, I2 GMM offers the flexibility to model each cluster data
by a mixture of potentially infinite number of components. Results on experiments with real and
artificial data sets favor I2 GMM over variational Bayes and collapsed Gibbs sampler versions of
DPMG in terms of clustering accuracy. Although execution time can be improved by sampling
component indicator variables in parallel, the amount of speed up that can be gained is limited with
the execution time of the sampling of the cluster indicator variables. As most time consuming part of
this task is the sequential computation of likelihoods for data instances, significant gains in execution
time can be achieved by parallelizing the computation of likelihoods. I2 GMM is implemented in
C++. The source files and executables are available on the web. 2
Acknowledgments
This research was sponsored by the National Science Foundation (NSF) under Grant Number IIS1252648 (CAREER), by the National Institute of Biomedical Imaging and Bioengineering (NIBIB)
under Grant Number 5R21EB015707, and by the PhRMA Foundation (2012 Research Starter Grant
in Informatics). The content is solely the responsibility of the authors and does not represent the
official views of NSF, NIBIB or PhRMA.
2
https://github.com/halidziya/I2GMM
8
References
[1] FlowCAP - flow cytometry: Critical assessment of population identification methods. http:
//flowcap.flowsite.org/.
[2] N. Aghaeepour, G. Finak, FlowCAP Consortium, DREAM Consortium, H. Hoos, T. R. Mosmann, R. Brinkman, R. Gottardo, and R. H. Scheuermann. Critical assessment of automated
flow cytometry data analysis techniques. Nature Methods, 10(3):228?238, mar 2013.
?
? e St Flour 1983, pages 1?198.
[3] D. J. Aldous. Exchangeability and related topics. In Ecole
d?Et?
Springer-Verlag, 1985. Lecture Notes in Math. 1117.
[4] K. Bache and M. Lichman. Uci machine learning repository, 2013.
[5] D. M. Blei and M. I. Jordan. Variational inference for dirichlet process mixtures. Bayesian
Analysis, 1(1):121?144, 2006.
[6] A. J. Cron and M. West. Efficient classification-based relabeling in mixture models. The
American Statistician, 65:16?20, 2011. PMC3110018.
[7] A. Dempster, N. Laird, and D. Rubin. Maximum likelihood from incomplete data via the EM
algorithm. Journal of the Royal Statistical Society, 39(1):1?38, 1977.
[8] T. S. Ferguson. A Bayesian analysis of some nonparametric problems. Annals of Statistics,
1(2):209?230, 1973.
[9] H. Ishwaran and L. F. James. Gibbs sampling methods for stick-breaking priors. Journal of
the American Statistical Association, 96(453):pp. 161?173, 2001.
[10] S. Kim and P. Smyth. Hierarchical Dirichlet processes with random effects. In B. Sch?olkopf,
J. C. Platt, and T. Hoffman, editors, Advances in Neural Information Processing Systems 19,
pages 697?704, Cambridge, MA, 2007. MIT Press.
[11] K. Kurihara, M. Welling, and N. Vlassis. Accelerated variational dirichlet process mixtures.
In Advances in Neural Information Processing Systems 19. 2002.
[12] S. Pyne, X. Hu, K. Wang, E. Rossin, T.-I. Lin, L. M. Maier, C. Baecher-Allan, G. J. McLachlan,
P. Tamayo, D. A. Hafler, P. L. De Jager, and J. P. Mesirov. Automated high-dimensional flow
cytometric data analysis. Proc Natl Acad Sci U S A, 106(21):8519?24, 2009.
[13] A. Rodriguez, D. B. Dunson, and A. E. Gelfand. The nested Dirichlet process. Journal of The
American Statistical Association, 103:1131?1154, 2008.
[14] E. B. Sudderth, A. B. Torralba, W. T. Freeman, and A. S. Willsky. Describing visual scenes
using transformed objects and parts. International Journal of Computer Vision, 77:291?330,
2008.
[15] Y. Teh, M. Jordan, M. Beal, and D. Blei. Hierarchical Dirichlet processes. Journal of the
American Statistical Association, 101(476):1566?1581, 2006.
9
| 5299 |@word repository:2 version:9 proportion:1 hu:1 tamayo:1 covariance:11 accommodate:1 contains:3 score:12 lichman:1 ecole:1 document:2 existing:3 current:2 comparing:1 com:2 scatter:3 readily:1 visible:1 shape:3 enables:1 eleven:1 plot:1 sponsored:1 update:1 generative:5 igmm:1 leaf:4 half:1 xk:1 core:1 blei:2 iterates:3 characterization:1 provides:2 location:4 node:1 math:1 org:1 five:9 direct:1 consists:1 fitting:1 introduce:2 manner:1 allan:1 expected:4 multi:13 freeman:1 spherical:1 actual:6 underlying:2 notation:1 campus:1 mass:1 impractical:1 every:1 ti:9 scaled:1 stick:4 control:1 unit:1 grant:3 platt:1 tends:1 acad:1 despite:1 ak:7 solely:1 black:1 burn:1 initialization:1 studied:1 challenging:1 limited:1 statistically:1 lafayette:1 practical:1 thirty:4 acknowledgment:1 significantly:5 matching:1 integrating:2 refers:1 suggest:3 consortium:2 cannot:2 onto:2 collapsed:5 restriction:2 imposed:2 center:1 splitting:1 adjusts:2 population:7 variation:1 updated:1 annals:1 construction:1 play:1 smyth:2 exact:3 us:3 associate:1 trend:1 recognition:3 updating:1 bache:1 infrared:1 predicts:3 labeled:4 bottom:4 role:1 observed:1 wang:1 thousand:1 region:2 morphological:1 removed:1 ran:1 balanced:1 dempster:1 iupui:4 rajwa:2 depend:1 predictive:7 technically:1 creates:2 vague:1 matchings:1 derivation:1 separated:1 distinct:2 fast:2 effective:1 artificial:3 choosing:1 lymphoma:8 whose:2 gelfand:1 supplementary:1 favor:2 statistic:1 laird:1 superscript:1 final:2 beal:1 propose:1 mesirov:1 macro:6 uci:2 loop:3 flexibility:2 achieve:1 competition:3 olkopf:1 cluster:107 empty:2 produce:4 perfect:1 object:1 fixing:1 measured:1 implemented:1 c:2 involves:3 predicted:4 indicate:3 exhibiting:1 differ:3 biopsy:1 drawback:1 saved:1 owing:4 stochastic:1 f1:13 biological:1 practically:1 considered:3 predict:2 major:1 vary:2 achieves:3 torralba:1 purpose:1 estimation:1 proc:1 label:3 iw:2 grouped:2 largest:1 repetition:3 create:1 hoffman:1 mclachlan:1 mit:1 clearly:1 gaussian:22 always:1 ck:28 hj:6 exchangeability:1 improvement:1 consistently:1 indicates:2 likelihood:3 kim:2 inference:14 dependent:1 ferguson:1 entire:3 transformed:3 pixel:3 overall:1 among:4 flexible:2 classification:1 denoted:1 extraneous:2 spatial:1 equal:1 once:2 shaped:1 atom:14 sampling:10 manually:3 report:1 micro:6 employ:1 simultaneously:1 preserve:1 national:2 individual:10 roof:2 relabeling:1 stu:1 statistician:1 freedom:4 screening:1 investigate:1 highly:3 evaluation:2 flour:1 mixture:48 light:1 natl:1 accurate:5 bioengineering:1 integral:3 partial:1 tree:1 incomplete:1 re:2 instance:30 modeling:11 cover:2 measuring:1 maximization:1 deviation:3 subset:1 rare:7 characterize:1 reported:2 dependency:8 perturbed:2 thanks:1 density:1 immunology:1 st:1 international:1 informatics:1 together:1 imagery:1 containing:2 opposed:1 expert:2 american:4 closedform:1 de:1 student:2 includes:2 depends:1 piece:1 performed:4 view:1 closed:1 responsibility:1 linked:1 bayes:7 recover:1 parallel:2 identifiability:1 formed:5 accuracy:3 variance:1 characteristic:2 maier:1 identify:1 identification:3 bayesian:2 accurately:2 produced:4 sharing:14 centering:3 against:1 pp:1 james:1 bioscience:1 associated:1 workstation:1 sampled:4 seventeen:1 gain:2 popular:1 color:1 indianapolis:2 dimensionality:2 scattering:2 higher:8 subclusters:2 modal:11 response:1 improved:1 unidentifiable:1 mar:1 just:1 stage:1 crp:2 biomedical:1 web:1 assessment:3 marker:1 rodriguez:2 google:1 defines:1 mode:2 immunological:1 name:1 effect:3 true:4 analytically:1 hence:1 assigned:2 fluorescently:1 iteratively:1 i2:37 white:1 skewed:12 during:2 cp:1 logd:1 image:2 variational:10 wise:1 association:4 discussed:1 expressing:1 significant:2 cambridge:1 gibbs:11 rd:1 similarly:1 longer:1 surface:1 gj:2 base:12 posterior:10 own:2 multivariate:3 exclusion:1 aldous:1 belongs:1 apart:2 wellknown:1 flowcap:6 certain:1 verlag:1 meta:1 arbitrarily:1 niw:1 captured:1 preserving:1 additional:1 minimum:1 parallelized:2 multiple:12 full:1 stem:2 academic:1 offer:3 clinical:1 lin:1 parenthesis:2 simplistic:1 cron:1 circumstance:1 expectation:1 patient:2 metric:1 natick:1 vision:1 represent:1 achieved:3 cell:10 whereas:4 addition:2 grow:1 sudderth:2 source:1 crucial:1 sch:1 parallelization:6 eliminates:1 unlike:3 fifty:1 file:1 dundar:2 flow:7 petal:2 jordan:2 revealed:1 split:1 relaxes:1 automated:2 affect:2 restaurant:1 restrict:1 suboptimal:2 reduce:1 six:2 suffer:1 speaking:1 nine:1 matlab:1 ignored:1 useful:3 clear:1 amount:3 nonparametric:1 band:1 ten:5 reduced:2 http:3 nsf:2 per:3 correctly:1 diverse:1 discrete:4 dropping:2 group:12 four:8 scheuermann:1 blood:1 drawn:1 gmm:34 discreteness:1 utilize:1 imaging:1 run:9 parameterized:1 letter:4 extends:2 family:1 almost:2 antibody:1 scaling:5 vb:19 layer:10 ddp:2 fold:1 adapted:1 scene:1 software:1 generates:2 speed:3 extremely:2 performing:2 optical:1 relatively:3 department:2 according:2 combination:1 poor:1 conjugate:2 belonging:3 across:16 kd:10 em:1 making:1 projecting:1 explained:1 equation:2 conjugacy:2 turn:3 discus:1 mechanism:2 describing:1 needed:2 end:1 available:4 gaussians:7 finak:1 eight:2 ishwaran:1 hierarchical:3 spectral:1 substitute:1 original:2 denotes:1 dirichlet:14 clustering:13 ensure:1 top:4 include:1 chinese:1 society:1 sweep:2 cytometer:2 concentration:3 traditional:2 diagonal:1 exhibit:1 dp:32 separate:1 sci:1 topic:1 cellular:1 consensus:1 reason:1 dream:1 willsky:1 hdp:5 length:1 modeled:5 ratio:1 dunson:1 potentially:3 implementation:2 murat:1 twenty:1 gated:1 teh:2 allowing:1 purdue:2 finite:2 benchmark:2 defining:1 situation:1 vlassis:1 oncology:1 cytometry:4 parallelizing:1 introduced:4 namely:1 mechanical:1 lymph:1 herein:1 starter:1 below:1 flower:8 challenge:1 including:1 royal:1 suitable:1 critical:3 treated:2 predicting:2 indicator:15 brinkman:1 representing:2 scheme:1 github:1 technology:1 executables:1 prior:12 literature:2 interdependent:1 sck:15 par:1 lecture:1 interesting:1 limitation:3 proportional:1 foundation:2 degree:4 offered:1 rubin:1 editor:1 share:3 heavy:2 land:2 repeat:1 copy:1 free:2 offline:1 allow:1 institute:1 characterizing:1 benefit:1 boundary:1 world:3 evaluating:1 contour:1 stand:1 author:1 collection:2 made:1 constituting:1 welling:1 sj:2 nibib:2 countably:1 preferred:1 consuming:1 xi:19 ndp:2 tailed:2 table:4 nature:3 mj:2 elastic:1 career:1 obtaining:1 official:1 intracellular:1 fig:2 site:2 west:1 fashion:1 slow:1 sub:7 breaking:3 third:1 abundance:1 removing:2 down:1 specific:2 covariate:1 bivariate:1 intrinsic:1 sequential:1 effectively:1 gained:1 hyperspectral:4 execution:6 conditioned:3 nk:4 sorting:1 phenotype:1 hoos:1 fc:2 forming:1 visual:1 expressed:2 contained:1 partially:1 springer:1 nested:2 ma:2 cti:2 conditional:2 identity:1 price:1 shared:2 feasible:1 change:1 content:1 included:1 infinite:12 sampler:9 kurihara:1 decouple:1 principal:1 called:1 total:4 experimental:1 scan:1 unbalanced:2 accelerated:2 evaluate:3 |
4,748 | 53 | 338
The Connectivity Analysis of Simple Association
- orHow Many Connections Do You Need!
Dan Hammerstrom *
Oregon Graduate Center, Beaverton, OR 97006
ABSTRACT
The efficient realization, using current silicon technology, of Very Large Connection
Networks (VLCN) with more than a billion connections requires that these networks exhibit
a high degree of communication locality. Real neural networks exhibit significant locality,
yet most connectionist/neural network models have little. In this paper, the connectivity
requirements of a simple associative network are analyzed using communication theory.
Several techniques based on communication theory are presented that improve the robustness of the network in the face of sparse, local interconnect structures. Also discussed are
some potential problems when information is distributed too widely.
INTRODUCTION
Connectionist/neural network researchers are learning to program networks that exhibit a broad range of cognitive behavior. Unfortunately, existing computer systems are limited in their ability to emulate such networks efficiently. The cost of emulating a network,
whether with special purpose, highly parallel, silicon-based architectures, or with traditional
parallel architectures, is directly proportional to the number of connections in the network.
This number tends to increase geometrically as the number of nodes increases. Even with
large, massively parallel architectures, connections take time and silicon area. Many existing neural network models scale poorly in learning time and connections, precluding large
implementations.
The connectivity 'costs of a network are directly related to its locality. A network
exhibits locality 01 communication 1 if most of its processing elements connect to other physically adjacent processing elements in any reasonable mapping of the elements onto a planar
surface. There is much evidence that real neural networks exhibit locality2. In this paper,
a technique is presented for analyzing the effects of locality on the process of association.
These networks use a complex node similar to the higher-order learning units of Maxwell et
al. 3
NETWORK MODEL
The network model used in this paper is now defined (see Figure 1).
Definition 1: A recursive neural network, called a c-graph is a graph structure,
r( V,E, e), where:
?
There is a set of CNs (network nodes), V, whose outputs can take a range of positive
real values, Vi, between 0 and 1. There are N. nodes in the set.
?
There is a set of codons, E, that can take a range of positive real values, eij (for
codon j of node i), between 0 and 1. There are Ne codons dedicated to each CN (the
output of each codon is only used by its local CN), so there are a total of Ne N. codons
in the network. The fan-in or order of a codon is Ie. It is assumed that leis the
same for each codon, and Ne is the same for each CN.
*This work was supported in part by the Semiconductor Research Corporation contract no. 86-10-097, and
jointly by the Office of Naval Research and Air Force Office of Scientific Research, ONR contract no. NOOO14 87 K
0259.
? American Institute of Physics 1988
339
Ie
codon j
Figure 1 - A ON
?
Cijk E C is a set of connections of ONs to codons, 1<i ,k<N. and 1<j <Ne , Cijk can
take two values {O,l} indicating the existence of a connection from ON k to codon j
of ON i . 0
Definition 2: The value of ON i is
Vi
=
F[8+~eijl
(1)
J-l
The function, F, is a continuous non-linear, monotonic function, such as the sigmoid function. 0
Definition 9: Define a mapping, D(i,j,x)_y, where x is an input vector to rand y is
the Ie element input vector of codon j of ON i. That is, y has as its elements those elements of Zk of x where Cijk=1, \;/ k. 0
The D function indicates the subset of x seen by codon j of ON i. Different input vectors may map to the same codon vectors, e.g., D(i,j,x)-y and D(i,j,Zj-y, where x~7.
Definition 4: The codon values eij are determined as follows. Let X( m) be input vector
m of the M learned input vectors for ON i. For codon eij of ON i, let Tij be the set of I cdimensional vectors such that lij(m)E T ij , and D(i,j,X(m))-lij(m). That is, each vector,
lij( m) in Tij consists of those subvectors of X( m) that are in codon ii's receptive field.
The variable 1 indexes the L ( i ,i) vectors of T ij . The number of distinct vectors in Tij
may be less than the total number of learned vectors (L(i,j)<M). Though the X(m) are
distinct, the subsets, lij(m), need not be, since there is a possible many to one mapping of
the x vectors onto each vector lij.
Let Xl be the subset of vectors where vi=l (ON i is supposed to output a 1), and
those vectors where vi=O, then define
0#(/) - .izeof {D(i,i ,Z'( m)) "
.,-q}
for q=O,1, and \;/ m that map to this I. That is, ni~(I) is the number of
.xo be
(2)
x vectors that map
340
into "Iij{l) where
tlj-O
and ni}{I) is the number of 7 vectors that map into "Iii (I), where
tI;-1.
The compreaaion of a codon for a vector "Iii(1) then is defined as
n.1.(/)
He.?( I) = _ _I.:....;;J- ' - - IJ
(3)
nj}(I)+nj~(I)
(Hqj(l)=O when both nt, nO-O.) The output of codon
I),
eii' is the maximum-likelyhood
decoding
(4)
Where He indicates the likely hood of t l j - l when a vector 7 that maps to , is input, and'
is that vector 1'(') where min[d.(1'('),y)] \I I, D(i,j,7)-V, and 7 is the current input vector. In other words, , is that vector (of the set of subset learned vectors that codon ij
receives) that is closest (using distance measure d.) to V (the subset input vector). 0
The output of a codon is the "most-likely" output according to its inputs. For example, when there is no code compression at a codon, eji-1, if the "closest" (in terms of some
measure of vector distance, e.g. Hamming distance) subvector in the receptive field of the
codon belongs to a learned vector where the CN is to output a 1. The codons described here
are very similar to those proposed by Marr 4 and implement ne!'Lrest-neighbor classification.
It is assumed that codon function is determined statically prior to network operation, that
is, the desired categories have already been learned.
To measure performance, network capacity is used.
Definition 5: The input noiae, Or, is the average d. between an input vector and the
closest (minimum d.) learned vector, where d. is a measure of the "difference" between two
vectors - for bit vectors this can be Hamming distance. The output noise, 0 0 , is the average
distance between network output and the learned output vector associated with the closest
learned input vector. The in/ormation gain, Gr , is just
Gt
=-10.[ ~~
I
(5)
o
Definition 6: The capacity of a network is the maximum number of learned vectors such
that the information gain, Gr , is strictly positive (>0). 0
COMMUNICATION ANALOGY
Consider a single connection network node, or CN. (The remainder of this paper will
be restricted to a single CN.) Assume that the CN output value space is restricted to two
values, 0 and 1. Therefore, the CN must decide whether the input it sees belongs to the
class of "0" codes, those codes for which it remains off, or the class of "I" codes, those codes
for which it becomes active. The inputs it sees in its receptive field constitute a subset of
the input vectors (the D( ... ) function) to the network. It is also assumed that the CN is an
ideal I-NN (Nearest Neighbor) classifier or feature detector. That is, given a particular set
of learned vectors, the CN will classify an arbitrary input according to the class of the
nearest (using d. as a measure of distance) learned vector. This situation is equivalent to
the case where a single CN has a single codon whose receptive field size is equivalent to that
of the CN.
Imagine a sender who wishes to send one bit of information over a noisy channel. The
sender has a probabilistic encoder that choses a code word (learned vector) according to
some probability distribution. The receiver knows this code set, though it has no knowledge
of which bit is being sent. Noise is added to the code word during its transmission over the
341
channel, which is analogous to applying an input vector to a network's inputs, where the
vector lies within some learned vector's region. The "noise" is represented by the distance
( d,,) between the input vector and the associated learned vector.
The code word sent over the channel consists of those bits that are seen in the receptive field of the ON being modeled. In the associative mapping of input vectors to output
vectors, each ON must respond with the appropriate output (0 or 1) for the associated
learned output vector. Therefore, a ON is a decoder that estimates in which class the
received code word belongs. This is a classic block encoding problem, where increasing the
field size is equivalent to increasing code length. As the receptive field size increases, the
performance of the decoder improves in the presence of noise. Using communication theory
then, the trade-off between interconnection costs as they relate to field size and the functionality of a node as it relates to the correctness of its decision making process (output
errors) can be characterized.
As the receptive field size of a node increases, so does the redundancy of the input,
though this is dependent on the particular codes being used for the learned vectors, since
there are situations where increasing the field size provides no additional information.
There is a point of diminishing returns, where each additional bit provides ever less reduction in output error. Another factor is that interconnection costs increase exponentially
with field size. The result of these two trends is a cost performance measure that has a single global maximum value. In other words, given a set of learned vectors and their probabilities, and a set of interconnection costs, a "best" receptive field size can be determined,
beyond which, increasing connectivity brings diminishing returns.
SINGLE CODON, WITH NO CODE COMPRESSION
A single neural element with a single codon and with no code compression can be
modelled exactly as a communication channel (see Figure 2). Each network node is assumed
to have a single codon whose receptive field size is equal to that of the receptive field size of
the node.
sender
I I
encoder
~
nOIsy
I
I
Ch.nne11~1 : ~
transmitter
receiver
I
decoder
ON
Figure 2 - A Transmission Channel
recelver
342
The operation of the channel is as follows. A bit is input into the channel encoder,
which selects a random code of length N and transmits that code over the channel. The
receiver then, using nearest neighbor classification, decides if the original message was either
a 0 or a 1.
Let M be the number of code words used by the encoder. The rate* then indicates the
density of the code space.
Definition 7: The rate, R, of a communication channel is
R = 10gM
-
(6)
N
o
The block length, N, corresponds directly to the receptive field size of the codon, i.e.,
N=/e. The derivations in later sections use a related measure:
Definition 8: The code utilization, b, is the number of learned vectors assigned to a particular code or
(7)
b can be written in terms of R
b
=
2N (R-l)
(8)
As b approaches 1, code compression increases. b is essentially unbounded, since M may be
significantly larger than 2N. 0
The decode error (information loss) due to code compression is a random variable that
depends on the compression rate and the a priori probabilities, therefore, it will be different
with different learned vector sets and codons within a set. As the average code utilization
for all codons approaches 1, code compression occurs more often and codon decode error is
unavoidable.
Let Zi be the vector output of the encoder, and the input to the channel, where each
element of Zi is either a 1 or o. Let Vi be the vector output of the channel, and the input to
the decoder, where each element is either a 1 or a o. The Noisy Channel Coding Theorem is
now presented for a general case, where the individual M input codes are to be distinguished. The result is then extended to a CN, where, even though M input codes are
used, the ON need only distinguish those codes where it must output a 1 from those where it
must output a o. The theorem is from Gallager (5.6.1)5. Random codes are assumed
throughout.
Theorem 1: Let a discrete memoryless channel have transition probabilities PNU/k)
and, for any positive integer N and positive number R, consider the ensemble of (N,R)
block codes in which each letter of each code word is independently selected according to
fe
l
the probability assignment Q(k). Then, for each message m, l<m< NR
and all p,
O<p<l, the ensemble average probability of decoding error using maximum-likelyhood
decoding satisfies
(9)
where
?In the definitions given here and the theorems below, the notation of Gall ager 6 is used. Many of the
definitions and theorems are also from Gallager.
343
Eo(p,Q)=-ln~ [ ~1 Q(k)PU/kp!p ]
i-il
l+P
(10)
k-il
o
These results are now adjusted ror our special case.
Theorem 2: For a single CN, the average channel error rate ror random code vectors is
Pc.,.~2q(l-q )Pe ? m
where q=Q(k)
\I k
(11)
is the probability or an input vector bit being a 1. 0
These results cover a wide range or models. A more easily computable expression can
be derived by recognizing some or the restrictions inherent in the CN model. First, assume
that all channel code bits are equally likely, that is, \I k, Q( k )=q, that the error model is
the Binary Symmetric Channel (BSC), and that the errors are identically distributed and
independent - that is, each bit has the same probability, f, or being in error, independent
or the code word and the bit position in the code word.
A simplified version or the above theorem can be derived. Maximizing P gives the
tightest bounds:
Pc.,.
< 0.5 O$p~l
maxPe(p)
(12)
where (letting codon input be the block length, N = I c)
P,(p)
:'> eXP{-f,IE,(P)-PR1}
(13)
The minimum value or this expression is obtained when p=1 (for q=0.5):
Eo; -log 2 [
(o.sV,+O.SVl-,)'
1
(14)
SINGLE-CODON WITH CODE COMPRESSION
Unfortunately, the implementation complexity of a codon grows exponentially with the
size or the codon, which limits its practical size. An alternative is to approximate single
codon function of a single CN with many smaller, overlapped codons. The goal is to maintain performance and reduce implementation costs, thus improving the cost/performance of
the decoding process. As codons get smaller, the receptive field size becomes smaller relative
to the number of CNs in the network. When this happens there is codon compression, or
vector alia6ing, that introduces its own errors into the decoding process due to information
loss. Networks can overcome this error by using multiple redundant codons (with overlapping receptive fields) that tend to correct the compression error.
Compression occurs when two code words requiring different decoder output share the
same representation (within the receptive field or the codon) . The following theorem gives
the probability of incorrect codon output with and without compression error.
Theorem 9: For a BSC model where q=0.5, the codon receptive field is Ic, the code utilization is b, and the channel bits are selected randomly and independently, the probability
of a codon decoding error when b > 1 is approximately
Pc.,.
< (l-f)"Pc- [1-(I-f)"
]0.5
where the expected compression error per codon is approximated by
(15)
344
Pc = 0.5
(16)
and from equations 13-14, when 6<1
P,,,, < exp { -
j, [-log [
[(O .?V.+O .?Vl-' J'
I-RI}
(17)
Proof is given in Hammerstrom6 . 0
As 6 grows, Pc approaches 0.5 asymptotically. Thus, the performance of a single codon
degrades rapidly in the presence of even small amounts of compression.
MULTIPLE CODONS WITH CODE COMPRESSION
The use or mUltiple small codons is more efficient than a few large codons, but there
are some fundamental performance constraints. When a codon is split into two or more
smaller codons (and the original receptive field is subdivided accordingly), there are several
effects to be considered. First, the error rate of each new codon increases due to a decrease
in receptive field size (the codon's block code length). The second effect is that the code
utilization, II, will increase for each codon, since the same number of learned vectors is
mapped into a smaller receptive field. This change also increases the error rate per codon
due to code compression. In fact, as the individual codon receptive fields get smaller,
significant code compression occurs. For higher-order input codes, there is an added error
that occurs when the order of the individual codons is decreased (since random codes are
being assumed, this effect is not considered here). The third effect is the mass action of
large numbers of codons. Even though individual codons may be in error, if the majority
are correct, then the ON will have correct output. This effect decreases the total error rate.
Assume that each ON has more than one codon, c>1. The union of the receptive fields
for these codons is the receptive field for the ON with no no restrictions on the degree of
overlap of the various codon receptive fields within or between ONs. For a ON with a large
number of codons, the codon overlap will generally be random and uniformly distributed.
Also assume that the transmission errors seen by different receptive fields are independent.
Now consider what happens to a codon's compression error rate (ignoring transmission
error for the time being) when a codon is replaced by two or more smaller co dons covering
the same receptive field. This replacement process can continue until there are only 1..
codons, which, incidentally, is analogous to most current neural models. For a multiple
codon ON, assume that each codon votes a 1 or o. The summation unit then totals this
information and outputs a 1 if the majority of codons vote for a 1, etc.
Theorem 4: The probability of a ON error due to compression error is
1
Pc = "'\7?';
00
J
21r c!2-cp.-l!2
V cP.(i-p.)
where
Pc
J.2
e 2 dy
(18)
is given in equation 16 and q=0.5.
Pc incorporates the two effects of moving to mUltiple smaller codons and adding more
codons. Using equation 17 gives the total error probability (per bit), PeN:
(19)
Proof is in Hammerstrom6 . 0
345
For networks that perform association as defined in this paper, the connection weights
rapidly approach a single uniform value as the size of the network grows. In information
theoretic terms, the information content of those weights approaches zero as the compression increases. Why then do simple non-conjunctive networks (1-codon equivalent) work at
alI? In the next section I define connectivity cost constraints and show that the answer to
the first question is that the general associative structures defined here do not scale costeffectively and more importantly that there are limits to the degree of distribution of information.
CONNECTIVITY COSTS
It is much easier to assess costs if some implementation medium is assumed. I have
chosen standard silicon, which is a two dimensional surface where ON's and codons take up
surface area according to their receptive field sizes. In addition, there is area devoted to
the metal lines that interconnect the ONs. A specific VLSI technology need not be assumed,
since the comparisons are relative, thus keeping ONs, codons, and metal in the proper proportions, according to a standard metal width, m. (which also includes the inter-metal
pitch). For the analyses performed here, it is assumed that
levels of metal are possible.
m,
In the previous section I established the relationship of network performance , in terms
of the transmission error rate, E, and the network capacity, M. In this section I present an
implementation cost, which is total silicon area, A. This figure can then be used to derive a
cost/performance figure that can be used to compare such factors as codon size and receptive field size. There are two components to the total area: A ON , the area of a ON, and
AMI, the area of the metal interconnect between ONs. AON consists of the silicon area
requirements of the codons for all ONs. The metal area for local, intra-ON interconnect is
considered to be much smaller than that of the codons themselves and of that of the more
global, inter-ON interconnect, and is not considered here. The area per ON is roughly
m.
AON = cfeme(-)
2
(20)
m,
where me is the maximum number of vectors that each codon must distinguish, for 6>1,
me = 2".
Theorem 5: Assume a rectangular, un6ounded* grid of ONs (all ONs are equi-distant
from their four nearest neighbors), where each ON has a bounded receptive field of its nON
nearest ONs, where "ON is the receptive field size for the ON, nON =
C~e
,
where c is the
number of codons, and R is the intra-ON redundancy, that is, the ratio of inputs to
synapses (e.g., when R=l each ON input is used once at the ON, when R=2 each input is
used on the average at two sites). The metal area required to support each ON's receptive
field is (proof is giving by Hammerstrom6 ):
AMI = [
----w-+
"ON3
3"ON
2
~
+9"ON
21 [ m.j2
m,
(21)
The total area per ON, A, then is
?Another implementation IItrategy ill to place &II eNII along a diagonal, which givell n 2 area. However, thill
technique only works ror a bounded number or eNII and when dendritic computation can be lipread over a large
area, which limits the range or p08llible eN implementationll. The theorem IItated here covers an infinite plane or
eNII each with a bounded receptive Held.
346
(22)
o
Even with the assumption of maximum locality, the total metal interconnect area
increases as the cube of the per CN receptive field size!
SINGLE CN SIMULATION
What do the bounds tell us about CN connectivity requirements? From simulations,
increasing the CN's receptive field size improves the performance (increases capacity), but
there is also an increasing cost, which increases faster than the performance! Another
observation is that redundancy is quite effective as a means for increasing the effectiveness
of a CN with constrained connectivity. (There are some limits to R, since it can reach a
point where the intra-CN connectivity approaches that of inter-CN for some situations.)
With a fixed nON, increasing cost-effectiveness (A 1m) is possible by increasing both order
and redundancy.
In order to verify the derived bounds, I also wrote a discrete event simulation of a CN,
where a random set of learned vectors were chosen and the CN's codons were programmed
according to the model presented earlier. Learned vectors were chosen randomly and subjected to random noise, L The CN then attempted to categorize these inputs into two
major groups (CN output = 1 and CN output = 0). For the most part the analytic bounds
agreed with the simulation, though they tended to be optimistic in slightly underestimating
the error. These differences can be easily explained by the simplifying assumptions that
were made to make the analytic bounds mathematically tractable.
DISTRmUTED VS. LOCALIZED
Throughout this paper, it has been tacitly assumed that representations are distributed
across a number of CNs, and that any single CN participates in a number of representations. In a local representation each CN represents a single concept or feature . It is the distribution of representation that makes the CN's decode job difficult, since it is the cause of
the code compression problem.
There has been much debate in the connectionist/neuromodelling community as to the
advantages and disadvantages of each approach; the interested reader is referred to Hinton7 , Baum et al. 8, and BallardQ ? Some of the results derived here are relevant to this
debate. A1s the distribution of representation increases, the compression per CN increases
accordingly. It was shown above that the mean error in a codon's response quickly
approaches 0.5, independent of the input noise . This result also holds at the CN level. For
each individual CN, this error can be offset by adding more codons, but this is expensive
and tends to obviate one of the arguments in favor of distributed representations, that is,
the multi-use advantage, where fewer CNs are needed because of more complex, redundant
encodings. A1s the degree of distribution increases, the required connectivity and the code
compression increases, so the added information that each codon adds to its CN's decoding
process goes to zero (equivalent to all weights approaching a uniform value) .
SUMMARY AND CONCLUSIONS
In this paper a single CN (node) performance model was developed that was based on
Communication Theory. Likewise, an implementation cost model was derived .
The communication model introduced the codon as a higher-order decoding element
and showed that for small codons (much less than total CN fan-in, or convergence) code
compression, or vector aliasing, within the codon's receptive field is a severe problem for
347
large networks. As code compression increases, the information added by any individual
codon to the CN's decoding task rapidly approaches zero .
The cost model showed that for 2-dimensional silicon, the area required for inter-node
metal connectivity grows as the cube of a CN's fan-in.
The combination of these two trends indicates that past a certain point, which is
highly dependent on the probability structure of the learned vector space, increasing the
fan-in of a CN (as is done, for example, when the distribution of representation is increased)
yields diminishing returns in terms of total cost-performance. Though the rate of diminishing returns can be decreased by the use of redundant, higher-order connections.
The next step is to apply these techniques to ensembles of nodes (CNs) operating in a
competitive learning or feature extraction environment.
REFERENCES
[I]
J. Bailey, "A VLSI Interconnect Structure for Neural Networks," Ph.D.
Dissertation, Department of Computer SciencejEngineering, OGC. In Preparation.
[2]
V. B. Mountcastle, "An Organizing Principle for Cerebral Function: The Unit
Module and the Distributed System," in The Mindful Brain, MIT Press, Cambridge,
MA,1977.
[3]
T. Maxwell, C . L. Giles, Y . C. Lee and H. H. Chen, "Transformation Invariance
Using High Order Correlations in Neural Net Architectures," Proceeding8
International Con! on SY8tem8, Man, and Cybernetic8, 1986.
[4]
D. Marr, "A Theory for Cerebral Neocortex," Proc. Roy. Soc . London, vol.
176(1970), pp . 161-234.
[5]
R. G. Gallager, Information Theory and Reliable Communication, John Wiley and
Sons, New York, 1968.
[6]
D. Hammerstrom, "A Connectivity Analysis of Recursive, Auto-Associative
Connection Networks," Tech. Report CS/E-86-009, Dept. of Computer
SciencejEngineering, Oregon Graduate Center, Beaverton, Oregon, August 1986.
[7]
G. E . Hinton, "Distributed Representations," Technical Report CMU-CS-84-157,
Computer Science Dept., Carnegie-Mellon University, Pittsburgh, PA 15213, 1984.
[8]
E. B. Baum, J. Moody and F . Wilczek, "Internal Representations for Associative
Memory," Technical Report NSF-ITP-86-138, Institute for Theoretical Physics,
Santa Barbara, CA, 1986.
[9]
D. H . Ballard, "Cortical Connections and Parallel Processing: Structure and
Function," Technical Report 133, Computer Science Department, Rochester, NY,
January 1985.
| 53 |@word version:1 compression:25 proportion:1 simulation:4 simplifying:1 reduction:1 itp:1 precluding:1 past:1 existing:2 current:3 nt:1 yet:1 conjunctive:1 must:5 written:1 john:1 distant:1 analytic:2 v:1 selected:2 fewer:1 accordingly:2 plane:1 dissertation:1 underestimating:1 provides:2 equi:1 node:13 unbounded:1 along:1 incorrect:1 consists:3 dan:1 inter:4 expected:1 roughly:1 themselves:1 pr1:1 behavior:1 multi:1 aliasing:1 codon:89 brain:1 little:1 subvectors:1 increasing:10 becomes:2 notation:1 bounded:3 mass:1 medium:1 what:2 cijk:3 developed:1 transformation:1 corporation:1 nj:2 ti:1 exactly:1 classifier:1 utilization:4 unit:3 positive:5 local:4 tends:2 limit:4 semiconductor:1 encoding:2 analyzing:1 approximately:1 co:1 limited:1 programmed:1 graduate:2 range:5 practical:1 hood:1 recursive:2 block:5 implement:1 union:1 area:16 significantly:1 word:11 get:2 onto:2 applying:1 restriction:2 equivalent:5 map:5 center:2 maximizing:1 send:1 baum:2 go:1 independently:2 rectangular:1 importantly:1 marr:2 obviate:1 classic:1 analogous:2 imagine:1 gm:1 decode:3 gall:1 overlapped:1 element:10 trend:2 approximated:1 expensive:1 roy:1 pa:1 module:1 region:1 ormation:1 trade:1 decrease:2 environment:1 complexity:1 tacitly:1 proceeding8:1 ror:3 ali:1 easily:2 emulate:1 represented:1 various:1 derivation:1 distinct:2 effective:1 london:1 kp:1 tell:1 whose:3 quite:1 widely:1 larger:1 interconnection:3 encoder:5 ability:1 favor:1 jointly:1 noisy:3 associative:5 advantage:2 net:1 remainder:1 j2:1 relevant:1 realization:1 rapidly:3 organizing:1 poorly:1 supposed:1 billion:1 convergence:1 requirement:3 transmission:5 incidentally:1 derive:1 nearest:5 ij:4 received:1 job:1 soc:1 c:2 correct:3 functionality:1 subdivided:1 dendritic:1 summation:1 mathematically:1 adjusted:1 strictly:1 hold:1 considered:4 ic:1 exp:2 mapping:4 major:1 purpose:1 proc:1 correctness:1 bsc:2 mit:1 office:2 derived:5 naval:1 transmitter:1 indicates:4 tech:1 dependent:2 interconnect:7 nn:1 vl:1 diminishing:4 vlsi:2 selects:1 interested:1 classification:2 ill:1 priori:1 constrained:1 special:2 cube:2 field:36 equal:1 once:1 extraction:1 represents:1 broad:1 choses:1 connectionist:3 report:4 inherent:1 few:1 randomly:2 individual:6 replaced:1 cns:5 replacement:1 maintain:1 message:2 highly:2 intra:3 severe:1 introduces:1 analyzed:1 pc:9 devoted:1 held:1 ager:1 desired:1 theoretical:1 increased:1 classify:1 earlier:1 giles:1 cover:2 disadvantage:1 assignment:1 cost:18 subset:6 uniform:2 recognizing:1 gr:2 too:1 connect:1 answer:1 sv:1 density:1 fundamental:1 international:1 ie:4 lee:1 contract:2 physic:2 off:2 decoding:9 probabilistic:1 participates:1 quickly:1 moody:1 connectivity:12 unavoidable:1 cognitive:1 american:1 return:4 potential:1 coding:1 includes:1 oregon:3 vi:5 depends:1 later:1 performed:1 optimistic:1 ogc:1 competitive:1 parallel:4 rochester:1 ass:1 air:1 ni:2 il:2 who:1 efficiently:1 ensemble:3 likewise:1 yield:1 modelled:1 researcher:1 detector:1 synapsis:1 reach:1 tended:1 definition:10 pp:1 associated:3 transmits:1 proof:3 hamming:2 con:1 gain:2 knowledge:1 improves:2 agreed:1 maxwell:2 higher:4 planar:1 response:1 rand:1 done:1 though:7 just:1 likelyhood:2 until:1 correlation:1 receives:1 wilczek:1 overlapping:1 brings:1 lei:1 grows:4 scientific:1 effect:7 requiring:1 verify:1 concept:1 nooo14:1 assigned:1 memoryless:1 symmetric:1 adjacent:1 during:1 width:1 covering:1 eii:1 theoretic:1 dedicated:1 cp:2 sigmoid:1 exponentially:2 cerebral:2 association:3 discussed:1 he:2 silicon:7 significant:2 cybernetic8:1 cambridge:1 mellon:1 grid:1 mindful:1 moving:1 surface:3 operating:1 gt:1 pu:1 etc:1 add:1 closest:4 own:1 showed:2 belongs:3 barbara:1 massively:1 certain:1 onr:1 binary:1 continue:1 seen:3 minimum:2 additional:2 eo:2 redundant:3 ii:3 relates:1 multiple:5 eji:1 faster:1 characterized:1 technical:3 equally:1 pitch:1 essentially:1 cmu:1 physically:1 addition:1 decreased:2 tend:1 sent:2 incorporates:1 effectiveness:2 integer:1 presence:2 ideal:1 iii:2 identically:1 split:1 zi:2 architecture:4 approaching:1 reduce:1 cn:40 computable:1 whether:2 expression:2 aon:2 york:1 cause:1 constitute:1 action:1 tij:3 generally:1 santa:1 amount:1 neocortex:1 ph:1 category:1 zj:1 nsf:1 per:7 discrete:2 carnegie:1 vol:1 group:1 redundancy:4 four:1 graph:2 asymptotically:1 geometrically:1 letter:1 you:1 respond:1 place:1 throughout:2 reasonable:1 decide:1 reader:1 decision:1 dy:1 bit:12 bound:5 distinguish:2 fan:4 constraint:2 ri:1 argument:1 min:1 statically:1 department:2 according:7 combination:1 smaller:9 slightly:1 across:1 son:1 making:1 happens:2 explained:1 restricted:2 xo:1 ln:1 equation:3 remains:1 needed:1 know:1 letting:1 subjected:1 tractable:1 operation:2 tightest:1 apply:1 appropriate:1 bailey:1 tlj:1 distinguished:1 alternative:1 robustness:1 hammerstrom:2 existence:1 original:2 beaverton:2 giving:1 already:1 added:4 occurs:4 question:1 receptive:33 degrades:1 traditional:1 nr:1 diagonal:1 exhibit:5 distance:7 mapped:1 capacity:4 decoder:5 majority:2 a1s:2 me:2 code:48 length:5 index:1 modeled:1 relationship:1 ratio:1 difficult:1 unfortunately:2 fe:1 relate:1 debate:2 implementation:7 proper:1 perform:1 observation:1 on3:1 january:1 situation:3 emulating:1 communication:11 ever:1 extended:1 hinton:1 arbitrary:1 august:1 community:1 introduced:1 subvector:1 required:3 connection:13 learned:23 established:1 beyond:1 below:1 program:1 reliable:1 memory:1 overlap:2 event:1 force:1 improve:1 technology:2 ne:5 auto:1 lij:5 prior:1 mountcastle:1 relative:2 loss:2 proportional:1 analogy:1 localized:1 degree:4 metal:10 principle:1 share:1 summary:1 supported:1 keeping:1 institute:2 neighbor:4 wide:1 face:1 sparse:1 distributed:7 overcome:1 cortical:1 transition:1 made:1 simplified:1 approximate:1 ons:9 wrote:1 global:2 active:1 decides:1 receiver:3 pittsburgh:1 assumed:10 don:1 continuous:1 svl:1 pen:1 why:1 ballard:1 channel:17 zk:1 ca:1 ignoring:1 improving:1 complex:2 noise:6 site:1 referred:1 en:1 ny:1 wiley:1 iij:1 position:1 wish:1 xl:1 lie:1 pe:1 third:1 theorem:12 specific:1 offset:1 pnu:1 evidence:1 adding:2 chen:1 easier:1 locality:6 eij:3 likely:3 sender:3 gallager:3 monotonic:1 distrmuted:1 ch:1 corresponds:1 satisfies:1 ma:1 goal:1 man:1 content:1 change:1 determined:3 infinite:1 uniformly:1 ami:2 called:1 total:11 invariance:1 attempted:1 vote:2 indicating:1 internal:1 support:1 categorize:1 preparation:1 dept:2 |
4,749 | 530 | The Effective Number of Parameters:
An Analysis of Generalization and Regularization
in Nonlinear Learning Systems
John E. Moody
Department of Computer Science, Yale University
P.O. Box 2158 Yale Station, New Haven, CT 06520-2158
Internet: [email protected], Phone: (203)432-1200
Abstract
We present an analysis of how the generalization performance (expected
test set error) relates to the expected training set error for nonlinear learning systems, such as multilayer perceptrons and radial basis functions. The
principal result is the following relationship (computed to second order)
between the expected test set and tlaining set errors:
(1)
e,
Here, n is the size of the training sample
u;f f is the effective noise
variance in the response variable( s), ,x is a regularization or weight decay
parameter, and Peff(,x) is the effective number of parameters in the nonlinear model. The expectations ( ) of training set and test set errors are
taken over possible training sets and training and test sets e' respectively. The effective number of parameters Peff(,x) usually differs from the
true number of model parameters P for nonlinear or regularized models;
this theoretical conclusion is supported by Monte Carlo experiments. In
addition to the surprising result that Peff(,x) ;/; p, we propose an estimate
of (1) called the generalized prediction error (GPE) which generalizes well
established estimates of prediction risk such as Akaike's F P E and AI C,
Mallows Cp, and Barron's PSE to the nonlinear setting.!
e
lCPE and Peff(>") were previously introduced in Moody (1991).
847
848
Moody
1
Background and Motivation
Many of the nonlinear learning systems of current interest for adaptive control,
adaptive signal processing, and time series prediction, are supervised learning systems of the regression type. Understanding the relationship between generalization
performance and training error and being able to estimate the generalization performance of such systems is of crucial importance. We will take the prediction risk
(expected test set error) as our measure of generalization performance.
2
Learning from Examples
Consider a set of n real-valued input/output data pairs ~(n) = {~i = (xi, yi); i =
1, ... , n} drawn from a stationary density 3(~). The observations can be viewed as
being generated according to the signal plus noise model 2
(2)
where yi is the observed response (dependent variable), Xl are the independent
variables sampled with input probability density O( x), Ei is independent, identicaIIydistributed (iid) noise sampled with density ~(E) having mean 0 and variance (72,3
and J.t(x) is the conditional mean, an unknown function. From the signal plus noise
perspective, the density 3(~) = 3(x, y) can be represented as the product of two
components, the conditional density w(ylx) and the input density O(x):
3(x, y)
w(ylx) O(x)
~(y - J.t(x? O(x)
(3)
The learning problem is then to find an estimate jJ,(x) of the conditional mean J.t(x)
on the basis of the training set ~(n).
In many real world problems, few a priori assumptions can be made about the
functional form of J.t(x). Since a parame~ric function class is usually not known,
one must resort to a nonparametric regression approach, whereby one constructs an
estimate jJ,(x) f(x) for J.t(x) from a large class of functions F known to have good
approximation properties (for example, F could be all possible radial basis function networks and multilayer perceptrons). The class of approximation functions is
usually the union of a countable set of subclasses (specific network architectures)4
A C F for which the elements of each subclass f(w, x) E A are continuously
parametrized by a set of p
p( A) weights w = {WCX; 0:
1, ... , p}. The task of
finding the estimate f( x) thus consists of two problems: choosing the best architecture A and choosing the best set of weights given the architecture. Note that in
=
=
=
w
2The assumption of additive noise ( which is independent of x is a standard assumption
and is not overly restrictive. Many other conceivable signal/noise models can be transformed into this form. For example, the multiplicative model y = /L(x)(l + () becomes
y' = /L'(x) + (' for the transformed variable y' = log(y).
3Note that we have made only a minimal assumption about the noise (, that it is has
finite variance (T2 independent of x. Specifically, we do not need to make the assumption
that the noise density <I>(() is of known form (e.g. gaussian) for the following development.
4For example, a "fully connected two layer perceptron with five internal units".
The Effective Number of Parameters
the nonparametric setting, there does not typically exist a function f( w'" , x) E F
with a finite number of parameters such that f(w"', x)
I1(X) for arbitrary l1(x).
For this reason, the estimators ji( x) = f( x) will be biased estimators of 11( x). 5
w,
=
The first problem (finding the architecture A) requires a search over possible architectures (e.g. network sizes and topologies), usually starting with small architectures and then considering larger ones. By necessity, the search is not usually
exhaustive and must use heuristics to reduce search complexity. (A heuristic search
procedure for two layer networks is presented in Moody and Utans (1992).)
The second problem (finding a good set of weights for f(w,x)) is accomplished by
minimizing an objective function:
(4)
WA = argminw U(A, w, e(n)) .
The objective function U consists of an error function plus a regularizer:
(5)
U(A, w,e(n)) = nEtrain(W,e(n)) + A S(w)
Here, the error Etrain(W,e(n)) measures the "distance" between the target response
values yi and the fitted values f(w,xi):
n
'.
.)] ,
Etrain(W,e(n)) = ~1 "
6 E[y"f(w,x'
(6)
i=l
and S( w) is a regularization or weight-decay function which biases the solution
toward functions with a priori "desirable" characteristics, such as smoothness. The
parameter A ~ 0 is the regularization or weight decay parameter and must itself be
optimized. 6
The most familiar example of an objective function uses the squared error 7
E[yi,f(w, xi)] = [yi - f(w,x i )]2 and a weight decay term:
n
U(A,w,~(n)) = L(yi - f(w,x i ))2
i=l
p
+ A Lg(w CY )
(7)
cy=l
The first term is the sum of squared errors (SSE) of the model f (w, x) with resp ect
to the training data, while the second term penalizes either small, medium, or
large weights, depending on the form of g(wCY). Two common examples of weight
decay functions are the ridge regression form g( wCY) = (w CY )2 (which penalizes large
weights) and the Rumelhart form g(w CY ) = (w CY )2/[(wO)2 + (w CY )2] (which penalizes
weights of intermediate values near wO).
J
5By biased, we mean that the mean squared bias is nonzero: MSB = p(x)((/:t(x))elL(x))2dx > o. Here, p(x) is some positive weighting function on the input space and
()e denotes an expected valued taken over possible training sets ?(n). For unbiasedness
(MSB = 0) to occur, there must exist a set of weights w* such that f(w"', x) = IL(X),
and the learned weights ill must be "close to" w*. For "near unbiasedness", we must have
w* argminwMSB(w) such that (MSB(w?)::::: 0) and ill "close to" w*.
6The optimization of..x will be discussed in Moody (1992).
7 Other error functions, such as those used in generalized linear models (see for example
McCullagh and NeIder 1983) or robust statistics (see for example Huber 1981) are more
appropriate than the squared error if the noise is known to be non-gaussian or the data
contains many outliers.
=
849
850
Moody
An example of a regularizer which is not explicitly a weight decay term is:
S(w) =
1
dxO(x)IIOxxf(w, x)112
.
(8)
This is a smoothing term which penalizes functional fits with high curvature.
3
Prediction Risk
=
With l1(x) f( w[c;( n)], x) denoting an estimate of the true regression function J.t(x)
trained on a data set c;( n), we wish to estimate the prediction risk P, which is the
exp ected error of 11( x) in predicting future data. In principle, we can either define
P for models l1(x) trained on arbitrary training sets of size n sampled from the
unknown density w(ylx )O( x) or for training sets of size n with input density equal
to the empirical density defined by the single training set available:
1 n
O'(x)
= -n L
8(x - xi) .
(9)
i=1
For such training sets, the n inputs xi are held fixed, but the response variables yi
are sampled with the conditional densities w(ylx i ). Since O'(x) is known, but O(x)
is generally not known a priori, we adopt the latter approach.
For a large ensemble of such training sets, the expected training set error is 8
(f...;n( A)),
/
~t
f[Y;, I(
J~ t.
\
w[~( n)], X;)])
E
1=1
f[lI ,J(
w[~( n)], x;)]
{g
wMx; )dll }
(10)
For a future exemplar (x,z) sampled with density w(zlx)O(x), the prediction risk
P is defined as:
P =
Jf[z,J(w[~(n)]'x)lw(zlx)n(x) {g
W(Y;IX;)d Y;} dzdx
(11)
Again, however, we don't assume that O(x) is known, so computing (11) is not
possible.
Following Akaike (1970), Barron (1984), and numerous other authors (see Eubank
1988), we can define the prediction risk P as the expected test set error for test sets
of size n e'(n)
{c;i,
(xi,zi); i
1, ... ,n} having the empirical input density
0' (x). This expected test set error has form:
=
(f.".(A)),<,
=
/
=
~ tf[i,J(w[~(n)l,x;)l)
J! t.
\
(12)
EE'
1=1
f[z; ,J(
w[~( n)], x;)I
{g
w(y; Ix; )w( z; Ix;)dy; dz; }
8Following the physics convention, we use angled brackets ( ) to denote expected values.
The subscripts denote the random variables being integrated over.
The Effective Number of Parameters
We take (12) as a proxy for the true prediction risk P.
In order to compute (12), it will not be necessary to know the precise functional
form of the noise density ~(f). Knowing just the noise variance (T2 will enable an
exact calculation for linear models trained with the SSE error and an approximate
calculation correct to second order for general nonlinear models. The results of
these calculations are presented in the next two sections.
4
The Expected Test Set Error for Linear Models
The relationship between expected training set and expected test set errors for linear
models trained using the SSE error function with no regularizer is well known in
statistics (Akaike 1970, Barron 1984, Eubank 1988). The exact relation for test and
training sets with density (9):
(13)
As pointed out by Barron (1984), (13) can also apply approximately to the case of
a nonlinear model f( w, x) trained by minimizing the sum of squared errors SSE.
This approximation can be arrived at in two ways. First, the model few, x) can be
treated as locally linear in a neighborhood of w. This approximation ignores the
hessian and higher order shape of f( w, x) in parameter space. Alternatively, the
model f( w, x) can be assumed to be locally quadratic in parameter space wand
unbiased.
However, the extension of (13) as an approximate relation for nonlinear models
breaks down if any of the following situations hold:
The SSE error function is not used. (For example, one may use a robust error
measure (Huber 1981) or log likelihood error measure instead.)
A regularization term is included in the objective function. (This introduces bias.)
The locally linear approximation for few, x) is not good.
The unbiasedness assumption for few, x) is incorrect.
5
The Expected Test Set Error for Nonlinear Models
For neural network models, which are typically nonparametric (thus biased) and
highly nonlinear, a new relationship is needed to replace (13). We have derived
such a result correct to second order for nonlinear models:
(14)
This result differs from the classical result (13) by the appearance of Pelf ()..) (the
effective number of parameters), (T;1f (the effective noise variance in the response
variable( s?, and a dependence on ).. (the regularization or weight decay parameter).
A full derivation of (14) will be presented in a longer paper (Moody 1992). The
result is obtained by considering the noise terms fi for both the training and test
851
852
Moody
sets as perturbations to an idealized model fit to noise free data. The perturbative
expansion is computed out to second order in the fi s subject to the constraint that
the estimated weights w minimize the perturbed objective function. Computing
expectation values and comparing the expansions for expected test and training
errors yields (14). It is important to re-emphasize that deriving (14) does not
require knowing the precise form of the noise density ~(f). Only a knowledge of u 2
is assumed.
The effective number of parameters Peff(>') usually differs from the true number
of model parameters P and depends upon the amount of model bias, model nonlinearity, and on our prior model preferences (eg. smoothness) as determined by
the regularization parameter A and the form of our regularizer. The precise form of
Peff(A) is
Peff(A)
1",
_
= trC = -2.
L..J1iaUaJTf3i
,
(15)
laf3
where C is the generalized influence matrix which generalizes the standard influence
or hat matrix of linear regression, 1ia is the n x p matrix of derivatives of the training
error function
1ia
=-88 . -88 nE(w,e(n)) ,
yl
wa
(16)
and U;;J is the inverse of the hessian of the total objective function
Uaf3
8
8
= 8w a 8wf3 U(A, w, e(n))
(17)
In the general case that u 2 (x) varies with location in the input space x, the effective
noise variance u;ff is a weighted average of the noise variances u 2 {xi). For the
uniform signal plus noise model model we have described above, u;f f = u 2 ?
6
The Effects of Regularization
In the neural network community, the most commonly used regularizers are weight
decay functions. The use of weight decay is motivated by the intuitive notion that
it removes unnecessary weights from the model. An analysis of Peff{A) with weight
decay (A > 0) confirms this intuitive notion. Furthermore, whenever u 2 > 0 and
n < 00, there exists some Aoptimal > 0 such that the expected test set error (12) is
minimized. This is because weight decay methods yield models with lower model
variance, even though they are biased. These effects will be discussed further in
Moody (1992).
For models trained with squared error ~SSE) and quadratic weight decay g(w a ) =
(w a )2, the assumptions of unbiasedness or local linearizability lead to the following
expression for Peff{A) which we call the linearized effective number of parameters
Plin{A):
(18)
9S trictly speaking, a model with quadratic weight decay is unbiased only if the "true"
weights are o.
The Effective Number of Parameters
mplied. Li nearized. and Full P-effectiv e
Linearized
. - -----.--~--
.... ..
,
'II(
~,
Full
..0 ?
~"
,
t
~
'"K
ImpJi d
E
~--
:i
~ ~-
1\'
u
::=> ,
i"
,.
,.
,.
1-
I.
"
Weight Decay Parameter (Lambda)
Figure 1: The full Peff(~) (15) agrees with the implied Pimp(~) (19) to within
exp erimental error, whereas the linearized Plin (~) (18) does not (except for very
large ~). These results verify the significance of (14) and (15) for nonlinear models.
Here,
",01
is the a th eigenvalue of the P x P matrix
J{
= TtT, with T
as defined in
(16).
The form of Pelf(~) can be computed easily for other weight decay functions, such
as the Rumelhart form g(w Ol ) = (w Ol )2/[(wO)2 + (w Ol )2]. The basic result for all
weight decay or regularization functions , however, is that Peff (~) is a decreasing
P and Pelf(oo)
0, as is evident in the special case
function of ~ with Pelf(O)
(18). If the model is nonlinear and biased, then Pelf (0) generally differs from p.
=
7
=
Testing the Theory
To test the result (14) in a nonlinear context, we computed the full Pej j(A) (15),
the linearized Plin(~) (18), and the implied number of parameters Pimp (A) (19) for a
nonlinear test problem. The value of Pimp (~) is obtained by computing the expected
training and test errors for an ensemble of training sets of size n with known noise
variance u 2 and solving for Pelf (~) in equation (14):
(19)
The """s indicate Monte Carlo estimates based on computations using a finite ensemble (10 in our experiments) of training sets. The test problem was to fit training
sets of size 50 generated as a sum of three sigmoids plus noise, with the noise sampled from the uniform density. The model architecture f(w , x) was also a sum of
three sigmoids and the weights w were estimated by minimizing (7) with quadratic
weight decay. See figure 1.
853
854
Moody
8
G PE: An Estimate of Prediction Risk for Nonlinear
Systems
A number of well established, closely related criteria for estimating the prediction
risk for linear or linearizable models are available. These include Akaike's F P E
(1970), Akaike's AlC (1973) Mallow's Cp (1973), and Barron's PSE (1984). (See
also Akaike (1974) and Eubank (1988).) These estimates are all based on equation
(13).
The generalized prediction error (G P E) generalizes the classical estimators F P E,
AIC, Cp, and PSE to the nonlinear setting by estimating (14) as follows:
-.
()
( ) = PGPE = &train n
GPE>'
+ 2u~2eff Peff(>')
n
.
(20)
The estimation process and the quality of the resulting GP E estimates will be
described in greater detail elsewhere.
Acknowledgements
The author wishes to thank Andrew Barron and Joseph Chang for helpful conversations.
This research was supported by AFOSR grant 89-0478 and ONR grant N00014-89-J-1228.
References
H. Akaike. (1970). Statistical predictor identification. Ann. Inst. Stat. Math., 22:203.
H. Akaike. (1973). Information theory and an extension of the maximum likelihood
principle. In 2nd Inti. Symp. on Information Theory, Akademia Kiado, Budapest, 267.
H. Akaike. (1974). A new look at the statistical model identification.
Auto. Control, 19:716-723.
IEEE Trans.
A. Barron. (1984). Predicted squared error: a criterion for automatic model selection. In
Self-Organizing Methods in Modeling, S. Farlow, ed., Marcel Dekker, New York.
R. Eubank. (1988). Spline Smoothing and Nonparametric Regression. Marcel Dekker,
New York.
P. J. Huber. (1981). Robust Statistics. Wiley, New York.
C. L. Mallows. (1973). Some comments on Cpo Technometrics 15:661-675.
P. McCullagh and J.A. NeIder. (1983). Generalized Linear Models. Chapman and Hall,
New York.
J. Moody. (1991). Note on Generalization, Regularization, and Architecture Selection
in Nonlinear Learning Systems. In B.H. Juang, S.Y. Kung, and C.A. Kamm, editors,
Neural Networks for Signal Processing, IEEE Press, Piscataway, N J.
J. Moody. (1992). Long version of this paper, in preparation.
J. Moody and J. Utans. (1992). Principled architecture selection for neural networks:
application to corporate bond rating prediction. In this volume.
| 530 |@word version:1 nd:1 dekker:2 confirms:1 linearized:4 necessity:1 series:1 contains:1 denoting:1 current:1 comparing:1 surprising:1 trc:1 dx:1 must:6 perturbative:1 john:1 additive:1 shape:1 remove:1 msb:3 stationary:1 math:1 location:1 preference:1 five:1 ect:1 incorrect:1 consists:2 symp:1 huber:3 expected:16 ol:3 decreasing:1 kamm:1 considering:2 becomes:1 estimating:2 medium:1 finding:3 subclass:2 control:2 unit:1 grant:2 positive:1 local:1 farlow:1 subscript:1 approximately:1 plus:5 testing:1 mallow:3 union:1 differs:4 procedure:1 empirical:2 radial:2 close:2 selection:3 risk:9 influence:2 context:1 dz:1 starting:1 estimator:3 deriving:1 notion:2 sse:6 resp:1 target:1 exact:2 akaike:9 us:1 element:1 rumelhart:2 observed:1 cy:6 connected:1 linearizable:1 principled:1 complexity:1 trained:6 solving:1 upon:1 basis:3 easily:1 represented:1 regularizer:4 derivation:1 train:1 effective:12 monte:2 choosing:2 neighborhood:1 exhaustive:1 heuristic:2 larger:1 valued:2 statistic:3 gp:1 itself:1 eigenvalue:1 propose:1 product:1 argminw:1 budapest:1 organizing:1 intuitive:2 juang:1 depending:1 oo:1 andrew:1 stat:1 exemplar:1 c:1 predicted:1 indicate:1 marcel:2 convention:1 closely:1 correct:2 enable:1 eff:1 require:1 generalization:6 extension:2 hold:1 hall:1 exp:2 adopt:1 angled:1 estimation:1 bond:1 agrees:1 tf:1 weighted:1 gaussian:2 derived:1 likelihood:2 helpful:1 inst:1 dependent:1 typically:2 integrated:1 relation:2 transformed:2 i1:1 ill:2 priori:3 development:1 smoothing:2 special:1 ell:1 equal:1 construct:1 having:2 chapman:1 look:1 future:2 minimized:1 t2:2 spline:1 haven:1 few:4 familiar:1 technometrics:1 interest:1 highly:1 introduces:1 bracket:1 regularizers:1 held:1 necessary:1 penalizes:4 re:1 theoretical:1 minimal:1 fitted:1 modeling:1 uniform:2 predictor:1 perturbed:1 varies:1 unbiasedness:4 density:17 physic:1 yl:1 continuously:1 moody:14 squared:7 again:1 lambda:1 resort:1 derivative:1 li:2 explicitly:1 idealized:1 depends:1 multiplicative:1 break:1 neider:2 ttt:1 dll:1 minimize:1 il:1 variance:9 characteristic:1 ensemble:3 yield:2 identification:2 eubank:4 iid:1 carlo:2 pej:1 whenever:1 ed:1 sampled:6 knowledge:1 conversation:1 higher:1 supervised:1 response:5 box:1 though:1 furthermore:1 just:1 ei:1 nonlinear:19 quality:1 effect:2 verify:1 true:5 pimp:3 unbiased:2 regularization:10 nonzero:1 eg:1 self:1 whereby:1 criterion:2 generalized:5 arrived:1 evident:1 ridge:1 cp:3 l1:3 fi:2 common:1 functional:3 ji:1 volume:1 discussed:2 ai:1 smoothness:2 automatic:1 pointed:1 nonlinearity:1 longer:1 curvature:1 perspective:1 phone:1 n00014:1 onr:1 yi:7 accomplished:1 greater:1 signal:6 ii:1 relates:1 full:5 desirable:1 corporate:1 calculation:3 long:1 prediction:13 regression:6 basic:1 multilayer:2 expectation:2 addition:1 background:1 whereas:1 crucial:1 biased:5 comment:1 subject:1 call:1 ee:1 near:2 intermediate:1 erimental:1 fit:3 zi:1 architecture:9 topology:1 reduce:1 knowing:2 motivated:1 expression:1 pse:3 wo:3 hessian:2 speaking:1 jj:2 york:4 generally:2 ylx:4 amount:1 nonparametric:4 locally:3 exist:2 utans:2 estimated:2 overly:1 drawn:1 sum:4 wand:1 inverse:1 ric:1 dy:1 layer:2 ct:1 internet:1 aic:1 yale:3 quadratic:4 occur:1 constraint:1 ected:1 department:1 according:1 piscataway:1 joseph:1 outlier:1 inti:1 taken:2 equation:2 previously:1 needed:1 know:1 generalizes:3 available:2 apply:1 barron:7 appropriate:1 pgpe:1 hat:1 denotes:1 include:1 restrictive:1 classical:2 implied:2 objective:6 dependence:1 conceivable:1 distance:1 thank:1 parametrized:1 parame:1 reason:1 toward:1 relationship:4 minimizing:3 lg:1 countable:1 unknown:2 observation:1 finite:3 situation:1 precise:3 perturbation:1 station:1 arbitrary:2 community:1 rating:1 introduced:1 pair:1 optimized:1 learned:1 established:2 zlx:2 trans:1 able:1 usually:6 ia:2 treated:1 etrain:2 regularized:1 predicting:1 numerous:1 ne:1 auto:1 dxo:1 prior:1 understanding:1 acknowledgement:1 afosr:1 cpo:1 fully:1 proxy:1 principle:2 editor:1 elsewhere:1 supported:2 free:1 bias:4 perceptron:1 world:1 ignores:1 author:2 made:2 adaptive:2 commonly:1 approximate:2 emphasize:1 assumed:2 unnecessary:1 xi:7 alternatively:1 don:1 search:4 robust:3 expansion:2 gpe:2 significance:1 motivation:1 noise:21 ff:1 wiley:1 wish:2 xl:1 pe:1 lw:1 weighting:1 ix:3 down:1 specific:1 decay:17 exists:1 alc:1 importance:1 sigmoids:2 appearance:1 chang:1 conditional:4 viewed:1 ann:1 jf:1 replace:1 mccullagh:2 included:1 specifically:1 determined:1 except:1 principal:1 called:1 total:1 perceptrons:2 internal:1 latter:1 kung:1 preparation:1 kiado:1 wcx:1 |
4,750 | 5,300 | Capturing Semantically Meaningful Word
Dependencies with an Admixture of Poisson MRFs
David I. Inouye
Pradeep Ravikumar
Inderjit S. Dhillon
Department of Computer Science
University of Texas at Austin
{dinouye,pradeepr,inderjit}@cs.utexas.edu
Abstract
We develop a fast algorithm for the Admixture of Poisson MRFs (APM) topic
model [1] and propose a novel metric to directly evaluate this model. The APM
topic model recently introduced by Inouye et al. [1] is the first topic model that
allows for word dependencies within each topic unlike in previous topic models
like LDA that assume independence between words within a topic. Research in
both the semantic coherence of a topic models [2, 3, 4, 5] and measures of model
fitness [6] provide strong support that explicitly modeling word dependencies?as
in APM?could be both semantically meaningful and essential for appropriately
modeling real text data. Though APM shows significant promise for providing
a better topic model, APM has a high computational complexity because O(p2 )
parameters must be estimated where p is the number of words ([1] could only
provide results for datasets with p = 200). In light of this, we develop a parallel alternating Newton-like algorithm for training the APM model that can handle p = 104 as an important step towards scaling to large datasets. In addition,
Inouye et al. [1] only provided tentative and inconclusive results on the utility
of APM. Thus, motivated by simple intuitions and previous evaluations of topic
models, we propose a novel evaluation metric based on human evocation scores
between word pairs (i.e. how much one word ?brings to mind? another word [7]).
We provide compelling quantitative and qualitative results on the BNC corpus
that demonstrate the superiority of APM over previous topic models for identifying semantically meaningful word dependencies. (MATLAB code available at:
http://bigdata.ices.utexas.edu/software/apm/)
1
Introduction and Related Work
In standard topic models such as LDA [8, 9], the primary representation for each topic is simply
a list of top 10 or 15 words. To understand a topic, a person must manually consider many of the
possible 10
relationships and attempt to
2 pairwise relationships as well as possibly larger m-wise
10
infer abstract meaning from this list of words. Of all the 2 pairwise relationships probably a
very small number of them are direct relationships. For example, a topic with the list of words
?money?, ?fund?, ?exchange? and ?company? can be understood as referring to investment but this
can only be inferred from a very high-level human abstraction of meaning. This problem has given
rise to research on automatically labeling topics with a topic word or phrase that summarizes the
topic [10, 11, 12]. [13] propose to evaluate topic models by randomly replacing a topic word with
a random word and evaluating whether a human can identify the intruding word. The intuition for
this metric is that the top words of a good topic will be related, and therefore, a person will be
able to easily identify the word that does not have any relationship to the other words. [2, 3, 5]
compute statistics related to Pointwise Mutual Information for all pairs of top words in a topic
and attempt to correlate this with human judgments. All of these metrics suggest that capturing
1
semantically meaningful relationships between pairs of words is fundamental to the interpretability
and usefulness of topic models as a document summarization and exploration tool.
In light of these metrics, [1] recently proposed a topic model called Admixture of Poisson MRFs
(APM) that relaxes the independence assumption for the topic distributions and explicitly models
word dependencies. This can be motivated in part by [6] who investigated whether the Multinomial (i.e. independent) assumption of word-topic distributions actually fits real-world text data.
Somewhat unsurprisingly, [6] found that the Multinomial assumption was often violated and thus
gives evidence that models with word dependencies?such as APM?may be a fundamentally more
appropriate model for text data.
Previous research in topic modeling has implicitly uncovered this issue with model misfit by finding
that models with 50, 100 or even 500 topics tend to perform better on semantic coherence experiments than smaller models with only 10 or 20 topics [4]. Though using more topics may allow topic
models to ignore the issue of word dependencies, using more topics can make the coherence of a
topic model more difficult as suggested by [4] who found that using 100 or 500 topics did not significantly improve the coherence results over 50 topics. Intuitively, a topic model with a much smaller
number of topics (e.g. 5 or 10) is easier to comprehend. For instance, if training on newspaper text,
the number of topics could roughly correspond to the number of sections in a newspaper such as
news, weather and sports. Or, if modeling an encyclopedia, the top-level topics could be art, history,
science, and society. Thus, rather than using more topics, APM opens the way for a promising topic
model that can overcome this model misfit issue while only using a small number of topics.
Even though APM shows promise for being a significantly more powerful and more realistic topic
model than previous models, the original paper acknowledged the significant computational complexity. Instead of needing to fit O(k(n + p)) parameters, APM needs to estimate O(k(n + p2 ))
parameters. [1] suggested that by using a sparsity prior (i.e. `1 regularization of the likelihood), this
computational complexity could be reduced. However, [1] could only produce some quantitative results on a very small dataset with only 200 words. In addition, the quantitative results from [1] were
tentative and inconclusive on whether APM could actually perform better than LDA in coherence
experiments.
Therefore, in this paper, we seek to answer two major open questions regarding APM: 1) Is there an
algorithm that can overcome the computational complexity of APM and handle real-world datasets?
2) Does the APM model actually capture more semantically interesting concepts that were not possible with previous topic models? We answer the first question by developing a parallel alternating
algorithm whose independent subproblems are solved using a Newton-like algorithm similar to the
algorithms developed for sparse inverse covariance estimation [14]. As in [14], this new APM
algorithm exploits the sparsity of the solution to significantly reduce the computational time for
computing the approximate Newton direction. However, unlike [14], the APM model is solving for
k Poisson MRFs simultaneously whereas [14] is only solving for a single Gaussian MRF. Another
difference from [14] is that the whole algorithm can be easily parallelized up to min(n, p).
For the second question about the semantic utility of APM, we develop a novel evaluation metric that
more directly evaluates the APM model against human judgments of semantic relatedness?a notion
called evocation introduced by [7]. Intuitively, the idea is that humans seek to understand traditional
topic models by looking at the list of top words. They will implicitly attempt to find how these
words are related and extract some more abstract meaning that generalizes the set of words. Thus,
this evaluation metric attempts to explicitly score how well pairs of words capture some semantically
meaningful word dependency. Previous research has evaluated topic models using word similarity
measures [4]. However, our work is different from [4] in three significant ways: 1) our metrics use
evocation rather than similarity (e.g. antonyms should have high evocation but low similarity), 2) we
evaluate top individual word pairs instead of rough aggregate statistics, and 3) we evaluate a topic
model that directly captures word dependencies (i.e. APM). We demonstrate that APM substantially
outperforms other topic models in both quantitative and qualitative ways.
2
Background on Admixture of Poisson MRFs (APM)
Admixtures The general notion of admixtures introduced by [1] generalizes many previous topic
models including PLSA [15], LDA [8], and the Spherical Admixture Model (SAM) [16]. Admix2
tures have also been known as mixed membership models [17]. In contrast to mixture distributions
which assume that each observation is drawn from 1 of k component distributions, admixture distributions assume that each observation is drawn from an admixed distribution whose parameters are
a mixture of component parameters. As examples of admixtures, PLSA and LDA are admixtures of
Multinomials whereas SAM is an admixture of Von-Mises Fisher distributions. In addition, because
of the connections between Poissons and Multinomials, PLSA and LDA can be seen as admixtures
of independent Poisson distributions [1].
Poisson MRFs (PMRF) Yang et al. [18] introduced a multivariate generalization of the Poisson
that assumes that the conditional distributions are univariate Poisson which is similar to a Gaussian
MRF whose conditionals are Gaussian (unlike a Guassian MRF, however, the marginals are not
univariate Poisson). A PMRF can be parameterized by a node vector ? and an edge matrix ? whose
non-zeros encode the direct dependencies between words: PrPMRF (x | ?, ?) = exp ? T x+xT ?x?
Pp
s=1 ln(xs !) ? A (?, ?) , where A (?, ?) is the log partition function needed for normalization.
This formulation needs to be slightly modified to allow for positive edges using the ideas from [19].
The log partition function can be approximated
Ppby using the pseudo log-likelihood instead of the
true likelihood, which means that A (?, ?) ? s=1 exp(?s + xT ?s ). The reader should note that
because this is an MRF distribution, all the properties of MRFs apply to PMRFs including that a
word is independent of all other words given the value of its neighbors. For example, in a chain
graph, all the variables are correlated with each other but they have a much simpler dependency
structure that can be encoded with O(n) parameters. Therefore, PMRFs more directly and succinctly
capture the dependencies between words as opposed to other simple statistics such as covariance or
pointwise mutual information.
Admixture of Poisson MRFs (APM) Inouye et al. [1] essentially constructed a new admixture model by using Poisson MRFs as the topic-word distributions instead of the usual Multinomial as in LDA. This allows for word dependencies within each topic. For example, if the word
?classification? appears in a document, ?supervised? is more likely to appear than in general documents. Given the admixture weights vector for a document the likelihood of a document
is sim
Pk
Pk
ply: PrAPM (x | w, ? 1...k , ?1...k ) = PrPMRF x | ? = j=1 wj ? j , ? = j=1 wj ?j (please see
Appendix A for notational conventions used throughout the paper). Inouye et al. [1] define a
Dirichlet(?) prior on the admixture weights and a conjugate prior with hyperparameter ? on the
PMRF parameters which can be easily incorporated as pseudo counts. For our experiments as described in Sec. 4.1, we set ? = 1 (i.e. a uniform prior on admixture weights) and ? = {0, 1}.
3
Parallel Alternating Newton-like Algorithm for APM
In the original APM paper [1], parameters were estimated by maximizing the joint approximate
posterior over all variables.1 Instead of maximizing jointly over all parameters, we split the problem
into alternating convex optimization problems. Let us denote the likelihood part (i.e. the smooth
part) of the optimization function as g(W, ? 1...k , ?1...k ) and the non-smooth `1 regularization term
as h where the full negative posterior is defined as f = g + h. The smooth part of the approximate
posterior can be written as:
g=?
n p
k
k
X
i
1 XX h X
wij xis (?sj + xTi ?js ) ? exp
wij (?sj + xTi ?js ) ,
n i=1 s=1 j=1
j=1
(1)
where xi is the word-count vector for the ith document, wi is the admixture weight vector for the
ith document, and ? j and ?j are the PMRF parameters for the jth component (see Appendix B for
derivation). By writing g in this form, it is straightforward to see that even though the whole optimization problem is not convex because of the interaction between the admixture weights w and the
PMRF parameters, the problem is convex if either the admixture weights W or the component parameters ? 1...k , ?1...k are held fixed. To simplify the notation in the following sections, we combine
1
This posterior approximation was based on the pseudo-likelihood while ignoring the symmetry constraint
so that nodewise regression parameters are independent. This leads to an overcomplete parameterization for
APM. For an overview of composite likelihood methods, see [20]. For a comparison of pseudo-likelihood
versus nodewise regressions, see [21].
3
the node (which is analogous to an intercept term in regression) and edge parameters by defining
zi = [1 xTi ]T , ?js = [?sj (?js )T ]T and ?s = [?1s ?2s ? ? ? ?ks ].
Thus, we can alternate between optimizing two similar optimization problems where one has a nonsmooth `1 regularization and the other has the constraint that wi must lie on the simplex ?k :
arg min
?1 ,?2 ,??? ,?p
arg min
w1 ,w2 ,??? ,wn ??k
p
p
n
i X
X
1 Xh
s s
T s
?
tr(? ? ) ?
exp(zi ? wi ) +
?kvec(?s )\1 k1
n s=1
s=1
i=1
?
p
n
i
X
1 Xh T
?i wi ?
exp(ziT ?s wi )) ,
n i=1
s=1
(2)
(3)
where ?i and ?s are constants in the optimization that can be computed from the data matrix X and
the other parameters that are being held fixed (see Alg. 2 in Appendix D for computation of ?s ).
This alternating scheme is analogous to Alternating Least Squares (ALS) for Non-negative Matrix
Factorization (NMF) [22] and EM-like algorithms such as k-means. By writing the optimization as
in Eq. 2 and Eq. 3, we also expose the simple independence between the subproblems because they
are simple summations. Thus, we can easily parallelize both optimization problems upto min(n, p)
with little overhead and simple changes to the code?in our MATLAB implementation, we only
changed a for loop to a parfor loop.
3.1
Newton-like Algorithms for Subproblems
For each of the subproblems, we develop Newton-like optimization algorithms. For the component
PMRFs, we borrow several important ideas from [14] including fixed and free sets of variables for
the `1 regularized optimization problem. The overall idea is to construct a quadratic approximation
around the current solution and approximately optimize this simpler function to find a step direction.
Usually, finding the Newton direction requires computing the Hessian for all the optimization variables but because of the `1 regularization, we only need to focus on variables that might be non-zero.
This set of free variables, denoted F, can be simply determined from the gradient and current iterate
[14]. Since usually there is only a small number of free variables compared to fixed variables (i.e. ?
is large enough), we can simply run coordinate descent on these free variables and only implicitly
calculate Hessian information as needed in each coordinate descent step. After finding an approximate Newton direction, we find a step size that satisfies the Armijo rule and then update the iterate
(see Alg. 2 in Appendix D).
We also employed a similar Newton-like algorithm for estimating the admixture weights. Instead of
the `1 regularization term, however, this subproblem has the constraint that the admixture weights
wi must lie on the simplex so that each document can be properly interpreted as a convex mixture
of over topic parameters. For this constraint, we used a dual-coordinate descent algorithm to find
the approximate Newton direction as in [23].
Finally, we put both subproblem algorithms together and alternate between the two (see Alg. 1 in
Appendix D). For tracing through different ? parameters, ? is initially set to ? so that the model
trains an independent APM model first. Then, the initial ? = ?max is found by computing the largest
gradient of the final independent iteration. Every time the alternating algorithm converges, the value
of ? is decreased so that a set of models is trained for decreasing values of ?.
3.2
Timing Results
We conducted two main timing experiments to show that the algorithm can be efficiently parallelized
and the algorithm can scale to reasonably large datasets. For the parallel timing experiment, we used
the BNC corpus described in Sec. 4.1 (n = 4049, p = 1646) and fixed k = 5, ? = 8 and a total of
30 alternating iterations. For the large data experiment, we used a Wikipedia dataset formed from a
recent Wikipedia dump by choosing the top 10k words neglecting stop words and then selecting the
longest documents. We ran several main iterations of the algorithm with this dataset while fixing
the parameters k = 5 and ? = 0.5. All timing experiments were conducted on the TACC Maverick
system with Intel Xeon E5-2680 v2 Ivy Bridge CPUs (2.80 GHz), 20 CPUs per node, and 12.8 GB
memory per CPU (https://www.tacc.utexas.edu/).
4
The parallel timing results can be seen in Fig. 1 (left) which shows that the algorithm does have
almost linear speedup when parallelizing across multiple workers. Though we only had access to a
single computer with 20 processors, substantially more speed up could be obtained by using more
processors on a distributed computing system. This simple parallelism makes this algorithm viable
for much larger datasets. The timing results for the Wikipedia can be seen in Fig. 1 (right). These
results give an approximate computational complexity of O(np2 ) which show that the proposed
algorithm has the potential for scaling to datasets where n is O(105 ) and p is O(104 ). The O(p2 )
comes from the fact that there are p subproblems and each subproblem needs to calculate the gradient
which is O(p) as well as approximate the Newton direction for a subset of the variables. The
first iteration takes longer because the initial parameter values are na??vely set to 0 whereas future
iterations start from reasonable initial value.
APM Training Time on Wikipedia Dataset
Parallel Speedup on BNC Dataset
4
Perfect Speedup
15
1st Iter.
Time (hrs)
Speedup
20
Actual Speedup
10
5
2.2
5
10
15
# of MATLAB Workers
20
2.2
1
1
0
3.4
3.1
2
0
0
Avg. Next 3 Iter.
3
0.6
n = 20,000
p = 5,000
# of Words = 50M
n = 100,000
p = 5,000
# of Words = 133M
n = 20,000
p = 10,000
# of Words = 57M
Figure 1: (left) The speedup on the BNC dataset shows that the algorithm scales approximately linearly with the number of workers because the subproblems are all independent. (right) The timing
results on the Wikipedia dataset show that the algorithm scales to larger datasets and has a computational complexity of approximately O(np2 ).
4
Evocation Metric
Boyd-Graber et al. [7] introduced the notion of evocation which denotes the idea of which words
?evoke? or ?bring to mind? other words. There can be many types of evocation including the following examples from [7]: [rose - flower] (example), [brave - noble] (kind), [yell - talk] (manner),
[eggs - bacon] (co-occurence), [snore - sleep] (setting), [wet - desert] (antonymy), [work - lazy] (exclusivity), and [banana - kiwi] (likeness). This is distinctive from word similarity or synonymy since
two words can have very different meanings but ?bring to mind? the other word (e.g. antonyms).
This notion of word relatedness is a much simpler but potentially more semantically meaningful and
interpretable than word similarity. For instance, ?work? and ?lazy? do not have similar meanings
but are related through the semantic meanings of the words. Another difference is that?unlike word
semantic similarity? words that generally appear in very different contexts yet mean the same thing
would probably not have a high evocation score. For example, ?networks? and ?graphs? both have a
definition that means a set of nodes and edges yet usually one word is chosen in a particular context.
Recent work in evaluating topic models [2, 3, 4, 5] formulate automated metrics based on automatically scoring all pairs of top words and noticing that they correlate with human judgment of overall
topic coherence. All of these metrics are based on the common assumption that a person should
be able to understand a topic by understanding the abstract semantic connections between the word
pairs. Thus, evocation is a reasonable notion for evaluating topic modeling because it directly evaluates the level of semantic connection between word pairs. In addition, this new evocation metric
provides a way to explicitly evaluate the edge matrices of APM, which would be ignored in previous
metrics because explicit word dependencies are not modeled in other topic models.
We now formally define our evocation metric. Given human-evaluated scores for a subset of word
pairs H and the corresponding weights given by a topic model for this subset of word pairs M, let
us define ?M (j) to be an ordering of the word pairs induced by M such that M?(1) ? M?(2) ?
? ? ? ? M?(|H|) . Then, the top-m evocation metric is simply:
m
X
Evocm (M, H) =
H?M(j) .
(4)
j=1
Note that the scaling of M is inconsequential because M is only needed to define an ordering or
? = ? exp(M) would yield the same evocation score for all scalar
ranking of H. For example, M
5
values ? > 0 because the ordering would be maintained. Essentially, M merely induces an ordering
of the word pairs and the evocation score is the sum of the human scores for these top m word pairs.
For APM, the word pair weights come primarily from the PMRF edge matrices ?1...k ?the PMRF
node vectors are only used to provide an ordering if there are not enough non-zeros in the edge
matrices. For the other Multinomial-based topic models which do not have parameters explicitly
associated with word-pairs, we can compute the most likely word pairs in a topic by multiplying
their corresponding marginal probabilities. This weighting corresponds to the probability that two
independent draws from the topic distribution produce the word pair and thus is the most obvious
choice for Multinomial-based topic models.
Since this metric only gives a way to evaluate one topic, we consider two ways of determining
Pk 1
j
the overall evocation score for the whole topic model: Evoc-1 =
j=1 k Evocm (M , H) and
Pk 1 j
Evoc-2 = Evocm ( j=1 k M , H). In words, these are ?average evocation of topics? and ?evocation of average topic? respectively. Evoc-1 measures whether all or at least most topics capture
meaningful word associations since it can be affected by uninteresting topics. Evoc-2 is reasonable
for measuring whether the topic model as a whole is capturing word semantics even if some of the
topics are not capturing interesting word associations. This second measure has some relation to
the word similarity measure of topic coherence in [4]. However, [4] uses similarity rather than evocation, does not directly evaluate top individual word pairs and does not evaluate any models with
word dependencies such as APM.
4.1
Experimental Setup
Human-Scored Evocation Dataset The original human-scored evocation dataset was produced
by a set of trained undergraduates in which 1,000 words were hand selected primarily based on
their frequency and usage in the British National Corpus (BNC) [7]. From the possible pairwise
evaluations, approximately 10% of the word pairs were randomly selected to be manually scored by
a set of trained undergraduates. The second dataset was constructed by predicting the pairs of words
that were likely to have a high evocation using a standard machine learning classifier. This new set of
pairs was scored using Amazon MTurk (mturk.com) by using the original dataset as a control [24].
Though these scores are between synsets?which are a word, part-of-speech and sense triplet?, we
mapped all of the synsets to word, part-of-speech pairs since that is the only information we have
for the BNC corpus. This led to a total of 1646 words. In addition, though the evocation dataset has
scores for directed relationships (i.e. word1 ? word2 could have a different score than word2 ?
word1), we averaged these two scores because the directionality of the relationship is not modeled
by APM or any other topic model.
BNC Corpus Because the evocation dataset was based on the BNC corpus, we used the BNC corpus for our experiments. We processed the BNC corpus by lemmatizing each word using the WordNetLemmatizer included in the nltk package (nltk.org) and then attaching the part-of-speech,
which is already included in the BNC corpus. We only retained the counts for the 1646 words that
occurred in the human-scored datasets but processed all 4049 documents in the corpus.
APM Model Parameters We trained APM on the BNC corpus with several different parameter
settings including various ? and ? parameter settings. We also trained two particular APM models
denoted APM-LowReg and APM-HeldOut. APM-LowReg uses a very small regularization parameter so that almost all edges are non-zero. APM-HeldOut automatically selects a reasonable value
for ? based on the likelihood of a held-out set of the documents. Thus, the APM-HeldOut model
does not require a user-specified ? parameter but?as seen in the following sections?still performs
reasonably well even compared to the APM model in which many different parameter settings are
attempted. In addition, the APM-HeldOut can stop the training early when the model begins to overfit the data rather than tracing through all the ? parameters?this could lead to a significant gain in
model training time. The authors suggest that APM-HeldOut is a simple baseline model for future
comparison if a user does not want to specify ?.
Other Models For comparison, we trained five other models: Correlated Topic Models (CTM),
Hierarchical Dirichlet Process (HDP), Latent Dirichlet Allocation (LDA), Replicated Softmax
(RSM), and a na??ve random baseline (RND). CTM models correlations between topics [25]. HDP
6
is a non-parametric Bayesian model that selects the number of topics based input data and hyperparameters [26]. The standard topic model LDA was trained using MALLET [27]. LDA was trained
for at least 5,000 iterations and HDP was trained for at least 300 iterations since HDP is computationally expensive. RSM is an undirected topic model based on Restricted Boltzmann Machines
(RBM) [28]. The random model is merely the expected evocation score if edges are ranked at random. We ran a full factorial experimental setting where all the combinations of a set of parameter
values were trained to give a fair comparison between models (see Appendix C for a summary of
parameter values). All these comparison models only indirectly model dependencies between words
through the latent variables since the topic distributions are Multinomials whereas APM can directly
model the dependencies between words since the topic distributions are Poisson MRFs.
Selecting Best Parameters We randomly split the human scores into a 50% tuning split and 50%
testing split. Note that we have a tuning split rather than a training split because the model training
algorithms are unsupervised (i.e. they never see the human scores) so the only supervision occurs in
selecting the final model parameters (i.e. during the tuning phase). Therefore, we selected the final
parameters based on the tuning split and computed the final evocation scores on the test split. Thus,
even when selecting the parameter settings, the modeling process never sees the test data.
4.2
Main Results
The Evoc-1 and Evoc-2 scores with m = 50 for all models can be seen in Fig. 2.2 For Evoc-1, the
APM models significantly outperform all other models for a small number of topics and even capture
many semantically meaningful word pairs with a single topic. For higher number of topics, the APM
models seem to perform only competitively with previous topic models. It seems that APM-LowReg
performs better with a small number of topics whereas APM-HeldOut?which generally chooses a
relatively high ??seems more robust for large number of topics. These trends are likely caused
because there is a relatively small number of documents (n = 4049) so the APM-LowReg begins to
significantly overfit the data as the number of topics increases whereas APM-HeldOut does not seem
to overfit as much. For all the APM models, the degradation in performance as the number of topics
increases is most likely caused by the fact that a Poisson MRF with O(p2 ) parameters is a much
more flexible distribution than a Multinomial, and thus, fewer topics are needed to appropriately
model the data. These results also give some evidence that APM can succinctly model the data with
a much smaller number of topics than is needed for independent topic models; this succinctness
could be particularly helpful for the interpretability and intuitions of topic models.
Evocation (m= 50)
APM
APM-LowReg
APM-HeldOut
CTM
HDP
LDA
RSM
RND
1600
1400
1200
1000
800
600
400
200
0
k=1
3
5
10
25
50
Evoc-1 (Avg. Evoc. of Topics)
k=1
3
5
10
25
50
Evoc-2 (Evoc. of Avg. Topic)
Figure 2: Both Evoc-1 scores (left) and Evoc-2 scores (right) demonstrate that APM usually significantly outperforms other topic models in capturing meaningful word pairs.
For the Evoc-2 score, the APM models?including the APM-HeldOut model which automatically
determines a ? from the data?significantly outperform previous topic models even for a large number of topics. This supports the idea that APM only needs a small number of topics to capture many
of the semantically meaningful word dependencies. Thus, when increasing the number of topics
beyond 5, the performance does not decrease as in Evoc-1. It is likely that this discrepancy is caused
by the fact that many of the edges are concentrated in a small number of topics even when the number of topics is 10 or 25. As expected because of previous research in topic models, most other topic
2
For simplicity and comparability, we grouped HDP into the topic number that was closest to its discovered
number of topics because HDP can select a variable number of topics.
7
models perform slightly better with a larger number of topics. Though it is possible that using 100
or 500 topics for these topic models might give an evocation score better than APM with 5 topics,
this would only enforce the idea that APM can perform better or at least competitively with previous
topic models while only using a comparatively small number of topics.
Qualitative Analysis of Top 20 Word Pairs for Best LDA and APM Models To validate the
intuition of using evocation as an human-standard evaluation metric, we present the top 20 word
pairs for the best standard topic model?in this case LDA?and the best APM model for the Evoc-2
metric as seen in Table 1. The best performing LDA model was trained with 50 topics, ? = 1 and
? = 0.0001. The best APM model was the APM-LowReg model trained with only 5 topics and a
small regularization parameter ? = 0.05. It is important to note that the best model for LDA has
50 topics while the best model for APM only has 5 topics. As before, this reinforces the theme that
APM can capture more semantically meaningful word pairs with a smaller number of topics than
previous topic models.
1:=Top
APM
(right)
LDALDA
Evocation
Evocation
ofTable
Avg.
of Avg.
Graph
Graph
967
= 96720 words for LDA (left) and
APM
APM
Evocation
Evocation
of Avg.
of Avg.
Graph
Graph
= 1627
= 1627
RankRank
Evoc.
Evoc.
1
2
3
4
5
6
7
8
9
10
1 38
2 0
3 13
4 69
5 0
6 82
7 38
8 35
9 7
10 38
EdgeEdge
RankRank
Evoc.
Evoc.
38 woman.n
woman.n
??
man.n
man.n 11
0 woman.n
woman.n
??
wife.n
wife.n
12
13
train.n
train.n
??
car.n
car.n
13
69 school.n
school.n
??
class.n
class.n 14
0
drive.v
drive.v
??
car.n
car.n
15
82 teach.v
teach.v
??
school.n
school.n 16
38 engine.n
engine.n
??
car.n
car.n
17
35 publish.v
publish.v
??
book.n
book.n 18
7 religious.a
religious.a
??
church.n
church.n 19
38 state.n
state.n
??
government.n
government.n
20
EdgeEdge
RankRank
Evoc.
Evoc.
11 0 0
car.n
car.n
??
bus.n
bus.n
12 31 31
year.n
year.n
??
day.n
day.n
13 25 25
car.n
car.n
??
seat.n
seat.n
14 50 50 teach.v
teach.v
??
student.n
student.n
15 0 0
tell.v
tell.v
??
get.v
get.v
16 38 38
wife.n
wife.n
??
man.n
man.n
17100 100
run.v
run.v
??
car.n
car.n
18 0 0
give.v
give.v
??
get.v
get.v
19 16 16 paper.n
paper.n
??
book.n
book.n
20 19 19 white.a
white.a
??
black.a
black.a
1
2
3
4
5
6
7
8
9
10
1 13
2 60
3 13
4 50
5 38
6 75
7 57
8 13
9 7
10 97
EdgeEdge
RankRank
Evoc.
Evoc.
13 smoke.v
smoke.v
??
cigarette.n
cigarette.n11
60
love.v
love.v
??
love.n
love.n
12
13
eat.v
eat.v
??
food.n
food.n 13
50 west.n
west.n
??
east.n
east.n
14
38 south.n
south.n
??
north.n
north.n 15
75
iron.n
iron.n
??
steel.n
steel.n 16
57question.n
question.n
??
answer.n
answer.n 17
13
boil.v
boil.v
??
potato.n
potato.n 18
7 religious.a
religious.a
??
church.n
church.n 19
97husband.n
husband.n
??
wife.n
wife.n
20
11 72
12 28
13 25
14 0
15 35
16 0
17 19
18 41
19 33
20 7
EdgeEdge
72
aunt.n
aunt.n
??
uncle.n
uncle.n
28
tea.n
tea.n
??
coffee.n
coffee.n
25operational.a
operational.a
??
aircraft.n
aircraft.n
0competition.n
competition.n
??
compete.v
compete.v
35 green.n
green.n
??
green.a
green.a
0
fox.n
fox.n
??
animal.n
animal.n
19 smoke.n
smoke.n
??
fire.n
fire.n
41 wine.n
wine.n
??
drink.v
drink.v
33 troop.n
troop.n
??
force.n
force.n
7
lock.n
lock.n
??
key.n
key.n
One interesting example is that LDA finds two word pairs [woman.n - wife.n] and [wife.n - man.n]
that capture some semantic notion of marriage. However, APM directly captures this semantic
meaning with [husband.n - wife.n]. APM also finds more intuitive verb-noun relationships that are
closely tied semantically and portray a particular context: [smoke.v - cigarette.n], [eat.v - food.n],
[boil.v - potato.n], and [drink.v - wine.n] whereas LDA tends to select less interesting verb-noun
relationships such as [run.v - car.n]. In addition, APM finds multiple semantically coherent yet high
level word pairs such as [iron.n - steel.n], [question.n - answer.n], and [aunt.n - uncle.n], whereas
LDA finds several low-level edges such as [year.n - day.n] and [tell.v - get.v]. These overall trends
become even more evident when looking at the top 50 edges as can be found in the Appendix E.
Both the quantitative evaluation metrics (i.e. Evoc-1 and Evoc-2) as well as a qualitative exploration
of the top word pairs give strong evidence that APM can succinctly capture both more interesting
and higher-level semantic concepts through word dependencies than independent topic models.
5
Conclusion and Future Work
We motivated the need for more expressive topic models that consider word dependencies?such
as APM?by considering previous work on topic model evaluation metrics. We overcame the significant computational barrier of APM by providing a fast alternating Newton-like algorithm which
can be easily parallelized. We proposed a new evaluation metric based on human evocation scores
that seeks to measure whether a model is capturing semantically meaningful word pairs. Finally,
we presented compelling quantitative and qualitative measures showing the superiority of APM in
capturing semantically meaningful word pairs. In addition, this metric suggests new evaluations
of topic models based on evaluating top word pairs rather than top words. One drawback with the
current human-scored data is that only a small portion of the word pairs have been scored. Thus,
one extension is to dynamically collect more human scores as needed for evaluation. This work also
opens the door for exciting new word-semantic applications for APM such as Word Sense Induction
using topic models [29], keyword expansion or suggestion, document summarization, and document
visualization because APM is capturing semantically meaningful relationships between words.
Acknowledgments
D. Inouye was supported by the NSF Graduate Research Fellowship via DGE-1110007. P. Ravikumar acknowledges support from ARO via W911NF-12-1-0390 and NSF via IIS-1149803, IIS1447574, and DMS-1264033. I. Dhillon acknowledges support from NSF via CCF-1117055.
8
References
[1] D. I. Inouye, P. Ravikumar, and I. S. Dhillon, ?Admixture of Poisson MRFs: A Topic Model with Word
Dependencies,? in International Conference on Machine Learning (ICML), 2014.
[2] D. Mimno, H. M. Wallach, E. Talley, M. Leenders, and A. McCallum, ?Optimizing semantic coherence
in topic models,? in EMNLP, pp. 262?272, 2011.
[3] D. Newman, Y. Noh, E. Talley, S. Karimi, and T. Baldwin, ?Evaluating topic models for digital libraries,?
in ACM/IEEE Joint Conference on Digital Libraries (JCDL), pp. 215?224, 2010.
[4] K. Stevens and P. Kegelmeyer, ?Exploring topic coherence over many models and many topics,? in
EMNLP-CoNLL, pp. 952?961, 2012.
[5] N. Aletras and R. Court, ?Evaluating Topic Coherence Using Distributional Semantics,? in International
Conference on Computational Semantics (IWCS 2013) - Long Papers, pp. 13?22, 2013.
[6] D. Mimno and D. Blei, ?Bayesian Checking for Topic Models,? in EMNLP, pp. 227?237, 2011.
[7] J. Boyd-graber, C. Fellbaum, D. Osherson, and R. Schapire, ?Adding Dense, Weighted Connections to
{WordNet},? in Proceedings of the Global {WordNet} Conference, 2006.
[8] D. Blei, A. Ng, and M. Jordan, ?Latent dirichlet allocation,? JMLR, vol. 3, pp. 993?1022, 2003.
[9] T. L. Griffiths and M. Steyvers, ?Finding scientific topics,? Proceedings of the National Academy of
Sciences of the United States of America, vol. 101, pp. 5228?35, Apr. 2004.
[10] J. H. Lau, K. Grieser, D. Newman, and T. Baldwin, ?Automatic Labelling of Topic Models,? in NAACL
HLT, pp. 1536?1545, 2011.
[11] D. Magatti, S. Calegari, D. Ciucci, and F. Stella, ?Automatic Labeling Of Topics,? in ISDA, 2009.
[12] X.-l. Mao, Z.-y. Ming, Z.-j. Zha, T.-s. Chua, H. Yan, and X. Li, ?Automatic Labeling Hierarchical Topics,?
in CIKM, pp. 2383?2386, 2012.
[13] J. Chang, J. Boyd-Graber, S. Gerrish, C. Wang, and D. Blei, ?Reading tea leaves: How humans interpret
topic models,? NIPS, pp. 1?9, 2009.
[14] C.-J. Hsieh, M. A. Sustik, I. S. Dhillon, and P. Ravkiumar, ?Sparse inverse covariance matrix estimation
using quadratic approximation,? NIPS, pp. 1?18, 2011.
[15] T. Hofmann, ?Probabilistic latent semantic analysis,? in Uncertainty in Artificial Intelligence (UAI),
pp. 289?296, Morgan Kaufmann Publishers Inc., 1999.
[16] J. Reisinger, A. Waters, B. Silverthorn, and R. J. Mooney, ?Spherical topic models,? in ICML, pp. 903?
910, 2010.
[17] E. M. Airoldi, D. M. Blei, S. E. Fienberg, and E. P. Xing, ?Mixed Membership Stochastic Blockmodels.,?
JMLR, vol. 9, pp. 1981?2014, Sept. 2008.
[18] E. Yang, P. Ravikumar, G. I. Allen, and Z. Liu, ?Graphical models via generalized linear models,? in
NIPS, pp. 1367?1375, 2012.
[19] E. Yang, P. Ravikumar, G. Allen, and Z. Liu., ?On poisson graphical models,? in NIPS, pp. 1718?1726,
2013.
[20] C. Varin, N. Reid, and D. Firth, ?An overview of composite likelihood methods,? STATISTICA SINICA,
vol. 21, pp. 5?42, 2011.
[21] J. D. Lee and T. J. Hastie, ?Structure Learning of Mixed Graphical Models,? in AISTATS, vol. 31, pp. 388?
396, 2013.
[22] D. D. Lee and H. S. Seung, ?Algorithms for Non-negative Matrix Factorization,? in NIPS, pp. 556?562,
2000.
[23] H.-F. Yu, F.-L. Huang, and C.-J. Lin, ?Dual coordinate descent methods for logistic regression and maximum entropy models,? Machine Learning, vol. 85, pp. 41?75, Nov. 2010.
[24] S. Nikolova, J. Boyd-graber, C. Fellbaum, and P. Cook, ?Better Vocabularies for Assistive Communication
Aids: Connecting Terms using Semantic Networks and Untrained Annotators,? in ACM Conference on
Computers and Accessibility, pp. 171?178, 2009.
[25] D. M. Blei and J. D. Lafferty, ?Correlated topic models,? in NIPS, pp. 147?154, 2005.
[26] Y. W. Teh, M. I. Jordan, M. J. Beal, and D. M. Blei, ?Hierarchical Dirichlet Processes,? Journal of the
American Statistical Association, vol. 101, pp. 1566?1581, Dec. 2006.
[27] A. K. McCallum, ?MALLET: A Machine Learning for Language Toolkit,? 2002.
[28] G. Hinton and R. Salakhutdinov, ?Replicated softmax: An undirected topic model,? NIPS, 2009.
[29] J. H. Lau, P. Cook, D. Mccarthy, D. Newman, and T. Baldwin, ?Word Sense Induction for Novel Sense
Detection,? in EACL, pp. 591?601, 2012.
9
| 5300 |@word aircraft:2 seems:2 plsa:3 open:3 seek:3 covariance:3 hsieh:1 tr:1 initial:3 liu:2 uncovered:1 score:23 selecting:4 united:1 document:14 outperforms:2 current:3 com:1 yet:3 must:4 written:1 realistic:1 partition:2 hofmann:1 interpretable:1 fund:1 update:1 intelligence:1 selected:3 fewer:1 leaf:1 parameterization:1 cook:2 mccallum:2 ith:2 chua:1 blei:6 provides:1 node:5 org:1 simpler:3 five:1 constructed:2 direct:2 become:1 viable:1 qualitative:5 combine:1 overhead:1 manner:1 pairwise:3 expected:2 roughly:1 love:4 salakhutdinov:1 ming:1 spherical:2 decreasing:1 company:1 automatically:4 xti:3 little:1 cpu:3 actual:1 increasing:1 considering:1 provided:1 intruding:1 xx:1 notation:1 estimating:1 begin:2 kind:1 interpreted:1 substantially:2 developed:1 finding:4 pseudo:4 quantitative:6 every:1 classifier:1 control:1 superiority:2 appear:2 kegelmeyer:1 reid:1 ice:1 positive:1 understood:1 timing:7 before:1 tends:1 parallelize:1 approximately:4 inconsequential:1 might:2 black:2 k:1 dynamically:1 suggests:1 collect:1 wallach:1 co:1 factorization:2 graduate:1 averaged:1 directed:1 acknowledgment:1 testing:1 investment:1 yan:1 significantly:7 weather:1 composite:2 boyd:4 word:110 synonymy:1 griffith:1 suggest:2 get:5 put:1 context:3 writing:2 intercept:1 optimize:1 www:1 troop:2 maximizing:2 straightforward:1 convex:4 formulate:1 amazon:1 identifying:1 simplicity:1 rule:1 seat:2 borrow:1 steyvers:1 handle:2 notion:6 coordinate:4 poissons:1 analogous:2 user:2 us:2 grieser:1 trend:2 approximated:1 expensive:1 particularly:1 distributional:1 exclusivity:1 baldwin:3 subproblem:3 solved:1 capture:11 wang:1 calculate:2 wj:2 pradeepr:1 news:1 keyword:1 ordering:5 decrease:1 ran:2 rose:1 intuition:4 leenders:1 complexity:6 seung:1 trained:12 solving:2 eacl:1 distinctive:1 easily:5 joint:2 kiwi:1 osherson:1 various:1 america:1 talk:1 assistive:1 derivation:1 train:3 fast:2 guassian:1 artificial:1 labeling:3 aggregate:1 tell:3 choosing:1 newman:3 varin:1 whose:4 encoded:1 larger:4 statistic:3 jointly:1 final:4 beal:1 jcdl:1 propose:3 aro:1 interaction:1 loop:2 ivy:1 academy:1 intuitive:1 validate:1 competition:2 produce:2 perfect:1 converges:1 develop:4 fixing:1 school:4 eq:2 strong:2 zit:1 sim:1 c:1 p2:4 come:2 convention:1 direction:6 closely:1 drawback:1 stevens:1 stochastic:1 exploration:2 human:19 brave:1 exchange:1 require:1 government:2 generalization:1 summation:1 extension:1 exploring:1 around:1 prpmrf:2 marriage:1 exp:6 major:1 early:1 wine:3 ctm:3 estimation:2 wet:1 expose:1 utexas:3 bridge:1 largest:1 grouped:1 tool:1 weighted:1 rough:1 gaussian:3 modified:1 rather:6 apm:82 encode:1 np2:2 focus:1 notational:1 properly:1 longest:1 likelihood:10 contrast:1 baseline:2 sense:4 helpful:1 mrfs:11 abstraction:1 membership:2 initially:1 relation:1 wij:2 selects:2 semantics:3 karimi:1 arg:2 issue:3 flexible:1 classification:1 denoted:2 dual:2 overall:4 noh:1 animal:2 noun:2 art:1 softmax:2 mutual:2 marginal:1 construct:1 never:2 religious:4 ng:1 manually:2 yu:1 unsupervised:1 icml:2 noble:1 future:3 simplex:2 nonsmooth:1 word1:2 fundamentally:1 simplify:1 primarily:2 word2:2 discrepancy:1 randomly:3 simultaneously:1 national:2 ve:1 individual:2 fitness:1 phase:1 fire:2 attempt:4 detection:1 evaluation:11 mixture:3 pradeep:1 light:2 held:3 chain:1 edge:12 potato:3 neglecting:1 worker:3 vely:1 fox:2 overcomplete:1 instance:2 xeon:1 modeling:6 compelling:2 w911nf:1 measuring:1 phrase:1 subset:3 uniform:1 usefulness:1 uninteresting:1 conducted:2 dependency:21 answer:5 chooses:1 referring:1 person:3 st:1 fundamental:1 international:2 probabilistic:1 lee:2 together:1 connecting:1 na:2 w1:1 von:1 opposed:1 huang:1 possibly:1 woman:5 emnlp:3 book:4 american:1 li:1 potential:1 sec:2 student:2 north:2 inc:1 explicitly:5 ranking:1 caused:3 portion:1 start:1 zha:1 xing:1 parallel:6 square:1 formed:1 kaufmann:1 who:2 efficiently:1 correspond:1 identify:2 judgment:3 yield:1 reisinger:1 misfit:2 bayesian:2 produced:1 multiplying:1 drive:2 mooney:1 processor:2 history:1 husband:3 hlt:1 definition:1 evaluates:2 against:1 pp:25 frequency:1 obvious:1 dm:1 associated:1 mi:1 rbm:1 boil:3 stop:2 gain:1 dataset:13 car:13 iron:3 actually:3 fellbaum:2 appears:1 higher:2 supervised:1 day:3 specify:1 formulation:1 evaluated:2 though:8 correlation:1 overfit:3 hand:1 replacing:1 expressive:1 smoke:5 logistic:1 brings:1 lda:19 scientific:1 dge:1 cigarette:3 usage:1 naacl:1 concept:2 true:1 succinctness:1 ccf:1 regularization:7 alternating:9 dhillon:4 semantic:15 white:2 comprehend:1 during:1 aunt:3 please:1 maintained:1 generalized:1 mallet:2 evident:1 demonstrate:3 performs:2 allen:2 bring:2 rsm:3 meaning:7 wise:1 likeness:1 novel:4 recently:2 wikipedia:5 common:1 multinomial:9 overview:2 food:3 association:3 occurred:1 marginals:1 interpret:1 significant:5 tuning:4 automatic:3 maverick:1 language:1 had:1 toolkit:1 access:1 similarity:8 money:1 longer:1 supervision:1 j:4 multivariate:1 posterior:4 recent:2 closest:1 mccarthy:1 optimizing:2 scoring:1 seen:6 morgan:1 somewhat:1 employed:1 parallelized:3 ii:1 full:2 multiple:2 needing:1 infer:1 smooth:3 long:1 lin:1 ravikumar:5 n11:1 mrf:5 regression:4 essentially:2 metric:22 poisson:16 mturk:2 publish:2 iteration:7 normalization:1 dec:1 addition:8 whereas:8 background:1 conditionals:1 decreased:1 want:1 fellowship:1 publisher:1 appropriately:2 w2:1 unlike:4 probably:2 south:2 induced:1 tend:1 undirected:2 thing:1 lafferty:1 seem:2 jordan:2 yang:3 door:1 silverthorn:1 split:8 enough:2 relaxes:1 wn:1 iterate:2 independence:3 fit:2 zi:2 automated:1 hastie:1 nikolova:1 reduce:1 regarding:1 idea:7 court:1 texas:1 whether:6 motivated:3 utility:2 gb:1 iwcs:1 speech:3 hessian:2 matlab:3 ignored:1 generally:2 factorial:1 encyclopedia:1 induces:1 processed:2 concentrated:1 reduced:1 http:2 tacc:2 outperform:2 schapire:1 nsf:3 estimated:2 cikm:1 per:2 reinforces:1 nodewise:2 hyperparameter:1 promise:2 tea:3 affected:1 vol:7 iter:2 key:2 acknowledged:1 drawn:2 graph:6 merely:2 year:3 sum:1 wife:9 run:4 inverse:2 parameterized:1 uncertainty:1 powerful:1 noticing:1 package:1 compete:2 throughout:1 reader:1 almost:2 reasonable:4 draw:1 coherence:10 summarizes:1 scaling:3 appendix:7 conll:1 capturing:8 drink:3 overcame:1 quadratic:2 sleep:1 constraint:4 software:1 speed:1 min:4 performing:1 eat:3 relatively:2 speedup:6 department:1 developing:1 alternate:2 combination:1 conjugate:1 smaller:4 slightly:2 em:1 sam:2 across:1 wi:6 intuitively:2 restricted:1 lau:2 fienberg:1 ln:1 computationally:1 visualization:1 bus:2 count:3 needed:6 mind:3 sustik:1 available:1 generalizes:2 competitively:2 apply:1 hierarchical:3 v2:1 appropriate:1 upto:1 indirectly:1 enforce:1 original:4 top:18 assumes:1 dirichlet:5 denotes:1 graphical:3 lock:2 newton:12 exploit:1 k1:1 coffee:2 society:1 comparatively:1 question:6 already:1 occurs:1 parametric:1 primary:1 usual:1 traditional:1 gradient:3 comparability:1 mapped:1 accessibility:1 topic:137 water:1 induction:2 hdp:7 code:2 pointwise:2 relationship:11 modeled:2 providing:2 retained:1 difficult:1 setup:1 sinica:1 potentially:1 subproblems:6 teach:4 negative:3 rise:1 steel:3 implementation:1 summarization:2 boltzmann:1 perform:5 teh:1 observation:2 datasets:8 descent:4 defining:1 hinton:1 looking:2 incorporated:1 banana:1 communication:1 discovered:1 verb:2 parallelizing:1 nmf:1 inferred:1 david:1 introduced:5 pair:34 specified:1 connection:4 tentative:2 engine:2 coherent:1 nip:7 able:2 suggested:2 beyond:1 usually:4 parallelism:1 flower:1 sparsity:2 reading:1 interpretability:2 including:6 max:1 memory:1 green:4 ranked:1 force:2 regularized:1 predicting:1 hr:1 bacon:1 scheme:1 improve:1 firth:1 library:2 admixture:23 church:4 acknowledges:2 portray:1 stella:1 inouye:7 extract:1 occurence:1 sept:1 text:4 prior:4 understanding:1 checking:1 determining:1 unsurprisingly:1 heldout:9 mixed:3 interesting:5 suggestion:1 tures:1 allocation:2 versus:1 annotator:1 digital:2 exciting:1 austin:1 succinctly:3 bnc:12 changed:1 summary:1 supported:1 free:4 jth:1 synset:2 allow:2 understand:3 neighbor:1 barrier:1 attaching:1 sparse:2 tracing:2 ghz:1 distributed:1 overcome:2 mimno:2 vocabulary:1 evaluating:6 world:2 kvec:1 author:1 avg:7 replicated:2 correlate:2 newspaper:2 sj:3 approximate:7 nov:1 ignore:1 implicitly:3 relatedness:2 evoke:1 global:1 uai:1 corpus:11 xi:2 latent:4 triplet:1 table:1 promising:1 reasonably:2 robust:1 ignoring:1 symmetry:1 uncle:3 operational:2 alg:3 e5:1 expansion:1 investigated:1 untrained:1 did:1 pk:4 statistica:1 main:3 linearly:1 dense:1 whole:4 apr:1 scored:7 hyperparameters:1 blockmodels:1 fair:1 graber:4 fig:3 intel:1 dump:1 west:2 egg:1 talley:2 aid:1 theme:1 mao:1 explicit:1 xh:2 lie:2 tied:1 ply:1 weighting:1 jmlr:2 aistats:1 admixed:1 british:1 nltk:2 xt:2 showing:1 list:4 x:1 evidence:3 inconclusive:2 essential:1 undergraduate:2 adding:1 airoldi:1 labelling:1 antonym:2 easier:1 entropy:1 pmrf:7 led:1 simply:4 univariate:2 likely:6 lazy:2 sport:1 inderjit:2 scalar:1 chang:1 rnd:2 corresponds:1 satisfies:1 determines:1 acm:2 gerrish:1 conditional:1 towards:1 fisher:1 man:5 change:1 directionality:1 included:2 determined:1 semantically:15 wordnet:2 degradation:1 called:2 total:2 experimental:2 attempted:1 meaningful:14 east:2 desert:1 formally:1 select:2 support:4 armijo:1 violated:1 bigdata:1 evaluate:8 correlated:3 |
4,751 | 5,301 | Dynamic Rank Factor Model for Text Streams
Shaobo Han?, Lin Du?, Esther Salazar and Lawrence Carin
Duke University, Durham, NC 27708
{shaobo.han, lin.du, esther.salazar, lcarin}@duke.edu
Abstract
We propose a semi-parametric and dynamic rank factor model for topic modeling, capable of (i) discovering topic prevalence over time, and (ii) learning contemporary multi-scale dependence structures, providing topic and word correlations as a byproduct. The high-dimensional and time-evolving ordinal/rank observations (such as word counts), after an arbitrary monotone transformation, are
well accommodated through an underlying dynamic sparse factor model. The
framework naturally admits heavy-tailed innovations, capable of inferring abrupt
temporal jumps in the importance of topics. Posterior inference is performed
through straightforward Gibbs sampling, based on the forward-filtering backwardsampling algorithm. Moreover, an efficient data subsampling scheme is leveraged
to speed up inference on massive datasets. The modeling framework is illustrated
on two real datasets: the US State of the Union Address and the JSTOR collection
from Science.
1
Introduction
Multivariate longitudinal ordinal/count data arise in many areas, including economics, opinion polls,
text mining, and social science research. Due to the lack of discrete multivariate distributions supporting a rich enough correlation structure, one popular choice in modeling correlated categorical
data employs the multivariate normal mixture of independent exponential family distributions, after
appropriate transformations. Examples include the logistic-normal model for compositional data
[1], the Poisson log-normal model for correlated count data [2], and the ordered probit model for
multivariate ordinal data [3]. Moreover, a dynamic Bayesian extension of the generalized linear
model [4] may be considered, for capturing the temporal dependencies of non-Gaussian data (such
as ordinal data). In this general framework, the observations are assumed to follow an exponential family distribution, with natural parameter related to a conditionally Gaussian dynamic model
[5], via a nonlinear transformation. However, these model specifications may still be too restrictive
in practice, for the following reasons: (i) Observations are usually discrete, non-negative and with
a massive number of zero values and, unfortunately, far from any standard parametric distributions
(e.g., multinomial, Poisson, negative binomial and even their zero-inflated variants). (ii) The number
of contemporaneous series can be large, bringing difficulties in sharing/learning statistical strength
and in performing efficient computations. (iii) The linear state evolution is not truly manifested after
a nonlinear transformation, where positive shocks (such as outliers and jumps) are magnified and
negative shocks are suppressed; hence, handling temporal jumps (up and down) is a challenge for
the above models.
We present a flexible semi-parametric Bayesian model, termed dynamic rank factor model (DRFM),
that does not suffer these drawbacks. We first reduce the effect of model misspecification by modeling the sampling distribution non-parametrically. To do so, we fit the observed data only after
some implicit monotone transformation, learned automatically via the extended rank likelihood [6].
Second, instead of treating panels of time series as independent collections of variables, we analyze
them jointly, with the high-dimensional cross-sectional dependencies estimated via a latent factor
?
contributed equally
1
model. Finally, by avoiding nonlinear transformations, both smooth transitions and sudden changes
(?jumps?) are better preserved in the state-space model, using heavy-tailed innovations.
The proposed model offers an alternative to both dynamic and correlated topic models [7, 8, 9],
with additional modeling facility of word dependencies, and improved ability to handle jumps. It
also provides a semi-parametric Bayesian treatment of dynamic sparse factor model. Further, our
proposed framework is applicable in the analysis of multiple ordinal time series, where the innovations follow either stationary Gaussian or heavy-tailed distributions.
2
Dynamic Rank Factor Model
We perform analysis of multivariate ordinal time series. In the most general sense, such ordinal
variables indicate a ranking of responses in the sample space, rather than a cardinal measure [10].
Examples include real continuous variables, discrete ordered variables with or without numerical
scales or, more specially, counts, which can be viewed as discrete variables with integer numeric
scales. Our goal is twofold: (i) discover the common trends that govern variations in observations,
and (ii) extract interpretable patterns from the cross-sectional dependencies.
Dependencies among multivariate non-normal variables may be induced through normally distributed latent variables. Suppose we have P ordinal-valued time series yp,t , p = 1, . . . , P ,
t = 1, . . . , T . The general framework contains three components:
yp,t ? g(zp,t ), zp,t ? p(? t ), ? t ? q(? t?1 ),
(1)
where g(?) is the sampling distribution, or marginal likelihood for the observations, the latent variable zp,t is modeled by p(?) (assumed to be Gaussian) with underlying system parameters ? t , and
q(?) is the system equation representing Markovian dynamics for the time-evolving parameter ? t .
In order to gain more model flexibility and robustness against misspecification, we propose a semiparametric Bayesian dynamic factor model for multiple ordinal time series analysis. The model is
based on the extended rank likelihood [6], allowing the transformation from the latent conditionally
Gaussian dynamic model to the multivariate observations, treated non-parametrically.
Extended rank likelihood (ERL): There exist many approaches for dealing with ordinal data, however, they all have some restrictions. For continuous variables, the underlying normality assumption
could be easily violated without a carefully chosen deterministic transformation. For discrete ordinal variables, an ordered probit model, with cut points, becomes computationally expensive if the
number of categories is large. For count variables, a multinomial model requires finite support on
the integer values. Poisson and negative binomial models lack flexibility from a practical viewpoint,
and often lead to non-conjugacy when employing log-normal priors.
Being aware of these issues, a natural candidate for consideration is the ERL [6]. With appropriate
monotone transformations learned automatically from data, it offers a unified framework for handling both continuous [11] and discrete ordinal variables. The ERL depends only on the ranks of the
observations (zero values in observations are further restricted to have negative latent variables),
zp,t ? D(Y ) ? {z p,t ? R : yp,t < yp0 ,t0 ? zp,t < zp0 ,t0 , and zp,t ? 0 if yp,t = 0}.
(2)
In particular, this offers a distribution-free approach, with relaxed assumptions compared to parametric models, such as Poisson log-normal [12]. It also avoids the burden of computing nuisance
parameters in the ordered probit model (cut points). The ERL has been utilized in Bayesian Gaussian
copula modeling, to characterize the dependence of mixed data [6]. In [13] a low-rank decomposition of the covariance matrix is further employed and efficient posterior sampling is developed in
[14]. The proposed work herein can be viewed as a dynamic extension of that framework.
2.1
Latent sparse dynamic factor model
In the forthcoming text, G(?, ?) denotes a gamma distribution with shape parameter ? and rate
parameter ?, TN(l,u) (?, ? 2 ) denotes a univariate truncated normal distribution within the interval
(l, u), and N+ (0, ? 2 ) is the half-normal distribution that only has non-negative support.
Assume z t ? N (0, ?t ), where ?t is usually a high-dimensional (P ? P ) covariance matrix.
To reduce the number of parameters, we assume a low rank factor model decomposition of the
covariance matrix ?t = ?V t ?T + R such that
z t = ?st + t , t ? N (0, R), R = I P .
(3)
2
Common trends (importance of topics) are captured by a low-dimensional factor score parameter
st . We assume autoregressive dynamics on sk,t ? AR(1|(?k , ?k,t )) with heavy-tailed innovations,
sk,t = ?k sk,t?1 + ?k,t , 0 < ?k < 1, ?k,t ? TPBN(e, f, ?), ? 1/2 ? C + (0, h),
(4)
where ?k,t follows the three-parameter beta mixture of normal TPBN(e, f, ?) distribution [15]. Parameter e controls the peak around zero, f controls the heaviness on the tails, and ? controls the
global sparsity with a half-Cauchy prior [16]. This prior encourages smooth transitions in general,
while jumps are captured by the heavy tails. The conjugate hierarchy may be equivalently represented as
?k,t ? N (0, ?k,t ), ?k,t ? G(e, ?k,t ), ?k,t ? G(f, ?) ? ? G(1/2, ?), ? ? G(1/2, h2 ).
Truncated normal priors are employed on ?k , ?k ? TN(0,1) (?0 , ?02 ), and assume s0,k ? N (0, ?s2 ).
Note that the extended rank likelihood is scale-free; therefore, we do not need to include a redundant
intercept parameter in (3). For the same reason, we set R = I P .
Model identifiability issues: Although
p the covariance matrix ?t is not identifiable [10], the related
correlation matrix C t = ?[i,j],t / ?[i,i],t ?[j,j],t , (i, j = 1, . . . , P ) may be identified, using the
parameter expansion technique [3, 13]. Further, the rank K in the low-rank decomposition of ?t is
also not unique. For the purpose of brevity, we do not explore this uncertainty here, but the tools
developed in the Bayesian factor analysis literature [17, 18, 19] can be easily adopted.
Identifiability is a key concern for factor analysis. Conventionally, for fixed K, a full-rank, lowertriangular structure in ? ensures identifiability [20]. Unfortunately, this assumption depends on the
ordering of variables. As a solution, we add nonnegative and sparseness constraints on the factor
loadings, to alleviate the inherit ambiguity, while also improving interpretability. Also, we add a
Procrustes post-processing step [21] on the posterior samples, to reduce this indeterminacy.
The nonnegative and (near) sparseness constraints are imposed by the following hierarchy,
1/2
?p,k ? N+ (0, lp,k ) lp,k ? G(a, up,k ), up,k ? G(b, ?k ), ?k ? C + (0, d).
(5)
Integrating out lp,k and up,k , we obtain a half-TPBN prior ?p,k ? TPBN+ (a, b, ?k ). The columnwise shrinkage parameters ?k enable factors to be of different sparsity levels [22]. We set hyperparameters a = b = e = f = 0.5, d = P , h = 1, ?s2 = 1. For weakly informative priors, we set
? = ? = 0.01; ?0 = 0.5, ?02 = 10.
2.2
Extension to handle multiple documents
nt
t
At each time point t we may have a corpus of documents {y nt t }N
nt =1 , where y t is a P -dimensional
observation vector, and Nt denotes the number of documents at time t. The model presented in
Section 2.1 is readily extended to handle this situation. Specifically, at each time point t, for each
nt
, is
document nt , the ERL representation for word count p, denoted by yp,t
nt
nt
yp,t
= g zp,t
, p = 1, . . . , P, t = 1, . . . , T, nt = 1, . . . , Nt ,
where z nt t ? RP and P is the vocabulary size. We assume a latent factor model for z nt t such that
z nt t = ?bnt t + nt t , nt t ? N (0, I P ), bnt t ? N (st , ?), ? = diag(?), ?k?1 ? G(?, ?),
?K
where ? ? RP
is the topic-word loading matrix, representing the K topics as columns of ?.
+
The factor score vector bnt t ? RK is the topic usage for each document y nt t , corresponding to locations in a low-dimensional RK space. The other parts of the model remain unchanged. The latent
trajectory s1:T represents the common trends for the K topics. Moreover, through the forward filtering backward sampling (FFBS) algorithm [23, 24], we also obtain time-evolving topic correlation
matrices ?t ? RK?K and word dependencies matrices C t ? RP ?P , offering a multi-scale graph
representation, a useful tool for document visualization.
2.3
Comparison with admixture topic models
Many topic models are unified in the admixture framework [25],
!
K
X
P (y n |w, ?) = P y n ?n =
wk,n ?k ,
Base
Admix
(6)
k=1
where y n is the P -dimensional observation vector of word counts in the n th document, and P denotes the vocabulary size. Traditionally, y n is generated from an admixture of base distributions, wn
is the admixture weight (topic proportion for document n), and ?k is the canonical parameter (word
3
distribution for topic k), which denotes the location of the kth topic on the P -1 dimensional simplex.
For example, latent Dirichlet allocation (LDA) [26] assumes the base distribution to be multinomial,
with ?k ? Dir(?0 ), wn ? Dir(? 0 ). The correlated topic model (CTM) [8] modifies the topic distribution, with wn ? Logistic Normal(?, ?). The dynamic topic model (DTM) [7] analyzes document collections in a known chronological order. In order to incorporate the state space model, both
the topic proportion and the word distribution are changed to logistic normal, with isotropic covariance matrices wt ? Logistic Normal(wt?1 , ? 2 I K ) and ?k,t ? Logistic Normal(?k,t?1 , vI P ),
respectively. To overcome the drawbacks of multinomial base, spherical topic models [27] assume
the von Mises-Fisher (vMF) distribution as its base distribution, with ?k ? vMF(?, ?) lying on a
unit P -1 dimensional sphere. Recently in [25] the base and word distribution are both replaced with
Poisson Markov random fields (MRFs), which characterizes word dependencies.
We present here a semi-parametric factor model formulation,
!
K
X
P(y n |s, ?) , P z n ? D(Y ) ?n =
sk,n ?k ,
(7)
k=1
with y n defined as above, ?k ? RP
+ is a vector of nonnegative weights, indicating the P vocabulary usage in each individual topics k, and sn ? RK is the topic usage. Note that the extended
rank likelihood does not depend on any assumptions about the data marginal distribution, making it
appropriate for a broad class of ordinal-valued observations, e.g., term frequency-inverse document
frequency (tf-idf) or rankings, beyond word counts. However, the proposed model here is not an
admixture model, as the topic usage is allowed to be either positive or negative.
The DRFM framework has some appealing advantages: (i) It is more natural and convenient to incorporate with sparsity, rank selection, and state-space model; (ii) it provides topic-correlations and
word-dependences as a byproduct; and (iii) computationally, this model is tractable and often leads
to locally conjugate posterior inference. DRFM has limitations. Since the marginal distributions
are of unspecified types, objective criteria (e.g. perplexity) is not directly computable. This makes
quantitative comparisons to other parametric baselines developed in the literature very difficult.
3
Conjugate Posterior Inference
Let ? = {?, S, L, U , ?, ?, ?, ? , ?, ?, ?} denote the set of parameters in basic model, and let Z be
the augmented data (from the ERL). We use Gibbs sampling to approximate the joint posterior distribution p(Z, ?|Z ? R(Y )). The algorithm alternates between sampling p(Z|?, Z ? R(Y )) and
p(?|Z, Z ? R(Y )) (reduced to p(?|Z)). The derivation of the Gibbs sampler is straightforward,
and for brevity here we only highlight the sampling steps for Z, and the forward filtering backward
sampling (FFBS) steps for the trajectory s1:T . The Supplementary Material contains further details
for the inference.
PK
? Sampling zp,t : p(zp,t |?, Z ? R(Y ), Z ?p,?t ) ? TN[zp,t ,zp,t ] ( k=1 ?p,k sk,t , 1), where zp,t =
max{zp0 ,t0 : yp0 ,t0 < yp,t } and zp,t = min{zp0 ,t0 : yp0 ,t0 > yp,t }.
This conditional sampling scheme is widely used in [6, 10, 13]. In [14] a novel Hamiltonian Monte
Carlo (HMC) approach has been developed recently, for a Gaussian copula extended rank likelihood
model, where ranking is only within each row of Z. This method simultaneously samples a column
vector of z i conditioned on other columns Z ?i , with higher computation but better mixing.
? Sampling st : we have the state model st |st?1 ? N (Ast?1 , Qt ), and the observation model
z t |st ? N (?st , R),1 where A = diag(?), Qt = diag(? t ), R = I P . for t = 1, . . . , T
1. Forward Filtering: beginning at t = 0 with s0 ? N (0, ?s2 I K ), for all t = 1, . . . , T , we
find the on-line posteriors at t, p(st |z 1:t ) = N (mt , V t ), where mt = V t {?T R?1 z t +
?1
T ?1
H ?1
?]?1 , and H t = Qt + AV t?1 AT .
t Amt?1 }, V t = [H t + ? R
2. Backward Sampling: starting from N (f
mt , Ve t ), the backward smoothing density, i.e., the
e t?1 ), where
conditional distribution of st?1 given st , is p(st?1 |st , z 1:(t?1) ) = N (e
?t?1 , ?
T ?1
?1
?1
T ?1
?1
e
e
e t?1 = ?t?1 {A Qt st + V t?1 mt?1 }, ?t?1 = (V t?1 + A Qt A) .
?
There exist different variants of FFBS schemes (see [28] for a detailed comparison); the method we
choose here enjoys fast decay in autocorrelation and reduced computation time.
1
For brevity, we omit the dependencies on ? in notation
4
3.1
Time-evolving topic and word dependencies
We also have the backward recursion density at t ? 1, p(st?1 |z 1:T ) = N (f
mt?1 , Ve t?1 ), where
T ?1
?1
T ?1 e
e
e
e
e
e
ft?1 = ?t?1 (A Qt m
ft + V t?1 mt?1 ) and V t?1 = ?t?1 + ?t?1 A Qt V t Q?1
m
t A?t?1 .
We perform inference on the K ? K time-evolving topic dependences in s1:T , using the posterior
p
covariances {Ve 1:T } (with topic correlation matrices ?1:T , ?[r,s],t = V[r,s],t / V[r,r],t V[s,s],t , r, s =
1, . . . , K), and further obtain the P ? P time-evolving word dependencies capsuled in {?1:T }
with ?t = ?Ve t ?T + I P . Essentially, this can be viewed as a dynamic Gaussian copula model,
et ? N (0, C t ), where g(?) is a non-decreasing function of a univariate marginal
yp,t = g(e
zp,t ), z
likelihood and C t (t = 1, . . . , T ) is the correlation matrix capturing the multivariate dependence.
We obtain a posterior distribution for C 1:T as a byproduct, without having to estimate the nuisance
parameters in marginal likelihoods g(?). This decoupling strategy resembles the idea of copula
models.
3.2
Accelerated MCMC via document subsampling
For large-scale datasets, recent approaches efficiently reduce the computational load of Monte Carlo
Markov chain (MCMC) by data subsampling [29, 30]. We borrow this idea of subsampling documents when considering a large corpora (e.g., in our experiments, we consider analysis of articles
in the magazine Science, composed of 139379 articles from years 1880 to 2002, and a vocabulary
size 5855). In our model, the augmented data z nt t (nt = 1, . . . , Nt ) for each document is relatively
expensive to sample. One simple method is random document sampling without replacement. However, by treating all likelihood contributions symmetrically, this method leads to a highly inefficient
MCMC chain with poor mixing [29].
Alternatively, we adopt the probability proportional-to-size (PSS) sampling scheme in [30], i.e.,
sampling the documents with inclusion probability proportional to the likelihood contributions. For
each MCMC iteration, the sub-sampling procedure for documents at time t is designed as follows:
? Step 1: Given a small subset Vt ? {1, . . . , Nt } of chosen documents, only sample {z dt } for all
d ? Vt and compute the augment log-likelihood contributions (with B t integrated out) `Vt (z dt ) =
e where R
e = ???T + I P . Note that, only a K-dimensional matrix inversion is
N (?st , R),
e ?1 = I P ? ?(??1 + ?T ?)T ?T .
required, by using the Woodbury matrix inversion formula R
? Step 2: Similar to [30], we use a Gaussian process [31] to predict the log-likelihood for
the remaining documents `Vtc (z dt ) = K(Vtc , Vt )K(Vt , Vt )?1 `Vt (z dt ), where K is a Nt ?
j
i
Nt squared-exponential
kernel,
which denotes the similarity of documents: K(y t , y t ) =
?f2 exp ?||y it ? y jt ||2 /(2s2 ) , i, j = 1, . . . , Nt , ?f2 = 1, s = 1.
P
ed = wd / d0 wd0 .
? Step 3: Calculate the inclusion probability wd ? exp [`(z dt )], d = 1, . . . , Nt , w
? Step 4: Sampling the next subset Vt of pre-specified size |Vt | with inclusion probability w
ed , and
store it for the use of the next MCMC iteration.
In practice, this adaptive design allows MCMC to run more efficiently on a full dataset of large
scale, often mitigating the need to do parallel MCMC implementation. Future work could also consider nonparametric function estimation subject to monotonicity constraint, e.g. Gaussian process
projections recently developed in [32].
4
Experiments
Different from DTM [7] , the proposed model has the jumps directly at the level of the factor scores
(no exponentiation or normalization needed), and therefore it proved more effective in uncovering
jumps in factor scores over time. Demonstrations of this phenomenon in a synthetic experiment are
detailed in the Supplementary Material. In the following, we present exploratory data analysis on
two real examples, demonstrating the ability of the proposed model to infer temporal jumps in topic
importance, and to infer correlations across topics and words.
4.1
Case Study I: State of the Union dataset
The State of the Union dataset contains the transcripts of T = 225 US State of the Union addresses,
from 1790 to 2014. We take each transcript as a document, i.e., we have one document per year.
5
After removing stop words, and removing terms that occur fewer than 3 times in one document and
less than 10 times overall, we have P = 7518 unique words. The observation yp,t corresponds to
the frequency of word p of the State of the Union transcript from year t.
We apply the proposed DRFM setting and learned K = 25 topics. To better understand the temporal
dynamic per topic, six topics are selected and the posterior mean of their latent trajectories sk,1:T
are shown in Figure 1 (with also the top 12 most probable words associated with each of the topics).
A complete table with all 25 learned topics and top 12 words is provided in the Supplementary
Material. The learned trajectory associated with every topic indicates different temporal patterns
across all the topics. Clearly, we can identify jumps associated with some key historical events. For
instance, for Topic 10, we observe a positive jump in 1846 associated with the Mexican-American
war. Topic 13 is related with the Spanish-American war of 1898, with a positive jump in that year.
In Topic 24, we observe a positive jump in 1914, when the Panama Canal was officially opened
(words Panana and canal are included). In Topic 18, the positive jumps observed from 1997 to
1999 seem to be associated with the creation of the State Children?s Health Insurance Program in
1997. We note that the words for this topic are explicitly related with this issue. Topic 25 appears to
be related to banking; the significant spike around 1836 appears to correspond to the Second Bank
of the United States, which was allowed to go out of existence, and end national banking that year.
In 1863 Congress passed the National Banking Act, which ended the ?free-banking? period from
1836-1863; note the spike around 1863 in Topic 25.
Topic 10
4
4
Topic 17
2
2
0
0
6
Topic 13
4
6
Topic 18
4
2
2
0
0
10
6
Topic 24
4
0
0
-5
1800
1850
Topic#10
Mexico
Government
Texas
United
War
Mexican
Army
Territory
Country
Peace
Policy
Lands
1900
Topic#13
Government
United
Islands
Commission
Island
Cuba
Spain
Act
General
Military
International
Officiers
1950
Topic 25
5
2
1800
2000
Topic#24
United
Treaty
Isthmus
Public
Panama
Law
Territory
America
Canal
Service
Banks
Colombia
Topic#17
Jobs
Country
Tax
American
Economy
Deficit
Americans
Energy
Businesses
Health
Plan
Care
1850
1900
Topic#18
Children
America
Americans
Care
Tonight
Support
Century
Health
Working
Challenge
Security
Families
1950
2000
Topic#25
Government
Public
Banks
Bank
Currency
Money
United
Federal
American
National
Duty
Institutions
Figure 1: (State of the Union dataset) Above: Time evolving from 1790 to 2014 for six selected
topics. The plotted values represent the posterior means. Below: Top 12 most probable words
associated with the above topics.
Our modeling framework is able to capture dynamic patterns of topics and word correlations. To
illustrate this, we select three years (associated with some meaningful historical events) and analyze
their corresponding topic and word correlations. Figure 2 (first row) shows graphs of the topic
correlation matrices, in which the nodes represent topics and the edges indicate positive (green) and
negative (red) correlations (we show correlations with absolute value larger than 0.01). We notice
that Topics 11 and 22 are positively correlated with those years. Some of the most probable words
associated with each of them are: increase, united, law and legislation (for Topic 11) and war,
Mexico, peace, army, enemy and military (for Topic 22). We also are interested in understanding
the time-varying correlation between words. To do so, and for the same years as before, in Figure 2
(second row) we plot the dendrogram associated with the learned correlation matrix for words. In
the plots, different colors indicate highly correlated word clusters defined by cutting the branches off
the dendrogram. Those figures reveal different sets of highly correlated words for different years. By
6
1846
Mexican-American War
T6
T5
T2
T22
T11
T5
T14
T13
T7
T14
T18
T18
T20
T12
T10
T25
T9
T18
T21
T19
T16
T10
T23
T21
T21
T1
tre
ex cu b asu
pe rr an r
nd en k y
itu cys
re
dofisca s
milllliars l
o
n
b
illion
constitut
ion
union
president
n
tio om
nareeed s
f re ion
f at rgy
nene
nt
lth
heeavelopmse
d ogram
pr
ic
nom
eco
program
country
general
pow
er
auth
pub ority
gfoederlic
v al
cict onv ernm
ize en e
ns tio nt
n
1.0
0.5
0.0
?0.5
?1.0
n
ai
ry
sp ct ritoy
ater at
tre ited l
n
unationaistratio
n dmin
a olicy
p bject
su sent
pre
r
n
we ns tio
po itizneven
cco ain
sp
y
actrritor
teeaty
tr ited
unexican
m s
texa
mi pea
for litar ce
ce y
arm
ys
islacnuba
islandds
mexico
texas
mexican
treasury
s
bank
ncy
curreident
prespolicyn
t
a io al
istr tion on
min natitutinion
ad
s
n u
co
pr
am subese
n
foer ricaject t
eig n
t
dep orrdaden
ar tm er
e
canant
cour l
americat
americans
service
increase
nu
ju mber
ta ne
carx
job e
t ch s
amoniigldre
am
h n
erericat
ica n
s
om
ed
frereetions
f a rt
n ou l
c ana r tment
c epa
d er
ordde
tra ign
fore
american
country
general
autho
po
ri
pubwer ty
gfoederlic
v al
cocitiz ernm
nv ens e
nt
en
tio
n
n
ai
ry
sp ctrritoy
ate at
tre ited
unx
taare
c bs
jo ren
childght
toni
pm
enhea ent
erglth
w
pe a y
milit acer
a
forcery
armys
islands
cuba
island
mexxicaos
te an
ic
mex gold
tes
noilvers
s ndve
boser
re
secretary
n
attentpioort
re ne
ju er
b se
num
rea aw
inc l essor r
sin ab a
bu l w
servicey
secrettiaron
atterneporet
jun er
mb e
nu rearsamic
incrog omms
p n ra
og
ec ro
p
secretary
n
attentpio
rt
re o
ent
pres ject
sub laws
es r
sin bo al
bu lation ion
t
na na
de
ve
lo
om
ed
frereetions
f a rt
n ou l
c ana r tment
c epa
d er
ordde
tra ign
fore
american
country
general
autho
pub rity
li
fe
g de c
p ov ral
u re ern
co nio side men
ns n nt t
tit
ut
io
n
T1
T9
T24
nation
labor
bu
lawsiness
ta
c x
jobare
tocnhilds
a
i
m
am e ghren
er ric t
ic an
a s
T1
m
bil illio
p wlio n
mil eacearn
forcitary
armes
island y
cubas
island
mexico
T23
T12
T2
T16
n go
silotesld
v
resbonder
trea erves
sury
b
anks
curren
cy
policy
administration
ent
pm
develo ra
m
og
pr
omic
ecoongramsh
pr earltgy
h e
en turesal
i c
nd fis llars
pe
do
ex
T16
T15
T15
T15
T24
T9
T20
T17
T23
T8
T11
T7
T20
T4
T22
T2
T24
T25
T11
T22
T14
T3
T17
T10
T12
T4
T8
T19
T3
T8
T3
T5
T13
T13
T6
T6
T4
T19
T17
2003
Iraq War
service
billion
fis
ex cal
millipenditure
s
o
d
o
r lla n
b es rs
si on erve
gonotlverds
e
ld s
T25 T7
1929
Economic Depression
Figure 2: (State of the Union dataset) First row: Inferred correlations between topics for some
specific years associated with some meaningful historical events. Green edges indicate positive
correlations and red edges indicate negative correlations. Second row: Learned dendrogram based
upon the correlation matrix between the top 10 words associated with each topic (we display 80
unique words in total).
inspecting all the words correlation, we noticed that the set of words {government, federal, public,
power, authority, general, country} are highly correlated across the whole period.
4.2
Case Study II: Analysis of Science dataset
We analyze a collection of scientific documents from the JSTOR Science journal [7]. This dataset
contains a collection of 139379 documents from 1880 to 2002 (T = 123), with approximately 1100
documents per year. After removing terms that occurred fewer than 25 times, the total vocabulary
size is P = 5855. We learn K = 50 topics from the inferred posterior distribution, for brevity and
simplicity, we only show 20 of them. We handle about 2700 documents per iteration (subsampling
rate: 2%). Table 1 shows the 20 selected topics and the top 10 most probable words associated with
each of them. By inspection, we notice that those topics are related with specific fields in science.
For instance, Topic 2 is more related to ?scientific research?, Topic 10 to ?natural resources?, and
Topic 15 to ?genetics?. Figure 3 shows the time-varying trend for some specific words, zbp,1:T , which
reveals the importance of those words across time. Finally, Figure 4 shows the correlation between
the selected 20 topics. For instance, in 1950 and 2000, topic 9 (related to mouse, cells, human,
transgenic) and topic 17 (related to virus, rna, tumor, infection) are highly correlated.
DNA
RNA
Gene
2.5
2
1
Cancer
Patients
Nuclear
0.7
0.6
0.8
1.5
Astronomy
Psychology
Brain
0.5
0.6
1
0.4
0.5
0.4
0
0.2
0.3
0.2
?0.5
0.1
0
?1
1880
1900
1920
1940
1960
1980
2000
1880
1900
1920
1940
1960
1980
2000
0
1880
1900
1920
1940
1960
1980
2000
Figure 3: (Science dataset) the inferred latent trend for variable zbp,1:T associated with words.
7
1900
1950
T14
T14
T8
T15
T12
T9
2000
T11
T11
T13
T20
T19
T18
T20
T17
T19
T20
T13
T3
T7
T15
T16
T15
T17
T2
T6
T5
T9
T4
T10
T16
T10
T5
T18
T3
T10
T4
T19
T12
T8
T9
T1
T1
T17
T6
T5
T2
T4
T2
T11
T16
T13
T12
T7
T3
T6
T14
T18
T8
T1
T7
1.0
0.5
0.0
?0.5
?1.0
Figure 4: (Science dataset) Inferred correlations between topics for some specific years. Green
edges indicate positive correlations and red edges indicate negative correlations.
Table 1: Selected 20 topics associated with the analysis of the Science dataset and top 10 most
probable words.
Topic#1
cells
cell
normal
two
growth
development
tissue
body
egg
blood
Topic#11
system
nuclear
new
systems
power
cost
computer
fuel
coal
plant
5
Topic#2
research
national
government
support
federal
development
new
program
scientific
basic
Topic#12
energy
theory
temperature
radiation
atoms
surface
atomic
mass
atom
time
Topic#3
field
magnetic
solar
energy
spin
state
electron
quantum
temperature
current
Topic#13
association
science
meeting
university
american
society
section
president
committee
secretary
Topic#4
animals
brain
neurons
activity
response
rats
control
fig
effects
days
Topic#14
protein
proteins
cell
membrane
amino
sequence
binding
acid
residues
sequences
Topic#5
energy
oil
percent
production
fuel
total
growth
states
electricity
coal
Topic#15
human
genome
sequence
chromosome
gene
genes
map
data
sequences
genetic
Topic#6
university
professor
college
president
department
research
institute
director
society
school
Topic#16
professor
university
society
department
college
president
director
american
appointed
medical
Topic#7
science
scientific
new
scientists
human
men
sciences
knowledge
meeting
work
Topic#17
virus
rna
viruses
particles
tumor
mice
disease
viral
human
infection
Topic#8
work
research
scientific
laboratory
made
university
results
science
survey
department
Topic#18
energy
electron
state
fig
two
structure
reaction
laser
high
temperature
Topic#9
mice
mouse
type
wild
fig
cells
human
transgenic
animals
mutant
Topic#19
stars
mass
star
temperature
solar
gas
data
density
surface
galaxies
Topic#10
water
surface
temperature
soil
pressure
sea
plants
solution
plant
air
Topic#20
rna
fig
mrna
protein
site
sequence
splicing
synthesis
trna
rnas
Discussion
We have proposed a DRFM framework that could be applied to a broad class of applications such
as: (i) dynamic topic model for the analysis of time-stamped document collections; (ii) joint analysis of multiple time series, with ordinal valued observations; and (iii) multivariate ordinal dynamic
factor analysis or dynamic copula analysis for mixed type of data. The proposed model is a semiparametric methodology, which offers modeling flexibilities and reduces the effect of model misspecification. However, as the marginal likelihood is distribution-free, we could not calculate the
model evidence or other evaluation metrics based on it (e.g. held-out likelihood). As a consequence,
we are lack of objective evaluation criteria, which allow us to perform formal model comparisons.
In our proposed setting, we are able to perform either retrospective analysis or multi-step ahead
forecasting (using the recursive equations derived in the FFBS algorithm). Finally, our inference
framework is easily adaptable for using sequential Monte Carlo (SMC) methods [33] allowing online learning.
Acknowledgments
The research reported here was funded in part by ARO, DARPA, DOE, NGA and ONR. The authors
are grateful to Jonas Wallin, Lund University, Sweden, for providing efficient package on simulation
of the GIG distribution.
8
References
[1] J. Aitchison. The statistical analysis of compositional data. J. Roy. Stat. Soc. Ser. B, 44(2):139?177, 1982.
[2] S. Chib and R. Winkelmann. Markov chain Monte Carlo analysis of correlated count data. Journal of
Business & Economic Statistics, 19(4), 2001.
[3] E. Lawrence, D. Bingham, C. Liu, and V. N. Nair. Bayesian inference for multivariate ordinal data using
parameter expansion. Technometrics, 50(2), 2008.
[4] M. West, P. J. Harrison, and H. S. Migon. Dynamic generalized linear models and Bayesian forecasting.
J. Am. Statist. Assoc., 80(389):73?83, 1985.
[5] C. Cargnoni, P. M?uller, and M. West. Bayesian forecasting of multinomial time series through conditionally Gaussian dynamic models. J. Am. Statist. Assoc., 92(438):640?647, 1997.
[6] P. D. Hoff. Extending the rank likelihood for semiparametric copula estimation. Ann. Appl. Statist.,
1(1):265?283, 2007.
[7] D. M. Blei and J. D. Lafferty. Dynamic topic models. In Int. Conf. Machine Learning, 2006.
[8] D. M. Blei and J. D. Lafferty. Correlated topic models. In Adv. Neural Inform. Processing Systems, 2006.
[9] A. Ahmed and E. P. Xing. Timeline: A dynamic hierarchical dirichlet process model for recovering
birth/death and evolution of topics in text stream. 2010.
[10] P. D. Hoff. A first course in Bayesian statistical methods. Springer, 2009.
[11] A. N. Pettitt. Inference for the linear model using a likelihood based on ranks. J. Roy. Stat. Soc. Ser. B,
44(2):234?243, 1982.
[12] J. Aitchison and C. H. Ho. The multivariate Poisson-log normal distribution. Biometrika, 76(4):643?653,
1989.
[13] J. S. Murray, D. B. Dunson, L. Carin, and J. E. Lucas. Bayesian Gaussian copula factor models for mixed
data. J. Am. Statist. Assoc., 108(502):656?665, 2013.
[14] A. Kalaitzis and R. Silva. Flexible sampling of discrete data correlations without the marginal distributions. In Adv. Neural Inform. Processing Systems, 2013.
[15] A. Armagan, M. Clyde, and D. B. Dunson. Generalized Beta mixtures of Gaussians. In Adv. Neural
Inform. Processing Systems, 2011.
[16] N. G. Polson and J. G. Scott. On the half-Cauchy prior for a global scale parameter. Bayesian Analysis,
7(4):887?902, 2012.
[17] H. F. Lopes and M. West. Bayesian model assessment in factor analysis. Statistica Sinica, 14(1):41?68,
2004.
[18] J. Ghosh and D. B. Dunson. Default prior distributions and efficient posterior computation in Bayesian
factor analysis. Journal of Computational and Graphical Statistics, 18(2):306?320, 2009.
[19] A. Bhattacharya and D. B. Dunson. Sparse Bayesian infinite factor models. Biometrika, 98(2):291?306,
2011.
[20] J. Geweke and G. Zhou. Measuring the pricing error of the arbitrage pricing theory. Review of Financial
Studies, 9(2):557?587, 1996.
[21] A. Christian, B. Jens, and P. Markus. Bayesian analysis of dynamic factor models: an ex-post approach
towards the rotation problem. Kiel Working Papers 1902, Kiel Institute for the World Economy, 2014.
[22] C. Gao and B. E. Engelhardt. A sparse factor analysis model for high dimensional latent spaces. In NIPS:
Workshop on Analysis Operator Learning vs. Dictionary Learning: Fraternal Twins in Sparse Modeling,
2012.
[23] C. K. Carter and R. Kohn. On Gibbs sampling for state space models. Biometrika, 81:541?553, 1994.
[24] S. Fr?uhwirth-Schnatter. Data augmentation and dynamic linear models. Journal of Times Series Analysis,
15:183?202, 1994.
[25] D. Inouye, P. Ravikumar, and I. Dhillon. Admixture of Poisson MRFs: A topic model with word dependencies. In Int. Conf. Machine Learning, 2014.
[26] D. M. Blei, A. Y. Ng, and M. I. Jordan. Latent Dirichlet allocation. J. Machine Learn. Res., 3:993?1022,
2003.
[27] J. Reisinger, A. Waters, B. Silverthorn, and R. J. Mooney. Spherical topic models. In Int. Conf. Machine
Learning, 2010.
[28] E. A. Reis, E. Salazar, and D. Gamerman. Comparison of sampling schemes for dynamic linear models.
International Statistical Review, 74(2):203?214, 2006.
[29] A. Korattikara, Y. Chen, and M. Welling. Austerity in MCMC land: cutting the Metropolis-Hastings
budget. In Int. Conf. Machine Learning, pages 181?189, 2014.
[30] M. Quiroz, M. Villani, and R. Kohn.
Speeding up MCMC by efficient data subsampling.
arXiv:1404.4178, 2014.
[31] C. E. Rasmussen. Gaussian processes in machine learning. Springer, 2004.
[32] L. Lin and D. B. Dunson. Bayesian monotone regression using Gaussian process projection. Biometrika,
101(2):303?317, 2014.
[33] A. Doucet, D. F. Nando, and N. Gordon. Sequential Monte Carlo methods in practice. Springer, 2001.
9
| 5301 |@word cu:1 inversion:2 proportion:2 loading:2 nd:2 villani:1 cco:1 simulation:1 r:1 covariance:6 decomposition:3 pressure:1 gamerman:1 tr:1 ld:1 liu:1 series:9 score:4 united:6 unx:1 t7:6 offering:1 contains:4 pub:2 genetic:1 document:27 longitudinal:1 reaction:1 current:1 wd:2 virus:3 nt:28 si:1 readily:1 numerical:1 informative:1 shape:1 christian:1 designed:1 treating:2 plot:2 interpretable:1 v:1 stationary:1 half:4 discovering:1 selected:5 fewer:2 asu:1 inspection:1 isotropic:1 beginning:1 hamiltonian:1 sudden:1 num:1 provides:2 institution:1 authority:1 blei:3 node:1 location:2 nom:1 kiel:2 beta:2 jonas:1 director:2 wild:1 autocorrelation:1 coal:2 ra:2 ica:1 multi:3 ry:2 brain:2 spherical:2 decreasing:1 automatically:2 considering:1 becomes:1 spain:1 provided:1 underlying:3 notation:1 discover:1 mass:2 fuel:2 panel:1 moreover:3 unspecified:1 developed:5 unified:2 astronomy:1 transformation:9 magnified:1 ghosh:1 ended:1 temporal:6 quantitative:1 every:1 act:2 nation:1 growth:2 chronological:1 ro:1 assoc:3 biometrika:4 ser:2 control:4 normally:1 medical:1 omit:1 unit:1 t13:6 service:3 scientist:1 before:1 t1:6 congress:1 positive:9 io:2 consequence:1 trna:1 approximately:1 resembles:1 appl:1 co:2 smc:1 unique:3 acknowledgment:1 woodbury:1 practical:1 atomic:1 union:8 practice:3 recursive:1 prevalence:1 lcarin:1 procedure:1 acer:1 area:1 lla:1 evolving:7 t10:6 convenient:1 projection:2 word:41 integrating:1 pre:2 protein:3 selection:1 cal:1 operator:1 transition:2 ast:1 intercept:1 restriction:1 deterministic:1 map:1 imposed:1 modifies:1 straightforward:2 economics:1 starting:1 go:2 mrna:1 survey:1 simplicity:1 abrupt:1 colombia:1 nuclear:2 borrow:1 erl:6 financial:1 century:1 handle:4 exploratory:1 traditionally:1 variation:1 president:4 hierarchy:2 suppose:1 magazine:1 massive:2 duke:2 trend:5 roy:2 expensive:2 utilized:1 iraq:1 cut:2 observed:2 ft:2 capture:1 calculate:2 cy:1 t12:6 ensures:1 adv:3 ordering:1 contemporary:1 disease:1 govern:1 dynamic:29 weakly:1 depend:1 grateful:1 tit:1 ov:1 creation:1 upon:1 f2:2 treaty:1 po:2 darpa:1 easily:3 joint:2 represented:1 america:2 derivation:1 laser:1 fast:1 effective:1 monte:5 birth:1 larger:1 valued:3 enemy:1 supplementary:3 widely:1 ability:2 statistic:2 jointly:1 online:1 advantage:1 sequence:5 rr:1 t21:3 propose:2 aro:1 mb:1 fr:1 korattikara:1 mixing:2 flexibility:3 uhwirth:1 tax:1 gold:1 cuba:3 rgy:1 ent:3 billion:1 cour:1 cluster:1 extending:1 sea:1 zp:14 illustrate:1 radiation:1 stat:2 school:1 qt:7 dep:1 indeterminacy:1 job:2 bnt:3 soc:2 recovering:1 transcript:3 indicate:7 inflated:1 drawback:2 pea:1 opened:1 human:5 nando:1 enable:1 opinion:1 material:3 ana:2 public:3 government:5 t5:6 alleviate:1 probable:5 inspecting:1 extension:3 lying:1 around:3 considered:1 ic:3 normal:16 exp:2 lawrence:2 predict:1 electron:2 dictionary:1 adopt:1 ctm:1 purpose:1 estimation:2 applicable:1 ain:1 pres:1 t22:3 tf:1 tool:2 uller:1 federal:3 clearly:1 gaussian:14 rna:5 rather:1 zhou:1 shrinkage:1 og:2 varying:2 mil:1 derived:1 mutant:1 rank:20 indicates:1 ral:1 likelihood:17 t14:6 ps:1 baseline:1 sense:1 am:6 esther:2 inference:9 economy:2 austerity:1 secretary:3 mrfs:2 integrated:1 fis:2 interested:1 mitigating:1 overall:1 issue:3 among:1 flexible:2 augment:1 uncovering:1 lucas:1 denoted:1 plan:1 smoothing:1 development:2 copula:7 hoff:2 animal:2 field:3 aware:1 marginal:7 having:1 ng:1 sampling:21 atom:2 represents:1 broad:2 carin:2 future:1 simplex:1 t2:6 gordon:1 cardinal:1 employ:1 bil:1 chib:1 composed:1 ve:5 national:4 gamma:1 simultaneously:1 individual:1 replaced:1 replacement:1 technometrics:1 ab:1 mining:1 highly:5 insurance:1 evaluation:2 mixture:3 truly:1 tonight:1 held:1 chain:3 heaviness:1 wd0:1 edge:5 byproduct:3 capable:2 sweden:1 accommodated:1 re:7 plotted:1 instance:3 military:2 column:3 modeling:9 markovian:1 ncy:1 ar:2 measuring:1 epa:2 electricity:1 cost:1 parametrically:2 subset:2 trea:1 too:1 characterize:1 commission:1 reported:1 dependency:11 aw:1 dir:2 author:1 synthetic:1 clyde:1 st:16 density:3 international:2 peak:1 ju:2 jump:14 bu:3 off:1 synthesis:1 mouse:4 t16:6 na:2 jo:1 augmentation:1 ambiguity:1 von:1 squared:1 choose:1 leveraged:1 conf:4 american:12 inefficient:1 yp:10 li:1 de:2 star:2 twin:1 wk:1 int:4 inc:1 tra:2 explicitly:1 ranking:3 depends:2 stream:2 vi:1 performed:1 ad:1 tion:1 analyze:3 characterizes:1 red:3 xing:1 parallel:1 identifiability:3 solar:2 contribution:3 om:3 air:1 spin:1 acid:1 efficiently:2 t3:6 correspond:1 identify:1 reisinger:1 bayesian:17 territory:2 ren:1 carlo:5 trajectory:4 fore:2 mooney:1 tissue:1 nene:1 inform:3 sharing:1 ed:4 infection:2 against:1 ty:1 energy:5 frequency:3 galaxy:1 naturally:1 associated:14 mi:2 stop:1 gain:1 dataset:10 treatment:1 popular:1 proved:1 onv:1 color:1 ut:1 knowledge:1 geweke:1 ou:2 carefully:1 adaptable:1 appears:2 higher:1 dt:5 day:1 follow:2 methodology:1 response:2 improved:1 ta:2 formulation:1 implicit:1 dendrogram:3 correlation:25 working:2 tre:3 hastings:1 su:1 nonlinear:3 assessment:1 eig:1 lack:3 logistic:5 ex:4 lda:1 reveal:1 scientific:5 pricing:2 usage:4 effect:3 oil:1 ize:1 evolution:2 hence:1 zbp:2 facility:1 dhillon:1 laboratory:1 death:1 illustrated:1 conditionally:3 sin:2 spanish:1 nuisance:2 encourages:1 t25:3 rat:1 criterion:2 generalized:3 complete:1 vmf:2 tn:3 eco:1 temperature:5 silva:1 percent:1 consideration:1 novel:1 recently:3 common:3 rotation:1 viral:1 multinomial:5 mt:6 association:1 occurred:1 tail:2 significant:1 gibbs:4 ai:2 pm:2 inclusion:3 particle:1 funded:1 specification:1 han:2 similarity:1 surface:3 money:1 add:2 base:6 multivariate:11 t19:6 recent:1 posterior:13 perplexity:1 termed:1 store:1 manifested:1 onr:1 vt:9 meeting:2 jens:1 captured:2 analyzes:1 additional:1 care:2 gig:1 relaxed:1 employed:2 redundant:1 salazar:3 period:2 ii:6 branch:1 currency:1 semi:4 multiple:4 reduces:1 mber:1 full:2 d0:1 infer:2 smooth:2 ahmed:1 offer:4 sphere:1 ited:3 lin:3 ign:2 cross:2 post:2 equally:1 y:1 ravikumar:1 peace:2 dtm:2 variant:2 basic:2 pow:1 regression:1 essentially:1 metric:1 patient:1 poisson:7 arxiv:1 iteration:3 kernel:1 represent:2 normalization:1 mex:1 ion:3 cell:5 rea:1 preserved:1 semiparametric:3 residue:1 interval:1 harrison:1 country:6 specially:1 bringing:1 nv:1 subject:1 induced:1 sent:1 lafferty:2 seem:1 jordan:1 integer:2 near:1 symmetrically:1 silverthorn:1 iii:3 enough:1 curren:1 wn:3 fit:1 psychology:1 forthcoming:1 identified:1 reduce:4 idea:2 tm:1 economic:2 computable:1 administration:1 texas:2 t0:6 six:2 war:6 duty:1 kohn:2 passed:1 ject:1 retrospective:1 forecasting:3 suffer:1 compositional:2 depression:1 useful:1 detailed:2 se:1 procrustes:1 nonparametric:1 officially:1 locally:1 statist:4 category:1 dna:1 reduced:2 carter:1 exist:2 canonical:1 notice:2 canal:3 estimated:1 per:4 aitchison:2 discrete:7 key:2 tpbn:4 demonstrating:1 poll:1 blood:1 stamped:1 ce:2 shock:2 backward:5 graph:2 monotone:4 year:12 nga:1 run:1 inverse:1 exponentiation:1 package:1 uncertainty:1 legislation:1 lope:1 family:3 splicing:1 ric:1 banking:4 capturing:2 ct:1 display:1 kalaitzis:1 jstor:2 identifiable:1 nonnegative:3 activity:1 strength:1 occur:1 ahead:1 constraint:3 idf:1 ri:1 t24:3 markus:1 speed:1 min:2 performing:1 relatively:1 ern:1 department:3 alternate:1 poor:1 conjugate:3 membrane:1 remain:1 ate:1 across:4 suppressed:1 appealing:1 metropolis:1 island:6 lp:3 b:1 making:1 s1:3 outlier:1 restricted:1 pr:4 handling:2 computationally:2 equation:2 visualization:1 resource:1 conjugacy:1 count:9 committee:1 needed:1 ordinal:16 tractable:1 end:1 adopted:1 gaussians:1 apply:1 observe:2 hierarchical:1 appropriate:3 magnetic:1 bhattacharya:1 alternative:1 robustness:1 ho:1 rp:4 existence:1 denotes:6 dirichlet:3 remaining:1 t11:6 include:3 graphical:1 binomial:2 assumes:1 subsampling:6 top:6 restrictive:1 murray:1 lation:1 society:3 unchanged:1 objective:2 noticed:1 spike:2 parametric:7 strategy:1 dependence:5 rt:3 kth:1 deficit:1 armagan:1 columnwise:1 topic:104 cauchy:2 reason:2 water:2 engelhardt:1 itu:1 modeled:1 ffbs:4 providing:2 demonstration:1 innovation:4 nc:1 equivalently:1 dunson:5 difficult:1 unfortunately:2 hmc:1 fe:1 mexico:4 sinica:1 negative:10 polson:1 design:1 implementation:1 policy:2 perform:4 contributed:1 dmin:1 av:1 neuron:1 observation:14 datasets:3 allowing:2 markov:3 finite:1 gas:1 truncated:2 tment:2 situation:1 extended:7 supporting:1 misspecification:3 arbitrary:1 inferred:4 cys:1 required:1 specified:1 security:1 learned:7 boser:1 herein:1 nu:2 nip:1 timeline:1 address:2 beyond:1 able:2 usually:2 below:1 pattern:3 scott:1 lund:1 sparsity:3 challenge:2 panama:2 program:3 t20:6 max:1 including:1 interpretability:1 green:3 bject:1 power:2 event:3 difficulty:1 treated:1 business:2 natural:4 recursion:1 arm:1 normality:1 representing:2 pettitt:1 yp0:3 t18:6 scheme:5 ne:2 auth:1 admixture:6 conventionally:1 jun:1 categorical:1 extract:1 health:3 inouye:1 sn:1 speeding:1 text:4 prior:8 literature:2 understanding:1 review:2 law:3 plant:3 probit:3 highlight:1 winkelmann:1 men:2 mixed:3 allocation:2 proportional:2 limitation:1 filtering:4 shaobo:2 h2:1 olicy:1 s0:2 article:2 viewpoint:1 bank:5 heavy:5 production:1 row:5 arbitrage:1 cancer:1 changed:1 land:2 soil:1 lo:1 genetics:1 course:1 free:4 enjoys:1 t6:6 formal:1 rasmussen:1 understand:1 side:1 institute:2 allow:1 absolute:1 sparse:6 distributed:1 overcome:1 default:1 vocabulary:5 numeric:1 avoids:1 genome:1 world:1 rich:1 forward:4 autoregressive:1 made:1 quantum:1 collection:6 adaptive:1 historical:3 far:1 ec:1 welling:1 employing:1 contemporaneous:1 social:1 approximate:1 cutting:2 gene:3 monotonicity:1 dealing:1 global:2 doucet:1 reveals:1 corpus:2 assumed:2 alternatively:1 bingham:1 latent:13 continuous:3 tailed:4 sk:6 table:3 learn:2 chromosome:1 decoupling:1 improving:1 du:2 expansion:2 t8:6 diag:3 sp:3 pk:1 statistica:1 inherit:1 whole:1 s2:4 arise:1 hyperparameters:1 child:2 allowed:2 toni:1 amino:1 positively:1 body:1 fig:4 site:1 schnatter:1 augmented:2 west:3 egg:1 en:5 n:3 sub:2 inferring:1 exponential:3 candidate:1 pe:3 rk:4 removing:3 formula:1 treasury:1 specific:4 load:1 jt:1 down:1 er:7 decay:1 admits:1 t9:6 evidence:1 burden:1 concern:1 workshop:1 sequential:2 importance:4 te:2 conditioned:1 tio:4 t4:6 sparseness:2 budget:1 chen:1 durham:1 army:3 univariate:2 gao:1 explore:1 sectional:2 labor:1 ordered:4 bo:1 binding:1 springer:3 ch:1 corresponds:1 amt:1 drfm:5 nair:1 conditional:2 lth:1 goal:1 viewed:3 ann:1 towards:1 twofold:1 fisher:1 professor:2 change:1 included:1 specifically:1 infinite:1 wt:2 sampler:1 tumor:2 mexican:4 total:3 e:2 t15:6 meaningful:2 indicating:1 select:1 college:2 transgenic:2 support:4 brevity:4 accelerated:1 violated:1 phenomenon:1 incorporate:2 mcmc:9 avoiding:1 correlated:11 |
4,752 | 5,302 | A provable SVD-based algorithm for learning topics
in dominant admixture corpus
Trapit Bansal?, C. Bhattacharyya??
Department of Computer Science and Automation
Indian Institute of Science
Bangalore -560012, India
[email protected]
[email protected]
Ravindran Kannan
Microsoft Research
India
[email protected]
Abstract
Topic models, such as Latent Dirichlet Allocation (LDA), posit that documents
are drawn from admixtures of distributions over words, known as topics. The
inference problem of recovering topics from such a collection of documents drawn
from admixtures, is NP-hard. Making a strong assumption called separability, [4]
gave the first provable algorithm for inference. For the widely used LDA model,
[6] gave a provable algorithm using clever tensor-methods. But [4, 6] do not learn
topic vectors with bounded l1 error (a natural measure for probability vectors).
Our aim is to develop a model which makes intuitive and empirically supported
assumptions and to design an algorithm with natural, simple components such as
SVD, which provably solves the inference problem for the model with bounded l1
error. A topic in LDA and other models is essentially characterized by a group of
co-occurring words. Motivated by this, we introduce topic specific Catchwords,
a group of words which occur with strictly greater frequency in a topic than any
other topic individually and are required to have high frequency together rather
than individually. A major contribution of the paper is to show that under this
more realistic assumption, which is empirically verified on real corpora, a singular value decomposition (SVD) based algorithm with a crucial pre-processing step
of thresholding, can provably recover the topics from a collection of documents
drawn from Dominant admixtures. Dominant admixtures are convex combination
of distributions in which one distribution has a significantly higher contribution
than the others. Apart from the simplicity of the algorithm, the sample complexity
has near optimal dependence on w0 , the lowest probability that a topic is dominant, and is better than [4]. Empirical evidence shows that on several real world
corpora, both Catchwords and Dominant admixture assumptions hold and the proposed algorithm substantially outperforms the state of the art [5].
1
Introduction
Topic models [1] assume that each document in a text corpus is generated from an ad-mixture of
topics, where, each topic is a distribution over words in a Vocabulary. An admixture is a convex
combination of distributions. Words in the document are then picked in i.i.d. trials, each trial has a
multinomial distribution over words given by the weighted combination of topic distributions. The
problem of inference, recovering the topic distributions from such a collection of documents, is
provably NP-hard. Existing literature pursues techniques such as variational methods [2] or MCMC
procedures [3] for approximating the maximum likelihood estimates.
?
http://mllab.csa.iisc.ernet.in/tsvd
1
Given the intractability of the problem one needs further assumptions on topics to derive polynomial
time algorithms which can provably recover topics. A possible (strong) assumption is that each
document has only one topic but the collection can have many topics. A document with only one
topic is sometimes referred as a pure topic document. [7] proved that a natural algorithm, based
on SVD, recovers topics when each document is pure and in addition, for each topic, there is a set
of words, called primary words, whose total frequency in that topic is close to 1. More recently,
[6] show using tensor methods that if the topic weights have Dirichlet distribution, we can learn
the topic matrix. Note that while this allows non-pure documents, the Dirichlet distribution gives
essentially uncorrelated topic weights.
In an interesting recent development [4, 5] gave the first provable algorithm which can recover topics
from a corpus of documents drawn from admixtures, assuming separability. Topics are said to be
separable if in every topic there exists at least one Anchor word. A word in a topic is said to be an
Anchor word for that topic if it has a high probability in that topic and zero probability in remaining
topics. The requirement of high probability in a topic for a single word is unrealistic.
Our Contributions: Topic distributions, such as those learnt in LDA, try to model the cooccurrence of a group of words which describes a theme. Keeping this in mind we introduce the
notion of Catchwords. A group of words are called Catchwords of a topic, if each word occurs
strictly more frequently in the topic than other topics and together they have high frequency. This
is a much weaker assumption than separability. Furthermore we observe, empirically, that posterior
topic weights assigned by LDA to a document often have the property that one of the weights is
significantly higher than the rest. Motivated by this observation, which has not been exploited by
topic modeling literature, we suggest a new assumption. It is natural to assume that in a text corpus,
a document, even if it has multiple themes, will have an overarching dominant theme. In this paper
we focus on document collections drawn from dominant admixtures. A document collection is said
to be drawn from a dominant admixture if for every document, there is one topic whose weight is
significantly higher than the other topics and in addition, for every topic, there is a small fraction of
documents which are nearly purely on that topic. The main contribution of the paper is to show that
under these assumptions, our algorithm, which we call TSVD, indeed provably finds a good approximation in total l1 error to the topic matrix. We prove a bound on the error of our approximation
which does not grow with dictionary size d, unlike [5] where the error grows linearly with d.
Empirical evidence shows that on semi-synthetic corpora constructed from several real world
datasets, as suggested by [5], TSVD substantially outperforms the state of the art [5]. In particular it is seen that compared to [5] TSVD gives 27% lower error in terms of l1 recovery on 90% of
the topics.
Problem Definition: d, k, s will denote respectively, the number of words in the dictionary, number of topics and number of documents. d, s are large,
P whereas, k is to be thought of as much
smaller. Let Sk = {x = (x1 , x2 , . . . , xk ) : xl ? 0; l xl = 1}. For each topic, there is a fixed
vector in Sk giving the probability of each word in that topic. Let M be the d ? k matrix with these
vectors as its columns.
Documents are picked in i.i.d. trials.
To pick document j, one first picks a k-vector
W1j , W2j , . . . , Wkj of topic weights according to a fixed distribution on Sk . Let P?,j = MW?,j
be the weighted combination of the topic vectors. Then the m words of the document are picked in
i.i.d. trials; each trial picks a word according to the multinomial distribution with P?,j as the probabilities. All that is given as data is the frequency of words in each document, namely, we are given
the d ? s matrix A, where Aij = Number of occurrencesmof word i in Document j . Note that E(A|W) = P,
where, the expectation is taken entry-wise.
In this paper we consider the problem of finding M given A.
2
Previous Results
In this section we review literature related to designing provable algorithms for topic models. For an
overview of topic models we refer the reader to the excellent survey of [1]. Provable algorithms for
recovering topic models was started by [7]. Latent Semantic Indexing (LSI) [8] remains a successful
method for retrieving similar documents by using SVD. [7] showed that one can recover M from a
2
collection of documents, with pure topics, by using SVD based procedure under the additional Primary Words assumption. [6] showed that in the admixture case, if one assumes Dirichlet distribution
for the topic weights, then, indeed, using tensor methods, one can learn M to l2 error provided some
added assumptions on numerical parameters like condition number are satisfied.
The first provably polynomial time algorithm for admixture corpus was given in [4, 5]. For a topic
l, a word i is an anchor word if: Mi,l ? p0 and Mi,l0 = 0 ?l0 6= l.
Theorem 2.1 [4] If every topic has an anchor word, there is a polynomial time algorithm that
? such that with high probability,
returns an M
6
4
k X
d
X
k log d
k
?
|Mil ? Mil | ? d? provided s ? Max O
,
,O
6
4
2
2
a ? p0 ? m
? 2 a2
i=1
l=1
where, ? is the condition number of E(W W T ), a is the minimum expected weight of a topic and m
is the number of words in each document.
Note that the error grows linearly in the dictionary size d, which is often large. Note also the
dependence of s on parameters p0 , which is, 1/p60 and on a, which is 1/a4 . If, say, the word ?run? is
an anchor word for the topic ?baseball? and p0 = 0.1, then the requirement is that every 10 th word
in a document on this topic is ?run?. This seems too strong to be realistic. It would be more realistic
to ask that a set of words like - ?run?, ?hit?, ?score?, etc. together have frequency at least 0.1 which
is what our catchwords assumption does.
3
Learning Topics from Dominant Admixtures
Informally, a document is said to be drawn from a Dominant Admixture if the document has one
dominant topic. Besides its simplicity, we show empirical evidence from real corpora to demonstrate
that topic dominance is a reasonable assumption. The Dominant Topic assumption is weaker than
the Pure Topic assumption. More importantly, SVD based procedures proposed by [7] will not
apply. Inspired by the Primary Words assumption we introduce the assumption that each topic has a
set of Catchwords which individually occur more frequently in that topic than others. This is again
a much weaker assumption than both Primary Words and Anchor Words assumptions and can be
verified experimentally. In this section we establish that by applying SVD on a matrix, obtained by
thresholding the word-document matrix, and subsequent k-means clustering can learn topics having
Catchwords from a Dominant Admixture corpus.
3.1
Assumptions: Catchwords and Dominant admixtures
Let ?, ?, ?, ?, ?0 be non-negative reals satisfying: ? + ? ? (1 ? ?)?,
? + 2? ? 0.5,
? ? 0.08
Dominant topic Assumption (a) For j = 1, 2, . . . , s, document j has a dominant topic l(j) such
that Wl(j),j ? ? and Wl0 j ? ?, ?l0 6= l(j).
(b) For each topic l, there are at least ?0 w0 s documents in each of which topic l has weight at least
1 ? ?.
Catchwords Assumption: There are k disjoint P
sets of words - S1 , S2 , . . . , Sk such that with ?
defined in (5), ?i ? Sl , ?l0 6= l, Mil0 ? ?Mil ,
i?Sl Mil ? p0 ,
20
?i ? Sl , m? 2 ?Mil ? 8 ln
.
(1)
?w0
Part (b) of the Dominant Topic Assumption is in a sense necessary for ?identifiability? - namely for
the model to have a set of k document vectors so that every document vector is in the convex hull of
these vectors. The Catchwords assumption is natural to describe a theme as it tries to model a unique
group of words which is likely to co-occur when a theme is expressed. This assumption is close to
topics discovered by LDA like models, which try to model co-occurence of words. If ?, ? ? ?(1),
then, the assumption (1) says Mil ? ?? (1/m). In fact if Mil ? o(1/m), we do not expect to see
word i (in topic l), so it cannot be called a catchword at all.
3
A slightly different (but equivalent) description of the model will be useful to keep in mind. What
is fixed (not stochastic) are the matrices M and the distribution of the weight matrix W. To pick
document j, we can first pick the dominant topic l in document j and condition the distribution of
W?,j on this being the dominant topic. One could instead also think of W?,j being picked from a
Pk
mixture of k distributions. Then, we let Pij = l=1 Mil Wlj and pick the m words of the document
in i.i.d multinomial trials as before. We will assume that
Tl = {j : l is the dominant topic in document j} satisfies |Tl | = wl s,
where, wl is the probability of topic l being dominant. This is only approximately valid, but the
error is small enough that we can disregard it.
For ? ? {0, 1, 2, . . . , m}, let pi (?, l) be the probability that j ? Tl and Aij = ?/m and qi (?, l) the
corresponding ?empirical probability?:
Z
m
pi (?, l) =
Pij? (1 ? Pij )m?? Prob(W?,j | j ? Tl ) Prob(j ? Tl ), where, P?,j = MW?,j .
?
W?,j
(2)
1
qi (?, l) = |{j ? Tl : Aij = ?/m}| .
(3)
s
Note that pi (?, l) is a real number, whereas, qi (?, l) is a random variable with E(qi (?, l)) = pi (?, l).
We need a technical assumption on the pi (?, l) (which is weaker than unimodality).
No-Local-Min Assumption: We assume that pi (?, l) does not have a local minimum, in the sense:
pi (?, l) > Min(pi (? ? 1, l), pi (? + 1, l)) ? ? ? {1, 2, . . . , m ? 1}.
(4)
The justification for this assumption is two-fold. First, generally, Zipf?s law kind of behaviour where
the number of words plotted against relative frequencies declines as a power function has often been
observed. Such a plot is monotonically decreasing and indeed satisfies our assumption. But for
Catchwords, we do not expect this behaviour - indeed, we expect the curve to go up initially as the
relative frequency increases, then reach a maximum and then decline. This is a unimodal function
and also satisfies our assumption.
Relative sizes of parameters: Before we close this section, a discussion on the values of the parameters is in order. Here, s is large. For asymptotic analysis, we can think of it as going to infinity.
1/w0 is also large and can be thought of as going to infinity. [In fact, if 1/w0 ? O(1), then, intuitively, we see that there is no use of a corpus of more than constant size - since our model has
i.i.d. documents, intuitively, the number of samples we need should depend mainly on 1/w0 ]. m is
(much) smaller, but need not be constant.
c refers to a generic constant independent of m, s, 1/w0 , ?, ?; its value may be different in different
contexts.
3.2
The TSVD Algorithm
Existing SVD based procedures for clustering on raw word-document matrices fail because the
spread of frequencies of a word within a topic is often more (at least not significantly less) than the
gap between the word?s frequencies in two different topics. Hypothetically, the frequency for the
word run, in the topic Sports, may range upwards of 0.01, say. But in other topics, it may range
from, say, 0 to 0.005. The success of the algorithm will lie on correctly identifying the dominant
topics such as sports by identifying that the word run has occurred with high frequency. In this
example, the gap (0.01-0.005) between Sports and other topics is less than the spread within Sports
(1.0-0.01), so a 2-clustering approach (based on SVD) will split the topic Sports into two. While
this is a toy example, note that if we threshold the frequencies at say 0.01, ideally, sports will be all
above and the rest all below the threshold, making the succeeding job of clustering easy.
There are several issues in extending beyond the toy case. Data is not one-dimensional. We will use
different thresholds for each word; word i will have a threshold ?i /m. Also, we have to compute
?i /m. Ideally we would not like to split any Tl , namely, we would like that for each l and and each
i, either most j ? Tl have Aij > ?i /m or most j ? Tl have Aij ? ?i /m. We will show that
4
our threshold procedure indeed achieves this. One other nuance: to avoid conditioning, we split
the data A into two parts A(1) and A(2) , compute the thresholds using A(1) and actually do the
thresholding on A(2) . We will assume that the intial A had 2s columns, so each part now has s
columns. Also, T1 , T2 , . . . , Tk partitions the columns of A(1) as well as those of A(2) . The columns
of thresholded matrix B are then clustered, by a technique we call Project and Cluster, namely,
we project the columns of B to its k?dimensional SVD subspace and cluster in the projection.
The projection before clustering has recently been proven [9] (see also [10]) to yield good starting
cluster centers. The clustering so found is not yet satisfactory. We use the classic Lloyd?s k-means
algorithm proposed by [12]. As we will show, the partition produced after clustering, {R1 , . . . , Rk }
of A(2) is close to the partition induced by the Dominant Topics, {T1 , . . . , Tk }. Catchwords of topic
l are now (approximately) identified as the most frequently occurring words in documents in Rl .
Finally, we identify nearly pure documents in Tl (approximately) as the documents in which the
catchwords occur the most. Then we get an approximation to M?,l by averaging these nearly pure
documents. We now describe the precise algorithm.
3.3
Topic recovery using Thresholded SVD
Threshold SVD based K-means (TSVD)
?
1 ?p0 ?0 ?p0 ?
? , .
? = Min
,
900c20 k 3 m 640m k
(5)
1. Randomly partition the columns of A into two matrices A(1) and A(2) of s columns each.
2. Thresholding
(a) Compute Thresholds on A(1) For each i, let ?i be the highest value of ? ?
(1)
(1)
?
?
{0, 1, 2, . . . , m} such that |{j : Aij > m
}| ? w20 s; |{j : Aij = m
}| ? 3?w0 s.
(?
(2)
?i if Aij > ?i /m and ?i ? 8 ln(20/?w0 )
(b) Do the thresholding on A(2) : Bij =
.
0
otherwise
3. SVD Find the best rank k approximation B(k) to B.
4. Identify Dominant Topics
(a) Project and Cluster Find (approximately) optimal k-means clustering of the columns
of B(k) .
(b) Lloyd?s Algorithm Using the clustering found in Step 4(a) as the starting clustering,
apply Lloyd?s k-means algorithm to the columns of B (B, not B(k) ).
(c) Let R1 , R2 , . . . , Rk be the k?partition of [s] corresponding to the clustering after
Lloyd?s. //*Will prove that Rl ? Tl *//
5. Identify Catchwords
(2)
(a) For each i, l, compute g(i, l) = the (b?0 w0 s/2c)th highest element of {Aij : j ? Rl }.
4
0
, where, ? =
(b) Let Jl = i : g(i, l) > Max m?
2 ln(20/?w0 ), Maxl0 6=l ? g(i, l )
1?2?
.
(1+?)(?+?)
P
(2)
6. Find Topic Vectors Find the b?0 w0 s/2c highest i?Jl Aij among all j ? [s] and return
? ?,l to M?,l .
the average of these A?,j as our approximation M
Theorem 3.1 Main Theorem Under the Dominant Topic, Catchwords and No-Local-Min assump? so that
tions, the algorithm succeeds with high probability in finding an M
6 2
2
X
? il | ? O(k?), provided 1 s ? ?? 1 k m + m k + d
|Mil ? M
.
w0 ?2 p20
?20 ? 2 ?p0
?0 ? 2
i,l
1
The superscript ? hides a logarithmic factor in dsk/?fail , where, ?fail > 0 is the desired upper bound on the
probability of failure.
5
A note on the sample complexity is in order. Notably, the dependence of s on w0 is best possible
(namely s ? ?? (1/w0 )) within logarithmic factors, since, if we had fewer than 1/w0 documents, a
topic which is dominant with probability only w0 may have none of the documents in the collection.
The dependence of s on d needs to be at least d/?0 w0 ? 2 : to see this, note that we only assume
that there are r = O(?0 w0 s) nearly pure documents on each topic. Assuming we ?
can find
? this
set (the algorithm approximately does), their average has standard deviation of about Mil / r in
coordinate i. If topic vector M?,l has O(d) entries, each of?size O(1/d), to get an approximation
of M?,l to l1 error ?, we need the per coordinate error 1/ dr to be at most ?/d which implies
s ? d/?0 w0 ? 2 . Note that to get comparable error in [4], we need a quadratic dependence on d.
There is a long sequence of Lemmas to prove the theorem. To improve the readability of the paper
we relegate the proofs to supplementary material [14]. The essence of the proof lies in proving
that the clustering step correctly identifies the partition induced by the dominant topics. For this,
we take advantage of a recent development on the k?means algorithm from [9] [see also [10]],
where, it is shown that under a condition called the Proximity Condition, Lloyd?s k means algorithm
starting with the centers provided by the SVD-based algorithm, correctly identifies almost all the
documents? dominant topics. We prove that indeed the Proximity Condition holds. This calls for
machinery from Random Matrix theory (in particular bounds on singular values). We prove that the
singular values of the thresholded word-document matrix are nicely bounded. Once the dominant
topic of each document is identified, we are able to find the Catchwords for each topic. Now, we
rely upon part (b.) of the Dominant Topic assumption : that is there is a small fraction of nearly Pure
Topic-documents for each topic. The Catchwords help isolate the nearly pure-topic documents and
hence find the topic vectors. The proofs are complicated by the fact that each step of the algorithm
induces conditioning on the data ? for example, after clustering, the document vectors in one cluster
are not independent anymore.
4
Experimental Results
We compare the thresholded SVD based k-means (TSVD2 ) algorithm 3.3 with the algorithms of
[5], Recover-KL and Recover-L2, using the code made available by the authors3 . We observed
the results of Recover-KL to be better than Recover-L2, and report here the results of Recover-KL
(abbreviated R-KL), full set of results can be found in supplementary section 5. We first provide
empirical support for the algorithm assumptions in Section 3.1, namely the dominant topic and the
catchwords assumption. Then we show on 4 different semi-synthetic data that TSVD provides as
good or better recovery of topics than the Recover algorithms. Finally on real-life datasets, we show
that the algorithm performs as well as [5] in terms of perplexity and topic coherence.
Implementation Details: TSVD parameters (w0 , ?, ?0 , ?) are not known in advance for real corpus. We tested empirically for multiple settings and the following values gave the best performance.
Thresholding parameters used were: w0 = k1 , ? = 61 . For finding the catchwords, ? = 1.1, ?0 = 13
in step 5. For finding the topic vectors (step 6), taking the top 50% (?0 w0 = k1 ) gave empirically
better results. The same values were used on all the datasets tested. The new algorithm is sensitive
to the initialization of the first k-means step in the projected SVD space. To remedy this, we run 10
independent random initializations of the algorithm with K-Means++ [13] and report the best result.
Datasets: We use four real word datasets in the experiments. As pre-processing steps we removed
standard stop-words, selected the vocabulary size by term-frequency and removed documents with
less than 20 words. Datasets used are: (1) NIPS4 : Consists of 1,500 NIPS full papers, vocabulary
of 2,000 words and mean document length 1023. (2) NYT4 : Consists of a random subset of 30,000
documents from the New York Times dataset, vocabulary of 5,000 words and mean document length
238. (3) Pubmed4 : Consists of a random subset of 30,000 documents from the Pubmed abstracts
dataset, vocabulary of 5,030 words and mean document length 58. (4) 20NewsGroup5 (20NG):
Consist of 13,389 documents, vocabulary of 7,118 words and mean document length 160.
2
Resources available at: http://mllab.csa.iisc.ernet.in/tsvd
http://www.cs.nyu.edu/?halpern/files/anchor-word-recovery.zip
4
http://archive.ics.uci.edu/ml/datasets/Bag+of+Words
5
http://qwone.com/?jason/20Newsgroups
3
6
Corpus
s
k
NIPS
NYT
Pubmed
20NG
1500
30000
30000
13389
50
50
50
20
% s with Dominant
Topics (? = 0.4)
56.6%
63.7%
62.2%
74.1%
% s with Pure
Topics (? = 0.05)
2.3%
8.5%
5.1%
39.5%
% Topics
with CW
96%
98%
78%
85%
CW Mean
Frequency
0.05
0.07
0.05
0.06
Table 1: Algorithm Assumptions. For dominant topic assumption, fraction of documents with satisfy
the assumption for (?, ?) = (0.4, 0.3) are shown. % documents with almost pure topics (? = 0.05,
i.e. 95% pure) are also shown. Last two columns show results for catchwords (CW) assumption.
4.1
Algorithm Assumptions
To check the dominant topic and catchwords assumptions, we first run 1000 iterations of Gibbs
sampling on the real corpus and learn the posterior document-topic distribution ({W.,j }) for each
document in the corpus (by averaging over 10 saved-states separated by 50 iterations after the 500
burn-in iterations). We will use this posterior document-topic distribution as the document generating distribution to check the two assumptions.
Dominant topic assumption: Table 1 shows the fraction of the documents in each corpus which
satisfy this assumption with ? = 0.4 (minimum probability of dominant topic) and ? = 0.3 (maximum probability of non-dominant topics). The fraction of documents which have almost pure topics
with highest topic weight at least 0.95 (? = 0.05) is also shown. The results indicate that the dominant topic assumption is well justified (on average 64% documents satisfy the assumption) and there
is also a substantial fraction of documents satisfying almost pure topic assumption.
Catchwords assumption: We first find a k-clustering of the documents {T1 , . . . , Tk } by assigning
all documents which have highest posterior probability for the same topic into one cluster. Then
we use step 5 of TSVD (Algorithm 3.3) to find the set of catchwords for each topic-cluster, i.e.
1
{S1 , . . . , Sk }, with the parameters: 0 w0 = 3k
, ? = 2.3 (taking into account constraints in Section
3.1, ? = 0.4, ? = 0.3, ? = 0.05, ? = 0.07). Table 1 reports the fraction of topics with non-empty
set of catchwords and the average per topic frequency of the catchwords6 . Results indicate that
most topics on real data contain catchwords (Table 1, second-last column). Moreover, the average
per-topic frequency of the group of catchwords for that topic is also quite high (Table 1, last column).
4.2
Empirical Results
Semi-synthetic Data: Following [5], we generate semi-synthetic corpora from LDA model trained
by MCMC, to ensure that the synthetic corpora retain the characteristics of real data. Gibbs sampling7 is run for 1000 iterations on all the four datasets and the final word-topic distribution is used
to generate varying number of synthetic documents with document-topic distribution drawn from a
symmetric Dirichlet with hyper-parameter 0.01. For NIPS, NYT and Pubmed we use k = 50 topics,
for 20NewsGroup k = 20, and mean document lengths of 1000, 300, 100 and 200 respectively. Note
that the synthetic data is not guaranteed to satisfy dominant topic assumption for every document
(on average about 80% documents satisfy the assumption for value of (?, ?) tested in Section 4.1).
? ) for the semiTopic Recovery on Semi-synthetic Data: We learn the word-topic distribution (M
synthetic corpora using TSVD and the Recover algorithms of [5]. Given these learned topic dis? by
tributions and the original data-generating distribution (M ), we align the topics of M and M
bipartite matching and evaluate the l1 distance between each pair of topics. We report the average
of l1 error across topics (called l1 reconstruction-error [5]) in Table 2 for TSVD and Recover-KL
(R-KL). TSVD has smaller error on most datasets than the R-KL algorithm. We observed performance of TSVD to be always better than Recover-L2 (see supplement Table 1 for full results). Best
performance is observed on NIPS which has largest mean document length, indicating that larger
m leads to better recovery. Results on 20NG are slightly worse than R-KL for smaller sample size,
but performance improves for larger number of documents. While the error-values in Table 2 are
6
P
k
1
k
7
1
l=1 |Tl |
P
i?Sl
P
j?Tl
Aij
Dirichlet hyperparameters used: document-topic = 0.03 and topic-word = 1
7
Corpus
NIPS
40
30
NIPS
20
Number of Topics
10
0
0.00 0.15 0.30 0.45 0.60 0.75 0.90 1.05 1.20
NYT
Algorithm
R?KL
TSVD
Pubmed
40
20NG
30
20
10
0
0.00 0.15 0.30 0.45 0.60 0.75 0.90 1.05 1.20
NYT
L1 Reconstruction Error
Figure 1: Histogram of l1 error across
topics (40,000 documents). TSVD(blue,
solid border) gets smaller error on most
topics than R-KL(green, dashed border).
Documents
R-KL
TSVD
40,000
50,000
60,000
40,000
50,000
60,000
40,000
50,000
60,000
40,000
50,000
60,000
0.308
0.308
0.311
0.332
0.326
0.328
0.120
0.114
0.110
0.208
0.206
0.200
0.115 (62.7%)
0.145 (52.9%)
0.131 (57.9%)
0.288 (13.3%)
0.280 (14.1%)
0.284 (13.4%)
0.124 (-3.3%)
0.113 (0.9%)
0.106 (3.6%)
0.195 (6.3%)
0.185 (10.2%)
0.194 (3.0%)
Table 2: l1 reconstruction error on various semi-synthetic
datasets. Brackets in the last column give percent improvement over R-KL (best performing Recover algorithm).
Full results in supplementary.
averaged values across topics, Figure 1 shows that TSVD algorithm achieves much better topic recovery for majority of the topics (>90%) on most datasets (overall average improvement of 27%,
full results in supplement Figure 1).
Topic Recovery on Real Data: To evaluate perplexity [2] on real data, the held-out sets consist of 350 documents for NIPS, 10000 documents for NYT and Pubmed, and 6780 documents for
20NewsGroup. TSVD achieved perplexity measure of 835 (NIPS), 1307 (Pubmed), 1555 (NYT),
2390 (20NG) while Recover-KL achieved 754 (NIPS), 1188 (Pubmed), 1579 (NYT), 2431 (20NG)
(refer to supplement Table 2 for complete results). TSVD gives comparable perplexity with RecoverKL, results being slightly better on NYT and 20NewsGroup which are larger datasets with moderately high mean document lengths. We also find comparable results on Topic Coherence [11] (see
Table 2 in supplementary for topic coherence results and Table 3 for list of top words of topics).
Summary: We evaluated the proposed algorithm, TSVD, rigorously on multiple datasets with respect to the state of the art [5] (Recover-KL and Recover-L2), following the evaluation methodology
of [5]. In Table 2 we show that the l1 reconstruction error for the new algorithm is small and on
average 19.6% better than the best results of the Recover algorithms [5]. In Figure 1, we show that
TSVD achieves significantly better recover on majority of the topics. We also demonstrate that on
real datasets the algorithm achieves comparable perplexity and topic coherence to Recover algorithms. Moreover, we show on multiple real world datasets that the algorithm assumptions are well
justified in practice.
Conclusion
Real world corpora often exhibits the property that in every document there is one topic dominantly
present. A standard SVD based procedure will not be able to detect these topics, however TSVD,
a thresholded SVD based procedure, as suggested in this paper, discovers these topics. While SVD
is time-consuming, there have been a host of recent sampling-based approaches which make SVD
easier to apply to massive corpora which may be distributed among many servers. We believe that
apart from topic recovery, thresholded SVD can be applied even more broadly to similar problems,
such as matrix factorization, and will be the basis for future research.
Acknowledgements TB was supported by a Department of Science and Technology (DST) grant.
References
[1] Blei, D. Introduction to probabilistic topic models. Communications of the ACM, pp. 77?84,
2012.
8
[2] Blei, D., Ng, A., and Jordan, M. Latent Dirichlet allocation. Journal of Machine Learning Research, pp. 3:993?1022, 2003. Preliminary version in Neural Information Processing Systems
2001.
[3] Griffiths, T. L. and Steyvers, M. Finding scientific topics. Proceedings of the National Academy
of Sciences, 101:5228?5235, 2004.
[4] Arora, S., Ge, R., and Moitra, A. Learning topic models ? going beyond SVD. In Foundations
of Computer Science, 2012b.
[5] Arora, S., Ge, R., Halpern, Y., Mimno, D., Moitra, A., Sontag, D., Wu, Y., and Zhu M. A
practical algorithm for topic modeling with provable guarantees. In Internation Conference on
Machine Learning, 2013
[6] Anandkumar, A., Foster, D., Hsu, D., Kakade, S., and Liu, Y. A Spectral Algorithm for Latent
Dirichlet Allocation In Neural Information Processing Systems, 2012.
[7] Papadimitriou, C., Raghavan, P., Tamaki H., and Vempala S. Latent semantic indexing: a probabilistic analysis. Journal of Computer and System Sciences, pp. 217?235, 2000. Preliminary
version in PODS 1998.
[8] Deerwester, S., Dumais, S., Landauer, T., Furnas, G., and Harshman, R. Indexing by latent
semantic analysis. Journal of the American Society for Information Science, pp. 391?407,
1990.
[9] Kumar, A., and Kannan, R. Clustering with spectral norm and the k-means algorithm. In
Foundations of Computer Science, 2010
[10] Awashti, P., and Sheffet, O. Improved spectral-norm bounds for clustering. In Proceedings of
Approx/Random, 2012.
[11] Mimno, D., Wallach, H., Talley, E., Leenders, M. and McCallum, A. Optimizing semantic
coherence in topic models. In Empirical Methods in Natural Language Processing, pp. 262?
272, 2011.
[12] Lloyd, Stuart P. Least squares quantization in PCM, IEEE Transactions on Information Theory
28 (2): 129137,1982.
[13] Arthur, D., and Vassilvitskii, S. K-means++: The advantages of careful seeding. In Proceedings of ACM-SIAM symposium on Discrete algorithms, pp. 1027?1035, 2007
[14] Supplementary material
9
| 5302 |@word trial:6 version:2 polynomial:3 seems:1 norm:2 decomposition:1 p0:8 sheffet:1 pick:6 solid:1 liu:1 score:1 document:89 bhattacharyya:1 outperforms:2 existing:2 com:3 gmail:1 yet:1 assigning:1 realistic:3 numerical:1 subsequent:1 partition:6 seeding:1 plot:1 succeeding:1 fewer:1 selected:1 xk:1 mccallum:1 blei:2 provides:1 readability:1 constructed:1 symposium:1 retrieving:1 prove:5 consists:3 introduce:3 notably:1 expected:1 ravindran:1 indeed:6 frequently:3 inspired:1 decreasing:1 iisc:3 provided:4 bounded:3 project:3 moreover:2 lowest:1 what:2 kind:1 substantially:2 finding:5 guarantee:1 every:8 hit:1 grant:1 harshman:1 before:3 t1:3 local:3 approximately:5 burn:1 initialization:2 wallach:1 co:3 factorization:1 range:2 averaged:1 unique:1 practical:1 trapit:1 practice:1 procedure:7 empirical:7 significantly:5 thought:2 projection:2 matching:1 word:62 pre:2 refers:1 griffith:1 suggest:1 get:4 cannot:1 clever:1 close:4 context:1 applying:1 www:1 equivalent:1 center:2 pursues:1 go:1 overarching:1 starting:3 pod:1 convex:3 survey:1 simplicity:2 recovery:9 identifying:2 pure:15 importantly:1 steyvers:1 classic:1 proving:1 notion:1 coordinate:2 justification:1 massive:1 designing:1 element:1 satisfying:2 tributions:1 observed:4 highest:5 removed:2 leenders:1 substantial:1 complexity:2 moderately:1 cooccurrence:1 ideally:2 rigorously:1 halpern:2 trained:1 depend:1 purely:1 baseball:1 upon:1 bipartite:1 basis:1 various:1 unimodality:1 separated:1 describe:2 w2j:1 hyper:1 whose:2 quite:1 widely:1 supplementary:5 larger:3 say:5 otherwise:1 think:2 superscript:1 final:1 sequence:1 advantage:2 reconstruction:4 uci:1 academy:1 intuitive:1 description:1 cluster:7 requirement:2 extending:1 r1:2 empty:1 generating:2 tk:3 tions:1 derive:1 develop:1 help:1 job:1 strong:3 solves:1 recovering:3 c:1 implies:1 indicate:2 posit:1 saved:1 hull:1 stochastic:1 raghavan:1 material:2 behaviour:2 clustered:1 preliminary:2 strictly:2 hold:2 proximity:2 ic:1 major:1 dictionary:3 achieves:4 a2:1 bag:1 sensitive:1 individually:3 largest:1 wl:3 weighted:2 always:1 aim:1 rather:1 avoid:1 varying:1 mil:10 l0:4 focus:1 improvement:2 rank:1 likelihood:1 mainly:1 check:2 sense:2 detect:1 inference:4 initially:1 going:3 provably:6 issue:1 among:2 overall:1 development:2 art:3 ernet:3 once:1 having:1 nicely:1 ng:7 sampling:2 stuart:1 nearly:6 future:1 papadimitriou:1 np:2 others:2 t2:1 bangalore:1 report:4 randomly:1 national:1 microsoft:2 evaluation:1 mixture:2 bracket:1 held:1 necessary:1 arthur:1 machinery:1 desired:1 plotted:1 column:14 modeling:2 deviation:1 entry:2 intial:1 subset:2 successful:1 too:1 learnt:1 synthetic:10 dumais:1 siam:1 retain:1 probabilistic:2 together:3 again:1 satisfied:1 moitra:2 dr:1 worse:1 american:1 return:2 toy:2 account:1 lloyd:6 automation:1 satisfy:5 ad:1 try:3 picked:4 jason:1 recover:20 complicated:1 identifiability:1 contribution:4 il:1 square:1 characteristic:1 yield:1 identify:3 raw:1 produced:1 none:1 reach:1 definition:1 against:1 failure:1 frequency:17 pp:6 proof:3 mi:2 recovers:1 stop:1 hsu:1 proved:1 dataset:2 ask:1 improves:1 actually:1 higher:3 methodology:1 improved:1 evaluated:1 furthermore:1 wlj:1 lda:7 scientific:1 believe:1 grows:2 contain:1 remedy:1 qwone:1 hence:1 assigned:1 symmetric:1 satisfactory:1 semantic:4 essence:1 bansal:1 complete:1 demonstrate:2 performs:1 l1:12 upwards:1 percent:1 variational:1 wise:1 discovers:1 recently:2 multinomial:3 empirically:5 overview:1 rl:3 conditioning:2 jl:2 occurred:1 refer:2 gibbs:2 zipf:1 approx:1 language:1 had:2 p20:1 etc:1 align:1 dominant:39 posterior:4 recent:3 showed:2 hide:1 optimizing:1 apart:2 perplexity:5 server:1 success:1 life:1 exploited:1 seen:1 minimum:3 greater:1 additional:1 nips4:1 zip:1 monotonically:1 dashed:1 semi:6 multiple:4 unimodal:1 full:5 technical:1 characterized:1 long:1 host:1 qi:4 essentially:2 expectation:1 iteration:4 sometimes:1 histogram:1 achieved:2 justified:2 addition:2 whereas:2 singular:3 grow:1 wkj:1 crucial:1 rest:2 unlike:1 archive:1 file:1 induced:2 isolate:1 jordan:1 call:3 anandkumar:1 near:1 mw:2 split:3 enough:1 easy:1 newsgroups:1 gave:5 identified:2 decline:2 vassilvitskii:1 motivated:2 sontag:1 york:1 dsk:1 useful:1 generally:1 informally:1 induces:1 http:5 generate:2 sl:4 lsi:1 disjoint:1 correctly:3 per:3 blue:1 broadly:1 discrete:1 tsvd:23 group:6 dominance:1 four:2 threshold:8 drawn:8 verified:2 thresholded:6 tamaki:1 nyt:8 fraction:7 deerwester:1 run:8 prob:2 dst:1 almost:4 reader:1 reasonable:1 wu:1 coherence:5 comparable:4 authors3:1 bound:4 guaranteed:1 fold:1 quadratic:1 occur:4 infinity:2 constraint:1 x2:1 min:4 c20:1 kumar:1 performing:1 separable:1 vempala:1 department:2 according:2 combination:4 mllab:2 describes:1 smaller:5 slightly:3 separability:3 across:3 kakade:1 making:2 s1:2 intuitively:2 indexing:3 taken:1 ln:3 resource:1 remains:1 abbreviated:1 fail:3 recoverkl:1 mind:2 ge:2 available:2 apply:3 observe:1 generic:1 spectral:3 anymore:1 original:1 assumes:1 dirichlet:8 remaining:1 clustering:16 top:2 a4:1 ensure:1 giving:1 k1:2 establish:1 approximating:1 w20:1 society:1 tensor:3 added:1 occurs:1 primary:4 dependence:5 said:4 exhibit:1 subspace:1 cw:3 distance:1 majority:2 w0:24 topic:165 provable:7 kannan:3 nuance:1 assuming:2 besides:1 code:1 length:7 negative:1 design:1 implementation:1 wl0:1 upper:1 observation:1 datasets:15 communication:1 precise:1 discovered:1 namely:6 required:1 kl:14 pair:1 learned:1 nip:9 beyond:2 suggested:2 able:2 below:1 tb:1 max:2 green:1 unrealistic:1 power:1 natural:6 rely:1 zhu:1 improve:1 technology:1 identifies:2 arora:2 admixture:16 started:1 occurence:1 text:2 review:1 literature:3 l2:5 acknowledgement:1 relative:3 law:1 asymptotic:1 expect:3 interesting:1 allocation:3 proven:1 foundation:2 pij:3 thresholding:6 foster:1 intractability:1 uncorrelated:1 pi:9 summary:1 supported:2 last:4 keeping:1 dis:1 aij:11 dominantly:1 weaker:4 india:2 institute:1 taking:2 distributed:1 mimno:2 curve:1 vocabulary:6 world:4 valid:1 collection:8 made:1 projected:1 transaction:1 keep:1 ml:1 anchor:7 corpus:22 consuming:1 landauer:1 w1j:1 latent:6 sk:5 table:13 learn:6 csa:3 excellent:1 pk:1 main:2 spread:2 linearly:2 s2:1 border:2 hyperparameters:1 x1:1 referred:1 tl:13 pubmed:7 talley:1 furnas:1 theme:5 xl:2 lie:2 bij:1 theorem:4 rk:2 specific:1 r2:1 nyu:1 list:1 evidence:3 exists:1 consist:2 quantization:1 supplement:3 occurring:2 gap:2 easier:1 logarithmic:2 likely:1 relegate:1 pcm:1 assump:1 expressed:1 sport:6 satisfies:3 acm:2 chiru:1 careful:1 internation:1 hard:2 experimentally:1 averaging:2 lemma:1 called:6 total:2 svd:23 disregard:1 succeeds:1 experimental:1 newsgroup:3 indicating:1 hypothetically:1 support:1 indian:1 evaluate:2 mcmc:2 tested:3 |
4,753 | 5,303 | Learning a Concept Hierarchy from
Multi-labeled Documents
1
Viet-An Nguyen1?, Jordan Boyd-Graber2 , Philip Resnik1,3,4 , Jonathan Chang5
2
5
Computer Science, 3 Linguistics, 4 UMIACS
Computer Science
Facebook
Univ. of Maryland, College Park, MD
Univ. of Colorado, Boulder, CO Menlo Park, CA
[email protected]
Jordan.Boyd.Graber [email protected]
[email protected]
@colorado.edu
Abstract
While topic models can discover patterns of word usage in large corpora, it is
difficult to meld this unsupervised structure with noisy, human-provided labels,
especially when the label space is large. In this paper, we present a model?Label
to Hierarchy (L 2 H)?that can induce a hierarchy of user-generated labels and the
topics associated with those labels from a set of multi-labeled documents. The
model is robust enough to account for missing labels from untrained, disparate
annotators and provide an interpretable summary of an otherwise unwieldy label
set. We show empirically the effectiveness of L 2 H in predicting held-out words
and labels for unseen documents.
1
Understanding Large Text Corpora through Label Annotations
Probabilistic topic models [4] discover the thematic structure of documents from news, blogs, and
web pages. Typical unsupervised topic models such as latent Dirichlet allocation [7, LDA] uncover
topics from unannotated documents. In many settings, however, documents are also associated
with additional data, which provide a foundation for joint models of text with continuous response
variables [6, 48, 27], categorical labels [37, 18, 46, 26] or link structure [9].
This paper focuses on additional information in the form of multi-labeled data, where each document
is tagged with a set of labels. These data are ubiquitous. Web pages are tagged with multiple
directories,1 books are labeled with different categories or political speeches are annotated with
multiple issues.2 Previous topic models on multi-labeled data focus on a small set of relatively
independent labels [25, 36, 46]. Unfortunately, in many real-world examples, the number of labels?
from hundreds to thousands?is incompatible with the independence assumptions of these models.
In this paper, we capture the dependence among the labels using a learned tree-structured hierarchy.
Our proposed model, L 2 H?Label to Hierarchy?learns from label co-occurrence and word usage to
discover a hierarchy of topics associated with user-generated labels. We show empirically that L 2 H
can improve over relevant baselines in predicting words or missing labels in two prediction tasks.
L 2 H is designed to explicitly capture the relationships among labels to discover a highly interpretable
hierarchy from multi-labeled data. This interpretable hierarchy helps improve prediction performance
and also provides an effective way to search, browse and understand multi-labeled data [17, 10, 8, 12].
?
Part of this work was done while the first author interned at Facebook.
Open Directory Project (http://www.dmoz.org/)
2
Policy Agenda Codebook (http://policyagendas.org/)
1
1
2
L 2 H:
Capturing Label Dependencies using a Tree-structured Hierarchy
Discovering a topical hierarchy from text has been the focus of much topic modeling research. One
popular approach is to learn an unsupervised hierarchy of topics. For example, hLDA [5] learns
an unbounded tree-structured hierarchy of topics from unannotated documents. One drawback of
hLDA is that documents only are associated with a single path in the topic tree. Recent work
relaxing this restriction include TSSB [1], nHDP [30], nCRF [2] and S H LDA [27]. Going beyond
tree structure, PAM [20] captures the topic hierarchy using a pre-defined DAG, inspiring more flexible
extensions [19, 24]. However, since only unannotated text is used to infer the hierarchical topics,
it usually requires an additional step of topic labeling to make them interpretable. This difficulty
motivates work leveraging existing taxonomies such as HSLDA [31] and hLLDA [32].
A second active area of research is constructing a taxonomy from multi-labeled data. For example,
Heymann and Garcia-Molina [17] extract a tag hierarchy using the tag network centrality; similar work has been applied to protein hierarchies [42]. Hierarchies of concepts have come from
seeded ontologies [39], crowdsourcing [29], and user-specified relations [33]. More sophisticated
approaches build domain-specific keyword taxonomies with adapting Bayesian Rose Trees [21].
These approaches, however, concentrate on the tags, ignoring the content the tags describe.
In this paper, we combine ideas from these two lines of research and introduce L 2 H, a hierarchical
topic model that discovers a tree-structured hierarchy of concepts from a collection of multi-labeled
documents. L 2 H takes as input a set of D documents {wd }, each tagged with a set of labels ld .
The label set L contains K unique, unstructured labels and the word vocabulary size is V . To
learn an interpretable taxonomy, L 2 H associates each label?a user-generated word/phrase?with a
topic?a multinomial distribution over the vocabulary?to form a concept, and infers a tree-structured
hierarchy to capture the relationships among concepts. Figure 1 shows the plate diagram for L 2 H,
together with its generative process.
?
?
?
??? , ???
??? , ???
??
????,?
??
????,?
????,?
??
??
???
?
???
?
1. Create label graph G and draw a uniform spanning tree T from G (? 2.1)
2. For each node k ? [1, K] in T
(a) If k is the root, draw background topic ?k ? Dir(?u)
(b) Otherwise, draw topic ?k ? Dir(???(k) ) where ?(k) is node k?s
parent.
3. For each document d ? [1, D] having labels ld
(a) Define L0d and L1d using T and ld (cf. ? 2.2)
(b) Draw ?d0 ? Dir(L0d ? ?) and ?d1 ? Dir(L1d ? ?)
(c) Draw a stochastic switching variable ?d ? Beta(?0 , ?1 )
(d) For each token n ? [1, Nd ]
i. Draw set indicator xd,n ? Bern(?d )
x
ii. Draw topic indicator zd,n ? Mult(?d d,n )
iii. Draw word wd,n ? Mult(?zd,n )
Figure 1: Generative process and the plate diagram notation of L 2 H.
2.1
Generating a labeled topic hierarchy
We assume an underlying directed graph G = (E, V) in which each node is a concept consisting of
(1) a label?observable user-generated input, and (2) a topic?latent multinomial distribution over
words.3 The prior weight of a directed edge from node i to node k is the fraction of documents tagged
with label k which are also tagged with label i: ti,k = Di,k /Dj . We also assume an additional
Background node. Edges to the Background node have prior zero weight and edges from the
Background node to node i have prior weight troot,i = Di /maxk Dk . Here, Di is the number of
documents tagged with label i, and Di,k is the number of documents tagged with both labels i and k.
The tree T is a spanning tree generated from G. The probability
of a tree given the graph G is thus
Q
the product of all its edge prior weights p(T | G) = e?E te . To capture the intuition that child
nodes in the hierarchy specialize the concepts of their parents, we model the topic ?k at each node
3
In this paper, we use node when emphasizing the structure discovered by the model. Each node corresponds
to a concept which consists of a label and a topic.
2
k using a Dirichlet distribution whose mean is centered at the topic of node k?s parent ?(k), i.e.,
?k ? Dir(???(k) ). The topic at the root node is drawn from a symmetric Dirichlet ?root ? Dir(?u),
where u is a uniform distribution over the vocabulary [1, 2]. This is similar to the idea of ?backoff? in
language models where more specific contexts inherit the ideas expressed in more general contexts;
i.e., if we talk about ?pedagogy? in education, there?s a high likelihood we?ll also talk about it in
university education [22, 41].
2.2
Generating documents
As in LDA, each word in a document is generated by one of the latent topics. L 2 H, however, also
uses the labels and topic hierarchy to restrict the topics a document uses. The document?s label set ld
identifies which nodes are more likely to be used. Restricting tokens of a document in this way?to
be generated only from a subset of the topics depending the document?s labels?creates specific,
focused, labeled topics [36, Labeled LDA].
Unfortunately, ld is unlikely to be an exhaustive enumeration: particularly when the label set is large,
users often forget or overlook relevant labels. We therefore depend on the learned topology of the
hierarchy to fill in the gaps of what users forget by expanding ld into a broader set, L1d , which is the
union of nodes on the paths from the root node to any of the document?s label nodes. We call this
the document?s candidate set. The candidate set also induces a complementary set L0d ? L \ L1d
(illustrated in Figure 2). Previous approaches such as LPAM [3] and Tree labeled LDA [40] also
leverage the label hierarchy to expand the original label set. However, these previous models require
that the label hierarchy is given rather than inferred as in our L 2 H.
0
1
3
2
4
5
6
Figure 2: Illustration of the candidate label set: given a document d
having labels ld = {2, 4} (double-circled nodes), the candidate label
set of d consists of nodes on all the paths from the root node to node
2 and node 4. L1d = {0, 1, 2, 4} and L0d = {3, 5, 6}. This allows
an imperfect label set to induce topics that the document should be
associated with even if they weren?t explicitly enumerated.
L 2 H replaces Labeled LDA?s absolute restriction to specific topics to a soft preference. To achieve this,
each document d has a switching variable ?d drawn from Beta(?0 , ?1 ), which effectively decides
how likely tokens in d are to be generated from L1d versus L0d . Token n in document d is generated by
first flipping the biased coin ?d to choose the set indicator xd,n . Given xd,n , the label zd,n is drawn
x
from the corresponding label distribution ?d d,n and the token is generated from the corresponding
topic wd,n ? Mult(?zd,n ).
3
Posterior Inference
Given a set of documents with observed words {wd } and labels {ld }, inference finds the posterior
distribution over the latent variables. We use a Markov chain Monte Carlo (MCMC) algorithm to
perform posterior inference, in which each iteration after initialization consists of the following
steps: (1) sample a set indicator xd,n and topic assignment zd,n for each token, (2) sample a word
distribution ?k for each node k in the tree, and (3) update the structure of the label tree.
Initialization: With the large number of labels, the space of hierarchical structures that MCMC
needs to explore is huge. Initializing the tree-structure hierarchy is crucial to help the sampler focus
on more important regions of the search space and help the sampler converge. We initialize the
hierarchy with the maximum a priori probability tree by running Chu-Liu/Edmonds? algorithm to
find the maximum spanning tree on the graph G starting at the background node.
Sampling assignments xd,n and zd,n : For each token, we need to sample whether it was generated
from the label set or not, xd,n . We choose label set i with probability
node in the chosen set i with probability
?d,n
Nd,k
+?
?d,n
Cd,i +?|Lid |
?d,n
Cd,i
+?i
?d,n
Cd,?
+?0 +?1
and we sample a
? ?k,wd,n . Here, Cd,i is the number of times
tokens in document d are assigned to label set i; Nd,k is the number of times tokens in document d
3
are assigned to node k. Marginal counts are denoted by ?, and ?d,n denotes the counts excluding the
assignment of token wd,n .
After we have the label set, we can sample the topic assignment. This is more efficient than sampling
jointly, as most tokens are in the label set, and there are a limited number of topics in the label set.
The probability of assigning node k to zd,n is
p(xd,n = i, zd,n = k | x?d,n , z ?d,n , ?, Lid ) ?
?d,n
Cd,i
+ ?i
?d,n
Cd,?
+ ?0 + ?1
?
?d,n
Nd,k
+?
?d,n
Cd,i
+ ?|Lid |
? ?k,wd,n
(1)
Sampling topics ?: As discussed in Section 2.1, topics on each path in the hierarchy form a
cascaded Dirichlet-multinomial chain where the multinomial ?k at node k is drawn from a Dirichlet
distribution with the mean vector being the topic ??(k) at the parent node ?(k). Given assignments
of tokens to nodes, we need to determine the conditional probability of a word given the token. This
can be done efficiently in two steps: bottom-up smoothing and top-down sampling [2].
? k,v of node k propagated from its children.
? Bottom-up smoothing: This step estimates the counts M
This can be approximated efficiently by using the minimal/maximal path assumption [11, 44]. For
? k,v if Mk0 ,v > 0.
the minimal path assumption, each child node k 0 of k propagates a value of 1 to M
0
? k,v .
For the maximal path assumption, each child node k of k propagates the full count Mk0 ,v to M
? k,v for each node from leaf to root, we sample the word
? Top-down sampling: After estimating M
distributions top-down using its actual counts mk , its children?s propagated counts m
? k and its
parent?s word distribution ??(k) as ?k ? Dir(mk + m
? k + ???(k) ).
Updating tree structure T : We update the tree structure by looping through each non-root
node, proposing a new parent node and either accepting or rejecting the proposed parent using
the Metropolis-Hastings algorithm. More specifically, given a non-root node k with current parent i,
we propose a new parent node j by sampling from all incoming nodes of k in graph G, with probability
proportional to the corresponding edge weights. If the proposed parent node j is a descendant of
k, we reject the proposal
to avoid a cycle.
If it is not a descendant, we accept the proposed move
Q(i?k) P (j?k)
with probability min 1, Q(j?k) P (i?k) , where Q and P denote the proposal distribution and the
model?s joint distribution respectively, and i ? k denotes the case where i is the parent of k.
Since we sample the proposed parent using the edge weights, the proposal probability ratio is
Q(i ? k)
ti,k
=
Q(j ? k)
tj,k
(2)
The joint probability of L 2 H?s observed and latent variables is:
P =
Y
e?E
p(e | G)
D
Y
p(xd | ?)p(zd | xd , ld , ?)p(wd | zd , ?)
d=1
K
Y
p(?l | ??(l) , ?)p(?root | ?)
(3)
l=1
When node k changes its parent from node i to j, the candidate set L1d changes for any document d
that is tagged with any label in the subtree rooted at k. Let 4k denote the subtree rooted at k and
D4k = {d | ?l ? 4k ? l ? ld } the set of documents whose candidate set might change when k?s
parent changes. Canceling unchanged quantities, the ratio of the joint probabilities is:
P (j ? k)
tj,k
=
P (i ? k)
ti,k
K
Y p(zd | j ? k) p(xd | j ? k) p(wd | j ? k) Y
p(?l | j ? k)
p(zd | i ? k) p(xd | i ? k) p(wd | i ? k)
p(?l | i ? k)
d?D4k
(4)
l=1
We now expand each factor in Equation 4. The probability of node assignments zd for document d is
computed by integrating out the document-topic multinomials ?d0 and ?d1 (for the candidate set and its
inverse):
Y
Y ?(Nd,l + ?)
?(?|Lxd |)
p(zd | xd , L0d , L1d ; ?) =
(5)
x
?(Cd,x + ?|Ld |)
?(?)
x
l?Ld
x?{0,1}
4
Similarly, we compute the probability of xd for each document d, integrating out ?d ,
p(xd | ?) =
?(?0 + ?1 )
?(Cd,? + ?0 + ?1 )
Y
x?{0,1}
?(Cd,x + ?i )
?(?x )
(6)
Since we explicitly sample the topic ?l at each node l, we need to re-sample all topics for the case that
QK
l | j?k)
j is the parent of i to compute the ratio l=1 p(?
p(?l | i?k) . Given the sampled ?, the word likelihood is
QNd
p(wd | zd , ?) = n=1
?zd,n ,wd,n .
However, re-sampling the topics for the whole hierarchy for every node proposal is inefficient. To
avoid that, we keep all ??s fixed and approximate the ratio as:
R
K
Y p(wd | j ? k) Y
? k | ?k ) p(?k | ?j ) d?k
p(mk + m
p(?l | j ? k)
?
? R k
(7)
? k | ?k ) p(?k | ?i ) d?k
p(wd | i ? k)
p(?l | i ? k)
p(mk + m
?k
d?D4k
l=1
? k is the word counts propagated from children of k.
where mk is the word counts at node k and m
Since ? is fixed and the node assignments z are unchanged, the word likelihoods cancel out except
for tokens assigned at k or any of its children. The integration in Equation 7 is
Z
? k | ?k ) p(?k | ?j ) d?k =
p(mk + m
?k
V
Y
? k,v + ??i,v )
?(?)
?(Mk,v + M
(8)
? k,? + ?)
?(??i,v )
?(Mk,? + M
v=1
Using Equations 2 and 4, we can compute the Metropolis-Hastings acceptance probability.
4
Experiments: Analyzing Political Agendas in U.S. Congresses
In our experiments, we focus on studying political attention in the legislative process, of interest to
both computer scientists [13, 14] and political scientists [15, 34]. GovTrack provides bill text from
the US Congress, each of which is assigned multiple political issues by the Congressional Research
Service. Examples of Congressional issues include Education, Higher Education, Health, Medicare, etc. To evaluate the effectiveness of L 2 H, we evaluate on two computational tasks: document
modeling?measuring perplexity on a held-out set of documents?and multi-label classification. We
also discuss qualitative results based on the label hierarchy learned by our model.
Data: We use the text and labels from GovTrack for the 109th through 112th Congresses (2005?
2012). For both quantitative tasks, we perform 5-fold cross-validation. For each fold, we perform
standard pre-processing steps on the training set including tokenization, removing stopwords, stemming, adding bigrams, and filtering using TF - IDF to obtain a vocabulary of 10,000 words (final
statistics in Figure 3).4 After building the vocabulary from training documents, we discard all
out-of-vocabulary words in the test documents. We ignore labels associated with fewer than 100 bills.
4.1
Document modeling
In the first quantitative experiment, we focus on the task of predicting the words in held-out test
documents, given their labels. This is measured by perplexity, a widely-used evaluation metric [7, 45].
To compute perplexity, we follow the ?estimating ?? method described in Wallach et al. [45, Sec.
5.1] and split each test document d into wdTE1 and wdTE2 . During training, we estimate all topics?
? During test, first we run Gibbs sampling using the learned topics
distributions over the vocabulary ?.
TE 1
on wd to estimate the topic proportions ??dTEfor each test document d. Then, we compute the
P
TE
? )
log(p(wd 2 | ld ,??dTE ,?)
perplexity on the held-out words wdTE2 as exp ? d
where N TE2 is the total
N TE2
number of tokens in wdTE2 .
4
We find bigram candidates that occur at least ten times in the training set and use a ?2 test to filter out those
having a ?2 value less than 5.0. We then treat selected bigrams as single word types in the vocabulary.
5
Setup We compare our proposed model L 2 H with the following methods:
? LDA [7]: unsupervised topic model with a flat topic structure. In our experiments, we set the
number of topics of LDA equal to the number of labels in each dataset.
? L - LDA [36]: associates each topic with a label, and a document is generated using the topics
associated with the document?s labels only.
? L 2 F (Label to Flat structure): a simplified version of L 2 H with a fixed, flat topic structure. The
major difference between L 2 F and L - LDA is that L 2 F allows tokens to be drawn from topics that are
not in the document?s label set via the use of the switching variable (Section 2.2). Improvements
of L 2 H over L 2 F show the importance of the hierarchical structure.
For all models, the number of topics is the number of labels in the dataset. We run for 1,000 iterations
on the training data with a burn-in period of 500 iterations. After the burn-in period, we store ten
sets of estimated parameters, one after every fifty iterations. During test time, we run ten chains
using these ten learned models on the test data and compute the perplexity after 100 iterations. The
perplexity of each fold is the average value over the ten chains [28].
Number of bills
Congress
Figure 3: Dataset statistics
109
110
250
111
112
?
300
Perplexity
10000
14034 13673 12274
5000 13067
0
109 110 111 112
Number of labels
400
300
200 418
375
243
205
100
0
109 110 111 112
?
LDA
?
?
200
? L?LDA
L2F
L2H
150
Figure 4: Perplexity on held-out documents, averaged over 5 folds
Results: Figure 4 shows the perplexity of the four models averaged over five folds on the four
datasets. LDA outperforms the other models with labels since it can freely optimize the likelihood
without additional constraints. L - LDA and L 2 F are comparable. However, L 2 H significantly outperforms both L - LDA and L 2 F. Thus, if incorporating labels into a model, learning an additional topic
hierarchy improves predictive power and generalizability of L - LDA.
4.2
Multi-label Classification
Multi-label classification is predicting a set of labels for a test document given its text [43, 23, 47].
The prediction is from a set of pre-defined K labels and each document can be tagged with any
of the 2K possible subsets. In this experiment, we use M 3 L?an efficient max-margin multi-label
classifier [16]?to study how features extracted from our L 2 H improve classification.
We use F1 as the evaluation metric. The F1 score is first computed for each document d as F1 (d) =
2?P (d)?R(d)
P (d)+R(d) , where P (d) and R(d) are the precision and recall for document d. After F1 (d) is
computed for all documents, the overall performance can be summarized by micro-averaging and
macro-averaging to obtain Micro-F1 and Macro-F1 respectively. In macro-averaging, F1 is first
computed for each document using its own confusion matrix and then averaged. In micro-averaging,
on the other hand, only a single confusion matrix is computed for all documents, and the F1 score is
computed based on this single confusion matrix [38].
Setup We use the following sets of features:
? TF: Each document is represented by a vector of term frequency of all word types in the vocabulary.
? TF - IDF: Each document is represented by a vector ?dTFIDF of TF - IDF of all word types.
? L - LDA & TF - IDF: Ramage et al. [35] combine L - LDA features and TF - IDF features to improve the
performance on recommendation tasks. Likewise, we extract a K-dimensional vector ??dL - LDA and
combine with TF - IDF vector ?dTFIDF to form the feature vector of L - LDA & TF - IDF.5
5
We run L - LDA on train for 1,000 iterations and ten models after 500 burn-in iterations. For each model, we
sample assignments for all tokens using 100 iterations and average over chains to estimate ??dL - LDA .
6
? L 2 H & TF - IDF: Similarly, we combine TF - IDF with the features ??dL 2 H = {??d0 , ??d1 } extracted using
L 2 H (same MCMC setup as L - LDA ).
One complication for L 2 H is the candidate label set L1d , which is not observed during test time. Thus,
during test time, we estimate L1d using TF - IDF. Let Dl be the set of documents tagged with label
l. For each l, we compute a TF - IDF vector ?lTFIDF = avgd?Dl ?dTFIDF . Then for each document d, we
generate the k nearest labels using cosine similarity, and add them to the candidate label set L1d of
d. Finally, we expand this initial set by adding all labels on the paths from the root of the learned
hierarchy to any of the k nearest labels (Figure 2). We explored different values of k ? {3, 5, 7, 9},
with similar results; the results in this section are reported with k = 5.
109
110
111
112
109
110
0.65
?
?
112
?
?
Micro F1
0.60
Macro F1
111
0.6
0.55
0.50
TF
?
L?LDA & TFIDF
L2H & TFIDF
?
0.5
?
? TFIDF
?
0.45
Figure 5: Multi-label classification results. The results are averaged over 5 folds.
Results Figure 5 shows classification results. For both Macro-F 1 and Micro-F 1, TF - IDF, L LDA & TF - IDF and L 2 H & TF - IDF significantly outperform TF . Also, L - LDA & TF - IDF performs better
than TF - IDF, which is consistent with Ramage et al. (2010) [35].
L 2 H & TF - IDF performs better than L - LDA & TF - IDF, which in turn performs better than TF - IDF. This
shows that features extracted from L 2 H are more predictive than those extracted from L - LDA, and
both improve classification. The improvements of L 2 H & TF - IDF and L - LDA & TF - IDF over TF - IDF are
clearer for Macro-F 1 compared with Micro-F 1. Thus, features from both topic models help improve
prediction, regardless of the frequencies of their tagged labels.
4.3
Learned label hierarchy: A taxonomy of Congressional issues
Terrorism
Military operations and
strategy
intellig, intellig_commun,
afghanistan, nation_intellig,
guantanamo_bai, qaeda,
central_intellig, detent, pakistan,
interrog, defens_intellig, detaine,
armi, air_forc, none, navi,
addit_amount, control_act,
emerg_deficit, fund_appropri,
balanc_budget, terror_pursuant,
transfer_author,marin_corp
International affairs
libya, unit_nation, intern_religi,
bahrain, religi_freedom,
religi_minor, freedom_act, africa,
violenc, secur_council,
benghazi, privileg_resolut, hostil,
Foreign aid and
international relief
International law and
treaties
Human rights
fund_appropri, foreign_assist,
remain_avail, regular_notif,
intern_develop, relat_program,
unit_nation, pakistan,
foreign_oper, usaid, prior_act
foreign_assist, intern_develop,
vessel, foreign_countri, sanit,
appropri_congression,
develop_countri, violenc, girl,
defens_articl, export
traffick, russian_feder,
traffick_victim, prison, alien,
visa, nation_act, victim, detent,
human_traffick, corrupt, russian,
foreign_labor, sex_traffick,
Military personnel and
dependents
Department of Defense
Asia
Int'l organizations &
cooperation
Middle East
coast_guard, vessel, command,
special_select, sexual_violenc,
academi, sexual_harass, navi,
former_offic, gulf_coast, haze,
port, marin, marin_debri
air_forc, militari_construct,
author_act, armi, nation_defens,
navi, militari_depart, aircraft,
congression_defens, command,
sexual_assault, activ_duti
china, vietnam, taiwan, republ,
chines, sea, north_korea,
tibetan, north_korean, refuge,
south_china, intern_religi, tibet,
enterpris, religi_freedom
export, arm_export, control_act,
foreign_assist, cuba,
defens_articl, foreign_countri,
foreign_servic, export_administr,
author_act, munit_list
syria, israel, iran, enterpris_fund,
unit_nation, egypt, palestinian,
cypru, tunisia, hezbollah,
lebanon, republ, hama, syrian,
violenc, weapon,
Armed forces and
national security
Department of
Homeland Security
Europe
Latin America
Sanctions
Religion
cemeteri, nation_guard, dog,
service_memb,
homeless_veteran, funer,
medic_center, militari_servic,
arlington_nation, armi, guard
cybersecur, inform_secur,
inform_system, cover_critic,
critic_infrastructur,
inform_infrastructur,
cybersecur_threat,
republ, belaru, turkei, nato,
holocaust_survivor, north_atlant,
holocaust, european_union,
albania, jew, china, macedonia,
treati_organ, albanian, greec
border_protect, haiti,
merchandis, evas, tariff_act,
cover_merchandis, export,
custom_territori, custom_enforc,,
countervail_duti, intern_trade
iran, sanction, syria,
comprehens_iran, north_korea,
financi_institut, presid_determin,
islam_republ, foreign_person,
weapon, iran_sanction
unit_nation, israel, iaea, harass,
syria, iran, peacekeep_oper,
regular_budget, unrwa,
palestinian, refuge, durban, bulli,
secur_council
Figure 6: A subtree in the hierarchy learned by L 2 H. The subtree root International Affairs is a child
node of the Background root node.
To qualitatively analyze the hierarchy learned by our model, Figure 6 shows a subtree whose root
is about International Affairs, obtained by running L 2 H on bills in the 112th U.S. Congress. The
learned topic at International Affairs shows the focus of 112th Congress on the Arab Spring?a
revolutionary wave of demonstrations and protests in Arab countries like Libya, Bahrain, etc. The
concept is then split into two distinctive aspects of international affairs: Military and Diplomacy.
7
We are working with domain experts to formally evaluate the learned concept hierarchy. A political
scientist (personal communication) comments:
The international affairs topic does an excellent job of capturing the key distinction
between military/defense and diplomacy/aid. Even more impressive is that it then
also captures the major policy areas within each of these issues: the distinction
between traditional military issues and terrorism-related issues, and the distinction
between thematic policy (e.g., human rights) and geographic/regional policy.
5
Conclusion
We have presented L 2 H, a model that discovers not just the interaction between overt labels and
the latent topics used in a corpus, but also how they fit together in a hierarchy. Hierarchies are a
natural way to organize information, and combining labels with a hierarchy provides a mechanism
for integrating user knowledge and data-driven summaries in a single, consistent structure. Our
experiments show that L 2 H yields interpretable label/topic structures, that it can substantially improve
model perplexity compared to baseline approaches, and that it improves performance on a multi-label
prediction task.
Acknowledgments
We thank Kristina Miler, Ke Zhai, Leo Claudino, and He He for helpful discussions, and thank
the anonymous reviewers for insightful comments. This research was supported in part by NSF
under grant #1211153 (Resnik) and #1018625 (Boyd-Graber and Resnik). Any opinions, findings,
conclusions, or recommendations expressed here are those of the authors and do not necessarily
reflect the view of the sponsor.
References
[1] Adams, R., Ghahramani, Z., and Jordan, M. (2010). Tree-structured stick breaking for hierarchical data. In
NIPS.
[2] Ahmed, A., Hong, L., and Smola, A. (2013). The nested Chinese restaurant franchise process: User tracking
and document modeling. In ICML.
[3] Bakalov, A., McCallum, A., Wallach, H., and Mimno, D. (2012). Topic models for taxonomies. In JCDL.
[4] Blei, D. M. (2012). Probabilistic topic models. Communications of the ACM, 55(4):77?84.
[5] Blei, D. M., Griffiths, T. L., Jordan, M. I., and Tenenbaum, J. B. (2003a). Hierarchical topic models and the
nested Chinese restaurant process. In NIPS.
[6] Blei, D. M. and McAuliffe, J. D. (2007). Supervised topic models. In NIPS.
[7] Blei, D. M., Ng, A., and Jordan, M. (2003b). Latent Dirichlet allocation. JMLR, 3.
[8] Bragg, J., Mausam, and Weld, D. S. (2013). Crowdsourcing multi-label classification for taxonomy creation.
In HCOMP.
[9] Chang, J. and Blei, D. M. (2010). Hierarchical relational models for document networks. The Annals of
Applied Statistics, 4(1):124?150.
[10] Chilton, L. B., Little, G., Edge, D., Weld, D. S., and Landay, J. A. (2013). Cascade: Crowdsourcing
taxonomy creation. In CHI.
[11] Cowans, P. J. (2006). Probabilistic Document Modelling. PhD thesis, University of Cambridge.
[12] Deng, J., Russakovsky, O., Krause, J., Bernstein, M. S., Berg, A., and Fei-Fei, L. (2014). Scalable
multi-label annotation. In CHI.
[13] Gerrish, S. and Blei, D. M. (2011). Predicting legislative roll calls from text. In ICML.
[14] Gerrish, S. and Blei, D. M. (2012). How they vote: Issue-adjusted models of legislative behavior. In NIPS.
[15] Grimmer, J. (2010). A Bayesian Hierarchical Topic Model for Political Texts: Measuring Expressed
Agendas in Senate Press Releases. Political Analysis, 18(1):1?35.
[16] Hariharan, B., Vishwanathan, S. V., and Varma, M. (2012). Efficient max-margin multi-label classification
with applications to zero-shot learning. Mach. Learn., 88(1-2):127?155.
8
[17] Heymann, P. and Garcia-Molina, H. (2006). Collaborative creation of communal hierarchical taxonomies
in social tagging systems. Technical Report 2006-10, Stanford InfoLab.
[18] Lacoste-Julien, S., Sha, F., and Jordan, M. I. (2008). DiscLDA: Discriminative learning for dimensionality
reduction and classification. In NIPS, pages 897?904.
[19] Li, W., Blei, D. M., and McCallum, A. (2007). Nonparametric Bayes Pachinko allocation. In UAI.
[20] Li, W. and McCallum, A. (2006). Pachinko allocation: DAG-structured mixture models of topic correlations.
In ICML.
[21] Liu, X., Song, Y., Liu, S., and Wang, H. (2012). Automatic taxonomy construction from keywords. In
KDD.
[22] Mackay, D. J. C. and Peto, L. C. B. (1995). A hierarchical Dirichlet language model. Natural Language
Engineering, 1(3):289?308.
[23] Madjarov, G., Kocev, D., Gjorgjevikj, D., and Deroski, S. (2012). An extensive experimental comparison
of methods for multi-label learning. Pattern Recogn., 45(9):3084?3104.
[24] Mimno, D., Li, W., and McCallum, A. (2007). Mixtures of hierarchical topics with Pachinko allocation. In
ICML.
[25] Mimno, D. M. and McCallum, A. (2008). Topic models conditioned on arbitrary features with Dirichletmultinomial regression. In UAI.
[26] Nguyen, V.-A., Boyd-Graber, J., and Resnik, P. (2012). SITS: A hierarchical nonparametric model using
speaker identity for topic segmentation in multiparty conversations. In ACL.
[27] Nguyen, V.-A., Boyd-Graber, J., and Resnik, P. (2013). Lexical and hierarchical topic regression. In NIPS.
[28] Nguyen, V.-A., Boyd-Graber, J., and Resnik, P. (2014). Sometimes average is best: The importance of
averaging for prediction using MCMC inference in topic modeling. In EMNLP.
[29] Nikolova, S. S., Boyd-Graber, J., and Fellbaum, C. (2011). Collecting Semantic Similarity Ratings to
Connect Concepts in Assistive Communication Tools. Studies in Computational Intelligence. Springer.
[30] Paisley, J. W., Wang, C., Blei, D. M., and Jordan, M. I. (2012). Nested hierarchical Dirichlet processes.
CoRR, abs/1210.6738.
[31] Perotte, A. J., Wood, F., Elhadad, N., and Bartlett, N. (2011). Hierarchically supervised latent Dirichlet
allocation. In NIPS.
[32] Petinot, Y., McKeown, K., and Thadani, K. (2011). A hierarchical model of web summaries. In HLT.
[33] Plangprasopchok, A. and Lerman, K. (2009). Constructing folksonomies from user-specified relations on
Flickr. In WWW.
[34] Quinn, K. M., Monroe, B. L., Colaresi, M., Crespin, M. H., and Radev, D. R. (2010). How to analyze
political attention with minimal assumptions and costs. American Journal of Political Science, 54(1):209?228.
[35] Ramage, D., Dumais, S. T., and Liebling, D. J. (2010). Characterizing microblogs with topic models. In
ICWSM.
[36] Ramage, D., Hall, D., Nallapati, R., and Manning, C. (2009). Labeled LDA: A supervised topic model for
credit attribution in multi-labeled corpora. In EMNLP.
[37] Rosen-Zvi, M., Griffiths, T. L., Steyvers, M., and Smyth, P. (2004). The author-topic model for authors
and documents. In UAI.
[38] Rubin, T. N., Chambers, A., Smyth, P., and Steyvers, M. (2012). Statistical topic models for multi-label
document classification. Mach. Learn., 88(1-2):157?208.
[39] Schmitz, P. (2006). Inducing ontology from Flickr tags. In WWW 2006.
[40] Slutsky, A., Hu, X., and An, Y. (2013). Tree labeled LDA: A hierarchical model for web summaries. In
IEEE International Conference on Big Data, pages 134?140.
[41] Teh, Y. W. (2006). A hierarchical Bayesian language model based on Pitman-Yor processes. In ACL.
[42] Tibely, G., Pollner, P., Vicsek, T., and Palla, G. (2013). Extracting tag hierarchies. PLoS ONE, 8(12):e84133.
[43] Tsoumakas, G., Katakis, I., and Vlahavas, I. P. (2010). Mining multi-label data. In Data Mining and
Knowledge Discovery Handbook.
[44] Wallach, H. M. (2008). Structured Topic Models for Language. PhD thesis, University of Cambridge.
[45] Wallach, H. M., Murray, I., Salakhutdinov, R., and Mimno, D. (2009). Evaluation methods for topic
models. In ICML.
[46] Wang, C., Blei, D., and Fei-Fei, L. (2009). Simultaneous image classification and annotation. In CVPR.
[47] Zhang, M.-L. and Zhou, Z.-H. (2014). A review on multi-label learning algorithms. IEEE TKDE, 26(8).
[48] Zhu, J., Ahmed, A., and Xing, E. P. (2009). MedLDA: maximum margin supervised topic models for
regression and classification. In ICML.
9
| 5303 |@word aircraft:1 version:1 middle:1 bigram:3 proportion:1 nd:5 open:1 hu:1 shot:1 ld:13 reduction:1 initial:1 liu:3 contains:1 score:2 document:63 outperforms:2 existing:1 africa:1 current:1 com:1 wd:16 assigning:1 chu:1 stemming:1 kdd:1 designed:1 interpretable:6 update:2 kristina:1 generative:2 discovering:1 leaf:1 fewer:1 selected:1 intelligence:1 directory:2 mccallum:5 affair:6 accepting:1 blei:10 provides:3 codebook:1 node:51 preference:1 complication:1 org:2 sits:1 zhang:1 five:1 unbounded:1 stopwords:1 guard:1 kocev:1 perotte:1 beta:2 descendant:2 specialize:1 consists:3 qualitative:1 combine:4 cowans:1 introduce:1 tagging:1 behavior:1 ontology:2 prison:1 multi:22 chi:2 salakhutdinov:1 palla:1 actual:1 enumeration:1 armed:1 little:1 l2f:1 provided:1 discover:4 project:1 notation:1 underlying:1 estimating:2 katakis:1 what:1 israel:2 substantially:1 proposing:1 finding:1 l2h:2 quantitative:2 every:2 collecting:1 ti:3 xd:14 classifier:1 stick:1 grant:1 organize:1 mcauliffe:1 service:1 scientist:3 engineering:1 treat:1 congress:6 switching:3 marin:1 mach:2 analyzing:1 path:8 might:1 pam:1 burn:3 initialization:2 terrorism:2 china:2 wallach:4 acl:2 relaxing:1 co:2 limited:1 averaged:4 directed:2 unique:1 qaeda:1 acknowledgment:1 union:1 gjorgjevikj:1 area:2 adapting:1 mult:3 boyd:7 reject:1 word:25 induce:2 pre:3 integrating:3 significantly:2 protein:1 griffith:2 qnd:1 context:2 www:3 restriction:2 hllda:1 bill:4 missing:2 optimize:1 reviewer:1 attribution:1 attention:2 starting:1 regardless:1 lexical:1 focused:1 ke:1 unstructured:1 fill:1 varma:1 steyvers:2 annals:1 hierarchy:39 construction:1 colorado:2 user:10 smyth:2 us:2 associate:2 approximated:1 particularly:1 updating:1 afghanistan:1 labeled:17 observed:3 bottom:2 export:3 initializing:1 capture:6 wang:3 thousand:1 tunisia:1 region:1 te2:2 news:1 cycle:1 keyword:1 plo:1 rose:1 intuition:1 personal:1 depend:1 predictive:2 creation:3 creates:1 distinctive:1 treaty:1 girl:1 joint:4 represented:2 america:1 talk:2 recogn:1 assistive:1 leo:1 train:1 univ:2 effective:1 describe:1 monte:1 labeling:1 sanction:2 exhaustive:1 whose:3 victim:1 widely:1 stanford:1 cvpr:1 otherwise:2 statistic:3 unseen:1 jointly:1 noisy:1 final:1 jcdl:1 mausam:1 propose:1 interaction:1 product:1 maximal:2 hslda:1 macro:6 relevant:2 bahrain:2 combining:1 achieve:1 cuba:1 inducing:1 parent:15 double:1 sea:1 generating:2 adam:1 franchise:1 mckeown:1 help:4 depending:1 clearer:1 measured:1 nearest:2 keywords:1 job:1 c:1 come:1 concentrate:1 drawback:1 annotated:1 filter:1 stochastic:1 centered:1 human:3 opinion:1 tsoumakas:1 education:4 require:1 f1:10 ncrf:1 weren:1 anonymous:1 tfidf:3 enumerated:1 adjusted:1 extension:1 hall:1 credit:1 exp:1 major:2 overt:1 label:96 create:1 tf:25 tool:1 schmitz:1 rather:1 avoid:2 zhou:1 command:2 broader:1 release:1 focus:7 improvement:2 modelling:1 likelihood:4 alien:1 political:10 baseline:2 helpful:1 inference:4 dependent:1 foreign:1 unlikely:1 accept:1 relation:2 expand:3 going:1 issue:8 among:3 flexible:1 classification:13 denoted:1 priori:1 overall:1 smoothing:2 integration:1 initialize:1 mackay:1 tokenization:1 marginal:1 equal:1 having:3 ng:1 sampling:8 park:2 unsupervised:4 cancel:1 icml:6 rosen:1 report:1 micro:6 national:1 consisting:1 relief:1 ab:1 organization:1 huge:1 acceptance:1 interest:1 mining:2 highly:1 evaluation:3 mixture:2 tj:2 held:5 chain:5 edge:7 tree:22 re:2 minimal:3 mk:8 military:5 modeling:5 soft:1 measuring:2 assignment:8 phrase:1 cost:1 nhdp:1 subset:2 hundred:1 uniform:2 mk0:2 zvi:1 avgd:1 reported:1 dependency:1 connect:1 dir:7 generalizability:1 dumais:1 international:9 probabilistic:3 together:2 thesis:2 reflect:1 choose:2 emnlp:2 book:1 expert:1 inefficient:1 american:1 li:3 account:1 sec:1 summarized:1 int:1 explicitly:3 unannotated:3 root:13 view:1 analyze:2 wave:1 bayes:1 xing:1 annotation:3 collaborative:1 hariharan:1 roll:1 qk:1 efficiently:2 likewise:1 yield:1 infolab:1 bayesian:3 l1d:11 rejecting:1 overlook:1 none:1 carlo:1 russakovsky:1 simultaneous:1 flickr:2 canceling:1 hlt:1 facebook:2 frequency:2 associated:7 di:4 propagated:3 sampled:1 dataset:3 popular:1 recall:1 knowledge:2 arab:2 infers:1 ubiquitous:1 improves:2 dimensionality:1 segmentation:1 conversation:1 uncover:1 sophisticated:1 fellbaum:1 higher:1 supervised:4 follow:1 asia:1 response:1 hama:1 diplomacy:2 done:2 just:1 smola:1 correlation:1 hand:1 hastings:2 working:1 web:4 radev:1 lda:31 russian:1 usage:2 building:1 concept:11 geographic:1 ramage:4 tagged:11 seeded:1 assigned:4 symmetric:1 semantic:1 illustrated:1 ll:1 during:5 rooted:2 speaker:1 cosine:1 hong:1 plate:2 confusion:3 performs:3 egypt:1 image:1 meld:1 discovers:2 multinomial:5 empirically:2 discussed:1 he:2 cambridge:2 gibbs:1 dag:2 chine:1 paisley:1 automatic:1 libya:2 similarly:2 language:5 dj:1 europe:1 similarity:2 impressive:1 etc:2 add:1 posterior:3 own:1 recent:1 driven:1 perplexity:10 discard:1 store:1 browse:1 blog:1 palestinian:2 molina:2 additional:6 deng:1 freely:1 converge:1 determine:1 period:2 ii:1 multiple:3 full:1 infer:1 d0:3 hcomp:1 legislative:3 technical:1 ahmed:2 cross:1 vicsek:1 sponsor:1 prediction:6 scalable:1 regression:3 metric:2 iteration:8 sometimes:1 dirichletmultinomial:1 proposal:4 background:6 krause:1 diagram:2 country:1 crucial:1 biased:1 fifty:1 umiacs:1 umd:2 weapon:2 regional:1 comment:2 leveraging:1 effectiveness:2 jordan:7 call:2 extracting:1 leverage:1 latin:1 iii:1 enough:1 congressional:3 split:2 bernstein:1 independence:1 fit:1 restaurant:2 nikolova:1 restrict:1 topology:1 imperfect:1 idea:3 whether:1 defense:2 bartlett:1 song:1 speech:1 backoff:1 nonparametric:2 iran:3 ten:6 tenenbaum:1 inspiring:1 induces:1 category:1 http:2 generate:1 outperform:1 nsf:1 estimated:1 tkde:1 zd:16 edmonds:1 medlda:1 key:1 four:2 elhadad:1 drawn:5 troot:1 lacoste:1 graph:5 fraction:1 wood:1 run:4 inverse:1 revolutionary:1 multiparty:1 draw:8 incompatible:1 comparable:1 disclda:1 capturing:2 vietnam:1 fold:6 replaces:1 slutsky:1 occur:1 constraint:1 idf:22 looping:1 fei:4 vishwanathan:1 flat:3 tag:6 govtrack:2 aspect:1 weld:2 min:1 spring:1 jew:1 relatively:1 structured:8 department:2 manning:1 protest:1 visa:1 metropolis:2 lid:3 folksonomies:1 boulder:1 intellig:1 haze:1 equation:3 discus:1 count:8 turn:1 mechanism:1 studying:1 operation:1 pakistan:2 hierarchical:17 quinn:1 chamber:1 occurrence:1 vlahavas:1 centrality:1 coin:1 original:1 denotes:2 dirichlet:9 linguistics:1 include:2 cf:1 running:2 top:3 ghahramani:1 especially:1 build:1 chinese:2 murray:1 unchanged:2 move:1 quantity:1 flipping:1 strategy:1 sha:1 dependence:1 md:1 traditional:1 link:1 maryland:1 thank:2 philip:1 topic:78 albanian:1 armi:3 dmoz:1 spanning:3 thadani:1 taiwan:1 relationship:2 illustration:1 ratio:4 demonstration:1 zhai:1 difficult:1 unfortunately:2 setup:3 taxonomy:10 disparate:1 agenda:3 motivates:1 policy:4 perform:3 teh:1 markov:1 datasets:1 maxk:1 relational:1 excluding:1 communication:3 peto:1 topical:1 discovered:1 arbitrary:1 grimmer:1 lxd:1 inferred:1 rating:1 dog:1 specified:2 extensive:1 security:2 homeland:1 learned:11 distinction:3 nip:7 beyond:1 usually:1 pattern:2 including:1 max:2 power:1 difficulty:1 force:1 natural:2 predicting:5 indicator:4 cascaded:1 senate:1 medicare:1 zhu:1 improve:7 julien:1 identifies:1 categorical:1 extract:2 health:1 text:9 prior:4 understanding:1 circled:1 discovery:1 review:1 law:1 allocation:6 proportional:1 filtering:1 versus:1 annotator:1 validation:1 foundation:1 consistent:2 propagates:2 port:1 rubin:1 corrupt:1 cd:10 summary:4 token:17 cooperation:1 supported:1 bern:1 viet:1 understand:1 characterizing:1 absolute:1 pitman:1 yor:1 mimno:4 vocabulary:9 world:1 fb:1 pachinko:3 author:4 collection:1 qualitatively:1 simplified:1 nguyen:3 social:1 lebanon:1 approximate:1 observable:1 ignore:1 nato:1 keep:1 active:1 decides:1 incoming:1 uai:3 corpus:4 handbook:1 discriminative:1 hlda:2 continuous:1 latent:8 search:2 heymann:2 learn:4 robust:1 ca:1 expanding:1 ignoring:1 menlo:1 dte:2 vessel:2 untrained:1 excellent:1 necessarily:1 constructing:2 domain:2 inherit:1 petinot:1 hierarchically:1 whole:1 cascade:1 big:1 nallapati:1 child:8 complementary:1 graber:6 syrian:1 pedagogy:1 syria:3 resnik:6 aid:2 precision:1 thematic:2 candidate:10 breaking:1 jmlr:1 learns:2 unwieldy:1 emphasizing:1 down:3 removing:1 specific:4 insightful:1 explored:1 dk:1 dl:5 incorporating:1 restricting:1 adding:2 effectively:1 importance:2 corr:1 phd:2 te:3 subtree:5 conditioned:1 margin:3 gap:1 monroe:1 forget:2 garcia:2 likely:2 explore:1 expressed:3 personnel:1 religion:1 tracking:1 recommendation:2 chang:1 springer:1 corresponds:1 nested:3 gerrish:2 extracted:4 acm:1 conditional:1 identity:1 content:1 change:4 typical:1 tssb:1 specifically:1 except:1 sampler:2 averaging:5 total:1 experimental:1 lerman:1 vote:1 east:1 formally:1 college:1 berg:1 icwsm:1 jonathan:1 evaluate:3 mcmc:4 d1:3 crowdsourcing:3 |
4,754 | 5,304 | On a Theory of Nonparametric Pairwise Similarity
for Clustering: Connecting Clustering to
Classification
Yingzhen Yang1 Feng Liang1 Shuicheng Yan2 Zhangyang Wang1 Thomas S. Huang1
1
University of Illinois at Urbana-Champaign, Urbana, IL 61801, USA
{yyang58,liangf,zwang119,t-huang1}@illinois.edu
2
National University of Singapore, Singapore, 117576
[email protected]
Abstract
Pairwise clustering methods partition the data space into clusters by the pairwise
similarity between data points. The success of pairwise clustering largely depends on the pairwise similarity function defined over the data points, where kernel similarity is broadly used. In this paper, we present a novel pairwise clustering
framework by bridging the gap between clustering and multi-class classification.
This pairwise clustering framework learns an unsupervised nonparametric classifier from each data partition, and search for the optimal partition of the data by
minimizing the generalization error of the learned classifiers associated with the
data partitions. We consider two nonparametric classifiers in this framework, i.e.
the nearest neighbor classifier and the plug-in classifier. Modeling the underlying data distribution by nonparametric kernel density estimation, the generalization error bounds for both unsupervised nonparametric classifiers are the sum of
nonparametric pairwise similarity terms between the data points for the purpose
of clustering. Under uniform distribution, the nonparametric similarity terms induced by both unsupervised classifiers exhibit a well known form of kernel similarity. We also prove that the generalization error bound for the unsupervised plugin classifier is asymptotically equal to the weighted volume of cluster boundary
[1] for Low Density Separation, a widely used criteria for semi-supervised learning and clustering. Based on the derived nonparametric pairwise similarity using
the plug-in classifier, we propose a new nonparametric exemplar-based clustering
method with enhanced discriminative capability, whose superiority is evidenced
by the experimental results.
1
Introduction
Pairwise clustering methods partition the data into a set of self-similar clusters based on the pairwise similarity between the data points. Representative clustering methods include K-means [2]
which minimizes the within-cluster dissimilarities, spectral clustering [3] which identifies clusters
of more complex shapes lying on low dimensional manifolds, and the pairwise clustering method
[4] using message-passing algorithm to inference the cluster labels in a pairwise undirected graphical model. Utilizing pairwise similarity, these pairwise clustering methods often avoid estimating
complex hidden variables or parameters, which is difficult for high dimensional data.
However, most pairwise clustering methods assume that the pairwise similarity is given [2, 3], or
they learn a more complicated similarity measure based on several given base similarities [4]. In
this paper, we present a new framework for pairwise clustering where the pairwise similarity is
derived as the generalization error bound for the unsupervised nonparametric classifier. The un1
supervised classifier is learned from unlabeled data and the hypothetical labeling. The quality of
the hypothetical labeling is measured by the associated generalization error of the learned classifier, and the hypothetical labeling with minimum associated generalization error bound is preferred.
We consider two nonparametric classifiers, i.e. the nearest neighbor classifier (NN) and the plug-in
classifier (or the kernel density classifier). The generalization error bounds for both unsupervised
classifiers are expressed as sum of pairwise terms between the data points, which can be interpreted
as nonparametric pairwise similarity measure between the data points. Under uniform distribution,
both nonparametric similarity measures exhibit a well known form of kernel similarity. We also
prove that the generalization error bound for the unsupervised plug-in classifier is asymptotically
equal to the weighted volume of cluster boundary [1] for Low Density Separation, a widely used
criteria for semi-supervised learning and clustering.
Our work is closely related to discriminative clustering methods by unsupervised classification,
which search for the cluster boundaries with the help of unsupervised classifier. For example, [5]
learns a max-margin two-class classifier to group unlabeled data in an unsupervised manner, known
as unsupervised SVM whose theoretical property is further analyzed in [6]. Also, [7] learns the
kernel logistic regression classifier, and uses the entropy of the posterior distribution of the class
label by the classifier to measure the quality of the learned classifier. More recent work presented in
[8] learns an unsupervised classifier by maximizing the mutual information between cluster labels
and the data, and the Squared-Loss Mutual Information is employed to produce a convex optimization problem. Although such discriminative methods produce satisfactory empirical results, the
optimization of complex parameters hampers their application in high-dimensional data. Following
the same principle of unsupervised classification using nonparametric classifiers, we derive nonparametric pairwise similarity and eliminate the need of estimating complicated parameters of the
unsupervised classifer. As an application, we develop a new nonparametric exemplar-based clustering method with the derived nonparametric pairwise similarity induced by the plug-in classifier, and
our new method demonstrates better empirical clustering results than the existing exemplar-based
clustering methods.
It should be emphasized that our generalization bounds are essentially different from the literature. As nonparametric classification methods, the generalization properties of the nearest neighbor
classifier (NN) and the plug-in classifier are extensively studied. Previous research focuses on the
average generalization error of the NN [9, 10], which is the average error of the NN over all the
random training data sets, or the excess risk of the plug-in classifier [11, 12]. In [9], it is shown that
the average generalization error of the NN is bounded by twice of the Bayes error. Assuming that
the class of the regression functions has a smooth parameter ?, [11] proves that the excess risk of
??
the plug-in classifier converges to 0 of the order n 2?+d where d is the dimension of the data. [12]
further shows that the plug-in classifier attains faster convergence rate of the excess risk, namely
1
n? 2 , under some margin assumption on the data distribution. All these generalization error bounds
depend on the unknown Bayes error. By virtue of kernel density estimation and generalized kernel density estimation [13], our generalization bounds are represented mostly in terms of the data,
leading to the pairwise similarities for clustering.
2
Formulation of Pairwise Clustering by Unsupervised Nonparametric
Classification
The discriminative clustering literature [5, 7] has demonstrated the potential of multi-class classification for the clustering problem. Inspired by the natural connection between clustering and
classification, we model the clustering problem as a multi-class classification problem: a classifier
is learned from the training data built by a hypothetical labeling, which is a possible cluster labeling.
The optimal hypothetical labeling is supposed to be the one such that its associated classifier has the
minimum generalization error bound. To study the generalization bound for the classifier learned
from the hypothetical labeling, we define the concept of classification model. Given unlabeled data
{xl }nl=1 , a classification model MY is constructed for any hypothetical labeling Y = {yl }nl=1 as
below:
Definition 1. The classification
model corresponding
to the hypothetical labeling Y = {yl }nl=1
(
)
Q
is defined as MY = S, PXY , {?i , fi }i=1 , F . S = {xl , yl }nl=1 are the labeled data by the
2
hypothetical labeling, and S are assumed to be i.i.d. samples drawn from the joint distribution PXY = PX|Y PY , where (X, Y ) is a random couple, X ? IRd represents the data and
Y ? {1, 2, ..., Q} is the class label of X, Q is the number of classes determined by the hypothetical
(i)
labeling. Furthermore, PXY is specified by {? (i) , f (i) }Q
is the class prior for
i=1 as follows: ?
(i)
class i, i.e. Pr [Y = i] = ? ; the conditional distribution PX|Y =i has probabilistic density function f (i) , i = 1, . . . , Q. F is a classifier trained using the training data S. The generalization error
of the classification model MY is defined as the generalization error of the classifier F in MY .
In this paper, we study two types of classification models with the nearest neighbor classifier and
the plug-in classifier respectively, and derive their generalization error bounds as sum of pairwise
similarity between the data. Given a specific type of classification model, the optimal hypothetical
labeling corresponds to the classification model with minimum generalization error bound. The
optimal hypothetical labeling also generates a data partition where the sum of pairwise similarity
between the data from different clusters is minimized, which is a common criteria for discriminative
clustering.
In the following text, we derive the generalization error bounds for the two types of classification
models. Before that, we introduce more notations and assumptions for the classification model.
Denote by PX the induced marginal distribution of X, and f is the probabilistic density function of
Q
?
PX which is a mixture of Q class-conditional densities: f =
? (i) f (i) . ? (i) (x) is the regression
i=1
(i)
(i)
function of Y on X = x, i.e. ? (i) (x) = Pr [Y = i |X = x ] = ? ff(x)(x) . For the sake of the
consistency of the kernel density estimators used in the sequel, there are further assumptions on
the marginal density and class-conditional densities in the classification model for any hypothetical
labeling:
(A) f is bounded from below, i.e. f ? fmin > 0
(i)
(B) {f (i) } is bounded from above, i.e. f (i) ? fmax , and f (i) ? ??,ci , 1 ? i ? Q.
where ??,c is the class of H?older-? smooth functions with H?older constant c:
??,c , {f : IRd ? IR | ?x, y, |f (x) ? f (y)| ? c?x ? y?? }, ? > 0
It follows from assumption (B) that f ? ??,c where c =
?
? (i) ci . Assumption (A) and (B) are
i
mild. The upper bound for the density functions is widely required for the consistency of kernel
density estimators [14, 15]; H?older-? smoothness is required to bound the bias of such estimators,
and it also appears in [12] for estimating the excess risk of the plug-in classifier. The lower bound
for the marginal density is used to derive the consistency of the estimator of the regression function
? (i) (Lemma 2) and the consistency of the generalized kernel density estimator (Lemma 3). We
denote by PX the collection of marginal distributions that satisfy assumption (A), and denote by
PX|Y the collection of class-conditional distributions that satisfy assumption (B). We then define
the collection of joint distributions PXY that PXY belongs to, which requires the marginal density
and class-conditional densities satisfy assumption (A)-(B):
PXY , {PXY | PX ? PX , {PX|Y =i } ? PX|Y , min{? (i) } > 0}
i
(1)
Given the joint distribution PXY , the generalization error of the classifier F learned from the training data S is:
R (FS ) , Pr [(X, Y ) : F (X) ?= Y ]
(2)
Nonparametric kernel density estimator (KDE) serves as the primary tool of estimating the underlying probabilistic density functions in our generalization analysis, and we introduce the KDE of f
as below:
n
1?
Khn (x ? xl )
f?n,hn (x) =
n
where Kh (x) =
1
e
(2?)d/2
?x?2
? 2
1
K
hd
(x)
h
(3)
l=1
is the isotropic Gaussian kernel with bandwidth h and K (x) ,
. We have the following VC property of the Gaussian kernel K. Define the class
3
of functions
(
F , {K
t??
h
)
, t ? IRd , h ?= 0}
(4)
The VC property appears in [14, 15, 16, 17, 18], and it is proved that F is a bounded VC class of
measurable functions with respect to the envelope function F such that |u| ? F for any u ? F (e.g.
d
F ? (2?)? 2 ). It follows that? there exist positive numbers A and v such that for every probability
measure P on IRd for which F 2 dP < ? and any 0 < ? < 1,
(
) ( A )v
N F , ???L2 (P ) , ? ?F ?L2 (P ) ?
?
(5)
(
)
? ? is defined as the minimal number of open d-balls
?
where N T , d,
of radius ? required to cover
(
)
T in the metric space T , d? . A and v are called the VC characteristics of F.
The VC property of K is required for the consistency of kernel density estimators shown in
Lemma 2. Also, we adopt the kernel estimator of ? (i) below
n
?
(i)
??n,hn (x) =
l=1
Khn (x ? xl )1I{yl =i}
nf?n,hn (x)
(6)
Before stating Lemma 2, we introduce several frequently used quantities throughout this paper. Let
L, C > 0 be constants which only depend on the VC characteristics of the Gaussian kernel K. We
define
f0 ,
Q
?
(i)
? (i) fmax
?02 , ?K?22 f0
(7)
i=1
Also, for all positive numbers ? ? C and ? > 0, we define
E?2 ,
log (1 + ?/4L)
?L? 2
(8)
Based on Corollary 2.2 in [14], Lemma 2 and Lemma 3 in the Appendix (more complete version
in the supplementary) show the strong consistency (almost sure uniformly convergence) of several
(i)
kernel density estimators, i.e. f?n,hn , {?
?n,hn } and the generalized kernel density estimator, and they
form the basis for the derivation of the generalization error bounds for the two types of classification
models.
3
Generalization Bounds
We derive the generalization error bounds for the two types of classification models with the nearest
neighbor classifier and the plug-in classifier respectively. Substituting these kernel density estimators for the corresponding true density functions, Theorem 1 and Theorem 2 present the generalization error bounds for the classification models with the plug-in classifier and the nearest neighbor
classifier. The dominant terms of both bounds are expressed as sum of pairwise similarity depending solely on the data, which facilitates the application of clustering. We also show the connection
between the error bound for the plug-in classifier and Low Density Separation in this section. The
detailed proofs are included in the supplementary.
3.1
Generalization Bound for the Classification Model with Plug-In Classifier
The plug-in classifier resembles the Bayes classifier while it uses the kernel density estimator of the
regression function ? (i) instead of the true ? (i) . It has the form
(i)
PI (X) = arg max ??n,hn (X)
(9)
1?i?Q
(i)
where ??n,hn is the nonparametric kernel estimator of the regression function ? (i) by (6). The
generalization capability of the plug-in classifier has been studied by the literature[11, 12]. Let
4
F ? be the Bayes classifier, it is proved that the excess risk of PIS , namely IES R (PIS ) ? R (F ? ),
??
converges to 0 of the order n 2?+d under some complexity assumption on the class of the regression
functions with smooth parameter ? that {? (i) } belongs to [11, 12]. However, this result cannot be
used to derive the generalization error bound for the plug-in classifier comprising of nonparametric
pairwise similarities in our setting.
We show the upper bound for the generalization error of PIS in Lemma 1.
Lemma 1. For any PXY ? PXY , there exists a n0 which depends on ?0 and VC characteristics
E? 2
of K, when n > n0 , with probability greater than 1 ? 2QLhn
plug-in classifier satisfies
R (PIS ) ?
PI
Rn
=
PI
Rn
+O
?
?
( log h?1
n
nhdn
0
+ h?n
, the generalization error of the
)
(10)
[
]
(i)
(j)
IEX ??n,hn (X) ??n,hn (X)
(11)
log h?1
n
nhd
n
? 0, ??n,hn is the kernel
i,j=1,...,Q,i?=j
where E?2 is defined by (8), hn is chosen such that hn ? 0,
(i)
(i)
estimator of the regression function. Moreover, the equality in (10) holds when ??n,hn ?
1 ? i ? Q.
1
Q
for
Based on Lemma 1, we can bound the error of the plug-in classifier from above by RnPI . Theorem 1
then gives the bound for the error of the plug-in classifier in the corresponding classification model
using the generalized kernel density estimator in Lemma 3. The bound has a form of sum of pairwise
similarity between the data from different classes.
Theorem
1. (Error of) the Plug-In Classifier) Given the classification model MY =
(
S, PXY , {?i , fi }Q
i=1 , PI such that PXY ? PXY , there exists a n1 which depends on ?0 , ?1 and
E? 2
the VC characteristics of K, when n > n1 , with probability greater than 1 ? 2QLhn
the generalization error of the plug-in classifier satisfies
? n (PIS ) + O
R (PIS ) ? R
? n (PIS ) =
where R
1
n2
?
l,m
(
?
)
log h?1
n
+ h?n
d
nhn
?K?22 fmax
,
fmin
?lm Glm,?2hn , ?12 =
0
E? 2
? QLhn
1
,
(12)
?lm = 1I{yl ?=ym } is a class indicator
function and
Glm,h = Gh (xl , xm ) , Gh (x, y) =
E?2 is defined by (8), hn is chosen such that hn ? 0,
estimator of f defined by (3).
Kh (x ? y)
1
1
2
2
f?n,h
(x)f?n,h
(y)
log h?1
n
nhd
n
,
(13)
? 0, f?n,hn is the kernel density
(?
)
log h?1
?
n
? n is the dominant term determined solely by the data and the excess error O
R
+
h
n
nhd
n
goes to 0 with infinite n. In the following subsection, we show the close connection between the
error bound for the plug-in classifier and the weighted volume of cluster boundary, and the latter is
proposed by [1] for Low Density Separation.
3.1.1
Connection to Low Density Separation
Low Density Separation [19], a well-known criteria for clustering, requires that the cluster boundary
should pass through regions of low density. It has been extensively studied in unsupervised learning
and semi-supervised learning [20, 21, 22]. Suppose the data {xl }nl=1 lies on a domain ? ? Rd .
Let f be the probability density function on ?, S be the cluster boundary which separates ? into
two parts S1 and S2 . Following the Low Density Separation assumption, [1] suggests that the
5
?
cluster boundary S with low weighted volume
f (s)ds should be preferable. [1] also proves that
S
a particular type of cut function converges to the weighted volume of S. Based on their study, we
obtain the following result relating the error of the plug-in classifier to the weighted volume of the
cluster boundary.
Corollary 1. Under the assumption of Theorem 1, for any kernel bandwidth sequence {hn }?
n=1
1
such that lim hn = 0 and hn > n? 4d+4 , with probability 1,
n??
?
lim
n??
? ?
Rn (PIS ) =
2hn
?
f (s)ds
(14)
S
3.2
Generalization Bound for the Classification Model with Nearest Neighbor Classifier
Theorem 2 shows the generalization error bound for the classification model with nearest neighbor
classifier (NN), which has a similar form as (12).
(
)
Theorem 2. (Error of the NN) Given the classification model MY = S, PXY , {?i , fi }Q
i=1 , NN
d
such that PXY ? PXY and the support of PX is bounded by [?M0 , M0 ] , there exists a n0 which
depends on ?0 and VC characteristics of K, when n > n0 , with probability greater than 1 ?
E? 2
2QLhn
0
? (2M0 )d ndd0 e?n
1?dd0
fmin
, the generalization error of the NN satisfies:
? n (NNS ) + c0
R (NNS ) ? R
? n (NN) =
where R
1
n
?
?
)
( log h?1
(? )?
n
?
?d0 ?
+
h
d n
+O
n
nhdn
(15)
Hlm,hn ?lm ,
1?l<m?n
Hlm,hn = Khn (xl ? xm )
(
?
Vl
f?n,hn (x) dx
f?n,hn (xl )
?
+
f?n,hn (x) dx )
,
f?n,hn (xm )
Vm
(16)
E?2 is defined by (8), d0 is a constant such that dd0 < 1, f?n,hn is the kernel density estimator of
log h?1
f defined by (3) with the kernel bandwidth hn satisfying hn ? 0, nhdn ? 0, Vl is the Voronoi
n
cell associated with xl , c0 is a constant, ?lm = 1I{yl ?=ym } is a class indicator function such that
?lm = 1 if xl and xm belongs to different classes, and 0 otherwise. Moreover, the equality in (15)
1
holds when ? (i) ? Q
for 1 ? i ? Q.
Glm,?2hn in (13) and Hlm,hn in (16) are the new pairwise similarity functions induced by the plugin classifier and the nearest neighbor classifier respectively. According to the proof of Theorem 1 and
Theorem 2, the kernel density estimator f? can be replaced by the true density f in the denominators
of (13) and (16), and the conclusions of Theorem 1 and 2 still hold. Therefore, both Glm,?2hn and
Hlm,hn are equal to ordinary Gaussian kernels (up to a scale) with different kernel bandwidth under
uniform distribution, which explains the broadly used kernel similarity in data clustering from an
angle of supervised learning.
4
Application to Exemplar-Based Clustering
We propose a nonparametric exemplar-based clustering algorithm using the derived nonparametric
pairwise similarity by the plug-in classifier. In exemplar-based clustering, each xl is associated with
a cluster indicator el (l ? {1, 2, ...n} , el ? {1, 2, ...n}), indicating that xl takes xel as the cluster
n
exemplar. Data from the same cluster share the same cluster exemplar. We define e , {el }l=1 .
Moreover, a configuration of the cluster indicators e is consistent iff el = l when em = l for any
l, m ? 1..n, meaning that xl should take itself as its exemplar if any xm take xl as its exemplar. It is
required that the cluster indicators e should always be consistent. Affinity Propagation (AP) [23], a
representative of the exemplar-based clustering methods, solves the following optimization problem
min
e
n
?
Sl,el
s.t.
l=1
6
e is consistent
(17)
Sl,el is the dissimilarity between xl and xel , and note that Sl,l is set to be nonzero to avoid the trivial
minimizer of (17).
Now we aim to improve the discriminative capability of the exemplar-based clustering (17) using
the nonparametric pairwise similarity derived by the unsupervised plug-in classifier. As mentioned
? is evaluated by the generalization error bound for
before, the quality of the hypothetical labeling y
? with minimum
the nonparametric plug-in classifier trained by Sy? , and the hypothetical labeling y
? n (PIS ) = arg miny? ? ?lm G ?
associated error bound is preferred, i.e. arg miny? R
lm, 2hn where
l,m?
?lm Glm,?2hn also
?lm = 1Iy?l ?=y?m and Glm,?2hn is defined in (13). By Lemma 3, minimizing
l,m
enforces minimization of the weighted volume of cluster boundary asymptotically. To avoid the
trivial clustering where all the data are grouped into a single cluster, we use the sum of within)
(
n
?
cluster dissimilarities term
exp ?Glel ,?2hn to control the size of clusters. Therefore, the
l=1
objective function of our pairwise clustering method is below:
? (e) =
n
?
(
)
)
?(
exp ?Glel ,?2hn + ?
??lm Glm,?2hn + ?lm (el , em )
l=1
(18)
l,m
where ?lm is a function to enforce the consistency of the cluster indicators:
{
?lm (el , em ) =
? em = l, el ?= l or el = m, em ?= m
,
0 otherwise
? is a balancing parameter. Due to the form of (18), we construct a pairwise Markov Random
Field (MRF) representing the unary term ul and the pairwise term ??lm Glm,?2hn + ?lm as the data
likelihood and prior respectively. The variables e are modeled as nodes and the unary term and
pairwise term in (18) are modeled as potential functions in the pairwise MRF. The minimization of
the objective function is then converted to a MAP (Maximum a Posterior) problem in the pairwise
MRF. (18) is minimized by Max-Product Belief Propagation (BP).
The computational complexity of our clustering algorithm is O(T EN ), where E is the number of
edges in the pairwise MRF, T is the number of iterations of message passing in the BP algorithm.
We call our new algorithm Plug-In Exemplar Clustering (PIEC), and compare it to representative
exemplar-based clustering methods, i.e. AP and Convex Clustering with Exemplar-Based Model
(CEB) [24], for clustering on three real data sets from UCI repository, i.e. Iris, Vertebral Column
(VC) and Breast Tissue (BT). We record the average clustering accuracy (AC) and the standard
deviation of AC for all the exemplar-based clustering methods when they produce the correct number
of clusters for each data set with different values of hn and ?, and the results are shown in Table 1.
Although AP produces better clustering accuracy on the VC data set, PIEC generates the correct
cluster numbers for much more times. The dash in Table 1 indicates that the corresponding clustering
method cannot produce the correct cluster number. The default value for the{kernel bandwidth} hn is
h?n , which is set as the variance of the pairwise distance between data points ?xl ? xm ?l<m . The
default value for the balancing parameter ? is 1. We let hn = ?h?n , ? varies between [0.2, 1] and
? varies between [0.2, 1.9] with step 0.2 and 0.05 respectively, resulting in 170 different parameter
settings. We also generate the same number of parameter settings for AP and CEB.
Table 1: Comparison Between Exemplar-Based Clustering Methods. The number in the bracket is
the number of times when the corresponding algorithm produces correct cluster numbers.
Data sets
AP
CEB
PIEC
5
Iris
0.8933 ? 0.0138 (16)
0.6929 ? 0.0168 (15)
0.9089 ? 0.0033 (15)
VC
0.6677 (14)
0.4748 ? 0.0014 (5)
0.5263 ? 0.0173 (35)
BT
0.4906 (1)
0.3868 ? 0.08 (2)
0.6585 ? 0.0103 (5)
Conclusion
We propose a new pairwise clustering framework where nonparametric pairwise similarity is derived by minimizing the generalization error unsupervised nonparametric classifier. Our framework
bridges the gap between clustering and multi-class classification, and explains the widely used kernel similarity for clustering. In addition, we prove that the generalization error bound for the unsupervised plug-in classifier is asymptotically equal to the weighted volume of cluster boundary for
7
Low Density Separation. Based on the derived nonparametric pairwise similarity using the plug-in
classifier, we propose a new nonparametric exemplar-based clustering method with enhanced discriminative capability compared to the exiting exemplar-based clustering methods.
Appendix
Lemma 2. (Consistency of Kernel Density Estimator) Let the kernel bandwidth hn of the Gaussian
log h?1
kernel K be chosen such that hn ? 0, nhdn ? 0. For any PX ? PX , there exists a n0 which
n
E? 2
depends on ?0 and VC characteristics of K, when n > n0 , with probability greater than 1 ? Lhn
over the data {xl },
?
fn,hn (x) ? f (x)
=O
?
?
( log h?1
n
nhdn
+ h?n
0
)
(19)
where f?n,hn is the kernel density estimator of f . Furthermore, for any PXY ? PXY , when n > n0 ,
E? 2
then with probability greater than 1 ? 2Lhn
over the data {xl },
0
(i)
?n,hn (x) ? ? (i) (x)
?
?
=O
?
( log h?1
n
nhdn
+ h?n
)
(20)
for each 1 ? i ? Q.
Lemma 3. (Consistency of the Generalized Kernel Density Estimator) Suppose f is the probabilistic
density function of PX ? PX . Let g be a bounded function defined on X and g ? ??,g0 , 0 < gmin ?
g ? gmax , and e = fg . Define the generalized kernel density estimator of e as
e?n,h ,
n
1 ? Kh (x ? xl )
n
g (xl )
(21)
l=1
Let ?g2 =
?K?22 fmax
.
2
gmin
There exists ng which depends on ?g and the VC characteristics of K, When
E?2
n > ng , with probability greater than 1 ? Lhn
g
over the data {xl },
??
en,hn (x) ? e (x)?? = O
where hn is chosen such that hn ? 0,
log h?1
n
nhd
n
?
( log h?1
n
nhdn
+ h?n
)
(22)
? 0.
Sketch of proof: For fixed h ?= 0, we consider the class of functions
Fg , {
K
( t?? )
h
g (?)
, t ? IRd }
It can be verified that Fg is also a bounded VC class with the envelope function Fg =
(
) ( A )v
N Fg , ???L2 (P ) , ? ?Fg ?L2 (P ) ?
?
F
,
gmin
and
(23)
Then (22) follows from similar argument in the proof of Lemma 2 and Corollary 2.2 in [14].
The generalized kernel density estimator (21) is also used in [13] to estimate the Laplacian PDF
Distance between two probabilistic density functions, and the authors only provide the proof of
pointwise weak consistency of this estimator in [13]. Under mild conditions, our Lemma 3 and
Lemma 2 show the strong consistency of the generalized kernel density estimator and the traditional
kernel density estimator under the same theoretical framework of the VC property of the kernel.
Acknowledgements. This material is based upon work supported by the National Science Foundation under Grant No. 1318971.
8
References
[1] Hariharan Narayanan, Mikhail Belkin, and Partha Niyogi. On the relation between low density separation,
spectral clustering and graph cuts. In NIPS, pages 1025?1032, 2006.
[2] J. A. Hartigan and M. A. Wong. A K-means clustering algorithm. Applied Statistics, 28:100?108, 1979.
[3] Andrew Y. Ng, Michael I. Jordan, and Yair Weiss. On spectral clustering: Analysis and an algorithm. In
NIPS, pages 849?856, 2001.
[4] Noam Shental, Assaf Zomet, Tomer Hertz, and Yair Weiss. Pairwise clustering and graphical models. In
NIPS, 2003.
[5] Linli Xu, James Neufeld, Bryce Larson, and Dale Schuurmans. Maximum margin clustering. In NIPS,
2004.
[6] Zohar Karnin, Edo Liberty, Shachar Lovett, Roy Schwartz, and Omri Weinstein. Unsupervised svms: On
the complexity of the furthest hyperplane problem. Journal of Machine Learning Research - Proceedings
Track, 23:2.1?2.17, 2012.
[7] Ryan Gomes, Andreas Krause, and Pietro Perona. Discriminative clustering by regularized information
maximization. In NIPS, pages 775?783, 2010.
[8] Masashi Sugiyama, Makoto Yamada, Manabu Kimura, and Hirotaka Hachiya. On informationmaximization clustering: Tuning parameter selection and analytic solution. In ICML, pages 65?72, 2011.
[9] T. Cover and P. Hart. Nearest neighbor pattern classification. Information Theory, IEEE Transactions on,
13(1):21?27, January 1967.
[10] Luc Devroye. A probabilistic theory of pattern recognition, volume 31. springer, 1996.
[11] Yuhong Yang. Minimax nonparametric classification - part i: Rates of convergence. IEEE Transactions
on Information Theory, 45(7):2271?2284, 1999.
[12] Jean-Yves Audibert and Alexandre B. Tsybakov. Fast learning rates for plug-in classifiers. The Annals of
Statistics, 35(2):pp. 608?633, 2007.
[13] Robert Jenssen, Deniz Erdogmus, Jos?e Carlos Pr??ncipe, and Torbj?rn Eltoft. The laplacian pdf distance:
A cost function for clustering in a kernel feature space. In NIPS, 2004.
[14] Evarist Gin?e and Armelle Guillou. Rates of strong uniform consistency for multivariate kernel density
estimators. Ann. Inst. H. Poincar?e Probab. Statist., 38(6):907?921, November 2002.
[15] Uwe Einmahl and David M. Mason. Uniform in bandwidth consistency of kernel-type function estimators.
The Annals of Statistics, 33:1380C1403, 2005.
[16] R. M. Dudley. Uniform Central Limit Theorems. Cambridge University Press, 1999.
[17] A.W. van der Vaart and J.A. Wellner. Weak Convergence and Empirical Processes. Springer series in
statistics. Springer, 1996.
[18] Deborah Nolan and David Pollard. U-Processes: Rates of convergence. The Annals of Statistics, 15(2),
1987.
[19] Olivier Chapelle and Alexander Zien. Semi-Supervised Classification by Low Density Separation. In
AISTATS, 2005.
[20] Markus Maier, Ulrike von Luxburg, and Matthias Hein. Influence of graph construction on graph-based
clustering measures. In NIPS, pages 1025?1032, 2008.
[21] Zenglin Xu, Rong Jin, Jianke Zhu, Irwin King, Michael R. Lyu, and Zhirong Yang. Adaptive regularization for transductive support vector machine. In NIPS, pages 2125?2133, 2009.
[22] Xiaojin Zhu, John Lafferty, and Ronald Rosenfeld. Semi-supervised learning with graphs. PhD thesis,
Carnegie Mellon University, Language Technologies Institute, School of Computer Science, 2005.
[23] Brendan J. Frey and Delbert Dueck. Clustering by passing messages between data points. Science,
315:972?977, 2007.
[24] Danial Lashkari and Polina Golland. Convex clustering with exemplar-based models. In NIPS, 2007.
9
| 5304 |@word mild:2 repository:1 version:1 c0:2 open:1 shuicheng:1 configuration:1 series:1 existing:1 dx:2 john:1 deniz:1 ronald:1 fn:1 partition:6 shape:1 analytic:1 n0:7 isotropic:1 yamada:1 record:1 node:1 constructed:1 prove:3 weinstein:1 assaf:1 introduce:3 manner:1 pairwise:45 torbj:1 frequently:1 multi:4 inspired:1 estimating:4 underlying:2 bounded:7 notation:1 moreover:3 interpreted:1 minimizes:1 kimura:1 dueck:1 every:1 hypothetical:15 nf:1 masashi:1 preferable:1 classifier:69 demonstrates:1 schwartz:1 control:1 grant:1 superiority:1 before:3 positive:2 frey:1 limit:1 plugin:2 hirotaka:1 solely:2 ap:5 twice:1 studied:3 resembles:1 suggests:1 enforces:1 poincar:1 empirical:3 einmahl:1 cannot:2 unlabeled:3 close:1 selection:1 risk:5 influence:1 py:1 wong:1 measurable:1 map:1 demonstrated:1 maximizing:1 go:1 convex:3 estimator:28 utilizing:1 hd:1 annals:3 enhanced:2 suppose:2 construction:1 olivier:1 us:2 roy:1 satisfying:1 recognition:1 cut:2 labeled:1 hlm:4 gmin:3 region:1 mentioned:1 deborah:1 lashkari:1 complexity:3 miny:2 trained:2 depend:2 classifer:1 upon:1 basis:1 joint:3 represented:1 derivation:1 fast:1 labeling:16 whose:2 jean:1 huang1:2 widely:4 supplementary:2 otherwise:2 nolan:1 niyogi:1 statistic:5 vaart:1 transductive:1 rosenfeld:1 itself:1 sequence:1 un1:1 neufeld:1 matthias:1 propose:4 product:1 uci:1 fmax:4 liangf:1 iff:1 supposed:1 kh:3 convergence:5 cluster:32 produce:6 converges:3 help:1 derive:6 develop:1 stating:1 depending:1 ac:2 exemplar:20 measured:1 school:1 nearest:10 andrew:1 solves:1 strong:3 liberty:1 radius:1 closely:1 correct:4 vc:16 material:1 explains:2 generalization:39 ryan:1 rong:1 hold:3 lying:1 exp:2 lyu:1 lm:15 substituting:1 m0:3 adopt:1 purpose:1 estimation:3 khn:3 label:4 makoto:1 bridge:1 grouped:1 tool:1 weighted:8 minimization:2 gaussian:5 always:1 aim:1 avoid:3 corollary:3 iex:1 derived:7 focus:1 eleyans:1 likelihood:1 indicates:1 brendan:1 attains:1 wang1:1 inst:1 inference:1 voronoi:1 el:10 nn:10 vl:2 eliminate:1 unary:2 bt:2 hidden:1 relation:1 perona:1 comprising:1 arg:3 classification:32 uwe:1 mutual:2 marginal:5 equal:4 construct:1 field:1 karnin:1 ng:3 informationmaximization:1 represents:1 unsupervised:20 icml:1 minimized:2 belkin:1 national:2 hamper:1 replaced:1 n1:2 message:3 analyzed:1 mixture:1 nl:5 bracket:1 edge:1 gmax:1 hein:1 theoretical:2 minimal:1 column:1 modeling:1 cover:2 maximization:1 ordinary:1 cost:1 deviation:1 uniform:6 zenglin:1 varies:2 my:6 nns:2 density:51 sequel:1 probabilistic:6 yl:6 vm:1 jos:1 michael:2 connecting:1 ym:2 iy:1 squared:1 central:1 von:1 thesis:1 hn:52 leading:1 potential:2 converted:1 satisfy:3 audibert:1 depends:6 nhn:1 ulrike:1 bayes:4 carlos:1 capability:4 complicated:2 partha:1 il:1 hariharan:1 ir:1 accuracy:2 variance:1 largely:1 characteristic:7 sy:1 maier:1 yves:1 weak:2 tissue:1 hachiya:1 edo:1 definition:1 pp:1 james:1 associated:7 proof:5 couple:1 proved:2 subsection:1 lim:2 appears:2 alexandre:1 supervised:7 wei:2 formulation:1 evaluated:1 guillou:1 furthermore:2 d:2 sketch:1 ncipe:1 propagation:2 logistic:1 quality:3 usa:1 concept:1 true:3 equality:2 regularization:1 nonzero:1 satisfactory:1 self:1 larson:1 iris:2 criterion:4 generalized:8 pdf:2 complete:1 gh:2 meaning:1 novel:1 fi:3 common:1 volume:9 relating:1 mellon:1 cambridge:1 danial:1 smoothness:1 rd:1 tuning:1 consistency:13 illinois:2 sugiyama:1 language:1 chapelle:1 f0:2 similarity:32 base:1 dominant:2 posterior:2 multivariate:1 recent:1 belongs:3 vertebral:1 zhirong:1 success:1 der:1 minimum:4 greater:6 employed:1 semi:5 zien:1 d0:2 jianke:1 champaign:1 smooth:3 faster:1 plug:32 hart:1 y:1 laplacian:2 mrf:4 regression:8 denominator:1 essentially:1 metric:1 breast:1 iteration:1 kernel:47 cell:1 golland:1 addition:1 krause:1 envelope:2 sure:1 induced:4 undirected:1 facilitates:1 xel:2 lafferty:1 jordan:1 call:1 yang:2 bandwidth:7 andreas:1 bridging:1 ul:1 wellner:1 ird:5 f:1 pollard:1 passing:3 linli:1 polina:1 detailed:1 nonparametric:31 tsybakov:1 extensively:2 statist:1 narayanan:1 svms:1 generate:1 sl:3 exist:1 singapore:2 track:1 broadly:2 carnegie:1 shental:1 group:1 drawn:1 hartigan:1 verified:1 asymptotically:4 graph:4 pietro:1 sum:7 luxburg:1 angle:1 throughout:1 almost:1 separation:10 appendix:2 bound:35 dash:1 pxy:18 bp:2 sake:1 markus:1 generates:2 fmin:3 argument:1 min:2 px:15 according:1 ball:1 hertz:1 em:5 s1:1 pr:4 glm:8 serf:1 spectral:3 enforce:1 dudley:1 yair:2 thomas:1 clustering:65 include:1 graphical:2 prof:2 feng:1 objective:2 g0:1 quantity:1 primary:1 traditional:1 exhibit:2 affinity:1 dp:1 gin:1 distance:3 separate:1 manifold:1 trivial:2 furthest:1 assuming:1 devroye:1 modeled:2 pointwise:1 minimizing:3 difficult:1 mostly:1 yingzhen:1 robert:1 kde:2 noam:1 unknown:1 upper:2 markov:1 urbana:2 november:1 jin:1 january:1 rn:4 tomer:1 exiting:1 david:2 evidenced:1 namely:2 required:5 specified:1 connection:4 learned:7 nu:1 nip:9 zohar:1 below:5 pattern:2 xm:6 built:1 max:3 eltoft:1 belief:1 dd0:2 natural:1 regularized:1 indicator:6 zhu:2 representing:1 older:3 improve:1 minimax:1 technology:1 identifies:1 bryce:1 xiaojin:1 text:1 prior:2 sg:1 literature:3 l2:4 acknowledgement:1 probab:1 loss:1 lovett:1 foundation:1 consistent:3 principle:1 pi:13 share:1 balancing:2 lhn:3 supported:1 bias:1 institute:1 neighbor:10 mikhail:1 fg:6 van:1 boundary:10 dimension:1 default:2 dale:1 author:1 collection:3 adaptive:1 transaction:2 excess:6 preferred:2 nhd:4 assumed:1 gomes:1 discriminative:8 search:2 table:3 learn:1 evarist:1 schuurmans:1 complex:3 domain:1 aistats:1 s2:1 n2:1 xu:2 representative:3 en:2 ff:1 delbert:1 xl:21 lie:1 omri:1 learns:4 theorem:11 specific:1 emphasized:1 yuhong:1 ceb:3 mason:1 svm:1 virtue:1 exists:5 ci:2 phd:1 dissimilarity:3 margin:3 gap:2 yang1:1 entropy:1 zhangyang:1 expressed:2 g2:1 springer:3 corresponds:1 minimizer:1 satisfies:3 jenssen:1 conditional:5 king:1 ann:1 erdogmus:1 luc:1 included:1 determined:2 infinite:1 uniformly:1 zomet:1 hyperplane:1 lemma:16 called:1 pas:1 experimental:1 indicating:1 support:2 latter:1 irwin:1 alexander:1 |
4,755 | 5,305 | Robust Bayesian Max-Margin Clustering
Changyou Chen?
Jun Zhu?
Xinhua Zhang]
?
Dept. of Electrical and Computer Engineering, Duke University, Durham, NC, USA
?
State Key Lab of Intelligent Technology & Systems; Tsinghua National TNList Lab;
?
Dept. of Computer Science & Tech., Tsinghua University, Beijing 100084, China
]
Australian National University (ANU) and National ICT Australia (NICTA), Canberra, Australia
[email protected]; [email protected]; [email protected]
Abstract
We present max-margin Bayesian clustering (BMC), a general and robust framework that incorporates the max-margin criterion into Bayesian clustering models,
as well as two concrete models of BMC to demonstrate its flexibility and effectiveness in dealing with different clustering tasks. The Dirichlet process max-margin
Gaussian mixture is a nonparametric Bayesian clustering model that relaxes the
underlying Gaussian assumption of Dirichlet process Gaussian mixtures by incorporating max-margin posterior constraints, and is able to infer the number of
clusters from data. We further extend the ideas to present max-margin clustering topic model, which can learn the latent topic representation of each document
while at the same time cluster documents in the max-margin fashion. Extensive
experiments are performed on a number of real datasets, and the results indicate
superior clustering performance of our methods compared to related baselines.
1
Introduction
Existing clustering methods fall roughly into two categories. Deterministic clustering directly optimises some loss functions, while Bayesian clustering models the data generating process and infers the clustering structure via Bayes rule. Typical deterministic methods include the well known
kmeans [1], nCut [2], support vector clustering [3], Bregman divergence clustering [4, 5], and the
methods built on the very effective max-margin principle [6?9]. Although these methods can flexibly incorporate constraints for better performance, it is challenging for them to finely capture hidden
regularities in the data, e.g., automated inference of the number of clusters and the hierarchies underlying the clusters. In contrast, Bayesian clustering provides favourable convenience in modelling
latent structures, and their posterior distributions can be inferred in a principled fashion. For example, by defining a Dirichlet process (DP) prior on the mixing probability of Gaussian mixtures,
Dirichlet process Gaussian mixture models [10] (DPGMM) can infer the number of clusters in the
dataset. Other priors on latent structures include the hierarchical cluster structure [11?13], coclustering structure [14], etc. However, Bayesian clustering is typically difficult to accommodate
external constraints such as max-margin. This is because under the standard Bayesian inference
designing some informative priors (if any) that satisfy these constraints is highly challenging.
To address this issue, we propose Bayesian max-margin clustering (BMC), which allows maxmargin constraints to be flexibly incorporated into a Bayesian clustering model. Distinct from the
traditional max-margin clustering, BMC is fully Bayesian and enables probabilistic inference of the
number of clusters or the latent feature representations of data. Technically, BMC leverages the
regularized Bayesian inference (RegBayes) principle [15], which has shown promise on supervised
learning tasks, such as classification [16, 17], link prediction [18], and matrix factorisation [19],
where max-margin constraints are introduced to improve the discriminative power of a Bayesian
1
model. However, little exploration has been devoted to the unsupervised setting, due in part to the
absence of true labels that makes it technically challenging to enforce max-margin constraints. BMC
constitutes a first extension of RegBayes to the unsupervised clustering task. Note that distinct from
the clustering models using maximum entropy principle [20, 21] or posterior regularisation [22],
BMC is more general due to the intrinsic generality of RegBayes [15].
We demonstrate the flexibility and effectiveness of BMC by two concrete instantiations. The first is
Dirichlet process max-margin Gaussian mixture (DPMMGM), a nonparametric Bayesian clustering
model that relaxes the Gaussian assumption underlying DPGMM by incorporating max-margin constraints, and is able to infer the number of clusters in the raw input space. To further discover latent
feature representations, we propose the max-margin clustering topic model (MMCTM). As a topic
model, it performs max-margin clustering of documents, while at the same time learns the latent
topic representation for each document. For both DPMMGM and MMCTM, we develop efficient
MCMC algorithms by exploiting data augmentation techniques. This avoids imposing restrictive
assumptions such as in variational Bayes, thereby facilitating the inference of the true posterior. Extensive experiments demonstrate superior clustering performance of BMC over various competitors.
2
Regularized Bayesian Inference
We first briefly overview the principle of regularised Bayesian inference (RegBayes) [15].
The motivation of RegBayes is to enrich the posterior of a probabilistic model by incorporating additional constraints, under an information-theoretical optimisation formulation. Formally, suppose a
probabilistic model has latent variables ?, endowed with a prior p(?) (examples of ? will be clear
soon later). We also have observations X := {x1 , ? ? ? , xn }, with xi ? Rp . Let p(X|?) be the
likelihood. Then, posterior inference via the Bayes? theorem is equivalent to solving the following
optimisation problem [15]:
inf
q(?)?P
KL(q(?) || p(?)) ? E??q(?) [log p(X|?)]
(1)
where P is the space of probability distribution1 , q(?) is the required posterior (here and afterwards
we will drop the dependency on X for notation simplicity). In other words, the Bayesian posterior
p(?|X) is identical to the optimal solution to (1). The power of RegBayes stems in part from
the flexibility of engineering P, which typically encodes constraints imposed on q(?), e.g., via
expectations of some feature functions of ? (and possibly the data X). Furthermore, the constraints
can be parameterised by some auxiliary variable ?. For example, ? may quantify the extent to which
the constraints are violated, then it is penalised in the objective through a function U . To summarise,
RegBayes can be generally formulated as
inf KL(q(?) || p(?))? E??q(?) [log p(X|?)]+ U (?)
?,q(?)
s.t. q(?) ? P(?).
(2)
To distinguish from the standard Bayesian posterior, the optimal q(?) is called post-data posterior.
Under mild regularity conditions, RegBayes admits a generic representation theorem to characterise
the solution q(?) [15]. It is also shown to be more general than the conventional Bayesian methods,
including those methods that introduce constraints on a prior. Such generality is essential for us
to develop a Bayesian framework of max-margin clustering. Note that like many sophisticated
Bayesian models, posterior inference remains as a key challenge of developing novel RegBayes
models. Therefore, one of our key technical contributions is on developing efficient and accurate
algorithms for BMC, as detailed below.
3
Robust Bayesian Max-margin Clustering
For clustering, one key assumption of our model is that X forms a latent cluster structure. In particular, let each cluster be associated with a latent projector ?k ? Rp , which is included in ? and has
prior distribution subsumed in p(?). Given any distribution q on ?, we then define the compatibility
score of xi with respect to cluster k by using the marginal distribution on ?k (as ?k ? ?):
Fk (xi ) = Eq(?k ) ?kT xi = Eq(?) ?kT xi .
(3)
1
In theory, we also require that q is absolutely continuous with respect to p to make the KL-divergence well
defined. The present paper treats this constraint as an implicit assumption for clarity.
2
For each example xi , we introduce a random variable yi valued in Z+ , which denotes its cluster
assignment and is also included in ?. Inspired by conventional multiclass SVM [7, 23], we utilize
P(?) in RegBayes (2) to encode the max-margin constraints based on Fk (xi ), with the slack variable
? penalised via their sum in U (?). This amounts to our Bayesian max-margin clustering (BMC):
X
inf L(q(?)) + 2c
?i
(4)
?i ?0,q(?)
s.t.
i
Fyi (xi ) ? Fk (xi ) ? ` I(yi 6= k) ? ?i ,
?i, k
where L(q(?)) = KL(q(?)||p(?)) ? E??q(?) [log p(X|?)] measures the KL divergence between
q and the original Bayesian posterior p(?|X) (up to a constant); I(?) = 1 if ? holds true, and
0 otherwise; ` > 0 is a constant scalar of margin. Note we found that the commonly adopted
balance constraints in max-margin clustering models [6] either are unnecessary or do not help in our
framework. We will address this issue in specific models.
Clearly by absorbing the slack variables ?, the optimisation problem (4) is equivalent to
X
inf L(q(?)) + 2c
max 0, max E??q(?) [?ik ]
q(?)
k:k6=yi
i
(5)
where ?ik := ` I(yi 6= k) ? (?yi ? ?k )T xi . Exact solution to (5) is hard to compute. An alternative
approach is to approximate the posterior by assuming independence between random variables, e.g.
variational inference. However, this is usually slow and susceptible to local optimal. In order to
obtain an analytic optimal distribution q that facilitates efficient Bayesian inference, we resort to the
technique of Gibbs classifier [17] which approximates (in fact, upper bounds due to the convexity
of max function) the second term in (5) by an expected hinge loss, i.e., moving the expectation out
of the max. This leads to our final formulation of BMC:
X
inf L(q(?)) + 2c
E??q(?) max 0, max ?ik .
(6)
q(?)
k:k6=yi
i
Problem (6) is still much more challenging than existing RegBayes models [17], which are restricted
to supervised learning with two classes only. Specifically, BMC allows multiple clusters/classes in
an unsupervised setting, and the latent cluster membership yi needs to be inferred. This complicates
the model and brings challenges for posterior inference, as addressed below. In a nutshell, our
inference algorithms rely on two key steps by exploring data augmentation techniques. First, in order
to tackle the multi-class case, we introduce auxiliary variables si := arg maxk:k6=yi ?ik . Applying
standard derivations in calculus of variation [24] and augmenting the model with {si }, we obtain an
analytic form of the optimal solution to (6) by augmenting ? (refer to Appendix A for details):
Y
q(?, {si }) ? p(?|X)
exp(?2c max(0, ?isi )) .
(7)
i
Second, since the max term in (7) obfuscates efficient sampling, we apply the augmentation technique introduced by [17], which showed that q(?, {si }) is identical to the marginal distribution of
the augmented post-data posterior
Y
??i (?i |?),
(8)
q(?, {si }, {?i }) ? p(?|X)
i
?1
?i 2
where ??i (?i |?) :=
exp
+ c?isi )2 . Here ?i is an augmented variable for xi that has
an generalised inverse Gaussian distribution [25] given ? and xi .
?1
2?i (?i
Note that our two steps of data augmentation are exact and incur no approximation. With the augmented variables ({si }, {?i }), we can develop efficient sampling algorithms for the augmented
posterior q(?, {si }, {?i }) without restrictive assumptions, thereby allowing us to approach the true
target posterior q(?) by dropping the augmented variables. The details will become clear soon in
our subsequent clustering models.
4
Dirichlet Process Max-margin Gaussian Mixture Models
In (4), we have left unspecified the prior p(?) and the likelihood p(X|?). This section presents an
instantiation of Bayesian nonparametric clustering for non-Gaussian data. We will present another
instantiation of max-margin document clustering based on topic models in next section.
3
w
yi
?k
v
?k
?, S
?0
?
N
?k
?k
K
?1
?0
xi
v
?
?
?
T
yi
?
?k
r, m
?t
?i
zil
K
wil
Ni
?
D
Figure 1: Left: Graphical model of DPMMGM. The part excluding ?k and v corresponds to
DPGMM. Right: Graphical model of MMCTM. The one excluding {?k } and the arrow between
yi and wil corresponds to CTM.
Here a convenient model of p(X, ?) is mixture of Gaussian. Let the mean and variance of the k-th
cluster component be ?k and ?k . In a nonparametric setting, the number of clusters is allowed to
be infinite, and the cluster yi that each data point belongs to is drawn from a Dirichlet process [10].
n
To summarize, the latent variables are ? = {?k , ?k , ?k }?
k=1 ? {yi }i=1 . The prior p(?) is specified
as: ?k and ?k employ a standard Normal-inverse Wishart prior [26]:
?k ? N (?k ; m, (r?k )?1 ), and ?k ? IW(?k ; S, ?).
(9)
+
yi ? Z has a Dirichlet process prior with parameter ?. ?k follows a normal prior with mean 0 and
variance vI, where I is the identity matrix. The likelihood p(xi |?) is N (xi ; ?yi , (r?yi )?1 ), i.e.
independent of ?k . The max-margin constraints take effects in the model via ??i ?s in (8). Note this
model of p(?, X), apart from ?k , is effectively the Dirichlet process Gaussian mixture model [10]
(DPGMM). Therefore, we call our post-data posterior q(?, {si }, {?i }) in (8) as Dirichlet process
max-margin Gaussian mixture model (DPMMGM). The hyperparameters include m, r, S, ?, ?, v.
Interpretation as a generalised DP mixture The formula of the augmented post-data posterior
in (8) reveals that, compared with DPGMM, each data point is associated with an additional factor
??i (?i |?). Thus we can interpret DPMMGM as a generalised DP mixture with Normal-inversed
Wishart-Normal as the base distribution, and a generalised pseudo likelihood that is proportional to
f (xi , ?i |yi , ?y , ?y , {?k }) := N (xi ; ?y , (r?y )?1 )??i (?i |?) .
(10)
i
i
i
i
To summarise, DPMMGM employs the following generative process with the graphical model
shown in Fig. 1 (left):
(?k , ?k , ?k ) ? N ?k ; m, (r?k )?1 ? IW (?k ; S, ?) ? N (?k ; 0, vI) , k = 1, 2, ? ? ?
w ? Stick-Breaking(?),
yi |w ? Discrete(w),
i ? [n]
(xi , ?i )|yi , {?k , ?k , ?k } ' f (xi , ?i |yi , ?yi , ?yi , {?k }).
i ? [n]
Here [n] := {1, ? ? ? , n} is the set of integers up to n and ' means that (xi , ?i ) is generative from a
distribution that is proportional to f (?). Since this normalisation constant is shared by all samples
xi , there is no need to deal with it by posterior inference. Another benefit of this interpretation
is that it allows us to use existing techniques for non-conjugate DP mixtures to sample the cluster
indicators yi efficiently, and to infer the number of clusters in the data. This approach is different
from previous work on RegBayes nonparametric models where truncated approximation is used to
deal with the infinite dimensional model space [15, 18]. In contrast, our method does not rely on any
approximation. Note that DPMMGM does not need the complicated class balance constraints [6]
because the Gaussians in the pseudo likelihood would balance the clusters to some extent.
Posterior inference Posterior inference for DPMMGM can be done by efficient Gibbs sampling. We integrate out the infinite dimension vector w, so the variables needed to be sampled
are {?k , ?k , ?k }k ? {yi , si , ?i }i . Conditional distributions are derived in Appendix B. Note that we
use an extension of the Reused Algorithm [27] to jointly sample (yi , si ), which allows it to allocate
to empty clusters in Bayesian nonparametric setting. The time complexity is almost the same as
DPGMM except for the additional step to sample ?k , with cost O(p3 ). So it would be necessary to
put the constraints on a subspace (e.g., by projection) of the original feature space when p is high.
4
5
Max-margin Clustering Topic Model
Although many applications exhibit clustering structures in the raw observed data which can be
effectively captured by DPMMGM, it is common that such regularities are more salient in terms
of some high-level but latent features. For example, topic distributions are often more useful than
word frequency in the task of document clustering. Therefore, we develop a max-margin clustering
topic model (MMCTM) in the framework of BMC, which allows topic discovery to co-occur with
document clustering in a Bayesian and max-margin fashion. To this end, the latent Dirichlet allocation (LDA) [28] needs to be extended by introducing a cluster label into the model, and define each
cluster as a mixture of topic distributions. This cluster-based topic model [29] (CTM) can then be
used in concert with BMC to enforce large margin between clusters in the posterior q(?).
Let V be the size of the word vocabulary, T be the number of topics, and K be the number of clusters,
1N be a N -dimensional one vector. Then the generative process of CTM for the documents goes as:
1. For each topic t, generate its word distribution ?t : ?t |? ? Dir(?1V ).
2. Draw a base topic distribution ?0 : ?0 |?0 ? Dir(?0 1T ). Then for each cluster k, generate its
topic distribution mixture ?k : ?k |?1 , ?0 ? Dir(?1 ?0 ).
3. Draw a base cluster distribution ?: ?|? ? Dir(?1K ). Then for each document i ? [D]:
? Generate a cluster label yi and a topic distribution ?i : yi |? ? Discrete(?), ?i |?, ?yi ?
Dir(??yi ).
? Generate the observed words wil : zil ? Discrete(?i ), wil ? Discrete(?zil ), ? l ? [Ni ].
Fig. 1 (right) shows the structure. We then augment CTM with max-margin constraints, and get
the same posterior as in Eq. (7), with the variables ? corresponding to {?t }Tt=1 ? {?k , ?k }K
k=1 ?
D,Ni
D
{?0 , ?} ? {?i , yi }i=1 ? {zil }i=1,l=1 .
Compared with the raw word space which is normally extremely high-dimensional and sparse, it
is more reasonable to characterise the clustering structure in the latent feature space?the empirical
latent topic distributions as in the MedLDA [16]. Specifically, we summarise the topic distribution
PNi
of document i by xi ? RT , whose t-th element is N1i l=1
I(zil = t). Then the compatibility score
for document i with respect to cluster k is defined similar to (3) as Fk (xi ) = Eq(?) ?kT xi . Note,
however, the expectation is also taken over xi since it is not observed.
Posterior inference To achieve fast mixing, we integrate out {?t }Tt=1 ? {?0 , ?} ? {?k }K
k=1 ?
D,Ni
D
D
K
{?i }i=1 in the posterior, thus ? = {yi }i=1 ? {?k }k=1 ? {zil }i=1,l=1 . The integration is straightforward by the Dirichlet-Multinomial conjugacy. The detailed form of the posterior and the conditional
distributions are derived in Appendix C. By extending CTM with max-margin, we note that many
of the the sampling formulas are extension of those in CTM [29], with additional sampling for ?k ,
thus the sampling can be done fairly efficiently.
Dealing with vacuous solutions Different from DPMMGM, the max-margin constraints in MMCTM do not interact with the observed words wil , but with the latent topic representations xi (or zil )
that are also inferred from the model. This easily makes the latent representation zi ?s collapse into a
single cluster, a vacuous solution plaguing many other unsupervised learning methods as well. One
remedy is to incorporate the cluster balance constraints into the model [7]. However, this does not
help in our Bayesian setting because apart from significant increase in computational cost, MCMC
often fails to converge in practice2 . Another solution is to morph the problem into a weakly semisupervised setting, where we assign to each cluster a few documents according to their true label
(we will refer to these documents as landmarks), and sample the rest as in the above unsupervised
setting. These ?labeled examples? can be considered as introducing constraints that are alternative
to the balance constraints. Usually only a very small number of labeled documents are needed, thus
barely increasing the cost in training and labelling. We will focus on this setting in experiment.
6
6.1
Experiments
Dirichlet Process Max-margin Gaussian Mixture
2
We observed the cluster sizes kept bouncing with sampling iterations. This is probably due to the highly
nonlinear mapping from observed word space to the feature space (topic distribution), making the problem
multi-modal, i.e., there are multiple optimal topic assignments in the post-data posterior (8). Also the balance
constraints might weaken the max-margin constraints too much.
5
We first show the distinction between our DPMMGM and DPGMM
by running both models on the non-Gaussian half-rings data set [30].
There are a number of hyperparameters to be determined, e.g.,
(?, r, S, ?, v, c, `); see Section 4. It turns out the cluster structure is
insensitive to (?, r, S, ?), and so we use a standard sampling method
to update ? [31], while r, ?, S are sampled by employing Gamma,
truncated Poisson, inverse Wishart priors respectively, as is done
in [32]. We set v = 0.01, c = 0.1, ` = 5 in this experiment. Note that
the clustering structure is sensitive to the values of c and `, which will
be studied below. Empirically we find that DPMMGM converges
much faster than DPGMM, both converging well within 200 iterations (see Appendix D.4 for examples). In Fig. 2, the clustering
structures demonstrate clearly that DPMMGM relaxes the Gaussian
assumption of the data distribution, and correctly finds the number of Figure 2: An illustration
clusters based on the margin boundary, whereas DPGMM produces of DPGMM (up) and DPMa too fragmented partition of the data for the clustering task.
MGM (bottom).
0
5
0
5
0
5
0
5
0
5
0
5
0
5
0
5
0
5
0
5
Parameter sensitivity We next study the sensitivity of hyperparameters c and `, with other hyperparameters sampled during inference as above. Intuitively the impact of these parameters is as follows.
c controls the weight that the max-margin constraint places on the posterior. If there were no other
constraint, the max-margin constraint would drive the data points to collapse into a single cluster.
As a result, we expect that a larger value of the weight c will result in fewer clusters. Similarly,
increasing the value of ` will lead to a higher loss for any violation of the constraints, thus driving
the data points to collapse as well. To test these implications, we run DPMMGM on a 2-dimensional
synthetic dataset with 15 clusters [33]. We vary c and ` to study how the cluster structures change
with respect to these parameter settings. As can be observed from Fig. 3, the results indeed follow
our intuition, providing a mean to control the cluster structure in applications.
8
8
8
8
8
6
6
6
6
6
4
4
4
4
4
2
2
2
2
2
0
0
0
0
0
8
8
8
8
8
6
6
6
6
6
4
4
4
4
4
2
2
2
2
2
(a) c :5e-6, ` :5e-1
(b) c :5e-4, ` :5e-1
(c) c :5e-3, ` :5e-1
(d) c :5e-2, ` :5e-1
(e) c :5e-1, ` :5e-1
8
8
8
8
8
6
6
6
6
6
4
4
4
4
4
2
2
2
2
2
0
0
0
0
0
8
8
8
8
8
6
6
6
6
6
4
4
4
4
4
2
2
2
2
2
(f) c :5e-3, ` :5e-4
(g) c :5e-3, ` :5e-2
(h) c :5e-3, ` :5e-1
(i) c :5e-3, ` :2
(j) c :5e-3, ` :5
Figure 3: Clustering structures with varied ` and c: (first row) fixed ` and increasing c; (second row)
fixed c and increasing `. Lines are ??s. Clearly the number cluster decreases with growing c and `.
Real Datasets. As other clustering models, we test DPMMGM on ten real datasets (small to moderate sizes) from the UCI repository [34]. Scaling up to large dataset is an interesting future. The
first three columns of Table 1 list some of the statistics of these datasets (we used random subsets of
the three large datasets ? Letter, MNIST, and Segmentation).
A heuristic approach for model selection. Model selection is generally hard for unsupervised
clustering. Most existing algorithms simply fix the hyperparameters without examining their impacts
on model performance [10, 35]. In DPMMGM, the hyperparameters c and ` are critical to clustering
quality since they control the number of clusters. Without training data in our setting they can not
be set using cross validation. Moreover, they are not feasible to be estimated use Bayesian sampling
as well because they are not parameters from a proper Bayesian model. we thus introduce a timeefficient heuristic approach to selecting appropriate values. Suppose the dataset is known to have
K clusters. Our heuristic goes as follows. First initialise c and ` to 0.1. Then at each iteration,
we compare the inferred number of clusters with K. If it is larger than K (otherwise we do the
converse), we choose c or ` randomly, and increase its value by nu , where u is a uniform random
variable in [0, 1] and n is the number of iterations so far. According to the parameter sensitivity
studied above, increasing c or ` tends to decrease the number of clusters, and the model eventually
6
Dataset
Glass
Half circle
Iris
Letter
MNIST
Satimage
Segment?n
Vehicle
Vowel
Wine
Data property
n
p
K
214
10
7
300
2
2
150
4
3
1000 16 10
1000 784 10
4435 36
6
1000 19
7
846
18
4
990
10 11
178
13
3
kmeans
0.37?0.04
0.43?0.00
0.72?0.08
0.33?0.01
0.50?0.01
0.57?0.06
0.52?0.03
0.10?0.00
0.42?0.01
0.84?0.01
nCut
0.22?0.00
1.00?0.00
0.61?0.00
0.04?0.00
0.38?0.00
0.55?0.00
0.34?0.00
0.14?0.00
0.44?0.00
0.46?0.00
NMI
DPGMM
0.37?0.05
0.49?0.02
0.73?0.00
0.19?0.09
0.55?0.03
0.21?0.05
0.23?0.09
0.02?0.02
0.28?0.03
0.56?0.02
DPMMGM
0.46?0.01
0.67?0.02
0.73?0.00
0.38?0.04
0.56?0.01
0.51?0.01
0.61?0.05
0.14?0.00
0.39?0.02
0.90?0.02
DPMMGM?
0.45?0.01
0.51?0.07
0.73?0.00
0.23?0.04
0.55?0.02
0.30?0.00
0.52?0.10
0.05?0.00
0.41?0.02
0.59?0.01
Table 1: Comparison for different methods on NMI scores. K: true number of clusters.
stabilises due to the stochastic decrement by nu . We denote the model learned from this heuristic
as DPMMGM. In the case where the true number of clusters is unknown, we can still apply this
strategy, except that the number of clusters K needs to be first inferred from DPGMM. This method
is denoted as DPMMGM? .
Comparison. We measure the quality of clustering results by using the standard normalised mutual
information (NMI) criterion [36]. We compare our DPMMGM with the well established KMeans,
nCut and DPGMM clustering methods3 . All experiments are repeated for five times with random
initialisation. The results are shown in Table 1. Clearly DPMMGM significantly outperforms other
models, achieving the best NMI scores. DPMMGM? , which is not informed of the true number of
clusters, still obtains reasonably high NMI scores, and outperforms the DPGMM model.
6.2
Max-margin Clustering Topic Model
Datasets. We test the MMCTM model on two document datasets: 20NEWS and Reuters-R8 . For
the 20NEWS dataset, we combine the training and test datasets used in [16], which ends up with 20
categories/clusters with roughly balanced cluster sizes. It contains 18,772 documents in total with a
vocabulary size of 61,188. The Reuters-R8 dataset is a subset of the Reuters-21578 dataset4 , with of
8 categories and 7,674 documents in total. The size of different categories is biased, with the lowest
number of documents in a category being 51 while the highest being 2,292.
Comparison We choose L ? {5, 10, 15, 20, 25} documents randomly from each category as the
landmarks, use 80% documents for training and the rest for testing. We set the number of topics
(i.e., T ) to 50, and set the Dirichlet prior in Section 5 to ? = 0.1, ? = 0.01, ? = ?0 = ?1
= 10, as clustering quality is not sensitive to them. For the other hyperparameters related to the
max-margin constraints, e.g., v in the Gaussian prior for ?, the balance parameter c, and the cost
parameter `, instead of doing cross validation which is computationally expensive and not helpful
for our scenario with few labeled data, we simply set v = 0.1, c = 9, ` = 0.1. This is found to
be a good setting and denoted as MMCTM. To test the robustness of this setting, we vary c over
{0.1, 0.2, 0.5, 0.7, 1, 3, 5, 7, 9, 15, 30, 50} and keep v = ` = 0.1 (` and c play similar roles and so
varying one is enough). We choose the best performance out of these parameter settings, denoted
as MMCTM? , which can be roughly deemed as the setting for the optimal performance. We compared MMCTM with state-of-the-art SVM and semi-supervised SVM (S3VM) models. They are
efficiently implemented in [37], and the related parameters are chosen by 5-fold cross validation.
As in [16], raw word frequencies are used as input features. We also compare MMCTM with a
Bayesian baseline?cluster based topic model (CTM) [29], the building block of MMCTM without
the max-margin constraints. Note we did not compare with the standard MedLDA [16] because it
is supervised. We measure the performance by cluster accuracy, which is the proportion of correctly clustered documents. To accelerate MMCTM, we simply initialise it with CTM, and find it
converges surprisingly fast in term of accuracy, e.g., usually within 30 iterations (refer to Appendix
3
We additionally show some comparison with some existing max-margin clustering models in Appendix D.2
on two-cluster data because their code only deals with the case of two clusters. Our method performs best.
4
Downloaded from csmining.org/index.php/r52-and-r8-of-reuters-21578.html.
7
5
10
15
20
25
17.22? 4.2
24.50? 4.5
22.76? 4.2
26.07? 7.2
27.20? 1.5
5
10
15
20
25
41.27? 16.7
42.63? 7.4
39.67? 9.9
58.24? 8.3
51.93? 5.9
MMCTM
MMCTM?
37.13? 2.9 39.36? 3.2
46.99? 2.4 47.91? 2.8
52.80? 1.2 52.49? 1.4
56.10? 1.5 54.44? 2.1
59.15? 1.4 57.45? 1.7
Reuters-R8
56.70? 1.9
54.92? 1.6
55.06? 2.7
56.62? 2.2
55.70? 2.4
57.86? 0.9
56.56? 1.3
57.80? 2.2
59.70? 1.4
61.92? 3.0
78.12? 1.1
80.69? 1.2
83.25? 1.7
85.66? 1.0
84.95? 0.1
79.18? 4.1
80.04? 5.3
85.48? 2.1
82.92? 1.7
86.56? 2.5
80.86? 2.9
83.48? 1.0
86.86? 2.5
83.82? 1.6
88.12? 0.5
SVM
S3VM
20NEWS
78.51? 2.3
79.15? 1.2
81.87? 0.8
73.95? 2.0
82.39? 1.8
Accuracy (%)
CTM
training test
60
40
20
0
10 20 30 50 70 100
Number of topics (#topic)
(a) 20NEWS dataset
Accuracy (%)
L
80
60
40
20
0
10 20 30 50 70 100
Number of topics
(b) Reuters-R8 dataset
Table
2: Clustering acc. (in %). Bold means significantly
different. Figure 4: Accuracy vs. #topic aaaaaaaaaaaaaaaa
60
60
40
40
20
20
0
0
?20
?20
?40
?40
?60
?60
Figure
5: 2-D
tSNE
20NEWS
MMCTM
and 0CTM
viewed
in
?40
?30 (left)
?20
?10
10 (right).
20
30 Best
40
50
?50
?40
?30
?20
?10embedding
0
10
20on 30
40
50 for?50
color. See Appendix D.3 for the results on Reuters-R8 datasets.
D.5). The accuracies are shown in Table 2, and we can see that MMCTM outperforms other models
(also see Appendix D.4), except for SVM when L = 20 on the Reuters-R8 dataset. In addition,
MMCTM performs almost as well as using the optimal parameter setting (MMCTM? ).
Sensitivity to the number of topics (i.e., T ). Note the above experiments simply set T = 50. To
validate the affect of T , we varied T from 10 to 100, and the corresponding accuracies are plotted
In Fig. 4 for the two datasets. In both cases, T = 50 seems to be a good parameter value.
Cluster embedding. We finally plot the clustering results by embedding them into the 2dimensional plane using tSNE [38]. In Fig. 5, it can be observed that compared to CTM, MMCTM
generates well separated clusters with much larger margin between clusters.
7
Conclusions
We propose a robust Bayesian max-margin clustering framework to bridge the gap between maxmargin learning and Bayesian clustering, allowing many Bayesian clustering algorithms to be directly equipped with the max-margin criterion. Posterior inference is done via two data augmentation techniques. Two models from the framework are proposed for Bayesian nonparametric maxmargin clustering and topic model based document clustering. Experimental results show our models significantly outperform existing methods with competitive clustering accuracy.
Acknowledgments
This work was supported by an Australia China Science and Research Fund grant (ACSRF-06283)
from the Department of Industry, Innovation, Climate Change, Science, Research and Tertiary Education of the Australian Government, the National Key Project for Basic Research of China (No.
2013CB329403), and NSF of China (Nos. 61322308, 61332007). NICTA is funded by the Australian Government as represented by the Department of Broadband, Communications and the Digital Economy and the Australian Research Council through the ICT Centre of Excellence program.
8
References
[1] J. MacQueen. Some methods of classification and analysis of multivariate observations. In Proc. 5th
Berkeley Symposium on Math., Stat., and Prob., page 281, 1967.
[2] J. Shi and J. Malik. Normalized cuts and image segmentation. TPAMI, 22(8):705?767, 2000.
[3] A. Ben-Hur, D. Horn, H. Siegelmann, and V. Vapnik. Support vector clustering. JMLR, 2:125?137, 2001.
[4] A. Banerjee, S. Merugu, I. S. Dhillon, and J. Ghosh. Clustering with Bregman divergences. JMLR, 6:
1705?1749, 2005.
[5] H. Cheng, X. Zhang, and D. Schuurmans. Convex relaxations of Bregman divergence clustering. In UAI,
2013.
[6] L. Xu, J. Neufeld, B. Larson, and D. Schuurmans. Max-margin clustering. In NIPS, 2005.
[7] B. Zhao, F. Wang, and C. Zhang. Efficient multiclass maximum margin clustering. In ICML, 2008.
[8] Y. F. Li, I. W. Tsang, J. T. Kwok, and Z. H. Zhou. Tighter and convex maximum margin clustering. In
AISTATS, 2009.
[9] G. T. Zhou, T. Lan, A. Vahdat, and G. Mori. Latent maximum margin clustering. In NIPS, 2013.
[10] C. E. Rasmussen. The infinite Gaussian mixture model. In NIPS, 2000.
[11] K. A. Heller and Z. Ghahramani. Bayesian hierarchical clustering. In ICML, 2005.
? III, and D. Roy. Bayesian agglomerative clustering with coalescents. In NIPS, 2008.
[12] Y. W. Teh, H. Daume
? and Z. I. Ying. Binary to bushy: Bayesian hierarchical clustering with
[13] Y. Hu, J. Boyd-Graber, H. Daume,
the Beta coalescent. In NIPS, 2013.
[14] P. Wang, K. B. Laskey, C. Domeniconi, and M. I. Jordan. Nonparametric Bayesian co-clustering ensembles. In SDM, 2011.
[15] J. Zhu, N. Chen, and E. P. Xing. Bayesian inference with posterior regularization and applications to
infinite latent SVMs. JMLR, 2014.
[16] J. Zhu, A. Ahmed, and E. P. Xing. MedLDA: Maximum margin supervised topic models. JMLR, 13(8):
2237?2278, 2012.
[17] J. Zhu, N. Chen, H. Perkins, and B. Zhang. Gibbs max-margin topic models with fast sampling algorithms.
In ICML, 2013.
[18] J. Zhu. Max-margin nonparametric latent feature models for link prediction. In ICML, 2012.
[19] M. Xu, J. Zhu, and B. Zhang. Fast max-margin matrix factorization with data augmentation. In ICML,
2013.
[20] L. Wang, X. Li, Z. Tu, and J. Jia. Discriminative cllustering via generative feature mapping. In AAAI,
2012.
[21] R. Gomes, A. Krause, and P. Perona. Discriminative clustering by regularized information maximization.
In NIPS, 2010.
[22] K.?Ganchev, J. Graa, J. Gillenwater, and B. Taskar. Posterior regularization for structured latent variable
models. JMLR, 11:2001?2049, 2010.
[23] K. Crammer and Y. Singer. On the algorithmic implementation of multiclass kernel-based vector machines. JMLR, 2:265?292, 2001.
[24] C. Fox. An introduction to the calculus of variations. Courier Dover Publications, 1987.
[25] B. Jorgensen. Statistical properties of the generalized inverse Gaussian distribution. Lecture Notes in
Statistics, 1982.
[26] K. P. Murphy. Conjugate Bayesian analysis of the Gaussian distribution. Technical report, UCB, 2007.
[27] S. Favaro and Y. W. Teh. MCMC for normalized random measure mixture models. Stat. Sci., 2013.
[28] D. M. Blei, A. Y. Ng, and M. I. Jordan. Latent Dirichlet allocation. JMLR, 3:993?1022, 2003.
[29] H. M. Wallach. Structured topic models for language. PhD thesis, University of Cambridge, 2008.
[30] A. Jain and M. Law. Data clustering: A user?s dilemma. Lecture Notes in Comp. Sci., 3776:1?10, 2005.
[31] Y.W. Teh, M.I. Jordan, M.J. Beal, and D.M. Blei. Hierarchical Dirichlet processes. J. Amer. Statist.
Assoc., 101(476):1566?1581, 2006.
[32] M. Davy and J. Y. Tourneret. Generative supervised classification using Dirichlet process priors. TPAMI,
32(10):1781?1794, 2010.
[33] P. Franti and O. Virmajoki. Iterative shrinking method for clustering problems. PR, 39(5):761?765, 2006.
[34] K. Bache and M. Lichman. UCI machine learning repository, 2013. URL http://archive.ics.
uci.edu/ml.
[35] A. Shah and Z. Ghahramani. Determinantal clustering process ? a nonparametric bayesian approach to
kernel based semi-supervised clustering. In UAI, 2013.
[36] N. X. Vinh, J. Epps, and J. Bailey. Information theoretic measures for clusterings comparison: variants,
properties, normalization and correction for chance. JMLR, (11):2837?2854, 2010.
[37] V. Sindhwani, P. Niyogi, and M. Belkin. SVMlin: Fast linear SVM solvers for supervised and semisupervised learning. In NIPS Workshop on Machine Learning Open Source Software, 2006. http:
//vikas.sindhwani.org/svmlin.html.
[38] L.J.P. van der Maaten and G.E. Hinton. Visualizing high-dimensional data using t-SNE. JMLR, 9(11):
2579?2605, 2008.
9
| 5305 |@word mild:1 repository:2 briefly:1 changyou:1 proportion:1 seems:1 reused:1 open:1 calculus:2 hu:1 thereby:2 accommodate:1 tnlist:1 contains:1 score:5 selecting:1 lichman:1 initialisation:1 document:22 outperforms:3 existing:6 com:1 si:10 gmail:1 determinantal:1 subsequent:1 partition:1 informative:1 enables:1 analytic:2 drop:1 concert:1 update:1 plot:1 v:1 fund:1 generative:5 half:2 fewer:1 plane:1 dover:1 tertiary:1 blei:2 provides:1 math:1 org:2 zhang:6 five:1 favaro:1 become:1 symposium:1 ik:4 beta:1 combine:1 introduce:4 excellence:1 indeed:1 expected:1 roughly:3 isi:2 growing:1 multi:2 inspired:1 little:1 equipped:1 solver:1 increasing:5 project:1 discover:1 underlying:3 notation:1 moreover:1 lowest:1 unspecified:1 informed:1 ghosh:1 jorgensen:1 pseudo:2 berkeley:1 tackle:1 nutshell:1 classifier:1 assoc:1 stick:1 control:3 normally:1 converse:1 grant:1 generalised:4 engineering:2 local:1 treat:1 tsinghua:3 tends:1 vahdat:1 might:1 au:1 china:4 studied:2 wallach:1 challenging:4 cchangyou:1 co:2 collapse:3 factorization:1 acknowledgment:1 horn:1 testing:1 block:1 empirical:1 significantly:3 convenient:1 projection:1 word:9 boyd:1 courier:1 davy:1 coclustering:1 get:1 convenience:1 selection:2 put:1 applying:1 equivalent:2 deterministic:2 imposed:1 conventional:2 projector:1 shi:1 go:2 straightforward:1 flexibly:2 convex:2 simplicity:1 factorisation:1 rule:1 initialise:2 embedding:3 variation:2 hierarchy:1 suppose:2 target:1 play:1 exact:2 duke:1 user:1 designing:1 regularised:1 fyi:1 element:1 expensive:1 roy:1 bache:1 cut:1 labeled:3 observed:8 bottom:1 role:1 taskar:1 electrical:1 capture:1 wang:3 tsang:1 news:5 decrease:2 highest:1 principled:1 intuition:1 balanced:1 convexity:1 wil:5 complexity:1 xinhua:2 weakly:1 solving:1 segment:1 incur:1 technically:2 mgm:1 dilemma:1 easily:1 accelerate:1 various:1 represented:1 derivation:1 separated:1 distinct:2 fast:5 effective:1 jain:1 whose:1 heuristic:4 larger:3 valued:1 otherwise:2 coalescents:1 statistic:2 niyogi:1 jointly:1 final:1 beal:1 tpami:2 sdm:1 neufeld:1 propose:3 tu:1 uci:3 mixing:2 flexibility:3 achieve:1 validate:1 exploiting:1 cluster:59 regularity:3 empty:1 extending:1 produce:1 generating:1 ring:1 converges:2 ben:1 help:2 develop:4 stat:2 augmenting:2 eq:4 auxiliary:2 implemented:1 indicate:1 australian:4 quantify:1 stochastic:1 exploration:1 coalescent:1 australia:3 education:1 require:1 government:2 assign:1 fix:1 clustered:1 tighter:1 extension:3 exploring:1 correction:1 hold:1 considered:1 ic:1 normal:4 exp:2 mapping:2 algorithmic:1 driving:1 vary:2 wine:1 ctm:11 proc:1 label:4 iw:2 sensitive:2 bridge:1 council:1 ganchev:1 clearly:4 gaussian:20 zhou:2 varying:1 publication:1 encode:1 derived:2 focus:1 modelling:1 likelihood:5 tech:1 contrast:2 baseline:2 obfuscates:1 glass:1 inference:20 helpful:1 economy:1 membership:1 typically:2 hidden:1 perona:1 compatibility:2 issue:2 classification:3 arg:1 html:2 augment:1 k6:3 denoted:3 enrich:1 art:1 integration:1 fairly:1 mutual:1 marginal:2 optimises:1 ng:1 sampling:10 bmc:15 identical:2 unsupervised:6 constitutes:1 icml:5 future:1 summarise:3 report:1 intelligent:1 employ:2 few:2 belkin:1 randomly:2 gamma:1 national:4 divergence:5 murphy:1 vowel:1 subsumed:1 normalisation:1 highly:2 violation:1 mixture:17 devoted:1 implication:1 accurate:1 kt:3 bregman:3 necessary:1 fox:1 circle:1 plotted:1 theoretical:1 weaken:1 complicates:1 column:1 industry:1 svmlin:2 assignment:2 maximization:1 cost:4 introducing:2 subset:2 uniform:1 examining:1 too:2 dependency:1 morph:1 dir:5 synthetic:1 sensitivity:4 probabilistic:3 concrete:2 augmentation:6 aaai:1 thesis:1 choose:3 possibly:1 wishart:3 cb329403:1 external:1 resort:1 zhao:1 li:2 bold:1 satisfy:1 vi:2 performed:1 later:1 vehicle:1 lab:2 doing:1 competitive:1 bayes:3 xing:2 complicated:1 jia:1 vinh:1 contribution:1 ni:4 accuracy:8 php:1 variance:2 merugu:1 efficiently:3 ensemble:1 bayesian:43 raw:4 comp:1 drive:1 acc:1 penalised:2 competitor:1 frequency:2 associated:2 sampled:3 dataset:10 hur:1 color:1 infers:1 segmentation:2 sophisticated:1 higher:1 supervised:8 follow:1 modal:1 formulation:2 done:4 amer:1 generality:2 furthermore:1 parameterised:1 implicit:1 nonlinear:1 banerjee:1 tsne:2 brings:1 lda:1 quality:3 laskey:1 semisupervised:2 usa:1 effect:1 building:1 normalized:2 true:8 remedy:1 regularization:2 dhillon:1 deal:3 climate:1 visualizing:1 during:1 larson:1 iris:1 criterion:3 generalized:1 tt:2 demonstrate:4 theoretic:1 performs:3 image:1 variational:2 novel:1 superior:2 absorbing:1 common:1 multinomial:1 empirically:1 overview:1 insensitive:1 extend:1 interpretation:2 approximates:1 interpret:1 refer:3 zil:7 significant:1 cambridge:1 imposing:1 gibbs:3 fk:4 similarly:1 centre:1 gillenwater:1 language:1 funded:1 moving:1 etc:1 base:3 posterior:32 multivariate:1 showed:1 inf:5 belongs:1 apart:2 moderate:1 scenario:1 binary:1 yi:31 der:1 captured:1 additional:4 converge:1 semi:2 afterwards:1 multiple:2 infer:4 stem:1 technical:2 faster:1 ahmed:1 cross:3 post:5 impact:2 prediction:2 converging:1 basic:1 variant:1 optimisation:3 expectation:3 poisson:1 iteration:5 kernel:2 normalization:1 whereas:1 addition:1 krause:1 addressed:1 source:1 biased:1 rest:2 finely:1 archive:1 probably:1 n1i:1 facilitates:1 incorporates:1 effectiveness:2 jordan:3 call:1 integer:1 leverage:1 iii:1 enough:1 relaxes:3 automated:1 independence:1 affect:1 zi:1 idea:1 cn:1 multiclass:3 allocate:1 url:1 generally:2 useful:1 clear:2 detailed:2 characterise:2 amount:1 nonparametric:10 ten:1 statist:1 svms:1 category:6 generate:4 http:2 outperform:1 nsf:1 estimated:1 correctly:2 discrete:4 promise:1 dropping:1 medlda:3 key:6 salient:1 lan:1 achieving:1 drawn:1 clarity:1 utilize:1 kept:1 relaxation:1 s3vm:2 bushy:1 sum:1 beijing:1 run:1 inverse:4 letter:2 prob:1 bouncing:1 place:1 almost:2 reasonable:1 p3:1 epps:1 draw:2 maaten:1 appendix:8 scaling:1 bound:1 distinguish:1 cheng:1 fold:1 occur:1 constraint:32 perkins:1 software:1 encodes:1 generates:1 extremely:1 department:2 developing:2 according:2 structured:2 conjugate:2 nmi:5 making:1 maxmargin:3 intuitively:1 restricted:1 pr:1 taken:1 computationally:1 mori:1 conjugacy:1 remains:1 slack:2 turn:1 eventually:1 needed:2 singer:1 end:2 adopted:1 gaussians:1 endowed:1 apply:2 kwok:1 hierarchical:4 enforce:2 generic:1 appropriate:1 bailey:1 alternative:2 robustness:1 shah:1 rp:2 franti:1 original:2 vikas:1 denotes:1 dirichlet:17 clustering:75 include:3 running:1 graphical:3 inversed:1 hinge:1 siegelmann:1 restrictive:2 ghahramani:2 objective:1 malik:1 strategy:1 rt:1 traditional:1 exhibit:1 dp:4 subspace:1 link:2 sci:2 landmark:2 topic:34 agglomerative:1 extent:2 barely:1 nicta:2 dataset4:1 assuming:1 code:1 index:1 illustration:1 providing:1 balance:7 innovation:1 nc:1 difficult:1 susceptible:1 ying:1 sne:1 implementation:1 proper:1 unknown:1 allowing:2 upper:1 teh:3 observation:2 datasets:10 macqueen:1 truncated:2 defining:1 maxk:1 incorporated:1 excluding:2 extended:1 communication:1 hinton:1 varied:2 inferred:5 introduced:2 vacuous:2 required:1 kl:5 extensive:2 specified:1 distinction:1 learned:1 established:1 nu:2 nip:7 address:2 able:2 regbayes:12 distribution1:1 below:3 usually:3 challenge:2 summarize:1 program:1 built:1 max:54 including:1 tourneret:1 power:2 critical:1 rely:2 regularized:3 indicator:1 zhu:6 improve:1 technology:1 deemed:1 jun:1 prior:15 ict:2 discovery:1 heller:1 regularisation:1 law:1 loss:3 fully:1 expect:1 lecture:2 interesting:1 proportional:2 allocation:2 validation:3 digital:1 integrate:2 downloaded:1 principle:4 row:2 surprisingly:1 supported:1 soon:2 rasmussen:1 normalised:1 fall:1 pni:1 sparse:1 benefit:1 fragmented:1 boundary:1 dimension:1 xn:1 vocabulary:2 avoids:1 van:1 commonly:1 employing:1 far:1 approximate:1 obtains:1 keep:1 dealing:2 ml:1 instantiation:3 reveals:1 uai:2 unnecessary:1 gomes:1 discriminative:3 xi:26 continuous:1 latent:22 iterative:1 timeefficient:1 table:5 additionally:1 learn:1 reasonably:1 robust:4 schuurmans:2 interact:1 did:1 aistats:1 stabilises:1 arrow:1 motivation:1 decrement:1 hyperparameters:7 reuters:8 daume:2 allowed:1 facilitating:1 repeated:1 graber:1 x1:1 augmented:6 fig:6 canberra:1 broadband:1 xu:2 fashion:3 slow:1 shrinking:1 fails:1 breaking:1 jmlr:9 learns:1 theorem:2 formula:2 specific:1 dpgmm:14 favourable:1 list:1 r8:7 admits:1 svm:6 incorporating:3 intrinsic:1 essential:1 mnist:2 vapnik:1 effectively:2 workshop:1 phd:1 labelling:1 anu:2 margin:54 chen:3 durham:1 gap:1 entropy:1 simply:4 ncut:3 scalar:1 sindhwani:2 corresponds:2 chance:1 dcszj:1 conditional:2 identity:1 formulated:1 kmeans:3 viewed:1 satimage:1 shared:1 absence:1 feasible:1 hard:2 change:2 included:2 typical:1 specifically:2 infinite:5 except:3 determined:1 called:1 total:2 domeniconi:1 experimental:1 ucb:1 formally:1 support:2 crammer:1 violated:1 absolutely:1 incorporate:2 dept:2 mcmc:3 |
4,756 | 5,306 | On Integrated Clustering and Outlier Detection
Linsey Pang
University of Sydney
[email protected]
Lionel Ott
University of Sydney
[email protected]
Sanjay Chawla
University of Sydney
[email protected]
Fabio Ramos
University of Sydney
[email protected]
Abstract
We model the joint clustering and outlier detection problem using an extension
of the facility location formulation. The advantages of combining clustering and
outlier selection include: (i) the resulting clusters tend to be compact and semantically coherent (ii) the clusters are more robust against data perturbations and (iii)
the outliers are contextualised by the clusters and more interpretable. We provide a
practical subgradient-based algorithm for the problem and also study the theoretical properties of algorithm in terms of approximation and convergence. Extensive
evaluation on synthetic and real data sets attest to both the quality and scalability
of our proposed method.
1
Introduction
Clustering and outlier detection are often studied as separate problems [1]. However, it is natural
to consider them simultaneously. For example, outliers can have a disproportionate impact on the
location and shape of clusters which in turn can help identify, contextualize and interpret the outliers.
Pelillo [2] proposed a game theoretic definition of clustering algorithms which emphasis the need for
methods that require as little information as possible while being capable of dealing with outliers.
The area of ?robust statistics? studies the design of statistical methods which are less sensitive to the
presence of outliers [3]. For example, the median and trimmed mean estimators are less sensitive
to outliers than the mean. Similarly, versions of Principal Component Analysis (PCA) have been
proposed [4, 5, 6] which are more robust against model mis-specification and outliers. An important
primitive in the area of robust statistics is the notion of Minimum Covariance Determinant (MCD):
Given a set of n multivariate data points and a parameter `, the objective is to identify a subset of
points which minimizes the determinant of the variance-covariance matrix over all subsets of size
n ? `. The resulting variance-covariance matrix can be integrated into the Mahalanobis distance and
used as part of a chi-square test to identify multivariate outliers [7].
In the theoretical computer science literature, similar problems have been studied in the context
of clustering and facility location. For example, Chen [8] has considered and proposed a constant
factor approximation algorithm for the k-median with outliers problem: Given n data points and
parameters k and `, the objective is to remove a set of ` points such that the cost of k-median
clustering on the remaining n ? ` points is minimized. Our model is similar to the one proposed by
Charikar et. al. [9] who have used a primal-dual formulation to derive an approximation algorithm
for the facility location with outlier problem.
More recently, Chawla and Gionis [10] have proposed k-means--, a practical and scalable algorithm
for the k-means with outlier problem. k-means-- is a simple extension of the k-means algorithm and
is guaranteed to converge to a local optima. However, the algorithm inherits the weaknesses of the
1
classical k-means algorithm. These are: (i) the requirement of setting the number of clusters k and
(ii) initial specification of the k centroids. It is well known that the choice of k and initial set of
centroids can have a disproportionate impact on the result.
In this paper we model clustering and outlier detection as an integer programming optimization task
and then propose a Lagrangian relaxation to design a scalable subgradient-based algorithm. The
resulting algorithm discovers the number of clusters and requires as input: the distance (discrepancy)
between pairs of points, the cost of creating a new cluster and the number ` of outliers to select.
The remainder of the paper is structured as follows. In Section 2 we formally describe the problem as an integer program. In Section 3, we describe the Lagrangian relaxation and details of the
subgradient algorithm. The approximation properties of the relaxation and the convergence of the
subgradient algorithm are discussed in Section 4. Experiments on synthetic and real data sets are
the focus of Section 5 before concluding with Section 6. The supplementary section derives an extension of the affinity propagation algorithm [11] to detect outliers (APOC) - which will be used for
comparison.
2
Problem Formulation
The Facility Location with Outliers (FLO) problem is defined as follows [9]. Given a set of data
points with distances D = {dij }, the cluster creation costs ci and the number of outliers `, we define
the task of clustering and outlier detection as the problem of finding the assignments to the binary
exemplar indicators yj , outlier indicators oi and point assignments xij that minimizes the following
objective function:
X
XX
FLO ? min
cj yj +
dij xij ,
(1)
j
i
j
subject to xij ? yj
X
oi +
xij = 1
X
(2)
(3)
j
oi = `
(4)
i
xij , yj , oi ? {0, 1}.
In order to obtain a valid solution a set of constraints have been imposed:
?
?
?
?
(5)
points can only be assigned to valid exemplars Eq. (2);
every point must be assigned to exactly one other point or declared an outlier Eq. (3);
exactly ` outliers have to be selected Eq. (4);
only integer solutions are allowed Eq. (5).
These constraints describe the facility location problem with outlier detection. This formulation will
allow the algorithm to select the number of clusters automatically and implicitly defines outliers as
those points whose presence in the dataset has the biggest negative impact on the overall solution.
The problem is known to be NP-hard and while approximation algorithms have been proposed, when
distances are assumed to be a metric, there is no known algorithm which is practical, scalable, and
comes with solution guarantees [9]. For example, a linear relaxation of the problem and a solution
using a linear programming solver is not scalable to large data sets as the number of variables is
O(n2 ). In fact we will show that the Lagrangian relaxation of the problem is exactly equivalent to a
linear relaxation and the corresponding subgradient algorithm scales to large data sets, has a small
memory footprint, can be easily parallelized, and does not require access to a linear programming
solver.
3
Lagrangian Relaxation of FLO
The Lagrangian relaxation is based on the following recipe and observations: (i) relax (or dualize)
?tough? constraints of the original FLO problem by moving them to the objective; (ii) associate
2
a Lagrange multiplier (?) with the relaxed constraints which intuitively captures the price of constraints not being satisfied; (iii) For any non-negative ?, FLO(?) is a lower-bound on the FLO
problem. As a function of ?, FLO(?) is a concave but non-differentiable; (iv) Use a subgradient
algorithm to maximize FLO(?) as a function of ? in order to close the gap between the primal and
the dual.
P
More specifically, we relax the constraint oi + j xij = 1 for each i and associate a Lagrange
multiplier ?i with each constraint. Rearranging the terms yields:
X
X
XX
FLO(?) = min
(1 ? oi )?i +
cj yj +
(dij ? ?i )xij .
(6)
|
i
{z
}
outliers
subject to xij ? yi
X
oi = `
|
j
i
j
{z
clustering
}
(7)
(8)
i
0 ? xij , yj , oi ? {0, 1}
?i, j
(9)
We can now solve the relaxed problem with a heuristic finding valid assignments that attempt to
minimize Eq. (6) without optimality guarantees [12]. The Lagrange multipliers ? act as a penalty
incurred for constraint violations which we try to minimize. From Eq. (6) we see that the penalty
influences two parts: outlier selection and clustering. The heuristic starts by selecting good outliers
by designating the ` points with largest ? as outliers, as this removes a large part of the penalty. For
the remaining N ? ` points clustering assignments are found by setting xij = 0 for all pairs for
which dij ? ?i ? 0. To select the exemplars we compute:
X
?j = cj +
(dij ? ?i ),
(10)
i:dij ??i <0
which represents the amortized cost of selecting point j as exemplar and assigning points to it. Thus,
if ?j < 0 we select point j as an exemplar and set yj = 1, otherwise we set yj = 0. Finally, we set
xij = yj if dij ? ?i < 0. From this complete assignment found by the heuristic we compute a new
subgradient st and update the Lagrangian multipliers ?t as follows:
X
sti = 1 ?
xij ? oi
(11)
j
?ti
= max(?it?1 + ?t si , 0),
(12)
where ?t is the step size at time t computed as
?t = ?0 pow(?, t) ? ? (0, 1),
(13)
where pow(a, b) = ab . To obtain the final solution we repeat the above steps until the changes
become small enough, at which point we extract a feasible solution. This is guaranteed to converge
if a step function is used for which the following holds [12]:
lim
n??
n
X
t=1
?t = ?
and
lim ?t = 0.
t??
(14)
A high level algorithm description is given in Algorithm 1.
4
Analysis of Lagrangian Relaxation
In this section, we analyze the solution obtained from using the Lagrangian relaxation (LR) method.
Our analysis will have two parts. In the first part, we will show that the Lagrangian relaxation is
exactly equivalent to solving the linear relaxation of the FLO problem. Thus if FLO(IP), FLO(LP)
and FLO(LR) are the optimal value of integer program, linear relaxation and linear programming
solution respectively, we will show that FLO(LR) = FLO(LP). In the second part, we will analyze
the convergence rate of the subgradient method and the impact of outliers.
3
Algorithm 1: LagrangianRelaxation()
Initialize ?0 , x0 , t
while not converged do
st ? ComputeSubgradient(xt?1 )
?t ? ComputeLambda(st )
xt ? FLO(?t )
(solve via heuristic)
t?t+1
end
?
?
?
?
?
?
?
?
?
?
?
A=?
?
?
?
?
?
?
?
?
?
?
?
?1
1
?1
0
?1
1
0
0
?1
1
1
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
Figure 1: Visualization of the building blocks of the A matrix. The top left is a n2 ? n2 identity
matrix which is followed by n row stacked blocks of n ? n negative identity matrices. To the right
of those is another n2 ? n block of zeros. The final row in the block matrix consists of n2 + n zeros
followed by n ones.
4.1
Quality of the Lagrangian Relaxation
P
2
Consider the constraint set L = {(x, y, o) ? Zn +2n |xij ? yj ? i oi ? ` ? i, j}. Then it is well
known that the optimal value of FLO(LR) of the Lagrangian relaxation is equal to the cost of the
following optimization problem [12]:
min
X
cj yj +
j
oi +
X
XX
i
xij dij
(15)
j
xij = 1
(16)
j
conv(L)
(17)
where conv(L) is the convex hull of the set L. We now show that L is integral and therefore
X
2
conv(L) = {(x, y, o) ? Rn +2n |xij ? yj ?
oi ? ` ? i, j}
i
This in turn will imply that FLO(LR) = FLO(LP). In order to show that L is integral, we will establish
that that the constraint matrix corresponding to the set L is totally unimodular (TU). For completeness, we recall several important definitions and theorems from integer program theory [12]:
Definition 1. A matrix A is totally unimodular if every square submatrix of A, has determinant in
the set {?1, 0, 1}.
n
Proposition 1. Given a linear program: min{cT x : Ax ? b, x ? R+
}, let b be the set of integer
vectors for which the problem instance has finite value. Then the optimal solution has integral
solutions if A is totally unimodular.
An equivalent definition of total unimodularity (TU) and often easier to establish is captured in the
following theorem.
Theorem 1. Let A be a matrix. Then A is TU iff for any subset of rows X of A, there exists a
coloring of rows of X, with 1 or -1 such that the weighted sum of every column (while restricting
the sum to rows in X) is -1, 0 or 1.
We are now ready to state and prove the main theorem in this section.
4
Theorem 2. The matrix corresponding to the constraint set L is totally unimodular.
Proof. We need to consider the constraints
xij ? yj ? i, j
n
X
i=1
(18)
oi ? `
(19)
We can express the above constraints in the form Au = b where u is the vector:
T
u = [x11 , . . . , x1n , . . . , xn1 , . . . , xnn , y1 , . . . , yn , o1 , . . . , on ]
The block matrix A is of the form:
I
A=
0
B
0
0
1
(20)
(21)
Here I is an n2 ? n2 identity matrix, B is stack of n matrices of size n ? n where each element of
the stack is a negative identity matrix, and 1 is an 1 ? n block of 10 s. See Figure 1 for a detailed
visualization.
Now to prove that A is TU, we will use Theorem 1. Take any subset X of rows of A. Whether we
color the rows of X by 1 or -1, the column sum (within X) of a column of I will be in {?1, 0, 1}.
A similar argument holds for columns of the block matrix 1. Now consider the submatrix B. We
can express X as
X = ?ni=1,i?B(X,:) Xi
(22)
where each Xi = {r ? X|X(r, i) = ?1}. Given that B is a stack of negative diagonal matrices,
Xi ? Xj = ? for i 6= j. Now consider a column j of B. If Xj has even number of ?10 s, then split
the elements of Xj evenly and color one half as 1 and the other as ?1. Then the sum of column j
(for rows in X) will be 0. On the other hand, if another set of rows Xk has odd number of ?1, color
the rows of Xk alternatively with 1 and ?1. Since Xj and Xk are disjoint their colorings can be
carried out independently. Then the sum of column j will be 1 or ?1. Thus we satisfy the condition
of Theorem 1 and conclude that A is TU.
4.2
Convergence of Subgradient Method
As noted above, the langrangian dual is given by max{FLO(?)|? ? 0}. Furthermore, we use a
gradient ascent method to update the ??s as [?ti ]ni=1 = max(?t?1
+ ?t si , 0) where sti = 1 ?
i
P
j xij ? oi and ?t is the step-size.
Now, assuming that the norm of the subgradients are bounded, i.e., ksk2 ? G and the distance
between the initial point and the optimal set, k?1 ? ?? k2 ? R, it is known that [13]:
Pt
R2 + G2 i=1 ?i2
t
?
|Z(? ) ? Z(? )| ?
Pt
2 i=1 ?i
This can be used to show that to obtain accuracy (for any step size), the number of iterations is
lower bounded by O(RG/2 ), We examine the impact of integrating clustering and outliers on the
convergence rate. We make the following observations:
P
Observation 1. At a given iteration t and for a given data point i, if oti = 1 then j xtij = 0 and
sti = 0 and therefore ?t+1
= ?ti .
i
Observation 2. At a given iteration tP
and for a given data point i, if oti = 0 and the point i is
assigned to exactly one exemplar, then j xtij = 1 and therefore sti = 0 and ?t+1
= ?ti .
i
In conjunction with the algorithm for solving FLO(?) and the above observations we can draw
important conclusions regarding the behavior of the algorithm including (i) the ? values associated
with outliers will be relatively larger and stabilize earlier and (ii) the ? values of the exemplars will
be relatively smaller and will take longer to stabilize.
5
5
Experiments
In this section we evaluate the proposed method on both synthetic and real data and compare it to
other methods. We first present experiments using synthetic data to show quantitative analysis of
the methods in a controlled environment. Then, we present clustering and outlier results obtained
on the MNIST image data set.
We compare our Langrangian Relaxation (LR) based method to two other methods, k-means-- and
an extension of affinity propagation [11] to outlier clustering (APOC) whose details can be found in
the supplementary material. Both LR and APOC require a cost for creating clusters. We obtain this
value as ? ? median(dij ), i.e. the median of all distances multiplied by a scaling factor ? which typically is in the range [1, 30]. The initial centroids required by k-means-- are found using k-means++
[14] and unless specified otherwise k-means-- is provided with the correct number of clusters k.
5.1
Synthetic Data
We use synthetic datasets for controlled performance evaluation and comparison between the different methods. The data is generated by randomly sampling k clusters with m points, each from
d-dimensional normal distributions N (?, ?) with randomly selected ? and ?. To these clusters
we add ` additional outlier points that have a low probability of belonging to any of the selected
clusters. The distance between points is computed using the Euclidean distance. We focus on 2D
distributions as they are more challenging then higher dimensional data due to the separability of
the data.
To assess the performance of the methods we use the following three metrics:
1. Normalized Jaccard index, measures how accurately a method selects the ground truth outliers. It is a coefficient computed between selected outliers O and ground-truth outliers O? .
The final coefficient is normalized with regards to the best possible coefficient obtainable
in the following way:
J(O, O? ) =
|O ? O? | min(|O|, |O? |)
/
.
|O ? O? | max(|O|, |O? |)
(23)
2. Local outlier factor [15] (LOF) measures the outlier quality of a point. We compute the
ratio between the average LOF of O and O? , which indicates the quality of the set of
selected outliers.
3. V-Measure [16] indicates the quality of the overall clustering solution. The outliers are
considered as an additional class for this measure.
For the Jaccard index and V-Measure a value of 1 is optimal, while for the LOF factor a larger value
is better.
Since the number of outliers `, required by all methods, is typically not known exactly we explore
how its misspecification affects the results. We generate 2D datasets with 2000 inliers and 200
outliers and vary the number of outliers ` selected by the methods. The results in Figure 2 show
that in general none of the methods fail completely if the value of ` is misspecified. Looking at the
Jaccard index, which indicates the percentage of true outliers selected, we see that if ` is smaller
then the true number of outliers all methods pick only outliers. When ` is greater then the true
number of outliers we can see a that LR and APOC improve with larger ` while k-means-- does only
sometimes. This is due to the formulation of LR which selects the largest outliers, which APOC
does to some extent as well. This means that if some outliers are initially missed they are more
likely to be selected if ` is larger then the true number of outliers. Looking at the LOF ratio we
can see that selecting more outliers then present in the data set reduces the score somewhat but not
dramatically, which provides the method with robustness. Finally, V-Measure results show that the
overall clustering results remain accurate, even if the number of outliers is misspecified.
We experimentally investigate the quality of the solution by comparing with the results obtained by
solving the LP relaxation using CPLEX. This comparison indicates what quality can be typically expected from the different methods. Additionally, we can evaluate the speed of these approximations.
We evaluate 100 datasets, consisting of 2D Gaussian clusters and outliers, with varying number of
6
0
100 200 300 400
Selected Outliers (`)
V-Measure
0.5
k-means-APOC
LR
1
1
LOF Ratio
Jaccard Index
1
0.5
0
100 200 300 400
Selected Outliers (`)
0.5
0
100 200 300 400
Selected Outliers (`)
2,000
20
10
APOC
LR
Time (s)
APOC
LR
30
Time (s)
Speedup
Figure 2: The impact of number of outliers specified (`) on the quality of the clustering and outlier detection performance. LR and APOC perform similarly with more stability and better outlier
choices compared to k-means--. We can see that overestimating ` is more detrimental to the overall
performance, as indicated by the LOF Ratio and V-Measure, then underestimating it.
1,000
0
0
500
1,000 1,500 2,000
Data Points
(a) Speedup over LP
100
10?2
?4
0
5,000
Data Points
(b) Total Runtime
10,000
10
APOC
LR
0
5,000
10,000
Data Points
(c) Time per Iteration
Figure 3: The graphs shows how the number of points influences different measures. In (a) we
compare the speedup of both LR and APOC over LP. (b) compares the total runtime needed to solve
the clustering problem for LR and APOC . Finally, (c) plots the time required (on a log scale) for a
single iteration for LR and APOC.
points. On average LR obtains 94% ? 5% of the LP objective value, APOC obtains an energy that is
95% ? 4% of the optimal solution found by LP and k-means--, with correct k, obtains 86% ? 12% of
the optimum. These results reinforce the previous analysis; LR and APOC perform similarly while
outperforming k-means--. Next we look at the speed-up of LR and APOC over LP. Figure 3 a) shows
both methods are significantly faster with the speed-up increasing as the number of points increases.
Overall for a small price in quality the two methods obtain a significantly faster solution. k-means-outperforms the other two methods easily with regards to speed but has neither the accuracy nor the
ability to infer the number of clusters directly from the data.
Next we compare the runtime of LR and APOC. Figure 3 b) shows the overall runtime of both
methods for varying number of data points. Here we observe that APOC is faster then LR, however,
by observing the time a single iteration takes, shown in Figure 3 c), we see that LR is much faster
on a per iteration basis compared to APOC. In practice LR requires several times the number of
iterations of APOC, which is affected by the step size function used. Using a more sophisticated
method of computing the step size will provide large gains to LR. Finally, the biggest difference
between LR and APOC is that the latter requires all messages and distances to be held in memory.
This obviously scales poorly for large datasets. Conversely, LR computes the distances at runtime
and only needs to store indicator vectors and a sparse assignment matrix, thus using much less
memory. This makes LR amenable to processing large scale datasets. For example, with single
precision floating point numbers, dense matrices and 10 000 points APOC requires around 2200 MB
of memory while LR only needs 370 MB. Further gains can be obtained by using sparse matrices
which is straight forward in the case of LR but complicated for APOC.
5.2
MNIST Data
The MNIST dataset, introduced by LeCun et al. [17], contains 28 ? 28 pixel images of handwritten
digits. We extract features from these images by representing them as 768 dimensional vectors which
is reduced to 25 dimensions using PCA. The distance between these vectors is computed using the
L2 norm. In Figure 4 we show exemplary results obtained when processing 10 000 digits with the
7
(a) Digit 1
(b) Digit 4
(c) Outliers
Figure 4: Each row in (a) and (b) shows a different appearance of a digit captured by a cluster. The
outliers shown in (c) tend to have heavier then usual stroke, are incomplete or are not recognizable
as a digit.
Table 1: Evaluation of clustering results of the MNIST data set with different cost scaling values ?
for LR and APOC as well as different settings for k-means--. We can see that increasing the cost
results in fewer clusters but as a trade off reduces the homogeneity of the clusters.
LR
?
V-Measure
Homogeneity
Completeness
Clusters
APOC
k-means--
5
15
25
15
n.a.
n.a.
0.52
0.78
0.39
120
0.67
0.74
0.61
13
0.54
0.65
0.46
27
0.53
0.72
0.42
51
0.51
0.50
0.52
10
0.58
0.75
0.47
40
LR method with ? = 5 and ` = 500. Each row in Figure 4 a) and b) shows examples of clusters
representing the digits 1 and 4, respectively. This illustrates how different the same digit can appear
and the separation induced by the clusters. Figure 4 c) contains a subset of the outliers selected by
the method. These outliers have different characteristics that make them sensible outliers, such as:
thick stroke, incomplete, unrecognizable or ambiguous meaning.
To investigate the influence the cluster creation cost has we run the experiment with different values
of ?. In Table 1 we show results for LR with values of cost scaling factor ? = {5, 15, 25}, APOC
with ? = 15 and k-means-- with k = {10, 40}. We can see that LR obtains the best V-Measure score
out of all methods with ? = 15. The homogeneity and completeness scores reflect this as well, while
homogeneity is similar to other settings the completeness value is much better. Looking at APOC we
see that it struggles to obtain the same quality as LR. In the case of k-means-- we can observed how
providing the algorithm with the actual number of clusters results in worse performance compared
to a larger number of clusters which highlights the advantage of methods capable of automatically
selecting the number of clusters from the data.
6
Conclusion
In this paper we presented a novel approach to joint clustering and outlier detection formulated
as an integer program. The method only requires pairwise distances and the number of outliers
as input and detects the number of clusters directly from the data. Using a Lagrangian relaxation
of the problem formulation, which is solved using a subgradient method, we obtain a method that
is provably equivalent to a linear programming relaxation. Our proposed algorithm is simple to
implement, highly scalable, and has a small memory footprint. The clusters and outliers found by
the algorithm are meaningful and easily interpretable.
8
References
[1] V. Chandola, A. Banerjee, and V. Kumar. Anomaly detection: A survey. ACM Computing Surveys, 2009.
[2] M. Pelillo. What is a Cluster? Perspectives from Game Theory. In Proc. of Advances in Neural Information Processing Systems, 2009.
[3] P. Huber and E. Ronchetti. Robust Statistics. Wiley, 2008.
[4] C. Croux and A. Ruiz-Gazen. A Fast Algorithm for Robust Principal Components Based on Projection
Pursuit. In Proc. in Computational Statistics, 1996.
[5] J. Wright, A. Ganesh, S. Rao, Y. Peng, and Y. Ma. Robust Principal Component Analysis: Exact Recovery
of Corrupted Low-Rank Matrices by Convex Optimization. In Proc. of Advances in Neural Information
Processing Systems, 2009.
[6] Emmanuel J. Cand`es, Xiaodong Li, Yi Ma, and John Wright. Robust principal component analysis? J.
ACM, 58(3):11:1?11:37, June 2011. ISSN 0004-5411.
[7] P.J. Rousseeuw and K.V. Driessen. A fast algorithm for the minimum covariance determinant estimator.
Technometrics, 1999.
[8] K. Chen. A constant factor approximation algorithm for k-median clustering with outliers. In Proc. of the
ACM-SIAM Symposium on Discrete Algorithms, 2008.
[9] M. Charikar, S. Khuller, D. M. Mount, and G. Narasimhan. Algorithms for Facility Location Problems
with Outliers. In Proc. of the ACM-SIAM Symposium on Discrete Algorithms, 2001.
[10] S. Chawla and A. Gionis. k-means?: A Unified Approach to Clustering and Outlier Detection. In SIAM
International Conference on Data Mining, 2013.
[11] B. Frey and D. Dueck. Clustering by Passing Messages Between Data Points. Science, 2007.
[12] D. Bertsimas and R. Weismantel. Optimization over Integers. Dynamic Ideas Belmont, 2005.
[13] Stephen Boyd and Lieven Vandenberghe. Convex Optimization. Cambridge University Press, New York,
NY, USA, 2004. ISBN 0521833787.
[14] D. Arthur and S. Vassilvitskii. k-means++: The Advantages of Careful Seeding. In ACM-SIAM Symposium on Discrete Algorithms, 2007.
[15] M. Breunig, H. Kriegel, R. Ng, and J. Sander. LOF: Identifying Density-Based Local Outliers. In
Int. Conf. on Management of Data, 2000.
[16] A. Rosenberg and J. Hirschberg. V-Measure: A conditional entropy-based external cluster evaluation
measure. In Proc. of the Joint Conference on Empirical Methods in Natural Language Processing and
Computational Natural Language Learning, 2007.
[17] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition.
Proceedings of the IEEE, 1998.
9
| 5306 |@word determinant:4 version:1 norm:2 covariance:4 pick:1 ronchetti:1 initial:4 contains:2 score:3 selecting:4 document:1 outperforms:1 comparing:1 si:2 assigning:1 must:1 john:1 belmont:1 shape:1 remove:2 plot:1 interpretable:2 update:2 seeding:1 half:1 selected:12 fewer:1 xk:3 underestimating:1 lr:36 completeness:4 provides:1 location:7 become:1 symposium:3 consists:1 prove:2 recognizable:1 pairwise:1 x0:1 peng:1 huber:1 expected:1 behavior:1 cand:1 examine:1 nor:1 chi:1 detects:1 automatically:2 little:1 actual:1 solver:2 totally:4 conv:3 provided:1 xx:3 bounded:2 increasing:2 what:2 minimizes:2 narasimhan:1 unified:1 finding:2 guarantee:2 dueck:1 quantitative:1 every:3 act:1 concave:1 ti:4 unimodular:4 runtime:5 exactly:6 k2:1 yn:1 appear:1 before:1 local:3 frey:1 struggle:1 mount:1 emphasis:1 au:5 studied:2 conversely:1 challenging:1 range:1 practical:3 lecun:2 yj:13 practice:1 block:7 implement:1 footprint:2 digit:8 area:2 empirical:1 significantly:2 projection:1 boyd:1 integrating:1 close:1 selection:2 context:1 influence:3 equivalent:4 imposed:1 lagrangian:12 primitive:1 independently:1 convex:3 survey:2 recovery:1 identifying:1 estimator:2 vandenberghe:1 stability:1 notion:1 pt:2 anomaly:1 programming:5 exact:1 designating:1 breunig:1 associate:2 amortized:1 element:2 recognition:1 observed:1 solved:1 capture:1 trade:1 environment:1 dynamic:1 solving:3 creation:2 completely:1 basis:1 easily:3 joint:3 stacked:1 fast:2 describe:3 whose:2 heuristic:4 supplementary:2 solve:3 larger:5 relax:2 otherwise:2 ability:1 statistic:4 final:3 ip:1 obviously:1 advantage:3 differentiable:1 exemplary:1 isbn:1 propose:1 mb:2 remainder:1 tu:5 combining:1 iff:1 poorly:1 description:1 flo:21 scalability:1 recipe:1 convergence:5 cluster:30 lionel:1 optimum:2 requirement:1 unrecognizable:1 help:1 derive:1 exemplar:7 odd:1 pelillo:2 eq:6 sydney:7 disproportionate:2 come:1 thick:1 correct:2 hull:1 material:1 require:3 proposition:1 extension:4 hold:2 around:1 considered:2 ground:2 normal:1 lof:7 wright:2 vary:1 proc:6 sensitive:2 largest:2 weighted:1 gaussian:1 varying:2 rosenberg:1 conjunction:1 ax:1 inherits:1 focus:2 june:1 rank:1 indicates:4 centroid:3 detect:1 integrated:2 typically:3 initially:1 selects:2 provably:1 pixel:1 overall:6 dual:3 x11:1 initialize:1 equal:1 ng:1 sampling:1 represents:1 look:1 discrepancy:1 minimized:1 np:1 overestimating:1 randomly:2 simultaneously:1 homogeneity:4 floating:1 consisting:1 cplex:1 attempt:1 ab:1 detection:10 technometrics:1 message:2 investigate:2 highly:1 mining:1 evaluation:4 weakness:1 violation:1 primal:2 inliers:1 held:1 amenable:1 accurate:1 integral:3 capable:2 arthur:1 unless:1 iv:1 euclidean:1 incomplete:2 theoretical:2 instance:1 column:7 earlier:1 rao:1 tp:1 zn:1 assignment:6 ott:1 cost:10 subset:5 dij:9 contextualize:1 corrupted:1 synthetic:6 st:3 density:1 langrangian:2 siam:4 international:1 off:1 reflect:1 satisfied:1 management:1 worse:1 conf:1 creating:2 external:1 li:1 stabilize:2 chandola:1 coefficient:3 gionis:2 int:1 satisfy:1 try:1 analyze:2 observing:1 attest:1 start:1 mcd:1 complicated:1 ass:1 pang:1 ni:2 square:2 oi:14 variance:2 characteristic:1 who:1 minimize:2 yield:1 identify:3 accuracy:2 handwritten:1 accurately:1 none:1 straight:1 converged:1 stroke:2 definition:4 against:2 energy:1 proof:1 mi:1 associated:1 xn1:1 gain:2 dataset:2 recall:1 lim:2 color:3 cj:4 obtainable:1 sophisticated:1 coloring:2 higher:1 formulation:6 furthermore:1 until:1 hand:1 dualize:1 ganesh:1 banerjee:1 propagation:2 defines:1 quality:10 indicated:1 building:1 xiaodong:1 usa:1 normalized:2 multiplier:4 true:4 facility:6 assigned:3 i2:1 mahalanobis:1 game:2 ambiguous:1 unimodularity:1 x1n:1 noted:1 theoretic:1 complete:1 image:3 meaning:1 discovers:1 recently:1 novel:1 misspecified:2 discussed:1 lieven:1 interpret:1 cambridge:1 similarly:3 language:2 moving:1 specification:2 access:1 longer:1 add:1 multivariate:2 perspective:1 store:1 binary:1 outperforming:1 yi:2 captured:2 minimum:2 additional:2 relaxed:2 greater:1 somewhat:1 parallelized:1 converge:2 maximize:1 ii:4 stephen:1 reduces:2 infer:1 faster:4 controlled:2 impact:6 scalable:5 pow:2 metric:2 iteration:8 sometimes:1 median:6 ascent:1 subject:2 tend:2 induced:1 tough:1 integer:8 presence:2 iii:2 enough:1 split:1 sander:1 bengio:1 xj:4 affect:1 regarding:1 idea:1 haffner:1 whether:1 vassilvitskii:1 pca:2 heavier:1 trimmed:1 penalty:3 passing:1 york:1 dramatically:1 detailed:1 rousseeuw:1 reduced:1 generate:1 xij:18 percentage:1 driessen:1 disjoint:1 per:2 discrete:3 affected:1 express:2 neither:1 graph:1 subgradient:10 relaxation:19 bertsimas:1 sum:5 sti:4 run:1 weismantel:1 missed:1 separation:1 draw:1 scaling:3 jaccard:4 submatrix:2 bound:1 ct:1 guaranteed:2 followed:2 croux:1 gazen:1 constraint:13 declared:1 speed:4 argument:1 min:5 concluding:1 optimality:1 subgradients:1 kumar:1 relatively:2 speedup:3 charikar:2 structured:1 belonging:1 smaller:2 remain:1 separability:1 lp:9 outlier:74 intuitively:1 visualization:2 turn:2 fail:1 needed:1 end:1 ksk2:1 pursuit:1 multiplied:1 observe:1 chawla:4 robustness:1 original:1 top:1 clustering:24 include:1 remaining:2 xtij:2 emmanuel:1 establish:2 classical:1 objective:5 usual:1 diagonal:1 affinity:2 usyd:1 fabio:2 distance:12 separate:1 gradient:2 detrimental:1 reinforce:1 sensible:1 evenly:1 extent:1 assuming:1 issn:1 o1:1 index:4 ratio:4 providing:1 negative:5 design:2 perform:2 observation:5 datasets:5 finite:1 looking:3 misspecification:1 y1:1 rn:1 perturbation:1 stack:3 introduced:1 pair:2 required:3 specified:2 extensive:1 coherent:1 kriegel:1 sanjay:2 program:5 max:4 memory:5 including:1 natural:3 ramos:2 indicator:3 representing:2 improve:1 imply:1 ready:1 carried:1 extract:2 literature:1 l2:1 highlight:1 incurred:1 row:12 repeat:1 allow:1 sparse:2 regard:2 dimension:1 valid:3 computes:1 forward:1 compact:1 uni:1 implicitly:1 obtains:4 dealing:1 assumed:1 conclude:1 xi:3 alternatively:1 table:2 additionally:1 robust:8 rearranging:1 bottou:1 main:1 dense:1 n2:7 allowed:1 biggest:2 ny:1 wiley:1 precision:1 oti:2 ruiz:1 theorem:7 xt:2 r2:1 derives:1 exists:1 mnist:4 restricting:1 ci:1 illustrates:1 chen:2 gap:1 easier:1 rg:1 entropy:1 explore:1 likely:1 appearance:1 lagrange:3 khuller:1 g2:1 xnn:1 truth:2 acm:5 ma:2 conditional:1 identity:4 formulated:1 careful:1 price:2 feasible:1 hard:1 change:1 experimentally:1 specifically:1 semantically:1 principal:4 total:3 e:1 meaningful:1 formally:1 select:4 latter:1 evaluate:3 |
4,757 | 5,307 | Convex Optimization Procedure for Clustering:
Theoretical Revisit
Changbo Zhu
Department of Electrical and Computer Engineering
Department of Mathematics
National University of Singapore
[email protected]
Chenlei Leng
Department of Statistics
University of Warwick
[email protected]
Huan Xu
Department of Mechanical Engineering
National University of Singapore
[email protected]
Shuicheng Yan
Department of Electrical and Computer Engineering
National University of Singapore
[email protected]
Abstract
In this paper, we present theoretical analysis of SON ? a convex optimization
procedure for clustering using a sum-of-norms (SON) regularization recently proposed in [8, 10, 11, 17]. In particular, we show if the samples are drawn from two
cubes, each being one cluster, then SON can provably identify the cluster membership provided that the distance between the two cubes is larger than a threshold
which (linearly) depends on the size of the cube and the ratio of numbers of samples in each cluster. To the best of our knowledge, this paper is the first to provide
a rigorous analysis to understand why and when SON works. We believe this may
provide important insights to develop novel convex optimization based algorithms
for clustering.
1
Introduction
Clustering is an important problem in unsupervised learning that deals with grouping observations
(data points) appropriately based on their similarities or distances [20]. Many clustering algorithms have been proposed in literature, including K-means, spectral clustering, Gaussian mixture models and hierarchical clustering, to solve problems with respect to a wide range of cluster
shapes. However, much research has pointed out that these methods all suffer from instabilities
[3, 20, 16, 15, 13, 19]. Taking K-means as an example, the formulation of K-means is NP-hard and
the typical way to solve it is the Lloyd?s method, which requires randomly initializing the clusters.
However, different initialization may lead to significantly different final cluster results.
1.1
A Convex Optimization Procedure for Clustering
Recently, Lindsten et al. [10, 11], Hocking et al. [8] and Pelckmans et al. [17] proposed the
following convex optimization procedure for clustering, which is termed as SON by Lindsten et al.
[11] (Also called Clusterpath by Hocking et al. [8]),
X
? = arg min kA ? Xk2 + ?
X
kXi? ? Xj? k2 .
(1)
F
X?Rn?p
i<j
Here A is a given data matrix of dimension n ? p where each row is a data point, ? is a tunable
parameter to determine the number of clusters, k ? kF denotes the Frobenius norm and Xi? denotes
the ith row of X.
1
The main idea of the algorithm is that if the i-th sample and the j-th sample belong to the same
? i? and X
? j? should be equal. Intuitively, this is due to the fact that the second term is a
cluster, then X
? to be the same, and can be seen as a generalization
regularization term that enforces the rows of X
? From another
of the fused Lasso penalty [18]. In particular, this penalty seeks to fuse the rows of X.
point of view, the regularization term can be seen as an `1,2 norm, i.e., the sum of `2 norm. Such a
norm is known to encourage block sparse (in this case row-sparse) solutions [1]. Thus, it is expected
? i? ? X
? j? = 0.
that for many (i, j) pairs, X
Mathematically, given c disjoint clusters {C1 , C2 , ? ? ? , Cc } with Ci ? Rp for i = 1, 2, ? ? ? , c, we
define the Cluster Membership Matrix of a given data matrix A as the following.
Definition 1. Given a data matrix A of dimension n ? p, for j = 1, 2, ? ? ? , c, set Ij = {i | Ai? ?
Cj , 1 ? i ? n}. We say that a matrix X of dimension n ? p is a Cluster Membership Matrix of A
if
Xi? = Xj?
if i ? Ik , j ? Ik and 1 ? k ? c
Xi? 6= Xj?
if i ? Im , j ? Il , 1 ? m ? c, 1 ? l ? c and m 6= l.
? of Problem (1) is a Cluster Membership Matrix
Given a data matrix A, if the optimal solution X
?
of A, then we can determine the cluster membership by simply grouping the identical rows of X
together. We say that SON successfully recovers the cluster membership of A in this case.
Notice that unlike previous approaches, SON does not suffer from the instability issue since it is a
strictly convex optimization problem and the solution is fixed once a data matrix A is given. Moreover, SON can easily be adapted to incorporate a priori knowledge of the clustering membership.
For example, if we have prior knowledge about which points are more likely to be in the same
cluster,
P we can appropriately weight the regularization term, i.e., change the regularization term to
? i<j ?ij kXi? ? Xj? k2 for some ?ij > 0.
The main contribution of this paper is to provide theoretic analysis of SON, in particular to derive
sufficient conditions when SON successfully recovers the clustering membership. We show that
if there are two clusters, each of which is a cube, then SON succeeds provided that the distance
between the cubes is larger than a threshold value that depends on the cube size and the ratio of
number of samples drawn in each cluster. Thus, the intuitive argument about why SON works is
made rigorous and mathematically solid. To the best of our knowledge, this is the first attempt to
theoretically quantify why and when SON succeeds.
Related Work: we briefly review the related works on SON. Hocking et al. [8] proposed SON,
arguing that it can be seen as a generalization of hierarchical clustering, and presented via numerical
simulations several situations in which SON works while K-means and average linkage hierarchical
clustering fail. They also developed R package called ?clusterpath? which can be used to solve
Problem (1). Independently, Lindsten et al. [10, 11] derived SON as a convex relaxation of Kmeans clustering. In the algorithmic aspect, Chi et al. [6] developed two methods to solve Problem
(1), namely, Alternating Direction Method of Multipliers (ADMM) and alternating minimization
algorithm (AMA). Marchetti et al. [14] generalized SON to the high-dimensional and noisy cases.
Yet, in all these works, no attempt has been made to study rigorously why and when SON succeeds.
Notation: in this paper, matrices are denoted by upper case boldface letters (e.g. A, B), sets are
denoted by blackboard bold characters (e.g. R, I, C) and operators are denoted by Fraktur characters
(e.g. D, M). Given a matrix A, we use Ai? to denote its ith row, and A?j to denote its jth column.
Its (i, j)th entry is denoted by Ai,j . Two norms are used: we use k ? kF to denote the Frobenius
norm and k ? k2 to denote the l2 norm of a vector. The space spanned by the rows of A is denoted
by Row(A). Moreover, given a matrix A of dimension n ? p and a function f : Rp 7? Rq , we use
the notation f (A) to denote the matrix whose ith row is f (Ai? ).
2
Main Result
In this section we present our main theoretic result ? a provable guarantee when SON succeeds in
identifying cluster membership.
2
2.1
Preliminaries
We first define some operators that will be frequently used in the remainder of the paper.
Definition 2. Given any two matrices E of dimension n1 ? p and F of dimension n2 ? p, define
the difference operator D1 on E, D2 on the two matrices E, F and D on the matrix constructed by
concatenating E and F vertically as
?
?
?
?
E1? ? E2?
E1? ? F1?
? E1? ? F2? ?
? E1? ? E3? ?
?
?
?
?
..
..
?
?
?
?
?
?
?
?
.
.
?
?
?
?
!
? E1? ? Fn2 ? ?
? E1? ? En1 ? ?
D1 (E)
? E ?F ?
?
? E ?E
E
2?
3?
? , D2 (E, F) = ? 2?
1? ? and D(
D1 (F)
.
)=
D1 (E) = ?
?
?
?
?
..
..
F
?
?
?
?
D
(E,
F)
2
.
.
?
?
?
?
?E ?F ?
?
? E ?E
2?
n1 ?
n2 ? ?
? 2?
?
?
?
?
?
?
..
..
?
?
?
?
.
.
E(n1 ?1)? ? En1 ?
En1 ? ? Fn2 ?
In words, the operator D1 calculates the difference between every two rows of a matrix and lists the
results in the order indicated in the definition. Similarly, given two matrices E and F, the operator
D2 (E, F) calculates the difference of any two rows between E and F, one from E and the other from
F. We also define the following average operation which calculates the mean of the row vectors.
Definition 3. Given any matrix E of dimension n ? p, define the average operator on E as
M(E) =
n
1 X
(
Ei? ).
n i=1
Definition 4. A matrix E is called column centered if M(E) = 0.
2.2
Theoretical Guarantees
Our main result essentially says that when there are two clusters, each of which is a cube, and
they are reasonably separated away from each other, then SON successfully recovers the cluster
membership. We now make this formal. For i = 1, 2, suppose Ci ? Rp is a cube with center
(?i1 , ?i2 , ? ? ? , ?ip ) and edge length si = 2(?i1 , ?i2 , ? ? ? , ?ip ) , i.e.,
Ci = [?i1 ? ?i1 , ?i1 + ?i1 ] ? ? ? ? ? [?ip ? ?ip , ?ip + ?ip ].
Definition 5. The distance d1,2 between cubes C1 and C2 is
d1,2 , inf{kx ? yk2 | x ? C1 , y ? C2 }.
Definition 6. The weighted size w1,2 with respect to C1 , C2 , n1 and n2 is defined as
2n2 (n1 ? 1)
2n1 (n2 ? 1)
w1,2 = max
+ 1 ks1 k2 ,
+ 1 ks2 k2 .
n21
n22
Theorem 1. Given a column centered data matrix A of dimension n ? p, where each row is arbitrarily picked from either cube C1 or cube C2 and there are totally ni rows chosen from Ci for
i = 1, 2, if w1,2 < d1,2 , then by choosing the parameter ? ? R such that w1,2 < n2 ? < d1,2 , we
have the following:
1. SON can correctly determine the cluster membership of A;
2. Rearrange the rows of A such that
?
?
Ai1?
1
? Ai2? ?
A
?
?
i
A=
and
A
=
? .. ? ,
A2
? . ?
Aini ?
3
(2)
where for i = 1, 2 and j = 1, 2, ? ? ? , ni , Aij? = (Aij,1 , Aij,2 , ? ? ? , Aij,p ) ? Ci . Then, the
? of Problem (1) is given by
optimal solution X
?
n?
? n2
1
?
D2 (A1 , A2 ) ,
if Ai? ? C1 ;
1
2
n1 +n2
2kM(D2 (A ,A ))k2 M
?
Xi? =
n?
1
2
? ? n1
if Ai? ? C2 .
n1 +n2 1 ? 2kM(D2 (A1 ,A2 ))k2 M D2 (A , A ) ,
The theorem essentially states that we need d1,2 to be large and w1,2 to be small for correct determination of the cluster membership of A. This is indeed intuitive. Notice that d1,2 is the distance
between the cubes and w1,2 is a constant that depends on the size of the cube as well as the ratio
between the samples in each cube. Obviously, if the cubes are too close with each other, i.e., d1,2 is
small, or if the sizes of the clusters are too big compared to their distance, it is difficult to determine
the cluster membership correctly. Moreover, when n1 n2 or n1 n2 , w1,2 is large, and the
theorem states that it is difficult to determine the cluster membership. This is also well expected,
since in this case one cluster will be overwhelmed by the other, and hence determining where the
data points are chosen from becomes problematic.
The assumption in Theorem 1 that the data matrix A is column centered can be easily relaxed, using
the following proposition which states that the result of SON is invariant to any isometry operation.
Definition 7. An isometry of Rn is a function f : Rn ? Rn that preserves the distance between
vectors, i.e.,
kf (u) ? f (w)k2 = ku ? wk2 , ? u, w ? Rn .
Proposition 1. (Isometry Invariance) Given a data matrix A of dimension n ? p where each row
is chosen from some cluster Ci , i = 1, 2, ? ? ? , c, and f (?) an isometry of Rp , we have
X
? = arg min kA ? Xk2 + ?
X
kXi? ? Xj? k2
F
n?p
X?R
i<j
? = arg min kf (A) ?
??f (X)
X?Rn?p
Xk2F
+?
X
kXi? ? Xj? k2 .
i<j
This further implies that if SON successfully determines the cluster membership of A, then it also
successfully determines the cluster membership of f (A).
3
Kernelization
SON can be easily kernelized as we show in this section. In the kernel clustering setup, instead
of clustering {Ai? } such that points within a cluster are closer in the original space, we want to
cluster {Ai? } such that points within a cluster are closer in the feature space. Mathematically, this
means we map Ai? to a Hilbert space H (the feature space) by the feature mapping function ?(?)
and perform clustering on {?(Ai? )}.
Notice that we can write Problem (1) in terms of the inner product hAi? , Aj? i , hAi? , Xj? i and
hXi? , Xj? i. Thus, for SON in the feature space, we only need to replace all these inner products
by h?(Ai? ), ?(Aj? )i , h?(Ai? ), Xj? i and hXi? , Xj? i. Thus, SON in the feature space can be formulated as
n
X
? = arg min
X
(h?(Ai? ), ?(Ai? )i ? 2 h?(Ai? ), Xi? i + hXi? , Xi? i)
n?q
X?R
i=1
+?
Xq
(3)
hXi? , Xi? i ? 2 hXi? , Xj? i + hXj? , Xj? i.
i<j
We have the following representation theorem about the optimal solution of (3).
Theorem 2. (Representation Theorem) Each row of the optimal solution of Problem (3) can be
written as a linear combination of rows of A, i.e.,
? i? =
X
n
X
aij ?(Aj? ).
j=1
4
Thus, to solve SON in the feature space reduces to finding the optimal weight {aij }. Define the
kernel function as K(x, y) = h?(x), ?(y)i. Then Problem (3) is equivalent to
!
n
n
n X
n
X
X
X
K(Ai? , Ai? ) ? 2
min
aik K(Ai? , Ak? ) +
aik ail K(Ak? , Al? )
{aij }
i=1
k=1
k=1 l=1
v
u n n
X uX
X
t
+?
K(Ak? , Al? )(aik ail ? 2aik ajl + ajk ajl ),
i<j
(4)
k=1 l=1
which is a second order cone program since the kernel is positive semi-definite. Notice that this
implies that solving SON in the feature space only requires knowing the kernel function rather than
the feature mapping ?(?).
4
Proof
We sketch the proof of Theorem 1 here. The detailed proof is given in the supplementary material.
4.1
Preliminaries
We first introduce some notations useful in the proof. We use In to denote an identity matrix of
dimension n ? n and use 1m?n to denote a matrix of dimension m ? n with all entries being 1.
Similarly, we use 0m?n to denote a matrix of dimension m ? n with all entries being 0.
We now define some special matrices. Let Hn denote a matrix of dimension (n ? 1) ? n which
is constructed by concatenating 1(n?1)?1 and ?In?1 horizontally, i.e., Hn = (1(n?1)?1 ?
In?1 ). For i = 1, 2, ? ? ? , n ? 2, we first concatenate matrices Hn?i and 0(n?1?i)?i horizontally to form a matrix (0(n?1?i)?i Hn?i ). Then, we construct Rn by concatenating
{Hn , (0(n?2)?1 Hn?1 ), ? ? ? , (01?(n?2) H2 )} vertically, i.e.,
?
?
Hn
?0(n?2)?1 Hn?1 ?
?0
?
(n?3)?2 Hn?2 ? .
Rn , ?
?
?
.
?
?
..
01?(n?2) H2
We concatenate m copies of ?In vertically to form a new matrix and denote it by Wmn?n . Let
Gm,n,i denote an m ? n dimensional matrix where the entries of the ith column all equal 1 and
all the other entries equal 0, i.e., Gm,n,i , (0m?(i?1) 1m?1 0m?(n?i) ). Then, we concatenate
{Gm,n,1 , Gm,n,2 , ? ? ? , Gm,n,n } vertically and denote it by Smn?n , i.e.,
?
?
?
?
?In
Gm,n,1
??In ?
? Gm,n,2 ?
?
?
?
Wmn?n , ?
? ... ? , Smn?n , ? ... ? .
?In
Finally, set ? ,
?
Rn1 ?1
? 0n
?
( 22 )?(n1 ?1)
S(n1 ?1)n2 ?(n1 ?1)
4.2
I(n1 ?1)
2
0(n2 )?(n1 ?1)
2
2
0(n1 ?1)n2 ?(n1 ?1)
2
Gm,n,n
0(n1 ?1)?(n2 )
2
2
I(n2 )
2
0(n1 ?1)n2 ?(n2 )
2
0(n1 ?1)?n2
2
R n2
W(n1 ?1)n2 ?n2
?
0(n1 ?1)?(n1 ?1)n2
2
0(n2 )?(n1 ?1)n2 ?
?.
2
I(n1 ?1)n2
Proof sketch of Theorem 1
The proof of Theorem 1 is based on the idea of ?lifting?. That is, we project Problem (1) into a
higher dimensional space (in particular, from n rows to n(n ? 1)/2 rows), which then allows us to
separate the regularization term into the sum of l2 norm of each row. Although this brings additional
5
linear constraints to the formulation, it facilitates the analysis. In the following, we divide the proof
into 3 steps and explain the main idea of each step.
Step 1: In this step, we derive an equivalent form of Problem (1) and give optimality conditions.
For convenience, set B(1,2) = D2 (A1 , A2 ), B1 = D1 (A1 ), B2 = D1 (A2 ) and V = {y ?
n
R( 2 ) | ?y = 0}. The following lemmas show that we can lift the original problem into an equivalent
problem that is easier to analyze.
? of problem (1) is
Lemma 1. If the data matrix A is column centered, then the optimal solution X
?
?
also column centered. Further more, set B = D(A) and Y = D(X), we have
n(n?1)
2
X 1
? i? k2 .
kBi? ? Y
2
n
i=1
? 2 =
kA ? Xk
F
Lemma 2. Given a column centered data matrix A, set B = D(A) and S = {Z ?
n
? is the optimal solution to Problem (1) iff
R( 2 )?p | ?Z?j = 0, 1 ? j ? p}. Then, X
n(n?1)
2
? = arg min
D(X)
Y?S
X
i=1
1
( kBi? ? Yi? k22 + ? kYi? k2 ).
n
(5)
? is the membership matrix of A by solving Problem (5). ComThus, we can determine whether X
pared to Problem (1), Problem (5) is more amenable to analyze as it is the sum of separable equations. That is, for i = 1, 2, ? ? ? , n(n?1)
, we can minimize each n1 kBi? ? Yi? k22 + ? kYi? k2 individ2
ually with the additional constraint ?Y = 0. Following standard convex analysis (Page 303 of [2]),
? and ?
? are an optimal primal and dual solution pair of Problem (5) if and only if
Y
and
? ?j ? V, (?
? ?j )T ? V? , j = 1, 2, ? ? ? , p,
Y
(6)
1
n
2
T
?
?
Yi? ? arg minp ( kBi? ? yk2 + ?kyk2 ? y?i? ), i = 1, 2, ? ? ? ,
.
y?R n
2
(7)
? Since A is constructed by concatenating matrices A1 and
Step 2: In this step, we construct ?.
? to be concatenated by two matrices vertically. Due to the fact that
A2 vertically, we also expect X
?
?
? and ?
? as the following
Y = D(X), for 1 ? l ? p, we write Y
? ?1 ?
? ?1 ?
??l
Y?l
? ?l = ? ?
? ?l = ? Y
? 2 ? and Y
?2 ?
?
?l
?l
(1,2)
(1,2)
?
?
??l
Y?l
n
?i , Y
? i ? R( 2i ) for i = 1, 2 and ?
? (1,2) , Y
? (1,2) ? Rn1 n2 , which are determined below.
where ?
?l
?l
?l
?l
? ?l )T ? V? is equivaBy the structure of ?, after some algebraic operations, it can be shown that (?
lent to the following equalities that hold,
(1,2)
? 1 = ?ST
?
RTn1 ?
?l
n1 n2 ?n1 ??l
(1,2)
? 2 = ?WT
?
RTn2 ?
?l
n1 n2 ?n2 ??l
,
.
(8)
? (1,2) . Set
We now construct ?
2 (1,2)
(1,2)
? (1,2)
M B
? Bm?
, 1 ? m ? n1 n2 .
?
m? =
n
(9)
? (1,2) is now fixed, we can bound the right hand sides of the two equalities in (8). In order to
Since ?
? 1 and ?
? 2 , we need the following lemma.
bound the entries of ?
?l
?l
n
P
Lemma 3. Given cn ? Rn , i.e., cn = (c1 , c2 , ? ? ? , cn )T , such that
ci = 0 and ?b ? R, |ci | ? b,
i=1
then ?x ? R
n(n?1)
2
, such that kxk? ?
2
nb
and RTn x = cn .
6
n1 = 25 & n2 = 50
n1 = n2 = 25
n1 = 25 & n2 = 75
40
Theoretical bounds w1,2
Empirical performance d?1,2
Theoretical bounds w1,2
Empirical performance d?1,2
30
Theoretical bounds w1,2
Empirical performance d?1,2
60
30
20
40
20
10
0
20
10
2
4
6
8
10
0
2
4
6
ksk2
ksk2
8
10
0
2
4
6
8
10
ksk2
Figure 1: Theoretical bounds and empirical performance. This figure illustrates the case in which
n1 , n2 are constants and ksk2 is increasing.
Then, because we can bound the right hand sides of the two equalities of (8), by using Lemma 3, we
?1,?
? 2 satisfying (8) such that the following holds
can show that there exist ?
?l
?l
(n1 ? 1)
2
? 2 k? ? 2 (n1 ) (n2 ? 1) (4?2l ).
(n2 )
(4?1l ) and k?
?l
n
n21
n
n22
? of dimension n ? p such that
To summarize this step, we have constructed ?
2
? 1 2
? ,?
? satisfies (10), 1 ? l ? p,
??
?l
?l
2 (1,2)
(1,2)
??
? (1,2)
M B
? Bm?
, 1 ? m ? n1 n2 .
m? =
n
? 1 k? ?
k?
?l
? Set
Step 3: Finally, we construct Y.
?
?1 = Y
? 2 = 0, 1 ? l ? p,
?
?Y
?l
?l
?
? (1,2)
?
?
?Ym?
=
n?
1?
2kM B(1,2) k2
!
(10)
M B(1,2) , 1 ? m ? n1 n2 .
? and Y
? constructed, it is easy to checked that conditions
Choosing w1,2 < n2 ? < d1,2 , according to ?
? and Y
? are an optimal primal and dual solution pair of Problem (5).
(6) and (7) are satisfied. So ?
5
Experiments
We now report some numerical experimental results. The empirical performance of SON has been
reported in numerous works [8, 10, 11]. It has been shown that SON outperforms traditional clustering methods like K-means in many situations. As such, we do not reproduce such results. Instead,
we conduct experiments to validate our theoretic results.
Recall that Theorem 1 states that when samples are drawn from two cubes, SON guarantees to
successfully recover the cluster membership if the distance between cubes is larger than a threshold
which is linear to the cube size ksi k and the ratio between n1 and n2 . To validate this, we randomly
draw a data matrix A where each row belongs to one of the two cubes, and find numerically the
largest distance d?1,2 between the cubes where the cluster membership is not correctly recovered.
Clearly, d?1,2 provides an empirical estimator of the minimal distance needed to successfully recover
the cluster membership. We compare the theoretic bound w1,2 with the empirical performance d?1,2
to validate our theorem. The specific procedures of the experiments are as follows.
1. Choose two cubes C1 and C2 from space Rp with size s1 = 2(?11 , ?12 , ? ? ? , ?1p ) and
s2 = 2(?21 , ?22 , ? ? ? , ?2p ), and the distance between C1 and C2 is d.
2. Choose arbitrarily n1 points from C1 and n2 points from C2 and form the data matrix Ad
of dimension n ? p. Repeat and sample m data matrices {Ad1 , Ad2 , ? ? ? , Adm }.
7
ks1 k2 = ks2 k2 = 2
ks1 k2 = ks2 k2 = 1
ks1 k2 = ks2 k2 = 3
20
50
Theoretical bounds w1,2
Empirical performance d?1,2
Theoretical bounds w1,2
Empirical performance d?1,2
30
15
20
Theoretical bounds w1,2
Empirical performance d?1,2
40
30
10
20
10
5
0
10
2
4
6
8
n2
n1
0
2
4
6
n2
n1
8
0
2
4
6
8
n2
n1
Figure 2: Theoretical bounds and empirical performance. This figure illustrates the case in which
ks1 k2 ,ks2 k2 are constants and the ratio nn12 is increasing.
3. Repeat for different d. Set
d?1,2 = max{d|?1 ? j ? m s.t. SON fails to determine the cluster membership of Adj }.
4. Repeat for different cube sizes ks1 k2 and ks2 k2 .
5. Repeat for different sample numbers n1 and n2 .
In the experiments, we focus on the samples chosen from R2 , i.e., p = 2, and use synthetic data to
obtain the empirical performance. The results are shown in Figure 1 and 2. Figure 1 presents the
situation where n1 and n2 are fixed and the cube sizes are increasing. In particular,
the two cubes are
?
both of size l ? l, i.e., both with edge length (l, l). Thus we have ksk2 = 2l. Clearly, we can see
that the empirical performance and the theoretical bounds are both linearly increasing with respect
to ksk2 , which implies that our theoretical results correctly predict how the performance of SON
depends on ksk2 . Figure 2 presents the situation in which ksk1 and ksk2 are fixed, while the ratio
n2
n1 is changing. Again, we observe that both the empirical performance and the theoretical bounds
are linearly increasing with respect to nn21 , which implies that our theoretical bounds w1,2 predict the
correct relation between the performance of SON and nn12 .
6
Conclusion
In this paper, we provided theoretical analysis for the recently presented convex optimization procedure for clustering, which we term as SON. We showed that if all samples are drawn from two
clusters, each being a cube, then SON is guaranteed to successfully recover the cluster membership
provided that the distance between the two cubes is greater than the ?weighted size? ? a term that
linearly depends on the cube size and the ratio between the numbers of the samples in each cluster.
Such linear dependence is also observed in our numerical experiment, which demonstrates (at least
qualitatively) the validity of our results.
The main thrust of this paper is to explore using techniques from high-dimensional statistics, in
particular regularization methods that extract low-dimensional structures such as sparsity or lowrankness, to tackle clustering problems. These techniques have recently been successfully applied to
graph clustering and subspace clustering [4, 7, 12, 5, 9], but not so much to distance-based clustering
tasks with the only exception of SON, to the best of our knowledge. This paper is the first attempt
to provide a rigorous analysis to derive sufficient conditions when SON succeeds. We believe this
not only helps to understand why SON works in practice as shown in previous works [8, 10, 11], but
also provides important insights to develop novel algorithms based on high-dimensional statistics
tools for clustering tasks.
Acknowledgments
The work of H. Xu was partially supported by the Ministry of Education of Singapore through
AcRF Tier Two grant R-265-000-443-112. This work is also partially supported by the grant from
Microsoft Research Asia with grant number R-263-000-B13-597.
8
References
[1] Andreas Argyriou, Theodoros Evgeniou, and Massimiliano Pontil. Convex multi-task feature
learning. Machine Learning, 73(3):243?272, 2008. 2
[2] Dimitri P Bertsekas. Convex Optimization Theory. Universities Press. 6
[3] S?ebastien Bubeck, Marina Meila, and Ulrike von Luxburg. How the initialization affects the
stability of the k-means algorithm. arXiv preprint arXiv:0907.5494, 2009. 1
[4] Yudong Chen, Shiau Hong Lim, and Huan Xu. Weighted graph clustering with non-uniform
uncertainties. In Proceedings of The 31st International Conference on Machine Learning,
2014. 8
[5] Yudong Chen, Sujay Sanghavi, and Huan Xu. Clustering sparse graphs. In NIPS, pages 2213?
2221, 2012. 8
[6] E. C. Chi and K. Lange. Splitting Methods for Convex Clustering. ArXiv e-prints, April 2013.
2
[7] Ehsan Elhamifar and Ren?e Vidal. Sparse subspace clustering. In Computer Vision and Pattern
Recognition, 2009. CVPR 2009. IEEE Conference on, pages 2790?2797. IEEE, 2009. 8
[8] Toby Hocking, Jean-Philippe Vert, Francis Bach, and Armand Joulin. Clusterpath: an algorithm for clustering using convex fusion penalties. In Lise Getoor and Tobias Scheffer, editors,
Proceedings of the 28th International Conference on Machine Learning (ICML-11), ICML
?11, pages 745?752, New York, NY, USA, June 2011. ACM. 1, 2, 7, 8
[9] Ali Jalali, Yudong Chen, Sujay Sanghavi, and Huan Xu. Clustering partially observed graphs
via convex optimization. In Proceedings of the 28th International Conference on Machine
Learning (ICML-11), pages 1001?1008, 2011. 8
[10] F. Lindsten, H. Ohlsson, and L. Ljung. Clustering using sum-of-norms regularization: With
application to particle filter output computation. In Statistical Signal Processing Workshop
(SSP), 2011 IEEE, pages 201?204, 2011. 1, 2, 7, 8
[11] Fredrik Lindsten, Henrik Ohlsson, and Lennart Ljung. Just relax and come clustering! : A
convexification of k-means clustering. Technical Report 2992, Linkping University, Automatic
Control, 2011. 1, 2, 7, 8
[12] Guangcan Liu, Zhouchen Lin, and Yong Yu. Robust subspace segmentation by low-rank representation. In Proceedings of the 27th International Conference on Machine Learning (ICML10), pages 663?670, 2010. 8
[13] Ulrike Luxburg. A tutorial on spectral clustering. Statistics and Computing, 17(4):395?416,
dec 2007. 1
[14] Yuliya Marchetti and Qing Zhou. Solution path clustering with adaptive concave penalty. arXiv
preprint arXiv:1404.6289, 2014. 2
[15] Geoffrey McLachlan and Thriyambakam Krishnan. The EM algorithm and extensions, volume
382. John Wiley & Sons, 2007. 1
[16] Geoffrey J McLachlan, Thriyambakam Krishnan, and See Ket Ng. The em algorithm. Technical report, Papers/Humboldt-Universit?at Berlin, Center for Applied Statistics and Economics
(CASE), 2004. 1
[17] Kristiaan Pelckmans, Joseph De Brabanter, JAK Suykens, and B De Moor. Convex clustering
shrinkage. In PASCAL Workshop on Statistics and Optimization of Clustering Workshop, 2005.
1
[18] Robert Tibshirani, Michael Saunders, Saharon Rosset, Ji Zhu, and Keith Knight. Sparsity and
smoothness via the fused lasso. Journal of the Royal Statistical Society: Series B (Statistical
Methodology), 67(1):91?108, 2005. 2
[19] Spaans Alexander M. J. Heiser van der Kloot, Willem A. and Willem J. Instability of hierarchical cluster analysis due to input order of the data: The permucluster solution. Psychological
Methods, 10(4):468?476, 2005. 1
[20] Rui Xu and II Wunsch, D. Survey of clustering algorithms. Neural Networks, IEEE Transactions on, 16(3):645?678, May 2005. 1
9
| 5307 |@word armand:1 briefly:1 norm:10 d2:8 shuicheng:1 seek:1 simulation:1 km:3 heiser:1 solid:1 mpexuh:1 liu:1 series:1 outperforms:1 ksk1:1 ka:3 recovered:1 adj:1 si:1 yet:1 written:1 john:1 numerical:3 concatenate:3 thrust:1 shape:1 pelckmans:2 xk:1 ith:4 provides:2 theodoros:1 c2:10 constructed:5 ik:2 n22:2 introduce:1 theoretically:1 indeed:1 expected:2 frequently:1 multi:1 chi:2 totally:1 becomes:1 provided:4 project:1 moreover:3 notation:3 increasing:5 ail:2 developed:2 lindsten:5 finding:1 guarantee:3 every:1 concave:1 tackle:1 universit:1 k2:24 demonstrates:1 uk:1 control:1 grant:3 bertsekas:1 positive:1 engineering:3 vertically:6 ak:3 path:1 pared:1 initialization:2 wk2:1 range:1 en1:3 acknowledgment:1 enforces:1 arguing:1 practice:1 block:1 definite:1 procedure:6 pontil:1 empirical:14 yan:1 significantly:1 vert:1 word:1 kbi:4 convenience:1 close:1 operator:6 nb:1 instability:3 equivalent:3 map:1 fn2:2 center:2 economics:1 independently:1 convex:15 survey:1 identifying:1 splitting:1 insight:2 estimator:1 d1:15 spanned:1 wunsch:1 stability:1 suppose:1 gm:8 aik:4 satisfying:1 ks2:6 recognition:1 convexification:1 observed:2 preprint:2 electrical:2 initializing:1 knight:1 rq:1 tobias:1 rigorously:1 solving:2 ali:1 f2:1 easily:3 separated:1 massimiliano:1 lift:1 choosing:2 saunders:1 whose:1 jean:1 larger:3 solve:5 warwick:2 say:3 supplementary:1 cvpr:1 relax:1 statistic:6 noisy:1 final:1 ip:6 obviously:1 brabanter:1 product:2 remainder:1 blackboard:1 iff:1 ama:1 intuitive:2 frobenius:2 validate:3 shiau:1 cluster:40 help:1 derive:3 develop:2 ac:1 ij:3 keith:1 fredrik:1 implies:4 come:1 quantify:1 direction:1 correct:2 filter:1 centered:6 material:1 ually:1 education:1 humboldt:1 f1:1 generalization:2 preliminary:2 proposition:2 mathematically:3 im:1 strictly:1 ad2:1 extension:1 hold:2 algorithmic:1 mapping:2 predict:2 a2:6 xk2:2 largest:1 successfully:9 tool:1 weighted:3 moor:1 minimization:1 nn12:2 mclachlan:2 clearly:2 gaussian:1 rather:1 zhou:1 shrinkage:1 derived:1 focus:1 lise:1 june:1 eleyans:1 rank:1 rigorous:3 thriyambakam:2 membership:22 kernelized:1 relation:1 reproduce:1 i1:6 provably:1 arg:6 issue:1 dual:2 adm:1 denoted:5 priori:1 pascal:1 special:1 cube:27 equal:3 once:1 construct:4 evgeniou:1 ng:1 identical:1 yu:1 unsupervised:1 icml:3 np:1 report:3 sanghavi:2 randomly:2 preserve:1 national:3 qing:1 n1:47 microsoft:1 attempt:3 ai1:1 mixture:1 primal:2 rearrange:1 amenable:1 edge:2 encourage:1 closer:2 huan:4 conduct:1 divide:1 theoretical:16 minimal:1 psychological:1 column:8 entry:6 uniform:1 too:2 reported:1 kxi:4 synthetic:1 rosset:1 st:2 international:4 michael:1 together:1 fused:2 ym:1 w1:16 again:1 von:1 satisfied:1 n21:2 rn1:2 hn:9 choose:2 ket:1 dimitri:1 de:2 lloyd:1 bold:1 b2:1 depends:5 ad:1 view:1 picked:1 analyze:2 francis:1 ulrike:2 recover:3 guangcan:1 contribution:1 minimize:1 il:1 ni:2 identify:1 xk2f:1 ohlsson:2 ren:1 cc:1 explain:1 checked:1 definition:8 e2:1 proof:7 recovers:3 tunable:1 recall:1 knowledge:5 lim:1 cj:1 hilbert:1 segmentation:1 higher:1 asia:1 methodology:1 april:1 formulation:2 just:1 sketch:2 lent:1 hand:2 lennart:1 ei:1 acrf:1 brings:1 aj:3 indicated:1 believe:2 usa:1 k22:2 validity:1 multiplier:1 regularization:8 hence:1 equality:3 alternating:2 ai2:1 i2:2 deal:1 kyk2:1 hong:1 generalized:1 theoretic:4 saharon:1 novel:2 recently:4 ji:1 volume:1 belong:1 numerically:1 ai:18 smoothness:1 automatic:1 meila:1 sujay:2 mathematics:1 similarly:2 pointed:1 particle:1 zhouchen:1 hxi:5 similarity:1 yk2:2 isometry:4 showed:1 inf:1 belongs:1 ajl:2 termed:1 arbitrarily:2 yi:3 der:1 seen:3 ministry:1 additional:2 relaxed:1 greater:1 determine:7 signal:1 semi:1 ii:1 reduces:1 technical:2 determination:1 bach:1 lin:1 e1:6 marina:1 a1:5 calculates:3 essentially:2 vision:1 arxiv:5 kernel:4 dec:1 c1:10 suykens:1 want:1 appropriately:2 unlike:1 facilitates:1 easy:1 krishnan:2 xj:12 affect:1 kristiaan:1 lasso:2 inner:2 idea:3 cn:4 knowing:1 andreas:1 lange:1 whether:1 linkage:1 penalty:4 suffer:2 algebraic:1 e3:1 york:1 useful:1 detailed:1 exist:1 problematic:1 singapore:4 revisit:1 notice:4 tutorial:1 disjoint:1 correctly:4 tibshirani:1 write:2 threshold:3 drawn:4 kyi:2 smn:2 changing:1 fuse:1 relaxation:1 graph:4 sum:5 cone:1 luxburg:2 package:1 letter:1 uncertainty:1 wmn:2 draw:1 chenlei:1 bound:15 guaranteed:1 adapted:1 constraint:2 yong:1 aspect:1 argument:1 min:6 optimality:1 hocking:4 separable:1 department:5 according:1 combination:1 son:41 character:2 em:2 joseph:1 ks1:6 s1:1 intuitively:1 invariant:1 tier:1 equation:1 fail:1 needed:1 ksk2:8 operation:3 willem:2 vidal:1 observe:1 hierarchical:4 away:1 spectral:2 rp:5 original:2 denotes:2 clustering:38 concatenated:1 society:1 print:1 dependence:1 traditional:1 jalali:1 hai:2 ssp:1 subspace:3 distance:13 separate:1 berlin:1 boldface:1 provable:1 length:2 ratio:7 difficult:2 setup:1 robert:1 marchetti:2 ebastien:1 perform:1 upper:1 observation:1 jak:1 philippe:1 situation:4 rn:9 hxj:1 pair:3 mechanical:1 namely:1 nu:3 nip:1 below:1 pattern:1 sparsity:2 summarize:1 program:1 including:1 max:2 royal:1 getoor:1 zhu:2 numerous:1 rtn:1 extract:1 xq:1 prior:1 sg:3 literature:1 review:1 kf:4 l2:2 determining:1 expect:1 ljung:2 geoffrey:2 h2:2 sufficient:2 minp:1 editor:1 row:23 repeat:4 supported:2 copy:1 jth:1 aij:7 formal:1 side:2 understand:2 wide:1 taking:1 sparse:4 van:1 yudong:3 dimension:15 made:2 qualitatively:1 adaptive:1 leng:2 bm:2 transaction:1 b1:1 clusterpath:3 xi:7 why:5 ku:1 reasonably:1 robust:1 ehsan:1 joulin:1 main:7 linearly:4 big:1 s2:1 toby:1 n2:48 xu:6 scheffer:1 ny:1 henrik:1 wiley:1 fails:1 concatenating:4 theorem:12 specific:1 list:1 r2:1 grouping:2 ad1:1 fusion:1 workshop:3 ci:8 lifting:1 overwhelmed:1 illustrates:2 elhamifar:1 kx:1 ksi:1 chen:3 easier:1 rui:1 simply:1 likely:1 explore:1 bubeck:1 horizontally:2 kxk:1 ux:1 partially:3 determines:2 satisfies:1 acm:1 identity:1 formulated:1 kmeans:1 b13:1 replace:1 admm:1 ajk:1 hard:1 change:1 typical:1 determined:1 wt:1 lemma:6 called:3 invariance:1 experimental:1 succeeds:5 exception:1 alexander:1 incorporate:1 kernelization:1 argyriou:1 |
4,758 | 5,308 | Greedy Subspace Clustering
Dohyung Park
Dept. of Electrical and Computer Engineering
The University of Texas at Austin
[email protected]
Constantine Caramanis
Dept. of Electrical and Computer Engineering
The University of Texas at Austin
[email protected]
Sujay Sanghavi
Dept. of Electrical and Computer Engineering
The University of Texas at Austin
[email protected]
Abstract
We consider the problem of subspace clustering: given points that lie on or near
the union of many low-dimensional linear subspaces, recover the subspaces. To
this end, one first identifies sets of points close to the same subspace and uses the
sets to estimate the subspaces. As the geometric structure of the clusters (linear
subspaces) forbids proper performance of general distance based approaches such
as K-means, many model-specific methods have been proposed. In this paper,
we provide new simple and efficient algorithms for this problem. Our statistical
analysis shows that the algorithms are guaranteed exact (perfect) clustering performance under certain conditions on the number of points and the affinity between
subspaces. These conditions are weaker than those considered in the standard
statistical literature. Experimental results on synthetic data generated from the
standard unions of subspaces model demonstrate our theory. We also show that
our algorithm performs competitively against state-of-the-art algorithms on realworld applications such as motion segmentation and face clustering, with much
simpler implementation and lower computational cost.
1
Introduction
Subspace clustering is a classic problem where one is given points in a high-dimensional ambient
space and would like to approximate them by a union of lower-dimensional linear subspaces. In
particular, each subspace contains a subset of the points. This problem is hard because one needs to
jointly find the subspaces, and the points corresponding to each; the data we are given are unlabeled.
The unions of subspaces model naturally arises in settings where data from multiple latent phenomena are mixed together and need to be separated. Applications of subspace clustering include motion
segmentation [23], face clustering [8], gene expression analysis [10], and system identification [22].
In these applications, data points with the same label (e.g., face images of a person under varying
illumination conditions, feature points of a moving rigid object in a video sequence) lie on a lowdimensional subspace, and the mixed dataset can be modeled by unions of subspaces. For detailed
description of the applications, we refer the readers to the reviews [10, 20] and references therein.
There is now a sizable literature on empirical methods for this particular problem and some statistical analysis as well. Many recently proposed methods, which perform remarkably well and have
theoretical guarantees on their performances, can be characterized as involving two steps: (a) finding a ?neighborhood? for each data point, and (b) finding the subspaces and/or clustering the points
given these neighborhoods. Here, neighbors of a point are other points that the algorithm estimates
to lie on the same subspace as the point (and not necessarily just closest in Euclidean distance).
1
Algorithm
What is guaranteed
Subspace
condition
SSC [4, 16]
LRR [14]
SSC-OMP [3]
TSC [6, 7]
LRSSC [24]
Correct neighborhoods
Exact clustering
Correct neighborhoods
Exact clustering
Correct neighborhoods
NSN+GSR
NSN+Spectral
Conditions for:
Fully random model
log(n/d)
O( log(nL) )
None
No intersection
No intersection
None
None
d
p
=
d
p
d
p
1
= O( log(nL)
)
1
= O( log(nL)
)
Exact clustering
None
Exact clustering
None
d
p
d
p
-
log n
= O( log(ndL)
)
log n
= O( log(ndL)
)
Semi-random
pmodel
log(n/d)
max a? = O( log(nL) )
1
max a? = O( log(nL)
)
q
log n
max a? = O( (log dL)?log(ndL)
)
-
Table 1: Subspace clustering algorithms with theoretical guarantees. LRR and SSC-OMP have only
deterministic guarantees, not statistical ones. In the two standard statistical models, there are n data
points on each of L d-dimensional subspaces in Rp . For the definition of max a?, we refer the
readers to Section 3.1.
Our contributions: In this paper we devise new algorithms for each of the two steps above; (a) we
develop a new method, Nearest Subspace Neighbor (NSN), to determine a neighborhood set for each
point, and (b) a new method, Greedy Subspace Recovery (GSR), to recover subspaces from given
neighborhoods. Each of these two methods can be used in conjunction with other methods for the
corresponding other step; however, in this paper we focus on two algorithms that use NSN followed
by GSR and Spectral clustering, respectively. Our main result is establishing statistical guarantees
for exact clustering with general subspace conditions, in the standard models considered in recent
analytical literature on subspace clustering. Our condition for exact recovery is weaker than the
conditions of other existing algorithms that only guarantee correct neighborhoods1 , which do not
always lead to correct clustering. We provide numerical results which demonstrate our theory. We
also show that for the real-world applications our algorithm performs competitively against those
of state-of-the-art algorithms, but the computational cost is much lower than them. Moreover, our
algorithms are much simpler to implement.
1.1
Related work
The problem was first formulated in the data mining community [10]. Most of the related work in
this field assumes that an underlying subspace is parallel to some canonical axes. Subspace clustering for unions of arbitrary subspaces is considered mostly in the machine learning and the computer
vision communities [20]. Most of the results from those communities are based on empirical justification. They provided algorithms derived from theoretical intuition and showed that they perform
empirically well with practical dataset. To name a few, GPCA [21], Spectral curvature clustering
(SCC) [2], and many iterative methods [1, 19, 26] show their good empirical performance for subspace clustering. However, they lack theoretical analysis that guarantees exact clustering.
As described above, several algorithms with a common structure are recently proposed with both
theoretical guarantees and remarkable empirical performance. Elhamifar and Vidal [4] proposed an
algorithm called Sparse Subspace Clustering (SSC), which uses `1 -minimization for neighborhood
construction. They proved that if the subspaces have no intersection2 , SSC always finds a correct
neighborhood matrix. Later, Soltanolkotabi and Candes [16] provided a statistical guarantee of the
algorithm for subspaces with intersection. Dyer et al. [3] proposed another algorithm called SSCOMP, which uses Orthogonal Matching Pursuit (OMP) instead of `1 -minimization in SSC. Another
algorithm called Low-Rank Representation (LRR) which uses nuclear norm minimization is proposed by Liu et al. [14]. Wang et al. [24] proposed an hybrid algorithm, Low-Rank and Sparse Subspace Clustering (LRSSC), which involves both `1 -norm and nuclear norm. Heckel and B?olcskei [6]
presented Thresholding based Subspace Clustering (TSC), which constructs neighborhoods based
on the inner products between data points. All of these algorithms use spectral clustering for the
clustering step.
The analysis in those papers focuses on neither exact recovery of the subspaces nor exact clustering
in general subspace conditions. SSC, SSC-OMP, and LRSSC only guarantee correct neighborhoods which do not always lead to exact clustering. LRR guarantees exact clustering only when
1
2
By correct neighborhood, we mean that for each point every neighbor point lies on the same subspace.
By no intersection between subspaces, we mean that they share only the null point.
2
the subspaces have no intersections. In this paper, we provide novel algorithms that guarantee exact
clustering in general subspace conditions. When we were preparing this manuscript, it is proved
that TSC guarantees exact clustering under certain conditions [7], but the conditions are stricter than
ours. (See Table 1)
1.2
Notation
There is a set of N data points in Rp , denoted by Y = {y1 , . . . , yN }. The data points are lying on
or near a union of L subspaces D = [L
i=1 Di . Each subspace Di is of dimension di which is smaller
than p. For each point yj , wj denotes the index of the nearest subspace. Let Ni denote the number
PN
of points whose nearest subspace is Di , i.e., Ni = j=1 Iwj =i . Throughout this paper, sets and
subspaces are denoted by calligraphic letters. Matrices and key parameters are denoted by letters
in upper case, and vectors and scalars are denoted by letters in lower case. We frequently denote
the set of n indices by [n] = {1, 2, . . . , n}. As usual, span{?}
Pn denotes a subspace spanned by a
set of vectors. For example, span{v1 , . . . , vn } = {v : v = i=1 ?i vi , ?1 , . . . , ?n 2 R}. ProjU y
is defined as the projection of y onto subspace U . That is, ProjU y = arg minu2U ky uk2 . I{?}
L
denotes the indicator function which is one if the statement is true and zero otherwise. Finally,
denotes the direct sum.
2
Algorithms
We propose two algorithms for subspace clustering as follows.
? NSN+GSR : Run Nearest Subspace Neighbor (NSN) to construct a neighborhood matrix
W 2 {0, 1}N ?N , and then run Greedy Subspace Recovery (GSR) for W .
? NSN+Spectral : Run Nearest Subspace Neighbor (NSN) to construct a neighborhood matrix W 2 {0, 1}N ?N , and then run spectral clustering for Z = W + W > .
2.1
Nearest Subspace Neighbor (NSN)
NSN approaches the problem of finding neighbor points most likely to be on the same subspace in
a greedy fashion. At first, given a point y without any other knowledge, the one single point that is
most likely to be a neighbor of y is the nearest point of the line span{y}. In the following steps, if
we have found a few correct neighbor points (lying on the same true subspace) and have no other
knowledge about the true subspace and the rest of the points, then the most potentially correct point
is the one closest to the subspace spanned by the correct neighbors we have. This motivates us to
propose NSN described in the following.
Algorithm 1 Nearest Subspace Neighbor (NSN)
Input: A set of N samples Y = {y1 , . . . , yN }, The number of required neighbors K, Maximum
subspace dimension kmax .
Output: A neighborhood matrix W 2 {0, 1}N ?N
yi
yi /kyi k2 , 8i 2 [N ]
. Normalize magnitudes
for i = 1, . . . , N do
. Run NSN for each data point
Ii
{i}
for k = 1, . . . , K do
. Iteratively add the closest point to the current subspace
if k ? kmax then
U
span{yj : j 2 Ii }
end if
j?
arg maxj2[N ]\Ii kProjU yj k2
Ii
Ii [ {j ? }
end for
Wij
Ij2Ii or yj 2U , 8j 2 [N ]
. Construct the neighborhood matrix
end for
NSN collects K neighbors sequentially for each point. At each step k, a k-dimensional subspace U
spanned by the point and its k 1 neighbors is constructed, and the point closest to the subspace is
3
newly collected. After k kmax , the subspace U constructed at the kmax th step is used for collecting neighbors. At last, if there are more points lying on U , they are also counted as neighbors. The
subspace U can be stored in the form of a matrix U 2 Rp?dim(U ) whose columns form an orthonormal basis of U . Then kProjU yj k2 can be computed easily because it is equal to kU > yj k2 . While
a naive implementation requires O(K 2 pN 2 ) computational cost, this can be reduced to O(KpN 2 ),
and the faster implementation is described in Section A.1. We note that this computational cost is
much lower than that of the convex optimization based methods (e.g., SSC [4] and LRR [14]) which
solve a convex program with N 2 variables and pN constraints.
NSN for subspace clustering shares the same philosophy with Orthogonal Matching Pursuit (OMP)
for sparse recovery in the sense that it incrementally picks the point (dictionary element) that is
the most likely to be correct, assuming that the algorithms have found the correct ones. In subspace
clustering, that point is the one closest to the subspace spanned by the currently selected points, while
in sparse recovery it is the one closest to the residual of linear regression by the selected points. In
the sparse recovery literature, the performance of OMP is shown to be comparable to that of Basis
Pursuit (`1 -minimization) both theoretically and empirically [18, 11]. One of the contributions of
this work is to show that this high-level intuition is indeed born out, provable, as we show that NSN
also performs well in collecting neighbors lying on the same subspace.
2.2
Greedy Subspace Recovery (GSR)
Suppose that NSN has found correct neighbors for a data point. How can we check if they are
indeed correct, that is, lying on the same true subspace? One natural way is to count the number
of points close to the subspace spanned by the neighbors. If they span one of the true subspaces,
then many other points will be lying on the span. If they do not span any true subspaces, few points
will be close to it. This fact motivates us to use a greedy algorithm to recover the subspaces. Using
the neighborhood constructed by NSN (or some other algorithm), we recover the L subspaces. If
there is a neighborhood set containing only the points on the same subspace for each subspace, the
algorithm successfully recovers the unions of the true subspaces exactly.
Algorithm 2 Greedy Subspace Recovery (GSR)
Input: N points Y = {y1 , . . . , yN }, A neighborhood matrix W 2 {0, 1}N ?N , Error bound ?
? = [L D
?
Output: Estimated subspaces D
?1 , . . . , w
?N
l=1 l . Estimated labels w
yi
yi /kyi k2 , 8i 2 [N ]
. Normalize magnitudes
Wi
Top-d{yj : Wij = 1}, 8i 2 [N ] . Estimate a subspace using the neighbors for each point
I
[N ]
while I =
6 ; do
. Iteratively pick the best subspace estimates
PN
i?
arg maxi2I j=1 I{kProjWi yj k2 1 ?}
?l
? i?
D
W
I
I \ {j : kProjWi? yj k2 1 ?}
end while
w
?i
arg maxl2[L] kProjD? l yi k2 , 8i 2 [N ]
. Label the points using the subspace estimates
Recall that the matrix W contains the labelings of the points, so that Wij = 1 if point i is assigned
to subspace j. Top-d{yj : Wij = 1} denotes the d-dimensional principal subspace of the set of
vectors {yj : Wij = 1}. This can be obtained by taking the first d left singular vectors of the
matrix whose columns are the vector in the set. If there are only d vectors in the set, Gram-Schmidt
orthogonalization will give us the subspace. As in NSN, it is efficient to store a subspace Wi in
the form of its orthogonal basis because we can easily compute the norm of a projection onto the
subspace.
Testing a candidate subspace by counting the number of near points has already been considered in
the subspace clustering literature. In [25], the authors proposed to run RANdom SAmple Consensus
(RANSAC) iteratively. RANSAC randomly selects a few points and checks if there are many other
points near the subspace spanned by the collected points. Instead of randomly choosing sample
points, GSR receives some candidate subspaces (in the form of sets of points) from NSN (or possibly
some other algorithm) and selects subspaces in a greedy way as specified in the algorithm above.
4
3
Theoretical results
We analyze our algorithms in two standard noiseless models. The main theorems present sufficient
conditions under which the algorithms cluster the points exactly with high probability. For simplicity
of analysis, we assume that every subspace is of the same dimension, and the number of data points
on each subspace is the same, i.e., d , d1 = ? ? ? = dL , n , N1 = ? ? ? = NL . We assume that d
is known to the algorithm. Nonetheless, our analysis can extend to the general case.
3.1
Statistical models
We consider two models which have been used in the subspace clustering literature:
? Fully random model: The subspaces are drawn iid uniformly at random, and the points are
also iid randomly generated.
? Semi-random model: The subspaces are arbitrarily determined, but the points are iid randomly generated.
Let Di 2 Rp?d , i 2 [L] be a matrix whose columns form an orthonormal basis of Di . An important
measure that we use in the analysis is the affinity between two subspaces, defined as
s
Pd
>
2 i,j
kDi Dj kF
k=1 cos ?k
p
a?(i, j) ,
=
2 [0, 1],
d
d
where ?ki,j is the kth principal angle between Di and Dj . Two subspaces Di and Dj are identical if
and only if a?(i, j) = 1. If a?(i, j) = 0, every vector on Di is orthogonal to any vectors on Dj . We
also define the maximum affinity as
max a? ,
max
i,j2[L],i6=j
a?(i, j) 2 [0, 1].
There are N = nL points, and there are n points exactly lying on each subspace. We assume that
each data point yi is drawn iid uniformly at random from Sp 1 \ Dwi where Sp 1 is the unit sphere
in Rp . Equivalently,
yi = D wi x i ,
xi ? Unif(Sd
1
),
8i 2 [N ].
As the points are generated randomly on their corresponding subspaces, there are no points lying on
an intersection of two subspaces, almost surely. This implies that with probability one the points are
clustered correctly provided that the true subspaces are recovered exactly.
3.2
Main theorems
The first theorem gives a statistical guarantee for the fully random model.
Theorem 1 Suppose L d-dimensional subspaces and n points on each subspace are generated in
the fully random model with n polynomial in d. There are constants C1 , C2 > 0 such that if
?
n
ne ?2
d
C2 log n
> C1 log
,
<
,
(1)
d
d
p
log(ndL 1 )
3L
then with probability at least 1
, NSN+GSR3 clusters the points exactly. Also, there are
1
0
0
other constants C1 , C2 > 0 such that if (1) with C1 and C2 replaced by C10 and C20 holds then
NSN+Spectral4 clusters the points exactly with probability at least 1 13L . e is the exponential
constant.
3
4
NSN with K = kmax = d followed by GSR with arbitrarily small ?.
NSN with K = kmax = d.
5
Our sufficient conditions for exact clustering explain when subspace clustering becomes easy or
difficult, and they are consistent with our intuition. For NSN to find correct neighbors, the points on
the same subspace should be many enough so that they look like lying on a subspace. This condition
is spelled out in the first inequality of (1). We note that the condition holds even when n/d is a
constant, i.e., n is linear in d. The second inequality implies that the dimension of the subspaces
should not be too high for subspaces to be distinguishable. If d is high, the random subspaces are
more likely to be close to each other, and hence they become more difficult to be distinguished.
However, as n increases, the points become dense on the subspaces, and hence it becomes easier to
identify different subspaces.
Let us compare our result with the conditions required for success in the fully random model in the
existing literature. In [16], it is required for SSC to have correct neighborhoods that n should be
superlinear in d when d/p fixed. In [6, 24], the conditions on d/p becomes worse as we have more
points. On the other hand, our algorithms are guaranteed exact clustering of the points, and the
sufficient condition is order-wise at least as good as the conditions for correct neighborhoods by the
existing algorithms (See Table 1). Moreover, exact clustering is guaranteed even when n is linear in
d, and d/p fixed.
For the semi-random model, we have the following general theorem.
Theorem 2 Suppose L d-dimensional subspaces are arbitrarily chosen, and n points on each
subspace are generated in the semi-random model with n polynomial in d. There are constants
C1 , C2 > 0 such that if
s
?
n
ne ?2
C2 log n
> C1 log
, max a? <
.
(2)
d
d
log(dL 1 ) ? log(ndL 1 )
then with probability at least 1
3L
1
, NSN+GSR5 clusters the points exactly.
In the semi-random model, the sufficient condition does not depend on the ambient dimension p.
When the affinities between subspaces are fixed, and the points are exactly lying on the subspaces,
the difficulty of the problem does not depend on the ambient dimension. It rather depends on
max a?, which measures how close the subspaces are. As they become closer to each other, it
becomes more difficult to distinguish the subspaces. The second inequality of (2) explains this intuition. The inequality also shows that if we have more data points, the problem becomes easier to
identify different subspaces.
Compared with other algorithms, NSN+GSR is guaranteed exact clustering, and more importantly,
the condition on max a? improves as n grows. This remark is consistent with the practical performance of the algorithm which improves as the number of data points increases, while the existing guarantees of otherpalgorithms are not. In [16], correct neighborhoods in SSC are guaranteed if max a? = O( log(n/d)/ log(nL)). In [6], exact clustering of TSC is guaranteed if
max a? = O(1/ log(nL)). However, these algorithms perform empirically better as the number of
data points increases.
4
Experimental results
In this section, we empirically compare our algorithms with the existing algorithms in terms of
clustering performance and computational time (on a single desktop). For NSN, we used the fast
implementation described in Section A.1. The compared algorithms are K-means, K-flats6 , SSC,
LRR, SCC, TSC7 , and SSC-OMP8 . The numbers of replicates in K-means, K-flats, and the K5
NSN with K = d 1 and kmax = d2 log de followed by GSR with arbitrarily small ?.
K-flats is similar to K-means. At each iteration, it computes top-d principal subspaces of the points with
the same label, and then labels every point based on its distances to those subspaces.
7
The MATLAB codes for SSC, LRR, SCC, and TSC are obtained from http://www.cis.
jhu.edu/?ehsan/code.htm,
https://sites.google.com/site/guangcanliu,
and
http://www.math.duke.edu/?glchen/scc.html,
http://www.nari.ee.ethz.ch/
commth/research/downloads/sc.html, respectively.
8
For each data point, OMP constructs a neighborhood for each point by regressing the point on the other
points up to 10 4 accuracy.
6
6
Ambient dimension (p)
SSC
SSC?OMP
LRR
TSC
NSN+Spectral
NSN+GSR
1
50
50
50
50
50
50
35
35
35
35
35
35
20
20
20
20
20
20
10
10
10
10
10
10
5
5
5
0.8
0.6
0.4
0.2
5
5
2 4 6 8 10
5
2 4 6 8 10
2 4 6 8 10
2 4 6 8 10
2 4 6 8 10
Number of points per dimension for each subspace (n/d)
2 4 6 8 10
0
Figure 1: CE of algorithms on 5 random d-dimensional subspaces and n random points on each
subspace. The figures shows CE for different numbers of n/d and ambient dimension p. d/p is
fixed to be 3/5. Brighter cells represent that less data points are clustered incorrectly.
Ambient dimension (p)
l1?minimization (SSC)
OMP (SSC?OMP)
Nuclear norm min. (LRR)
Nearest neighbor (TSC)
NSN
1
50
50
50
50
50
35
35
35
35
35
20
20
20
20
20
10
10
10
10
10
5
5
5
5
5
0.8
0.6
0.4
0.2
2
4
6
8
10
2
4
6 8 10
2 4 6 8 10
2 4 6
Number of points per dimension for each subspace (n/d)
8
10
2
4
6
8
10
0
Figure 2: NSE for the same model parameters as those in Figure 1. Brighter cells represent that
more data points have all correct neighbors.
100?dim ambient space, five 10?dim subspaces
100?dim ambient space, 10?dim subspaces, 20 points/subspace
5
5
3
4
Time (sec)
Time (sec)
4
l1?minimization (SSC)
OMP (SSC?OMP)
Nuclear norm min. (LRR)
Thresholding (TSC)
NSN
2
1
0
20
3
2
1
40
60
80
Number of data points per subspace (n)
0
100
5
10
15
20
Number of subspaces (L)
25
Figure 3: Average computational time of the neighborhood selection algorithms
means used in the spectral clustering are all fixed to 10. The algorithms are compared in terms of
Clustering error (CE) and Neighborhood selection error (NSE), defined as
(CE) = min
?2?L
N
1 X
I(wi 6= ?(w
?i )),
N i=1
(NSE) =
N
1 X
I(9j : Wij 6= 0, wi 6= wj )
N i=1
where ?L is the permutation space of [L]. CE is the proportion of incorrectly labeled data points.
Since clustering is invariant up to permutation of label indices, the error is equal to the minimum
disagreement over the permutation of label indices. NSE measures the proportion of the points
which do not have all correct neighbors.9
4.1
Synthetic data
We compare the performances on synthetic data generated from the fully random model. In Rp ,
five d-dimensional subspaces are generated uniformly at random. Then for each subspace n unitnorm points are generated iid uniformly at random on the subspace. To see the agreement with the
theoretical result, we ran the algorithms under fixed d/p and varied n and d. We set d/p = 3/5 so
that each pair of subspaces has intersection. Figures 1 and 2 show CE and NSE, respectively. Each
error value is averaged over 100 trials. Figure 1 indicates that our algorithm clusters the data points
better than the other algorithms. As predicted in the theorems, the clustering performance improves
9
For the neighborhood matrices from SSC, LRR, and SSC-OMP, the d points with the maximum weights
are regarded as neighbors for each point. For TSC, the d nearest neighbors are collected for each point.
7
L
2
3
Algorithms
Mean CE (%)
Median CE (%)
Avg. Time (sec)
Mean CE (%)
Median CE (%)
Avg. Time (sec)
K-means
19.80
17.92
26.10
20.48
-
K-flats
13.62
10.65
0.80
14.07
14.18
1.89
SSC
1.52
0.00
3.03
4.40
0.56
5.39
LRR
2.13
0.00
3.42
4.03
1.43
4.05
SCC
2.06
0.00
1.28
6.37
0.21
2.16
SSC-OMP(8)
16.92
12.77
0.50
27.96
30.98
0.82
TSC(10)
18.44
16.92
0.50
28.58
29.67
1.15
NSN+Spectral(5)
3.62
0.00
0.25
8.28
2.76
0.51
Table 2: CE and computational time of algorithms on Hopkins155 dataset. L is the number of
clusters (motions). The numbers in the parentheses represent the number of neighbors for each
point collected in the corresponding algorithms.
L
2
3
5
10
Algorithms
Mean CE (%)
Median CE (%)
Avg. Time (sec)
Mean CE (%)
Median CE (%)
Avg. Time (sec)
Mean CE (%)
Median CE (%)
Avg. Time (sec)
Mean CE (%)
Median CE (%)
Avg. Time (sec)
K-means
45.98
47.66
62.55
63.54
73.77
74.06
82.68
82.97
-
SSC
1.77
0.00
37.72
5.77
1.56
49.45
4.79
2.97
74.91
9.43
8.75
157.5
K-flats
37.62
39.06
15.78
45.81
47.92
27.91
55.51
56.25
52.90
62.72
62.89
134.0
SSC-OMP
4.45
1.17
0.45
6.35
2.86
0.76
8.93
5.00
1.41
15.32
17.11
5.26
TSC
11.84
1.56
0.33
20.02
15.62
0.60
11.90
33.91
1.17
39.48
39.45
3.17
NSN+Spectral
1.71
0.78
0.78
3.63
3.12
3.37
5.81
4.69
5.62
9.82
9.06
14.73
Table 3: CE and computational time of algorithms on Extended Yale B dataset. For each number of
clusters (faces) L, the algorithms ran over 100 random subsets drawn from the overall 38 clusters.
as the number of points increases. However, it also improves as the dimension of subspaces grows in
contrast to the theoretical analysis. We believe that this is because our analysis on GSR is not tight.
In Figure 2, we can see that more data points obtain correct neighbors as n increases or d decreases,
which conforms the theoretical analysis.
We also compare the computational time of the neighborhood selection algorithms for different
numbers of subspaces and data points. As shown in Figure 3, the greedy algorithms (OMP, Thresholding, and NSN) are significantly more scalable than the convex optimization based algorithms
(`1 -minimization and nuclear norm minimization).
4.2
Real-world data : motion segmentation and face clustering
We compare our algorithm with the existing ones in the applications of motion segmentation and
face clustering. For the motion segmentation, we used Hopkins155 dataset [17], which contains
155 video sequences of 2 or 3 motions. For the face clustering, we used Extended Yale B dataset
with cropped images from [5, 13]. The dataset contains 64 images for each of 38 individuals in
frontal view and different illumination conditions. To compare with the existing algorithms, we
used the set of 48 ? 42 resized raw images provided by the authors of [4]. The parameters of the
existing algorithms were set as provided in their source codes.10 Tables 2 and 3 show CE and average
computational time.11 We can see that NSN+Spectral performs competitively with the methods with
the lowest errors, but much faster. Compared to the other greedy neighborhood construction based
algorithms, SSC-OMP and TSC, our algorithm performs significantly better.
Acknowledgments
The authors would like to acknowledge NSF grants 1302435, 0954059, 1017525, 1056028 and
DTRA grant HDTRA1-13-1-0024 for supporting this research. This research was also partially
supported by the U.S. Department of Transportation through the Data-Supported Transportation
Operations and Planning (D-STOP) Tier 1 University Transportation Center.
10
As SSC-OMP and TSC do not have proposed number of parameters for motion segmentation, we found
the numbers minimizing the mean CE. The numbers are given in the table.
11
The LRR code provided by the author did not perform properly with the face clustering dataset that we
used. We did not run NSN+GSR since the data points are not well distributed in its corresponding subspaces.
8
References
[1] P. S. Bradley and O. L. Mangasarian. K-plane clustering. Journal of Global Optimization, 16(1):23?32,
2000.
[2] G. Chen and G. Lerman. Spectral curvature clustering. International Journal of Computer Vision, 81(3):
317?330, 2009.
[3] E. L. Dyer, A. C. Sankaranarayanan, and R. G. Baraniuk. Greedy feature selection for subspace clustering.
The Journal of Machine Learning Research (JMLR), 14(1):2487?2517, 2013.
[4] E. Elhamifar and R. Vidal. Sparse subspace clustering: Algorithm, theory, and applications. Pattern
Analysis and Machine Intelligence, IEEE Transactions on, 35(11):2765?2781, 2013.
[5] A. S. Georghiades, P. N. Belhumeur, and D. J. Kriegman. From few to many: Illumination cone models
for face recognition under variable lighting and pose. IEEE Trans. Pattern Anal. Mach. Intelligence, 23
(6):643?660, 2001.
[6] R. Heckel and H. B?olcskei. Subspace clustering via thresholding and spectral clustering. In IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), May 2013.
[7] R. Heckel and H. B?olcskei.
arXiv:1307.4891v2, 2014.
Robust subspace clustering via thresholding.
arXiv preprint
[8] J. Ho, M.-H. Yang, J. Lim, K.-C. Lee, and D. Kriegman. Clustering appearances of objects under varying
illumination conditions. In IEEE conference on Computer Vision and Pattern Recognition (CVPR), 2003.
[9] T. Inglot. Inequalities for quantiles of the chi-square distribution. Probability and Mathematical Statistics,
30(2):339?351, 2010.
[10] H.-P. Kriegel, P. Kr?oger, and A. Zimek. Clustering high-dimensional data: A survey on subspace clustering, pattern-based clustering, and correlation clustering. ACM Transactions on Knowledge Discovery
from Data (TKDD), 3(1):1, 2009.
[11] S. Kunis and H. Rauhut. Random sampling of sparse trigonometric polynomials, ii. orthogonal matching
pursuit versus basis pursuit. Foundations of Computational Mathematics, 8(6):737?763, 2008.
[12] M. Ledoux. The concentration of measure phenomenon, volume 89. AMS Bookstore, 2005.
[13] K. C. Lee, J. Ho, and D. Kriegman. Acquiring linear subspaces for face recognition under variable
lighting. IEEE Trans. Pattern Anal. Mach. Intelligence, 27(5):684?698, 2005.
[14] G. Liu, Z. Lin, S. Yan, J. Sun, Y. Yu, and Y. Ma. Robust recovery of subspace structures by low-rank
representation. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 35(1):171?184, 2013.
[15] V. D. Milman and G. Schechtman. Asymptotic Theory of Finite Dimensional Normed Spaces: Isoperimetric Inequalities in Riemannian Manifolds. Lecture Notes in Mathematics. Springer, 1986.
[16] M. Soltanolkotabi and E. J. Candes. A geometric analysis of subspace clustering with outliers. The Annals
of Statistics, 40(4):2195?2238, 2012.
[17] R. Tron and R. Vidal. A benchmark for the comparison of 3-d motion segmentation algorithms. In IEEE
conference on Computer Vision and Pattern Recognition (CVPR), 2007.
[18] J. A. Tropp and A. C. Gilbert. Signal recovery from random measurements via orthogonal matching
pursuit. Information Theory, IEEE Transactions on, 53(12):4655?4666, 2007.
[19] P. Tseng. Nearest q-flat to m points. Journal of Optimization Theory and Applications, 105(1):249?252,
2000.
[20] R. Vidal. Subspace clustering. Signal Processing Magazine, IEEE, 28(2):52?68, 2011.
[21] R. Vidal, Y. Ma, and S. Sastry. Generalized principal component analysis. In IEEE conference on Computer Vision and Pattern Recognition (CVPR), 2003.
[22] R. Vidal, S. Soatto, Y. Ma, and S. Sastry. An algebraic geometric approach to the identification of a
class of linear hybrid systems. In Decision and Control, 2003. Proceedings. 42nd IEEE Conference on,
volume 1, pages 167?172. IEEE, 2003.
[23] R. Vidal, R. Tron, and R. Hartley. Multiframe motion segmentation with missing data using power
factorization and GPCA. International Journal of Computer Vision, 79(1):85?105, 2008.
[24] Y.-X. Wang, H. Xu, and C. Leng. Provable subspace clustering: When LRR meets SSC. In Advances in
Neural Information Processing Systems (NIPS), December 2013.
[25] A. Y. Yang, S. R. Rao, and Y. Ma. Robust statistical estimation and segmentation of multiple subspaces.
In IEEE conference on Computer Vision and Pattern Recognition (CVPR), 2006.
[26] T. Zhang, A. Szlam, Y. Wang, and G. Lerman. Hybrid linear modeling via local best-fit flats. International
journal of computer vision, 100(3):217?240, 2012.
9
| 5308 |@word trial:1 polynomial:3 norm:7 proportion:2 nd:1 unif:1 d2:1 pick:2 liu:2 contains:4 born:1 zimek:1 ours:1 existing:8 bradley:1 current:1 recovered:1 com:1 numerical:1 greedy:11 selected:2 intelligence:4 desktop:1 plane:1 gpca:2 math:1 simpler:2 zhang:1 five:2 mathematical:1 constructed:3 direct:1 c2:6 become:3 theoretically:1 indeed:2 nor:1 frequently:1 planning:1 chi:1 becomes:5 provided:6 moreover:2 underlying:1 notation:1 null:1 what:1 lowest:1 finding:3 pmodel:1 guarantee:14 every:4 collecting:2 stricter:1 exactly:8 k2:8 control:1 unit:1 grant:2 szlam:1 yn:3 engineering:3 local:1 sd:1 mach:2 establishing:1 meet:1 downloads:1 therein:1 collect:1 co:1 factorization:1 lrr:14 averaged:1 practical:2 acknowledgment:1 yj:11 testing:1 union:8 implement:1 empirical:4 yan:1 jhu:1 significantly:2 matching:4 projection:2 onto:2 close:5 unlabeled:1 superlinear:1 selection:4 kmax:7 www:3 gilbert:1 deterministic:1 transportation:3 center:1 missing:1 normed:1 convex:3 survey:1 simplicity:1 recovery:11 importantly:1 nuclear:5 spanned:6 orthonormal:2 regarded:1 classic:1 justification:1 annals:1 construction:2 suppose:3 magazine:1 exact:19 duke:1 us:4 agreement:1 element:1 recognition:6 labeled:1 preprint:1 electrical:3 wang:3 wj:2 sun:1 decrease:1 ran:2 intuition:4 pd:1 kriegman:3 isoperimetric:1 depend:2 tight:1 basis:5 easily:2 htm:1 georghiades:1 icassp:1 caramanis:1 separated:1 fast:1 sc:1 neighborhood:28 choosing:1 whose:4 solve:1 cvpr:4 otherwise:1 statistic:2 jointly:1 sequence:2 ledoux:1 analytical:1 propose:2 lowdimensional:1 product:1 j2:1 trigonometric:1 description:1 normalize:2 ky:1 cluster:9 perfect:1 spelled:1 object:2 develop:1 pose:1 nearest:11 c10:1 sizable:1 predicted:1 involves:1 implies:2 hartley:1 correct:22 explains:1 clustered:2 hold:2 lying:10 considered:4 dictionary:1 estimation:1 label:7 currently:1 utexas:3 successfully:1 minimization:8 always:3 rather:1 pn:5 resized:1 varying:2 conjunction:1 ax:1 focus:2 derived:1 properly:1 rank:3 check:2 indicates:1 contrast:1 sense:1 am:1 dim:5 rigid:1 minu2u:1 wij:6 labelings:1 selects:2 arg:4 overall:1 html:2 denoted:4 art:2 field:1 construct:5 equal:2 sampling:1 preparing:1 identical:1 park:1 look:1 yu:1 hdtra1:1 sanghavi:2 few:5 randomly:5 individual:1 replaced:1 n1:1 mining:1 regressing:1 replicates:1 nl:9 ambient:8 closer:1 conforms:1 orthogonal:6 euclidean:1 theoretical:9 column:3 modeling:1 rao:1 cost:4 subset:2 too:1 stored:1 synthetic:3 oger:1 person:1 international:4 lee:2 together:1 containing:1 multiframe:1 possibly:1 ssc:29 worse:1 de:1 sec:8 nari:1 vi:1 depends:1 later:1 view:1 analyze:1 recover:4 parallel:1 hopkins155:2 candes:2 contribution:2 square:1 ni:2 accuracy:1 identify:2 identification:2 raw:1 iid:5 none:5 rauhut:1 lighting:2 ndl:5 explain:1 definition:1 against:2 nonetheless:1 naturally:1 di:9 recovers:1 riemannian:1 stop:1 newly:1 dataset:8 proved:2 recall:1 knowledge:3 lim:1 improves:4 segmentation:9 manuscript:1 scc:5 just:1 correlation:1 hand:1 receives:1 tropp:1 lack:1 incrementally:1 google:1 believe:1 grows:2 name:1 true:8 hence:2 assigned:1 soatto:1 iteratively:3 generalized:1 tsc:13 demonstrate:2 dohyung:1 tron:2 performs:5 motion:10 l1:2 orthogonalization:1 image:4 wise:1 novel:1 recently:2 mangasarian:1 common:1 heckel:3 empirically:4 volume:2 extend:1 refer:2 measurement:1 sujay:1 sastry:2 mathematics:2 i6:1 soltanolkotabi:2 dj:4 moving:1 iwj:1 add:1 curvature:2 closest:6 recent:1 showed:1 constantine:2 store:1 certain:2 inequality:6 calligraphic:1 arbitrarily:4 success:1 yi:7 devise:1 minimum:1 dtra:1 omp:18 belhumeur:1 surely:1 determine:1 signal:3 semi:5 ii:6 multiple:2 faster:2 characterized:1 sphere:1 lin:1 lrssc:3 parenthesis:1 involving:1 regression:1 ransac:2 scalable:1 vision:8 noiseless:1 arxiv:2 iteration:1 represent:3 cell:2 c1:6 cropped:1 remarkably:1 singular:1 median:6 source:1 rest:1 december:1 ee:1 near:4 counting:1 yang:2 easy:1 enough:1 fit:1 brighter:2 inner:1 texas:3 expression:1 algebraic:1 speech:1 remark:1 matlab:1 detailed:1 reduced:1 http:4 canonical:1 nsf:1 uk2:1 estimated:2 correctly:1 per:3 key:1 drawn:3 kyi:2 neither:1 ce:22 v1:1 sum:1 cone:1 realworld:1 run:7 letter:3 angle:1 baraniuk:1 throughout:1 reader:2 almost:1 vn:1 decision:1 comparable:1 bound:1 ki:1 guaranteed:7 followed:3 distinguish:1 yale:2 milman:1 constraint:1 flat:6 span:7 c20:1 min:3 department:1 smaller:1 wi:5 kpn:1 outlier:1 invariant:1 tier:1 kdi:1 count:1 dyer:2 end:5 pursuit:6 operation:1 competitively:3 vidal:7 k5:1 v2:1 spectral:13 disagreement:1 distinguished:1 schmidt:1 ho:2 rp:6 assumes:1 clustering:67 include:1 denotes:5 top:3 already:1 concentration:1 usual:1 affinity:4 kth:1 subspace:151 distance:3 maxj2:1 mail:1 manifold:1 collected:4 consensus:1 tseng:1 provable:2 assuming:1 code:4 modeled:1 index:4 minimizing:1 equivalently:1 difficult:3 mostly:1 statement:1 potentially:1 implementation:4 anal:2 proper:1 motivates:2 perform:4 upper:1 benchmark:1 acknowledge:1 finite:1 incorrectly:2 supporting:1 sscomp:1 extended:2 y1:3 varied:1 arbitrary:1 community:3 pair:1 required:3 specified:1 acoustic:1 nip:1 trans:2 kriegel:1 pattern:9 program:1 max:11 video:2 power:1 natural:1 hybrid:3 difficulty:1 indicator:1 residual:1 ne:2 identifies:1 naive:1 review:1 geometric:3 literature:7 discovery:1 kf:1 asymptotic:1 fully:6 lecture:1 permutation:3 mixed:2 versus:1 remarkable:1 foundation:1 sufficient:4 consistent:2 thresholding:5 share:2 austin:3 nse:5 supported:2 last:1 weaker:2 neighbor:28 face:10 taking:1 sparse:7 distributed:1 dimension:12 world:2 gram:1 computes:1 author:4 dwi:1 avg:6 leng:1 counted:1 transaction:4 approximate:1 gene:1 global:1 sequentially:1 gsr:14 xi:1 forbids:1 latent:1 iterative:1 table:7 ku:1 robust:3 ehsan:1 nsn:38 necessarily:1 tkdd:1 sp:2 did:2 main:3 dense:1 xu:1 site:2 quantiles:1 fashion:1 exponential:1 lie:4 candidate:2 jmlr:1 theorem:7 bookstore:1 specific:1 dl:3 sankaranarayanan:1 olcskei:3 kr:1 ci:1 magnitude:2 illumination:4 elhamifar:2 chen:1 easier:2 intersection:7 distinguishable:1 likely:4 appearance:1 partially:1 scalar:1 acquiring:1 ch:1 springer:1 acm:1 ma:4 formulated:1 hard:1 determined:1 uniformly:4 principal:4 called:3 experimental:2 lerman:2 schechtman:1 arises:1 ethz:1 philosophy:1 frontal:1 dept:3 d1:1 phenomenon:2 |
4,759 | 5,309 | Graph Clustering With Missing Data : Convex
Algorithms and Analysis
Ramya Korlakai Vinayak, Samet Oymak, Babak Hassibi
Department of Electrical Engineering
California Institute of Technology, Pasadena, CA 91125
{ramya, soymak}@caltech.edu, [email protected]
Abstract
We consider the problem of finding clusters in an unweighted graph, when the
graph is partially observed. We analyze two programs, one which works for dense
graphs and one which works for both sparse and dense graphs, but requires some a
priori knowledge of the total cluster size, that are based on the convex optimization
approach for low-rank matrix recovery using nuclear norm minimization. For
the commonly used Stochastic Block Model, we obtain explicit bounds on the
parameters of the problem (size and sparsity of clusters, the amount of observed
data) and the regularization parameter characterize the success and failure of the
programs. We corroborate our theoretical findings through extensive simulations.
We also run our algorithm on a real data set obtained from crowdsourcing an
image classification task on the Amazon Mechanical Turk, and observe significant
performance improvement over traditional methods such as k-means.
1
Introduction
Clustering [1] broadly refers to the problem of identifying data points that are similar to each other.
It has applications in various problems in machine learning, data mining [2, 3], social networks [4?
6], bioinformatics [7, 8], etc. In this paper we focus on graph clustering [9] problems where the data
is in the form of an unweighted graph. Clearly, to observe the entire graph on n nodes requires n2
measurements. In most practical scenarios this is infeasible and we can only expect to have partial
observations. That is, for some node pairs we know whether there exists an edge between them
or not, whereas for the rest of the node pairs we do not have this knowledge. This leads us to the
problem of clustering graphs with missing data.
Given the adjacency matrix of an unweighted graph, a cluster is defined as a set of nodes that are
densely connected to each other when compared to the rest of the nodes. We consider the problem of
identifying such clusters when the input is a partially observed adjacency matrix. We use the popular
Stochastic Block Model (SBM) [10] or Planted Partition Model [11] to analyze the performance of
the proposed algorithms. SBM is a random graph model where the edge probability depends on
whether the pair of nodes being considered belong to the same cluster or not. More specifically, the
edge probability is higher when both nodes belong to the same cluster. Further, we assume that each
entry of the adjacency matrix of the graph is observed independently with probability r. We will
define the model in detail in Section 2.1.
1.1
Clustering by Low-Rank Matrix Recovery and Completion
The idea of using convex optimization for clustering has been proposed in [12?21]. While each of
these works differ in certain ways, and we will comment on their relation to the current paper in
Section 1.3, the common approach they use for clustering is inspired by recent work on low-rank
matrix recovery and completion via regularized nuclear norm (trace norm) minimization [22?26].
1
In the case of unweighted graphs, an ideal clustered graph is a union of disjoint cliques. Given
the adjacency matrix of an unweighted graph with clusters (denser connectivity inside the clusters
compared to outside), we can interpret it as an ideal clustered graph with missing edges inside the
clusters and erroneous edges in between clusters. Recovering the low-rank matrix corresponding to
the disjoint cliques is equivalent to finding the clusters.
We will look at the following well known convex program which aims to recover and complete the
low-rank matrix (L) from the partially observed adjacency matrix (Aobs ):
Simple Convex Program:
minimize kLk? + ?kSk1
(1.1)
subject to
1 ? Li,j ? 0 for all i, j ? {1, 2, . . . n}
(1.2)
L,S
obs
L
obs
+S
=A
obs
(1.3)
where ? ? 0 is the regularization parameter, k.k? is the nuclear norm (sum of the singular values
of the matrix), and k.k1 is the l1 -norm (sum of absolute values of the entries of the matrix). S is
the sparse error matrix that accounts for the missing edges inside the clusters and erroneous edges
outside the clusters on the observed entries. Lobs and Sobs denote entries of L and S that correspond
to the observed part of the adjacency matrix.
Program 1.1 is very simple and intuitive. Further, it does not require any information other than
the observed part of the adjacency matrix. In [13], the authors analyze Program 1.1 without the
constraint (1.2). While dropping (1.2) makes the convex program less effective, it does allow [13] to
make use of low-rank matrix completion results for its analysis. In [16] and [21], the authors analyze
Program 1.1 when the entire adjacency matrix is observed. In [17], the authors study a slightly more
general program, where the regularization parameter is different for the extra edges and the missing
edges. However, the adjacency matrix is completely observed.
It is not difficult to see that, when the edge probability inside the cluster is p < 1/2, that (as n ? ?)
Program 1.1 will return L0 = 0 as the optimal solution (since if the cluster is not dense enough it is
more costly to complete the missing edges). As a result our analysis of Program 1.1, and the main
result of Theorem 1, assumes p > 1/2. Clearly, there are many instances of graphs we would like
to cluster where p < 1/2. If the total size of the cluster region (i.e, the total number of edges in
the cluster, denoted by |R|) is known, then the following convex program can be used, and can be
shown to work for p < 1/2 (see Theorem 2).
Improved Convex Program:
minimize kLk? + ?kSk1
(1.4)
L,S
subject to
1 ? Li,j ? Si,j ? 0 for all i, j ? {1, 2, . . . n}
Li,j = Si,j whenever
sum(L) ? |R|
Aobs
i,j
=0
(1.5)
(1.6)
(1.7)
As before, L is the low-rank matrix corresponding to the ideal cluster structure and ? ? 0 is the
regularization parameter. However, S is now the sparse error matrix that accounts only for the
missing edges inside the clusters on the observed part of adjacency matrix. [16] and [19] study
programs similar to Program 1.4 for the case of a completely observed adjacency matrix. In [19],
the constraint 1.7 is a strict equality. In [15] the authors analyze a program close to Program 1.4 but
without the l1 penalty.
If R is not known, it is possible to solve Problem 1.4 for several values of R until the desired
performance is obtained. Our empirical results reported in Section 3, suggest that the solution is not
very sensitive to the choice of R.
1.2
Our Contributions
? We analyze the Simple Convex Program 1.1 for the SBM with partial observations. We provide
explicit bounds on the regularization parameter as a function of the parameters of the SBM, that
2
characterizes the success and failure conditions of Program 1.1 (see results in Section 2.2). We
show that clusters that are either too small or too sparse constitute the bottleneck. Our analysis is
helpful in understanding the phase transition from failure to success for the simple approach.
? We also analyze the Improved Convex Program 1.4. We explicitly characterize the conditions on
the parameters of the SBM and the regularization parameter for successfully recovering clusters
using this approach (see results in Section 2.3).
? Apart from providing theoretical guarantees and corroborating them with simulation results (Section 3), we also apply Programs 1.1 and 1.4 on a real data set (Section 3.3) obtained by crowdsourcing an image labeling task on Amazon Mechanical Turk.
1.3
Related Work
In [13], the authors consider the problem of identifying clusters from partially observed unweighted graphs. For the SBM with partial observations, they analyze Program 1.1 without constraint
p (1.2), and show that under certain conditions, the minimum cluster size must be at least
O( n(log(n))4 /r) for successful recovery of the clusters. Unlike our analysis, the exact requirement on the cluster size is not known (since the constant of proportionality is not known). Also
they do not provide conditions under which the approach fails to identify the clusters. Finding the
explicit bounds on the constant of proportionality is critical to understanding the phase transition
from failure to successfully identifying clusters.
In [14?19], analyze convex programs similar to the Programs
1.1 and 1.4 for the SBM and show
?
that the minimum cluster size should be at least O( n) for successfully recovering the clusters.
However, the exact requirement on the cluster size is not known. Also, they do not provide explicit
conditions for failure, and except for [16] they do not address the case when the data is missing.
In contrast, we consider the problem of clustering with missing data. We explicitly characterize
the constants by providing bounds on the model parameters that decide if Programs 1.1 and 1.4
can successfully identify clusters. Furthermore, for Program 1.1, we also explicitly characterize the
conditions under which the program fails.
In [16], the authors extend their results to partial observations by scaling the edge probabilities by r
(observation probability), which will not work for r < 1/2 or 1/2 < p < 1/2r in Program 1.1 . [21]
analyzes Program 1.1 for the SBM and provides conditions for success and failure of the program
when the entire adjacency matrix is observed. The dependence on the number of observed entries
emerges non-trivially in our analysis. Further, [21] does not address the drawback of Program 1.1,
which is p > 1/2, whereas in our work we analyze Program 1.4 that overcomes this drawback.
2
2.1
Partially Observed Unweighted Graph
Model
Definition 2.1 (Stochastic Block Model). Let A = AT be the adjacency matrix of a graph on n
nodes with K disjoint clusters of size ni each, i = 1, 2, ? ? ? , K. Let 1 ? pi ? 0, i = 1, ? ? ? , K and
1 ? q ? 0. For l > m,
1 w.p. pi , if both nodes l, m are in the same cluster i.
Al,m =
(2.1)
1 w.p. q, if nodes l, m are not in the same cluster.
If pi > q for each i, then we expect the density of edges to be higher inside the clusters compared to
outside. We will say the random variable Y has a ?(r, ?) distribution, for 0 ? ?, r ? 1, written as
Y ? ?(r, ?), if
?
?1, w.p. r?
Y = 0, w.p. r(1 ? ?)
?
?, w.p. (1 ? r)
where ? denotes unknown.
Definition 2.2 (Partial Observation Model). Let A be the adjacency matrix of a random graph
generated according to the Stochastic Block Model of Definition 2.1. Let 0 < r ? 1. Each entry of
3
the adjacency matrix A is observed independently with probability r. Let Aobs denote the observed
adjacency matrix. Then for l > m: (Aobs )l,m ? ?(r, pi ) if both the nodes l and m belong to the
same cluster i. Otherwise, (Aobs )l,m ? ?(r, q).
2.2
Results : Simple Convex Program
Let [n] = {1, 2, ? ? ? , n}. Let R be the union of regions induced by the clusters and Rc = [n] ? [n] ?
PK
PK
R its complement. Note that |R| = i=1 n2i and |Rc | = n2 ? i=1 n2i . Let nmin := min ni ,
1?i?K
pmin := min pi and nmax := max ni .
1?i?K
1?i?K
The following definitions are important to describe our results.
? Define Di := ni r (2pi ? 1) as the effective density of cluster i and Dmin = min Di .
1?i?K
? ?succ
?
q
PK
:= max 2r ni 2( 1r ? 1) + 4 (q(1 ? q) + pi (1 ? pi )) and ?fail := i=1
1?i?K
n2i
n
p
? q1
?1
? ??1
rq(n ? ?fail ).
succ := 2r n
r ? 1 + 4q(1 ? q) + ?succ and ?fail :=
We note that the thresholds, ?succ and ?fail depend only the parameters of the model. Some simple
algebra shows that ?succ < ?fail .
Theorem 1 (Simple Program). Consider a random graph generated according to the Partial ObserK
vation Model of Definition (2.2) with K disjoint clusters of sizes {ni }K
i=1 , and probabilities {pi }i=1
1
0
0
and q, such that pmin > 2 > q > 0. Given > 0, there exists positive constants c1 , c2 such that,
1. If ? ? (1 + )?fail , then Program 1.1 fails to correctly recover the clusters with probability
1 ? c01 exp(?c02 |Rc |).
2. If 0 < ? ? (1 ? )?succ ,
? If Dmin ? (1 + ) ?1 , then Program 1.1 succeeds in correctly recovering the clusters with
probability 1 ? c01 n2 exp(?c02 nmin ).
? If Dmin ? (1 ? ) ?1 , then Program 1.1 fails to correctly recover the clusters with probability
1 ? c01 exp(?c02 nmin ).
Discussion:
1. Theorem 1 characterizes the success and failure of Program 1.1 as a function of the regularization
parameter ?. In particular, if ? > ?fail , Program 1.1 fails with high probability. If ? < ?succ ,
Program 1.1 succeeds with high probability if and only if Dmin > ?1 . However, Theorem 1 has
nothing to say about ?succ < ? < ?fail .
? q 1
n
2. Small Cluster Regime: When nmax = o(n), we have ??1
=
2r
succ
r ? 1 + 4q(1 ? q) .
For simplicity let pi = p, ? i, which yields Dmin = nmin r(2p ? 1). Then Dmin > ??1
succ implies,
s
?
2 n
1
nmin >
? 1 + 4q(1 ? q) ,
(2.2)
2p ? 1
r
giving a lower bound on the minimum cluster size that is sufficient for success.
2.3
Results: Improved Convex Program
The following definitions are critical to describe our results.
? i := ni r (pi ? q) as the effective density of cluster i and D
? min = min D
? i.
? Define D
1?i?K
? ??succ
?
q
:= 2 max r ni (1 ? pi )( 1r ? 1 + pi ) + (1 ? q)( 1r ? 1 + q)
1?i?K
4
Observation Probability (r)
Observation Probability (r)
1
Success
Failure
0.8
0.6
0.4
0.2
0.6
0.7
0.8
0.9
1
Edge Probability inside the cluster (p)
(a)
1
Success
Failure
0.8
0.6
0.4
0.2
50
100
150
Minimum Cluster Size
200
(b)
Figure 1: Region of success (white region) and failure (black region) of Program 1.1 with ? =
1.01D?1
min . The solid red curve is the threshold for success (? < ?succ ) and the dashed green line
which is the threshold for failure (? > ?fail ) as predicted by Theorem 1.
? ?1 := 2r?n
? ?
succ
q
( 1r ? 1 + q)(1 ? q) + ??succ .
? succ depends only on the parameters of the model.
We note that the threshold, ?
Theorem 2 (Improved Program). Consider a random graph generated according to the Partial
Observation Model of Definition 2.2, with K disjoint clusters of sizes {ni }K
i=1 , and probabilities
0
0
{pi }K
i=1 and q, such that pmin > q > 0. Given > 0, there exists positive constants c1 , c2 such
1
?
?
that: If 0 < ? ? (1 ? )?succ and Dmin ? (1 + ) ? , then Program 1.4 succeeds in recovering the
clusters with probability 1 ? c01 n2 exp(?c02 nmin ).
Discussion:1
1. Theorem 2 gives a sufficient condition for the success of Program 1.4 as a function of ?. In
? ?1 < ? < ?
? succ .
particular, for any ? > 0, we succeed if D
min
q
1
? ?1 = 2r?n
2. Small Cluster Regime: When nmax = o(n), we have ?
succ
r ? 1 + q (1 ? q). For
? min = nmin r(p ? q). Then D
? min > ?
? ?1 implies,
simplicity let pi = p, ? i, which yields D
succ
s
?
2 n
1
? 1 + q (1 ? q),
(2.3)
nmin >
p?q
r
which gives a lower bound on the minimum cluster size that is sufficient for success.
3. (p, q) as a function of n: We now briefly discuss the regime in which cluster sizes are large
(i.e. O(n)) and we are interested in the parameters (p, q) as a function of n that allows proposed
approaches to be successful. Critical to Program 1.4 is the constraint (1.6): Li,j = Si,j when
obs
Aobs
). With missing data,
i,j = 0 (which is the only constraint involving the adjacency A
obs
Ai,j = 0 with probability r(1 ? p) inside the clusters and r(1 ? q) outside the clusters. Defining
p? = rp + 1 ? r and q? = rq + 1 ? r, the number of constraints in (1.6) becomes statistically
equivalent to those of a fully observed graph where p and q are replaced by p? and q?. Consequently,
for a fixed r > 0, from (2.3), we require p ? p ? q & O( ?1n ) for success. However, setting
the unobserved entries to 0, yields Ai,j = 0 with probability 1 ? rp inside the clusters and
1 ? rq outside the clusters. This is equivalent to a fully observed graph where p and q are
replaced by rp and rq. In this case, we can allow p ? O( n1 ) for success which is order-wise
better, and matches the results in McSherry [27]. Intuitively, clustering a fully observed graph
with parameters p? = rp + 1 ? r and q? = rq + 1 ? r is much more difficult than one with rp
and rq, since the links are more noisy in the former case. Hence, while it is beneficial to leave
the unobserved entries blank in Program 1.1, for Program 1.4 it is in fact beneficial to set the
unobserved entries to 0.
1
The proofs for Theorems 1 and 2 are provided in the supplementary material.
5
Probability of Success
Observation Probability (r)
1
Success
0.8
0.6
0.4
0.2
0.2
0.4
0.6
0.8
1
Edge Probability inside the cluster (p)
1
0.5
Simple
Improved
0
0.2
0.4
0.6
0.8
1
Edge Probability inside the clusters (p)
(b) Comparison range of edge probability p for Simple Program 1.1 and Improved Program 1.4.
(a) Region of success (white region) and failure
? succ .
(black region) of Program 1.4 with ? = 0.49?
The solid red curve is the threshold for success
? min > ??1 ) as predicted by Theorem 2.
(D
Figure 2: Simulation results for Improved Program.
3
Experimental Results
We implement Program 1.1 and 1.4 using the inexact augmented Lagrange method of multipliers [28]. Note that this method solves the Program 1.1 and 1.4 approximately. Further, the numerical
imprecisions will prevent the entries of the output of the algorithms from being strictly equal to 0 or
1. We use the mean of all the entries of the output as a hard threshold to round each entry. That is,
if an entry is less than the threshold, it is rounded to 0 and to 1 otherwise. We compare the output
of the algorithm after rounding to the optimal solution (L0 ), and declare success if the number of
wrong entries is less than 0.1%.
Set Up: We consider at an unweighted graph on n = 600 nodes with 3 disjoint clusters. For
simplicity the clusters are of equal size n1 = n2 = n3 , and the edge probability inside the clusters
are same p1 = p2 = p3 = p. The edge probability outside the clusters is fixed, q = 0.1. We generate
the adjacency matrix randomly according to the Stochastic Block Model 2.1 and Partial Observation
Model 2.2. All the results are an average over 20 experiments.
3.1
Simulations for Simple Convex Program
Dependence between r and p: In the first set of experiments we keep n1 = n2 = n3 = 200, and
vary p from 0.55 to 1 and r from 0.05 to 1 in steps of 0.05.
Dependence between nmin and r: In the second set of experiments we keep the edge probability
inside the clusters fixed, p = 0.85. The cluster size is varied from nmin = 20 to nmin = 200 in steps
of 20 and r is varied from 0.05 to 1 in steps of 0.05.
In both the experiments, we set the regularization parameter ? = 1.01D?1
min , ensuring that Dmin >
1/?, enabling us to focus on observing the transition around ?succ and ?fail . The outcome of the
experiments are shown in the Figures 1a and 1b. The experimental region of success is shown in
white and the region of failure is shown in black. The theoretical region of success is about the solid
red curve (? < ?succ ) and the region of failure is below dashed green curve (? > ?fail ). As we can
see the transition indeed occurs between the two thresholds ?succ and ?fail .
3.2
Simulations for Improved Convex Program
We keep the cluster size, n1 = n2 = n3 = 200 and vary p from 0.15 to 1 and r from 0.05 to 1 in
? succ , ensuring that ? < ?
? succ , enabling
steps of 0.05. We set the regularization parameter, ? = 0.49?
?
us to focus on observing the condition of success around Dmin . The outcome of this experiment is
shown in the Figure 2a. The experimental region of success is shown in white and region of failure
is shown in black. The theoretical region of success is above solid red curve.
Comparison with the Simple Convex Program: In this experiment, we are interested in observing
the range of p for which the Programs 1.1 and 1.4 work. Keeping the cluster size n1 = n2 = n3 =
6
Matrix Recovered by Simple Program Matrix Recovered by Improved Program
Ideal Clusters
0.5
1
50
50
100
100
150
150
200
200
250
250
300
300
350
350
400
400
1.5
200
(a)
300
400
150
200
250
300
350
400
450
400
450
400
450
400
450
1
1.5
50
100
150
200
250
300
350
Clusters Identified from Simple Program
0.5
1
50
100
150
200
250
300
350
Clusters Identified from Improved Program
0.5
1
450
100
100
Clusters identifyed by k?means on A
1.5
450
50
0.5
1.5
100
200
300
400
(b)
50
100
150
200
250
300
350
(c) Comparing with k-means clustering.
Figure 3: Result of using (a) Program 1.1 (Simple) and (b) Program 1.4 (Improved) on the real data
set. (c) Comparing the clustering output after running Program 1.1 and Program 1.4 with the output
of applying k-means clustering directly on A (with unknown entries set to 0).
200 and r = 1, we vary the edge probability inside the clusters from p = 0.15 to p = 1 in steps
of 0.05. For each instance of the adjacency matrix, we run both Program 1.1 and 1.4. We plot the
probability of success of both the algorithms in Figure 2b. As we can observe, Program 1.1 starts
succeeding only after p > 1/2, whereas for Program 1.4 it starts at p ? 0.35.
3.3
Labeling Images: Amazon MTurk Experiment
Creating a training dataset by labeling images is a tedious task. It would be useful to crowdsource
this task instead. Consider a specific example of a set of images of dogs of different breeds. We want
to cluster them such that the images of dogs of the same breed are in the same cluster. One could
show a set of images to each worker, and ask him/her to identify the breed of dog in each of those
images. But such a task would require the workers to be experts in identifying the dog breeds. A
relatively reasonable task is to ask the workers to compare pairs of images, and for each pair, answer
whether they
think the dogs in the images are of the same breed or not. If we have n images, then
there are n2 distinct pairs of images, and it will pretty quickly become unreasonable to compare all
possible pairs. This is an example where we could obtain a subset of the data and try to cluster the
images based on the partial observations.
Image Data Set: We used images of 3 different breeds of dogs : Norfolk Terrier (172 images), Toy
Poodle (151 images) and Bouvier des Flandres (150 images) from the Standford Dogs Dataset [29].
We uploaded all the 473 images of dogs on an image hosting server (we used imgur.com).
MTurk Task: We used Amazon Mechanical Turk [30] as the platformfor crowdsourcing. For
each worker, we showed 30 pairs of images chosen randomly from the n2 possible pairs. The task
assigned to the worker was to compare each pair of images, and answer whether they think the dogs
belong to the same breed or not. If the worker?s response is a ?yes?, then there we fill the entry of
the adjacency matrix corresponding to the pair as 1, and 0 if the answer is a ?no?.
Collected Data: We recorded around 608 responses. We were able to fill 16, 750 out of 111, 628
entries in A. That is, we observed 15% of the total number of entries. Compared with true answers
(which we know a priori), the answers given by the workers had around 23.53% errors (3941 out of
16750). The empirical parameters for the partially observed graph thus obtained is shown Table 1.
?
We ran Program 1.1 and Program 1.4 with regularization parameter,
? = 1/ n. Further, for Pro
gram 1.4, we set the size of the cluster region, R to 0.125 times n2 . Figure 3a shows the recovered
matrices. Entries with value 1 are depicted by white and 0 is depicted by black. In Figure 3c we
compare the clusters output by running the k-means algorithm directly on the adjacency matrix
A (with unknown entries set to 0) to that obtained by running k-means algorithm on the matrices
recovered after running Program 1.1 (Simple Program) and Program 1.4 (Improved Program) respectively. The overall error with k-means was 40.8% whereas the error significantly reduced to
15.86% and 7.19% respectively when we used the matrices recoverd from Programs 1.1 and 1.4
respectively (see Table 2). Further, note that for running the k-means algorithm we need to know
the exact number of clusters. A common heuristic is to identify the top K eigenvalues that are much
7
Table 1: Empirical Parameters from the real data.
Table 2: Number of miss-classified images
Params
n
K
n1
n2
n3
Value
473
3
172
151
150
Params
r
q
p1
p2
p3
Value
0.1500
0.1929
0.7587
0.6444
0.7687
Clusters?
K-means
Simple
Improved
1
39
9
1
2
150
57
29
3
4
8
4
Total
193
74
34
larger than the rest. In Figure 4 we plot the sorted eigenvalues for the adjacency matrix A and the
recovered matrices. We can see that the top 3 eigen values are very easily distinguished from the
rest for the matrix recovered after running Program 1.4.
A sample of the data is shown in Figure 5. We observe that factors such as color, grooming, posture,
face visibility etc. can result in confusion while comparing image pairs. Also, note that the ability
of the workers to distinguish the dog breeds is neither guaranteed nor uniform. Thus, the edge
probability inside and outside clusters are not uniform. Nonetheless, Programs 1.1 and Program 1.4,
especially Program 1.4, are quite successful in clustering the data with only 15% observations.
30
300
300
A
Simple
20
200
10
100
100
0
0
0
?10
0
200
400
600
?100
0
200
200
400
600
?100
0
Improved
200
400
600
Figure 4: Plot of sorted eigen values for (1) Adjacency matrix with unknown entries filled by 0, (2)
Recovered adjacency matrix from Program 1.1, (3) Recovered adjacency matrix from Program 1.4
Norfolk Terrier
Toy Poodle
Bouvier des Flandres
Figure 5: Sample images of three breeds of dogs that were used in the MTurk experiment.
The authors thank the anonymous reviewers for their insightful comments. This work was supported
in part by the National Science Foundation under grants CCF-0729203, CNS-0932428 and CIF1018927, by the Office of Naval Research under the MURI grant N00014-08-1-0747, and by a grant
from Qualcomm Inc. The first author is also supported by the Schlumberger Foundation Faculty for
the Future Program Grant.
References
[1] A. K. Jain, M. N. Murty, and P. J. Flynn. Data clustering: A review. ACM Comput. Surv., 31(3):264?323,
September 1999.
[2] M. Ester, H.-P. Kriegel, and X. Xu. A database interface for clustering in large spatial databases. In
Proceedings of the 1st international conference on Knowledge Discovery and Data mining (KDD?95),
pages 94?99. AAAI Press, August 1995.
[3] Xiaowei Xu, Jochen J?ager, and Hans-Peter Kriegel. A fast parallel clustering algorithm for large spatial
databases. Data Min. Knowl. Discov., 3(3):263?290, September 1999.
[4] Nina Mishra, Robert Schreiber, Isabelle Stanton, and Robert Tarjan. Clustering Social Networks. In Anthony Bonato and Fan R. K. Chung, editors, Algorithms and Models for the Web-Graph, volume 4863 of
Lecture Notes in Computer Science, chapter 5, pages 56?67. Springer Berlin Heidelberg, Berlin, Heidelberg, 2007.
8
[5] Pedro Domingos and Matt Richardson. Mining the network value of customers. In Proceedings of the
seventh ACM SIGKDD international conference on Knowledge discovery and data mining, KDD ?01,
pages 57?66, New York, NY, USA, 2001. ACM.
[6] Santo Fortunato. Community detection in graphs. Physics Reports, 486(3-5):75 ? 174, 2010.
[7] Ying Xu, Victor Olman, and Dong Xu. Clustering gene expression data using a graph-theoretic approach:
an application of minimum spanning trees. Bioinformatics, 18(4):536?545, 2002.
[8] Qiaofeng Yang and Stefano Lonardi. A parallel algorithm for clustering protein-protein interaction networks. In CSB Workshops, pages 174?177. IEEE Computer Society, 2005.
[9] Satu Elisa Schaeffer. Graph clustering. Computer Science Review, 1(1):27 ? 64, 2007.
[10] Paul W. Holland, Kathryn Blackmond Laskey, and Samuel Leinhardt. Stochastic blockmodels: First
steps. Social Networks, 5(2):109 ? 137, 1983.
[11] Anne Condon and Richard M. Karp. Algorithms for graph partitioning on the planted partition model.
Random Struct. Algorithms, 18(2):116?140, 2001.
[12] Huan Xu, Constantine Caramanis, and Sujay Sanghavi. Robust pca via outlier pursuit. In John D. Lafferty,
Christopher K. I. Williams, John Shawe-Taylor, Richard S. Zemel, and Aron Culotta, editors, NIPS, pages
2496?2504. Curran Associates, Inc., 2010.
[13] Ali Jalali, Yudong Chen, Sujay Sanghavi, and Huan Xu. Clustering partially observed graphs via convex
optimization. In Lise Getoor and Tobias Scheffer, editors, Proceedings of the 28th International Conference on Machine Learning (ICML-11), ICML ?11, pages 1001?1008, New York, NY, USA, June 2011.
ACM.
[14] Brendan P. W. Ames and Stephen A. Vavasis. Convex optimization for the planted k-disjoint-clique
problem. Math. Program., 143(1-2):299?337, 2014.
[15] Brendan P. W. Ames and Stephen A. Vavasis. Nuclear norm minimization for the planted clique and
biclique problems. Math. Program., 129(1):69?89, September 2011.
[16] S. Oymak and B. Hassibi. Finding Dense Clusters via ?Low Rank + Sparse? Decomposition.
arXiv:1104.5186, April 2011.
[17] Yudong Chen, Sujay Sanghavi, and Huan Xu. Clustering sparse graphs. In Peter L. Bartlett, Fernando
C. N. Pereira, Christopher J. C. Burges, Lon Bottou, and Kilian Q. Weinberger, editors, NIPS, pages
2213?2221, 2012.
[18] Yudong Chen, Ali Jalali, Sujay Sanghavi, and Constantine Caramanis. Low-rank matrix recovery from
errors and erasures. IEEE Transactions on Information Theory, 59(7):4324?4337, 2013.
[19] Brendan P. W. Ames. Robust convex relaxation for the planted clique and densest k-subgraph problems.
2013.
[20] Nir Ailon, Yudong Chen, and Huan Xu. Breaking the small cluster barrier of graph clustering. CoRR,
abs/1302.4549, 2013.
[21] Ramya Korlakai Vinayak, Samet Oymak, and Babak Hassibi. Sharp performance bounds for graph clustering via convex optimizations. In Proceedings of the 39th International Conference on Acoustics, Speech
and Signal Processing, ICASSP ?14, 2014.
[22] Emmanuel J. Candes and Justin Romberg. Quantitative robust uncertainty principles and optimally sparse
decompositions. Found. Comput. Math., 6(2):227?254, April 2006.
[23] Emmanuel J. Candes and Benjamin Recht. Exact matrix completion via convex optimization. Found.
Comput. Math., 9(6):717?772, December 2009.
[24] Emmanuel J. Cand`es, Xiaodong Li, Yi Ma, and John Wright. Robust principal component analysis? J.
ACM, 58(3):11:1?11:37, June 2011.
[25] Venkat Chandrasekaran, Sujay Sanghavi, Pablo A. Parrilo, and Alan S. Willsky. Rank-sparsity incoherence for matrix decomposition. SIAM Journal on Optimization, 21(2):572?596, 2011.
[26] Venkat Chandrasekaran, Pablo A. Parrilo, and Alan S. Willsky. Rejoinder: Latent variable graphical
model selection via convex optimization. CoRR, abs/1211.0835, 2012.
[27] Frank McSherry. Spectral partitioning of random graphs. In FOCS, pages 529?537. IEEE Computer
Society, 2001.
[28] Zhouchen Lin, Minming Chen, and Yi Ma. The Augmented Lagrange Multiplier Method for Exact
Recovery of Corrupted Low-Rank Matrices. Mathematical Programming, 2010.
[29] Aditya Khosla, Nityananda Jayadevaprakash, Bangpeng Yao, and Li Fei-Fei. Novel dataset for finegrained image categorization. In First Workshop on Fine-Grained Visual Categorization, IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, June 2011.
[30] Michael Buhrmester, Tracy Kwang, and Samuel D. Gosling. Amazon?s Mechanical Turk: A new source
of inexpensive, yet high-quality, data? Perspectives on Psychological Science, 6(1):3?5, January 2011.
9
| 5309 |@word faculty:1 briefly:1 norm:6 tedious:1 proportionality:2 simulation:5 condon:1 decomposition:3 q1:1 minming:1 solid:4 klk:2 mishra:1 ksk1:2 current:1 blank:1 recovered:8 comparing:3 com:1 si:3 anne:1 yet:1 must:1 written:1 john:3 numerical:1 partition:2 kdd:2 visibility:1 plot:3 succeeding:1 santo:1 provides:1 math:4 node:12 ames:3 rc:3 mathematical:1 c2:2 become:1 focs:1 inside:15 indeed:1 p1:2 nor:1 cand:1 inspired:1 becomes:1 provided:1 c01:4 finding:5 unobserved:3 flynn:1 guarantee:1 quantitative:1 wrong:1 partitioning:2 grant:4 before:1 positive:2 engineering:1 declare:1 incoherence:1 approximately:1 black:5 co:1 range:2 statistically:1 practical:1 union:2 block:5 implement:1 erasure:1 empirical:3 significantly:1 murty:1 refers:1 suggest:1 nmax:3 protein:2 close:1 selection:1 romberg:1 applying:1 equivalent:3 reviewer:1 missing:10 customer:1 uploaded:1 williams:1 independently:2 convex:22 amazon:5 recovery:6 identifying:5 simplicity:3 sbm:8 nuclear:4 fill:2 colorado:1 exact:5 densest:1 programming:1 kathryn:1 curran:1 domingo:1 surv:1 associate:1 recognition:1 muri:1 database:3 observed:24 electrical:1 region:16 culotta:1 connected:1 kilian:1 ran:1 rq:6 benjamin:1 tobias:1 babak:2 depend:1 algebra:1 ali:2 completely:2 easily:1 icassp:1 succ:25 various:1 chapter:1 caramanis:2 norfolk:2 distinct:1 jain:1 effective:3 describe:2 fast:1 zemel:1 labeling:3 outside:7 outcome:2 quite:1 heuristic:1 supplementary:1 solve:1 denser:1 say:2 larger:1 otherwise:2 ability:1 qualcomm:1 richardson:1 breed:9 think:2 noisy:1 eigenvalue:2 leinhardt:1 interaction:1 subgraph:1 intuitive:1 cluster:83 requirement:2 categorization:2 leave:1 completion:4 p2:2 solves:1 recovering:5 predicted:2 implies:2 differ:1 drawback:2 stochastic:6 material:1 adjacency:25 require:3 samet:2 clustered:2 anonymous:1 strictly:1 around:4 considered:1 wright:1 exp:4 vary:3 standford:1 knowl:1 sensitive:1 him:1 schreiber:1 successfully:4 minimization:3 clearly:2 aim:1 karp:1 office:1 l0:2 focus:3 lise:1 naval:1 improvement:1 june:3 rank:11 lon:1 contrast:1 sigkdd:1 brendan:3 helpful:1 entire:3 pasadena:1 relation:1 her:1 interested:2 overall:1 classification:1 denoted:1 priori:2 platform:1 spatial:2 equal:2 look:1 icml:2 jochen:1 future:1 report:1 sanghavi:5 richard:2 randomly:2 densely:1 national:1 replaced:2 phase:2 cns:1 n1:6 schlumberger:1 ab:2 detection:1 biclique:1 mining:4 mcsherry:2 edge:24 partial:9 worker:8 huan:4 ager:1 filled:1 tree:1 taylor:1 desired:1 theoretical:4 psychological:1 korlakai:2 instance:2 corroborate:1 vinayak:2 entry:21 subset:1 uniform:2 successful:3 rounding:1 seventh:1 too:2 optimally:1 characterize:4 reported:1 vation:1 answer:5 corrupted:1 params:2 st:1 density:3 international:4 oymak:3 recht:1 siam:1 physic:1 dong:1 rounded:1 michael:1 quickly:1 yao:1 connectivity:1 aaai:1 recorded:1 ester:1 poodle:2 creating:1 expert:1 sob:1 return:1 pmin:3 li:6 toy:2 account:2 chung:1 parrilo:2 de:2 inc:2 explicitly:3 depends:2 aron:1 try:1 analyze:10 characterizes:2 red:4 observing:3 recover:3 start:2 parallel:2 candes:2 contribution:1 minimize:2 ni:9 blackmond:1 correspond:1 identify:4 yield:3 yes:1 classified:1 whenever:1 definition:7 inexact:1 failure:15 inexpensive:1 nonetheless:1 turk:4 proof:1 di:2 schaeffer:1 dataset:3 popular:1 ask:2 lonardi:1 finegrained:1 knowledge:4 emerges:1 color:1 elisa:1 higher:2 response:2 improved:14 april:2 furthermore:1 until:1 nmin:11 web:1 christopher:2 bouvier:2 quality:1 laskey:1 xiaodong:1 matt:1 usa:2 multiplier:2 true:1 ccf:1 former:1 regularization:10 equality:1 hence:1 imprecision:1 assigned:1 white:5 round:1 lob:1 samuel:2 complete:2 theoretic:1 confusion:1 l1:2 interface:1 pro:1 stefano:1 image:26 wise:1 novel:1 common:2 volume:1 belong:4 extend:1 interpret:1 significant:1 measurement:1 isabelle:1 ai:2 sujay:5 trivially:1 zhouchen:1 shawe:1 had:1 han:1 etc:2 recent:1 showed:1 perspective:1 constantine:2 apart:1 scenario:1 certain:2 server:1 n00014:1 success:25 yi:2 caltech:2 victor:1 minimum:6 analyzes:1 fernando:1 dashed:2 stephen:2 signal:1 alan:2 match:1 lin:1 discov:1 ensuring:2 involving:1 mturk:3 vision:1 arxiv:1 c1:2 tracy:1 whereas:4 want:1 fine:1 singular:1 source:1 extra:1 rest:4 unlike:1 jayadevaprakash:1 strict:1 comment:2 subject:2 induced:1 december:1 lafferty:1 yang:1 ideal:4 enough:1 identified:2 idea:1 bottleneck:1 whether:4 recoverd:1 expression:1 pca:1 bartlett:1 penalty:1 peter:2 speech:1 york:2 constitute:1 useful:1 hosting:1 amount:1 reduced:1 generate:1 vavasis:2 terrier:2 disjoint:7 correctly:3 broadly:1 dropping:1 threshold:8 prevent:1 neither:1 graph:38 relaxation:1 sum:3 run:2 uncertainty:1 c02:4 decide:1 reasonable:1 chandrasekaran:2 p3:2 ob:5 scaling:1 bound:7 guaranteed:1 distinguish:1 fan:1 constraint:6 fei:2 n3:5 min:12 spring:1 relatively:1 department:1 ailon:1 according:4 beneficial:2 slightly:1 intuitively:1 outlier:1 discus:1 fail:12 know:3 pursuit:1 unreasonable:1 apply:1 observe:4 spectral:1 distinguished:1 weinberger:1 struct:1 eigen:2 rp:5 bangpeng:1 assumes:1 clustering:24 denotes:1 running:6 top:2 graphical:1 giving:1 k1:1 especially:1 emmanuel:3 society:2 occurs:1 posture:1 planted:5 costly:1 dependence:3 traditional:1 jalali:2 september:3 link:1 thank:1 berlin:2 collected:1 spanning:1 nina:1 willsky:2 providing:2 ying:1 difficult:2 robert:2 frank:1 trace:1 fortunato:1 unknown:4 dmin:9 observation:13 enabling:2 january:1 defining:1 varied:2 csb:1 sharp:1 august:1 tarjan:1 community:1 pablo:2 complement:1 pair:12 mechanical:4 dog:11 extensive:1 california:1 acoustic:1 nip:2 address:2 able:1 kriegel:2 justin:1 below:1 pattern:1 regime:3 sparsity:2 program:85 max:3 green:2 critical:3 getoor:1 regularized:1 stanton:1 technology:1 ramya:3 nir:1 review:2 understanding:2 discovery:2 fully:3 expect:2 lecture:1 grooming:1 rejoinder:1 foundation:2 sufficient:3 principle:1 editor:4 pi:15 supported:2 keeping:1 infeasible:1 allow:2 burges:1 institute:1 face:1 barrier:1 kwang:1 absolute:1 sparse:7 curve:5 yudong:4 aobs:6 transition:4 gram:1 unweighted:8 author:8 commonly:1 social:3 transaction:1 crowdsource:1 overcomes:1 clique:5 keep:3 gene:1 corroborating:1 latent:1 khosla:1 pretty:1 table:4 robust:4 ca:1 heidelberg:2 bottou:1 anthony:1 pk:3 dense:4 main:1 blockmodels:1 paul:1 n2:12 nothing:1 xu:8 augmented:2 scheffer:1 venkat:2 ny:2 hassibi:4 fails:5 pereira:1 explicit:4 comput:3 breaking:1 grained:1 theorem:10 erroneous:2 specific:1 insightful:1 exists:3 workshop:2 corr:2 chen:5 depicted:2 visual:1 lagrange:2 aditya:1 n2i:3 partially:7 holland:1 springer:1 pedro:1 xiaowei:1 acm:5 ma:2 succeed:1 sorted:2 consequently:1 hard:1 specifically:1 except:1 miss:1 principal:1 total:5 experimental:3 e:1 succeeds:3 bioinformatics:2 crowdsourcing:3 |
4,760 | 531 | Node Splitting: A Constructive Algorithm for
Feed-Forward Neural Networks
Mike Wynne-Jones
Research Initiative in Pattern Recognition
St. Andrews Road, Great Malvern
WR14 3PS, UK
[email protected]
Abstract
A constructive algorithm is proposed for feed-forward neural networks,
which uses node-splitting in the hidden layers to build large networks from
smaller ones. The small network forms an approximate model of a set of
training data, and the split creates a larger more powerful network which is
initialised with the approximate solution already found. The insufficiency
of the smaller network in modelling the system which generated the data
leads to oscillation in those hidden nodes whose weight vectors cover regions in the input space where more detail is required in the model. These
nodes are identified and split in two using principal component analysis,
allowing the new nodes t.o cover the two main modes of each oscillating
vector. Nodes are selected for splitting using principal component analysis
on the oscillating weight vectors, or by examining the Hessian matrix of
second derivatives of the network error with respect to the weight.s. The
second derivat.ive method can also be applied to the input layer, where it
provides a useful indication of t.he relative import.ances of parameters for
the classification t.ask. Node splitting in a standard Multi Layer Percept.ron is equivalent to introducing a hinge in the decision boundary to allow
more detail to be learned. Initial results were promising, but further evaluation indicates that the long range effects of decision boundaries cause
the new nodes to slip back to the old node position, and nothing is gained.
This problem does not occur in networks of localised receptive fields such
as radial basis functions or gaussian mixtures, where the t.echnique appears
to work well.
1072
Node Splitting: A Contructive Algorithm for Feed-Forward Neural Networks
1
Introduction
To achieve good generalisation in neural networks and other techniques for inferring
a model from data, we aim to match the number of degrees of freedom of the model
to that of the system generating the data. With too small a model we learn an
incomplete solution, while too many free parameters capture individual training
samples and noise.
Since the optimum size of network is seldom known in advance, there are two alternative ways of finding it. The constructive algorithm aims to build an approximate
model, and then add new nodes to learn more detail, thereby approaching the optimum network size from below. Pruning algorithms, on the other hand, start with
a network which is known to be too big, and then cut out nodes or weights which
do not contribute to the model. A review of recent techniques [\VJ91a] has led the
author to favour the constructive approach, since pruning still requires an estimate
of the optimum size, and the initial large net.works can take a long time t.o train.
Constructive algorithms offer fast training of the initial small networks, with the
network size and training slowness reflecting the amount of information already
learned. The best approach of all would be a constructive algorithm which also
allowed the pruning of unnecessary nodes or weights from the net.work.
The constructive algorithm trains a net.work until no further detail of the training
data can be learned, and then adds new nodes to t.he network. New nodes can be
added with random weights, or with pre-determined weight.s. Random weights are
likely to disrupt the approximate solut.ion already found, and are unlikely to be
initially placed in parts of the weight space where they can learn something useful,
although encouraging results have been reported in t.his ar~a.[Ash89] This problem
is likely to be accentuated in higher dimensional spaces. Alt.ernatively, weights can
be pre-determined by measurements on the performance of the seed network, and
this is the approach adopted here. One node is turned into two, each wit.h half the
output weight. A divergence is introduced in the weights into the nodes which is
sufficient for them behave independently in future training without disrupting the
approximate solution already found.
2
Node-Splitting
A network is trained using standard techniques until no furt.her improvement on
training set performance is achieved. Since we begin with a small network, we have
an approximate model of the data, which captures the dominant properties of the
generating system but lacks detail. We now freeze the weight.s in the network, and
calculate the updates which would be made them, using simple gradient descent,
by each separate t.raining pattern. Figure 1 shows t.he frozen vector of weights into
a single hidden node, and the scatter of proposed updates around the equilibrium
posit.ion.
The picture shows the case of a hidden node where there is one clear direction
of oscillation. This might be caused by two clusters of data within a class, each
trying to use the node in its own area of the input space, or by a decision boundary
pulled clockwise by some patterns and anti clockwise by others. If the oscillation
is strong, either in its exhibition of a clear direction or in comparison with other
1073
1074
Wynne-Jones
New Node
#1 --~U(
Weight Update
Vectors
Figure 1: A hidden node weight vector and updates proposed hy individual t.raining
patterns
nodes in the same layer, then the node is split in two. The new nodes are placed
one standard deviation either side of the old position. \Vhile this divergence gives
the nodes a push in the right direction, allowing them t.o continue to diverge in later
t.raining, the overall effect on the network is small. In most cases t.here is very little
degradation in performance as a result of the split.
The direction and size of oscillation are calculated by principal component analysis of the weight updates. By a traditional method, we are required to make a
cova.riance matrix of the weight updat.es for the weight vector int.o each node:
c
= L6w6wT
(1)
p
where p is the number of patterns. The mat.rix is then decomposed to a set of eigenvalues and eigenvectors; the largest. eigenvalue is the variance of oscillation and the
corresponding eigenvector is it.s direction. Suitable techniques for performing this
decomposition include Singular Value Dewmposition and Householder Reduction.
[Vet86] A much more suit.able way of calculating the principal components of a
stream of continuous measurements such as weight updat.es is iterative est.imation.
An est.imate is stored for each required principal component. vector, and the estimat.es are updated using each sample. [Oja83, San89] By Oja's method, the scalar
product of t.he current sample vector wit.h each current est.imate of the eigenvectors
is used as a mat.ching coefficient., M. The matching coefficient is used to re-estima.te
the eigenvalues and eigenvectors, in conjunction wit.h a gain term). which decays
as the number of patterns seen increases. The eigenvectors are updated by a proportion )'M of the current sample, and t.he eigenvalues hy ).lU 2 . The trace (sum of
eigenvalues) can also be est.imated simply as the mean of the traces (sum of diagonal
elements) of t.he individual sample covariance mat.rices. The principal component
vectors are renormalised and orthogonalised after every few updat.es. This algorithm
is of order n, the number of eigenvalues required, for the re-estimation, and O(n2)
for the orthogonalisation; the matrix decomposition method can take exponential
Node Splitting: A Contructive Algorithm for Feed-Forward Neural Networks
time, and is always much slower in practice.
In a recent paper on Ateiosis Networks, Hanson introduced stochastic weights in the
multi layer perceptron, with the aim of avoiding local minima in training.[Han90]
A sample was taken from a gaussian distribution each time a weight was used;
the mean was updated by gradient descent, and the variance reflected the network
convergence. The variance was allowed to decay with time, so that the network
would approach a deterministic state, but was increased in proportion to the updates
made to the mean. \Vhile the network wa.g far from convergence these updates were
large, and the variance remained large. Node splitting wa.g implemented in this
system, in nodes where the variances on the weights were large compared with the
means. In such cases, two new nodes were created with the weights one standard
deviation either side of the old mean: one SD is added to all weights to one node,
and subtracted for all weights to the other. Preliminary results were promising, but
there appear to be two problems with this approach for node-splitting. First, the
splitting criterion is not good: a useless node with all weights close to zero could
have comparatively large variances on the weights owing to noise. This node would
be split indefinit.ely. Secondly and more interestingly, the split is made wit.hout
regard to the correlations in sign between the weight updates, shown as dots in the
scatter plot.s of figure 2. In figure 2a, Meiosis would correctly place new nodes in the
positions marked with crosses, while in figure 2b, the new nodes would he placed
in completely the wrong places. This problem does not occur in the node splitting
scheme based on principal component analysis.
(a)
(b)
??
?.~
?
??
?
????? ? ?
'-.
~
~
???
?
?
?
?
? ???? ?
?
? ?
.~
??
??
? ? ??
?
? ????
? ? ???? ?
? ??
??? ? - ?
~.
? ?? ?
?? ?
?? ?
??
............
-.~
..
..... ....
~
X
.. .
X
Figure 2: Meiosis networks split correctly if the weight. updates are correlated in
sign (a), but fail when they are not (b).
3
Selecting nodes for splitting
Node splitting is carried out in t.he direct.ion of maximum variance of the scatter plot
of weight updates proposed by individual training samples. The hidden layer nodes
most likely t.o benefit from splitting are those for which the non-spherical nature
1075
1076
Wynne-Jones
of the scatter plot is most pronounced. In later implementations this criterion was
measured by comparing the largest eigenvalue with the sum of the eigenvalues,
both these quantities being calculated by the iterative method. This is less simple
in cases where there are a number of dominant directions of variance; the scatter
plot might, for example be a four dimensional disk in a ten dimensional space, and
hence present the possibility of splitt.ing one node into eight. It is hoped that these
more complicat.ed splits will be the suhject of further research.
An alternative approach in determining the need of nodes to be split, in comparison
with other nodes in the same layer, is to use the second derivat.ives of t.he network
error with respect to a parameter of the nodes which is normalised across all nodes
in a given layer of the network. Such a parameter wa.c;; proposed by Mozer and
Smolensky in [Sm089]: a multiplicative gat.ing function is applied to the outputs of
the nodes, with its gating parameter set to one. Small incrempnt.s in this parameter
can be used to characterise the error surface around the unity value, with the result
that derivatives are normalised a.cross all nodes in a given layer of the network.
Mozer and Smolensky rpplaced the sum squared error crit.erion with a modulus error criterion to preserve non-zero gradients close to the local minimum reached in
training; we prefer to characterise the t.rue error surface by mpans of second derivat.ives, which can be calculated by repeated use of the chain rule (hackpropagat.ion).
Backpropagat.ion of second derivat.ivps has previously been rpport.ed in [So190] and
[Hea90].
Since a high curvat.ure error minimum in t.he space of t.he gat.ing parampt.er for a
particular nocie indicat.es st.eep gradipnt.s surrounding thp minimum, it is t.hese nodes
which exhibit. t.he great.est instability in their weight-space position. In t.he weight
space, if the curvat.ure is high only in cert.ain directions, we have the situat.ion in
figure 1, where the node is oscillating, and is in need of splitt.ing. If the curvature is
high in all directions in comparison with other nodes, the network is highly sensitive
to changes in t.he node or it.s weights, and again it will benefit from splitting.
At t.he ot.her end of the scale of curvat.ure sensitivity, a node or weight wit.h very low
curvat.ure is one to which t.he network error is quit.e insensit.ive, and the parameter
is a suitable candidate for pruning. This scheme has previously been used for weight
pruning by Le Cun, Denker et a1. [SoW 0] , and offers the pot.ential for an int.egrated
syst.em of splitting and pruning - a truly adapt.ive net.work archit.ecture.
3.1
Applying the sensitivity measnre to inpnt nodes
In a.ddit.ion to using t.he ga.ting parameter sensit.ivit.y to select nodes for pruning,
Mozer and Smolensky mention the possibility of using it on the input nodes to
indicate those inputs to which the c1a.<;sification is most sensitive. This has been
implemented in our syst.em wit.h the second derivat.ive sensitivity measure, and applied to a large financial classification prohlem supplied by THORN El\JI Research.
The analysis was carried out. on the 78-dimensional dat.a, and the input sensitivities
varied over several orders of magnit.ude. The inputs were grouped into four sets according to sensitivit.y, and MLPs of 10 hidden nodes were trained on each subset of
the dat.a. \Vhile the low sensitivit.y groups failed to learn anyt.hing at all, t.he higher
sensit.ivit.y groups quickly attained a reasonable classification rat.e. Ident.ification of
useless inputs leads t.o greatly increased training speed in fut.ure analysis, and can
Node Splitting: A Contructive Algorithm for Feed-Forward Neural Networks
yield valuable economies in future data collection. This work is reported in more
detail in [WJ91b].
4
Evaluation in Multi Layer Percept ron networks
Despite the promising results from initial evaluations, further testing showed that
the splitter technique was often unable to improve on the performance of the network used as a seed for the first split. These test were carried out on a number of
different classification problems, where large numbers of hidden nodes were already
known to be required, and with a number of different splitting criteria. Prolonged
experimentation and consideration of this failure lead to the hypothesis that a split
might be made to correct some miscla.<;sified patterns in one region of the input
space but, owing to the long range effects of MLP decision boundaries, the changed
positions of the planes might cause a much greater number of misclassifications
elsewhere. These would tend to cause the newly creat.ed nodes to slip back to the
position of the node from which they were created, with no overall benefit. This
possibility was tested hy re-implementing the splitter technique in a gaussian mixture modeling system, which uses a network of localised receptive fields, and hence
does not have the long range effects which occurred in the multi layer perceptron.
5
Implementation of the splitter in a Gaussian Mixture
Model, and the results
The Gaussian Mixt.ures Model [Cox91] is a clustering algorithm, which attempts
to model the distribution of a points in a data set. It consists of a numher of
mult.ivariate gaussian dist.rihut.ions in different posit.ions in t.he input space, and
wit.h different variances in different direct.ions. The responses of t.hese recept.ive
fields (humps) are weighted and summed together; the weights are calculated to
sat.isfy the PDF const.raint. that t.he responses should sum to one over the data set.
For the experiment.s on node splitting, the variance was the same in all direct.ions for
a particular bump, leading to a model which is a sum of weight.ed spherical gaussian
distribut.ions of different sizes and in different. positions. The model is t.rained by
gradient ascent in the likelihood of the model fitting the data, which leads t.o a
set of learning rules for re-estimat.ing the weights, then t.he cent.re positions of the
recept.ive fields, then their variances.
For t.he splitter, a small model is trained until nothing more can be learned, and
the paramet.ers are frozen. The training set is run t.hrough once more, and the
updat.es are calculated which each pattern attempts to make to the centre position
of each receptive field. The first principal component and trace of these updates are
calculated by the iterative met.hod, and any nodes for which t.he principal component
variance is a large proportion of the trace is split in two.
The algorithm is quick to converge, and is slowed down only a. lit.tle by the oV('fhead
of computing the principal component and trace. Figure 3 shows the application of
t.he gaussian mixture splitter to modelling a circle and an enclosing annulus; in the
circle (a) there is no dominant. principa.l component direction in the data ('Overed by
the receptive field of each node (shown at. one st.anda.rd deviation by a circle), while
1077
1078
Wynne-Jones
in (b) three nodes are clearly insufficient to model the annulus, and one has just
undergone a split. (c) shows the same data set. and model a little later in t.raining
after a number of splits have taken place. The technique has been evaluated on a
number of other simple problems, with no negat.ive results to date.
Figure 3: Gaussian mixt.ure model with node-splitting applied to a circle and surrounding annulus
6
Conclusions
The split.ter t.echnique based on taking the principal component. of the influences
on hidden nodes in a network, ha.g been shown to be useful in the multi layer
perceptron in only a very limited number of cases. The split in this kind of net.work
corresponds to a hinge in the decision boundary, which corrects the errors for which
it was calculated, but usually caused for more errors in other parts of the input
space. This problem does not occur in networks of localised receptive fields such
as radial ba."is funct.ions of gaussian mixture distributions, where it appears to
work very well. Further studies will include splitting nodes into more than two, in
cases where there is more than one dominant principal component, and applying
node-split.t.ing to different. modelling algorithms, and to gaussian mixtures in hidden
markov models for speech recognition.
The analysis of the sensit.ivity of the net.work error to individual nodes gives an
ordered list which can be used for both splitting and pruning in the same network,
although splitting does not generally work in the MLP. This measure has been
demonstrated in t.he input layer, to identify which network inputs are more or less
useful in the classification t.ask.
Acknowledgements
The author is greatly indebted to John Bridle and Steve Luttrell of RSRE, Neil
Thacker of Sheffield University, and colleagues in the Research Initiative in Pattern
Node Splitting: A Contructive Algorithm for Feed-Forward Neural Networks
Recognition and its member companies for helpful comments and advice; also to
David Bounds of Aston University and RIPR for advice and encouragement.
References
[Ash89] Timur Ash. Dynamic node creation in backpropagation networks. Technical Report 8901, Institute for Cognitive Science, UCSD, La Jolla, California 92093, February 1989.
[Cox91] John S Bridle & Stephen J Cox. Recnorm: Simultaneous normalisation
and classification applied to speech recognition. In Richard P Lippmann
& John E Moody & David S Touretzky, editor, Advances in Neural Information Processing Systems 3, pages 234-240, San Mateo, CA, September
1991. Morgan Kaufmann Publishers.
[Han90] Stephen Jose Hanson. Meiosis networks. In David S Touretzky, editor,
Adllances in Nellral Information Processing Systems 2, pages 533-541, San
Mateo, CA, April 1990. Morgan Kaufmann Puhlishers.
[IIea90]
Anthony JR Heading. An analysis of noise tolerance in multi-layer perceptrons. Research Note SP4 122, R.oyal Signals and Radar Estahlishment,
St Andrews Road, Malvern, Worcestershire, WR14 3PS, UK, July 1990.
[Oja83]
E Oja. Sllhspace Methods of Pattern Recognition. Research Studit's Press
Ltd, Letchworth, UK, 1983.
[San89]
TD Sanger. Optimalunsupervispd learningin a single-Iayn linear feed forward neural network. Neural Networks, 2:459-473, 1989.
[SmoR9] MC Mozer & P Smolensky. Skeletonization: A tedlllique for trimming the
fat from a neural network. In DS Touretzky, editor, Advances in Neural
Information Processing Systems 1, pages 107-115, San Mateo, CA, April
1989. Morgan Kaufmann Publishers.
[SoWO]
Yann Le Cun & John S Denker & Sara A Solla. Optimal brain damage.
In David S Touretzky, editor, Adt'ances in Neural Information Processing
Systems 2, pages 598-605, San Mateo, CA, April 1990. Morgan Kaufmann
Publishers.
[Vet86]
WII Prpss & BP Flannery & SA Teukolsky & \VT Vette-rling. Numerical
Recipes in C: The A rt of Scientific Compttting. Camhrigde University
Press, 1986.
[WJ91a] Mike Wynne-Jones. Constructive algorithms and pruning: Improving the
multi layer perceptron. In R Vichnevetsky & JJII 1\filler, editor, Proceedings of the 18th IMACS World Congress on Computation and Applied
Mathemetics, pages 747-750, Duhlin, July 1991. IMACS '91, IMACS.
[\VJ91b] Mike Wynne-Jones. Self-configuring neural nptworks, a new constructive
algorithm, and assessing the importance of individual inputs. Technical Report X2345!1, Thorn BMI Central Research Lahoratories, Dawley
Road, Hayes, Middlesex, UB3 lHH, UK, March 1991.
1079
| 531 |@word cox:1 proportion:3 ivit:2 disk:1 decomposition:2 covariance:1 thereby:1 mention:1 reduction:1 initial:4 selecting:1 interestingly:1 current:3 comparing:1 scatter:5 import:1 creat:1 john:4 numerical:1 wynne:6 plot:4 update:11 recept:2 half:1 selected:1 plane:1 provides:1 node:73 ron:2 contribute:1 direct:3 initiative:2 consists:1 fitting:1 prohlem:1 dist:1 multi:7 brain:1 decomposed:1 spherical:2 company:1 prolonged:1 encouraging:1 little:2 td:1 begin:1 kind:1 eigenvector:1 finding:1 every:1 estimat:2 fat:1 wrong:1 uk:5 configuring:1 appear:1 local:2 sd:1 insufficiency:1 congress:1 despite:1 ure:6 might:4 mateo:4 sara:1 limited:1 recnorm:1 range:3 exhibition:1 renormalised:1 testing:1 practice:1 backpropagation:1 area:1 mult:1 matching:1 pre:2 road:3 radial:2 close:2 ga:1 applying:2 instability:1 influence:1 isfy:1 equivalent:1 deterministic:1 quick:1 demonstrated:1 independently:1 wit:7 splitting:24 rule:2 his:1 financial:1 updated:3 us:2 slip:2 hypothesis:1 element:1 recognition:5 cut:1 mike:3 capture:2 calculate:1 region:2 solla:1 valuable:1 mozer:4 nellral:1 dynamic:1 hese:2 radar:1 trained:3 ov:1 crit:1 funct:1 creation:1 creates:1 basis:1 estima:1 completely:1 surrounding:2 train:2 fast:1 adt:1 whose:1 larger:1 ive:7 neil:1 indication:1 frozen:2 hermes:1 net:6 vhile:3 eigenvalue:8 orthogonalisation:1 hout:1 product:1 luttrell:1 turned:1 date:1 achieve:1 pronounced:1 recipe:1 convergence:2 cluster:1 p:2 optimum:3 hrough:1 oscillating:3 generating:2 assessing:1 andrew:2 measured:1 sa:1 strong:1 implemented:2 pot:1 indicate:1 met:1 direction:9 posit:2 sensit:3 correct:1 owing:2 stochastic:1 accentuated:1 implementing:1 preliminary:1 secondly:1 quit:1 paramet:1 around:2 great:2 seed:2 equilibrium:1 bump:1 estimation:1 ain:1 sensitive:2 largest:2 grouped:1 weighted:1 ivp:1 clearly:1 gaussian:11 always:1 aim:3 conjunction:1 improvement:1 modelling:3 indicates:1 likelihood:1 greatly:2 helpful:1 economy:1 el:1 unlikely:1 initially:1 hidden:10 her:2 overall:2 classification:6 distribut:1 summed:1 field:7 once:1 lit:1 jones:6 future:2 others:1 report:2 richard:1 few:1 oja:2 preserve:1 divergence:2 individual:6 suit:1 attempt:2 freedom:1 mlp:2 trimming:1 ential:1 imated:1 possibility:3 highly:1 hing:1 evaluation:3 hump:1 normalisation:1 mixture:6 truly:1 chain:1 furt:1 incomplete:1 old:3 re:5 circle:4 increased:2 modeling:1 eep:1 cover:2 ar:1 introducing:1 deviation:3 subset:1 wr14:2 examining:1 thacker:1 too:3 reported:2 stored:1 st:4 sensitivity:4 corrects:1 diverge:1 together:1 quickly:1 moody:1 squared:1 again:1 central:1 cognitive:1 derivative:2 leading:1 syst:2 ivity:1 int:2 coefficient:2 caused:2 ely:1 stream:1 later:3 multiplicative:1 reached:1 start:1 thp:1 mlps:1 cert:1 kaufmann:4 variance:12 percept:2 yield:1 identify:1 ecture:1 lu:1 annulus:3 mc:1 indebted:1 simultaneous:1 touretzky:4 ed:4 failure:1 colleague:1 initialised:1 bridle:2 gain:1 newly:1 ask:2 back:2 reflecting:1 appears:2 feed:7 steve:1 higher:2 attained:1 reflected:1 response:2 april:3 evaluated:1 just:1 correlation:1 until:3 hand:1 d:1 ident:1 lack:1 mode:1 scientific:1 modulus:1 effect:4 hence:2 self:1 echnique:2 rat:1 criterion:4 trying:1 pdf:1 disrupting:1 consideration:1 mixt:2 ji:1 he:25 occurred:1 measurement:2 freeze:1 encouragement:1 rd:1 seldom:1 sp4:1 centre:1 dot:1 surface:2 indicat:1 add:2 something:1 dominant:4 curvature:1 own:1 recent:2 showed:1 jolla:1 slowness:1 continue:1 vt:1 seen:1 minimum:4 greater:1 morgan:4 rained:1 converge:1 clockwise:2 stephen:2 signal:1 july:2 ing:6 technical:2 match:1 adapt:1 offer:2 long:4 cross:2 a1:1 sheffield:1 imacs:3 achieved:1 ion:13 ures:1 singular:1 publisher:3 ot:1 ascent:1 comment:1 tend:1 negat:1 member:1 mod:1 ter:1 split:17 misclassifications:1 identified:1 approaching:1 favour:1 ltd:1 speech:2 hessian:1 cause:3 rsre:1 backpropagat:1 useful:4 generally:1 clear:2 eigenvectors:4 characterise:2 amount:1 ten:1 supplied:1 sign:2 correctly:2 mat:3 group:2 four:2 sum:6 run:1 jose:1 powerful:1 thorn:2 place:3 reasonable:1 yann:1 oscillation:5 decision:5 rix:1 prefer:1 layer:15 bound:1 occur:3 bp:1 hy:3 meiosis:3 speed:1 performing:1 dawley:1 magnit:1 according:1 march:1 jr:1 smaller:2 riance:1 across:1 em:2 unity:1 cun:2 slowed:1 taken:2 derivat:5 imate:2 previously:2 fail:1 imation:1 end:1 adopted:1 wii:1 experimentation:1 eight:1 denker:2 skeletonization:1 subtracted:1 alternative:2 slower:1 cent:1 clustering:1 include:2 hinge:2 sanger:1 const:1 calculating:1 archit:1 ting:1 build:2 february:1 comparatively:1 dat:2 ude:1 already:5 added:2 quantity:1 receptive:5 damage:1 rt:1 traditional:1 diagonal:1 exhibit:1 gradient:4 september:1 separate:1 unable:1 numher:1 useless:2 insufficient:1 ching:1 localised:3 trace:5 ba:1 implementation:2 enclosing:1 allowing:2 markov:1 descent:2 behave:1 anti:1 varied:1 ucsd:1 householder:1 introduced:2 david:4 required:5 hanson:2 california:1 learned:4 able:1 below:1 pattern:10 usually:1 smolensky:4 sified:1 solut:1 timur:1 suitable:2 scheme:2 improve:1 aston:1 splitter:5 picture:1 created:2 carried:3 review:1 acknowledgement:1 determining:1 relative:1 ash:1 degree:1 sufficient:1 undergone:1 editor:5 elsewhere:1 changed:1 placed:3 free:1 heading:1 side:2 allow:1 pulled:1 perceptron:4 normalised:2 institute:1 sification:1 taking:1 benefit:3 regard:1 boundary:5 raining:4 calculated:7 tolerance:1 world:1 forward:7 author:2 made:4 collection:1 san:4 far:1 approximate:6 pruning:9 lippmann:1 ances:2 anda:1 hayes:1 sat:1 unnecessary:1 inpnt:1 disrupt:1 continuous:1 iterative:3 promising:3 learn:4 nature:1 ca:4 improving:1 anthony:1 rue:1 main:1 bmi:1 erion:1 big:1 noise:3 n2:1 nothing:2 allowed:2 repeated:1 advice:2 malvern:2 position:9 inferring:1 exponential:1 candidate:1 down:1 remained:1 gating:1 oyal:1 er:2 list:1 decay:2 alt:1 gained:1 importance:1 gat:2 te:1 hoped:1 push:1 hod:1 updat:4 flannery:1 led:1 simply:1 likely:3 failed:1 ordered:1 scalar:1 corresponds:1 tle:1 teukolsky:1 rice:1 ives:2 marked:1 raint:1 fut:1 change:1 generalisation:1 determined:2 principal:12 degradation:1 e:6 la:1 est:5 perceptrons:1 select:1 filler:1 constructive:9 tested:1 avoiding:1 correlated:1 |
4,761 | 5,310 | Dimensionality Reduction with Subspace Structure
Preservation
Ifeoma Nwogu
Department of Computer Science
SUNY Buffalo
Buffalo, NY 14260
[email protected]
Devansh Arpit
Department of Computer Science
SUNY Buffalo
Buffalo, NY 14260
[email protected]
Venu Govindaraju
Department of Computer Science
SUNY Buffalo
Buffalo, NY 14260
[email protected]
Abstract
Modeling data as being sampled from a union of independent subspaces has been
widely applied to a number of real world applications. However, dimensionality reduction approaches that theoretically preserve this independence assumption
have not been well studied. Our key contribution is to show that 2K projection
vectors are sufficient for the independence preservation of any K class data sampled from a union of independent subspaces. It is this non-trivial observation that
we use for designing our dimensionality reduction technique. In this paper, we
propose a novel dimensionality reduction algorithm that theoretically preserves
this structure for a given dataset. We support our theoretical analysis with empirical results on both synthetic and real world data achieving state-of-the-art results
compared to popular dimensionality reduction techniques.
1
Introduction
A number of real world applications model data as being sampled from a union of independent
subspaces. These applications include image representation and compression [7], systems theory
[13], image segmentation [16], motion segmentation [14], face clustering [8, 6] and texture segmentation [9], to name a few. Dimensionality reduction is generally used prior to applying these
methods because most of these algorithms optimize expensive loss functions like nuclear norm, `1
regularization, e.t.c. Most of these applications simply apply off-the-shelf dimensionality reduction
techniques or resize images (in case of image data) as a pre-processing step.
The union of independent subspace model can be thought of as a generalization of the traditional
approach of representing a given set of data points using a single low dimensional subspace (e.g.
Principal Component Analysis). For the application of algorithms that model data at hand with this
independence assumption, the subspace structure of the data needs to be preserved after dimensionality reduction. Although a number of existing dimensionality reduction techniques [11, 4, 1, 5] try
to preserve the spacial geometry of any given data, no prior work has tried to explicitly preserve the
independence between subspaces to the best of our knowledge.
In this paper, we propose a novel dimensionality reduction technique that preserves independence
between multiple subspaces. In order to achieve this, we first show that for any two disjoint subspaces with arbitrary dimensionality, there exists a two dimensional subspace such that both the
1
subspaces collapse to form two lines. We then extend this non-trivial idea to multi-class case and
show that 2K projection vectors are sufficient for preserving the subspace structure of a K class
dataset. Further, we design an efficient algorithm that finds the projection vectors with the aforementioned properties while being able to handle corrupted data at the same time.
2
Preliminaries
Let S1 , S2 . . . SK be K subspaces in Rn . We say that these K subspaces are independent if there
does not exist any non-zero vector in Si which is a linear combination of vectors in the other K ? 1
subspaces. Let the columns of the matrix Bi ? Rn?d denote the support of the ith subspace of d
dimensions. Then any vector in this subspace can be represented as x = Bi w ?w ? Rd . Now we
define the notion of margin between two subspaces.
Definition 1 (Subspace Margin) Subspaces Si and Sj are separated by margin ?ij if
?ij =
max
u?Si ,v?Sj
hu, vi
kuk2 kvk2
(1)
Thus margin between any two subspaces is defined as the maximum dot product between two unit
vectors (u, v), one from either subspace. Such a vector pair (u, v) is known as the principal vector
pair between the two subspaces while the angle between these vectors is called the principal angle.
With these definitions of independent subspaces and margin, assume that we are given a dataset
which has been sampled from a union of independent linear subspaces. Specifically, each class in
this dataset lies along one such independent subspace. Then our goal is to reduce the dimensionality
of this dataset such that after projection, each class continues to lie along a linear subspace and that
each such subspace is independent of all others. Formally, let X = [X1 , X2 . . . , XK ] be a K class
dataset in Rn such that vectors from class i (x ? Xi ) lie along subspace Si . Then our goal is to
? i := {P T x : x ? Xi }
find a projection matrix (P ? Rn?m ) such that the projected data vectors X
?
?
(i ? {1 . . . K}) are such that data vectors Xi belong to a linear subspace (Si in Rm ). Further, each
subspace S?i (i ? {1 . . . K}) is independent of all others.
3
Proposed Approach
In this section, we propose a novel subspace learning approach applicable to labeled datasets that
theoretically guarantees independent subspace structure preservation. The number of projection
vectors required by our approach is not only independent of the size of the dataset but is also fixed,
depending only on the number of classes. Specifically, we show that for any K class labeled dataset
with independent subspace structure, only 2K projection vectors are required for structure preservation.
The entire idea of being able to find a fixed number of projection vectors for the structure preservation of a K class dataset is motivated by theorem 2. This theorem states a useful property of any
pair of disjoint subspaces.
Theorem 2 Let unit vectors v1 and v2 be the ith principal vector pair for any two disjoint subspaces
S1 and S2 in Rn . Let the columns of the matrix P ? Rn?2 be any two orthonormal vectors in the
span of v1 and v2 . Then for all vectors x ? Sj , P T x = ?tj (j ? {1, 2}), where ? ? R depends on
x and tj ? R2 is a fixed vector independent of x. Further,
tT
1 t2
kt1 k2 kt2 k2
= v1T v2
Proof: We use the notation (M )j to denote the j th column vector of matrix M for any arbitrary
matrix M . We claim that tj = P T vj (j ? {1, 2}). Also, without any loss of generality, assume that
(P )1 = v1 . Then in order to prove theorem 2, it suffices to show that ?x ? S1 , (P )T2 x = 0. By
symmetry, ?x ? S2 , P T x will also lie along a line in the subspace spanned by the columns of P .
Let the columns of B1 ? Rn?d1 and B2 ? Rn?d2 be the support of S1 and S2 respectively, where d1
and d2 are the dimensionality of the two subspaces. Then we can represent v1 and v2 as v1 = B1 w1
and v2 = B2 w2 for some w1 ? Rd1 and w2 ? Rd2 . Let B1 w be any arbitrary vector in S1 where
2
(a) Independent subspaces in 3
dimensions
(b) Subspaces after projection
Figure 1: A three dimensional example of the application of theorem 2. See text in section 3 for
details.
w ? Rd1 . Then we need to show that T := (B1 w)T (P )2 = 0?w. Notice that,
T = (B1 w)T (B2 w2 ? (w1T B1T B2 w2 )B1 w1 )
= wT (B1T B2 w2 ? (w1T B1T B2 w2 )w1 ) ?w
(2)
Let U SV T be the svd of B1T B2 . Then w1 and w2 are the ith columns of U and V respectively, and
v1T v2 is the ith diagonal element of S if v1 and v2 are the ith principal vectors of S1 and S2 . Thus,
T = wT (U SV T w2 ? Sii (U )i )
(3)
= wT (Sii (U )i ? Sii (U )i ) = 0
Geometrically, this theorem says that after projection on the plane (P ) defined by any one of the
principal vector pairs between subspaces S1 and S2 , both the entire subspaces collapse to just two
lines such that points from S1 lie along one line while points from S2 lie along the second line.
Further, the angle that separates these lines is equal to the angle between the ith principal vector pair
between S1 and S2 if the span of the ith principal vector pair is used as P .
We apply theorem 2 on a three dimensional example as shown in figure 1. In figure 1 (a), the first
subspace (y-z plane) is denoted by red color while the second subspace is the black line in x-y axis.
Notice that for this setting, the x-y plane (denoted by blue color) is in the span of the 1st (and only)
principal vector pair between the two subspaces. After projection of both the entire subspaces onto
the x-y plane, we get two lines (figure 1 (b)) as stated in the theorem.
Finally, we now show that for any K class dataset with independent subspace structure, 2K projection vectors are sufficient for structure preservation.
n
Theorem 3 Let X = {x}N
i=1 be a K class dataset in R with Independent Subspace structure. Let
n?2K
P = [P1 . . . PK ] ? R
be a projection matrix for X such that the columns of the matrix Pk ?
Rn?2 consists
of
orthonormal
vectors in the span of any principal vector pair between subspaces
P
Sk and j6=k Sj . Then the Independent Subspace structure of the dataset X is preserved after
projection on the 2K vectors in P .
Before stating the proof of this theorem, we first state lemma 4 which we will use later in our proof.
This lemma states that if two vectors are separated by a non-zero angle, then after augmenting these
vectors with any arbitrary vectors, the new vectors remain separated by some non-zero angle as
well. This straightforward idea will help us extend the two subspace case in theorem 2 to multiple
subspaces.
Lemma 4 Let x1 , y1 be any two fixed vectors of same dimensionality with respect to each other
xT y
such that kx1 k12 ky11 k2 = ? where ? ? [0, 1). Let x2 , y2 be any two arbitrary vectors of same
dimensionality with respect to each other. Then there exists a constant ?? ? [0, 1) such that vectors
0T 0
y
x0 = [x1 ; x2 ] and y 0 = [y1 ; y2 ] are also separated such that kx0xk2 ky
? ?? .
0k
2
Proof of theorem 3:
3
Algorithm 1 Computation of projection matrix P
INPUT: X,K,?, itermax
for k=1 to K do
?
w2? ? random vector in RNk
while iter < itermax or ? not converged do
? w?
X
w1? ? maxw1 kXk w1 ? kX? kkw?2k2 k2 + ?kw1 k2
2
w1? ? w1? /norm(w1? )
X w?
? k w2 k2 + ?kw2 k2
w2? ? maxw2 k kXkkw?1k2 ? X
1
w2? ? w2? /norm(w2? )
? k w? )
? ? (Xk w1? )T (X
2
end while
? k w? ]
Pk ? [Xk w1? , X
2
end for
P ? ? [P1 . . . PK ]
OUTPUT: P ?
P
For the proof of theorem 3, it suffices to show that data vectors from subspaces Sk and j6=k Sj
(for any k ? {1 . . . K}) are
P separated by margin less than 1 after projection using P . Let x and y
be any vectors in Sk and j6=k Sj respectively and the columns of the matrix Pk be in the span of
the ith (say) principal vector pair between these subspaces. Using theorem 2, the projected vectors
by an angle equal to the the angle between the ith principal vector
PkT x and PkT y are separated
P
pair between Sk and j6=k Sj . Let the cosine of this angle be ?. Then, using lemma 4, the added
dimensions in the vectors PkT x and PkT y to form the vectors P T x and P T y are also separated by
some margin ?? < 1. As the same argument holds for vectors from all classes, the Independent
Subspace Structure of the dataset remains preserved after projection.
For any two disjoint subspaces, theorem 2 tells us that there is a two dimensional plane in which
the entire projected subspaces form two lines. It can be argued that after adding arbitrary valued
finite dimensions to the basis of this plane, the two projected subspaces will also remain disjoint
(see proof of theorem 3). Theorem 3 simply applies this argument to each subspace and the sum of
the remaining subspaces one at a time. Thus for K subspaces, we get 2K projection vectors.
Finally, our approach projects data to 2K dimensions which could be a concern if the original
feature dimension itself is less than 2K. However, since we are only concerned with data that has
underlying independent subspace assumption, notice that the feature dimension must be at least
K. This is because each class must lie on at least 1 dimension which is linearly independent of
others. However, this is too strict an assumption and it is straight forward to see that if we relax this
assumption to 2 dimensions for each class, the feature dimensions are already at 2K.
3.1
Implementation
A naive approach to finding projection vectors (say for a binary class case) would be to compute
the SVD of the matrix X1T X2 , where the columns of X1 and X2 contain vectors from class 1 and
class 2 respectively. For large datasets this would not only be computationally expensive but also
be incapable of handling noise. Thus, even though theorem 3 guarantees the structure preservation
of the dataset X after projection using P as specified, this does not solve the problem of dimensionality reduction. The reason is that given a labeled dataset sampled from a union of independent
subspaces, we do not have any information about the basis or even the dimensionality of the underlying subspaces. Under these circumstances, constructing the projection matrix P as specified
in theorem 3 itself becomes a problem. To solve this problem, we propose
P an algorithm that tries
to find the underlying principal vector pair between subspaces Sk and j6=k Sj (for k = 1 to K)
given the labeled dataset X. The assumption behind this attempt is that samples from each subspace
(class) are not heavily corrupted and that the underlying subspaces are independent.
Notice that we are not specifically interested in a particular principal vector pair between any two
subspaces for the computation of the projection matrix. This is because we have assumed independent subspaces and so each principal vector pair is separated by some margin ? < 1. Hence we
4
need an algorithm that computes any arbitrary principal vector pair, given data from two independent subspaces. These vectors can then be used to form one of the K submatrices in P as specified
in theorem 3 . ForPcomputing the submatrix Pk , we need to find a principal vector pair between
subspaces Sk and j6=k Sj . In terms of dataset X, we estimate the vector pair using data in Xk
? k where X
? k := X \ {Xk }. We repeat this process for each class to finally form the entire
and X
?
matrix P . Our approach is stated in algorithm 1. For each class k, the idea is to start with a random
? k and find the vector in Xk closest to this vector. Then fix this vector and
vector in the span of X
? k . Repeating this process till the convergence of the cosine between
search of the closest vector in X
these 2 vectors leads to a principal vector pair. In order to estimate the closest vector from opposite subspace, we have used a quadratic program in 1 that minimizes the reconstruction error of the
fixed vector (of one subspace) using vectors from the opposite subspace. The regularization in the
optimization is to handle noise in data.
3.2
Justification
The definition 1 for margin ? between two subspaces S1 and S2 can be equivalently expressed as
1 ? ? = min
w1 ,w2
1
kB1 w1 ? B2 w2 k2 s.t. kB1 w1 k2 = 1, kB2 w2 k2 = 1
2
(4)
where the columns of B1 ? Rn?d1 and B2 ? Rn?d2 are the basis of the subspaces S1 and S2
respectively such that B1T B1 and B2T B2 are both identity matrices.
Proposition 5 Let B1 ? Rn?d1 and B2 ? Rn?d2 be the basis of two disjoint subspaces S1 and
S2 . Then for any principal vector pair (ui , vi ) between the subspaces S1 and S2 , the corresponding
vector pair (w1 ? Rd1 ,w2 ? Rd2 ), s.t. ui = B1 w1 and vi = B2 w2 , is a local minima to the
objective in equation (4).
Proof: The Lagrangian function for the above objective is:
L(w1 , w2 , ?) =
1
1 T T
w B B1 w1 + w2T B2T B2 w2 ?w1T B1T B2 w2 +?1 (kB1 w1 k2 ?1)+?2 (kB2 w2 k2 ?1)
2 1 1
2
(5)
Then setting the gradient w.r.t. w1 to zero we get
?w1 L = (1 + ?1 )w1 ? B1T B2 w2 = 0
(6)
Let U SV T be the SVD of B1T B2 and w1 and w2 be the ith columns of U and V respectively. Then
equation (6) becomes
?w L = (1 + ?1 )w1 ? U SV T w2
= (1 + ?1 )w1 ? Sii w1 = 0
(7)
Thus the gradient w.r.t. w1 is zero when ?1 = 1 ? Sii . Similarly, it can be shown that the gradient
w.r.t. w2 is zero when ?2 = 1 ? Sii . Thus the gradient of the Lagrangian L is 0 w.r.t. both w1 and
w2 for every corresponding principal vector pair. Thus vector pair (w1 , w2 ) corresponding to any
of the principal vector pairs between subspaces S1 and S2 is a local minima to the objective 4.
Since (w1 , w2 ) corresponding to any principal vector pair between two disjoint subspaces form a
local minima to the objective given by equation (4), one can alternatively minimize equation (4)
w.r.t. w1 and w2 and reach one of the local minima. Thus, by assuming independent subspace
structure for all the K classes in algorithm 1 and setting ? to zero, it is straight forward to see that
the algorithm yields a projection matrix that satisfies the criteria specified by theorem 3.
Finally, real world data do not strictly satisfy the independent subspace assumption in general and
even a slight corruption in data may easily lead to the violation of this independence. In order to
tackle this problem, we add a regularization (? > 0) term while solving for the principal vector
pair in algorithm 1. If we assume that the corruption is not heavy, reconstructing a sample using
vectors belonging to another subspace would require a large coefficient over those vectors. The
regularization avoids reconstructing data from one class using vectors from another class that are
slightly corrupted by assigning such vectors small coefficients.
5
0.5
0.4
0.3
0.2
0.3
0.1
0.25
0.05
0.1
0
0
0.2
?0.1
?0.2
?0.05
0.15
?0.3
?0.4
?0.5
?0.5
?0.4
?0.3
?0.2
?0.1
0
0.1
0.2
0.3
0.4
0.5
(a) Data projected
using Pa
0.1
?0.1
0.05
?0.15
?0.2
0
?0.05
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
?0.25
?1
?0.9
?0.8
?0.7
(a)
?0.6
?0.5
?0.4
?0.3
?0.2
?0.1
0
(b)
0.6
0.4
0.05
0.05
0
0
0.2
0
?0.05
?0.05
?0.1
?0.1
?0.15
?0.15
?0.2
?0.4
?0.6
?0.8
?0.8
?0.6
?0.4
?0.2
0
0.2
0.4
?0.2
?0.25
?0.3
?0.3
0.6
?0.35
?1
(b) Data projected
using Pb
?0.9
?0.8
?0.7
?0.6
?0.5
(c)
?0.4
?0.3
?0.2
?0.1
0
?0.35
?1
?0.9
?0.8
?0.7
?0.6
?0.5
?0.4
?0.3
?0.2
?0.1
0
(d)
Figure 3: Four different pairs of classes from
the Extended Yale dataset B projected onto
two dimensional subspaces using proposed
approach. See section 4.1.1 for details.
Figure 2: Qualitative comparison between
(a) true projection matrix and (b) projection
matrix from the proposed approach on high
dimensional synthetic two class data. See
section 4.1.1 for details.
3.3
?0.2
?0.25
Complexity
Solving algorithm 1 requires solving an unconstrained quadratic program within a while-loop. Assume that we run this while loop for T iterations and that we use conjugate gradient descent to solve
the quadratic program in each iteration. Also, it is known that for any matrix A ? Ra?b and vector
b ? Ra , conjugate gradient applied to a problem of the form
arg minkAx ? bk2
(8)
w
?
takes time O(ab K), where K is the condition number of AT A. Thus it is straight forward to
see that the?time required to compute the projection matrix for a K class problem in our case is
O(KT nN K), where n is the dimensionality of feature space, N is the total number of samples
and K is the condition number of the matrix (XkT Xk + ?I). Here I is the identity matrix.
4
Empirical Analysis
In this section, we present empirical evidence to support our theoretical analysis of our subspace
learning approach. For real world data, we use the following datasets:
1. Extended Yale dataset B [3]: It consists of ? 2414 frontal face images of 38 individuals (K = 38)
with 64 images per person. These images were taken under constrained but varying illumination
conditions.
2. AR dataset [10]: This dataset consists of more than 4000 frontal face images of 126 individuals
with 26 images per person. These images were taken under varying illumination, expression and
facial disguise. For our experiments, similar to [15], we use images from 100 individuals (K =
100) with 50 males and 50 females. We further use only 14 images per class which correspond to
illumination and expression changes. This corresponds to 7 images from Session 1 and rest 7 from
Session 2.
3. PIE dataset [12]: The pose, illumination, and expression (PIE) database is a subset of CMU PIE
dataset consisting of 11, 554 images of 68 people (K = 68).
We crop all the images to 32 ? 32, and concatenate all the pixel intensity to form our feature vectors.
Further, we normalize all data vectors to have unit `2 norm.
6
(a) Yale dataset B
(b) AR dataset
(c) PIE dataset
Figure 4: Multi-class separation after projection using proposed approach for different
datasets. See section 4.1.2 for details.
4.1 Qualitative Analysis
4.1.1 Two Subspaces-Two Lines
We test both the claim of theorem 2 and the quality of approximation achieved by algorithm 1 in
this section. We perform these tests on both synthetic and real data.
1. Synthetic Data: We generate two random subspaces in R1000 of dimensionality 20 and 30 (notice
that these subspaces will be independent with probability 1). We randomly generate 100 data vectors
from each subspace and normalize them to have unit length. We then compute the 1st principal
vector pair between the two subspaces using their basis vectors by performing SVD of B1T B2 ,
where B1 and B2 are the basis of the two subspaces. We orthonormalize the vector pair to form the
projection matrix Pa . Next, we use the labeled dataset of 200 points generated to form the projection
matrix Pb by applying algorithm 1. The entire dataset of 200 points is then projected onto Pa and Pb
separately and plotted in figure 2. The green and red points denote data from either subspace. The
results not only substantiate our claim in theorem 2 but also suggest that the proposed algorithm for
estimating the projection matrix is a good approximation.
2. Real Data: Here we use Extended Yale dataset B for analysis. Since we are interested in projection of two class data in this experimental setup, we randomly choose 4 different pairs of classes
from the dataset and use the labeled data from each pair to generate the two dimensional projection
matrix (for that pair) using algorithm 1. The resulting projected data from the 4 pairs can be seen in
figure 3. As is evident from the figure, the projected two class data for each pair approximately lie
along two different lines.
4.1.2
Multi-class separability
We analyze the separation between the K classes of a given K-class dataset after dimensionality
reduction. First we compute the projection matrix for that dataset using our approach and project the
data. Second, we compute the top principal vector for each class separately from the projected data.
This gives us K vectors. Let the columns of the matrix Z ? R2K?K contain these vectors. Then
in order to visualize inter-class separability, we simply take the dot product of the matrix Z with
itself, i.e. Z T Z. Figure 4 shows this visualization for the three face datasets. The diagonal elements
represent self-dot product; thus the value is 1 (white). The off-diagonal elements represent interclass dot product and these values are consistently small (dark) for all the three datasets reflecting
between class separability.
4.2
Quantitative Analysis
In order to evaluate theorem 3, we perform a classification experiment on all the three real world
datasets mentioned above after projecting the data vectors using different dimensionality reduction
techniques. We compare our quantitative results against PCA, Linear discriminant analysis (LDA),
Regularized LDA and Random Projections (RP) 1 . We make use of sparse coding [15] for classification.
1
We also used LPP (Locality Preserving Projections) [4], NPE (Neighborhood Preserving Embedding) [5],
and Laplacian Eigenmaps [1] for dimensionality reduction on Extended Yale B dataset. However, because
the best performing of these reduction techniques yielded a result of only 73% compared to the close to 98%
accuracy from our approach, we do not report results from these methods.
7
For Extended Yale dataset B, we use all 38 classes for evaluation with 50% ? 50% train-test split 1
and 70% ? 30% train-test split 2. Since our method is randomized, we perform 50 runs of computing the projection matrix using algorithm 1 and report the mean accuracy with standard deviation.
Similarly for RP, we generate 50 different random matrices and then perform classification. Since
all other methods are deterministic, there is no need for multiple runs.
Table 1: Classification Accuracy on Extended Yale dataset B with 50%-50% train-test split.
See section 4.2 for details.
Method
Ours
PCA LDA Reg-LDA
RP
dim
76
76
37
37
76
acc
98.06 ? 0.18 92.54 83.68
95.77
93.78 ? 0.48
Table 2: Classification Accuracy on Extended Yale dataset B with 70%-30% train-test split.
See section 4.2 for details.
Method
Ours
PCA LDA Reg-LDA
RP
dim
76
76
37
37
76
acc
99.45 ? 0.20 93.98 93.85
97.47
94.72 ? 0.66
Table 3: Classification Accuracy on AR dataset. See section 4.2 for details.
Method
Ours
PCA LDA Reg-LDA
RP
dim
200
200
99
99
200
acc
92.18 ? 0.08 85.00
88.71
84.76 ? 1.36
Table 4: Classification Accuracy on a subset of CMU PIE dataset. See section 4.2 for details.
Method
Ours
PCA LDA Reg-LDA
RP
dim
136
136
67
67
136
acc
93.65 ? 0.08 87.76 86.71
92.59
90.46 ? 0.93
Table 5: Classification Accuracy on a subset of CMU PIE dataset. See section 4.2 for details.
Method
Ours
PCA LDA Reg-LDA
RP
dim
20
20
9
9
20
acc
99.07 ? 0.09 97.06 95.88
97.25
95.03 ? 0.41
For AR dataset, we take the 7 images from Session 1 for training and the 7 images from Session 2
for testing. The results are shown in table 3. The result using LDA is not reported because we found
that the summed within class covariance was degenerate and hence LDA was not applicable. It can
be clearly seen that our approach significantly outperforms other dimensionality reduction methods.
Finally for PIE dataset, we perform experiments on two different subsets. First, we take all the
68 classes and for each class, we randomly choose 25 images for training and 25 for testing. The
performance for this subset is shown in table 4. Second, we take only the first 10 classes of the
dataset and of all the 170 images per class, we randomly split the data into 70% ? 30% train-test set.
The performance for this subset is shown in table 5.
Evidently, our approach consistently yields the best performance on all the three datasets compared
to other dimensionality reduction methods.
5
Conclusion
We proposed a theoretical analysis on the preservation of independence between multiple subspaces.
We show that for K independent subspaces, 2K projection vectors are sufficient for independence
preservation (theorem 3). This result is motivated from our observation that for any two disjoint
subspaces of arbitrary dimensionality, there exists a two dimensional plane such that after projection, the entire subspaces collapse to just two lines (theorem 2). Resulting from this analysis, we
proposed an efficient iterative algorithm (1) that tries to exploit these properties for learning a projection matrix for dimensionality reduction that preserves independence between multiple subspaces.
Our empirical results on three real world datasets yield state-of-the-art results compared to popular
dimensionality reduction methods.
8
References
[1] Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps for dimensionality reduction and data
representation. Neural Comput., 15(6):1373?1396, June 2003.
[2] Deng Cai, Xiaofei He, and Jiawei Han. Efficient kernel discriminant analysis via spectral
regression. In Data Mining, 2007. ICDM 2007. Seventh IEEE International Conference on,
pages 427?432. IEEE, 2007.
[3] A.S. Georghiades, P.N. Belhumeur, and D.J. Kriegman. From few to many: Illumination cone
models for face recognition under variable lighting and pose. IEEE Trans. Pattern Anal. Mach.
Intelligence, 23(6):643?660, 2001.
[4] X. He and P. Niyogi. Locality preserving projections (lpp). Proc. of the NIPS, Advances in
Neural Information Processing Systems. Vancouver: MIT Press, 103, 2004.
[5] Xiaofei He, Deng Cai, Shuicheng Yan, and Hong-Jiang Zhang. Neighborhood preserving
embedding. In Computer Vision, 2005. ICCV 2005. Tenth IEEE International Conference on,
volume 2, pages 1208?1213 Vol. 2, Oct 2005.
[6] Jeffrey Ho, Ming-Husang Yang, Jongwoo Lim, Kuang-Chih Lee, and David Kriegman. Clustering appearances of objects under varying illumination conditions. In Computer Vision and
Pattern Recognition, 2003. Proceedings. 2003 IEEE Computer Society Conference on, volume 1, pages I?11?I?18. IEEE, 2003.
[7] Wei Hong, John Wright, Kun Huang, and Yi Ma. Multiscale hybrid linear models for lossy
image representation. Image Processing, IEEE Transactions on, 15(12):3655?3671, 2006.
[8] Guangcan Liu, Zhouchen Lin, and Yong Yu. Robust subspace segmentation by low-rank representation. In ICML, 2010.
[9] Yi Ma, Harm Derksen, Wei Hong, John Wright, and Student Member. Segmentation of multivariate mixed data via lossy coding and compression. IEEE Transactions on Pattern Analysis
and Machine Intelligence, 3, 2007.
[10] Aleix Mart??nez and Robert Benavente. AR Face Database, 1998.
[11] Sam T. Roweis and Lawrence K. Saul. Nonlinear dimensionality reduction by locally linear
embedding. Science, 290:2323?2326, December 2000.
[12] Terence Sim, Simon Baker, and Maan Bsat. The cmu pose, illumination, and expression (pie)
database. In Automatic Face and Gesture Recognition, 2002. Proceedings. Fifth IEEE International Conference on, pages 46?51. IEEE, 2002.
[13] Ren?e Vidal, Stefano Soatto, Yi Ma, and Shankar Sastry. An algebraic geometric approach to the
identification of a class of linear hybrid systems. In Decision and Control, 2003. Proceedings.
42nd IEEE Conference on, volume 1, pages 167?172. IEEE, 2003.
[14] Ren?e Vidal, Roberto Tron, and Richard Hartley. Multiframe motion segmentation with missing
data using powerfactorization and gpca. International Journal of Computer Vision, 79(1):85?
105, 2008.
[15] J. Wright, A.Y. Yang, A. Ganesh, S.S. Sastry, and Yi Ma. Robust face recognition via sparse
representation. IEEEE TPAMI, 31(2):210 ?227, Feb. 2009.
[16] Allen Y Yang, John Wright, Yi Ma, and S Shankar Sastry. Unsupervised segmentation
of natural images via lossy data compression. Computer Vision and Image Understanding,
110(2):212?225, 2008.
9
| 5310 |@word compression:3 norm:4 nd:1 hu:1 d2:4 shuicheng:1 tried:1 covariance:1 lpp:2 reduction:21 liu:1 ours:5 outperforms:1 existing:1 si:5 assigning:1 must:2 john:3 concatenate:1 rd2:2 intelligence:2 plane:7 xk:7 ith:10 gpca:1 zhang:1 along:7 kvk2:1 sii:6 qualitative:2 prove:1 consists:3 x0:1 theoretically:3 inter:1 ra:2 p1:2 multi:3 v1t:2 ming:1 becomes:2 project:2 estimating:1 notation:1 underlying:4 baker:1 minimizes:1 finding:1 guarantee:2 quantitative:2 every:1 tackle:1 rm:1 k2:14 control:1 unit:4 before:1 local:4 mach:1 jiang:1 r1000:1 approximately:1 black:1 studied:1 collapse:3 bi:2 kw2:1 testing:2 union:6 b1t:9 empirical:4 yan:1 submatrices:1 thought:1 significantly:1 projection:40 pre:1 suggest:1 get:3 onto:3 close:1 shankar:2 applying:2 kb1:3 optimize:1 deterministic:1 lagrangian:2 missing:1 straightforward:1 nuclear:1 orthonormal:2 spanned:1 embedding:3 handle:2 notion:1 justification:1 heavily:1 designing:1 pa:3 element:3 expensive:2 recognition:4 continues:1 labeled:6 database:3 powerfactorization:1 mentioned:1 ui:2 complexity:1 kriegman:2 solving:3 basis:6 easily:1 georghiades:1 represented:1 train:5 separated:8 tell:1 neighborhood:2 widely:1 valued:1 solve:3 say:4 relax:1 niyogi:2 itself:3 tpami:1 evidently:1 cai:2 propose:4 reconstruction:1 product:4 loop:2 till:1 degenerate:1 achieve:1 kx1:1 roweis:1 normalize:2 ky:1 x1t:1 convergence:1 object:1 help:1 depending:1 stating:1 pose:3 augmenting:1 ij:2 sim:1 hartley:1 govind:1 argued:1 require:1 suffices:2 generalization:1 fix:1 preliminary:1 proposition:1 strictly:1 hold:1 wright:4 lawrence:1 visualize:1 claim:3 proc:1 applicable:2 mit:1 clearly:1 shelf:1 varying:3 june:1 devansh:1 consistently:2 rank:1 dim:5 nn:1 entire:7 jiawei:1 kkw:1 interested:2 pixel:1 arg:1 aforementioned:1 classification:8 denoted:2 art:2 constrained:1 summed:1 equal:2 itermax:2 yu:1 icml:1 unsupervised:1 others:3 t2:2 report:2 richard:1 few:2 belkin:1 randomly:4 preserve:6 individual:3 geometry:1 consisting:1 jeffrey:1 attempt:1 ab:1 mining:1 evaluation:1 kt2:1 violation:1 male:1 behind:1 tj:3 kt:1 facial:1 plotted:1 theoretical:3 column:12 modeling:1 ar:5 deviation:1 subset:6 kuang:1 eigenmaps:2 seventh:1 too:1 reported:1 corrupted:3 sv:4 synthetic:4 st:2 person:2 international:4 randomized:1 lee:1 off:2 terence:1 w1:32 benavente:1 choose:2 huang:1 multiframe:1 disguise:1 b2:18 coding:2 student:1 coefficient:2 satisfy:1 explicitly:1 vi:3 depends:1 later:1 try:3 analyze:1 red:2 start:1 simon:1 guangcan:1 contribution:1 minimize:1 partha:1 accuracy:7 yield:3 correspond:1 identification:1 ren:2 lighting:1 corruption:2 j6:6 straight:3 converged:1 acc:5 r2k:1 reach:1 definition:3 against:1 proof:7 sampled:5 dataset:42 govindaraju:1 popular:2 knowledge:1 color:2 dimensionality:29 lim:1 segmentation:7 reflecting:1 wei:2 though:1 generality:1 just:2 hand:1 multiscale:1 nonlinear:1 ganesh:1 lda:14 quality:1 lossy:3 xkt:1 name:1 contain:2 y2:2 true:1 regularization:4 hence:2 soatto:1 white:1 self:1 substantiate:1 cosine:2 criterion:1 hong:3 evident:1 tt:1 tron:1 motion:2 stefano:1 allen:1 image:23 novel:3 volume:3 extend:2 belong:1 slight:1 he:3 rd:1 unconstrained:1 automatic:1 zhouchen:1 similarly:2 session:4 sastry:3 dot:4 kw1:1 han:1 add:1 feb:1 closest:3 multivariate:1 female:1 incapable:1 binary:1 yi:5 preserving:5 minimum:4 seen:2 deng:2 belhumeur:1 preservation:9 multiple:5 gesture:1 lin:1 icdm:1 laplacian:2 crop:1 regression:1 circumstance:1 cmu:4 vision:4 iteration:2 represent:3 kernel:1 achieved:1 preserved:3 separately:2 w2:31 rest:1 strict:1 member:1 december:1 yang:3 split:5 concerned:1 independence:9 opposite:2 reduce:1 idea:4 arpit:1 motivated:2 expression:4 pca:6 algebraic:1 generally:1 useful:1 repeating:1 dark:1 locally:1 generate:4 exist:1 npe:1 notice:5 disjoint:8 per:4 blue:1 vol:1 key:1 iter:1 four:1 pb:3 achieving:1 suny:3 tenth:1 v1:6 geometrically:1 sum:1 cone:1 jongwoo:1 run:3 angle:9 chih:1 separation:2 decision:1 resize:1 submatrix:1 rnk:1 b2t:2 yale:8 quadratic:3 yielded:1 x2:5 yong:1 argument:2 span:6 min:1 performing:2 department:3 combination:1 belonging:1 conjugate:2 remain:2 slightly:1 reconstructing:2 separability:3 derksen:1 sam:1 s1:14 projecting:1 iccv:1 taken:2 computationally:1 equation:4 visualization:1 remains:1 bsat:1 end:2 vidal:2 apply:2 v2:7 spectral:1 ho:1 rp:7 original:1 top:1 clustering:2 include:1 remaining:1 exploit:1 kt1:1 society:1 objective:4 added:1 already:1 traditional:1 diagonal:3 gradient:6 subspace:99 venu:1 separate:1 discriminant:2 trivial:2 reason:1 assuming:1 length:1 equivalently:1 pie:8 setup:1 kun:1 robert:1 stated:2 design:1 implementation:1 anal:1 perform:5 observation:2 datasets:9 finite:1 descent:1 buffalo:9 xiaofei:2 extended:7 y1:2 rn:13 orthonormalize:1 interclass:1 arbitrary:8 intensity:1 david:1 pair:33 required:3 specified:4 w2t:1 nip:1 trans:1 able:2 pattern:3 program:3 max:1 green:1 natural:1 hybrid:2 regularized:1 representing:1 axis:1 naive:1 kb2:2 roberto:1 text:1 prior:2 geometric:1 understanding:1 vancouver:1 loss:2 mixed:1 sufficient:4 bk2:1 heavy:1 repeat:1 saul:1 face:8 mikhail:1 sparse:2 k12:1 fifth:1 dimension:10 world:7 avoids:1 computes:1 forward:3 projected:11 transaction:2 sj:9 b1:12 harm:1 assumed:1 xi:3 alternatively:1 search:1 iterative:1 spacial:1 sk:7 table:8 pkt:4 robust:2 symmetry:1 constructing:1 vj:1 pk:6 linearly:1 s2:13 noise:2 w1t:3 x1:4 ny:3 comput:1 lie:8 theorem:26 kuk2:1 xt:1 r2:1 concern:1 evidence:1 exists:3 adding:1 texture:1 illumination:7 margin:9 kx:1 locality:2 rd1:3 simply:3 appearance:1 nez:1 kxk:1 expressed:1 applies:1 corresponds:1 satisfies:1 ma:5 mart:1 oct:1 goal:2 identity:2 change:1 specifically:3 wt:3 principal:25 lemma:4 called:1 total:1 svd:4 experimental:1 formally:1 support:4 people:1 frontal:2 evaluate:1 reg:5 d1:4 handling:1 |
4,762 | 5,311 | Distributed Balanced Clustering via Mapping
Coresets
Aditya Bhaskara
Google NYC
[email protected]
MohammadHossein Bateni
Google NYC
[email protected]
Vahab Mirrokni
Google NYC
[email protected]
Silvio Lattanzi
Google NYC
[email protected]
Abstract
Large-scale clustering of data points in metric spaces is an important problem in
mining big data sets. For many applications, we face explicit or implicit size constraints for each cluster which leads to the problem of clustering under capacity
constraints or the ?balanced clustering? problem. Although the balanced clustering problem has been widely studied, developing a theoretically sound distributed
algorithm remains an open problem. In this paper we develop a new framework
based on ?mapping coresets? to tackle this issue. Our technique results in first
distributed approximation algorithms for balanced clustering problems for a wide
range of clustering objective functions such as k-center, k-median, and k-means.
1
Introduction
Large-scale clustering of data points in metric spaces is an important problem in mining big data
sets. Many variants of such clustering problems have been studied spanning, for instance, a wide
range of `p objective functions including the k-means, k-median, and k-center problems. Motivated
by a variety of big data applications, distributed clustering has attracted significant attention over the
literature [11, 4, 5]. In many of these applications, an explicit or implicit size constraint is imposed
for each cluster; e.g., if we cluster the points such that each cluster fits on one machine, the size
constraint is enforced by the storage constraint on each machine. We refer to this as balanced clustering. In the setting of network location problems, these are referred to as capacitated clustering
problems [6, 16, 17, 10, 3]. The distributed balanced clustering problem is also well-studied and
several distributed algorithms have been developed for it in the context of large-scale graph partitioning [21, 20]1 . Despite this extensive literature, none of the distributed algorithms developed for
the balanced version of the problem have theoretical approximation guarantees. The present work
presents the first such distributed algorithms for a wide range of balanced clustering problems with
provable approximation guarantees. To acheive this goal, we develop a new technique based on
mapping coresets.
A coreset for a set of points in a metric space is a subset of these points with the property that an
approximate solution to the whole point-set can be obtained given the coreset alone. An augmented
concept for coresets is the notion of composable coresets which have the following property: for a
collection of sets, the approximate solution to the union of the sets in the collection can be obtained
given the union of the composable coresets for the point sets in the collection. This notion was
1
A main difference between the balanced graph partitioning problems and balanced clustering problems
considered here is that in the graph partitioning problems a main objective function is to minimize the cut
function.
1
MapReduce model
Problem
Approximation
L-balanced k-center
O(1)
k-clustering in `p
O(p)
L-balanced k-clustering in `p
(O(p),2)
Streaming model
Problem
Approximation
L-balanced k-center
O(1)
k-clustering in `p
O(p)
L-balanced k-clustering in `p
(O(p),2)
Rounds
O(1)
O(1)
O(1)
Passes
O(1)
O(1)
O(1)
Table 1: Our contributions, all results hold for k < n1/2? , for constant > 0. We notice that for
the L-balanced k-clustering (p) general we get a bicriteria optimization (we can potentially open 2k
centers in our solutions).
formally defined in a recent paper by Indyk et al [14]. In this paper, we augment the notion of
composable coresets further, and introduce the concept of mapping coresets. A mapping coreset is a
coreset with an additional mapping of points in the original space to points in the coreset. As we will
see, this will help us solve balanced clustering problems for a wide range of objective functions and
a variety of massive data processing applications, including streaming algorithms and MapReduce
computations. Roughly speaking, this is how a mapping coreset is used to develop a distributed
algorithm for the balanced clustering problems: we first partition the data set into several blocks in
a specific manner. We then compute a coreset for each block. In addition, we compute a mapping
of points in the original space to points in the coreset. Finally, we collect all these coresets, and then
solve the clustering problem for the union of the coresets. We can them use the (inverse) map to get
back a clustering for the original points.
Our Contributions. In this paper, we introduce a framework for solving distributed clustering
problems. Using the concept of mapping coresets as described above, our framework applies to
balanced clustering problems, which are much harder than their unrestricted counterparts in terms
of approximation.
The rough template of our results is the following: given a single machine ?-approximation algorithm for a clustering problem (with or without balance constraints), we give a distributed algorithm
for the problem that has an O(?) approximation guarantee. Our results also imply streaming algorithms for such clustering problems, using sublinear memory and constant number of passes. More
precisely, we consider balanced clustering problems with an `p objective. For specific choice of p, it
captures the commonly used k-center, k-median and k-means objectives. Our results are also very
robust?for instance, bicriteria approximations (violating either the number of clusters or the cluster
sizes) on a single machine can be used to give distributed bicriteria approximation algorithms, with
a constant loss in the cost. This is particularly important for balanced versions of k-median and
k-means, for which we know of constant factor approximation to the cost only if we allow violating
one of the constraints. (Moreover, mild violation might not be terribly bad in certain applications,
as long as we obtain small cost.)
Finally, other than presenting the first distributed approximations for balanced clustering, our general framework also implies constant-factor distributed approximations for a general class of uncapacitated clustering problems (for which we are not aware of distributed algorithms with formal
guarantees). We summarize our new results in Table 1.
Related Work. The notion of coresets has been introduced in [2]. In this paper, we use the term
coresets to refer to an augmented notion of coresets, referred to as ?composable coresets? [14].
The notion of (composable) coresets are also related to the concept of mergeable summaries that
have been studied in the literature [1]. The main difference between the two is that aggregating
mergeable summaries does not increase the approximation error, while in the case of coresets the
error amplifies. The idea of using coresets has been applied either explicitly or implicitly in the
streaming model [12, 2] and in the MapReduce framework [15, 18, 5, 14]. However, none of the
previous work applies these ideas for balanced clustering problems.
2
There has been a lot of work on designing efficient distributed algorithms for clustering problems in
metric spaces. A formal computation model for the MapReduce framework has been introduced by
Karloff et al. [15]. The first paper that studied clustering problems in this model is by Ene et al. [11],
where the authors prove that one can use an ? approximation algorithm for the k-center or k-median
problem to obtain a 4? + 2 and a 10? + 3 approximation respectively for the k-center or k-median
problems in the MapReduce model. Subsequently Bahmani et al. [4] showed how to implement kmeans++ efficiently in the MapReduce model. Finally, very recently, Balcan et al. [5] demonstrate
how one can use an ? approximation algorithm for the k-means or k-median problem to obtain
coresets in the distributed (and MapReduce) setting. They however do not consider the balanced
clustering problems or the general set of clustering problems with the `p objective function.
The literature of clustering in the streaming model is also very rich. The first paper we are aware
of is due to Charikar et al. [7], who study the k-center problem in the classic streaming setting.
Subsquently Guha et al. [12] give the first single pass constant approximation algorithm to the kmedian problem. Following up on this, the memory requirements and the approximation factors of
their result were further improved by Charikar et al. in [8].
Finally, capacitated (or balanced) clustering is well studied in approximation algorithms [6, 16,
9], with constant factors known in some cases and only bicriteria in others. Our results may be
interpreted as saying that the capacity constraints may be a barrier to approximation, but are not a
barrier to parallelizability. This is the reason our approximation guarantees are bicriteria.
2
Preliminaries
In all the problems we study, we will denote by (V, d) the metric space we are working with. We
will denote n = |V |, the number of points in V . We will also write duv as short hand for d(u, v).
Given points u, v, we assume we have an oracle access to duv (or can compute it, as in geometric
settings). Formally, a clustering C of a set of points V is a collection of sets C1 , C2 , . . . , Cr which
partition V . Each cluster Ci has a center vi , and we define the ?`p cost? of this clustering as
!1/p
XX
p
costp (C) :=
d(v, vi )
.
(1)
i
v?Ci
When p is clear from the context, we will simply refer to this quantity as the cost of the clustering
and denote it cost(C).
Let us now define the L-balanced k-clustering problem with `p cost.
Definition 1 (L-balanced k-clustering (p)). Given (V, d) and a size bound L, find a clustering C of
V which has at most k clusters, at most L points in each cluster, and cluster centers v1 , . . . , vk so
as to minimize costp (C), the `p cost defined in Eq. (1).
The case p = 1 is the capacitated k-median and with p = ? is also known as the capacitated
k-center problem (with uniform capacities).
Definition 2 (Mapping and mapping cost). Given a multiset S and a set V , we call
P a bijective function f : V ? S a mapping from V to S and we define the cost of a mapping as v?V d(v, f (v))p .
Definition 3 (Clustering and optimal solution). Given a clustering problem P with an `p objective,
we define OP TP as the cost of the optimal solution to P.
3
Mapping coreset framework
The main idea behind our distributed framework is a new family of coresets that help in dealing with
balanced clustering.
Definition 4 (?-mapping coreset). Given a set of points V , a ?-mapping coreset for a clustering
problem P consists of a multiset S with elements from V , and a mapping from V to S such that the
total cost of the mapping is upper bounded by ? ? OP TPp . We define the size of a ?-mapping coreset
as the number of distinct elements in S.
Note that our definition does not prescribe the size of the mapping coreset ? this can be a parameter
we choose. We now define the composability of coresets.
3
Definition 5 (Composable ?-mapping coreset). Given disjoint sets of points V1 , V2 , . . . , Vm , and
corresponding ?-mapping coresets S1 , S2 , . . . , Sm , the coresets are said to be composable if we have
that ?i Si is a 2p ?-mapping coreset for ?i Vi (the overall map is the union of those for V1 , . . . , Vm ).
Remark. The non-trivial aspect of showing that coresets compose comes from the fact that we
compare the cost of mapping to the cost of OPTP on the union of Vi (which we need to show is not
too small). Our main theorem is now the following
Theorem 1. Let V be a set of points and suppose L, k, p ? 1 are parameters. Then for any
U ? V , there exists an algorithm that takes U as input, and produces a 2p -mapping coreset for the
2
?
L-balanced k-clustering (p) problem for U . The size of this coreset is O(k),
and the algorithm
uses space that is quadratic in |U |. Furthermore, for any partition V1 , V2 , . . . , Vr of V , the mapping
coresets produced by the algorithm on V1 , V2 , . . . , Vr compose.
Clustering via ?-mapping coresets
3.1
The theorem implies a simple general framework for distributed clustering:
1. Split the input into m chunks arbitrarily (such that each chunk fits on a machine), and
compute a (composable) 2p -mapping coreset for each of the chunks. For each point in the
coreset, assign a multiplicity equal to the number of points mapped to it (including itself).
2. Gather all the coresets (and multiplicities of their points) into one machine, and compute a
k-clustering of this multiset.
3. Once clusters for the points (and their copies) are found, we can ?map back?, and find a
clustering of the original points.
The idea is that in each chunk, the size of the coreset will be small, thus the union of the coresets
is small (and hence fits on one machine). The second step requires care: the clustering algorithm
should work when the points have associated multiplicities, and use limited memory. This is captured as follows.
Definition 6 (Space-efficient algorithm). Given an instance (V, d) for a k-clustering problem in
which V has N distinct points, each with some multiplicity, a sequential ?-approximation algorithm
is called space-efficient if the space used by the algorithm is O(N 2 ? poly(k)).
The framework itself is a very natural one, thus the key portions are the step of finding the mapping
coresets that (a) have small mapping cost and (b) compose well on arbitrary partition of the input,
and that of finding space efficient algorithms. Sections 4 and 5 give details of these two steps.
Further, because the framework is general, we can apply many ?levels? of it. This is illustrated
below in Section 3.2.
To prove the correctness of the framework, we also need to prove that moving from the original
points in a chunk to a coreset with multiplicities (as described in (1)) does not affect us too much in
the approximation. We prove this using a general theorem:
Theorem 2. Let f : V 7? S be a bijection. Let C be any clustering of V , and let C 0 denote the clustering of S obtained by applying a bijection f to the clustering C. Then there exists
P a choice of centers for C 0 such that cost(C 0 )p ? 22p?1 (cost(C)p + ?total ), where ?total denotes v?V d(v, f (v))p .
In our case, if we consider the set of points in the coreset with multiplicities, the mapping gives a
bijection, thus the above theorem applies in showing that the cost of clustering is not much more than
the ?mapping cost? given by the bijection. The theorem can also be used in the opposite direction,
as will be crucial in obtaining an approximation guarantee.
Preserving balanced property. The above theorem allows us to move back and forth (algorithmically) between clusterings of V and (the coreset with multiplicities) S as long as there is a small-cost
mapping. Furthermore, since f is a bijection, we have the property that if the clustering was balanced
in V , the corresponding one in S will be balanced as well, and vice versa.
Putting things together. Let us now see how to use the theorems to obtain approximation guarantees. Suppose we have a mapping f from V to the union of the coresets of the chunks (called
2
? is used to hide a logarithmic factor.
Here and elsewhere below, O(?)
4
S, which is a multi set), with total mapping cost ?total . Suppose also that we have an ? spaceefficient approximation algorithm for clustering S. Now we can use the Theorem 2 to show that in
S, there exists a clustering whose cost, raised to the p-th power, is at most 22p?1 (cost(C)p + ?total ).
This means that the approximation algorithm on S gives a clustering of cost (to the pth power)
? 22p?1 ?p (cost(C)p + ?total ). Finally, using Theorem 2 in the opposite direction, we can map
back the clusters from S to V and get a an upper bound on the clustering cost (to the pth power) of
22p?1 (22p?1 ?p (cost(C)p + ?total ) + ?total ). But now using Theorem 1, we know that for the f in our
algorithm, ?total ? 2p cost(C)p . So plugging this into the bound above, and after some manipulations
(and taking pth roots) we obtain that the cost of the final clustering is ? 32?cost(C). The details of
this calculation can be found in the supplementary material.
Remark. The approximation ratio (i.e., 32?) seems quite pessimistic. In our experiments, we have
observed (if we randomly partition the points initially) that the constants are much better (often at
most 1.5). The slack in our analysis arises mainly because of Theorem 2, in which the worst case in
the analysis is very unlikely to occur in practice.
3.2
Mapping Coresets for Clustering in MapReduce
The above distributed algorithm can be placed in the formal model for MapReduce introduced by
Karloff et al. [15].
The model has two main restrictions, one on the total number of machines and another on the
memory available on each machine. In particular, given an input of size N , and a sufficiently
small ? > 0, in the model there are N 1?? machines, each with N 1?? memory available for the
computation. As a result, the total amount of memory available to the entire system is O(N 2?2? ).
In each round a computation is executed on each machine in parallel and then the outputs of the
computation are shuffled between the machines.
In this model the efficiency of an algorithm is measured by the number of the ?rounds? of MapReduce
in the algorithm. A class of algorithms of particular interest are the ones that run in a constant
number of rounds. This class of algorithms are denoted MRC 0 .
The high level idea is to use coreset construction and a sequential space-efficient ?-approximation
algorithm (as outlined above). Unfortunately, this approach does not work as such in the MapReduce model because both the coreset construction algorithm, and the space-efficient algorithm, require memory quadratic in the size of their input. Therefore we perform multiple ?levels? of our
framework.
Given an instance (V, d), the MapReduce algorithm proceeds as follows:
1. Partition the points arbitrarily into 2n(1+?)/2 sets.
2. Compute the composable 2p -mapping coreset on each of the machines (in parallel) to obtain
e
f and the multisets S1 , S2 , . . . , S2n(1+?)/2 , each with roughly O(k)
distinct points.
3. Partition the computed coreset again into n1/4 sets.
4. Compute composable 2p -mapping coresets on each of the machines (in parallel) to obtain
e
f 0 , and multisets S10 , S20 , . . . , Sn0 1/4 , each with O(k)
distinct points.
5. Merge all the S10 , S20 , . . . , Sn0 1/4 on a single machine and compute a clustering using the
sequential space-efficient ?-approximation algorithm.
6. Map back the points in S10 , S20 , . . . , Sn0 1/4 to the points in S1 , S2 , . . . , S2n(1+?)/2 using the
function f 0?1 and obtain a clustering of the points in S1 , S2 , . . . , S2n(1+?)/2 .
7. Map back the points in S1 , S2 , . . . , S2n(1+?)/2 to the points in V using the function f ?1 and
thus obtain a clustering of the initial set of points.
Note that if k < n1/4? , for constant > ?, at every step of the MapReduce, the input size on
each machine is bounded by n(1??)/2 and thus we can run our coreset reduction and a space-efficient
algorithm (in which we think of the poly(k) as constant ? else we need minor modification). Furthermore if n1/4? ? k < n(1?)/2 , for constant > ?, we can exploit the trade-off between number
of rounds and approximation factor to get a similar result (refer to the supplement for details).
5
Figure 1: We split the input into m parts, compute mapping coresets for each part, and aggregate
them. We then compute a solution to this aggregate and map the clustering back to the input.
We are now ready to state our main theorem in the MapReduce framework:
Theorem 3. Given an instance (V, d) for a k-clustering problem, with |V | = n and a sequential
space-efficient ? approximation algorithm for the (L-balanced) k-clustering (p) problem, there exists a MapReduce algorithm that runs in O(1) rounds and obtains an O(?) approximation for the
(L-balanced) k-clustering (p) problem, for L, p ? 1 and 0 < k < n(1?)/2 (constant > 0).
The previous theorem combined with the results of Section 5 gives us the results presented in Table 1. Furthermore it is possible to extend this approach to obtain streaming algorithms via the same
techniques. We defer the details of this to the supplementary material.
4
Coresets and Analysis
We now come to the proof of our main result?Theorem 1. We give an algorithm to construct
coresets, and then show that coresets constructed this way compose.
Constructing composable coresets.
Suppose we are given a set of points V . We first show how to select a set of points S that are
close to each vertex in V , and use this set as a coreset with a good mapping f . The selection of
S uses a modification of the algorithm of Lin and Vitter [19] for k-median. We remark that any
approximation algorithm for k-median with `p objective can be used in place of the linear program
(as we did in our experiments, for p = ?, in which a greedy farthest point traversal can be used).
Consider a solution (x, y) to the following linear programming (LP) relaxation:
XX
min
d(u, v)p xuv
subject to
u
v
X
xuv = 1
for all u
(every u assigned to a center)
v
xuv ? yv
X
yu ? k
for all u, v
(assigned only to center)
(at most k centers)
u
0 ? xuv , yu ? 1
for all u, v.
In the above algorithms, we can always treat p ? log n, and in particular the case p = ?, as
p = log n. This introduces only negligible error in our computations but make them tractable. More
specifically, when working with p = log n, the power operators do not increase the size of the input
by more than a factor log n.
6
Rounding We perform a simple randomized rounding with weights scaled up by O(log n): round
each yu to 1 with a probability equal to min{1, yu (4 log n)/}. Let us denote this probability by yu0 ,
and the set of ?centers? thus obtained, by S. We prove the following (proof in the supplement)
Lemma 4. With probability (1?1/n), the set S of selected centers satisfies the following properties.
1. Each vertex has a relatively close selected center. In particular, for every u ? V , there is a
h
i1/p
P
center opened at distance at most (1 + ) v d(u, v)p xuv
.
2. Not too many centers are selected; i.e., |S| <
8k log n
.
Mapping and multiplicity. Once we have a set S of centers, we map every v ? V the center closest
to it, i.e., f (v) = arg mins?S d(v, s). If ms points in V are mapped to some s ? S, we set its
multiplicty to ms . This defines a bijection from V to the resulting multiset.
Composability of the coresets.
We now come to the crucial step, the proof of composability for the mapping coresets constructed
earlier, i.e., the ?furthermore? part of Theorem 1.
To show this, we consider any vertex sets V1 , V2 , . . . , Vm , and mapping coresets S1 , S2 , . . . , Sm
obtained by the rounding algorithm above. We have to prove that the total moving cost is at most
(1 + )2p OP TP , where the optimum value is for the instance ?i Vi . We denote by LP (Vi ) the
optimum value of the linear program
above, when the set of points involved is Vi . Finally, we write
P
?v := d(v, fv )p , and ?total := v?V ?v . We now have:
Lemma 5. Let LPi denote the objective value of the optimum solution to LP (Vi ), i.e., the LP
relaxation written earlier when only vertices in Vi are considered. Then we have
X
?total ? (1 + )
LPi .
i
The proof follows directly from Lemma 4 and the definition of f . The next lemma is crucial: it
shows that LP (V ) cannot be too small. The proof is deferred to the supplement.
P
Lemma 6. In the notation above, we have i LPi ? 2p ? LP (V ).
The two lemmas imply that the total mapping cost is at most (1 + )2p OP TP , because LP (V ) is
clearly ? OP TP . This completes the proof of Theorem 1.
5
Space efficient algorithms on a single machine
Our framework ultimately reduces distributed computation to a sequential computation on a compressed instance. For this, we need to adapt the known algorithms on balanced k-clustering, in order
to handle compressed instances. We now give a high level overview and defer the details to the
supplementary material.
For balanced k-center, we modify the linear programming (LP) based algorithm of [16], and its analysis to deal with compressed instances. This involves the following trick: if we have a compressed
instance with N points, since there are only k centers to open, at most k ?copies? of each point are
candidate centers. We believe this trick can be applied more generally to LP based algorithms.
For balanced k-clustering with other `p objectives (even p = 1), it is not known how to obtain
constant factor approximation algorithms (even without the space efficient restriction). Thus we
consider bicriteria approximations, in which we respect the cluster size constraints, but have up to
2k clusters. This can be done for all `p objectives as follows: first solve the problem approximately
without enforcing the balanced constraint, then post-process the clusters obtained. If a cluster contains ni points for ni > L, then subdivide the cluster into dni /Le many clusters. The division
should be done carefully (see supplement).
The post-processing step only involves the counts of the vertices in different clusters, and hence can
be done in a space efficient manner. Thus the crucial part is to find the ?unconstrained? k-clustering
in a space efficient way. For this, the typical algorithms are either based on local search (e.g., due
7
Graph
US
World
Relative size of
sequential instance
0.33%
0.1%
Relative increase
in radius
+52%
+58%
Table 2: Quality degradation due to the two-round approach.
Figure 2: Scalability of parallel implementation.
to [13]), or based on rounding linear programs. The former can easily be seen to be space efficient
(we only need to keep track of the number of centers picked at each location). The latter can be
made space efficient using the same trick we use for k-center.
6
Empirical study
In order to gauge its practicality, we implement our algorithm. We are interested in measuring its
scalability in addition to the effect of having several rounds on the quality of the solution.
In particular, we compare the quality of the solution (i.e., the maximum radius from the k-center
objective) produced by the parallel implementation to that of the sequential one-machine implementation of the farthest seed heuristic. In some sense, our algorithm is a parallel implementation of
this algorithm. However, the instance is too big for the sequential algorithm to be feasible. As a
result, we run the sequential algorithm on a small sample of the instance, hence a potentially easier
instance.
Our experiments deal with two instances to test this effect: the larger instance is the world graph
with hundreds of millions of nodes, and the smaller one is the graph of US road networks with tens
of millions of nodes. Each node has the coordinate locations, which we use to compute great-circle
distances?the closest distance between two points on the surface of the earth. We always look for
1000 clusters, and run our parallel algorithms on a few hundred machines.
Table 2 shows that the quality of the solution does not degrade substantially if we use the tworound algorithm, more suited to parallel implementation. The last column shows the increase in the
maximum radius of clusters due to computing the k-centers in two rounds as described in the paper.
Note that the radius increase numbers quoted in the table are upper bounds since the sequential
algorithm could only be run on a simpler instance. In reality, the quality reduction may be even less.
1
subset of the actual
In case of the US Graph, the sequential algorithm was run on a random 300
1
graph, whereas a random 1000 subset was used for the World Graph.
We next investigate how the running time of our algorithm scales with the size of the instance. We
focus on the bigger instance (World Graph) and once again take its random samples of different
sizes (10% up to 100%). This yields to varying instance sizes, but does not change the structure of
the problem significantly, and is perfect for measuring scalability. Figure 2 shows the increase in
running time is sublinear. In particular, a ten-fold increase in instance size only leads to a factor 3.6
increase in running time.
8
References
[1] P. K. AGARWAL , G. C ORMODE , Z. H UANG , J. P HILLIPS , Z. W EI , AND K. Y I, Mergeable
summaries, in Proceedings of the 31st symposium on Principles of Database Systems, ACM,
2012, pp. 23?34.
[2] P. K. AGARWAL , S. H AR -P ELED , AND K. R. VARADARAJAN, Approximating extent measures of points, Journal of the ACM (JACM), 51 (2004), pp. 606?635.
[3] H.-C. A N , A. B HASKARA , AND O. S VENSSON, Centrality of trees for capacitated k-center,
CoRR, abs/1304.2983 (2013).
[4] B. BAHMANI , B. M OSELEY, A. VATTANI , R. K UMAR , AND S. VASSILVITSKII, Scalable
k-means++, PVLDB, 5 (2012), pp. 622?633.
[5] M.-F. BALCAN , S. E HRLICH , AND Y. L IANG, Distributed clustering on graphs, in NIPS,
2013, p. to appear.
[6] J. BAR -I LAN , G. KORTSARZ , AND D. P ELEG, How to allocate network centers, J. Algorithms, 15 (1993), pp. 385?415.
[7] M. C HARIKAR , C. C HEKURI , T. F EDER , AND R. M OTWANI, Incremental clustering and
dynamic information retrieval, in Proceedings of the Twenty-ninth Annual ACM Symposium
on Theory of Computing, STOC ?97, New York, NY, USA, 1997, ACM, pp. 626?635.
[8] M. C HARIKAR , L. O?C ALLAGHAN , AND R. PANIGRAHY, Better streaming algorithms for
clustering problems, in In Proc. of 35th ACM Symposium on Theory of Computing (STOC,
2003, pp. 30?39.
[9] J. C HUZHOY AND Y. R ABANI, Approximating k-median with non-uniform capacities, in
SODA, 2005, pp. 952?958.
[10] M. C YGAN , M. H AJIAGHAYI , AND S. K HULLER, LP rounding for k-centers with nonuniform hard capacities, in FOCS, 2012, pp. 273?282.
[11] A. E NE , S. I M , AND B. M OSELEY, Fast clustering using mapreduce, in KDD, 2011, pp. 681?
689.
[12] S. G UHA , N. M ISHRA , R. M OTWANI , AND L. O?C ALLAGHAN, Clustering data streams,
STOC, (2001).
[13] A. G UPTA AND K. TANGWONGSAN, Simpler analyses of local search algorithms for facility
location, CoRR, abs/0809.2554 (2008).
[14] P. I NDYK , S. M AHABADI , M. M AHDIAN , AND V. M IRROKNI, Composable core-sets for
diversity and coverage maximization, in unpublished, 2014.
[15] H. J. K ARLOFF , S. S URI , AND S. VASSILVITSKII, A model of computation for mapreduce, in
SODA, 2010, pp. 938?948.
[16] S. K HULLER AND Y. J. S USSMANN, The capacitated k-center problem, SIAM J. Discrete
Math., 13 (2000), pp. 403?418.
[17] M. R. KORUPOLU , C. G. P LAXTON , AND R. R AJARAMAN, Analysis of a local search heuristic for facility location problems, in SODA, 1998, pp. 1?10.
[18] S. L ATTANZI , B. M OSELEY, S. S URI , AND S. VASSILVITSKII, Filtering: a method for solving graph problems in mapreduce, in SPAA, 2011, pp. 85?94.
[19] J.-H. L IN AND J. S. V ITTER, Approximation algorithms for geometric median problems, Inf.
Process. Lett., 44 (1992), pp. 245?249.
[20] F. R AHIMIAN , A. H. PAYBERAH , S. G IRDZIJAUSKAS , M. J ELASITY, AND S. H ARIDI, Jabe-ja: A distributed algorithm for balanced graph partitioning, in SASO, 2013, pp. 51?60.
[21] J. U GANDER AND L. BACKSTROM, Balanced label propagation for partitioning massive
graphs, in WSDM, 2013, pp. 507?516.
9
| 5311 |@word mild:1 version:2 seems:1 open:3 bicriteria:6 harder:1 bahmani:2 reduction:2 initial:1 contains:1 silviol:1 com:4 si:1 attracted:1 written:1 partition:7 kdd:1 alone:1 greedy:1 selected:3 pvldb:1 short:1 core:1 multiset:4 bijection:6 location:5 node:3 math:1 simpler:2 c2:1 constructed:2 symposium:3 focs:1 prove:6 consists:1 compose:4 vitter:1 manner:2 introduce:2 theoretically:1 roughly:2 multi:1 wsdm:1 actual:1 xx:2 moreover:1 bounded:2 notation:1 interpreted:1 substantially:1 developed:2 finding:2 guarantee:7 every:4 tackle:1 scaled:1 partitioning:5 farthest:2 appear:1 negligible:1 aggregating:1 treat:1 modify:1 local:3 despite:1 haskara:1 merge:1 approximately:1 might:1 studied:6 collect:1 limited:1 range:4 union:7 block:2 implement:2 practice:1 empirical:1 significantly:1 road:1 varadarajan:1 get:4 cannot:1 close:2 selection:1 operator:1 storage:1 context:2 applying:1 restriction:2 imposed:1 map:8 center:34 attention:1 coreset:29 classic:1 handle:1 notion:6 coordinate:1 construction:2 suppose:4 massive:2 programming:2 sn0:3 us:2 designing:1 prescribe:1 trick:3 element:2 particularly:1 cut:1 database:1 observed:1 capture:1 worst:1 trade:1 balanced:39 traversal:1 dynamic:1 ultimately:1 solving:2 division:1 efficiency:1 easily:1 distinct:4 fast:1 aggregate:2 whose:1 quite:1 widely:1 solve:3 supplementary:3 heuristic:2 larger:1 compressed:4 think:1 itself:2 indyk:1 final:1 kmedian:1 bateni:2 forth:1 scalability:3 amplifies:1 cluster:21 requirement:1 optimum:3 produce:1 perfect:1 incremental:1 help:2 develop:3 measured:1 minor:1 op:5 eq:1 coverage:1 involves:2 implies:2 come:3 direction:2 radius:4 subsequently:1 opened:1 terribly:1 material:3 require:1 ja:1 assign:1 preliminary:1 pessimistic:1 hold:1 sufficiently:1 considered:2 great:1 seed:1 mapping:44 earth:1 proc:1 label:1 correctness:1 vice:1 gauge:1 rough:1 clearly:1 always:2 mergeable:3 cr:1 varying:1 focus:1 vk:1 mainly:1 sense:1 streaming:8 unlikely:1 entire:1 initially:1 i1:1 interested:1 saso:1 issue:1 overall:1 arg:1 augment:1 denoted:1 raised:1 equal:2 aware:2 once:3 construct:1 having:1 yu:4 look:1 others:1 few:1 randomly:1 n1:4 ab:2 interest:1 mining:2 investigate:1 deferred:1 violation:1 introduces:1 behind:1 tree:1 circle:1 theoretical:1 vahab:1 instance:21 earlier:2 column:1 ar:1 tp:4 measuring:2 maximization:1 cost:32 vertex:5 subset:3 uniform:2 hundred:2 rounding:5 guha:1 too:5 combined:1 chunk:6 st:1 randomized:1 siam:1 vm:3 off:1 together:1 again:2 choose:1 vattani:1 xuv:5 diversity:1 coresets:40 explicitly:1 vi:9 stream:1 root:1 lot:1 picked:1 portion:1 yv:1 parallel:8 spaceefficient:1 defer:2 contribution:2 minimize:2 ni:2 who:1 efficiently:1 yield:1 mohammadhossein:1 produced:2 none:2 mrc:1 definition:8 iang:1 pp:16 involved:1 associated:1 proof:6 ormode:1 carefully:1 back:7 parallelizability:1 violating:2 improved:1 done:3 furthermore:5 implicit:2 working:2 hand:1 ei:1 propagation:1 google:8 defines:1 quality:5 believe:1 usa:1 effect:2 concept:4 counterpart:1 former:1 hence:3 assigned:2 shuffled:1 facility:2 illustrated:1 deal:2 round:10 m:2 presenting:1 bijective:1 demonstrate:1 balcan:2 recently:1 duv:2 overview:1 million:2 extend:1 significant:1 refer:4 versa:1 nyc:4 unconstrained:1 outlined:1 moving:2 access:1 surface:1 closest:2 recent:1 showed:1 hide:1 inf:1 manipulation:1 certain:1 arbitrarily:2 upta:1 captured:1 preserving:1 additional:1 unrestricted:1 optp:1 care:1 seen:1 multiple:1 sound:1 reduces:1 adapt:1 calculation:1 long:2 lin:1 retrieval:1 post:2 bigger:1 plugging:1 variant:1 scalable:1 metric:5 lpi:3 agarwal:2 c1:1 addition:2 whereas:1 else:1 median:12 completes:1 crucial:4 pass:2 acheive:1 subject:1 tangwongsan:1 thing:1 call:1 split:2 variety:2 affect:1 fit:3 tpp:1 opposite:2 karloff:2 idea:5 vassilvitskii:3 motivated:1 allocate:1 speaking:1 york:1 remark:3 generally:1 clear:1 amount:1 ten:2 laxton:1 capacitated:6 notice:1 disjoint:1 algorithmically:1 track:1 harikar:2 write:2 discrete:1 key:1 putting:1 lan:1 v1:6 graph:14 relaxation:2 enforced:1 run:7 inverse:1 soda:3 place:1 saying:1 family:1 bound:4 fold:1 quadratic:2 oracle:1 annual:1 occur:1 constraint:10 precisely:1 s10:3 aspect:1 min:3 relatively:1 charikar:2 developing:1 smaller:1 lp:10 backstrom:1 modification:2 s1:6 multiplicity:8 ene:1 remains:1 slack:1 count:1 know:2 tractable:1 available:3 apply:1 v2:4 s2n:4 centrality:1 subdivide:1 original:5 denotes:1 clustering:82 running:3 umar:1 exploit:1 practicality:1 approximating:2 objective:13 move:1 eled:1 quantity:1 mirrokni:2 said:1 distance:3 mapped:2 capacity:5 degrade:1 extent:1 trivial:1 spanning:1 provable:1 reason:1 enforcing:1 panigrahy:1 ratio:1 balance:1 executed:1 unfortunately:1 potentially:2 stoc:3 implementation:5 twenty:1 perform:2 upper:3 sm:2 nonuniform:1 ninth:1 arbitrary:1 introduced:3 unpublished:1 extensive:1 s20:3 uang:1 fv:1 nip:1 bar:1 proceeds:1 below:2 summarize:1 program:3 including:3 memory:7 power:4 natural:1 imply:2 ne:1 multisets:2 ready:1 literature:4 mapreduce:18 geometric:2 relative:2 loss:1 sublinear:2 filtering:1 composable:12 gather:1 principle:1 elsewhere:1 summary:3 placed:1 last:1 copy:2 formal:3 allow:1 wide:4 template:1 face:1 barrier:2 taking:1 dni:1 distributed:23 lett:1 world:4 rich:1 author:1 collection:4 commonly:1 made:1 pth:3 approximate:2 obtains:1 implicitly:1 keep:1 dealing:1 quoted:1 costp:2 search:3 table:6 reality:1 robust:1 spaa:1 obtaining:1 poly:2 constructing:1 did:1 main:8 big:4 whole:1 s2:6 lattanzi:1 augmented:2 referred:2 ny:1 vr:2 explicit:2 candidate:1 uncapacitated:1 bhaskara:1 composability:3 theorem:19 bad:1 specific:2 showing:2 exists:4 sequential:11 corr:2 ci:2 supplement:4 uri:2 easier:1 suited:1 logarithmic:1 simply:1 jacm:1 aditya:1 applies:3 satisfies:1 acm:5 goal:1 kmeans:1 feasible:1 change:1 hard:1 specifically:1 typical:1 lemma:6 degradation:1 silvio:1 total:16 pas:1 called:2 eder:1 formally:2 select:1 gander:1 latter:1 arises:1 |
4,763 | 5,312 | Zeta Hull Pursuits:
Learning Nonconvex Data Hulls
?
Yuanjun Xiong? Wei Liu? Deli Zhao Xiaoou Tang?
Information Engineering Department, The Chinese University of Hong Kong, Hong Kong
?
IBM T. J. Watson Research Center, Yorktown Heights, New York, USA
Advanced Algorithm Research Group, HTC, Beijing, China
{yjxiong,xtang}@ie.cuhk.edu.hk [email protected] deli [email protected]
Abstract
Selecting a small informative subset from a given dataset, also called column sampling, has drawn much attention in machine learning. For incorporating structured
data information into column sampling, research efforts were devoted to the cases
where data points are ?tted with clusters, simplices, or general convex hulls. This
paper aims to study nonconvex hull learning which has rarely been investigated in
the literature. In order to learn data-adaptive nonconvex hulls, a novel approach
is proposed based on a graph-theoretic measure that leverages graph cycles to
characterize the structural complexities of input data points. Employing this measure, we present a greedy algorithmic framework, dubbed Zeta Hulls, to perform
structured column sampling. The process of pursuing a Zeta hull involves the
computation of matrix inverse. To accelerate the matrix inversion computation
and reduce its space complexity as well, we exploit a low-rank approximation to
the graph adjacency matrix by using an ef?cient anchor graph technique. Extensive experimental results show that data representation learned by Zeta Hulls can
achieve state-of-the-art accuracy in text and image classi?cation tasks.
1
Introduction
In the era of big data, a natural idea is to select a small subset of m samples Ce = {xe1 , . . . , xem }
from a whole set of n data points X = {x1 , . . . , xn } such that the selected points Ce can capture
the underlying properties or structures of X . Then machine learning and data mining algorithms can
be carried out with Ce instead of X , thereby leading to signi?cant reductions in computational and
space complexities. Let us write the matrix forms of Ce and X as C = [xe1 , . . . , xem ] ? Rd?m
and X = [x1 , . . . , xn ] ? Rd?n , respectively. Here d is the dimensions of input data points. In
other words, C is a column subset selection of X. The task of selecting C from X is also called by
column sampling in the literature, and maintains importance in a variety of ?elds besides machine
learning, such as signal processing, geoscience and remote sensing, and applied mathematics. This
paper concentrates on solving the column sampling problem by means of graph-theoretic methods.
Existing methods in column sampling fall into two main categories according to their objectives: 1)
approximate the data matrix X, and 2) discover the underlying data structures. For machine learning
methods using kernel or similar ?N-Body? techniques, the Nystr?om matrix approximation is usually
applied to approximate large matrices. Such circumstances include fast training of nonlinear kernel
support vector machines (SVM) in the dual form [30], spectral clustering [8], manifold learning [25],
etc. Minimizing a relative approximation error is typically harnessed as the objective of column sampling, by which the most intuitive solution is to perform uniform sampling [30]. Other non-uniform
sampling schemes choose columns via various criteria, such as probabilistic samplings according
to diagonal elements of a kernel matrix [7], reconstruction errors [15], determinant measurements
[1], cluster centroids [33], and statistical leverage scores [21]. On the other hand, column sampling
1
may be cast into a combinatorial optimization problem, which can be tackled by using greedy strategies in polynomial time [4] and boosted by using advanced sampling strategies to further reduce the
relative approximation error [14].
From another perspective, we are aware that data points may form some interesting structures. Understanding these structures has been proven bene?cial to approximate or represent data inputs [11].
One of the most famous algorithms for dimensionality reduction, Non-negative Matrix Factorization
(NMF) [16], learns a low-dimensional convex hull from data points through a convex relaxation [3].
This idea was extended to signal separation by pursuing a convex hull with a maximized volume
[27] to enclose input data points. Assuming that vertices are equally distant, the problem of ?tting
a simplex with a maximized volume to data reduces to a simple greedy column selection procedure
[26]. The simplex ?tting approach demonstrated its success in face recognition tasks [32]. Parallel research in geoscience and remote sensing is also active, where the vertices of a convex hull
are coined as endmembers or extreme points, leading to a classic ?N-Finder? algorithm [31]. The
above approaches tried to learn data structures that are usually characterized by convexity. Hence,
they may fail to reveal the intrinsic data structures when the distributions of data points are diverse,
e.g., data being on manifolds or concave structures. Probabilistic models like Determinantal Point
Process (DPP) [13] measure data densities, so they are likely to overcome the convexity issue. However, few previous work accessed structural information of possibly nonconvex data for column
sampling/subset selection tasks.
This paper aims to address the issue of learning nonconvex structures of data in the case where
the data distributions can be arbitrary. More speci?cally, we learn a nonconvex hull to encapsulate
the data structure. The on-hull points tightly enclose the dataset but do not need to form a convex
set. Thus, nonconvex hulls can be more adaptive to capture practically complex data structures.
Akin to convex hull learning, our proposed approach also extracts extreme points from an input
dataset. To complete this task, we start with exploring the property of graph cycles in a neighborhood
graph built over the input data points. Using cycle-based measures to characterize data structures
has been proven successful in clustering data of multiple types of distributions [34]. To induce a
measure of structural complexities stemming from graph cycles, we introduce the Zeta Function
which applies the integration of graph cycles to model the linkage properties of the neighborhood
graph. The key advantage of the Zeta function is uniting both global and local connection properties
of the graph. As such, we are able to learn a hull which encompasses almost all input data points
but is not necessary to be convex. With structural complexities captured in the form of the Zeta
function, we present a leave-one-out strategy to ?nd the extreme points. The basic idea is that
removing the on-hull points only has weak impact on structural complexities of the graph. The
decision of removal will be based on extremeness of a data point. Our model, dubbed Zeta Hulls, is
derived by computing and analyzing the extremeness of data points. The greedy pursuit of the Zeta
Hull model requires the computation of the inversion of a matrix obtained from the graph af?nity
matrix, which is computationally prohibitive for massive-scale data. To accelerate such a matrix
manipulation, we employ the Anchor Graph [18] technique in the sense that the original graph can
be approximated with respect to the anchors originating from a randomly sampled data subset. Our
model is testi?ed through extensive experiments on toy data and real-world text and image datasets.
Experimental results show that in terms of unsupervised data representation learning, the Zeta Hull
based methods outperform the state-of-the-art methods used in convex hull learning, clustering,
matrix factorization, and dictionary learning.
2
Nonconvex Hull Learning
To elaborate on our approach, we ?rst introduce and de?ne the point extremeness. It measures the
degree of a data point being prone to lie on or near a nonconvex hull by virtue of a neighborhood
graph drawn from an input dataset. As an intuitive criterion, the data point with strong connections
in the graph should have the low point extremeness. To obtain the extremeness measure, we need to
explore the underlying structure of the graph, where graph cycles are employed.
2.1
Zeta Function and Structural Complexity
We model graph cycles by means of a sum-product rule and then integrate them using a Zeta function. There are many variants of original Riemann Zeta Function, one of which is specialized in
2
(b) Remaining Graph
(a) Original Graph
Figure 1: An illustration of pursuing on-hull points using the graph measure. (a) shows a point set
with a k-nearest neighbor graph. Points in red are ones lying on the hull of the point set, e.g., the
points we tend to select by the Zeta Hull Pursuit algorithm. (b) shows the remaining point set and the
graph after removing the on-hull points together with their corresponding edges. We observe that
the removal of the on-hull (i.e., ?extreme?) points yields little impact on the structural complexity
of the graph.
weighted adjacency graphs. Applying the theoretical results of Zeta functions provides us a powerful tool for characterizing structural complexities of graphs. The numerical description of graph
structures will play a critical role in column sampling/subset selection tasks.
Formally, given a graph G(X , E) with n nodes being data points in X = {xi }ni=1 , let the n ? n
matrix W denote the weighted adjacency (or af?nity) matrix of the graph G built over the dataset X .
Usually the graph af?nities are calculated with a proper distance metric. To be generic, we assume
that G is directed. Then an edge leaving from xi to xj is denoted as eij . A path of length from
xi to xj is de?ned as P (i, j, ) = {ehk tk }k=1 with h1 = i and t = j. Note that the nodes in
this path can be duplicate. A graph cycle, as a special case of paths of length , is also de?ned as
? = P (i, i, ) (i = 1, . . . , n). The sum-product path af?nity ? for all -length cycles can then
?1
be computed by ? = ? ?? ?? = ? ?? wt?1 h1 k=1 whk tk , where ? denotes the set of all
possible cycles of length and whk tk denotes the (hk , tk )-entry of W, i.e., the af?nity from node
xhk to node xtk . The edge et?1 h1 is the last edge that closes the cycle. The computed compound
af?nity ? provides a measure for all cyclic connections of length . Then we integrate such af?nities
for the cycles of lengths being from one to in?nity to derive the graph Zeta function as follows,
?
z
?z (G) = exp
,
(1)
?
=1
where z is a constant. We only consider the situation where z is real-valued. The Zeta function
in Eq. (1) has been proven to enjoy a closed form. Its convergence is also guaranteed when z <
1/?(W), where ?(W) is the spectral radius of W. These lead to Theorem 1 [23].
Theorem 1. Let I be the identity matrix and ?(W) be the spectral radius of the matrix W, respectively. If 0 < z < 1/?(W), then ?z (G) = 1/ det(I ? zW).
Note that W can be asymmetric, implying that ?i can be complex. In this case, Theorem 1 still
holds. Theorem 1 indicates that the graph Zeta function we formulate in Eq. (1) provides a closedform expression for describing the structural complexity of a graph. The next subsection will give
the de?nition of the point extremeness by analyzing the structural complexity.
2.2
Zeta Hull Pursuits
From now on, for simplicity we use G = ?z (G) to represent the structural complexity of the original
graph G. To measure the point extremeness numerically, we perform a leave-one-out strategy in the
sense that each point in C is successively left out and the variation of G is investigated. This is a
natural way to pursue extreme points, because if a point xj lies on the hull it has few communications
with the other points. After removing this point and its corresponding edges, the reductive structural
complexity of the remaining graph G/xj , which we denote as G/xj , will still be close to G . Hence,
the point extremeness ?xj is modeled as the relative change of the structural complexity G , that is
G
?xj = G/x
. Now we have the following theorem.
j
Theorem 2. Given G and G/xj as in Theorem 1, the point extremeness measure ?xj of point xj
satis?es ?xj = (I ? zW)?1
(jj) , i.e., the point extremeness measure of point xj is equal to the j-th
diagonal entry of the matrix (I ? zW)?1 .
3
Algorithm 1 Zeta Hull Pursuits
Input: A dataset X , the number m of data points to be selected, and free parameters z, ? and k.
Output: The hull of sampled columns Ce := Cm+1 .
Initialize: construct W, C1 ? ?, X1 = X , c1 = 0, and W1 = W
for i = 1 to m do
?xj := (I ? zWi )?1
(jj) , for xj ? Xi
xei := arg minxj ?Xi (?xj + ?i e
j Wci )
Ci+1 := Ci ? xei
ci+1 := ci + eei
Xi+1 := Ci /xei
Wi+1 := Wi with the ei -th row and column removed
end for
According to previous analysis, the data point with a small ?xj tends to be on the hull and therefore
has a strong extremeness. To seek the on-hull points, we need to select a subset of m points Ce =
{xe1 , . . . , xem } from X such that they have the strongest point extremenesses. We formulate this
goal into the following optimization problem:
Ce = arg min g(C) + ?c Wc,
(2)
C?X
where c is a selection vector with m nonzero elements cei = 1 (i = 1, . . . , m), and g(C) is the
function which measures
mthe impact on the structural complexity after removing the extracted points.
In our case, g(C) = i=1 ?xci . The second term in Eq. (2) is a regularization term enforcing that
the selected data points do not intersect with each other. It will enable the selection process to have
a better representative capability. The parameter ? controls the extent of the regularization.
Naively solving the combinatorial optimization problem in Eq. (2) requires exponential time. By
adopting a greedy strategy, we can solve this optimization problem in an iterative manner and with
a feasible time complexity. Speci?cally, in each iteration we extract one point from the current data
set and add it to the subset of the selected points. Sticking to this greedy strategy, we will attain the
desired m on-hull points after m iterations. In the i-th iteration, we extract the point xei according
to the criterion
?
xei = arg min ?xj + e
Wci?1 ,
(3)
i j
xj ?Xi?1
where ej is the j-th standard basis vector, and ci?1 is the selection vector according to i ? 1 selected
points before the i-th iteration.
We name our algorithm Zeta Hull Pursuits in order to emphasize that we use the Zeta function to
pursue the nonconvex data hull. Algorithm 1 summarizes the Zeta Hull Pursuits algorithm.
3
Zeta Hull Pursuits via Anchors
Algorithm 1 is applicable to small to medium-scale data X due to its cubical time complexity and
quadratic space complexity with respect to the data size |X |. Here we propose a scalable algorithm
facilitated by a reasonable prior to tackle the nonconvex hull learning problem ef?ciently. The idea is
to build a low-rank approximation to the graph adjacency matrix W with a small number of sampled
data points, namely anchor points. We resort to the Anchor Graph technique [18], which has been
successfully applied to handle large-scale hashing[20] and semi-supervised learning problems.
3.1
Anchor Graphs
The anchor graph framework is an elegant way to approximate neighborhood graphs. It ?rst chooses
a subset of l anchor points U = {uj }lj=1 from X . Then for each data point in X , its s nearest anchors
in U are sought, thereby forming an s-nearest anchor graph. The anchor graph theory assumes that
the original graph af?nity matrix W can be reconstructed from the anchor graph with a small number
of anchors (l n). Anchor points can be selected by random sampling or a rough clustering
process. Many algorithms are available to embed a data point to its s nearest anchor points, as
?
suggested in
we adopt the simplest approach to build the anchor embedding matrix H;
[18]. Here
2
2
? ij = exp ?dij /? , uj ? {s nearest anchors of xi } , where dij is the distance from data
say, h
0,
otherwise
4
Algorithm 2
Anchor-based Zeta Hull Pursuits
Input: A dataset X , the number m of data points to be sampled, the number l of anchors, the
number s of nearest anchors, and a free parameter z.
Output: The hull of sampled columns Ce := Cm+1 .
Initialize: construct H, X1 = X , C1 = ?, and H1 = H
for i = 1 to m do
perform SVD to obtain Hi := U?VT
l
?2j
2
?xj := z k=1 1?z?
2 (Ujk ) , for xj ? Xi
k
?
xei := arg minxj ?Xi (?xj +
i hj ht )
xt ?Ci
Ci+1 := Ci ? xei
Xi+1 := Xi /xei
Hi+1 := Hi with the ei -th row removed
end for
point xi to anchor uj , and ? is a parameter controlling the bandwidth of the exponential function.
? is then normalized so that its every row sums to one. In doing so, we can approximate
The matrix H
? = H?
? ?1 H
? , where ? is a diagonal matrix whose i-th
the af?nity matrix of the original graph as W
? As a result, all matrix manipulations
diagonal element is equal to the sum of the i-th column of H.
upon the original graph af?nity matrix W can be approximated by substituting the anchor graph
? for W.
af?nity matrix W
3.2
Extremeness Computation via Anchors
Note that the computation of the point extremeness for ?xj depends on the diagonal elements of
(I ? zW)?1 . Using the anchor graph technique, we can write (I ? zW)?1 = (I ? zHH )?1 ,
? ? 12 . Thus we have the following theorem that enables an ef?cient computation of
where H = H?
?xj . The proof is detailed in the supplementary material.
Theorem 3. Let the singular vector decomposition of H be H = U?V , where ? =
l
?2k
2
diag(?1 , . . . , ?l ). If H H is not singular, then ??1
xj = 1 + z
k=1 1?z?2 (Ujk ) , where U =
HV??1 and Ujk denotes the (i, j)-th entry of U.
k
Theorem 3 reveals that the major computation of ?xj will reduce to the eigendecomposition of a
much smaller matrix H H, which results in a direct acceleration of the Zeta hull pursuit process.
At the same time,
term of Eq. (3) encountered in the i-th iteration can be estimated by
the second
1
h
h
,
where hj denotes the j-th row of H and ci?1 is the selection vector
e
j
t
j Wci = i
xt ?Ci
of the extracted point set before the i-th iteration. These lead to the Anchor-based Zeta Hull Pursuits
algorithm shown in Algorithm 2.
3.3
Downdating SVD
In Algorithm 2, the singular value decomposition dominates the total time cost. We notice that
reusing information in previous iterations can save the computation time. The removal of one row
from H is equivalent to a rank-one modi?cation to the original matrix. Downdating SVD [10] was
proposed to handle this operation. Given the diagonal singular value matrix ?i and the point xei
chosen in the i-th iteration, the singular value matrix ?i+1 for the next iteration can be calculated
1
by the eigendecomposition of an l ? l matrix D derived from ?i , where D = (I ? 1+?
h ei h
ei )?i ,
2
2
2
and ? + hei 2 = 1. The decomposition of D can be ef?ciently performed in O(l ) time [10].
Then the computation of Ui+1 is achieved by a multiplication of Ui with an l ? l matrix produced
by the decomposition operation on D, which permits a natural parallelism. Consequently, we can
further accelerate Algorithm 2 by using a parallel computing scheme.
3.4
Complexity Analysis
We now analyze the complexities of Algorithms 1 and 2. For Algorithm 1, the most time-consuming
step is to solve the matrix inverse of n ? n size, which costs a time complexity of O(n3 ). The
overall time complexity is thus O(mn3 ) for extracting m points. In the implementation we can use
5
(a) m = 20, ZHP
(b) m = 40, ZHP
(c) m = 80, ZHP
(d) m = 200, ZHP
(e) m = 20, A-ZHP
(f) m = 40, A-ZHP
(g) m = 80, A-ZHP
(h) m = 200, A-ZHP
(i) m = 40, Leverage Score
(j) m = 40, Simplex
(k) m = 40, CUR
(l) m = 40, K-medoids
Figure 2: Zeta hull pursuits on the two-moon toy dataset. We select m data points from the dataset
with various methods. In the sub-?gures, blue dots are data points. The selected samples are surrounded with red circles. The caption of each sub-?gure describes the number of selected points m
and the method used to select those data points. First two rows shows the results of our algorithms
with different m. The third row illustrates the comparisons with other methods when m = 40. For
the leverage score approach, we follow the steps in [21].
the sparse matrix computation to reduce the constant factor [5]. For Algorithm 2, the most timeconsuming step is to perform SVD over H, so the overall time complexity is O(mnl2 ). Leveraging
downdating SVD, we only need to calculate the full SVD of H once in O(nl2 ) time and iteratively
update the decomposition in O(l2 ) time per iteration. The matrix multiplication operation then
dominates the total time cost. Also, it can be parallelized using a multi-core CPU or a modern GPU,
resulting in a very small constant factor in the time complexity. Since l is usually less than 10% of n,
Algorithm 2 is orders of magnitude faster than Algorithm 1. For cases where l needs to be relatively
large (20% of n for example), the computational cost will not show a considerable increase since H
is usually a very sparse matrix.
4
Experiments
The Zeta Hull model aims at learning the structures of dataset. We evaluate how well our model
achieves this goal by performing classi?cation experiments. For simplicity, we abbreviate our algorithms as follows: the original Zeta Hull Pursuit algorithm (Algorithm 1), ZHP and its anchor
version (Algorithm 2), A-ZHP. To compare with the state-of-the-art, we choose some renowned
methods: K-medoids, CUR matrix factorization (CUR) [29], simplex volume maximization (Simplex) [26], sparse dictionary learning (DictLearn) [22] and convex non-negative matrix factorization
(C-NMF) [6]. Basically, we use the extracted data points to learn a representation for each data
point in an unsupervised manner. Classi?cation is done by feeding the representation into a classi?er. The representation will be built in two ways: 1) the sparse coding [22] and 2) the locality
simplex coding [26]. To differentiate our algorithms from the original anchor graph framework, we
conduct a set of experiments using the left singular vectors of the anchor embedding matrix H as
the representation. In these experiments, anchors used in the anchor graph technique are randomly
selected from the training set. To compare with existing low-dimension embedding approaches, we
run the Large-Scale Manifold method [24] using the same number of landmarks as that of extracted
points.
4.1 Toy Dataset
First we illustrate our algorithms on a toy dataset. The dataset, commonly known as ?the two
moons?, consists of 2000 data points on the 2D plane which are manifold-structured and comprise
nonconvex distributions. This experiment on the two moons provides illustrative results of our
algorithms in the presence of nonconvexity. We select different numbers of column subsets m =
{20, 40, 80, 200} and compare with various other methods. A visualization of the results is shown
in Figure 2. We can see that our algorithms can extract the nonconvex hull of the data cloud more
accurately.
4.2 Text and Image Datasets
For the classi?cation experiments in this section, we derive the two types of data representations (the
sparse coding and the local simplex coding) from the points/columns extracted by compared meth6
Table 1: Classi?cation error rates in percentage (%) on texts (TDT2 and Newsgroups) and handwritten number datasets (MNIST). The numbers in bold font highlight best results under the settings.
In this table, ?SC? refers to the results using the sparse coding to form the representation, while
?LSC? refers to the results using local simplex coding. The cells with ?-? indicate that the ZHP
method is too expensive to be performed under the associated settings. The ?Anchor Graph? refers
to the additional experiments using the original anchor graph framework [18].
Methods
ZHP
A-ZHP
Simplex [26]
DictLearn [22]
C-NMF [6]
CUR [29]
K-medoids [12]
Anchor Graph [18]
TDT2
m = 500
m = 1000
SC
LSC
SC
LSC
2.31
2.52
3.79
3.73
4.83
6.82
9.14
1.97
2.68
1.73
5.62
3.46
3.73
7.87
0.48
0.96
1.51
2.57
2.31
1.52
4.69
1.53
2.08
1.77
1.18
2.07
2.37
3.73
5.81
2.68
Newsgroups
m = 500
m = 1000
SC
LSC
SC
LSC
11.79
10.77
13.55
10.41
9.51
10.76
11.68
11.83
15.32
11.44
19.73
12.02
12.32
7.1
6.58
8.16
8.04
6.72
9.63
7.72
7.42
12.38
9.47
19.67
10.04
8.76
MNIST
m = 500
m = 2000
SC
LSC
SC
LSC
3.45
3.07
5.79
5.79
3.16
3.16
5.07
5.27
10.13
10.13
9.28
9.28
3.17
1.43
2.27
1.36
3.01
3.79
2.72
1.19
1.51
2.11
3.04
5.27
2.31
2.33
Table 2: Recognition error rates in percentage (%) on object and face datasets. We select L samples
for each class in the training set for training or forming the gallery. The numbers in bold font
highlight best results under the settings. In this table, ?SC? refers to the results using the sparse
coding to form the representation, while ?LSC? refers to the results using local simplex coding. The
?Raw Feature? refers to the experiments conducted on the raw features vectors. The face recognition
process is described in Sec. (4.2).
Methods
A-ZHP
Simplex [26]
DictLearn [22]
C-NMF [6]
CUR [21]
K-medoids [12]
Anchor Graph [18]
Large Manifold [24]
Raw Feature [28]
Caltech101
d = 21504, L = 30
m = 500
m = 1000
SC
LSC
SC
LSC
25.77
26.82
29.83
26.16
26.95
29.73
30.66
27.83
29.74
28.77
27.82
27.64
26.32
28.71
23.13
25.81
26.83
25.18
26.73
29.51
28.72
27.62
26.16
26.81
26.09
25.73
25.15
27.92
Caltech101
d = 5120, L = 30
m = 500
m = 1000
SC
LSC
SC
LSC
29.61
28.95
32.43
29.66
29.15
31.83
32.57
31.13
31.69
32.57
29.85
29.63
30.53
32.67
26.7
25.62
26.59
30.62
27.47
28.93
29.67
31.15
28.73
30.72
31.13
28.97
28.28
28.14
30.19
31.18
MultiPIE
d = 2000, L = 30
m = 500
m = 2000
SC
LSC
SC
LSC
14.2
15.8
20.8
17.5
21.9
19.8
20.8
19.9
19.6
20.4
21.3
29.7
11.3
13.7
19.7
14.8
21.6
17.7
19.6
17.7
18.5
19.9
20.7
25.4
17.6
31.4
14.4
30.1
27.6
ods. By measuring the performance of applying these representations to solving the classi?cation
tasks, we can evaluate the representative power of the compared point/column selection methods.
The sparse coding is widely used for obtaining the representation for classi?cation. Here a standard
1 -regularized projection algorithm (LASSO) [22] is adopted to learn the sparse representation from
the extracted data points. LASSO will deliver a sparse coef?cient vector, which is applied as the
representation of the data point. We use ?SC? to indicate the related results in Table 1 and Table 2.
The local simplex coding reconstructs one data point as a convex combination of a set of nearest
exemplar points, which form local simplexes [26]. Imposing this convex reconstruction constraint
leads to non-negative combination coef?cients. The sparse coef?cients vector will be used as data
representation. ?LSC? indicates the related results in Table 1 and Table 2.
The classi?cation pipeline is as follows. After extracting m points/columns from the training set,
all data points will be represented with these selected points using the two approaches above. Then
we feed the representations into a linear SVM for the training and testing. The better classi?cation
accuracy will reveal the stronger representative power of the column selection algorithm. In all
experiments, the parameter z is ?xed at 0.05 to guarantee the convergence of the Zeta function. We
?nd that ?nal results are robust to z once the convergence is guaranteed. For the A-ZHP algorithm,
the parameter s is ?xed at 10 and the number of anchor points l is set as 10% of the training set
size. The bandwidth parameter ? of the exponential function is tuned on the training set to obtain a
reasonable anchor embedding.
The classi?cation of text contents relies on the informative representation of the plain words or sentences. Two text datasets are adopted for classi?cation, i.e. the TDT2 dataset and the Newsgroups
dataset [2]. In experiments, a subset of TDT2 is used (TDT2-30). It has 9394 samples from 30
classes. Each feature vector is of 36771 dimensions and normalized into unit length. The training
set contains 6000 samples randomly selected from the dataset and rest of the samples are used for
7
testing. The parameter m is set to be 500 and 1000 on this dataset. The Newsgroups dataset contains 18846 samples from 20 classes. The training set contains 11314, while the testing set has 7532.
The two sets are separated in advance [2] and ordered in time sequence to be more challenging for
classi?ers. The parameter m is set to be 500 and 1000 on this dataset. The classi?cation results are
reported in Table 1.
For object and face recognition tasks we conduct experiments under three classic scenarios, the
hand-written digits classi?cation, the image recognition, and the human face recognition. Related
experimental results are reported in Table 1 and Table 2.
The MNIST dataset serves as a standard benchmark for machine learning algorithms. It contains 10
classes of images corresponding to hand-written numbers from 0 to 9. The training set has 60000
images and the testing set has 10000 images. Each sample is a 784-dimensional vector.
The Caltech101 dataset [17] is a widely used benchmark for object recognition systems. It consists
of images from 102 classes of objects (101 object classes and one background class). We randomly
select 30 labeled images from every class for training the classi?er and 3000 images for testing.
The recognition rates averaged over all classes are reported. Every image is processed into a feature
vector of 21504 dimensions by the method in [28]. We also conduct experiment on a feature subset
of the top 5000 dimensions (Caltech101-5k). In this experiment, m is set to be 500 and 1000.
On-hull points are extracted on the training set.
The MultiPIE human face dataset is a widely applied benchmark for face recognition [9]. We follow
a standard gallery-probe protocol of face recognition. The testing set is divided into the gallery set
and the probe set. The identity predication of a probe image comes from its nearest neighbor of
Euclidean distance in the gallery. We randomly select 30, 000 images of 200 subjects as the training
set for learning the data representation. Then we pick out 3000 images of the other 100 subjects
(L = 30) to form the gallery set and 6000 images as the probes. The head poses of all these faces
are between ?15 degrees. Each face image is processed into a vector of 5000 dimensions using the
local binary pattern descriptor and PCA. We vary the parameter m from 500 to 2000 to evaluate the
in?uence of number of sampled points.
Discussion. For the experiments on these high-dimensional datasets, the methods based on the
Zeta Hull model outperform most compared methods and also show promising performance improvements over raw data representation. When the number of extracted points grows, the resulting
classi?cation accuracy increases. This corroborates that the Zeta Hull model can effectively capture
intrinsic structures of given datasets. More importantly, the discriminative information is preserved
through learning these Zeta hulls. The representation yielded by the Zeta Hull model is sparse and of
manageable dimensionality (500-2000), which substantially eases the workload of classi?er training. This property is also favorable for tackling other large-scale learning problems. Due to the
graph-theoretic measure that uni?es the local and global connection properties of a graph, the Zeta
Hull model leads to better data representation compared against existing graph-based embedding
and manifold learning methods. For the comparison with the Large-Scale Manifold method [24]
on the MultiPIE dataset, we ?nd that even using 10K landmarks, its accuracy is still inferior to our
methods relying on the Zeta Hull model. We also notice that noise may also affect the quality of Zeta
hulls. This dif?culty can be circumvented by running a number of well-established outlier removal
methods such as [19].
5
Conclusion
In this paper, we proposed a geometric model, dubbed Zeta Hulls, for column sampling through
learning nonconvex hulls of input data. The Zeta Hull model was built upon a novel graph-theoretic
measure which quanti?es the point extremeness to unify local and global connection properties of
individual data point in an adjacency graph. By means of the Zeta function de?ned on the graph,
the point extremeness measure amounts to the diagonal elements of a matrix related to the graph
adjacency matrix. We also reduced the time and space complexities for computing a Zeta hull by
incorporating an ef?cient anchor graph technique. A synthetic experiment ?rst showed that the Zeta
Hull model can detect appropriate hulls for non-convexly distributed data. The extensive real-world
experiments conducted on benchmark text and image datasets further demonstrated the superiority
of the Zeta Hull model over competing methods including convex hull learning, clustering, matrix
factorization, and dictionary learning.
Acknowledgement This research is partially supported by project #MMT-8115038 of the Shun
Hing Institute of Advanced Engineering, The Chinese University of Hong Kong.
8
References
[1] M.-A. Belabbas and P. J. Wolfe. Spectral methods in machine learning and new strategies for very large
datasets. PNAS, 106(2):369?374, 2009.
[2] D. Cai, X. Wang, and X. He. Probabilistic dyadic data analysis with local and global consistency. In Proc.
ICML, 2009.
[3] M. Chu and M. Lin. Low dimensional polytope approximation and its application to nonnegative matrix
factorization. SIAM Journal of Computing, pages 1131?1155, 2008.
[4] A. Das and D. Kempe. Submodular meets spectral: greedy algorithms for subset selection, sparse approximation and dictionary selection. In Proc. ICML, 2011.
[5] T. Davis. SPARSEINV: a MATLAB toolbox for computing the sparse inverse subset using the Takahashi
equations, 2011.
[6] C. Ding, T. Li, and M. Jordan. Convex and semi-nonnegative matrix factorizations. TPAMI, 32(1):45?55,
2010.
[7] P. Drineas and M. Mahoney. On the Nystr?om method for approximating a gram matrix for improved
kernel-based learning. JMLR, 6:2153?2175, 2005.
[8] C. Fowlkes, S. Belongie, F. Chung, and J. Malik. Spectral grouping using the Nystr?om method. TPAMI,
26:214?225, 2004.
[9] R. Gross, I. Matthews, J. Cohn, T. Kanade, and S. Baker. Multi-pie. In Proc. Automatic Face Gesture
Recognition, pages 1?8, Sept 2008.
[10] M. Gu and S. C. Eisenstat. Downdating the singular value decomposition. SIAM Journal on Matrix
Analysis and Applications, 16(3):793?810, 1995.
[11] T. Hastie, R. Tibshirani, and J. J. H. Friedman. The elements of statistical learning, volume 1. Springer
New York, 2001.
[12] L. Kaufman and P. J. Rousseeuw. Finding groups in data: an introduction to cluster analysis, volume
344. John Wiley & Sons, 2009.
[13] A. Kulesza and B. Taskar. Determinantal point processes for machine learning. Foundations and Trends
in Machine Learning, 5(2?3), 2012.
[14] S. Kumar, M. Mohri, and A. Talwalkar. Ensemble Nystr?om method. In NIPS 23, 2009.
[15] S. Kumar, M. Mohri, and A. Talwalkar. Sampling methods for the Nystr?om method. JMLR, 13(1):981?
1006, 2012.
[16] D. D. Lee and H. S. Seung. Learning the parts of objects by non-negative matrix factorization. Nature,
401(6755):788?791, 1999.
[17] F. Li, B. Fergus, and P. Perona. Learning generative visual models from few training examples: An
incremental bayesian approach tested on 101 object categories. CVIU, 106(1):59?70, 2007.
[18] W. Liu, J. He, and S.-F. Chang. Large graph construction for scalable semi-supervised learning. In Proc.
ICML, 2010.
[19] W. Liu, G. Hua, and J. Smith. Unsupervised one-class learning for automatic outlier removal. In Proc.
CVPR, 2014.
[20] W. Liu, J. Wang, S. Kumar, and S.-F. Chang. Hashing with graphs. In Proc. ICML, 2011.
[21] M. W. Mahoney and P. Drineas. Cur matrix decompositions for improved data analysis. PNAS,
106(3):697?702, 2009.
[22] J. Mairal, F. Bach, J. Ponce, and G. Sapiro. Online learning for matrix factorization and sparse coding.
JMLR, 11:19?60, 2010.
[23] S. Savchenko. The Zeta-function and gibbs measures. Russian Mathematical Surveys, 48(1):189?190,
1993.
[24] A. Talwalkar, S. Kumar, M. Mohri, and H. Rowley. Large-scale SVD and manifold learning. JMLR,
14(1):3129?3152, 2013.
[25] A. Talwalkar, S. Kumar, and H. Rowley. Large-scale manifold learning. In Proc. CVPR, 2008.
[26] C. Thurau, K. Kersting, and C. Bauckhage. Yes we can: simplex volume maximization for descriptive
web-scale matrix factorization. In Proc. CIKM, 2010.
[27] F. Wang, C. Chi, T. Chan, and Y. Wang. Nonnegative least correlated component analysis for separation
of dependent sources by volume maximization. TPAMI, 32:875?888, 2010.
[28] J. Wang, J. Yang, K. Yu, F. Lv, T. Huang, and Y. Gong. Locality-constrained linear coding for image
classi?cation. In Proc. CVPR, 2010.
[29] S. Wang and Z. Zhang. A scalable cur matrix decomposition algorithm: lower time complexity and tighter
bound. In NIPS 26, 2012.
[30] C. Williams and M. Seeger. Using the Nystr?om method to speed up kernel machines. In NIPS 14, 2000.
[31] M. E. Winter. N-?nder: an algorithm for fast autonomous spectral end-member determination in hyperspectral data. In SPIE?s International Symposium on Optical Science, Engineering, and Instrumentation.
International Society for Optics and Photonics, 1999.
[32] Y. Xiong, W. Liu, D. Zhao, and X. Tang. Face recognition via archetype hull ranking. In Proc. ICCV,
2013.
[33] K. Zhang and J. Kwok. Density weighted Nystr?om method for computing large kernel eigensystems.
Neural Computation, 21:121?146, 2009.
[34] D. Zhao and X. Tang. Cyclizing clusters via Zeta function of a graph. In NIPS 22, 2008.
9
| 5312 |@word kong:3 determinant:1 version:1 inversion:2 polynomial:1 stronger:1 manageable:1 nd:3 seek:1 tried:1 decomposition:8 pick:1 eld:1 thereby:2 nystr:7 reduction:2 liu:5 cyclic:1 score:3 selecting:2 contains:4 tuned:1 existing:3 current:1 com:2 od:1 tackling:1 chu:1 written:2 gpu:1 determinantal:2 stemming:1 john:1 distant:1 numerical:1 informative:2 cant:1 enables:1 update:1 implying:1 greedy:7 selected:11 prohibitive:1 generative:1 plane:1 smith:1 core:1 gure:1 provides:4 node:4 accessed:1 zhang:2 height:1 mathematical:1 direct:1 symposium:1 consists:2 manner:2 introduce:2 multi:2 chi:1 xhk:1 relying:1 riemann:1 reductive:1 little:1 cpu:1 project:1 discover:1 underlying:3 baker:1 medium:1 xe1:3 xed:2 cm:2 kaufman:1 pursue:2 ehk:1 substantially:1 finding:1 dubbed:3 guarantee:1 cial:1 sapiro:1 every:3 concave:1 tackle:1 control:1 unit:1 enjoy:1 superiority:1 encapsulate:1 before:2 engineering:3 local:10 tends:1 era:1 analyzing:2 meet:1 path:4 china:1 challenging:1 dif:1 factorization:10 averaged:1 directed:1 testing:6 digit:1 procedure:1 intersect:1 attain:1 projection:1 word:2 induce:1 refers:6 close:2 selection:12 applying:2 equivalent:1 demonstrated:2 center:1 xci:1 timeconsuming:1 williams:1 attention:1 convex:14 survey:1 formulate:2 unify:1 simplicity:2 eisenstat:1 rule:1 importantly:1 classic:2 handle:2 embedding:5 variation:1 autonomous:1 tting:2 controlling:1 play:1 construction:1 massive:1 caption:1 element:6 wolfe:1 recognition:12 approximated:2 expensive:1 trend:1 asymmetric:1 labeled:1 role:1 cloud:1 taskar:1 ding:1 wang:6 capture:3 hv:1 calculate:1 cycle:12 remote:2 removed:2 gross:1 convexity:2 complexity:26 ui:2 rowley:2 seung:1 solving:3 htc:2 deliver:1 upon:2 basis:1 gu:1 drineas:2 accelerate:3 workload:1 xiaoou:1 various:3 represented:1 separated:1 fast:2 sc:15 neighborhood:4 whose:1 supplementary:1 valued:1 solve:2 say:1 widely:3 otherwise:1 belabbas:1 cvpr:3 online:1 differentiate:1 advantage:1 sequence:1 tpami:3 descriptive:1 cai:1 reconstruction:2 propose:1 product:2 cients:2 culty:1 nity:10 achieve:1 intuitive:2 description:1 sticking:1 rst:3 convergence:3 cluster:4 incremental:1 leave:2 tk:4 object:7 yuanjun:1 derive:2 illustrate:1 gong:1 pose:1 exemplar:1 ij:1 nearest:8 eq:5 strong:2 involves:1 signi:1 enclose:2 indicate:2 come:1 concentrate:1 radius:2 hull:65 human:2 enable:1 material:1 adjacency:6 shun:1 feeding:1 nities:2 tighter:1 exploring:1 hold:1 practically:1 lying:1 exp:2 algorithmic:1 thurau:1 matthew:1 substituting:1 major:1 achieves:1 dictionary:4 sought:1 vary:1 adopt:1 favorable:1 proc:10 applicable:1 combinatorial:2 successfully:1 tool:1 weighted:3 rough:1 aim:3 ej:1 hj:2 boosted:1 kersting:1 mn3:1 derived:2 downdating:4 ponce:1 improvement:1 rank:3 indicates:2 hk:2 seeger:1 centroid:1 talwalkar:4 sense:2 detect:1 dependent:1 typically:1 lj:1 perona:1 originating:1 issue:2 dual:1 arg:4 overall:2 denoted:1 art:3 integration:1 special:1 initialize:2 kempe:1 equal:2 construct:2 aware:1 once:2 comprise:1 sampling:17 yu:1 unsupervised:3 icml:4 simplex:13 duplicate:1 few:3 employ:1 modern:1 randomly:5 modi:1 winter:1 tightly:1 individual:1 friedman:1 satis:1 hing:1 mining:1 mahoney:2 photonics:1 extreme:5 devoted:1 edge:5 necessary:1 conduct:3 euclidean:1 desired:1 circle:1 uence:1 theoretical:1 column:23 measuring:1 maximization:3 cost:4 vertex:2 subset:14 entry:3 uniform:2 successful:1 dij:2 conducted:2 too:1 characterize:2 reported:3 synthetic:1 chooses:1 density:2 international:2 siam:2 ie:1 eas:1 probabilistic:3 lee:1 zeta:46 together:1 w1:1 successively:1 reconstructs:1 choose:2 possibly:1 huang:1 resort:1 zhao:4 leading:2 chung:1 toy:4 closedform:1 reusing:1 takahashi:1 li:2 de:5 coding:12 bold:2 sec:1 xem:3 eei:1 ranking:1 depends:1 performed:2 h1:4 closed:1 doing:1 analyze:1 red:2 start:1 maintains:1 parallel:2 capability:1 om:7 ni:1 accuracy:4 moon:3 descriptor:1 maximized:2 yield:1 ensemble:1 yes:1 weak:1 famous:1 handwritten:1 raw:4 accurately:1 produced:1 basically:1 multipie:3 bayesian:1 cubical:1 cation:16 strongest:1 nl2:1 coef:3 ed:1 against:1 simplexes:1 proof:1 associated:1 spie:1 cur:7 sampled:6 dataset:23 subsection:1 minxj:2 dimensionality:2 feed:1 hashing:2 supervised:2 follow:2 wei:1 improved:2 done:1 hand:3 web:1 ei:4 cohn:1 nonlinear:1 quality:1 reveal:2 grows:1 russian:1 usa:1 name:1 normalized:2 hence:2 regularization:2 nonzero:1 iteratively:1 inferior:1 davis:1 illustrative:1 yorktown:1 hong:3 criterion:3 theoretic:4 complete:1 image:18 novel:2 ef:5 specialized:1 harnessed:1 volume:7 he:2 numerically:1 measurement:1 imposing:1 gibbs:1 rd:2 automatic:2 consistency:1 mathematics:1 submodular:1 dot:1 etc:1 add:1 showed:1 chan:1 perspective:1 instrumentation:1 manipulation:2 compound:1 scenario:1 nonconvex:14 binary:1 watson:1 success:1 vt:1 renowned:1 nition:1 captured:1 additional:1 speci:2 employed:1 parallelized:1 cuhk:1 signal:2 semi:3 multiple:1 full:1 pnas:2 reduces:1 eigensystems:1 faster:1 characterized:1 af:11 gesture:1 bach:1 lin:1 determination:1 divided:1 equally:1 finder:1 impact:3 variant:1 basic:1 scalable:3 circumstance:1 metric:1 iteration:10 kernel:6 represent:2 adopting:1 achieved:1 cell:1 c1:3 preserved:1 background:1 singular:7 leaving:1 source:1 zw:5 rest:1 subject:2 tend:1 elegant:1 member:1 leveraging:1 jordan:1 ciently:2 structural:14 near:1 leverage:4 extracting:2 presence:1 yang:1 variety:1 xj:25 newsgroups:4 ujk:3 affect:1 hastie:1 bandwidth:2 lasso:2 competing:1 reduce:4 idea:4 cyclizing:1 det:1 expression:1 pca:1 linkage:1 effort:1 akin:1 york:2 jj:2 matlab:1 detailed:1 amount:1 mthe:1 rousseeuw:1 processed:2 category:2 simplest:1 reduced:1 outperform:2 percentage:2 wci:3 notice:2 deli:2 estimated:1 per:1 tibshirani:1 cikm:1 blue:1 diverse:1 write:2 group:2 key:1 drawn:2 ce:8 nal:1 ht:1 nonconvexity:1 graph:68 relaxation:1 sum:4 beijing:1 run:1 inverse:3 facilitated:1 powerful:1 xei:9 almost:1 reasonable:2 pursuing:3 separation:2 decision:1 summarizes:1 bound:1 hi:3 guaranteed:2 tackled:1 savchenko:1 quadratic:1 encountered:1 yielded:1 nonnegative:3 optic:1 constraint:1 n3:1 wc:1 endmembers:1 speed:1 min:2 kumar:5 performing:1 optical:1 xtk:1 relatively:1 ned:3 circumvented:1 structured:3 department:1 according:5 combination:2 smaller:1 describes:1 son:1 wi:2 constrained:1 outlier:2 iccv:1 medoids:4 pipeline:1 computationally:1 equation:1 visualization:1 hei:1 describing:1 fail:1 end:3 serf:1 adopted:2 pursuit:13 available:1 operation:3 permit:1 probe:4 observe:1 kwok:1 spectral:7 generic:1 appropriate:1 fowlkes:1 xiong:2 save:1 original:11 denotes:4 clustering:5 include:1 remaining:3 assumes:1 top:1 running:1 cally:2 exploit:1 coined:1 chinese:2 build:2 uj:3 approximating:1 society:1 objective:2 malik:1 font:2 strategy:7 diagonal:7 distance:3 landmark:2 manifold:9 polytope:1 extent:1 gallery:5 enforcing:1 assuming:1 besides:1 length:7 modeled:1 illustration:1 tdt2:5 minimizing:1 pie:1 negative:4 implementation:1 proper:1 perform:5 datasets:9 benchmark:4 predication:1 situation:1 extended:1 communication:1 head:1 arbitrary:1 nmf:4 cast:1 namely:1 bene:1 extensive:3 connection:5 sentence:1 toolbox:1 learned:1 established:1 nip:4 address:1 able:1 suggested:1 usually:5 parallelism:1 pattern:1 kulesza:1 encompasses:1 built:4 including:1 power:2 critical:1 natural:3 regularized:1 abbreviate:1 advanced:3 scheme:2 ne:1 carried:1 extract:4 sept:1 text:7 prior:1 literature:2 understanding:1 removal:5 l2:1 multiplication:2 geometric:1 relative:3 acknowledgement:1 highlight:2 interesting:1 proven:3 lv:1 eigendecomposition:2 integrate:2 foundation:1 degree:2 surrounded:1 ibm:2 row:7 prone:1 caltech101:4 mohri:3 supported:1 last:1 free:2 institute:1 fall:1 neighbor:2 face:12 characterizing:1 sparse:15 distributed:1 dpp:1 dimension:6 xn:2 overcome:1 world:2 calculated:2 plain:1 gram:1 commonly:1 adaptive:2 employing:1 reconstructed:1 approximate:5 emphasize:1 uni:1 global:4 active:1 reveals:1 anchor:38 mairal:1 xtang:1 belongie:1 consuming:1 xi:13 corroborates:1 discriminative:1 fergus:1 iterative:1 nder:1 table:11 promising:1 nature:1 learn:6 robust:1 kanade:1 obtaining:1 lsc:15 whk:2 investigated:2 complex:2 protocol:1 diag:1 quanti:1 da:1 main:1 mmt:1 big:1 whole:1 cei:1 noise:1 dyadic:1 x1:4 body:1 representative:3 cient:4 elaborate:1 simplices:1 wiley:1 sub:2 exponential:3 lie:2 jmlr:4 third:1 learns:1 uniting:1 tang:3 removing:4 theorem:10 embed:1 xt:2 er:4 sensing:2 svm:2 virtue:1 dominates:2 grouping:1 convexly:1 incorporating:2 intrinsic:2 naively:1 mnist:3 effectively:1 importance:1 ci:11 hyperspectral:1 magnitude:1 illustrates:1 cviu:1 locality:2 eij:1 likely:1 explore:1 forming:2 visual:1 ordered:1 geoscience:2 partially:1 chang:2 applies:1 springer:1 hua:1 bauckhage:1 gures:1 relies:1 extracted:8 identity:2 goal:2 acceleration:1 consequently:1 tted:1 feasible:1 change:1 considerable:1 content:1 wt:1 classi:19 called:2 total:2 experimental:3 e:3 svd:7 extremeness:16 rarely:1 select:9 formally:1 support:1 evaluate:3 tested:1 correlated:1 |
4,764 | 5,313 | The Bayesian Case Model: A Generative Approach
for Case-Based Reasoning and Prototype
Classification
Been Kim, Cynthia Rudin and Julie Shah
Massachusetts Institute of Technology
Cambridge, Massachusetts 02139
{beenkim, rudin, julie a shah}@csail.mit.edu
Abstract
We present the Bayesian Case Model (BCM), a general framework for Bayesian
case-based reasoning (CBR) and prototype classification and clustering. BCM
brings the intuitive power of CBR to a Bayesian generative framework. The BCM
learns prototypes, the ?quintessential? observations that best represent clusters in
a dataset, by performing joint inference on cluster labels, prototypes and important features. Simultaneously, BCM pursues sparsity by learning subspaces, the
sets of features that play important roles in the characterization of the prototypes.
The prototype and subspace representation provides quantitative benefits in interpretability while preserving classification accuracy. Human subject experiments
verify statistically significant improvements to participants? understanding when
using explanations produced by BCM, compared to those given by prior art.
1 Introduction
People like to look at examples. Through advertising, marketers present examples of people we
might want to emulate in order to lure us into making a purchase. We might ignore recommendations
made by Amazon.com and look instead at an Amazon customer?s Listmania to find an example of a
customer like us. We might ignore medical guidelines computed from a large number of patients in
favor of medical blogs where we can get examples of individual patients? experiences.
Numerous studies have demonstrated that exemplar-based reasoning, involving various forms of
matching and prototyping, is fundamental to our most effective strategies for tactical decisionmaking ([26, 9, 21]). For example, naturalistic studies have shown that skilled decision makers
in the fire service use recognition-primed decision making, in which new situations are matched to
typical cases where certain actions are appropriate and usually successful [21]. To assist humans in
leveraging large data sources to make better decisions, we desire that machine learning algorithms
provide output in forms that are easily incorporated into the human decision-making process.
Studies of human decision-making and cognition provided the key inspiration for artificial intelligence Case-Based Reasoning (CBR) approaches [2, 28]. CBR relies on the idea that a new situation
can be well-represented by the summarized experience of previously solved problems [28]. CBR
has been used in important real-world applications [24, 4], but is fundamentally limited, in that it
does not learn the underlying complex structure of data in an unsupervised fashion and may not
scale to datasets with high-dimensional feature spaces (as discussed in [29]).
In this work, we introduce a new Bayesian model, called the Bayesian Case Model (BCM), for
prototype clustering and subspace learning. In this model, the prototype is the exemplar that is most
representative of the cluster. The subspace representation is a powerful output of the model because
we neither need nor want the best exemplar to be similar to the current situation in all possible ways:
1
for instance, a moviegoer who likes the same horror films as we do might be useful for identifying
good horror films, regardless of their cartoon preferences. We model the underlying data using a
mixture model, and infer sets of features that are important within each cluster (i.e., subspace). This
type of model can help to bridge the gap between machine learning methods and humans, who use
examples as a fundamental part of their decision-making strategies.
We show that BCM produces prediction accuracy comparable to or better than prior art for standard
datasets. We also verify through human subject experiments that the prototypes and subspaces
present as meaningful feedback for the characterization of important aspects of a dataset. In these
experiments, the exemplar-based output of BCM resulted in statistically significant improvements
to participants? performance of a task requiring an understanding of clusters within a dataset, as
compared to outputs produced by prior art.
2 Background and Related Work
People organize and interpret information through exemplar-based reasoning, particularly when they
are solving problems ([26, 7, 9, 21]). AI Cased-Based Reasoning approaches are motivated by
this insight, and provide example cases along with the machine-learned solution. Studies show
that example cases significantly improve user confidence in the resulting solutions, as compared to
providing the solution alone or by also displaying a rule that was used to find the solution [11].
However, CBR requires solutions (i.e. labels) for previous cases, and does not learn the underlying
structure of the data in an unsupervised fashion. Maintaining transparency in complex situations
also remains a challenge [29]. CBR models designed explicitly to produce explanations [1] rely on
the backward chaining of the causal relation from a solution, which does not scale as complexity
increases. The cognitive load of the user also increases with the complexity of the similarity measure
used for comparing cases [14]. Other CBR models for explanations require the model to be manually
crafted in advance by experts [25].
Alternatively, the mixture model is a powerful tool for discovering cluster distributions in an unsupervised fashion. However, this approach does not provide intuitive explanations for the learned
clusters (as pointed out in [8]). Sparse topic models are designed to improve interpretability by reducing the number of words per topic [32, 13]. However, using the number of features as a proxy for
interpretability is problematic, as sparsity is often not a good or complete measure of interpretability
[14]. Explanations produced by mixture models are typically presented as distributions over features. Even users with technical expertise in machine learning may have a difficult time interpreting
such output, especially when the cluster is distributed over a large number of features [14].
Our approach, the Bayesian Case Model (BCM), simultaneously performs unsupervised clustering
and learns both the most representative cases (i.e., prototypes) and important features (i.e., subspaces). BCM preserves the power of CBR in generating interpretable output, where interpretability
comes not only from sparsity but from the prototype exemplars.
In our view, there are at least three widely known types of interpretable models: sparse linear
classifiers ([30, 8, 31]); discretization methods, such as decision trees and decision lists (e.g.,
[12, 32, 13, 23, 15]); and prototype- or case-based classifiers (e.g., nearest neighbors [10] or a supervised optimization-based method [5]). (See [14] for a review of interpretable classification.) BCM is
intended as the third model type, but uses unsupervised generative mechanisms to explain clusters,
rather than supervised approaches [16] or by focusing myopically on neighboring points [3].
3 The Bayesian Case Model
Intuitively, BCM generates each observation using the important pieces of related prototypes. The
model might generate a movie profile made of the horror movies from a quintessential horror movie
watcher, and action movies from a quintessential action moviegoer.
BCM begins with a standard discrete mixture model [18, 6] to represent the underlying structure
of the observations. It augments the standard mixture model with prototypes and subspace feature
indicators that characterize the clusters. We show in Section 4.2 that prototypes and subspace feature
indicators improve human interpretability as compared to the standard mixture model output. The
graphical model for BCM is depicted in Figure 1.
2
?, c
N
q
ps
?s
?s
S
?
?i
zij
xij
F
N
Figure 1: Graphical model for the Bayesian Case Model
We start with N observations, denoted by x = {x1 , x2 , . . . , xN }, with each xi represented as a random mixture over clusters. There are S clusters, where S is assumed to be known in advance. (This
assumption can easily be relaxed through extension to a non-parametric mixture model.) Vector ?i
are the mixture weights over these clusters for the ith observation xi , ?i ? RS+ . Each observation
has P features, and we denote the j th feature of the ith observation as xij . Each feature j of the
observation xi comes from one of the clusters, the index of the cluster for xij is denoted by zij and
the full set of cluster assignments for observation-feature pairs is denoted by z. Each zij takes on the
value of a cluster index between 1 and S. Hyperparameters q, ?, c, and ? are assumed to be fixed.
The explanatory power of BCM results from how the clusters are characterized. While a standard
mixture model assumes that each cluster take the form of a predefined parametric distribution (e.g.,
normal), BCM characterizes each cluster by a prototype, ps , and a subspace feature indicator, ?s .
Intuitively, the subspace feature indicator selects only a few features that play an important role in
identifying the cluster and prototype (hence, BCM clusters are subspace clusters). We intuitively
define these latent variables below.
Prototype, ps : The prototype ps for cluster s is defined as one observation in x that maximizes
p(ps |?s , z, x), with the probability density and ?s as defined below. Our notation for element j of
ps is psj . Since ps is a prototype, it is equal to one of the observations, so psj = xij for some i.
Note that more than one maximum may exist per cluster; in this case, one prototype is arbitrarily
chosen. Intuitively, the prototype is the ?quintessential? observation that best represents the cluster.
Subspace feature indicator ?s : Intuitively, ?s ?turns on? the features that are important for characterizing cluster s and selecting the prototype, ps . Here, ?s ? {0, 1}P is an indicator variable that
is 1 on the subset of features that maximizes p(?s |ps , z, x), with the probability for ?s as defined
below. Here, ?s is a binary vector of size P , where each element is an indicator of whether or not
feature j belongs to subspace s.
The generative process for BCM is as follows: First, we generate the subspace clusters. A subspace cluster can be fully described by three components: 1) a prototype, ps , generated by sampling
uniformly over all observations, 1 . . . N ; 2) a feature indicator vector, ?s , that indicates important
features for that subspace cluster, where each element of the feature indicator (?sj ) is generated
according to a Bernoulli distribution with hyperparameter q; and 3) the distribution of feature outcomes for each feature, ?s , for subspace s, which we now describe.
Distribution of feature outcomes ?s for cluster s: Here, ?s is a data structure wherein each ?row?
?sj is a discrete probability distribution of possible outcomes for feature j. Explicitly, ?sj is a vector
of length Vj , where Vj is the number of possible outcomes of feature j. Let us define ? as a vector
of the possible outcomes of feature j (e.g., for feature ?color?, ? = [red, blue, yellow]), where ?v
represents a particular outcome for that feature (e.g., ?v = blue). We will generate ?s so that it
mostly takes outcomes from the prototype ps for the important dimensions of the cluster. We do this
by considering the vector g, indexed by possible outcomes v, as follows:
gpsj ,?sj ,? (v) = ?(1 + c1[wsj =1 and psj =?v ] ),
where c and ? are constant hyperparameters that indicate how much we will copy the prototype in
order to generate the observations. The distribution of feature outcomes will be determined by g
through ?sj ? Dirichlet(gpsj ,?sj ,? ). To explain at an intuitive level: First, consider the irrelevant
dimensions j in subspace s, which have wsj = 0. In that case, ?sj will look like a uniform distribu3
tion over all possible outcomes for features j; the feature values for the unimportant dimensions are
generated arbitrarily according to the prior. Next, consider relevant dimensions where wsj = 1. In
this case, ?sj will generally take on a larger value ? + c for the feature value that prototype ps has on
feature j, which is called ?v . All of the other possible outcomes are taken with lower probability ?.
As a result, we will be more likely to select the outcome ?v that agrees with the prototype ps . In the
extreme case where c is very large, we can copy the cluster?s prototype directly within the cluster?s
relevant subspace and assign the rest of the feature values randomly.
An observation is then a mix of different prototypes, wherein we take the most important pieces of
each prototype. To do this, mixture weights ?i are generated according to a Dirichlet distribution,
parameterized by hyperparameter ?. From there, to select a cluster and obtain the cluster index zij
for each xij , we sample from a multinomial distribution with parameters ?i . Finally, each feature for
an observation, xij , is sampled from the feature distribution of the assigned subspace cluster (?zij ).
(Note that Latent Dirichlet Allocation (LDA) [6] also begins with a standard mixture model, though
our feature values exist in a discrete set that is not necessarily binary.) Here is the full model, with
hyperparameters c, ?, q, and ?:
?sj ? Bernoulli(q) ?s, j
ps ? Uniform(1, N ) ?s
?sj ? Dirichlet(gpsj ,?sj ,? ) ?s, j
where gpsj ,?sj ,? (v) = ?(1 + c1[wsj =1 and psj =?v ] )
?i ? Dirichlet(?) ?i
zij ? Multinomial(?i ) ?i, j
xij ? Multinomial(?zij j ) ?i, j.
Our model can be readily extended to different similarity measures, such as standard kernel methods
or domain specific similarity measures, by modifying the function g. For example, we can use the
least squares loss i.e., for fixed threshold ?, gpsj ,?sj ,? (v) = ?(1 + c1[wsj =1 and (psj ??v )2 ??] ); or,
more generally, gpsj ,?sj ,? (v) = ?(1 + c1[wsj =1 and ?(psj ,?v )??] ).
In terms of setting hyperparameters, there are natural settings for ? (all entries being 1). This
means that there are three real-valued parameters to set, which can be done through cross-validation,
another layer of hierarchy with more diffuse hyperparameters, or plain intuition. To use BCM for
classification, vector ?i is used as S features for a classifier, such as SVM.
3.1 Motivating example
This section provides an illustrative example for prototypes, subspace feature indicators and subspace clusters, using a dataset composed of a mixture of smiley faces. The feature set for a smiley
face is composed of types, shapes and colors of eyes and mouths. For the purpose of this example,
assume that the ground truth is that there are three clusters, each of which has two features that are
important for defining that cluster. In Table 1, we show the first cluster, with a subspace defined by
the color (green) and shape (square) of the face; the rest of the features are not important for defining
the cluster. For the second cluster, color (orange) and eye shape define the subspace. We generated
240 smiley faces from BCM?s prior with ? = 0.1 for all entries, and q = 0.5, ? = 1 and c = 50.
Data in assigned to cluster
LDA
Top 3 words and probabilities
1
0.26
0.23
0.24
0.27
color (
) and shape (
are important.
)
color (
) and eye (
are important.
)
eye (
) and mouth (
are important.
)
0.16
3
0.35
BCM
Subspaces
0.12
2
0.26
Prototype
0.15
Table 1: The mixture of smiley faces for LDA and BCM
4
BCM works differently to Latent Dirichlet Allocation (LDA) [6], which presents its output in a very
different form. Table 1 depicts the representation of clusters in both LDA (middle column) and BCM
(right column). This dataset is particularly simple, and we chose this comparison because the two
most important features that both LDA and BCM learn are identical for each cluster. However, LDA
does not learn prototypes, and represents information differently. To convey cluster information
using LDA (i.e., to define a topic), we must record several probability distributions ? one for each
feature. For BCM, we need only to record a prototype (e.g., the green face depicted in the top row,
right column of the figure), and state which features were important for that cluster?s subspace (e.g.,
shape and color). For this reason, BCM is more succinct than LDA with regard to what information
must be recorded in order to define the clusters. One could define a ?special? constrained version
of LDA with topics having uniform weights over a subset of features, and with ?word? distributions
centered around a particular value. This would require a similar amount of memory; however, it loses
information, with respect to the fact that BCM carries a full prototype within it for each cluster.
A major benefit of BCM over LDA is that the ?words? in each topic (the choice of feature values) are
coupled and not assumed to be independent ? correlations can be controlled depending on the choice
of parameters. The independence assumption of LDA can be very strong, and this may be crippling
for its use in many important applications. Given our example of images, one could easily generate
an image with eyes and a nose that cannot physically occur on a single person (perhaps overlapping).
BCM can also generate this image, but it would be unlikely, as the model would generally prefer to
copy the important features from a prototype.
BCM performs joint inference on prototypes, subspace feature indicators and cluster labels for observations. This encourages the inference step to achieve solutions where clusters are better represented by prototypes. We will show that this is beneficial in terms of predictive accuracy in Section 4.1. We will also show through an experiment involving human subjects that BCM?s succinct
representation is very effective for communicating the characteristics of clusters in Section 4.2.
3.2 Inference: collapsed Gibbs sampling
We use collapsed Gibbs sampling to perform inference, as this has been observed to converge
quickly, particularly in mixture models [17]. We sample ?sj , zij , and ps , where ? and ? are integrated out. Note that we can recover ? by simply counting the number of feature values assigned
to each subspace. Integrating out ? and ? results in the following expression for sampling zij :
p(zij = s|zi?j , x, p, ?, ?, ?) ?
g(psj , ?sj , ?) + n(s,?,j,xij )
?/S + n(s,i,?j,?)
,
?P
?+n
s g(psj , ?sj , ?) + n(s,?,j,?)
(1)
where n(s,i,j,v) = 1(zij = s, xij = v). In other words, if xij takes feature value v for feature j
and is assigned to cluster s, then n(s,i,j,v) = 1, or 0 otherwise. Notation n(s,?,j,v) is the number of
times that the j th feature of an observation
takes feature value v and that observation is assigned to
P
subspace cluster s (i.e., n(s,?,j,v) = i 1(zij = s, xij = v)). Notation n(s,?,j,?) means sum over
i and v. We use n(s,i,?j,v) to denote a count that does not include the feature j. The derivation is
similar to the standard collapsed Gibbs sampling for LDA mixture models [17].
Similarly, integrating out ? results in the following expression for sampling ?sj :
?
B(g(psj , 1, ?) + n(s,?,j,?) )
?
?
?q ?
B(g(psj , 1, ?))
p(?sj = b|q, psj , ?, ?, x, z, ?) ?
B(g(p
?
sj , 0, ?) + n(s,?,j,?) )
?
?1 ? q ?
B(g(psj , 0, ?))
b=1
(2)
b = 0,
where B is the Beta function and comes from integrating out ? variables, which are sampled from
Dirichlet distributions.
4 Results
In this section, we show that BCM produces prediction accuracy comparable to or better than LDA
for standard datasets. We also verify the interpretability of BCM through human subject experiments
involving a task that requires an understanding of clusters within a dataset. We show statistically
5
(a) Accuracy and standard deviation
with SVM
(b) Unsupervised accuracy
for BCM
(c) Sensitivity analysis for BCM
Figure 2: Prediction test accuracy reported for the Handwritten Digit [19] and 20 Newsgroups
datasets [22]. (a) applies SVM for both LDA and BCM, (b) presents the unsupervised accuracy
of BCM for Handwritten Digit (top) and 20 Newsgroups (bottom) and (c) depicts the sensitivity
analysis conducted for hyperparameters for Handwritten Digit dataset. Datasets were produced by
randomly sampling 10 to 70 observations of each digit for the Handwritten Digit dataset, and 100450 documents per document class for the 20 Newsgroups dataset. The Handwritten Digit pixel
values (range from 0 to 255) were rescaled into seven bins (range from 0 to 6). Each 16-by-16 pixel
picture was represented as a 1D vector of pixel values, with a length of 256. Both BCM and LDA
were randomly initialized with the same seed (one half of the labels were incorrect and randomly
mixed), The number of iterations was set at 1,000. S = 4 for 20 Newsgroups and S = 10 for
Handwritten Digit. ? = 0.01, ? = 1, c = 50, q = 0.8.
significant improvements in objective measures of task performance using prototypes produced by
BCM, compared to output of LDA. Finally, we visually illustrate that the learned prototypes and subspaces present as meaningful feedback for the characterization of important aspects of the dataset.
4.1 BCM maintains prediction accuracy.
We show that BCM output produces prediction accuracy comparable to or better than LDA, which
uses the same mixture model (Section 3) to learn the underlying structure but does not learn explanations (i.e., prototypes and subspaces). We validate this through use of two standard datasets:
Handwritten Digit [19] and 20 Newsgroups [22]. We use the implementation of LDA available from
[27], which incorporates Gibbs sampling, the same inference technique used for BCM.
Figure 2a depicts the ratio of correctly assigned cluster labels for BCM and LDA. In order to compare the prediction accuracy with LDA, the learned cluster labels are provided as features to a support vector machine (SVM) with linear kernel, as is often done in the LDA literature on clustering [6]. The improved accuracy of BCM over LDA, as depicted in the figures, is explained in part
by the ability of BCM to capture dependencies among features via prototypes, as described in Section 3. We also note that prediction accuracy when using the full 20 Newsgroups dataset acquired
by LDA (accuracy: 0.68? 0.01) matches that reported previously for this dataset when using a combined LDA and SVM approach [33]. Also, LDA accuracy for the full Handwritten Digit dataset
(accuracy: 0.76 ? 0.017) is comparable to that produced by BCM using the subsampled dataset (70
samples per digit, accuracy: 0.77 ? 0.03).
As indicated by Figure 2b, BCM achieves high unsupervised clustering accuracy as a function of
iterations. We can compute this measure for BCM because each cluster is characterized by a prototype ? a particular data point with a label in the given datasets. (Note that this is not possible for
LDA.) We set ? to prefer each ?i to be sparse, so only one prototype generates each observation,
6
Figure 3: Web-interface for the human subject experiment
and we use that prototype?s label for the observation. Sensitivity analysis in Figure 2c indicates that
the additional parameters introduced to learn prototypes and subspaces (i.e., q, ? and c) are not too
sensitive within the range of reasonable choices.
4.2 Verifying the interpretability of BCM
We verified the interpretability of BCM by performing human subject experiments that incorporated
a task requiring an understanding of clusters within a dataset. This task required each participant
to assign 16 recipes, described only by a set of required ingredients (recipe names and instructions
were withheld), to one cluster representation out of a set of four to six. (This approach is similar
to those used in prior work to measure comprehensibility [20].) We chose a recipe dataset1 for this
task because such a dataset requires clusters to be well-explained in order for subjects to be able to
perform classification, but does not require special expertise or training.
Our experiment incorporated a within-subjects design, which allowed for more powerful statistical
testing and mitigated the effects of inter-participant variability. To account for possible learning
effects, we blocked the BCM and LDA questions and balanced the assignment of participants into
the two ordering groups: Half of the subjects were presented with all eight BCM questions first,
while the other half first saw the eight LDA questions. Twenty-four participants (10 females, 14
males, average age 27 years) performed the task, answering a total of 384 questions. Subjects were
encouraged to answer the questions as quickly and accurately as possible, but were instructed to take
a 5-second break every four questions in order to mitigate the potential effects of fatigue.
Cluster representations (i.e., explanations) from LDA were presented as the set of top ingredients
for each recipe topic cluster. For BCM we presented the ingredients of the prototype without the
name of the recipe and without subspaces. The number of top ingredients shown for LDA was set as
the number of ingredients from the corresponding BCM prototype and ran Gibbs sampling for LDA
with different initializations until the ground truth clusters were visually identifiable.
Using explanations from BCM, the average classification accuracy was 85.9%, which was statistically significantly higher (c2 (1, N = 24) = 12.15, p ? 0.001) than that of LDA, (71.3%). For
both LDA and BCM, each ground truth label was manually coded by two domain experts: the first
author and one independent analyst (kappa coefficient: 1). These manually-produced ground truth
labels were identical to those that LDA and BCM predicted for each recipe. There was no statistically significant difference between BCM and LDA in the amount of time spent on each question
(t(24) = 0.89, p = 0.37); the overall average was 32 seconds per question, with 3% more time spent
on BCM than on LDA. Subjective evaluation using Likert-style questionnaires produced no statistically significant differences between reported preferences for LDA versus BCM. Interestingly, this
suggests that participants did not have insight into their superior performance using output from
BCM versus that from LDA.
1
Computer Cooking Contest: http://liris.cnrs.fr/ccc/ccc2014/
7
Prototype (Recipe names)
Ingredients ( Subspaces )
Herbs and Tomato in Pasta
basil, garlic, Italian seasoning, oil
pasta pepper salt, tomato
Generic chili recipe
Microwave brownies
beer chili powder cumin, garlic, meat, oil, onion, pepper, salt,
tomato
baking powder
sugar, butter,
Spiced-punch
chocolate chopped pecans, eggs,
flour, salt, vanilla
cinnamon stick,
lemon juice
orange juice
pineapple juice
sugar, water, whole cloves
(b) Recipe dataset
(a) Handwritten Digit dataset
Figure 4: Learned prototypes and subspaces for the Handwritten Digit and Recipe datasets.
Overall, the experiment demonstrated substantial improvement to participants? classification accuracy when using BCM compared with LDA, with no degradation to other objective or subjective
measures of task performance.
4.3 Learning subspaces
Figure 4a illustrates the learned prototypes and subspaces as a function of sampling iterations for the
Handwritten Digit dataset. For the later iterations, shown on the right of the figure, the BCM output
effectively characterizes the important aspects of the data. In particular, the subspaces learned by
BCM are pixels that define the digit for the cluster?s prototype.
Interestingly, the subspace highlights the absence of writing in certain areas. This makes sense: For
example, one can define a ?7? by showing the absence of pixels on the left of the image where the
loop of a ?9? might otherwise appear. The pixels located where there is variability among digits of
the same cluster are not part of the defining subspace for the cluster.
Because we initialized randomly, in early iterations, the subspaces tend to identify features common
to the observations that were randomly initialized to the cluster. This is because ?s assigns higher
likelihood to features with the most similar values across observations within a given cluster. For
example, most digits ?agree? (i.e., have the same zero pixel value) near the borders; thus, these are
the first areas that are refined, as shown in Figure 4a. Over iterations, the third row of Figure 4a
shows how BCM learns to separate the digits ?3? and ?5,? which tend to share many pixel values in
similar locations. Note that the sparsity of the subspaces can be customized by hyperparameter q.
Next, we show results for BCM using the Computer Cooking Contest dataset in Figure 4b. Each prototype consists of a set of ingredients for a recipe, and the subspace is a set of important ingredients
that define that cluster, highlighted in red boxes. For instance, BCM found a ?chili? cluster defined
by the subspace ?beer,? ?chili powder,? and ?tomato.? A recipe called ?Generic Chili Recipe? was
chosen as the prototype for the cluster. (Note that beer is indeed a typical ingredient in chili recipes.)
5 Conclusion
The Bayesian Case Model provides a generative framework for case-based reasoning and prototypebased modeling. Its clusters come with natural explanations; namely, a prototype (a quintessential
exemplar for the cluster) and a set of defining features for that cluster. We showed the quantitative
advantages in prediction quality and interpretability resulting from the use of BCM. Exemplar-based
modeling (nearest-neighbors, case-based reasoning) has historical roots dating back to the beginning
of artificial intelligence; this method offers a fresh perspective on this topic, and a new way of
thinking about the balance of accuracy and interpretability in predictive modeling.
8
References
[1] A. Aamodt. A knowledge-intensive, integrated approach to problem solving and sustained learning.
Knowledge Engineering and Image Processing Group. University of Trondheim, pages 27?85, 1991.
[2] A. Aamodt and E. Plaza. Case-based reasoning: Foundational issues, methodological variations, and
system approaches. AI communications, 1994.
[3] D. Baehrens, T. Schroeter, S. Harmeling, M. Kawanabe, K. Hansen, and K.R. M?uller. How to explain
individual classification decisions. JMLR, 2010.
[4] I. Bichindaritz and C. Marling. Case-based reasoning in the health sciences: What?s next? AI in medicine,
2006.
[5] J. Bien, R. Tibshirani, et al. Prototype selection for interpretable classification. AOAS, 2011.
[6] D.M. Blei, A.Y. Ng, and M.I. Jordan. Latent dirichlet allocation. JMLR, 2003.
[7] J.S. Carroll. Analyzing decision behavior: The magician?s audience. Cognitive processes in choice and
decision behavior, 1980.
[8] J. Chang, J.L. Boyd-Graber, S. Gerrish, C. Wang, and D.M. Blei. Reading tea leaves: How humans
interpret topic models. In NIPS, 2009.
[9] M.S. Cohen, J.T. Freeman, and S. Wolf. Metarecognition in time-stressed decision making: Recognizing,
critiquing, and correcting. Human Factors, 1996.
[10] T. Cover and P. Hart. Nearest neighbor pattern classification. Information Theory, 1967.
[11] P. Cunningham, D. Doyle, and J. Loughrey. An evaluation of the usefulness of case-based explanation.
In CBRRD. Springer, 2003.
[12] G. De?ath and K.E. Fabricius. Classification and regression trees: a powerful yet simple technique for
ecological data analysis. Ecology, 2000.
[13] J. Eisenstein, A. Ahmed, and E. Xing. Sparse additive generative models of text. In ICML, 2011.
[14] A. Freitas. Comprehensible classification models: a position paper. ACM SIGKDD Explorations, 2014.
[15] S. Goh and C. Rudin. Box drawings for learning with imbalanced data. In KDD, 2014.
[16] A. Graf, O. Bousquet, G. R?atsch, and B. Sch?olkopf. Prototype classification: Insights from machine
learning. Neural computation, 2009.
[17] T.L. Griffiths and M. Steyvers. Finding scientific topics. PNAS, 2004.
[18] T. Hofmann. Probabilistic latent semantic indexing. In ACM SIGIR, 1999.
[19] J.J. Hull. A database for handwritten text recognition research. TPAMI, 1994.
[20] J. Huysmans, K. Dejaeger, C. Mues, J. Vanthienen, and B. Baesens. An empirical evaluation of the
comprehensibility of decision table, tree and rule based predictive models. DSS, 2011.
[21] G.A. Klein. Do decision biases explain too much. HFES, 1989.
[22] K. Lang. Newsweeder: Learning to filter netnews. In ICML, 1995.
[23] B. Letham, C. Rudin, T. McCormick, and D. Madigan. Interpretable classifiers using rules and Bayesian
analysis. Technical report, University of Washington, 2014.
[24] H. Li and J. Sun. Ranking-order case-based reasoning for financial distress prediction. KBSI, 2008.
[25] J.W. Murdock, D.W. Aha, and L.A. Breslow. Assessing elaborated hypotheses: An interpretive case-based
reasoning approach. In ICCBR. Springer, 2003.
[26] A. Newell and H.A. Simon. Human problem solving. Prentice-Hall Englewood Cliffs, 1972.
[27] X. Phan and C. Nguyen. GibbsLDA++, AC/C++ implementation of latent dirichlet allocation using gibbs
sampling for parameter estimation and inference, 2013.
[28] S. Slade. Case-based reasoning: A research paradigm. AI magazine, 1991.
[29] F. S?rmo, J. Cassens, and A. Aamodt. Explanation in case-based reasoning?perspectives and goals. AI
Review, 2005.
[30] R. Tibshirani. Regression shrinkage and selection via the lasso. JRSS, 1996.
[31] B. Ustun and C. Rudin. Methods and models for interpretable linear classification. ArXiv, 2014.
[32] S. Williamson, C. Wang, K. Heller, and D. Blei. The IBP compound dirichlet process and its application
to focused topic modeling. 2010.
[33] J. Zhu, A. Ahmed, and E.P. Xing. MedLDA: maximum margin supervised topic models. JMLR, 2012.
9
| 5313 |@word version:1 middle:1 instruction:1 r:1 chili:6 carry:1 zij:12 selecting:1 document:2 interestingly:2 psj:12 subjective:2 freitas:1 current:1 com:1 comparing:1 discretization:1 lang:1 yet:1 must:2 readily:1 additive:1 kdd:1 shape:5 hofmann:1 designed:2 interpretable:6 alone:1 generative:6 rudin:5 discovering:1 intelligence:2 half:3 leaf:1 beginning:1 ith:2 record:2 blei:3 characterization:3 provides:3 murdock:1 location:1 preference:2 along:1 skilled:1 c2:1 beta:1 incorrect:1 consists:1 sustained:1 introduce:1 acquired:1 inter:1 indeed:1 behavior:2 nor:1 freeman:1 considering:1 provided:2 begin:2 matched:1 underlying:5 maximizes:2 notation:3 mitigated:1 what:2 finding:1 quantitative:2 every:1 mitigate:1 classifier:4 stick:1 medical:2 cooking:2 organize:1 appear:1 service:1 engineering:1 analyzing:1 chocolate:1 cliff:1 might:6 lure:1 chose:2 initialization:1 suggests:1 limited:1 range:3 statistically:6 harmeling:1 testing:1 digit:17 foundational:1 area:2 empirical:1 significantly:2 matching:1 boyd:1 confidence:1 word:5 integrating:3 griffith:1 madigan:1 get:1 naturalistic:1 cannot:1 selection:2 prentice:1 collapsed:3 writing:1 customer:2 demonstrated:2 pursues:1 regardless:1 sigir:1 focused:1 amazon:2 identifying:2 assigns:1 correcting:1 communicating:1 insight:3 rule:3 financial:1 steyvers:1 variation:1 hierarchy:1 play:2 user:3 magazine:1 us:2 hypothesis:1 element:3 recognition:2 particularly:3 kappa:1 located:1 database:1 observed:1 role:2 bottom:1 solved:1 capture:1 verifying:1 wang:2 sun:1 ordering:1 rescaled:1 ran:1 balanced:1 intuition:1 questionnaire:1 substantial:1 complexity:2 sugar:2 solving:3 predictive:3 easily:3 joint:2 differently:2 emulate:1 various:1 represented:4 derivation:1 effective:2 describe:1 artificial:2 netnews:1 outcome:12 refined:1 brownie:1 film:2 widely:1 larger:1 valued:1 drawing:1 otherwise:2 favor:1 ability:1 highlighted:1 advantage:1 tpami:1 fr:1 neighboring:1 relevant:2 loop:1 ath:1 horror:4 achieve:1 powder:3 intuitive:3 validate:1 olkopf:1 recipe:14 cluster:76 p:15 decisionmaking:1 assessing:1 produce:4 generating:1 wsj:6 help:1 depending:1 illustrate:1 spent:2 ac:1 exemplar:8 nearest:3 ibp:1 strong:1 predicted:1 come:4 indicate:1 critiquing:1 modifying:1 hull:1 filter:1 centered:1 human:14 exploration:1 pineapple:1 bin:1 require:3 assign:2 extension:1 ds:1 around:1 hall:1 ground:4 normal:1 visually:2 seed:1 cognition:1 major:1 achieves:1 early:1 purpose:1 estimation:1 label:10 maker:1 hansen:1 bridge:1 sensitive:1 agrees:1 saw:1 liris:1 tool:1 uller:1 mit:1 primed:1 rather:1 shrinkage:1 improvement:4 methodological:1 bernoulli:2 indicates:2 likelihood:1 sigkdd:1 kim:1 sense:1 inference:7 cnrs:1 typically:1 unlikely:1 explanatory:1 integrated:2 onion:1 relation:1 italian:1 cunningham:1 selects:1 pixel:8 overall:2 classification:15 among:2 issue:1 denoted:3 art:3 special:2 orange:2 constrained:1 equal:1 having:1 ng:1 cartoon:1 manually:3 sampling:11 represents:3 identical:2 look:3 unsupervised:8 encouraged:1 icml:2 thinking:1 purchase:1 report:1 fundamentally:1 few:1 baking:1 randomly:6 composed:2 simultaneously:2 resulted:1 preserve:1 individual:2 doyle:1 subsampled:1 intended:1 fire:1 ecology:1 cbr:9 englewood:1 evaluation:3 flour:1 male:1 mixture:17 extreme:1 predefined:1 microwave:1 experience:2 tree:3 indexed:1 aha:1 initialized:3 goh:1 causal:1 instance:2 column:3 modeling:4 herb:1 cover:1 pecan:1 assignment:2 deviation:1 subset:2 entry:2 uniform:3 usefulness:1 recognizing:1 successful:1 conducted:1 too:2 characterize:1 motivating:1 reported:3 dependency:1 answer:1 combined:1 person:1 density:1 fundamental:2 sensitivity:3 csail:1 probabilistic:1 quickly:2 recorded:1 cognitive:2 expert:2 style:1 li:1 account:1 potential:1 de:1 summarized:1 tactical:1 coefficient:1 explicitly:2 ranking:1 piece:2 tion:1 view:1 performed:1 break:1 root:1 later:1 characterizes:2 red:2 start:1 recover:1 participant:8 maintains:1 xing:2 simon:1 elaborated:1 square:2 accuracy:21 who:2 crippling:1 characteristic:1 identify:1 yellow:1 bayesian:11 handwritten:12 accurately:1 produced:8 advertising:1 expertise:2 explain:4 cased:1 sampled:2 dataset:20 massachusetts:2 color:7 knowledge:2 back:1 focusing:1 higher:2 supervised:3 wherein:2 improved:1 done:2 though:1 box:2 correlation:1 until:1 web:1 bichindaritz:1 overlapping:1 brings:1 lda:40 perhaps:1 indicated:1 quality:1 scientific:1 name:3 effect:3 oil:2 verify:3 requiring:2 hence:1 inspiration:1 assigned:6 semantic:1 encourages:1 illustrative:1 chaining:1 eisenstein:1 fatigue:1 complete:1 performs:2 interpreting:1 interface:1 reasoning:14 image:5 superior:1 common:1 juice:3 multinomial:3 salt:3 cohen:1 discussed:1 interpret:2 significant:5 blocked:1 cambridge:1 gibbs:6 ai:5 vanilla:1 similarly:1 pointed:1 contest:2 distress:1 similarity:3 carroll:1 imbalanced:1 showed:1 female:1 perspective:2 belongs:1 irrelevant:1 compound:1 certain:2 ecological:1 blog:1 arbitrarily:2 binary:2 preserving:1 additional:1 relaxed:1 converge:1 paradigm:1 full:5 mix:1 pnas:1 infer:1 transparency:1 technical:2 match:1 characterized:2 ahmed:2 cross:1 offer:1 hart:1 coded:1 controlled:1 prediction:9 involving:3 regression:2 patient:2 physically:1 iteration:6 represent:2 kernel:2 arxiv:1 c1:4 audience:1 background:1 want:2 chopped:1 source:1 myopically:1 sch:1 rest:2 comprehensibility:2 subject:10 tend:2 leveraging:1 incorporates:1 jordan:1 near:1 counting:1 likert:1 newsgroups:6 independence:1 zi:1 pepper:2 lasso:1 idea:1 prototype:58 intensive:1 whether:1 motivated:1 expression:2 six:1 assist:1 beenkim:1 action:3 useful:1 generally:3 unimportant:1 amount:2 augments:1 generate:6 http:1 xij:11 exist:2 problematic:1 punch:1 per:5 correctly:1 tibshirani:2 blue:2 klein:1 discrete:3 hyperparameter:3 tea:1 medlda:1 group:2 key:1 four:3 threshold:1 basil:1 neither:1 verified:1 backward:1 garlic:2 schroeter:1 sum:1 year:1 parameterized:1 powerful:4 reasonable:1 decision:14 prefer:2 comparable:4 layer:1 identifiable:1 plaza:1 lemon:1 smiley:4 occur:1 x2:1 diffuse:1 bousquet:1 generates:2 aspect:3 performing:2 according:3 gibbslda:1 jr:1 beneficial:1 across:1 making:6 intuitively:5 explained:2 indexing:1 taken:1 trondheim:1 agree:1 previously:2 remains:1 turn:1 count:1 mechanism:1 nose:1 available:1 eight:2 kawanabe:1 appropriate:1 generic:2 washington:1 shah:2 comprehensible:1 assumes:1 clustering:5 dirichlet:10 top:5 include:1 graphical:2 maintaining:1 medicine:1 especially:1 aamodt:3 objective:2 question:8 strategy:2 parametric:2 ccc:1 subspace:45 separate:1 topic:11 seven:1 reason:1 water:1 fresh:1 analyst:1 length:2 index:3 providing:1 ratio:1 balance:1 marling:1 difficult:1 mostly:1 implementation:2 guideline:1 design:1 twenty:1 perform:2 mccormick:1 observation:24 datasets:8 withheld:1 situation:4 extended:1 incorporated:3 defining:4 variability:2 communication:1 introduced:1 pair:1 required:2 namely:1 bcm:69 learned:7 nip:1 able:1 usually:1 below:3 prototyping:1 pattern:1 sparsity:4 challenge:1 bien:1 reading:1 interpretability:11 green:2 explanation:11 memory:1 mouth:2 power:3 natural:2 rely:1 indicator:11 customized:1 zhu:1 improve:3 movie:4 technology:1 eye:5 numerous:1 picture:1 coupled:1 health:1 dating:1 text:2 prior:6 understanding:4 review:2 literature:1 heller:1 graf:1 fully:1 loss:1 highlight:1 mixed:1 allocation:4 versus:2 ingredient:9 age:1 validation:1 proxy:1 beer:3 displaying:1 tomato:4 share:1 row:3 copy:3 bias:1 institute:1 neighbor:3 characterizing:1 face:6 julie:2 sparse:4 benefit:2 distributed:1 feedback:2 dimension:4 xn:1 world:1 plain:1 regard:1 dataset1:1 instructed:1 made:2 author:1 historical:1 nguyen:1 sj:20 ignore:2 meat:1 assumed:3 quintessential:5 xi:3 alternatively:1 latent:6 table:4 learn:7 pasta:2 williamson:1 complex:2 necessarily:1 domain:2 vj:2 did:1 whole:1 border:1 hyperparameters:6 profile:1 succinct:2 allowed:1 convey:1 graber:1 x1:1 interpretive:1 crafted:1 representative:2 depicts:3 fashion:3 egg:1 position:1 answering:1 jmlr:3 third:2 learns:3 load:1 specific:1 showing:1 cynthia:1 list:1 svm:5 effectively:1 illustrates:1 baehrens:1 margin:1 gap:1 phan:1 depicted:3 simply:1 likely:1 desire:1 newsweeder:1 recommendation:1 chang:1 applies:1 marketer:1 springer:2 wolf:1 truth:4 loses:1 relies:1 gerrish:1 acm:2 newell:1 goal:1 absence:2 butter:1 typical:2 determined:1 reducing:1 uniformly:1 degradation:1 called:3 total:1 meaningful:2 atsch:1 select:2 people:3 support:1 stressed:1 |
4,765 | 5,314 | Consistency of Spectral Partitioning of Uniform
Hypergraphs under Planted Partition Model
Debarghya Ghoshdastidar
Ambedkar Dukkipati
Department of Computer Science & Automation
Indian Institute of Science
Bangalore ? 560012, India
{debarghya.g,ad}@csa.iisc.ernet.in
Abstract
Spectral graph partitioning methods have received significant attention from both
practitioners and theorists in computer science. Some notable studies have been
carried out regarding the behavior of these methods for infinitely large sample size
(von Luxburg et al., 2008; Rohe et al., 2011), which provide sufficient confidence
to practitioners about the effectiveness of these methods. On the other hand, recent
developments in computer vision have led to a plethora of applications, where the
model deals with multi-way affinity relations and can be posed as uniform hypergraphs. In this paper, we view these models as random m-uniform hypergraphs
and establish the consistency of spectral algorithm in this general setting. We develop a planted partition model or stochastic blockmodel for such problems using
higher order tensors, present a spectral technique suited for the purpose and study
its large sample behavior. The analysis reveals that the algorithm is consistent for
m-uniform hypergraphs for larger values of m, and also the rate of convergence
improves for increasing m. Our result provides the first theoretical evidence that
establishes the importance of m-way affinities.
1
Introduction
The central theme in approaches like kernel machines [1] and spectral clustering [2, 3] is the use
of symmetric matrices that encode certain similarity relations between pairs of data instances. This
allows one to use the tools of matrix theory to design efficient algorithms and provide theoretical
analysis for the same. Spectral graph theory [4] provides classic examples of this methodology,
where various hard combinatorial problems pertaining to graphs are relaxed to problems of matrix
theory. In this work, we focus on spectral partitioning, where the aim is to group the nodes of a graph
into disjoint sets using the eigenvectors of the adjacency matrix or the Laplacian operator. A statistical framework for this partitioning problem is the planted partition or stochastic blockmodel [5].
Here, one assumes the existence of an unknown map that partitions the nodes of a random graph,
and the probability of occurrence of any edge follows the partition rule. In a recent work, Rohe et
al. [6] studied normalized spectral clustering under the stochastic blockmodel and proved that, for
this method, the fractional number of misclustered nodes goes to zero as the sample size grows.
However, recent developments in signal processing, computer vision and statistical modeling have
posed numerous problems, where one is interested in computing multi-way similarity functions that
compute similarity among more than two data points. A few applications are listed below.
Example 1. In geometric grouping, one is required to cluster points sampled from a number of
geometric objects or manifolds [7]. Usually, these objects are highly overlapping, and one cannot
use standard distance based pairwise affinities to retrieve the desired clusters. Hence, one needs to
construct multi-point similarities based on the geometric structure. A special case is the subspace
clustering problem encountered in motion segmentation [7], face clustering [8] etc.
1
Example 2. The problem of point-set matching [9] underlies several problems in computer vision
including image registration, object recognition, feature tracking etc. The problem is often formulated as finding a strongly connected component in a uniform hypergraph [9, 10], where the strongly
connected component represents the correct matching. This formulation has the flavor of the standard problem of detecting cliques in random graphs.
Both of the above problems are variants of the classic hypergraph partitioning problem, that arose in
the VLSI community [11] in 1980s, and has been an active area of research till date [12]. Spectral approaches for hypergraph partitioning also exist in the literature [13, 14, 15], and various definitions of
the hypergraph Laplacian matrix has been proposed based on different criteria. Recent studies [16]
suggest an alternative representation of uniform hypergraphs in terms of the ?affinity tensor?. Tensors have been popular in machine learning and signal processing for a considerable time (see [17]),
and have even found use in graph partitioning and detecting planted partitions [17, 18]. But their
role in hypergraph partitioning have been mostly overlooked in the literature. Recently, techniques
have emerged in computer vision that use such affinity tensors in hypergraph partitioning [8, 9].
This paper provides the first consistency result on uniform hypergraph partitioning by analyzing the
spectral decomposition of the affinity tensor. The main contributions of this work are the following.
(1) We propose a planted partition model for random uniform hypergraphs similar to that of
graphs [5]. We show that the above examples are special cases of the proposed partition model.
(2) We present a spectral technique to extract the underlying partitions of the model. This method
relies on a spectral decomposition of tensors [19] that can be computed in polynomial time, and
hence, it is computationally efficient than the tensorial approaches in [10, 8].
(3) We analyze the proposed approach and provide almost sure bounds on the number of misclustered nodes. Our analysis reveals that the presented method is consistent almost surely in the grouping problem and for detection of a strongly connected component, whenever one uses m-way affinities for any m ? 3 and m ? 4, respectively. The derived rate of convergence also shows that the
use of higher order affinities lead to a faster decay in the number of misclustered nodes.
(4) We numerically demonstrate the performance of the approach on benchmark datasets.
2
Planted partitions in random uniform hypergraphs
We describe the planted partition model for an undirected unweighted graph. Let ? : {1, . . . , n} ?
{1, . . . , k} be an (unknown) partition of n nodes into k disjoint groups, i.e., ?i = ?(i) denotes the
partition in which node-i belongs. We also define an assignment matrix Zn ? {0, 1}n?k such that
(Zn )ij = 1 if j = ?i , and 0 otherwise. For some unknown symmetric matrix B ? [0, 1]k?k , the
random graph on the n nodes contains the edge (i, j) with probability B?i ?j . Let the symmetric
matrix An ? {0, 1}n?n be a realization of the affinity matrix of the random graph on n nodes. The
aim is to identify Zn given the matrix An . In some cases, one also needs to estimate the entries in
B. One can hope to achieve this goal for the following reason: If An ? Rn?n contains the expected
values of the entries in An conditioned on B and ?, then one can write An as An = Zn BZnT [6].
Thus, if one can find An , then this relation can be used to find Zn .
We generalize the partition model to uniform hypergraphs. A hypergraph is a structure on n nodes
with multi-way connections or hyperedges. Formally, each hyperedge in an undirected unweighted
hypergraph is a collection of an arbitrary number of vertices. A special case is that of m-uniform
hypergraph, where each hyperedge contains exactly m nodes. One can note that a graph is a 2uniform hypergraph. An often cited example of uniform hypergraph is as follows [10]. Let the
nodes be representative of points in an Euclidean space, where a hyperedge exists if the points
are collinear. For m = 2, we obtain a complete graph that does not convey enough information
about the nodes. However, for m = 3, the constructed hypergraph is a union of several connected
components, each component representing a set of collinear points. The affinity relations of an muniform hypergraph can be represented in the form of an mth -order tensor An ? {0, 1}n?n?...?n ,
which we call an affinity tensor. The entry (An )i1 ...im = 1 if there exists a hyperedge on nodes
i1 , . . . , im . One can observe that the tensor is symmetric, i.e., invariant under any permutation of
indices. In some works [16], the tensor is scaled by a factor of 1/(m ? 1)! for certain reasons.
Let ? and Zn be as defined above, and B ? [0, 1]k?...?k be an mth -order k-dimensional symmetric
tensor. The random m-uniform hypergraph on the n nodes is constructed such that a hyperedge
occurs on nodes i1 , . . . , im with probability B?i1 ...?im . If An is a random affinity tensor of the
2
hypergraph, our aim is to find Zn or ? from An . Notice that if An ? Rn?...?n contains the
expected values of the entries in An , then one can write the entries in An as
(An )i1 ...im = B?i1 ...?im =
k
X
Bj1 ...jm (Zn )i1 j1 . . . (Zn )im jm .
(1)
j1 ,...,jm =1
The subscript n in the above terms emphasizes their dependence on the number of nodes. We now
describe how two standard applications in computer vision can be formulated as the problem of
detecting planted partitions in uniform hypergraphs.
2.1
Subspace clustering problem
In motion segmentation [7, 20] or illumination invariant face clustering [8], the data belong to a
high dimensional space. However, the instances belonging to each cluster approximately span a
low-dimensional subspace (usually, of dimension 3 or 4). Here, one needs to check whether m
points approximate such a subspace, where this information is useful only when m is larger than the
dimension of the underlying subspace of interest. The model can be represented as an m-uniform
hypergraph, where a hyperedge occurs on m nodes whenever they approximately span a subspace.
The partition model for this problem is similar to the standard four parameter blockmodel [6]. The
number of partitions is k, and each partition contains s nodes, i.e., n = ks. There exists probabilities
p ? (0, 1] and q ? [0, p) such that any set of m vectors span a subspace with probability p if all m
vectors belong to the same group, and with probability q if they come from different groups. Thus,
the tensor B has the form Bi...i = p for all i = 1, . . . , k, and Bi1 ...im = q for all the other entries.
2.2
Point set matching problem
We consider a simplified version of the matching problem [10], where one is given two sets of
points of interest, each of size s. In practice, these points may come from two different images
of the same object or scene, and the goal is to match the corresponding points. One can see that
there are s2 candidate matches. However, if one considers m correct matches then certain properties
are preserved. For instance, let i1 , . . . , im be some points from the first image, and i01 , . . . , i0m be
the corresponding points in the second image, then the angles or ratio of areas of triangles formed
among these points are more or less preserved [9]. Thus, the set of matches (i1 , i01 ), . . . , (im , i0m )
have a certain connection, which is usually not present if the matches are not exact.
The above model is an m-uniform hypergraph on n = s2 nodes, each node representing a candidate match, and a hyperedge is formed if properties (like preservation
of angles) is satisfied by m
?
candidate matches. Here, one can see that there are only s = n correct matches, which have a
large number of hyperedges among them, whereas very few hyperedges
may
? be present for other
?
combinations. Thus, the partition model has two groups of size n and (n ? n), respectively. For
p, q ? [0, 1], p q, p denotes the probability of a hyperedge among m correct matches and for any
other m candidates, there is a hyperedge with probability q. Thus, if the first partition is the strongly
connected component, then we have B ? [0, 1]2?...?2 with B1...1 = p and Bi1 ...im = q otherwise.
3
Spectral partitioning algorithm and its consistency
Before presenting the algorithm, we provide some background on spectral decomposition of tensors.
In the related literature, one can find a number of significantly different characterizations of the
spectral properties of tensors. While the work in [16] builds on a variational characterization, De
Lathauwer et al. [19] provide an explicit decomposition of a tensor in the spirit of the singular
value decomposition of matrices. The second line of study is more appropriate for our work since
our analysis significantly relies on the use of Davis-Kahan perturbation theorem [21] that uses an
explicit decomposition, and has been often used to analyze spectral clustering [2, 6].
The work in [19] provides a way of expressing any mth -order n-dimensional symmetric tensor,
An , as a mode-k product [19] of a certain core tensor with m orthonormal matrices, where each
bn ? {0, 1}n?nm?1 ,
orthonormal matrix is formed from the orthonormal left singular vectors of A
3
whose entries, for all i = 1, . . . , n and j = 1, . . . , nm?1 , are defined as
bn )ij = (An )i i ...i ,
(A
1 2
m
if i = i1 and j = 1 +
m
X
(il ? 1)nl?2 .
(2)
l=2
bn , often called the mode-1 flattened matrix, forms a key component of the
The above matrix A
bn contain inforpartitioning algorithm. Later, we show that the leading k left singular vectors of A
mation about the true partitions in the hypergraph. It is easier to work with the symmetric matrix
bn A
bTn ? Rn?n , whose eigenvectors correspond to the left singular vectors of A
bn . The
Wn = A
spectral partitioning algorithm is presented in Algorithm 1, which is quite similar to the normalized
spectral clustering [2]. Such a tensor based approach was first studied in [7] for geometric grouping. Subsequent improvements of the algorithm were proposed in [22, 20]. However, we deviate
from these methods as we do not normalize the rows of the eigenvector matrix. The method in [9]
also uses the largest eigenvector of the flattened matrix for the point set matching problem. This is
computed via tensor power iterations. To keep the analysis simple, we do not use such iterations.
The complexity of Algorithm 1 is O(nm+1 ), which can be significantly improved using sampling
techniques as in [7, 9, 20]. The matrix Dn is used for normalization as in spectral clustering.
Algorithm 1 Spectral partitioning of m-uniform hypergraph
bn using (2).
1. From the mth -order affinity tensor An , construct A
bn A
bT , and Dn ? Rn?n be diagonal with (Dn )ii = Pn (Wn )ij .
2. Let Wn = A
n
j=1
?1/2
3. Set Ln = Dn
?1/2
Wn Dn
.
4. Compute leading k orthonormal eigenvectors of Ln , denoted by matrix Xn ? Rn?k .
5. Cluster the rows of Xn into k clusters using k-means clustering.
6. Assign node-i of hypergraph to j th partition if ith row of Xn is grouped in j th cluster.
An alternative technique of using eigenvectors of Laplacian matrix is often preferred in graph parbn , in
titioning [3], and has been extended to hypergraphs [13, 15]. Unlike the flattened matrix, A
Algorithm 1, such Laplacians do not preserve the spectral properties of a higher-order structure such
as the affinity tensor that accurately represents the affinities of the hypergraph. Hence, we avoid the
use of hypergraph Laplacian.
3.1
Consistency of above algorithm
We now comment on the error incurred by Algorithm 1. For this, let Mn be the set of nodes that
are incorrectly clustered by Algorithm 1. It is tricky to formalize the definition of Mn in clustering
problems. We follow the definition of Mn given in [6] that requires some details of the analysis
and hence, a formal definition is postponed till Section 4. In addition, we need the following terms.
b ?
The analysis depends on the tensor B ? [0, 1]k?...?k of the underlying random model. Let B
k?k
k?km?1
be the flattening of tensor B using (2). We also define a matrix Cn ? R
as
[0, 1]
b nT Zn )?(m?1) B
b T (ZnT Zn )1/2 ,
Cn = (ZnT Zn )1/2 B(Z
(3)
where (ZnT Zn )?(m?1) is the (m ? 1)-times Kronecker product of ZnT Zn with itself. Use of such
Kronecker product is quite common in tensor decompositions (see [19]). Observe that the positive
semi-definite matrix Cn contains information regarding the connectivity of clusters (stored in B)
and the cluster sizes (diagonal entries of ZnT Zn ). Let ?k (Cn ) be the smallest eigenvalue of Cn ,
which is non-negative. In addition, define Dn ? Rn?n as the expectation of the diagonal matrix
Dn . One can see that (Dn )ii ? nm for all i = 1, . . . , n. Let Dn and Dn be the smallest and largest
values in Dn . Also, let S n and S n be the sizes of the smallest and largest partitions, respectively.
We have the following bound on the number of misclustered nodes.
Theorem 1. If there exists N such that for all n > N ,
r
?k (Cn ) 2nm?1
2
m
>0
and
Dn ? n (m ? 1)!
?n :=
?
,
Dn
log n
Dn
4
m?1
and if (log n)3/2 = o ?n n 2 , then the number of misclustered nodes
|Mn | = O
S n (log n)2 nm+1
?n2 D2n
almost surely.
The above result is too general to provide conclusive remarks about consistency of the algorithm.
Hence, we focus on two examples, precisely the ones described in Sections 2.1 and 2.2. However,
without loss of generality, we assume here that q > 0 since otherwise, the problem of detecting the
partitions is trivial (at least for reasonably large n) as we can construct the partitions only based on
the presence of hyperedges. The following results are proved in the appendix. The proofs mainly
depend on computation of ?k (Cn ), which can be derived for the first example, while for the second,
it is enough to work with a lower bound of ?k (Cn ). Further, in the first example, we make the result
general by allowing the number of clusters, k, to grow with n under certain conditions.
Corollary 2. Consider the setting
1 of subspace
clustering described in Section 2.1. If the number
?1
of clusters k satisfy k = O n 2m (log n)
, then the conditions in Theorem 1 are satisfied and
2m?1
2
3?2m
k
(log n)
(log n)
|Mn | = O
=
O
almost surely. Hence, for m > 2, |Mn | ? 0
1
nm?2
nm?3+ 2m
|Mn |
a.s. as n ? ?, i.e., the algorithm is consistent. For m = 2, we can only conclude
? 0 a.s.
n
From the above result, it is evident that the rate of convergence improves as m increases, indicating
that, ignoring practical considerations, one should prefer the use of higher order affinities. However, the condition of number of clusters becomes more strict in such cases. We note here that our
result and conditions are quite similar to those given in [6] for the case of four-parameter blockmodel. Thus, Algorithm 1 is comparable to spectral clustering [6]. Next, we consider the setting of
Section 2.2.
Corollary 3. For the problem of point set matching
described
in Section 2.2, the conditions in
(log n)2
Theorem 1 are satisfied for m ? 3 and |Mn | = O
a.s. Hence, for m > 3, |Mn | ? 0
nm?3
|Mn |
? 0 a.s.
a.s. as n ? ?, i.e., the algorithm is consistent. For m = 3, we can only conclude
n
The above result shows, theoretically, why higher order matching provides high accuracy in practice [9]. It also suggests that increase in the order of tensor will lead to a better convergence rate.
We note that the following result does not
? hold for graphs (m = 2). In Corollary 3, we used the fact
that the smaller partition is of size s = n. The result can be made more general in terms of s, i.e.,
for m > 4, if s ? 3p
q 3 eventually, then Algorithm 1 is consistent.
Before providing the detailed analysis (proof of Theorem 1), we briefly comment on the model
considered here. In Section 2, we have followed the lines of [6] to define the model with An =
Zn BZnT . However, this would mean that the diagonal entries in An are non-negative, and hence,
there is a non-zero probability of formation of self loops that is not common in practice. The same
issue exists for hypergraphs. To avoid this, one can add a correction term to An so that the entries
with repeated indices become zero. Under this correction, conditions in Theorem 1 should not
change significantly. This is easy to verify for graphs, but it is not straightforward for hypergraphs.
4
Analysis of partitioning algorithm
In this section, we prove Theorem 1. The result follows from a series of lemmas. The proof requires
cn be the flattening of the tensor An defined in (1). Then we can
defining certain terms. Let A
T ?(m?1)
c
b
write An = Zn B(Zn )
, where (ZnT )?(m?1) is (m ? 1)-times Kronecker product of ZnT
with itself. Along with the definitions in Section 3, let Wn ? Rn?n be the expectation of Wn , and
T
?1/2
?1/2
cn A
cn + Pn , where Pn is
Ln = Dn Wn Dn . One can see that Wn can be written as Wn = A
cn . The proof contains the following steps:
a diagonal matrix defined in terms of the entries in A
(1) For any fixed n, we show that if ?n > 0 (stated in Theorem 1), the leading k orthonormal
5
eigenvectors of Ln has k distinct rows, where each row is a representative of a partition.
(2) Since, Ln is not the expectation of Ln , we derive a bound on the Frobeniusqnorm of their
difference. The bound holds almost surely for all n if eventually Dn ? nm (m ? 1)! log2 n .
(3) We use a version of Davis-Kahan sin-? theorem given in [6] that
surely bounds the
almost
m?1
difference in the leading eigenvectors of Ln and Ln if (log n)3/2 = o ?n n 2 .
(4) Finally, we rely on [6, Lemma 3.2], which holds in our case, to define the set of misclustered
nodes Mn , and its size is bounded almost surely using the previously derived bounds.
We now present the statements for the above results. The proofs can be found in the appendix.
Lemma 4. Fix n and let ?n be as defined in Theorem 1. If ?n > 0, then there exists ?n ? Rk?k such
that the columns of Zn ?n are the leading k orthonormal eigenvectors of Ln . Moreover, for nodes i
and j, ?i = ?j if and only if the ith and j th rows of Zn ?n are identical.
Thus, clustering the rows of Zn ?n into k clusters will provide the true partitions, and the cluster
centers will precisely be these rows. The condition ?n > 0 is required to ensure that the eigenvalues
corresponding to the columns of Zn ?n are strictly greater than other eigenvalues. The requirement
of a positive eigen-gap is essential for analysis of any spectral partitioning method [2, 23]. Next, we
focus on deriving the upper bound for kLn ? Ln kF .
q
Lemma 5. If there exists N such that Dn ? nm (m ? 1)! log2 n for all n > N , then
kLn ? Ln kF ?
4n
m+1
2
log n
,
Dn
(4)
almost surely.
The condition in the above result implies that each vertex is reasonably connected to other vertices
of the hypergraph, i.e., there are no outliers. It is easy to satisfy this condition in the stated examples
as Dn ? q 2 nm and
hence, it holds for all q > 0. Under the condition, one can also see that the
bound in (4) is O
(log n)3/2
n
m?1
2
and hence goes to zero as n increases. Note that in Lemma 4, ?n > 0
need not hold for all n, but if it holds eventually, then we can choose N such that the conditions in
Lemmas 4 and 5 both hold for all n > N . Under such a case, we use the Davis-Kahan perturbation
theorem [21] as stated in [6, Theorem 2.1] to claim the following.
Lemma
6. Let Xn ? Rn?k contain the leading k orthonormal eigenvectorsq
of Ln . If (log n)3/2 =
o ?n n
m?1
2
2
log n
and there exists N such that ?n > 0 and Dn ? nm (m ? 1)!
then there exists an orthonormal (rotation) matrix On ? R
kXn ? Zn ?n On kF ?
16n
k?k
for all n > N ,
such that
m+1
2
log n
,
?n D n
(5)
almost surely.
m?1
is crucial as it ensures that the difference in eigenvalues
The condition (log n)3/2 = o ?n n 2
of Ln and Ln decays much faster than the eigen-gap in Ln . This condition requires the eigen-gap
(lower bounded by ?n ) to decay at a relatively slow rate, and is necessary for using [6, Theorem 2.1].
The bound (5) only says that rows of Xn converges to some rotation of the rows of Zn ?n . However,
this is not an issue since the k-means algorithm is expected to perform well as long as the rows of
Xn corresponding to each partition are tightly clustered, and the k clusters are well-separated. Now,
let z1 , . . . , zn be the rows of Zn , and let ci be the center of the cluster in which ith row of Xn is
grouped for each i ? {1, . . . , n}. We use a key result from [6] that is applicable in our setting.
Lemma 7. [6, Lemma 3.2] For the matrix On from Lemma 6, if kci ? zi ?n On k2 < ? 1 , then
2S n
kci ? zi ?n On k2 < kci ? zj ?n On k2 for all zj 6= zi .
This result hints that one may use the definition of correct clustering as follows. Node-i is correctly
clustered if its center ci is closer to zi ?n On than the rows corresponding to other partitions. A sufficient condition to satisfy this definition is kci ? zi ?n On k2 < ? 1 . Hence, the set of misclustered
2S n
nodes is defined as [6]
(
Mn =
1
i ? {1, . . . , n} : kci ? zi ?n On k2 ? p
2S n
6
)
.
(6)
It is easy to see that if Mn is empty, i.e., all nodes satisfy the condition kci ? zi ?n On k2 < ? 1
2S n
,
then the clustering leads to true partitions, and does not incur any error. Hence, for statements, where
|Mn | is small (at least compared to n), one can always use such a definition for misclustered nodes.
The next result provides a simple bound on |Mn |, that immediately leads to Theorem 1.
Lemma 8. If the k-means algorithm achieves its global optimum, then the set Mn satisfies
|Mn | ? 8S n kXn ? Zn ?n On k2F .
(7)
In practice, k-means algorithm tries to find a local minimum, and hence, one should run this step
with multiple initializations to achieve a global minimum. However, empirically we found that
good performance is achieved even if we use a single run of k-means. From above lemma, it is
straightforward to arrive at Theorem 1 by using the bound in Lemma 6.
5
5.1
Experiments
Validation of Corollaries 2 and 3
We demonstrate the claims of Corollaries 2 and 3, where we stated that for higher order tensors, the
number of misclustered nodes decays to zero at a faster rate. We run Algorithm 1 on both the models
of subspace clustering and point-set matching, varying the number of nodes n, the results for each n
being averaged over 10 trials. For the clustering model (Section 2.1), we choose p = 0.6, q = 0.4,
and consider two cases of k = 2 and 3 cluster problems. Figure 1 (top row) shows that in this model,
the number of errors eventually decreases for all m, even m = 2. This observation is similar to the
one in [6]. However, the decrease is much faster for m = 3, where accurate partitioning is often
observed for n ? 100. We also observe that error rises for larger k, thus validating the dependence
of the bound on k. A similar inference can be drawn from Figure 1 (second row) for the matching
?
problem (Section 2.2), where we use p = 0.9, q = 0.1 and the number of correct matches as n.
5.2
Motion Segmentation on Hopkins 155 dataset
We now turn to practical applications, and test the performance of Algorithm 1 in motion segmentation. We perform the experiments on the Hopkins 155 dataset [24], which contains 120 videos with
2 independent affine motions. Figure 1 (third row) shows two cases, where Algorithm 1 correctly
clusters the trajectories into their true groups. We used 4th -order tensors in the approach, where the
bn is tackled by using only 500 uniformly sampled columns of A
bn for comlarge dimensionality of A
puting Wn . We also compare the performance of Algorithm 1, averaged over 20 runs, with some
standard approaches. The results for other methods have been taken from [20]. We observe that Algorithm 1 performs reasonably well, while the best performance is obtained using Sparse Grassmann
Clustering (SGC) [20], which is expected as SGC is an iterative improvement of Algorithm 1.
5.3
Matching point sets from the Mpeg-7 shape database
We now consider a matching problem using points sampled from images in Mpeg-7 database [25].
This problem has been considered in [10]. We use 70 random images, one from each shape class.
Ten points were sampled from the boundary of each shape, which formed one point set. The other
set of points was generated by adding Gaussian noise of variance ? 2 to the original points and then
using a random affine transformation on the points. In Figure 1 (last row), we compare performance
of Algorithm 1 with the methods in [9, 10], which have been shown to outperform other methods.
We use 4-way similarities based on ratio of areas of two triangles. We show the variation in the
number of correctly detected matches and the F1-score for all methods as ? increases from 0 to
0.2. The results show that Algorithm 1 is quite robust compared to [10] in detecting true matches.
However, Algorithm 1 does not use additional post-processing as in [9], and hence, allows high
number of false positives that reduces F1-score, whereas [9, 10] show similar trends in both plots.
6
Concluding remarks
In this paper, we presented a planted partition model for unweighted undirected uniform hypergraphs. We devised a spectral approach (Algorithm 1) for detecting the partitions from the affinity
7
The plots show variation in the
number (left) and fraction (right)
of misclustered nodes as n increases in k = 2 and 3 cluster problems for 2 and 3-uniform
hypergraphs. Black lines are for
m = 2 and red for m = 3. Solid
lines for k = 2, and dashed lines
for k = 3.
The plots show variation in number (left) and fraction (right) of
incorrect matches as n increases
in matching problem for 2 and
3-uniform hypergraphs. Black
lines are for m = 2 and red for
m = 3.
Percentage error in clustering
LSA
4.23 %
SCC
2.89 %
LRR-H
2.13 %
LRSC
3.69 %
SSC
1.52 %
SGC
1.03 %
Algorithm 1
1.83 %
Figure 1: First row: Number of misclustered nodes in clustering problem as n increases.
Second row: Number of misclustered nodes in matching problem as n increases.
Third row: Grouping two affine motions with Algorithm 1 (left), and performance comparison of
Algorithm 1 with other methods (right).
Fourth row: Variation in number of correct matches detected (left) and F1-score (middle) as noise
level, ? increases. (right) A pair of images where Algorithm 1 correctly matches all sampled points.
tensor of the corresponding random hypergraph. The above model is appropriate for a number of
problems in computer vision including motion segmentation, illumination-invariant face clustering,
point-set matching, feature tracking etc. We analyzed the approach to provide an almost sure upper
bound on the number of misclustered nodes (c.f. Theorem 1). Using this bound, we conclude that
for the problems of subspace clustering and point-set matching, Algorithm 1 is consistent for m ? 3
and m ? 4, respectively. To the best of our knowledge, this is the first theoretical study of the above
problems in a probabilistic setting, and also the first theoretical evidence that shows importance of
m-way affinities.
Acknowledgement
D. Ghoshdastidar is supported by Google Ph.D. Fellowship in Statistical Learning Theory.
8
References
[1] B. Scholk?opf and A. J. Smola. Learning with Kernels. MIT Press, 2002.
[2] A. Ng, M. Jordan, and Y. Weiss. On spectral clustering: analysis and an algorithm. In Advances
in Neural Information Processing Systems, pages 849?856, 2002.
[3] U. von Luxburg. A tutorial on spectral clustering. Statistics and computing, 17(4):395?416,
2007.
[4] F. R. K. Chung. Spectral graph theory, volume 92. American Mathematical Soc., 1997.
[5] F. McSherry. Spectral partitioning of random graphs. In IEEE Symposium on Foundations of
Computer Science, pages 529?537, 2001.
[6] K. Rohe, S. Chatterjee, and B. Yu. Spectral clustering and the high-dimensional stochastic
blockmodel. Annals of Statistics, 39(4):1878?1915, 2011.
[7] V. M. Govindu. A tensor decomposition for geometric grouping and segmentation. In IEEE
Conference on Computer Vision and Pattern Recognition, pages 1150?1157, 2005.
[8] S. Rota Bulo and M. Pelillo. A game-theoretic approach to hypergraph clustering. IEEE
Transactions on Pattern Analysis and Machine Intelligence, 35(6):1312?1327, 2013.
[9] M. Chertok and Y. Keller. Efficient high order matching. IEEE Trans. on Pattern Analysis and
Machine Intelligence, 32(12):2205?2215, 2010.
[10] H. Liu, L. J. Latecki, and S. Yan. Robust clustering as ensembles of affinity relations. In
Advances in Neural Information Processing Systems, pages 1414?1422, 2010.
[11] G. Schweikert and B. W. Kernighan. A proper model for the partitioning of electrical circuits.
In Proceedings of 9th Design Automation Workshop, pages 57?62, Dallas, 1979.
[12] N. Selvakkumaran and G. Karypis. Multi-objective hypergraph partitioning algorithms for cut
and maximum subdomain degree minimization. IEEE Trans. on CAD, 25(3):504?517, 2006.
[13] M. Bolla. Spectra, euclidean representations and clusterings of hypergraphs. Discrete Mathematics, 117(1):19?39, 1993.
[14] S. Agarwal, K. Branson, and S. Belongie. Higher order learning with graphs. In Proceedings
of the International Conference on Machine Learning, pages 17?24, 2006.
[15] J. A. Rodriguez. Laplacian eigenvalues and partition problems in hypergraphs. Applied Mathematics Letters, 22(6):916?921, 2009.
[16] J. Cooper and A. Dutle. Spectra of uniform hypergraphs. Linear Algebra and its Applications,
436(9):3268?3292, 2012.
[17] A. Anandkumar, R. Ge, D. Hsu, and S.M. Kakade. A tensor spectral approach to learning
mixed membership community models. In Conference on Learning Theory (Expanded version
at arXiv:1210.7559v3), 2013.
[18] A. Frieze and R. Kannan. A new approach to the planted clique problem. In IARCS Annual Conference on Foundations of Software Technology and Theoretical Computer Science,
volume 2, pages 187?198, 2008.
[19] L. De Lathauwer, B. De Moor, and J. Vandewalle. A multilinear singular value decomposition.
SIAM Journal on Matrix Analysis and Appl., 21(4):1253?1278, 2000.
[20] S. Jain and V. M. Govindu. Efficient higher-order clustering on the grassmann manifold. In
IEEE International Conference on Computer Vision, 2013.
[21] G. W. Stewart and J. Sun. Matrix Perturbation Theory. Academic Press, 1990.
[22] G. Chen and G. Lerman. Foundations of a multi-way spectral clustering framework for hybrid
linear modeling. Foundations of Computational Mathematics, 9:517?558, 2009.
[23] U. von Luxburg, M. Belkin, and O. Bousquet. Consistency of spectral clustering. Annals of
Statistics, 36(2):555?586, 2008.
[24] R. Tron and R. Vidal. A benchmark for the comparison of 3-D motion segmentation algorithms. In IEEE Conference on Computer Vision and Pattern Recognition, 2007.
[25] L. J. Latecki, R. Lakamper, and T. Eckhardt. Shape descriptors for non-rigid shapes with
a single closed contour. In IEEE Conference on Computer Vision and Pattern Recognition,
volume 1, pages 424?429, 2000.
9
| 5314 |@word trial:1 briefly:1 middle:1 version:3 polynomial:1 tensorial:1 km:1 bn:10 decomposition:9 solid:1 liu:1 contains:8 series:1 score:3 nt:1 cad:1 written:1 subsequent:1 partition:34 j1:2 shape:5 plot:3 intelligence:2 ith:3 core:1 provides:6 detecting:6 node:37 characterization:2 mathematical:1 lathauwer:2 constructed:2 dn:21 become:1 along:1 symposium:1 incorrect:1 prove:1 theoretically:1 pairwise:1 expected:4 behavior:2 multi:6 jm:3 increasing:1 becomes:1 iisc:1 latecki:2 underlying:3 bounded:2 moreover:1 circuit:1 bj1:1 eigenvector:2 finding:1 transformation:1 exactly:1 scaled:1 k2:6 tricky:1 partitioning:19 lsa:1 before:2 positive:3 puting:1 local:1 dallas:1 analyzing:1 subscript:1 approximately:2 black:2 initialization:1 studied:2 k:1 suggests:1 appl:1 branson:1 karypis:1 bi:1 lrr:1 averaged:2 practical:2 union:1 practice:4 definite:1 area:3 yan:1 significantly:4 matching:16 confidence:1 rota:1 suggest:1 cannot:1 operator:1 map:1 center:3 go:2 attention:1 straightforward:2 keller:1 immediately:1 rule:1 orthonormal:8 deriving:1 retrieve:1 classic:2 variation:4 annals:2 exact:1 us:3 trend:1 recognition:4 cut:1 database:2 observed:1 role:1 electrical:1 ensures:1 connected:6 sun:1 decrease:2 complexity:1 hypergraph:27 dukkipati:1 depend:1 algebra:1 incur:1 triangle:2 various:2 represented:2 separated:1 distinct:1 jain:1 describe:2 pertaining:1 detected:2 formation:1 whose:2 emerged:1 posed:2 larger:3 quite:4 say:1 otherwise:3 statistic:3 kahan:3 itself:2 eigenvalue:5 propose:1 product:4 loop:1 realization:1 date:1 till:2 achieve:2 normalize:1 convergence:4 cluster:18 requirement:1 plethora:1 misclustered:13 empty:1 optimum:1 mpeg:2 converges:1 object:4 derive:1 develop:1 ij:3 received:1 pelillo:1 soc:1 come:2 implies:1 correct:7 stochastic:4 adjacency:1 assign:1 fix:1 clustered:3 f1:3 bi1:2 multilinear:1 im:11 strictly:1 correction:2 hold:7 considered:2 claim:2 achieves:1 smallest:3 purpose:1 applicable:1 combinatorial:1 largest:3 grouped:2 establishes:1 tool:1 moor:1 hope:1 minimization:1 mit:1 gaussian:1 always:1 aim:3 mation:1 arose:1 pn:3 avoid:2 varying:1 corollary:5 encode:1 derived:3 focus:3 improvement:2 check:1 mainly:1 blockmodel:6 inference:1 rigid:1 membership:1 bt:1 mth:4 relation:5 vlsi:1 interested:1 i1:10 issue:2 among:4 denoted:1 development:2 special:3 ernet:1 construct:3 ng:1 sampling:1 identical:1 represents:2 yu:1 k2f:1 govindu:2 bangalore:1 few:2 hint:1 belkin:1 frieze:1 preserve:1 tightly:1 detection:1 interest:2 highly:1 analyzed:1 nl:1 mcsherry:1 accurate:1 edge:2 closer:1 necessary:1 euclidean:2 desired:1 theoretical:5 instance:3 column:3 modeling:2 d2n:1 ghoshdastidar:2 stewart:1 zn:27 assignment:1 bolla:1 vertex:3 entry:11 uniform:22 vandewalle:1 too:1 stored:1 cited:1 international:2 siam:1 probabilistic:1 hopkins:2 connectivity:1 von:3 central:1 satisfied:3 nm:13 choose:2 ssc:1 american:1 chung:1 leading:6 de:3 automation:2 satisfy:4 notable:1 ad:1 depends:1 later:1 view:1 try:1 closed:1 analyze:2 red:2 contribution:1 formed:4 il:1 accuracy:1 variance:1 ambedkar:1 descriptor:1 ensemble:1 correspond:1 identify:1 generalize:1 accurately:1 emphasizes:1 trajectory:1 whenever:2 definition:8 i01:2 proof:5 sampled:5 hsu:1 proved:2 dataset:2 popular:1 knowledge:1 fractional:1 improves:2 dimensionality:1 segmentation:7 formalize:1 scc:1 higher:8 follow:1 methodology:1 improved:1 wei:1 formulation:1 strongly:4 generality:1 smola:1 hand:1 overlapping:1 google:1 rodriguez:1 kernighan:1 lrsc:1 mode:2 grows:1 normalized:2 contain:2 true:5 verify:1 hence:14 kxn:2 symmetric:7 deal:1 sin:1 game:1 self:1 davis:3 criterion:1 scholk:1 presenting:1 evident:1 complete:1 demonstrate:2 theoretic:1 tron:1 performs:1 motion:8 image:7 variational:1 consideration:1 subdomain:1 recently:1 common:2 rotation:2 empirically:1 volume:3 belong:2 hypergraphs:18 numerically:1 significant:1 expressing:1 theorist:1 consistency:7 mathematics:3 similarity:5 etc:3 add:1 recent:4 belongs:1 certain:7 hyperedge:9 postponed:1 sgc:3 minimum:2 greater:1 relaxed:1 additional:1 surely:8 v3:1 dashed:1 signal:2 preservation:1 ii:2 semi:1 multiple:1 reduces:1 faster:4 match:15 academic:1 long:1 devised:1 post:1 grassmann:2 laplacian:5 underlies:1 variant:1 vision:10 expectation:3 arxiv:1 iteration:2 kernel:2 normalization:1 agarwal:1 achieved:1 preserved:2 whereas:2 background:1 addition:2 fellowship:1 eckhardt:1 singular:5 hyperedges:4 grow:1 crucial:1 unlike:1 sure:2 comment:2 strict:1 validating:1 undirected:3 spirit:1 effectiveness:1 jordan:1 practitioner:2 call:1 anandkumar:1 presence:1 enough:2 wn:10 easy:3 zi:7 regarding:2 cn:12 whether:1 collinear:2 remark:2 useful:1 detailed:1 eigenvectors:7 listed:1 ten:1 ph:1 outperform:1 exist:1 percentage:1 zj:2 tutorial:1 notice:1 disjoint:2 correctly:4 write:3 discrete:1 group:6 key:2 four:2 kci:6 drawn:1 bulo:1 registration:1 graph:19 znt:7 fraction:2 luxburg:3 angle:2 run:4 fourth:1 letter:1 arrive:1 almost:10 schweikert:1 appendix:2 prefer:1 comparable:1 bound:15 followed:1 i0m:2 tackled:1 encountered:1 annual:1 kronecker:3 precisely:2 scene:1 software:1 bousquet:1 span:3 concluding:1 expanded:1 relatively:1 department:1 combination:1 belonging:1 smaller:1 kakade:1 outlier:1 invariant:3 taken:1 computationally:1 ln:15 previously:1 turn:1 eventually:4 ge:1 vidal:1 observe:4 spectral:32 appropriate:2 occurrence:1 alternative:2 eigen:3 existence:1 original:1 kln:2 assumes:1 clustering:32 denotes:2 ensure:1 top:1 log2:2 build:1 establish:1 tensor:32 objective:1 occurs:2 planted:10 dependence:2 diagonal:5 affinity:19 subspace:10 distance:1 manifold:2 considers:1 trivial:1 reason:2 kannan:1 index:2 ratio:2 providing:1 btn:1 mostly:1 statement:2 negative:2 stated:4 rise:1 design:2 proper:1 unknown:3 perform:2 allowing:1 upper:2 observation:1 datasets:1 benchmark:2 incorrectly:1 defining:1 extended:1 rn:8 perturbation:3 arbitrary:1 community:2 overlooked:1 pair:2 required:2 connection:2 conclusive:1 z1:1 trans:2 below:1 usually:3 pattern:5 laplacians:1 including:2 video:1 power:1 rely:1 hybrid:1 lakamper:1 mn:17 representing:2 technology:1 numerous:1 carried:1 extract:1 deviate:1 geometric:5 literature:3 acknowledgement:1 kf:3 opf:1 loss:1 permutation:1 mixed:1 validation:1 foundation:4 incurred:1 degree:1 affine:3 sufficient:2 consistent:6 row:22 supported:1 last:1 formal:1 india:1 institute:1 face:3 sparse:1 boundary:1 dimension:2 xn:7 unweighted:3 contour:1 collection:1 made:1 simplified:1 transaction:1 approximate:1 preferred:1 keep:1 clique:2 global:2 active:1 reveals:2 b1:1 conclude:3 belongie:1 spectrum:2 iterative:1 why:1 reasonably:3 robust:2 ignoring:1 csa:1 flattening:2 main:1 s2:2 noise:2 n2:1 repeated:1 convey:1 representative:2 slow:1 cooper:1 theme:1 explicit:2 candidate:4 third:2 theorem:16 rk:1 rohe:3 decay:4 evidence:2 grouping:5 exists:9 essential:1 workshop:1 false:1 adding:1 importance:2 flattened:3 ci:2 illumination:2 conditioned:1 chatterjee:1 gap:3 flavor:1 easier:1 suited:1 chen:1 led:1 infinitely:1 tracking:2 satisfies:1 relies:2 goal:2 formulated:2 considerable:1 hard:1 change:1 uniformly:1 lemma:13 called:1 lerman:1 indicating:1 formally:1 indian:1 |
4,766 | 5,315 | Projecting Markov Random Field Parameters for
Fast Mixing
Justin Domke
NICTA, The Australian National University
[email protected]
Xianghang Liu
NICTA, The University of New South Wales
[email protected]
Abstract
Markov chain Monte Carlo (MCMC) algorithms are simple and extremely powerful techniques to sample from almost arbitrary distributions. The flaw in practice
is that it can take a large and/or unknown amount of time to converge to the stationary distribution. This paper gives sufficient conditions to guarantee that univariate
Gibbs sampling on Markov Random Fields (MRFs) will be fast mixing, in a precise sense. Further, an algorithm is given to project onto this set of fast-mixing
parameters in the Euclidean norm. Following recent work, we give an example use
of this to project in various divergence measures, comparing univariate marginals
obtained by sampling after projection to common variational methods and Gibbs
sampling on the original parameters.
1 Introduction
Exact inference in Markov Random Fields (MRFs) is generally intractable, motivating approximate
algorithms. There are two main classes of approximate inference algorithms: variational methods
and Markov chain Monte Carlo (MCMC) algorithms [13].
Among variational methods, mean-field approximations [9] are based on a ?tractable? family of
distributions, such as the fully-factorized distributions. Inference finds a distribution in the tractable
set to minimize the KL-divergence from the true distribution. Other methods, such as loopy belief
propagation (LBP), generalized belief propagation [14] and expectation propagation [10] use a less
restricted family of target distributions, but approximate the KL-divergence. Variational methods
are typically fast, and often produce high-quality approximations. However, when the variational
approximations are poor, estimates can be correspondingly worse.
MCMC strategies, such as Gibbs sampling, simulate a Markov chain whose stationary distribution is
the target distribution. Inference queries are then answered by the samples drawn from the Markov
chain. In principle, MCMC will be arbitrarily accurate if run long enough. The principal difficulty
is that the time for the Markov chain to converge to its stationary distribution, or the ?mixing time?,
can be exponential in the number of variables.
This paper is inspired by a recent hybrid approach for Ising models [3]. This approach minimizes
the divergence from the true distribution to one in a tractable family. However, the tractable family
is a ?fast mixing? family where Gibbs sampling is guaranteed to quickly converge to the stationary
distribution. They observe that an Ising model will be fast mixing if the spectral norm of a matrix
containing the absolute values of all interactions strengths is controlled. An algorithm projects
onto this fast mixing parameter set in the Euclidean norm, and projected gradient descent (PGD)
can minimize various divergence measures. This often leads to inference results that are better
than either simple variational methods or univariate Gibbs sampling (with a limited time budget).
However, this approach is limited to Ising models, and scales poorly in the size of the model, due to
the difficulty of projecting onto the spectral norm.
1
The principal contributions of this paper are, first, a set of sufficient conditions to guarantee that
univariate Gibbs sampling on an MRF will be fast-mixing (Section 4), and an algorithm to project
onto this set in the Euclidean norm (Section 5). A secondary contribution of this paper is considering
an alternative matrix norm (the induced ?-norm) that is somewhat looser than the spectral norm,
but more computationally efficient. Following previous work [3], these ideas are experimentally
validated via a projected gradient descent algorithm to minimize other divergences, and looking at
the accuracy of the resulting marginals. The ability to project onto a fast-mixing parameter set may
also be of independent interest. For example, it might be used during maximum likelihood learning
to ensure that the gradients estimated through sampling are more accurate.
2 Notation
We consider discrete pairwise MRFs with n variables, where the i-th variable takes values in
{1, ..., Li }, E is the set of edges, and ? are the potentials on each edge. Each edge in E is an ordered
pair (i, j) with i ? j. The parameters are a set of matrices ? := {?ij |?ij ? RLi ?Lj , ?(i, j) ? E}.
When i > j, and (j, i) ? E, we let ?ij denote the transpose of ?ji . The corresponding distribution is
?
?
#
p(x; ?) = exp ?
?ij (xi , xj ) ? A(?)? ,
(1)
(i,j)?E
where A(?) := log
&
x exp
'&
(
ij
?
(x
,
x
)
is the log-partition function, and ?ij (xi , xj )
i
j
(i,j)?E
denotes the entry in the xi -th row and xj -th column of ?ij . It is easy to show that any parametrization
of a pairwise MRF can be converted into this form. ?Self-edges? (i, i) can be included in E if one
wishes to explicitly represent univariate terms.
It is sometimes convenient to work with the exponential family representation
p(x; ?) = exp{f (x) ? ? ? A(?)},
(2)
where f (x) is the sufficient statistics for configuration x. If these are indicator functions for all
configurations of all pairs in E, then the two representations are equivalent.
3 Background Theory on Rapid Mixing
This section reviews background on mixing times that will be used later in the paper.
Definition 1. Given two finite
& distributions p and q, the total variation distance ? ? ?T V is defined
as ?p(X) ? q(X)?T V = 21 x |p(X = x) ? q(X = x)|.
Next, one must define a measure of how fast a Markov chain converges to the stationary distribution.
Let the state of the Markov chain after t iterations be X t . Given a constant ?, this is done by finding
some number of iterations ? (?) such that the induced distribution p(X t |X 0 = x) will always have a
distance of less than ? from the stationary distribution, irrespective of the starting state x.
Definition 2. Let {X t } be the sequence of random variables corresponding to running Gibbs sampling on a distribution p. The mixing time ? (?) is defined as ? (?) = min{t : d(t) < ?}, where
d(t) = maxx ?P(X t |X 0 = x) ? p(X)?T V is the maximum distance at time t when considering all
possible starting states x.
Now, we are interested in when Gibbs sampling on a distribution p can be shown to have a fast
mixing time. The central property we use is the dependency of one variable on another, defined
informally as how much the conditional distribution over Xi can be changed when all variables
other than Xj are the same.
Definition 3. Given a distribution p, the dependency matrix R is defined by
Rij = ? max ? ?p(Xi |x?i ) ? p(Xi |x??i )?T V .
x,x :x?j =x?j
Here, the constraint x?j = x??j indicates that all variables in x and x? are identical except xj . The
central result on rapid mixing is given by the following Theorem, due to Dyer et al. [5], generalizing
the work of Hayes [7]. Informally, it states that if ?R? < 1 for any sub-multiplicative norm ? ? ?,
then mixing will take on the order of n ln n iterations, where n is the number of variables.
2
Theorem 4. [5, Lemma 17] If ? ? ? is any sub-multiplicative matrix norm and ||R|| < 1, the mixing
time of univariate!Gibbs sampling
on a system with n variables with random updates is bounded by
"
?1n ? ?1T
n
n?
? (?) ? 1??R? ln
.
?
Here, ?1n ? denotes the same matrix norm applied to a matrix of ones of size n ? 1, and similarly
for 1Tn . In particular, if ? ? ? induced by a vector p-norm, then ?1n ? ?1Tn ? = n.
Since this result is true for a variety of norms, it is natural to ask, for a given matrix R, which norm
will give the strongest result. It can be shown that for symmetric matrices (such as the dependency
matrix), the spectral norm ? ? ?2 is always superior.
Theorem 5. [5, Lemma 13] If A is a symmetric matrix and ? ? ? is any sub-multiplicative norm,
then ?A?2 ? ?A?.
Unfortunately, as will be discussed below, the spectral norm can be more computationally expensive
than other norms. As such, we will also consider the use of the ?-norm ? ? ?? . This leads to
additional looseness in the bound in general, but is limited in some cases. In particular if R = rG
where G is the adjacency matrix for some regular graph with degree d, then for all induced p-norms,
?R? = rd, since ?R? = maxx?=0 ?Rx?/?x| = r maxx?=0 ?Gx?/?x? = r?Go?/?o? = rd, where
o is a vector of ones. Thus, the extra looseness from using, say, ? ? ?? instead of ? ? ?2 will tend to
be minimal when the graph is close to regular, and the dependency is close to a constant value. For
irregular graphs with highly variable dependency, the looseness can be much larger.
4 Dependency for Markov Random Fields
In order to establish that Gibbs sampling on a given MRF will be fast mixing, it is necessary to
compute (a bound on) the dependency matrix R, as done in the following result. The proof of this
result is fairly long, and so it is postponed to the Appendix. Note that it follows from several bounds
on the dependency that are tighter, but less computationally convenient.
Theorem 6. The dependency matrix for a pairwise Markov random field is bounded by
1 ij
ij
Rij (?) ? max ???a
? ??b
?? .
a,b 2
ij
Here, ??a
indicates the a?th column of ?ij . Note that the MRF can include univariate terms as selfedges with no impact on the dependency bound, regardless of the strength of the univariate terms. It
can be seen easily that from the definition of R (Definition 3), for any i the entry Rii for self-edges
(i, i) should always be zero. One can, without loss of generality, set each column of ?ii to be the
same, meaning that Rii = 0 in the above bound.
5 Euclidean Projection Operator
The Euclidean distance between two MRFs parameterized respectively by ? and ? is ?? ? ??2 :=
#
ij
ij 2
(i,j)?E ?? ? ? ?F . This section considers projecting a given vector ? onto the fast mixing set
or, formally, finding a vector ? with minimum Euclidean distance to ?, subject to the constraint
that a norm ? ? ?? applied to the bound on the dependency matrix R is less than some constant c.
Euclidean projection is considered because, first, it is a straightforward measure of the closeness
between two parameters and, second, it is the building block of the projected gradient descent for
projection in other distance measures. To begin with, we do not specify the matrix norm ? ? ?? , as it
could be any sub-multiplicative norm (Section 3).
Thus, in principle, we would like to find ? to solve
projc (?) := argmin ?? ? ??2 .
(3)
?:?R(?)?? ?c
Unfortunately, while convex, this optimization turns out to be somewhat expensive to solve, due to
a lack of smoothness Instead, we introduce a matrix Z, and constrain that Zij ? Rij (?), where
Rij (?) is the bound on dependency in Thm 6 (as an equality). We add an extra quadratic term
3
??Z ? Y ?2F to the objective, where Y is an arbitrarily given matrix and ? > 0 is trade-off between
the smoothness and the closeness to original problem (3). The smoothed projection operator is
projC (?, Y ) := argmin ?? ? ??2 + ??Z ? Y ?2F , C = {(?, Z) : Zij ? Rij (?), ?Z?? ? c}. (4)
(?,Z)?C
If ? = 0, this yields a solution that is identical to that of Eq. 3. However, when ? = 0, the objective
in Eq. 4 is not strongly convex as a function of Z, which results in a dual function which is nonsmooth, meaning it must be solved with a method like subgradient descent, with a slow convergence
rate. In general, of course, the optimal point of Eq. 4 is different to that of Eq. 3. However,
the main usage of the Euclidean projection operator is the projection step in the projected gradient
descent algorithm for divergence minimization. In these tasks the smoothed projection operator can
be directly used in the place of the non-smoothed one without changing the final result. In situations
when the exact Euclidean projection is required, it can be done by initializing Y1 arbitrarily and
repeating (?k+1 , Yk+1 ) ? projC (?, Yk ), for k = 1, 2, . . . until convergence.
5.1 Dual Representation
Theorem 7. Eq. 4 has the dual representation
maximize
?,?,?,?
g(?, ?, ?, ?)
,
(5)
subject to ?ij (a, b, c) ? 0, ?ij (a, b, c) ? 0, ?(i, j) ? E, a, b, c
where
g(?, ?, ?, ?) = min h1 (Z; ?, ?, ?, ?) + min h2 (?; ?, ?)
Z
?
h1 (Z; ?, ?, ?, ?) = ?tr(Z?T ) + I(?Z?? ? c) + ??Z ? Y ?2F
# ij
1 ! !"
ij
h2 (?; ?, ?) = ?? ? ??2 +
?ij (a, b, c) ? ?ij (a, b, c) (?c,a
? ?c,b
),
2
i,j?E a,b,c
$
? ij +
?
in which ?ij := ?ij Dij + ?
:=
a,b,c ?ij (a, b, c) + ?ij (a, b, c), where ?ij
%
?ij
if (i, j) ? E
, and D is an indicator matrix with Dij = 0 if (i, j) ? E or (j, i) ? E,
??ij if (j, i) ? E
and Dij = 1 otherwise. The dual variables ?ij and ?ij are arrays of size Lj ? Li ? Li for all pairs
(i, j) ? E while ? and ? are of size n ? n.
The proof of this is in the Appendix. Here, I(?) is the indicator function with I(x) = 0 when x is
true and I(x) = ? otherwise.
Being a smooth optimization problem with simple bound constraints, Eq. 5 can be solved with
LBFGS-B [2]. For a gradient-based method like this to be practical, it must be possible to quickly
evaluate g and its gradient. This is complicated by the fact that g is defined in terms of the minimization of h1 with respect to Z and h2 with respect to ?. We discuss how to solve these problems
now. We first consider the minimization of h2 . This is a quadratic function of ? and can be solved
?
analytically via the condition that ??
h2 (?; ?, ?) = 0. The closed form solution is
&
'
!
!
!
1 !
ij
ij
?c,a = ?c,a ?
?ij (a, b, c) ?
?ij (b, a, c) ?
?ij (a, b, c) +
?ij (b, a, c)
4
b
b
b
b
?(i, j) ? E, 1 ? a, c ? m.. The time complexity is linear in the size of ?.
Minimizing h1 is more involved. We assume to start that there exists an algorithm to quickly project
a matrix onto the set {Z : ?Z?? ? c}, i.e. to solve the optimization problem of
min ?Z ? A?2F .
(6)
?Z?? ?c
Then, we observe that arg minZ h1 is equal to
arg min ?tr(Z?T ) + I(?Z?? ? c) + ??Z ? Y ?2F = arg min ?Z ? (Y +
Z
?Z?? ?c
4
1
?)?2F .
2?
For different norms ? ? ?? , the projection algorithm will be different and can have a large impact on
efficiency. We will discuss in the followings sections the choices of ? ? ?? and an algorithm for the
?-norm.
Finally, once h1 and h2 have been solved, the gradient of g is (by Danskin?s theorem [1])
?g
?g
= ? Dij Z?ij ,
=Z?ji ? Z?ij ,
??ij
??ij
?g
1
?g
ij
= (??ij ? ??c,b
) ? Z?ij ,
= ? ??ij (a,b,c) g,
??ij (a, b, c) 2 c,a
??ij (a, b, c)
where Z? and ?? represent the solutions to the subproblems.
5.2 Spectral Norm
When ? ??? is set to the spectral norm, i.e. the largest singular value of a matrix, the projection in Eq.
6 can be performed by thresholding the singular values of A [3]. Theoretically, using spectral norm
will give a tighter bound on Z than other norms (Section 3). However, computing a full singular
value decomposition can be impractically slow for a graph with a large number of variables.
5.3 ?-norm
!
Here, we consider setting ? ? ?? to the ?-norm, ?A?? = maxi j |Aij |, which measures the
maximum l1 norm of the rows of A. This norm has several computational advantages. Firstly, to
project a matrix onto a ?-norm ball {A : ?A? ? ? c}, we can simply project each row ai of the
matrix onto the l1 -norm ball {a : ?a?1 ? c}. Duchi et al. [4] provide a method linear in the number
of nonzeros in a and logarithmic in the length of a. Thus, if Z is an n ? n, matrix, Eq. 6 for the
?-norm can be solved in time n2 and, for sufficiently sparse matrices, in time n log n.
A second advantage of the ?-norm is that (unlike the spectral norm) projection in Eq. 6 preserves
the sparsity of the matrix. Thus, one can disregard the matrix D and dual variables ? when solving
the optimization in Theorem 7. This means that Z itself can be represented sparsely, i.e. we only
need variables for those (i, j) ? E. These simplifications significantly improve the efficiency of
projection, with some tradeoff in accuracy.
6 Projection in Divergences
In this section, we want to find a distribution p(x; ?) in the fast mixing family closest to a target
distribution p(x; ?) in some divergence D(?, ?). The choice of divergence depends on convenience
of projection, the approximate family and the inference task. We will first present a general algorithmic framework based on projected gradient descent (Algorithm 1), and then discuss the details
of several previously proposed divergences [11, 3].
6.1 General algorithm framework for divergence minimization
The problem of projection in divergences is formulated as
min D(?, ?),
??C?
(7)
D(?, ?) is some divergence measure, and C? := {? : ?Z, s.t.(?, Z) ? C}, where C is the feasible set
in Eq. 4. Our general strategy for this is to use projected gradient descent to solve the optimization
min D(?, ?),
(?,Z)?C
(8)
using the joint operator to project onto C described in Section 5.
For different divergences, the only difference in projection algorithm is the evaluation of the gradient
?? D(?, ?). It is clear that if (?? , Z ? ) is the solution of Eq. 8, then ?? is the solution of 7.
6.2 Divergences
5
Algorithm 1 Projected gradient descent for divergence projection
Initialize (?1 , Z1 ), k ? 1.
repeat
?? ? ?k ? ??? D(?, ?k )
(?k+1 , Zk+1 ) ? projC (?? , Zk )
k ?k+1
until convergence
In this section, we will discuss the different choices of divergences and corresponding projection algorithms.
Grid, Attractive only
0.7
0.6
6.2.1 KL-divergence
Marginal Error
KL-divergence KL(???)
:=
p(x;?)
is arguably the
x p(x; ?) log p(x;?)
optimal divergence for marginal inference
because it strives to preserve the marginals
of p(x; ?) and p(x; ?). However, projection
in KL-divergence is intractable here because
the evaluation of the gradient ?? KL(???)
requires the marginals of distribution ?.
The
!
0.5
0.3
0.2
0.1
0
0
6.2.2 Piecewise KL-divergence
0.5
1
1.5
2
2.5
3
Interaction Strength
Edge density = 0.50, Mixed
3.5
4
3.5
4
0.45
LBP
TRW
Mean?Field
Original Parameters
Euclidean
Piecewise KL(?||?)
KL(?||?)
0.4
0.35
0.3
Marginal Error
One tractable surrogate of KL(???) is
the piecewise KL-divergence [3] defined over some tractable subgraphs.
Here, D(?, ?) := maxT ?T KL(?T ??T ),
where T is a set of low-treewidth subgraphs. The gradient can be evaluated as
?? D(?, ?) = ?? KL(?T ? ??T ? ) where
T ? = arg maxT ?T KL(?T ??T ). For any
T in T , KL(?T ??T ) and its gradient can be
evaluated by the junction-tree algorithm.
0.4
LBP
TRW
Mean?Field
Original Parameters
Euclidean
Piecewise KL(?||?) (TW 1)
Piecewise KL(?||?) (TW 2)
KL(?||?)
0.25
0.2
0.15
0.1
0.05
0
0
6.2.3 Reversed KL-divergence
0.5
1
1.5
2
2.5
Interaction Strength
3
The ?reversed? KL-divergence KL(???) is
minimized by mean-field methods.
In Figure 1: Mean univariate marginal error on 16 ? 16
general KL(???) is inferior to KL(???) grids (top) with attractive interactions and medianfor marginal inference since it tends to density random graphs (bottom) with mixed interacunderestimate the support of the distri- tions, comparing 30k iterations of Gibbs sampling afbution [11]. Still, it often works well ter projection (onto the l? norm) to variational methin practice. ?? KL(???)
can computed ods. The6 original parameters also show a lower curve
!
as ?"? KL(???) =
p(x;
?)(? ? ?) ? with 10 samples.
x
#
f (x) f (x) ? ?(?) , which can be approximated by samples generated from p(x; ?) [3]. In implementation, we maintain a ?pool? of samples,
each of which is updated by a single Gibbs step after each iteration of Algorithm 1.
7 Experiments
The experiments below take two stages: first, the parameters are projected (in some divergence) and
then we compare the accuracy of sampling with the resulting marginals. We focus on this second
aspect. However, we provide a comparison of the computation time for various projection algorithms
in Table 1, and when comparing the accuracy of sampling with a given amount of time, provide two
6
curves for sampling with the original parameters, where one curve has an extra amount of sampling
effort roughly approximating the time to perform projection in the reversed KL divergence.
7.1 Synthetic MRFs
Interaction strength = 2.00, Attractive Only
0.55
LBP
TRW
Mean?Field
Original Parameters
Euclidean
Piecewise KL(?||?) (TW 1)
Piecewise KL(?||?) (TW 2)
KL(?||?)
0.5
Our first experiment follows that of [8, 3]
in evaluating the accuracy of approximation
methods in marginal inference. In the experiments, we approximate randomly generated
MRF models with rapid-mixing distributions
using the projection algorithms described previously. Then, the marginals of fast mixing
approximate distributions are estimated by running a Gibbs chain on each distribution. These
are compared against exact marginals as computed by the junction tree algorithm. We use
the mean absolute difference of the marginals
|p(Xi = 1) ? q(Xi = 1)| as the accuracy measure. We compare to Naive mean-field (MF),
Gibbs sampling on original parameters (Gibbs),
and Loopy belief propagation (LBP). Many
other methods have been compared against a
similar benchmark [6, 8].
0.45
Marginal Error
0.4
0.35
0.3
0.25
0.2
0.15
0.1
0.05 0
10
1
2
10
10
3
4
5
10
10
Number of Samples
6
10
10
Edge density = 0.50, Interaction strength = 3.00, Mixed
0.5
LBP
TRW
Mean?Field
Original Parameters
Euclidean
Piecewise KL(?||?)
KL(?||?)
0.45
0.4
Marginal Error
0.35
0.3
While our methods are for general MRFs, we
0.25
test on Ising potentials because this is a stan0.2
dard benchmark. Two graph topologies are
used: two-dimensional 16 ? 16 grids and 10
0.15
node random graphs, where each edge is in0.1
dependently present with probability pe ?
0.05
{0.3, 0.5, 0.7}. Node parameters ?i are uni10
10
10
10
10
10
form from [?dn , dn ] with fixed field strength
Number of Samples
dn = 1.0. Edge parameters ?ij are uniform
from [?de , de ] or [0, de ] to obtain mixed or at- Figure 2: Examples of the accuracy of obtained
tractive interactions respectively, with interac- marginals vs. the number of samples. Top:
tion strengths de ? {0, 0.5, . . . , 4}. Figure 1 Grid graphs. Bottom: Median-Density Random
shows the average marginal error at different graphs.
interaction strengths. Error bars show the standard error normalized by the number of samples, which can be interpreted as a 68.27% confidence
interval. We also include time-accuracy comparisons in Figure 2. All results are averaged over 50
random trials. We run Gibbs long enough ( 106 samples) to get a fair comparison in terms of running
time.
0
1
2
3
4
5
Except where otherwise stated, parameters are projected onto the ball {? : ?R(?)?? ? c}, where
c = 2.5 is larger than the value of c = 1 suggested by the proofs above. Better results are obtained
by using this larger constraint set, presumably because of looseness in the bound. For piecewise
projection, grids use simple vertical and horizontal chains of treewidth either one or two. For random
graphs, we randomly generate spanning trees until all edges are covered. Gradient descent uses a
fixed step size of ? = 0.1. A Gibbs step is one ?systematic-scan? pass over all variables between.
The reversed KL divergence maintains a pool of 500 samples, each of which is updated by a single
Gibbs step in each iteration.
We wish to compare the trade-off between computation time and accuracy represented by the choice
between the use of the ? and spectral norms. We measure the running time on 16 ? 16 grids in
Table 1, and compare the accuracy in Figure 3.
The appendix contains results for a three-state Potts model on an 8 ? 8 grid, as a test of the multivariate setting. Here, the intractable divergence KL(???) is included for reference, with the projection
computed with the help of the junction tree algorithm for inference.
7
Table 1: Running times on 16 ? 16 grids with attractive interactions. Euclidean projection converges
in around 5 LBFGS-B iterations. Piecewise projection (with a treewidth of 1) and reversed KL
projection use 60 gradient descent steps. All results use a single core of a Intel i7 860 processor.
Gibbs
Euclidean
Piecewise
Reversed-KL
30k Steps 106 Steps l? norm l2 norm l? norm l2 norm l? norm l2 norm
de = 1.5 0.67s
22.42s
1.50s
25.63s 12.87s 45.26s 13.13s 66.81s
de = 3.0 0.67s
22.42s
3.26s 164.34s 20.73s 211.08s 20.12s 254.25s
7.2 Berkeley binary image denoising
Grid, Mixed
Marginal Error
This experiment evaluates various methods
0.35
Euclidean SP
for denoising binary images from the BerkePiecewise KL(?||?) (TW 1) SP
0.3
ley segmentation dataset downscaled from
Piecewise KL(?||?) (TW 2) SP
KL(?||?) SP
300 ? 200 to 120 ? 80. The images are
Euclidean Inf
0.25
binarized by setting Yi = 1 if pixel i is above
Piecewise KL(?||?) (TW 1) Inf
Piecewise KL(?||?) (TW 2) Inf
the average gray scale in the image, and
0.2
KL(?||?) Inf
Yi = ?1. The noisy image X is created by
0.15
i 1.25
setting: Xi = Yi2+1 i (1 ? t1.25
) + 1?Y
,
i
2 ti
in which ti is sampled uniformly from [0, 1].
0.1
For inference purposes, the conditional
0.05
distribution
as P (Y |X)
?
! " Y is modeled
#
"
?
exp ? ij Yi Yj + 2 i (2Xi ? 1)Yi ,
0
0
0.5
1
1.5
2
2.5
3
3.5
4
Interaction Strength
where the pairwise strength ? > 0 encourages
smoothness.
On this attractive-only Ising
potential, the Swendsen-Wang method [12] Figure 3: The marginal error using ?-norm promixes rapidly, and so we use the resulting jection (solid lines) and spectral-norm projection
samples to estimate the ground truth. The (dotted lines) on 16x16 Ising grids.
parameters ? and ? are heuristically chosen to
be 0.5 and 0.7 respectively.
Marginal Error
Figure 4 shows the decrease of average
marginal error. To compare running time, Euclidean and K(???) projection cost approximately the same as sampling 105 and 4.8 ? 105
samples respectively. Gibbs sampling on the
original parameter converges very slowly. Sampling the approximate distributions from our
projection algorithms converge quickly in less
than 104 samples.
0.5
LBP
Mean?Field
Original Parameter
Euclidean
Piecewise KL(?||?) (TW 1)
KL(?||?)
0.45
Marginal Error
0.4
0.35
0.3
0.25
0.2
0.15
8 Conclusions
0.1 0
10
1
10
2
10
3
10
4
10
5
10
6
10
We derived sufficient conditions on the parameNumber of samples
ters of an MRF to ensure fast-mixing of univariate Gibbs sampling, along with an algorithm to Figure 4: Average marginal error on the Berkeley
project onto this set in the Euclidean norm. As segmentation dataset.
an example use, we explored the accuracy of
samples obtained by projecting parameters and
then sampling, which is competitive with simple variational methods as well as traditional Gibbs
sampling. Other possible applications of fast-mixing parameter sets include constraining parameters during learning.
Acknowledgments
NICTA is funded by the Australian Government through the Department of Communications and
the Australian Research Council through the ICT Centre of Excellence Program.
8
References
[1] Dimitri Bertsekas. Nonlinear Programming. Athena Scientific, 2004. 5.1
[2] Richard H. Byrd, Peihuang Lu, Jorge Nocedal, and Ciyou Zhu. A limited memory algorithm
for bound constrained optimization. SIAM J. Sci. Comput., 16(5):1190?1208, 1995. 5.1
[3] Justin Domke and Xianghang Liu. Projecting Ising model parameters for fast mixing. In NIPS,
2013. 1, 5.2, 6, 6.2.2, 6.2.3, 7.1
[4] John C. Duchi, Shai Shalev-Shwartz, Yoram Singer, and Tushar Chandra. Efficient projections
onto the l1 -ball for learning in high dimensions. In ICML, 2008. 5.3
[5] Martin E. Dyer, Leslie Ann Goldberg, and Mark Jerrum. Matrix norms and rapid mixing for
spin systems. Ann. Appl. Probab., 19:71?107, 2009. 3, 4, 5
[6] Amir Globerson and Tommi Jaakkola. Approximate inference using conditional entropy decompositions. In UAI, 2007. 7.1
[7] Thomas P. Hayes. A simple condition implying rapid mixing of single-site dynamics on spin
systems. In FOCS, pages 39?46, 2006. 3
[8] Tamir Hazan and Amnon Shashua. Convergent message-passing algorithms for inference over
general graphs with convex free energies. In UAI, pages 264?273, 2008. 7.1
[9] D. Koller and N. Friedman. Probabilistic Graphical Models: Principles and Techniques. MIT
Press, 2009. 1
[10] Thomas Minka. Expectation propagation for approximate bayesian inference. In UAI, 2001. 1
[11] Thomas Minka. Divergence measures and message passing. Technical report, 2005. 6, 6.2.3
[12] Robert H. Swendsen and Jian-Sheng Wang. Nonuniversal critical dynamics in monte carlo
simulations. Phys. Rev. Lett., 58:86?88, Jan 1987. 7.2
[13] Martin Wainwright and Michael Jordan. Graphical models, exponential families, and variational inference. Found. Trends Mach. Learn., 1(1-2):1?305, 2008. 1
[14] Jonathan Yedidia, William Freeman, and Yair Weiss. Constructing free energy approximations
and generalized belief propagation algorithms. IEEE Transactions on Information Theory,
51:2282?2312, 2005. 1
9
| 5315 |@word trial:1 norm:50 heuristically:1 simulation:1 decomposition:2 tr:2 solid:1 liu:3 configuration:2 contains:1 zij:2 com:2 comparing:3 od:1 must:3 john:1 partition:1 update:1 v:1 stationary:6 implying:1 amir:1 parametrization:1 core:1 node:2 gx:1 firstly:1 dn:3 along:1 focs:1 wale:1 downscaled:1 introduce:1 theoretically:1 excellence:1 pairwise:4 rapid:5 roughly:1 inspired:1 freeman:1 byrd:1 considering:2 project:10 begin:1 notation:1 bounded:2 distri:1 factorized:1 argmin:2 interpreted:1 minimizes:1 finding:2 guarantee:2 berkeley:2 binarized:1 ti:2 arguably:1 bertsekas:1 t1:1 tends:1 mach:1 approximately:1 might:1 au:2 appl:1 limited:4 averaged:1 practical:1 acknowledgment:1 globerson:1 yj:1 practice:2 block:1 jan:1 maxx:3 significantly:1 projection:33 convenient:2 confidence:1 regular:2 get:1 onto:14 close:2 convenience:1 operator:5 equivalent:1 go:1 regardless:1 starting:2 straightforward:1 convex:3 subgraphs:2 array:1 variation:1 updated:2 target:3 exact:3 programming:1 us:1 goldberg:1 trend:1 expensive:2 approximated:1 sparsely:1 ising:7 bottom:2 rij:5 solved:5 initializing:1 wang:2 trade:2 decrease:1 yk:2 complexity:1 dynamic:2 solving:1 efficiency:2 easily:1 joint:1 various:4 represented:2 fast:18 monte:3 query:1 shalev:1 whose:1 larger:3 solve:5 say:1 otherwise:3 ability:1 statistic:1 jerrum:1 itself:1 noisy:1 final:1 sequence:1 advantage:2 interaction:10 rapidly:1 mixing:26 poorly:1 convergence:3 produce:1 converges:3 tions:1 help:1 ij:47 eq:11 treewidth:3 australian:3 tommi:1 ley:1 adjacency:1 government:1 tighter:2 sufficiently:1 considered:1 around:1 swendsen:2 exp:4 presumably:1 ground:1 algorithmic:1 purpose:1 council:1 largest:1 minimization:4 mit:1 always:3 jaakkola:1 validated:1 focus:1 derived:1 potts:1 likelihood:1 indicates:2 sense:1 inference:15 flaw:1 mrfs:6 typically:1 lj:2 koller:1 interested:1 pixel:1 arg:4 among:1 dual:5 constrained:1 fairly:1 initialize:1 marginal:15 field:14 equal:1 once:1 sampling:24 identical:2 icml:1 minimized:1 nonsmooth:1 report:1 piecewise:15 richard:1 randomly:2 preserve:2 national:1 divergence:31 maintain:1 william:1 friedman:1 interest:1 message:2 highly:1 evaluation:2 chain:9 accurate:2 edge:10 necessary:1 tree:4 euclidean:20 minimal:1 column:3 leslie:1 loopy:2 cost:1 entry:2 uniform:1 dij:4 motivating:1 interac:1 dependency:12 synthetic:1 density:4 siam:1 systematic:1 off:2 probabilistic:1 pool:2 michael:1 quickly:4 central:2 containing:1 slowly:1 worse:1 dimitri:1 li:3 potential:3 converted:1 de:6 jection:1 explicitly:1 depends:1 later:1 multiplicative:4 h1:6 closed:1 performed:1 tion:1 hazan:1 shashua:1 start:1 competitive:1 maintains:1 complicated:1 shai:1 contribution:2 minimize:3 spin:2 accuracy:11 yield:1 bayesian:1 lu:1 carlo:3 rx:1 processor:1 strongest:1 phys:1 definition:5 against:2 evaluates:1 energy:2 involved:1 minka:2 proof:3 sampled:1 dataset:2 ask:1 segmentation:2 trw:4 specify:1 wei:1 done:3 evaluated:2 strongly:1 generality:1 stage:1 until:3 sheng:1 horizontal:1 nonlinear:1 propagation:6 lack:1 quality:1 gray:1 scientific:1 usage:1 building:1 normalized:1 true:4 equality:1 analytically:1 symmetric:2 dependently:1 attractive:5 during:2 self:2 encourages:1 inferior:1 generalized:2 tn:2 duchi:2 l1:3 meaning:2 variational:9 image:5 common:1 superior:1 ji:2 discussed:1 peihuang:1 marginals:9 gibbs:22 projc:4 ai:1 smoothness:3 rd:2 grid:10 similarly:1 centre:1 funded:1 add:1 closest:1 multivariate:1 recent:2 inf:4 binary:2 arbitrarily:3 jorge:1 postponed:1 yi:4 seen:1 minimum:1 additional:1 somewhat:2 converge:4 maximize:1 ii:1 full:1 nonzeros:1 smooth:1 technical:1 long:3 controlled:1 impact:2 mrf:6 expectation:2 chandra:1 iteration:7 represent:2 sometimes:1 irregular:1 lbp:7 background:2 want:1 interval:1 singular:3 median:1 jian:1 extra:3 unlike:1 south:1 induced:4 tend:1 subject:2 jordan:1 ter:1 constraining:1 enough:2 easy:1 variety:1 xj:5 nonuniversal:1 topology:1 idea:1 tradeoff:1 i7:1 rli:1 amnon:1 effort:1 passing:2 generally:1 clear:1 informally:2 covered:1 amount:3 repeating:1 generate:1 dotted:1 estimated:2 discrete:1 drawn:1 changing:1 nocedal:1 graph:11 subgradient:1 run:2 parameterized:1 powerful:1 place:1 almost:1 family:9 looser:1 appendix:3 bound:11 guaranteed:1 simplification:1 convergent:1 quadratic:2 strength:11 constraint:4 constrain:1 aspect:1 simulate:1 answered:1 extremely:1 min:8 martin:2 department:1 ball:4 poor:1 strives:1 tw:9 rev:1 projecting:5 restricted:1 computationally:3 ln:2 previously:2 turn:1 discus:4 singer:1 dyer:2 tractable:6 junction:3 yedidia:1 observe:2 spectral:11 alternative:1 yair:1 xianghang:3 original:11 thomas:3 denotes:2 running:6 ensure:2 include:3 top:2 graphical:2 in0:1 yoram:1 establish:1 approximating:1 objective:2 strategy:2 traditional:1 surrogate:1 gradient:17 distance:6 reversed:6 sci:1 athena:1 considers:1 spanning:1 nicta:5 length:1 modeled:1 minimizing:1 unfortunately:2 robert:1 subproblems:1 stated:1 danskin:1 implementation:1 rii:2 unknown:1 looseness:4 perform:1 vertical:1 markov:12 benchmark:2 finite:1 descent:10 situation:1 looking:1 precise:1 communication:1 y1:1 smoothed:3 arbitrary:1 thm:1 pair:3 required:1 kl:44 z1:1 nip:1 justin:3 bar:1 suggested:1 below:2 sparsity:1 program:1 max:2 memory:1 belief:4 wainwright:1 critical:1 difficulty:2 hybrid:1 natural:1 indicator:3 ciyou:1 zhu:1 improve:1 irrespective:1 created:1 naive:1 review:1 ict:1 l2:3 probab:1 fully:1 loss:1 mixed:5 h2:6 degree:1 sufficient:4 principle:3 thresholding:1 maxt:2 row:3 course:1 changed:1 repeat:1 transpose:1 free:2 aij:1 correspondingly:1 absolute:2 sparse:1 curve:3 dimension:1 lett:1 evaluating:1 tamir:1 dard:1 projected:9 transaction:1 approximate:9 hayes:2 uai:3 xi:10 shwartz:1 table:3 learn:1 zk:2 constructing:1 sp:4 main:2 yi2:1 n2:1 fair:1 site:1 intel:1 x16:1 slow:2 sub:4 wish:2 exponential:3 comput:1 pe:1 minz:1 theorem:7 maxi:1 explored:1 closeness:2 intractable:3 exists:1 budget:1 mf:1 rg:1 generalizing:1 logarithmic:1 entropy:1 simply:1 univariate:10 lbfgs:2 ordered:1 ters:1 truth:1 conditional:3 formulated:1 ann:2 feasible:1 experimentally:1 included:2 pgd:1 except:2 uniformly:1 domke:3 impractically:1 denoising:2 principal:2 lemma:2 total:1 tushar:1 secondary:1 pas:1 disregard:1 formally:1 support:1 mark:1 scan:1 jonathan:1 evaluate:1 mcmc:4 |
4,767 | 5,316 | Blossom Tree Graphical Models
John Lafferty
Department of Statistics
Department of Computer Science
University of Chicago
Zhe Liu
Department of Statistics
University of Chicago
Abstract
We combine the ideas behind trees and Gaussian graphical models to form a new
nonparametric family of graphical models. Our approach is to attach nonparanormal ?blossoms?, with arbitrary graphs, to a collection of nonparametric trees.
The tree edges are chosen to connect variables that most violate joint Gaussianity.
The non-tree edges are partitioned into disjoint groups, and assigned to tree nodes
using a nonparametric partial correlation statistic. A nonparanormal blossom is
then ?grown? for each group using established methods based on the graphical
lasso. The result is a factorization with respect to the union of the tree branches
and blossoms, defining a high-dimensional joint density that can be efficiently estimated and evaluated on test points. Theoretical properties and experiments with
simulated and real data demonstrate the effectiveness of blossom trees.
1
Introduction
Let p? (x) be a probability density on Rd corresponding to a random vector X = (X1 , . . . , Xd ). The
undirected graph G = (V, E) associated with p? has d = |V | vertices corresponding to X1 , . . . , Xd ,
and missing edges (i, j) 6? E whenever Xi and Xj are conditionally independent given the other
variables. The undirected graph is a useful way of exploring and modeling the distribution.
In this paper we are concerned with building graphical models for continuous variables, under
weaker assumptions than those imposed by existing methods. If p? (x) > 0 is strictly positive,
the Hammersley-Clifford theorem implies that the density has the form
!
Y
X
?
p (x) ?
?C (xC ) = exp
fC (xC ) .
(1.1)
C?C
C?C
In this expression, C denotes the set of cliques in the graph, and ?C (xC ) = exp(fC (xC )) > 0
denotes arbitrary potential functions. This represents a very large and rich set of nonparametric
graphical models. The fundamental difficulty is that it is in general intractable to compute the
normalizing constant. A compromise must be made to achieve computationally tractable inference,
typically involving strong assumptions on the functions fC , on the graph G = {C}, or both.
The default model for graphical modeling of continuous data is the multivariate Gaussian. When
the Gaussian has covariance matrix ?, the graph is encoded in the sparsity pattern of the precision
matrix ? = ??1 . Specifically, edge (i, j) is missing if and only if ?ij = 0. Recent work has
focused on sparse estimates of the precision matrix [8, 10]. In particular, an efficient algorithm for
computing the estimator using a graphical version of the lasso is developed in [3]. The nonparanormal [5], a form of Gaussian copula, weakens the Gaussian assumption by imposing Gaussianity
on the transformed random vector f (X) = (f1 (X1 ), . . . , fd (Xd )), where each fj is a monotonic
function. This allows arbitrary single variable marginal probability distributions in the model [5].
1
Both the Gaussian graphical model and the nonparanormal maintain tractable inference without
placing limitations on the independence graph. But they are limited in their ability to flexibly model
the bivariate and higher order marginals. At another extreme, forest-structured graphical models
permit arbitrary bivariate marginals, but maintain tractability by restricting to acyclic graphs. An
nonparametric approach based on forests and trees is developed in [7] as a nonparametric method
for estimating the density in high-dimensional settings. However, the ability to model complex
independence graphs is compromised.
In this paper we bring together the Gaussian, nonparanormal, and forest graphical models, using
what we call blossom tree graphical models. Informally, a blossom tree consists of a forest of trees,
and a collection of subgraphs?the blossoms?possibly containing many cycles. The vertex sets
of the blossoms are disjoint, and each blossom contains at most one node of a tree. We estimate
nonparanormal graphical models over the blossoms, and nonparametric bivariate densities over the
branches (edges) of the trees. Using the properties of the nonparanormal, these components can
be combined, or factored, to give a valid joint density for X = (X1 , . . . , Xd ). The details of our
construction are given in Section 2. We develop an estimation procedure for blossom tree graphical models, including an algorithm for selecting tree branches, partition the remaining vertices into
potential blossoms, and then estimating the graphical structures of the blossoms. Since an objective is to relax the Gaussian assumption, our criterion for selecting tree branches is deviation from
Gaussianity. Toward this end, we use the negentropy, showing that it has strong statistical properties
in high dimensions. In order to partition the nodes into blossoms, we employ a nonparametric partial correlation statistic. We use a data-splitting scheme to select the optimal blossom tree structure
based on held-out risk.
In the following section, we present the details of our method, including definitions of blossom tree
graphs, the associated family of graphical models, and our estimation methods. In Sections 3 and 4,
we present experiments with simulated and real data. Finally, we conclude in Section 5. Statistical
properties, detailed proofs, and further experimental results are collected in a supplement.
2
Blossom Tree Graphs and Estimation Methods
To unify the Gaussian, nonparanormal and forest graphical models we make the following definition.
Definition 2.1. A blossom tree on a node set V = {1, 2, . . . , d} is a graph G = (V, E), together
with a decomposition of the edge set E as E = F ? {?B?B B} satisfying the following properties:
1. F is acyclic;
2. V (B) ? V (B 0 ) = ?, for B, B 0 ? B with B 6= B 0 , where V (B) denotes the vertex set of
B.
3. |V (B) ? V (F )| ? 1 for each B ? B;
S
4. V (F ) ? B V (B) = V .
The subgraphs B ? B are called blossoms. The unique node ?(B) ? V (B) ? V (F ), which may be
empty, is called the pedicel of the blossom. The set of pedicels is denoted P(F ) ? V (F ).
Property 1 says that the set of edges F forms a union of trees?a forest. Property 2 says that distinct
blossoms share no vertices or edges in common. Property 3 says that each blossom is connected to at
most one tree node. Property 4 says that every node in the graph is either in a tree or a blossom. Note
that the blossoms are not required to be connected, but must have at most one vertex in common
with the forest?this is the pedicel node.
2
(a) blossom tree
(b) violation
(c) blossom tree
(d) violation
Figure 1: Four graphs, two blossom trees. The tree edges are colored blue, the blossom edges are
colored black, and pedicels are orange. Graphs (a) and (c) correspond to blossom trees. Graphs (b)
and (d) violate the restriction that each blossom has only a single pedicel, or attachment to a tree.
Suppose that p(x) = p(x1 , . . . , xd ) is the density of a distribution that has an independence graph
given by a blossom tree F ? {?B B}. Then from the blossom tree properties we have that
p(x) = p(XV (F ) )
Y
B?B
= p(XV (F ) )
Y
B?B
= p(XV (F ) )
p(XV (B) | XV (F ) )
(2.1)
p(XV (B) | X?(B) )
(2.2)
Y p(XV (B) )
p(X?(B) )
(2.3)
B?B
=
Y
(s,t)?F
=
Y
(s,t)?F
p(Xs , Xt )
p(Xs )p(Xt )
p(Xs , Xt )
p(Xs )p(Xt )
Y
p(Xs )
s?V (F )
Y
s?V (F )\P(F )
Y p(XV (B) )
p(X?(B) )
(2.4)
B?B
p(Xs )
Y
p(XV (B) ).
(2.5)
B?B
The first equality follows from disjointness of the blossoms. The second equality follows from the
existence of a single pedicel node attaching the blossom to a tree. The fourth equality follows from
the standard factorization of forests, and the last equality follows from the fact that each non-empty
pedicel for a blossom is unique. We call the set of distributions that factor in this way the family of
blossom tree graphical models.
A key property of the nonparanormal [5] is that the single node marginal probabilities p(Xs ) are
arbitrary. This property allows us to form graphical models where each blossom distribution satisfies
XV (B) ? NPN(?B , ?B , fB ), while enforcing that the single node marginal of the pedicel ?(B)
agrees with the marginals of this node defined by the forest. This allows us to define and estimate
distributions that are consistent with the factorization (2.5).
Let X (1) , . . . , X (n) be n i.i.d. Rd -valued data vectors sampled from p? (x) where X (l) =
(l)
(l)
(X1 , . . . , Xd ). Our goal is to derive a method for high-dimensional undirected graph estimation and density estimation, using a family of semiparametric estimators based on the blossom tree
structure. Let FB denote the blossom tree structure F ? {?B B}. Our estimation procedure is the
following.
First, randomly partition the data X (1) , . . . , X (n) into two sets D1 and D2 of sample size n1 and n2 .
Then apply the following steps.
1. Using D1 , estimate the bivariate densities p? (xi , xj ) using kernel density estimation. Also,
estimate the covariance ?ij for each pair of variables. Apply Kruskal?s algorithm on the
estimated pairwise negentropy matrix to construct a family of forests {Fb(k) } with k =
0, . . . , d ? 1 edges;
2. Using D1 , for each forest Fb(k) obtained in Step 1, build the blossom tree-structured graph
(k)
FbBb . The forest structure Fb(k) is modeled by nonparametric kernel density estimators,
b (k) is modeled by the graphical lasso or nonparanormal. A family of
while each blossom B
i
3
graphs is obtained by computing regularization paths for the blossoms, using the graphical
lasso.
(b
k)
3. Using D2 , choose FbBb from this family of blossom tree models that maximizes the heldout log-likelihood.
The details of each step are presented below.
2.1 Step 1: Construct A Family of Forests
In information theory and statistics, negentropy is used as a measure of distance to normality. The
negentropy is zero for Gaussian densities and is always nonnegative. The negentropy between variables Xi and Xj is defined as
J(Xi ; Xj ) = H(?(xi , xj )) ? H(p? (xi , xj )),
(2.6)
where H(?) denotes the differential entropy of a density, and ?(xi , xj ) is an Gaussian density with
the same mean and covariance matrix as p? (xi , xj ).
Kruskal?s algorithm [4] is a greedy algorithm to find a maximum weight spanning tree of a weighted
graph. At each step it includes an edge connecting the pair of nodes with the maximum weight
among all unvisited pairs, if doing so does not form a cycle. The algorithm also results in the best
k-edge weighted forest after k < d edges have been included.
In our setting, we define the weight w(i, j) of nodes i and j as the negentropy between Xi and Xj ,
and use Kruskal?s algorithm to build the maximum weight spanning forest Fb(k) with k edges where
k < d. In such a way, the pairs of nodes that are less likely to be a bivariate Gaussian are included
in the forest and then are modeled nonparametrically.
Since the true density p? is unknown, we replace the population negentropy J(Xi ; Xj ) by the estimate
b pn (xi , xj )),
Jbn (Xi ; Xj ) = H(?bn (xi , xj )) ? H(b
(2.7)
1
1
1
where ?bn1 (xi , xj ) is an estimate of the Gaussian density ?(xi , xj ) for Xi and Xj using D1 ,
b
pbn1 (xi , xj ) is a bivariate kernel density estimate for Xi and Xj , and H(?)
denotes the empirical
ij
b ij
differential entropy. In particular, let ? be the covariance matrix of Xi and Xj . Denote ?
n1 as the
empirical covariance matrix of Xi and Xj based on D1 , then the plug-in estimate
1
b ij
(2.8)
H(?bn1 (xi , xj )) = 1 + log(2?) + logdet(?
n1 ).
2
Let K(?) be a univariate kernel function. Then given an evaluation point (xi , xj ), the bivariate kernel
(l)
(l)
density estimate for (Xi , Xj ) based on observations {Xi , Xj }l?D1 is given by
!
!
(l)
(l)
Xj ? xj
1 X
1
Xi ? xi
pbn1 (xi , xj ) =
K
K
,
(2.9)
n1
h2i h2j
h2i
h2j
l?D1
where h2i and h2j are bandwidth parameters for (Xi , Xj ). To compute the empirical differential
b pn (xi , xj )), we numerically evaluate a two-dimensional integral.
entropy H(b
1
h
i
Once the estimated negentropy matrix Jbn1 (Xi ; Xj )
is obtained, we apply Kruskal?s algorithm
d?d
to construct a family of forests {Fb(k) }k=0...d?1 .
2.2 Step 2: Build and Model the Blossom Tree Graphs
Suppose that we have a forest-structured graph F with |V (F )| < d vertices. Then for each remaining non-forest node, we need to determine which blossom it belongs to. We exploit the following
basic fact.
4
Proposition 2.1. Suppose that X ? p? is a density for a blossom tree graphical model with forest
F . Let i 6? V (F ) and s ? V (F ). Then node i is not in a blossom attached to tree node s if and only
if
Xi ?? Xs | Xt for some node t ? V (F ) such that (s, t) ? E(F ).
(2.10)
We use this property, together with a measure of partial correlation, in order to partition the nonforest nodes into blossoms. Partial correlation measures the degree of association between two
random variables, with the effect of a set of controlling random variables removed. Traditionally,
the partial correlation between variables Xi and Xs given a controlling variable Xt is the correlation
between the residuals i\t and s\t resulting from the linear regression of Xi with Xt and of Xs
with Xt , respectively. However, if the underlying joint Gaussian or nonparanormal assumption is
not satisfied, linear regression cannot remove all of the effects of the controlling variable.
We thus use a nonparametric version of partial correlation. Following [1], suppose Xi = g(Xt )+i\t
and Xs = h(Xt )+s\t , for certain functions g and h such that E(i\t | Xt ) = 0 and E(s\t | Xt ) =
0. Define the nonparametric partial correlation as
.q
?is?t = E(i\t s\t )
E(2i\t ) E(2s\t ).
(2.11)
It is shown in [1] that if Xi ?
? Xs | Xt , then ?is?t = 0. We thus conclude the following.
Proposition 2.2. If ?is?t 6= 0 for all t such that (s, t) ? E(F ), node i is in a blossom attached to
node s.
(l)
(l)
(l)
(l)
(l)
Let gb and b
h be local polynomial estimators of g and h, and b
i\t = Xi ? gb(Xt ), b
s\t = Xs ?
(l)
b
h(X ) for any l ? D1 , then an estimate of ?is?t is given by
t
?bis?t =
X
l?D1
.s X
(l) (l)
(b
i\t b
s\t )
(l)
(b
i\t )2
l?D1
X
(l)
(b
s\t )2 .
(2.12)
l?D1
Based on Proposition 2.2, for each forest Fb(k) obtained in Step 1, we then assign each non-forest
node i to the blossom with the pedicel given by
sbi = argmax
min
b(k) )}
b(k) ) {t: (s,t)?E(F
s?V (F
|b
?is?t |.
(2.13)
(k)
After iterating over all non-forest nodes, we obtain a blossom tree-structured graph FbBb . Then
the forest structure is nonparametrically modeled by the bivariate and univariate kernel density estimations, while each blossom is modeled with the graphical lasso or nonparanormal. In particular,
when k = 0 that there is no forest node, our method is reduced to modeling the entire graph by the
graphical lasso or nonparanormal.
Alternative testing procedures based on nonparametric partial correlations could be adopted for partitioning nodes into blossoms. However, such methods may have large computational cost, and low
power for small sample sizes.
Note that while each non-forest node is associated with a pedicel in this step, after graph estimation
for the blossoms, the node may well become disconnected from the forest.
2.3 Step 3: Optimize the Blossom Tree Graphs
(d?1)
The full blossom tree graph FbBb
obtained in Steps 1 and 2 might result in overfitting in the density
estimate. Thus we need to choose an optimal graph with the number of forest edges k ? d ? 1.
Besides, the tuning parameters involved in the fitting of each blossom by the graphical lasso or
nonparanormal also induce a bias-variance tradeoff.
5
(k)
To optimize the blossom tree structures over {FbBb }k=0...d?1 , we choose the complexity parameter
b
k as the one that maximizes the log-likelihood on D2 , using the factorization (2.5):
?
(l)
(l)
X
Y
pbn1 (Xi , Xj )
1
b
log ?
?
k = argmax
(l)
(l)
k?{0,...,d?1} n2 l?D
bn1 (Xi )b
pn1 (Xj )
b(k) ) p
2
(i,j)?E(F
?
k
Y
Y
(k)
?
(l)
(2.14)
pbn1 (Xs(l) )
?bni1 X b(k) ? ,
b(k) )\P(F
b(k) )
s?V (F
(k)
?
i=1
V (Bi
)
where ?bni1 is the density estimate for blossoms by the graphical lasso or nonparanormal, with the
(k)
tuning parameter ?i selected to maximize the held-out log-likelihood. That is, the complexity of
each blossom is also optimized on D2 .
Thus the final blossom tree density estimator is given by
p b(k)
b (x) =
Fb
B
Y
b
b(k)
(i,j)?E(F
)
pbn1 (xi , xj )
pbn1 (xi )b
pn1 (xj )
Y
b
b
b(k)
b(k)
s?V (F
)\P(F
)
pbn1 (Xs(l) )
b
k
Y
i=1
b
(k)
?
?bni1 (x b(k)
b ).
Bi
(2.15)
In practice, Step 3 can be carried out simultaneously with Steps 1 and 2. Whenever a new edge is
added to the current forest in Kruskal?s algorithm, the blossoms are re-constructed and re-modeled.
Then the held-out log-likelihood of the obtained density estimator can be immediately computed.
In addition, since there are no overlapping nodes between different blossoms, the sparsity tuning
parameters are selected separately for each blossom, which reduces the computational cost considerably.
3
Analysis of Simulated Data
Here we present numerical results based on simulations. We compare the blossom tree density
estimator with the graphical lasso [3] and forest density estimator [7]. To evaluate the performance
of these estimators, we compute and compare the log-likelihood of each method on held-out data.
We simulate high-dimensional data which are consistent with an undirected graph. We generate multivariate non-Gaussian data using a sequence of mixtures of two Gaussian distributions with contrary
correlation and equal weights. Then a subset of variables are chosen to generate the blossoms that
are distributed as multivariate Gaussians. In dimensional d = 80, we sample n1 = n2 = 400 data
points from this synthetic distribution.
A typical run showing the held-out log-likelihood and estimated graphs is provided in Figures 2 and
3. The term ?trunk? is used to represent the edge added to the forest in a blossom tree graph. We can
see that the blossom tree density estimator is superior to other methods in terms of generalization
performance. In particular, the graphical lasso is unable to uncover the edges that are generated
nonparametrically. This is expected, since different blossoms have zero correlations among each
other and are thus regarded as independent by the algorithm of graphical lasso. For the modeling of
the variables that are contained in a blossom and are thus multivariate Gaussian distributed, there is
an efficiency loss in the forest density estimator, compared to the graphical lasso. This illustrates the
advantage of blossom tree density estimator. As is seen from the number of selected edges by each
method shown in Figure 2, the blossom tree density estimator selects a graph with a similar sparsity
pattern as the true graph.
4
Analysis of Cell Signalling Data
We analyze a flow cytometry dataset on d = 11 proteins from [9]. A subset of n = 853 cells were
chosen. A nonparanormal transformation was estimated and the observations, for each variable,
6
80
60
50
40
true
glasso
glasso
30
Number of selected edges
70
?108
?109
?110
?113
20
?112
Held out log?likelihood
?111
true
0
20
40
60
80
0
Number of trunks
20
40
60
80
Number of trunks
Figure 2: Results on simulations. Left: Held-out log-likelihood of the graphical lasso (horizontal
dotted line), forest density estimator (horizontal dashed line), and blossom tree density estimator
(circles); Right: Number of selected edges by these methods. The horizontal solid line indicates the
number of edges in the true graph, and the solid triangle indicates the best blossom tree graph. The
first circle for blossom tree refers to the 1-trunk case.
true
true
glasso
(a) true
(b) glasso
glasso
forest
forest
(c) forest
forest?blossom
forest?blossom
(d) blossom tree
Figure 3: Results on simulations. Graph (a) corresponds to the true graph. Graphs (b), (c) and (d)
forest
forest?blossom
forest?blossom
correspond
toforest
the estimated graphs
by the graphical
lasso, forest density estimator, and blossom tree
density estimator, respectively. The tree edges are colored red, and the blossom edges are colored
black.
were replaced by their respective normal scores, subject to a Winsorized truncation [5]. We study
the associations among the proteins using the graphical lasso, forest density estimator, and blossom
tree forest density estimator. The maximum held-out log-likelihood for glasso, forest, and blossom
tree are -14.3, -13.8, and -13.7, respectively. We can see that blossom tree is slighter better than
forest in terms of the generalization performance, both of which outperform glasso. Results of
estimated graphs are provided in Figures 4. When the maximum of held-out log-likelihood curve
is achieved, glasso selects 28 edges, forest selects 7 edges, and blossom tree selects 10 edges. The
two graphs uncovered by forest and blossom tree agree on most edges, although the latter contains
cycles.
5
Conclusion
We have proposed a combination of tree-based graphical models and Gaussian graphical models to
form a new nonparametric approach for high dimensional data. Blossom tree models relax the normality assumption and increase statistical efficiency by modeling the forest with nonparametric kernel density estimators and modeling each blossom with the graphical lasso or nonparanormal. Our
experimental results indicate that this method can be a practical alternative to standard approaches
to graph and density estimation.
7
(a) graph reported in [9]
(b) glasso
(c) forest
(d) blossom tree
Figure 4: Results on cell signalling data. Graph (a) refers to the fitted graph reported in [9]. Graphs
(b), (c) and (d) correspond to the estimated graphs by the graphical lasso, forest density estimator,
and blossom tree density estimator, respectively.
Acknowledgements
Research supported in part by NSF grant IIS-1116730, AFOSR grant FA9550-09-1-0373, ONR
grant N000141210762, and an Amazon AWS in Education Machine Learning Research grant.
References
[1] Wicher Bergsma. A note on the distribution of the partial correlation coefficient with nonparametrically estimated marginal regressions. arXiv:1101.4616, 2011.
[2] T. Tony Cai, Tengyuan Liang, and Harrison H. Zhou. Law of log determinant of sample
covariance matrix and optimal estimation of differential entropy for high-dimensional gaussian
distributions. arXiv:1309.0482, 2013.
[3] Jerome H. Friedman, Trevor Hastie, and Robert Tibshirani. Sparse inverse covariance estimation with the graphical lasso. Biostatistics, 9(3):432?441, 2008.
[4] Joseph B. Kruskal. On the shortest spanning subtree of a graph and the traveling salesman
problem. In Proceedings of the American Mathematical Society, volume 7, pages 48?50,
1956.
[5] Han Liu, John Lafferty, and Larry Wasserman. The nonparanormal: Semiparametric estimation
of high dimensional undirected graphs. Journal of Machine Learning Research, 10:2295?2328,
2009.
[6] Han Liu, Larry Wasserman, and John D. Lafferty. Exponential concentration for mutual information estimation with application to forests. In Advances in Neural Information Processing
Systems (NIPS), 2012.
[7] Han Liu, Min Xu, Haijie Gu, Anupam Gupta, John Lafferty, and Larry Wasserman. Forest
density estimation. Journal of Machine Learning Research, 12:907?951, 2011.
[8] Nicolai Meinshausen and Peter B?uhlmann. High dimensional graphs and variable selection
with the lasso. Annals of Statistics, 34(3), 2006.
[9] Karen Sachs, Omar Perez, Dana Pe?er, Douglas A. Lauffenburger, and Garry P. Nolan.
Causal protein-signaling networks derived from multiparameter single-cell data. Science,
308(5721):523?529, 2003.
[10] Ming Yuan and Yi Lin. Model selection and estimation in the Gaussian graphical model.
Biometrika, 94(1):19?35, 2007.
8
| 5316 |@word determinant:1 version:2 polynomial:1 d2:4 simulation:3 bn:1 covariance:7 decomposition:1 solid:2 liu:4 contains:2 score:1 selecting:2 uncovered:1 nonparanormal:18 existing:1 current:1 nicolai:1 negentropy:8 must:2 john:4 chicago:2 partition:4 numerical:1 remove:1 greedy:1 selected:5 signalling:2 fa9550:1 colored:4 node:29 mathematical:1 constructed:1 differential:4 become:1 yuan:1 consists:1 combine:1 fitting:1 pairwise:1 expected:1 ming:1 provided:2 estimating:2 underlying:1 maximizes:2 biostatistics:1 what:1 developed:2 transformation:1 every:1 xd:6 biometrika:1 partitioning:1 grant:4 positive:1 local:1 xv:10 path:1 black:2 might:1 meinshausen:1 factorization:4 limited:1 bi:3 unique:2 practical:1 testing:1 union:2 practice:1 signaling:1 procedure:3 empirical:3 h2i:3 induce:1 refers:2 protein:3 cannot:1 selection:2 risk:1 restriction:1 optimize:2 imposed:1 missing:2 flexibly:1 focused:1 unify:1 amazon:1 splitting:1 immediately:1 wasserman:3 subgraphs:2 estimator:22 factored:1 regarded:1 population:1 traditionally:1 annals:1 construction:1 suppose:4 controlling:3 satisfying:1 cycle:3 connected:2 sbi:1 removed:1 complexity:2 compromise:1 efficiency:2 triangle:1 gu:1 joint:4 haijie:1 grown:1 distinct:1 encoded:1 valued:1 say:4 relax:2 nolan:1 ability:2 statistic:6 multiparameter:1 final:1 sequence:1 advantage:1 cai:1 achieve:1 empty:2 weakens:1 develop:1 derive:1 ij:5 bn1:3 strong:2 implies:1 indicate:1 larry:3 education:1 assign:1 f1:1 generalization:2 proposition:3 exploring:1 strictly:1 normal:1 exp:2 kruskal:6 estimation:16 uhlmann:1 agrees:1 weighted:2 gaussian:20 always:1 pn:2 zhou:1 derived:1 likelihood:10 indicates:2 inference:2 typically:1 entire:1 transformed:1 selects:4 among:3 denoted:1 copula:1 orange:1 marginal:4 equal:1 construct:3 once:1 mutual:1 represents:1 placing:1 employ:1 randomly:1 simultaneously:1 replaced:1 argmax:2 maintain:2 n1:5 friedman:1 fd:1 evaluation:1 violation:2 mixture:1 extreme:1 perez:1 behind:1 held:9 edge:29 integral:1 partial:9 respective:1 tree:67 re:2 circle:2 causal:1 theoretical:1 fitted:1 modeling:6 tractability:1 cost:2 vertex:7 deviation:1 subset:2 reported:2 connect:1 considerably:1 combined:1 synthetic:1 density:41 fundamental:1 together:3 connecting:1 clifford:1 satisfied:1 containing:1 choose:3 possibly:1 american:1 unvisited:1 potential:2 disjointness:1 gaussianity:3 includes:1 coefficient:1 doing:1 analyze:1 red:1 wicher:1 variance:1 efficiently:1 correspond:3 whenever:2 trevor:1 definition:3 involved:1 associated:3 proof:1 sampled:1 dataset:1 n000141210762:1 uncover:1 higher:1 evaluated:1 correlation:12 jerome:1 traveling:1 horizontal:3 overlapping:1 nonparametrically:4 building:1 effect:2 true:9 equality:4 assigned:1 regularization:1 conditionally:1 criterion:1 demonstrate:1 bring:1 fj:1 common:2 superior:1 attached:2 volume:1 association:2 marginals:3 numerically:1 pn1:2 imposing:1 rd:2 tuning:3 han:3 multivariate:4 bergsma:1 recent:1 belongs:1 certain:1 onr:1 yi:1 seen:1 determine:1 maximize:1 shortest:1 dashed:1 ii:1 branch:4 violate:2 full:1 reduces:1 plug:1 lin:1 involving:1 basic:1 regression:3 arxiv:2 kernel:7 represent:1 achieved:1 cell:4 addition:1 semiparametric:2 separately:1 aws:1 harrison:1 subject:1 undirected:5 contrary:1 lafferty:4 flow:1 effectiveness:1 tengyuan:1 call:2 concerned:1 npn:1 xj:34 independence:3 hastie:1 lasso:19 bandwidth:1 idea:1 tradeoff:1 expression:1 gb:2 peter:1 karen:1 logdet:1 useful:1 iterating:1 detailed:1 informally:1 nonparametric:14 reduced:1 generate:2 outperform:1 nsf:1 dotted:1 estimated:9 disjoint:2 tibshirani:1 blue:1 group:2 key:1 four:1 douglas:1 graph:51 run:1 inverse:1 fourth:1 family:9 nonnegative:1 simulate:1 min:2 department:3 structured:4 combination:1 disconnected:1 partitioned:1 joseph:1 computationally:1 agree:1 trunk:4 tractable:2 end:1 adopted:1 salesman:1 gaussians:1 lauffenburger:1 permit:1 apply:3 alternative:2 anupam:1 existence:1 denotes:5 remaining:2 tony:1 graphical:39 xc:4 exploit:1 build:3 society:1 objective:1 added:2 concentration:1 distance:1 unable:1 simulated:3 omar:1 collected:1 toward:1 enforcing:1 spanning:3 besides:1 modeled:6 liang:1 robert:1 unknown:1 observation:2 defining:1 arbitrary:5 cytometry:1 pair:4 required:1 optimized:1 established:1 nip:1 below:1 pattern:2 sparsity:3 hammersley:1 including:2 power:1 difficulty:1 attach:1 residual:1 normality:2 scheme:1 h2j:3 attachment:1 carried:1 acknowledgement:1 garry:1 afosr:1 law:1 loss:1 glasso:9 heldout:1 limitation:1 acyclic:2 dana:1 degree:1 consistent:2 share:1 supported:1 last:1 truncation:1 bias:1 blossom:92 weaker:1 attaching:1 sparse:2 distributed:2 curve:1 default:1 dimension:1 valid:1 rich:1 fb:9 collection:2 made:1 clique:1 overfitting:1 winsorized:1 conclude:2 xi:40 zhe:1 continuous:2 compromised:1 forest:53 complex:1 sachs:1 n2:3 x1:6 xu:1 precision:2 exponential:1 pe:1 theorem:1 xt:14 showing:2 er:1 x:15 gupta:1 normalizing:1 bivariate:8 intractable:1 restricting:1 supplement:1 subtree:1 illustrates:1 entropy:4 fc:3 likely:1 univariate:2 contained:1 monotonic:1 corresponds:1 satisfies:1 goal:1 replace:1 included:2 specifically:1 typical:1 called:2 experimental:2 select:1 latter:1 evaluate:2 d1:11 |
4,768 | 5,317 | Distributed Parameter Estimation
in Probabilistic Graphical Models
Yariv D. Mizrahi1 Misha Denil2 Nando de Freitas2,3,4
1
University of British Columbia, Canada
2
University of Oxford, United Kingdom
3
Canadian Institute for Advanced Research
4
Google DeepMind
[email protected]
{misha.denil,nando}@cs.ox.ac.uk
Abstract
This paper presents foundational theoretical results on distributed parameter estimation for undirected probabilistic graphical models. It introduces a general
condition on composite likelihood decompositions of these models which guarantees the global consistency of distributed estimators, provided the local estimators
are consistent.
1
Introduction
Undirected probabilistic graphical models, also known as Markov Random Fields (MRFs), are a
natural framework for modelling in networks, such as sensor networks and social networks [24, 11,
20]. In large-scale domains there is great interest in designing distributed learning algorithms to
estimate parameters of these models from data [27, 13, 19]. Designing distributed algorithms in
this setting is challenging because the distribution over variables in an MRF depends on the global
structure of the model.
In this paper we make several theoretical contributions to the design of algorithms for distributed
parameter estimation in MRFs by showing how the recent works of Liu and Ihler [13] and of Mizrahi
et al. [19] can both be seen as special cases of distributed composite likelihood. Casting these two
works in a common framework allows us to transfer results between them, strengthening the results
of both works.
Mizrahi et al. introduced a theoretical result, known as the LAP condition, to show that it is possible
to learn MRFs with untied parameters in a fully-parallel but globally consistent manner. Their result
led to the construction of a globally consistent estimator, whose cost is linear in the number of cliques
as opposed to exponential as in centralised maximum likelihood estimators. While remarkable, their
results apply only to a specific factorisation, with the cost of learning being exponential in the size of
the factors. While their factors are small for lattice-MRFs and other models of low degree, they can
be as large as the original graph for other models, such as fully-observed Boltzmann machines [1]. In
this paper, we introduce the Strong LAP Condition, which characterises a large class of composite
likelihood factorisations for which it is possible to obtain global consistency, provided the local
estimators are consistent. This much stronger condition enables us to construct linear and globally
consistent distributed estimators for a much wider class of models than Mizrahi et al., including
fully-connected Boltzmann machines.
Using our framework we also show how the asymptotic theory of Liu and Ihler applies more generally to distributed composite likelihood estimators. In particular, the Strong LAP Condition provides
a sufficient condition to guarantee the validity of a core assumption made in the theory of Liu and
Ihler, namely that each local estimate for the parameter of a clique is a consistent estimator of the
1
1
2
3
1
2
3
1
2
3
1
2
3
4
5
6
4
5
6
4
5
6
4
5
6
7
8
9
7
8
9
7
8
9
7
8
9
Figure 1: Left: A simple 2d-lattice MRF to illustrate our notation. For node j = 7 we have N (xj ) =
{x4 , x8 }. Centre left: The 1-neighbourhood of the clique q = {x7 , x8 } including additional edges
(dashed lines) present in the marginal over the 1-neighbourhood. Factors of this form are used by
the LAP algorithm of Mizrahi et. al. Centre right: The MRF used by our conditional estimator of
Section 5 when using the same domain as Mizrahi et. al. Right: A smaller neighbourhood which
we show is also sufficient to estimate the clique parameter of q.
corresponding clique parameter in the joint distribution. By applying the Strong LAP Condition to
verify the assumption of Liu and Ihler, we are able to import their M-estimation results into the LAP
framework directly, bridging the gap between LAP and consensus estimators.
2
Background
Our goal is to estimate the D-dimensional parameter vector ? of an MRF with the following Gibbs
density or mass function:
X
1
p(x | ?) =
exp(
E(xc | ? c ))
(1)
Z(?)
c
Here c 2 C is an index over the cliques of an undirected graph G = (V, E), E(xc | ? c ) is known as
the energy or Gibbs potential, and Z(?) is a normalizing term known as the partition function.
When E(xc | ? c ) = ? Tc c (xc ), where c (xc ) is a local sufficient statistic derived from the values
of the local data vector xc , this model is known as a maximum entropy or log-linear model. In
this paper we do not restrict ourselves to a specific form for the potentials, leaving them as general
functions; we require only that their parameters are identifiable. Throughout this paper we focus
on the case where the xj ?s are discrete random variables, however generalising our results to the
continuous case is straightforward.
The j-th node of G is associated with the random variable xj for j = 1, . . . , M , and the edge connecting nodes j and k represents the statistical interaction between xj and xk . By the HammersleyClifford Theorem [10], the random vector x satisfies the Markov property with respect to the graph
G, i.e., p(xj |x j ) = p(xj |xN (xj ) ) for all j where x j denotes all variables in x excluding xj ,
and xN (xj ) are the variables in the neighbourhood of node j (variables associated with nodes in G
directly connected to node j).
2.1
Centralised estimation
The standard approach to parameter estimation in statistics is through maximum likelihood, which
chooses parameters ? by maximising
LM L (?) =
N
Y
n=1
p(xn | ?)
(2)
(To keep the notation light, we reserve n to index the data samples. In particular, xn denotes the
n-th |V|-dimensional data vector and xmn refers to the n-th observation of node m.)
This estimator has played a central role in statistics as it has many desirable properties including
consistency, efficiency and asymptotic normality. However, applying maximum likelihood estimation to an MRF is generally intractable since computing the value of log LM L and its derivative
require evaluating the partition function, and an expectation over the model, respectively. Both of
these values involve a sum over exponentially many terms.
2
To surmount this difficulty it is common to approximate p(x | ?) as a product over more tractable
terms. This approach is known as composite likelihood and leads to an objective of the form
LCL (?) =
N Y
I
Y
f i (xn , ? i )
(3)
n=1 i=1
where ? i denote the (possibly shared) parameters of each composite likelihood factor f i .
Composite likelihood estimators are both well studied and widely applied [6, 14, 12, 7, 16, 2, 22,
4, 21]. In practice the f i terms are chosen to be easy to compute, and are typically local functions,
depending only on some local region of the underlying graph G.
An early and influential variant of composite likelihood is pseudo-likelihood (PL) [3], where
f i (x, ? i ) is chosen to be the conditional distribution of xi given its neighbours,
LP L (?) =
N Y
M
Y
n=1 m=1
p(xmn | xN (xm )n , ? m )
(4)
Since the joint distribution has a Markov structure with respect to the graph G, the conditional
distribution for xm depends only on its neighbours, namely xN (xm ) . In general more statistically
efficient composite likelihood estimators can be obtained by blocking, i.e. choosing the f i (x, ? i ) to
be conditional or marginal likelihoods over blocks of variables, which may be allowed to overlap.
Composite likelihood estimators are often divided into conditional and marginal variants, depending
on whether the f i (x, ? i ) are formed from conditional or marginal likelihoods. In machine learning
the conditional variant is quite popular [12, 7, 16, 15, 4] while the marginal variant has received less
attention. In statistics, both the marginal and conditional variants of composite likelihood are well
studied (see the comprehensive review of Varin et. al. [26]).
An unfortunate difficulty with composite likelihood is that the estimators cannot be computed in
parallel, since elements of ? are often shared between the different factors. For a fixed value of ?
the terms of log LCL decouple over data and over blocks of the decomposition; however, if ? is not
fixed then the terms remain coupled.
2.2
Consensus estimation
Seeking greater parallelism, researchers have investigated methods for decoupling the sub-problems
in composite likelihood. This leads to the class of consensus estimators, which perform parameter
estimation independently in each composite likelihood factor. This approach results in parameters
that are shared between factors being estimated multiple times, and a final consensus step is required
to force agreement between the solutions from separate sub-problems [27, 13].
Centralised estimators enforce sub-problem agreement throughout the estimation process, requiring
many rounds of communication in a distributed setting. Consensus estimators allow sub-problems
to disagree during optimisation, enforcing agreement as a post-processing step which requires only
a single round of communication.
Liu and Ihler [13] approach distributed composite likelihood by optimising each term separately
!
N
Y
i
? = arg max
?
f i (xAi ,n , ? )
(5)
i
?
i
i
n=1
where Ai denotes the group of variables associated with block i, and ? i is the corresponding set of
parameters. In this setting the sets i ? V are allowed to overlap, but the optimisations are carried
out independently, so multiple estimates for overlapping parameters are obtained. Following Liu
and Ihler we have used the notation ? i = ? i to make this interdependence between factors explicit.
i
? into a degenerate estiThe analysis of this setting proceeds by embedding each local estimator ?
i
i
i
?
?
mator ? for the global parameter vector ? by setting ? c = 0 for c 2
/ i . The degenerate estimators
are combined into a single non-degenerate global estimate using different consensus operators, e.g.
?i.
weighted averages of the ?
3
The analysis of Liu and Ihler assumes that for each sub-problem i and for each c 2
i
p
? )c ! ? c
(?
i
i
(6)
i.e., each local estimate for the parameter of clique c is a consistent estimator of the corresponding
clique parameter in the joint distribution. This assumption does not hold in general, and one of the
contributions of this work is to give a general condition under which this assumption holds.
The analysis of Liu and Ihler [13] considers the case where the local estimators in Equation 5 are arbitrary M -estimators [25], however their experiments address only the case of pseudo-likelihood. In
Section 5 we prove that the factorisation used by pseudo-likelihood satisfies Equation 6, explaining
the good results in their experiments.
2.3
Distributed estimation
Consensus estimation dramatically increases the parallelism of composite likelihood estimates by
relaxing the requirements on enforcing agreement between coupled sub-problems. Recently Mizrahi
et. al. [19] have shown that if the composite likelihood factorisation is constructed correctly then
consistent parameter estimates can be obtained without requiring a consensus step.
In the LAP algorithm of Mizrahi et al. [19] the domain of each composite likelihood factor (which
they call the auxiliary MRF) is constructed by surrounding each maximal clique q with the variables
in its 1-neighbourhood
[
Aq =
c
c\q6=;
which contains all of the variables of q itself as well as the variables with at least one neighbour in
q; see Figure 1 for an example. For MRFs of low degree the sets Aq are small, and consequently
maximum likelihood estimates for parameters of MRFs over these sets can be obtained efficiently.
The parametric form of each factor in LAP is chosen to coincide with the marginal distribution over
Aq .
The factorisation of Mizrahi et al. is essentially the same as in Equation 5, but the domain of each
term is carefully selected, and the LAP theorems are proved only for the case where f i (xAq , ? q ) =
p(xAq , ? q ).
As in consensus estimation, parameter estimation in LAP is performed separately and in parallel for
each term; however, agreement between sub-problems is handled differently. Instead of combining
parameter estimates from different sub-problems, LAP designates a specific sub-problem as authoritative for each parameter (in particular the sub-problem with domain Aq is authoritative for the
parameter ? q ). The global solution is constructed by collecting parameters from each sub-problem
for which it is authoritative and discarding the rest.
In order to obtain consistency for LAP, Mizrahi et al. [19] assume that both the joint distribution and
each composite likelihood factor are parametrised using normalized potentials.
Definition 1. A Gibbs potential E(xc |? c ) is said to be normalised with respect to zero if E(xc |? c ) =
0 whenever there exists t 2 c such that xt = 0.
A perhaps under-appreciated existence and uniqueness theorem [9, 5] for MRFs states that there
exists one and only one potential normalized with respect to zero corresponding to a Gibbs distribution. This result ensures a one to one correspondence between Gibbs distributions and normalised
potential representations of an MRF.
The consistency of LAP relies on the following observation. Suppose we have a Gibbs distribution
p(xV | ?) that factors according to the clique system C, and suppose that the parametrisation is
chosen so that the potentials are normalised with respect to zero. For a particular clique of interest
q, the marginal over xAq can be written as follows (see Appendix A for a detailed derivation)
p(xAq | ?) =
1
exp( E(xq | ? q )
Z(?)
4
X
c2Cq \{q}
E(xc | ? V\q ))
(7)
where Cq denotes the clique system of the marginal, which in general includes cliques not present in
the joint. The same distribution can also be written in terms of different parameters ?
X
1
p(xAq | ?) =
exp( E(xq | ?q )
E(xc | ?c ))
(8)
Z(?)
c2Cq \{q}
which are also assumed to be normalised with respect to zero. As shown in Mizrahi et. al. [19], the
uniqueness of normalised potentials can be used to obtain the following result.
Proposition 2 (LAP argument [19]). If the parametrisations of p(xV | ?) and p(xAq | ?) are chosen to be normalized with respect to zero, and if the parameters are identifiable with respect to the
potentials, then ? q = ?q .
This proposition enables Mizrahi et. al. [19] to obtain consistency for LAP under the standard
smoothness and identifiability assumptions for MRFs [8].
3
Contributions of this paper
The strength of the results of Mizrahi et al. [19] is to show that it is possible to perform parameter
estimation in a completely distributed way without sacrificing global consistency. They prove that
through careful design of a composite likelihood factorisation it is possible to obtain estimates for
each parameter of the joint distribution in isolation, without requiring even a final consensus step
to enforce sub-problem agreement. Their weakness is that the LAP algorithm is very restrictive,
requiring a specific composite likelihood factorisation.
The strength of the results of Liu and Ihler [13] is that they apply in a very general setting (arbitrary
M -estimators) and make no assumptions about the underlying structure of the MRF. On the other
hand they assume the convergence in Equation 6, and do not characterise the conditions under which
this assumption holds.
The key to unifying these works is to notice that the specific decomposition used in LAP is chosen
essentially to ensure the convergence of Equation 6. This leads to our development of the Strong
LAP Condition and an associated Strong LAP Argument, which is a drop in replacement for the LAP
argument of Mizrahi et al. and holds for a much larger range of composite likelihood factorisations
than their original proof allows.
Since the purpose of the Strong LAP Condition is to guarantee the convergence of Equation 6, we
are able to import the results of Liu and Ihler [13] into the LAP framework directly, bridging the
gap between LAP and consensus estimators. The same Strong LAP Condition also provides the
necessary convergence guarantee for the results of Liu and Ihler to apply.
Finally we show how the Strong LAP Condition can lead to the development of new estimators, by
developing a new distributed estimator which subsumes the distributed pseudo-likelihood and gives
estimates that are both consistent and asymptotically normal.
4
Strong LAP argument
In this section we present the Strong LAP Condition, which provides a general condition under
which the convergence of Equation 6 holds. This turns out to be intimately connected to the structure
of the underlying graph.
Definition 3 (Relative Path Connectivity). Let G = (V, E) be an undirected graph, and let A be a
given subset of V. We say that two nodes i, j 2 A are path connected with respect to V \ A if there
exists a path P = {i, s1 , s2 , . . . , sn , j} =
6 {i, j} with none of the sk 2 A. Otherwise, we say that
i, j are path disconnected with respect to V \ A.
in
For a given A ? V we partition the clique system of G into two parts, CA
that contains all of the
out
in
cliques that are a subset of A, and CA = C \ CA that contains the remaining cliques of G. Using
this notation we can write the marginal distribution over xA as
X
X
X
1
p(xA | ?) =
exp(
E(xc | ? c ))
exp(
E(xc | ? c ))
(9)
Z(?)
out
in
x
c2CA
V\A
5
c2CA
(b)
(a)
1
3
6
4
2
j
k
5
i
(c)
2
3
(d)
2
3
0
1
4
0
1
4
5
2
3
0
1
4
5
5
Figure 2: (a) Illustrating the concept of relative path connectivity. Here, A = {i, j, k}. While (k, j)
are path connected via {3, 4} and (k, i) are path connected via {2, 1, 5}, the pair (i, j) are path
disconnected with respect to V \ A. (b)-(d) Illustrating the difference between LAP and Strong LAP.
(b) Shows a star graph with q highlighted. (c) Shows Aq required by LAP. (d) Shows an alternative
neighbourhood allowed by Strong LAP. Thus, if the root node is a response variable and the leafs
are covariates, Strong LAP states we can estimate each parameter separately and consistently.
P
P
Up to a normalisation constant, xV\A exp(
out E(xc | ? c )) induces a Gibbs density (and
c2CA
therefore an MRF) on A, which we refer to as the induced MRF. (For example, as illustrated in
Figure 1 centre-left, the induced MRF involves all the cliques over the nodes 4, 5 and 9.) By the
Hammersley-Clifford theorem this MRF has a corresponding graph which we refer to as the induced
graph and denote GA . Note that the induced graph does not have the same structure as the marginal,
it contains only edges which are created by summing over xV\A .
Remark 4. To work in the general case, we assume throughout that that if an MRF contains the
path {i, j, k} then summing over j creates the edge (i, k) in the marginal.
Proposition 5. Let A be a subset of V, and let i, j 2 A. The edge (i, j) exists in the induced graph
GA if and only if i and j are path connected with respect to V \ A.
Proof. If i and j are path connected then there is a path P = {i, s1 , s2 , . . . , sn , j} =
6 {i, j} with
none of the sk 2 A. Summing over sk forms an edge (sk 1 , sk+1 ). By induction, summing over
s1 , . . . , sn forms the edge (i, j).
If i and j are path disconnected with respect to V \ A then summing over any s 2 V \ A cannot
form the edge (i, j) or i and j would be path connected through the path {i, s, j}. By induction, if
the edge (i, j) is formed by summing over s1 , . . . , sn this implies that i and j are path connected via
{i, s1 , . . . , sn , j}, contradicting the assumption.
Corollary 6. B ? A is a clique in the induced graph GA if and only if all pairs of nodes in B are
path connected with respect to V \ A.
Definition 7 (Strong LAP condition). Let G = (V, E) be an undirected graph and let q 2 C be a
clique of interest. We say that a set A such that q ? A ? V satisfies the strong LAP condition for q
if there exist i, j 2 q such that i and j are path-disconnected with respect to V \ A.
Proposition 8. Let G = (V, E) be an undirected graph and let q 2 C be a clique of interest. If
Aq satisfies the Strong LAP condition for q then the joint distribution p(xV | ?) and the marginal
p(xAq | ?) share the same normalised potential for q.
Proof. If Aq satisfies the Strong LAP Condition for q then by Corollary 6 the induced MRF contains
no potential for q. Inspection of Equation 9 reveals that the same E(xq | ? q ) appears as a potential
in both the marginal and the joint distributions. The result follows by uniqueness of the normalised
potential representation.
We now restrict our attention to a set Aq which satisfies the Strong LAP Condition for a clique of
interest q. The marginal over p(xAq | ?) can be written as in Equation 9 in terms of ?, or in terms of
auxiliary parameters ?
X
1
p(xAq | ?) =
exp(
E(xc | ?c ))
(10)
Z(?)
c2Cq
Where Cq is the clique system over the marginal. We will assume both parametrisations are normalised with respect to zero.
Theorem 9 (Strong LAP Argument). Let q be a clique in G and let q ? Aq ? V. Suppose p(xV | ?)
and p(xAq | ?) are parametrised so that their potentials are normalised with respect to zero and the
parameters are identifiable with respect to the potentials. If Aq satisfies the Strong LAP Condition
for q then ? q = ?q .
6
Proof. From Proposition 8 we know that p(xV | ?) and p(xAq | ?) share the same clique potential
for q. Alternatively we can write the marginal distribution as in Equation 10 in terms of auxiliary
variables ?. By uniqueness, both parametrisations must have the same normalised potentials.
Since the potentials are equal, we can match terms between the two parametrisations. In particular
since E(xq | ? q ) = E(xq | ?q ) we see that ? q = ?q by identifiability.
4.1
Efficiency and the choice of decomposition
Theorem 9 implies that distributed composite likelihood is consistent for a wide class of decompositions of the joint distribution; however it does not address the issue of statistical efficiency.
This question has been studied empirically in the work of Meng et. al. [17, 18], who introduce a
distributed algorithm for Gaussian random fields and consider neighbourhoods of different sizes.
Meng et. al. find the larger neighbourhoods produce better empirical results and the following theorem confirms this observation.
Theorem 10. Let A be set of nodes which satisfies the Strong LAP Condition for q. Let ??A be the
ML parameter estimate of the marginal over A. If B is a superset of A, and ??B is the ML parameter
estimate of the marginal over B. Then (asymptotically):
|?q
(??B )q | ? |?q
(??A )q |.
Proof. Suppose that |?q (??B )q | > |?q (??A )q |. Then the estimates ??A over the various subsets A
of B improve upon the ML estimates of the marginal on B. This contradicts the Cramer-Rao lower
bound achieved by the ML estimate of the marginal on B.
In general the choice of decomposition implies a trade-off in computational and statistical efficiency.
Larger factors are preferable from a statistical efficiency standpoint, but increase computation and
decrease the degree of parallelism.
5
Conditional LAP
The Strong LAP Argument tells us that if we construct composite likelihood factors using marginal
distributions over domains that satisfy the Strong LAP Condition then the LAP algorithm of Mizrahi
et. al. [19] remains consistent. In this section we show that more can be achieved.
Once we have satisfied the Strong LAP Condition we know it is acceptable to match parameters
between the joint distribution p(xV | ?) and the auxiliary distribution p(xAq | ?). To obtain a consistent LAP algorithm from this correspondence all that is required is to have a consistent estimate
of ?q . Mizrahi et. al. [19] achieve this by applying maximum likelihood estimation to p(xAq | ?),
but any consistent estimator is valid.
We exploit this fact to show how the Strong LAP Argument can be applied to create a consistent
conditional LAP algorithm, where conditional estimation is performed in each auxiliary MRF. This
allows us to apply the LAP methodology to a broader class of models. For some models, such as
large densely connected graphs, we cannot rely on the LAP algorithm of Mizrahi et. al. [19]. For
example, for a restricted Boltzmann machine (RBM) [23], the 1-neighbourhood of any pairwise
clique includes the entire graph. Hence, the complexity of LAP is exponential in the size of V.
However, it is linear for conditional LAP, without sacrificing consistency.
Theorem 11. Let q be a clique in G and let xj 2 q ? Aq ? V. If Aq satisfies the Strong LAP
Condition for q then p(xV | ?) and p(xj | xAq \{xj } , ?) share the same normalised potential for q.
Proof. We can write the conditional distribution of xj given Aq \ {xj } as
p(xAq | ?)
p(xj | xAq \{xj } , ?) = P
xj p(xAq | ?)
(11)
Both the numerator and the denominator of Equation 11 are Gibbs distributions, and can therefore
be expressed in terms of potentials over clique systems.
7
Since Aq satisfies the Strong LAP Condition for qPwe know that p(xAq | ?) and p(xV | ?) have the
same potential for q. Moreover, the domain of xj p(xAq | ?) does not include q, so it cannot
contain a potential for q. We conclude that the potential for q in p(xj | xAq \{xj } , ?) must be shared
with p(xV | ?).
Remark 12. There exists a Gibbs representation normalised with respect to zero for
p(xj | xAq \{xj } , ?). Moreover, the clique potential for q is unique in that representation.
Existence in the above remark is an immediate result of the the existence of normalized representation both for the numerator and denominator of Equation 11, and the fact that difference
of normalised potentials is a normalized potential. For uniqueness, first note that p(xAq | ?) =
p(xj | xAq \{xj } , ?)p(xAq \{xj } , ?) The variable xj is not part of p(xAq \{xj } , ?) and hence this distribution does not contain the clique q. Suppose there were two different normalised representations
for the conditional p(xj | xAq \{xj } , ?). This would then imply two normalised representations for
the joint, which contradicts the fact that the joint has a unique normalized representation.
We can now proceed as in the original LAP construction from Mizrahi et al. [19]. For a clique of
interest q we find a set Aq which satisfies the Strong LAP Condition for q. However, instead of
creating an auxiliary parametrisation of the marginal we create an auxiliary parametrisation of the
conditional in Equation 11.
X
1
p(xj | xAq \{xj } , ?) =
exp(
E(xc | ?c ))
(12)
Zj (?)
c2CAq
From Theorem 11 we know that E(xq | ?q ) = E(xq | ? q ). Equality of the parameters is also
obtained, provided they are identifiable.
Corollary 13. If Aq satisfies the Strong LAP Condition for q then any consistent estimator of ?q in
p(xj | xAq \{xj } , ?) is also a consistent estimator of ? q in p(xV | ?).
5.1
Connection to distributed pseudo-likelihood and composite likelihood
Theorem 11 tells us that if Aq satisfies the Strong LAP Condition for q then to estimate ? q in
p(xV | ?) it is sufficient to have an estimate of ?q in p(xj | xAq \{xj } , ?) for any xj 2 q. This tells
us that it is sufficient to use pseudo-likelihood-like conditional factors, provided that their domains
satisfy the Strong LAP Condition. The following remark completes the connection by telling us
that the Strong LAP Condition is satisfied by the specific domains used in the pseudo-likelihood
factorisation.
Remark 14. Let q = {x1 , x2 , .., xm } be a clique of interest, with 1-neighbourhood Aq = q [
{N (xi )}xi 2q . Then for any xj 2 q, the set q [ N (xj ) satisfies the Strong LAP Condition for q.
Moreover, q [ N (xj ) satisfies the Strong LAP Condition for all cliques in the graph that contain xj .
Importantly, to estimate every unary clique potential we need to visit each node in the graph. However, to estimate pairwise clique potentials, visiting all nodes is redundant because the parameters of
each pairwise clique are estimated twice. If a parameter is estimated more than once it is reasonable
from a statistical standpoint to apply a consensus operator to obtain a single estimate. The theory of
Liu and Ihler tells us that the consensus estimates are consistent and asymptotically normal, provided
Equation 6 is satisfied. In turn, the Strong LAP Condition guarantees the convergence in Equation 6.
We can go beyond pseudo-likelihood and consider either marginal or conditional factorisations over
larger groups of variables. Since the asymptotic results of Liu and Ihler [13] apply to any distributed composite likelihood estimator where the convergence of Equation 6 holds, it follows that
any distributed composite likelihood estimator where each factor satisfies the Strong LAP Condition
(including LAP and the conditional composite likelihood estimator from Section 5) immediately
gains asymptotic normality and variance guarantees as a result of their work and ours.
6
Conclusion
We presented foundational theoretical results for distributed composite likelihood. The results provide us with sufficient conditions to apply the results of Liu and Ihler to a broad class of distributed
estimators. The theory also led us to the construction of a new globally consistent estimator, whose
complexity is linear even for many densely connected graphs. We view extending these results to
model selection, tied parameters, models with latent variables, and inference tasks as very important
avenues for future research.
8
References
[1] D. H. Ackley, G. Hinton, and T. Sejnowski. A learning algorithm for Boltzmann machines. Cognitive
Science, 9:147?169, 1985.
[2] A. Asuncion, Q. Liu, A. Ihler, and P. Smyth. Learning with blocks: Composite likelihood and contrastive
divergence. In Artificial Intelligence and Statistics, pages 33?40, 2010.
[3] J. Besag. Spatial interaction and the statistical analysis of lattice systems. Journal of the Royal Statistical
Society, Series B, 36:192?236, 1974.
[4] J. K. Bradley and C. Guestrin. Sample complexity of composite likelihood. In Artificial Intelligence and
Statistics, pages 136?160, 2012.
[5] P. Bremaud. Markov Chains: Gibbs Fields, Monte Carlo Simulation, and Queues. Springer-Verlag, 2001.
[6] B. Cox. Composite likelihood methods. Contemporary Mathematics, 80:221?239, 1988.
[7] J. V. Dillon and G. Lebanon. Stochastic composite likelihood. Journal of Machine Learning Research,
11:2597?2633, 2010.
[8] S. E. Fienberg and A. Rinaldo. Maximum likelihood estimation in log-linear models. The Annals of
Statistics, 40(2):996?1023, 2012.
[9] D. Griffeath. Introduction to random fields. In Denumerable Markov Chains, volume 40 of Graduate
Texts in Mathematics, pages 425?458. Springer, 1976.
[10] J. M. Hammersley and P. Clifford. Markov fields on finite graphs and lattices. 1971.
[11] D. Koller and N. Friedman. Probabilistic Graphical Models: Principles and Techniques. MIT Press,
2009.
[12] P. Liang and M. I. Jordan. An asymptotic analysis of generative, discriminative, and pseudolikelihood
estimators. In International Conference on Machine Learning, pages 584?591, 2008.
[13] Q. Liu and A. Ihler. Distributed parameter estimation via pseudo-likelihood. In International Conference
on Machine Learning, 2012.
[14] K. V. Mardia, J. T. Kent, G. Hughes, and C. C. Taylor. Maximum likelihood estimation using composite
likelihoods for closed exponential families. Biometrika, 96(4):975?982, 2009.
[15] B. Marlin and N. de Freitas. Asymptotic efficiency of deterministic estimators for discrete energy-based
models: Ratio matching and pseudolikelihood. In Uncertainty in Artificial Intelligence, pages 497?505,
2011.
[16] B. Marlin, K. Swersky, B. Chen, and N. de Freitas. Inductive principles for restricted Boltzmann machine
learning. In Artificial Intelligence and Statistics, pages 509?516, 2010.
[17] Z. Meng, D. Wei, A. Wiesel, and A. O. Hero III. Distributed learning of Gaussian graphical models via
marginal likelihoods. In Artificial Intelligence and Statistics, pages 39?47, 2013.
[18] Z. Meng, D. Wei, A. Wiesel, and A. O. Hero III. Marginal likelihoods for distributed parameter estimation
of Gaussian graphical models. Technical report, arXiv:1303.4756, 2014.
[19] Y. Mizrahi, M. Denil, and N. de Freitas. Linear and parallel learning of Markov random fields. In
International Conference on Machine Learning, 2014.
[20] K. P. Murphy. Machine Learning: A Probabilistic Perspective. The MIT Press, 2012.
[21] S. Nowozin. Constructing composite likelihoods in general random fields. In ICML Workshop on Inferning: Interactions between Inference and Learning, 2013.
[22] S. Okabayashi, L. Johnson, and C. Geyer. Extending pseudo-likelihood for Potts models. Statistica Sinica,
21(1):331?347, 2011.
[23] P. Smolensky. Information processing in dynamical systems: foundations of harmony theory. Parallel
distributed processing: explorations in the microstructure of cognition, 1:194?281, 1986.
[24] D. Strauss and M. Ikeda. Pseudolikelihood estimation for social networks. Journal of the American
Statistical Association, 85(409):204?212, 1990.
[25] A. W. van der Vaart. Asymptotic statistics. Cambridge University Press, 1998.
[26] C. Varin, N. Reid, and D. Firth. An overview of composite likelihood methods. Statistica Sinica, 21:5?42,
2011.
[27] A. Wiesel and A. Hero III. Distributed covariance estimation in Gaussian graphical models. IEEE Transactions on Signal Processing, 60(1):211?220, 2012.
9
| 5317 |@word illustrating:2 cox:1 wiesel:3 stronger:1 confirms:1 simulation:1 decomposition:6 kent:1 contrastive:1 covariance:1 liu:16 contains:6 series:1 united:1 ours:1 freitas:3 bradley:1 import:2 written:3 must:2 ikeda:1 partition:3 enables:2 drop:1 intelligence:5 selected:1 leaf:1 generative:1 inspection:1 xk:1 geyer:1 core:1 provides:3 math:1 node:14 constructed:3 prove:2 introduce:2 manner:1 interdependence:1 pairwise:3 globally:4 provided:5 notation:4 underlying:3 moreover:3 mass:1 denumerable:1 deepmind:1 marlin:2 guarantee:6 pseudo:10 every:1 collecting:1 preferable:1 biometrika:1 uk:1 reid:1 local:10 xv:13 oxford:1 meng:4 path:18 twice:1 studied:3 challenging:1 relaxing:1 range:1 statistically:1 graduate:1 unique:2 yariv:2 practice:1 block:4 hughes:1 foundational:2 empirical:1 composite:36 matching:1 okabayashi:1 refers:1 cannot:4 ga:3 selection:1 operator:2 applying:3 deterministic:1 straightforward:1 attention:2 go:1 independently:2 immediately:1 factorisation:10 estimator:37 importantly:1 embedding:1 annals:1 construction:3 suppose:5 smyth:1 designing:2 agreement:6 element:1 blocking:1 observed:1 role:1 ackley:1 region:1 ensures:1 connected:13 trade:1 decrease:1 contemporary:1 complexity:3 covariates:1 creates:1 upon:1 efficiency:6 completely:1 joint:12 differently:1 various:1 surrounding:1 derivation:1 sejnowski:1 monte:1 artificial:5 varin:2 tell:4 choosing:1 lcl:2 whose:2 quite:1 widely:1 larger:4 say:3 otherwise:1 statistic:10 vaart:1 highlighted:1 itself:1 final:2 interaction:3 product:1 strengthening:1 maximal:1 combining:1 degenerate:3 achieve:1 convergence:7 requirement:1 extending:2 produce:1 wider:1 illustrate:1 depending:2 ac:1 received:1 strong:36 auxiliary:7 c:1 involves:1 implies:3 stochastic:1 exploration:1 nando:2 require:2 microstructure:1 proposition:5 pl:1 hold:6 cramer:1 normal:2 exp:8 great:1 cognition:1 lm:2 reserve:1 early:1 purpose:1 uniqueness:5 estimation:23 harmony:1 mizrahi:18 create:2 weighted:1 mit:2 sensor:1 gaussian:4 denil:2 casting:1 broader:1 corollary:3 derived:1 focus:1 consistently:1 modelling:1 likelihood:57 potts:1 besag:1 inference:2 mrfs:8 unary:1 typically:1 entire:1 koller:1 arg:1 issue:1 development:2 spatial:1 special:1 marginal:26 field:7 construct:2 equal:1 once:2 x4:1 represents:1 optimising:1 broad:1 icml:1 future:1 report:1 bremaud:1 neighbour:3 densely:2 comprehensive:1 divergence:1 murphy:1 ourselves:1 replacement:1 friedman:1 interest:7 normalisation:1 weakness:1 introduces:1 misha:2 light:1 parametrised:2 chain:2 edge:9 necessary:1 taylor:1 sacrificing:2 theoretical:4 rao:1 lattice:4 cost:2 subset:4 johnson:1 chooses:1 combined:1 density:2 international:3 probabilistic:5 off:1 connecting:1 parametrisation:3 connectivity:2 clifford:2 central:1 centralised:3 satisfied:3 opposed:1 possibly:1 cognitive:1 creating:1 american:1 derivative:1 potential:28 de:4 star:1 subsumes:1 includes:2 dillon:1 satisfy:2 depends:2 performed:2 root:1 view:1 closed:1 parallel:5 inferning:1 identifiability:2 asuncion:1 contribution:3 formed:2 mator:1 who:1 efficiently:1 variance:1 none:2 carlo:1 q6:1 researcher:1 whenever:1 definition:3 energy:2 associated:4 ihler:16 proof:6 rbm:1 gain:1 proved:1 popular:1 carefully:1 appears:1 methodology:1 response:1 wei:2 ox:1 xa:2 hand:1 overlapping:1 google:1 perhaps:1 validity:1 verify:1 requiring:4 concept:1 contain:3 inductive:1 hence:2 equality:1 normalized:6 illustrated:1 round:2 during:1 numerator:2 hammersleyclifford:1 recently:1 common:2 empirically:1 overview:1 exponentially:1 volume:1 association:1 refer:2 cambridge:1 gibbs:10 ai:1 smoothness:1 consistency:8 mathematics:2 centre:3 aq:18 recent:1 perspective:1 verlag:1 der:1 seen:1 guestrin:1 additional:1 greater:1 xaq:29 redundant:1 dashed:1 signal:1 multiple:2 desirable:1 technical:1 match:2 divided:1 post:1 visit:1 mrf:15 variant:5 denominator:2 optimisation:2 expectation:1 essentially:2 arxiv:1 achieved:2 background:1 separately:3 completes:1 leaving:1 standpoint:2 rest:1 induced:7 undirected:6 jordan:1 call:1 canadian:1 iii:3 easy:1 superset:1 xj:40 isolation:1 restrict:2 avenue:1 whether:1 handled:1 bridging:2 queue:1 proceed:1 remark:5 dramatically:1 generally:2 detailed:1 involve:1 characterise:1 induces:1 exist:1 zj:1 notice:1 estimated:3 correctly:1 discrete:2 write:3 group:2 key:1 graph:21 asymptotically:3 sum:1 surmount:1 uncertainty:1 swersky:1 throughout:3 reasonable:1 family:1 appendix:1 acceptable:1 bound:1 played:1 correspondence:2 identifiable:4 strength:2 untied:1 x2:1 x7:1 argument:7 influential:1 developing:1 according:1 disconnected:4 smaller:1 remain:1 intimately:1 contradicts:2 lp:1 s1:5 restricted:2 fienberg:1 equation:16 remains:1 turn:2 know:4 hero:3 tractable:1 apply:7 enforce:2 neighbourhood:10 alternative:1 existence:3 original:3 denotes:4 assumes:1 ensure:1 remaining:1 unfortunate:1 graphical:7 include:1 unifying:1 xc:15 exploit:1 restrictive:1 society:1 seeking:1 objective:1 question:1 parametric:1 said:1 visiting:1 separate:1 considers:1 consensus:13 enforcing:2 induction:2 maximising:1 index:2 cq:2 ratio:1 liang:1 kingdom:1 sinica:2 design:2 boltzmann:5 perform:2 disagree:1 observation:3 markov:7 finite:1 immediate:1 hinton:1 excluding:1 communication:2 arbitrary:2 canada:1 introduced:1 namely:2 required:3 pair:2 connection:2 address:2 able:2 xmn:2 proceeds:1 parallelism:3 beyond:1 xm:4 dynamical:1 smolensky:1 hammersley:2 including:4 max:1 royal:1 overlap:2 natural:1 difficulty:2 force:1 rely:1 advanced:1 normality:2 improve:1 firth:1 imply:1 created:1 carried:1 x8:2 coupled:2 columbia:1 sn:5 xq:7 text:1 review:1 characterises:1 asymptotic:7 relative:2 fully:3 remarkable:1 foundation:1 authoritative:3 degree:3 sufficient:6 consistent:19 principle:2 nowozin:1 share:3 appreciated:1 allow:1 normalised:15 telling:1 institute:1 explaining:1 wide:1 pseudolikelihood:3 distributed:27 van:1 xn:7 evaluating:1 valid:1 made:1 coincide:1 social:2 transaction:1 lebanon:1 approximate:1 keep:1 clique:35 ml:4 global:7 xai:1 reveals:1 generalising:1 summing:6 assumed:1 conclude:1 xi:3 discriminative:1 alternatively:1 continuous:1 latent:1 designates:1 sk:5 learn:1 transfer:1 ca:4 decoupling:1 parametrisations:4 investigated:1 constructing:1 domain:9 statistica:2 s2:2 contradicting:1 allowed:3 x1:1 sub:12 explicit:1 exponential:4 tied:1 mardia:1 british:1 theorem:11 specific:6 discarding:1 xt:1 showing:1 normalizing:1 intractable:1 exists:5 workshop:1 strauss:1 gap:2 chen:1 entropy:1 tc:1 led:2 lap:66 rinaldo:1 expressed:1 applies:1 springer:2 ubc:1 satisfies:16 relies:1 conditional:18 goal:1 consequently:1 careful:1 shared:4 decouple:1 |
4,769 | 5,318 | Elementary Estimators for Graphical Models
Eunho Yang
IBM T.J. Watson Research Center
[email protected]
Aur?elie C. Lozano
IBM T.J. Watson Research Center
[email protected]
Pradeep Ravikumar
University of Texas at Austin
[email protected]
Abstract
We propose a class of closed-form estimators for sparsity-structured graphical
models, expressed as exponential family distributions, under high-dimensional
settings. Our approach builds on observing the precise manner in which the classical graphical model MLE ?breaks down? under high-dimensional settings. Our estimator uses a carefully constructed, well-defined and closed-form backward map,
and then performs thresholding operations to ensure the desired sparsity structure.
We provide a rigorous statistical analysis that shows that surprisingly our simple
class of estimators recovers the same asymptotic convergence rates as those of the
`1 -regularized MLEs that are much more difficult to compute. We corroborate
this statistical performance, as well as significant computational advantages via
simulations of both discrete and Gaussian graphical models.
1
Introduction
Undirected graphical models, also known as Markov random fields (MRFs), are a powerful class
of statistical models, that represent distributions over a large number of variables using graphs, and
where the structure of the graph encodes Markov conditional independence assumptions among the
variables. MRFs are widely used in a variety of domains, including natural language processing [1],
image processing [2, 3, 4], statistical physics [5], and spatial statistics [6]. Popular instances of
this class of models include Gaussian graphical models (GMRFs) [7, 8, 9, 10], used for modeling
continuous real-valued data, and discrete graphical models including the Ising model where each
variable takes values in a discrete set [10, 11, 12]. In this paper, we consider the problem of highdimensional estimation, where the number of variables p may exceed the number of observations n.
In such high-dimensional settings, it is still possible to perform consistent estimation by leveraging
low-dimensional structure. Sparse and group-sparse structural constraints, where few parameters (or
parameter groups) are non-zero, are particularly pertinent in the context of MRFs as they translate
into graphs with few edges.
A key class of estimators for learning graphical models has thus been based on maximum likelihood
estimators (MLE) with sparsity-encouraging regularization. For the task of sparse GMRF estimation, the state-of-the-art estimator minimizes the Gaussian negative log-likelihood regularized by the
`1 norm of the entries (or the off-diagonal entries) of the concentration matrix (see [8, 9, 10]). Strong
statistical guarantees for this estimator have been established (see [13] and references therein). The
resulting optimization problem is a log-determinant program, which can be solved in polynomial
time with interior point methods [14], or by co-ordinate descent algorithms [9, 10]. In a computationally simpler approach for sparse GMRFs, [7] proposed the use of neighborhood selection,
which consists of estimating conditional independence relationships separately for each node in the
graph, via `1 -regularized linear regression, or LASSO [15]. They showed that the procedure can
1
consistently recover the sparse GMRF structure even under high-dimensional settings. The neighborhood selection approach has also been successfully applied to discrete Markov random fields.
In particular, for binary graphical models, [11] showed that consistent neighborhood selection can
be performed via `1 -regularized logistic regression. These results were generalized to general discrete graphical models (where each variable can take m
2 values) by [12] through node-wise
multi-class logistic regression with group sparsity. A related regularized convex program to solve
for sparse GMRFs is the CLIME estimator [16], which reduces the estimation problem to solving
linear programs. Overall, while state of the art optimization methods have been developed to solve
all of these regularized (and consequently non-smooth) convex programs, their iterative approach is
very expensive for large scale problems. Indeed, scaling such regularized convex programs to very
large scale settings has attracted considerable recent research and attention.
In this paper, we investigate the following leading question: ?Can one devise simple estimators
with closed-form solutions that are yet consistent and achieve the sharp convergence rates of the
aforementioned regularized convex programs?? This question was originally considered in the
context of linear regression by [17] and to which they had given a positive answer. It is thus natural
to wonder whether an affirmative response can be provided for the more complicated statistical
modeling setting of MRFs as well.
Our key idea is to revisit the vanilla MLE for estimating a graphical model, and consider where it
?breaks down? in the case of high-dimensions. The vanilla graphical model MLE can be expressed
as a backward mapping [18] in an exponential family distribution that computes the model parameters corresponding to some given (sample) moments. There are however two caveats with this
backward mapping: it is not available in closed form for many classes of models, and even if it were
available in closed form, it need not be well-defined in high-dimensional settings (i.e. could lead to
unbounded model parameter estimates). Accordingly, we consider the use of carefully constructed
proxy backward maps that are both available in closed-form, and well-defined in high-dimensional
settings. We then perform simple thresholding operations on these proxy backward maps to obtain
our final estimators. Our class of algorithms is thus both computationally practical and highly scalable. We provide a unified statistical analysis of our class of algorithms for graphical models arising
from general exponential families. We then instantiate our analysis for the specific cases of GMRFs
and DMRFs, and show that the resulting algorithms come with strong statistical guarantees achieving near-optimal convergence rates, but doing so computationally much faster than the regularized
convex programs. These surprising results are confirmed via simulation for both GMRFs and DMRFs. There has been considerable recent interest in large-scale statistical model estimation, and in
particular, in scaling these to very large-scale settings. We believe our much simpler class of closedform graphical model estimators have the potential to be estimators of choice in such large-scale
settings, particularly if it attracts research on optimizing and scaling its closed-form operations.
2
Background and Problem Setup
Since most popular graphical model families can be expressed as exponential families (see [18]), we
consider general exponential family distributions for a random variable X 2 Rp :
n
o
P(X; ?) = exp h?, (X)i A(?)
(1)
where ? 2 ? ? Rd is the canonical parameter to be estimated, (X) denotes the sufficient statistics
with feature function : Rp 7! Rd , and A(?) is the log-partition function.
An alternative parameterization of the exponential family, to the canonical parameterization above,
def
is via the vector of ?mean parameters? ?(?) = E? [ (X)], which are the moments of the sufficient
statistics (X) under the exponential family distribution. We denote the set of all possible moments
by the moment polytope: M = {? : 9 distribution p s.t. Ep ( ) = ?}, which consist of moments
of the sufficient statistics under all possible distributions. The problem of computing the mean
parameters ?(?) 2 M given the canonical parameters ? 2 ? constitutes the key machine learning
problem of inference in graphical models (expressed in exponential family form (1)). Let us denote
this computation via a so-called forward mapping A : ? 7! M. By properties of exponential family
distributions, the forward mapping A can actually be expressed in terms of the first derivative of
the log-partition function A(?): A : ? 7! ? = rA(?). It can be shown that this map is injective
(one-to-one with its range) if the exponential family is minimal. Moreover, it is onto the interior of
2
M, denoted by Mo . Thus, for any mean parameter ? 2 Mo , there exists a canonical parameter
?(?) 2 ? such that E?(?) [ (X)] = ?. Unless the exponential family is minimal, the corresponding
canonical parameter ?(?) however need not be unique. Thus while there will always exist a so-called
backward mapping A? : Mo 7! ?, that computes the canonical parameters corresponding to given
moments, it need not be unique. A candidate backward map can be constructed via the conjugate of
the log-partition function A? (?) = sup?2? h?, ?i A(?): A? : ? 7! ? = rA? (?).
2.1
High-dimensional Graphical Model Selection
We focus on the high-dimensional setting, where the number of variables p may greatly exceed
the sample size n. Under such high-dimensional settings, it is now well understood that consistent
estimation is possible if structural constraints are imposed on the model parameters ?. In this paper,
we focus on the structural constraint of sparsity, for which the `1 norm is known to be well-suited.
Given n samples {X (i) }ni=1 from P(X; ?? ) that belongs to an exponential family (1), a popular
class of M -estimators for recovering the sparse model parameter ?? is the `1 -regularized maximum
likelihood estimators:
minimize h ?, b i A(?) + n k?k1
(2)
?
Pn
where b := n1 i=1 (X (i) ) is the empirical mean of the sufficient statistics. Since the log partition
function A(?) in (1) is convex, the problem (2) is convex as well.
We now briefly review the two most popular examples of exponential families in the context of
graphical models.
Gaussian Graphical Models. Consider a random vector X = (X1 , . . . , Xp ) with associated pvariate Gaussian distribution N (X|?, ?), mean vector ? and covariance matrix ?. The probability
density function of X can be formulated as an instance of (1):
? 1
?
P(X|?, ?) = exp
hh?, XX > ii + h?, Xi A(?, ?)
(3)
2
where hhA, Bii denotes the trace inner product tr(A B T ). Here, the canonical parameters are the
precision matrix ? and a vector ?, with domain ? := {(?, ?) 2 Rp ?Rp?p : ? 0, ? = ?T }. The
corresponding moment parameters of the graphical model distribution are given by the mean ? =
E? [X], and the covariance matrix ? = E? [XX T ] of the Gaussian. The forward map A : (?, ?) 7!
(?, ?) computing these from the canonical parameters can be written as: ? = ? 1 and ? = ? 1 ?.
The moment polytope can be seen to be given by M = {(?, ?) 2 Rp ? Rp?p : ? ??T ? 0, ? ?
0}, with interior Mo = {(?, ?) 2 Rp ?Rp?p : ? ??T 0, ? 0}. The corresponding backward
map A? : (?, ?) 7! (?, ?) for (?, ?) 2 Mo can be computed as: ? = ? 1 and ? = ? 1 ?.
Without loss of generality, assume that ? = 0 (and hence ? = 0). Then the set of non-zero entries
in the precision matrix ? corresponds to the set of edges in an associated Gaussian Markov random
field (GMRF). In cases where the graph underlying the GMRF has relatively few edges, it thus
makes sense to impose `1 regularization on the off-diagonal entries of ?. Given n i.i.d. random
vectors X (i) 2 Rp sampled from N (0, ?? ), the corresponding `1 -regularized maximum likelihood
estimator (MLE) is given by:
minimize hh?, Sii
? 0
log det ? +
n k?k1,off
Pn
(i)
where S is the sample covariance matrix defined as
i=1 X
P
n
1
(i)
i=1 X , and k ? k1,off is the off-diagonal element-wise `1 norm.
n
(4)
,
X X (i)
X
>
, X :=
Discrete Graphical Models. Let X = (X1 , . . . , Xp ) be a random vector where each variable Xi
takes values in a discrete set X of cardinality m. Given a graph G = (V, E), a pairwise Markov
random field over X is specified via nodewise functions ?s : X 7! R for all s 2 V , and pairwise
functions ?st : X ? X 7! R for
(s, t) 2 E, as
n all
o
P
P
P(X) = exp
?
(X
)
+
?
(X
,
X
)
A(?)
.
(5)
s
s
st
s
t
s2V
(s,t)2E
This family of probability distributions can be represented using the so-called overcomplete representations [18] as follows. For each random variable Xs and j 2 {1, . . . , m}, define nodewise
3
indicators I[Xs = j] equal to 1 if Xs = j and 0 otherwise. Then pairwise MRFs in (5) can be
rewritten as ?
X
X
P(X) = exp
?s;j I[Xs = j] +
?st;jk I[Xs = j, Xt = k] A(?)
(6)
s2V ;j2[m]
(s,t)2E;j,k2[m]
for a set of parameters ? := {?s;j , ?st;jk : s, t 2 V ; (s, t) 2 E; j, k 2 [m]}. Given these sufficient
statistics, the mean/moment parameters are given by the moments ?s;j := E? I[Xs = j] =
P(Xs = j; ?), and ?st;jk := E? I[Xs = j, Xt = k] = P(Xs = j, Xt = k; ?), which precisely
correspond to nodewise and pairwise marginals of the discrete graphical model. Thus, the forward
mapping A : ? 7! ? would correspond to the inference task of computing nodewise and pairwise
marginals of the discrete graphical model given the canonical parameters. A backward mapping
A? : ? 7! ? corresponds to computing a set of canonical parameters such that the corresponding
graphical model distribution would yield the given set of nodewise and pairwise marginals. The
moment polytope in this case consists of the set of all nodewise and pairwise marginals of any
distribution over the random vector X, and hence is termed the marginal polytope; it is typically
intractable to characterize in high-dimensions [18].
Given n i.i.d. samples from an unknown distribution (6) with parameter ?? , one could consider
b
estimating the graphical model structure with an `1 -regularized MLE: ?b 2 minimizeP
? h?, i +
A(?) + k?k1,E , where k ? k1,E is the `1 norm of the edge-parameters: k?k1,E =
s6=t k?st k,
and where we have collated the edgewise parameters {?st;jk }m
for
an
edge
(s,
t)
2
E
into the
j,k=1
vector ?st . However, there is an critical caveat to actually computing this regularized MLE: the
computation of the log-partition function A(?) is intractable (see [18] for details). To overcome this
issue, one might consider instead the following class of M -estimators, discussed in [19]:
?b 2 minimize h?, bi + B(?) + k?k1,E .
?
(7)
Here B(?) is a variational approximation to the log-partition function A(?) of the form: B(?) =
sup?2L h?, ?i B ? (?), where L is a tractable bound on the marginal polytope M, and B ? (?) is a
tractable approximation to the graphical model entropy A? (?). An example of such approximation,
?
which
we shall P
later leverage in this paper, is the tree-reweighted entropy [20] given by Btrw
(?) =
P
s Hs (?s )
st ?st Ist (?st ), where Hs (?s ) is the entropy for node s, Ist (?st ) is the mutual
information for an edge (s, t), and ?st denote the edge-weights that lie in a so-called spanning tree
polytope. If all ?st are set to 1, this boils down to the Bethe approximation [21].
3
Closed-form Estimators for Graphical Models
The state-of-the-art `1 -regularized MLE estimators discussed in the previous section enjoy strong
statistical guarantees but involve solving difficult non-smooth programs. Scaling them to very largescale problems is thus an important and challenging ongoing research area.
In this paper we tackle the scalability issue at the source by departing from regularized MLE approaches and proposing instead a family of closed-form estimators for graphical models.
Elem-GM Estimation:
(8)
minimize k?k1
?
s. t. ?
B ? ( b)
where B ? (?) is the proxy of backward mapping A? , and
1
n
?
n
is a regularization parameter as in (2).
One of the most important properties of (8) is that the estimator is available in closed-form: ?b =
S n B ? ( b) , where [S (u)]i = sign(ui ) max(|ui |
, 0) is the element-wise soft-thresholding
function. This can be shown by the fact that the optimization problem (8) is decomposable into
independent element-wise sub-problems, where each sub-problem corresponds to soft-thresholding.
To get some intuition on our approach, let us first revisit classical MLE estimators for graphical models as in (1), and see where they ?break down? in a high-dimensional setting: minimize? h ?, b i
A(?). By the stationary condition of this optimization problem, the MLE estimator can be simply
expressed as a backward mapping A? ( b). There are two caveats here in high-dimensional settings.
4
The first is that this backward mapping need not have a simple closed-form, and is typically intractable to compute for a large number of variables p. The second is that the backward mapping is
well-defined only for mean parameters that are in the interior Mo of the marginal polytope, whereas
the sample moments b might well lie on the boundary of the marginal polytope. We will illustrate
these two caveats in the next two examples.
Our key idea is to use instead a well-defined proxy function B ? (?) in lieu of the MLE backward
map A? (?) so that B ? ( b) is both well-defined under high-dimensional settings, as well as with a
simple closed-form. The optimization problem (8) seeks an estimator with minimum complexity in
terms of regularizer k ? k1 while being close enough to some ?initial estimator? B ? ( b) in terms of
element-wise `1 norm; ensuring that the final estimator has the desired sparse structure.
3.1
Strong Statistical Guarantees of Closed-form Estimators
We now provide a statistical analysis of estimators in (8) under the following structural constraint:
(C-Sparsity) The ?true? canonical exponential family parameter ?? is exactly sparse with k nonzero elements indexed by the support set S. All other elements in S c are zeros.
Theorem 1. Consider any graphical model in (1) with sparse canonical parameter ?? as stated in
(C-Sparsity). Suppose we solve (8) setting the constraint bound n such that n
?? B ? ( b) 1 .
(A) Then the optimal solution ?b satisfies the following error bounds:
p
?b ?? 1 ? 2 n , k?b ?? k2 ? 4 k n , and
?b ?? 1 ? 8k n .
(B) The support set of the estimate ?b correctly excludes all true zero coordinates of ?? . Moreover,
under the additional assumption that mins2S |?s? |
3 ?? B ? ( b) 1 , it correctly includes all
non-zero coordinates of ?? .
Remarks. Theorem 1 is a non-probabilistic result, and holds deterministically for any selection of
?
n and any selection of B (?). We would then use a probabilistic analysis when we applying the
theorem to specific distributional settings and choices of the backward map B ? (?).
We note that while the theorem analyses the case of sparsity structured parameters, our class of
estimators as well as analyses can be seamlessly extended to more general structures (such as group
sparsity and low rank), by substituting appropriate regularization functions in (8).
A key ingredient in our class of closed-form estimators is the proxy backward map B ? ( b). The
conditions of the theorem require that this backward map has to be carefully constructed in order
for the error bounds and sparsistency guarantees to hold. In the following sections, we will see how
to precisely construct such backward maps B ? (?) for specific problem instances, and then derive the
corresponding consequences of our abstract theorem as corollaries.
4
Closed-form Estimators for Inverse Covariance Estimation in Gaussian
Graphical Models
In this section, we derive a class of closed-form estimators for the multivariate Gaussian setting
in Section 2.1. From our discussion of Gaussian graphical models in Section 2.1, the backward
mapping from moments to the canonical parameters can be simply computed as A? (?) = ? 1 , but
only provided ? 2 Mo := {? 2 Rp?p : ?
0}. However, given the sample covariance, we
cannot just compute the MLE as A? (S) = S 1 since the sample covariance matrix is rank-deficient
and hence does not belong the Mo under high-dimensional settings where p > n.
In our estimation framework (8), we thus use an alternative backward mapping B ? (?) via a thresholding operator. Specifically, for any matrix M 2 Rp?p , we consider the family of thresholding operators T? (M ) : Rp?p ! Rp?p with thresholding parameter ?, defined as [T? (M )]ij := ?? (Mij )
where ?? (?) is an element-wise thresholding operator. Soft-thresholding is a natural option, however,
along the lines of [22], we can use arbitrary sparse thresholding operators satisfying the conditions:
(C-Thresh) For any input a 2 R, (i) |?? (a)| ? |a|, (ii) |?? (a)| = 0 for |a| ? ?, and finally (iii)
|?? (a) a| ? ?.
5
As long as T? (S) is invertible (which we shall examine in section 4.1), we can define B ? (S) :=
[T? (S)] 1 and obtain the following class of estimators:
Elem-GGM Estimation:
(9)
minimize k?k1,off
?
s. t. ?
[T? (S)]
1
1,off
?
n
where k ? k1,off is the off-diagonal element-wise `1 norm as the dual of k ? k1,off .
Comparison with related work.
Note that [16] suggest a Dantzig-like estimator :
minimize? k?k1 s. t. kS? Ik1 ? n where both k ? k1 and k ? k1 are entry-wise (`1 and `1 ,
respectively) norms for a matrix. This estimator applies penalty functions even for the diagonal elements so that the problem can be decoupled into multiple but much simpler optimization problems.
It still requires solving p linear programs with 2p linear constraints for each. On the other hand, the
estimator from (9) has a closed-form solution as long as T? (S) is invertible.
4.1
Convergence Rates for Elem-GGM
In this section we derive a corollary of theorem 1 for Elem-GGM. A prerequisite is to show that
B ? (S) := [T? (S)] 1 is well-defined and ?well-behaved?. The following conditions define a broad
class of Gaussian graphical models that satisfy this requirement.
(C-MinInf?) The true canonical parameter ?? of (3) has bounded induced operator norm such that
?
1
|||?? |||1 := supw6=02Rp k?kwkwk
? ?1 .
1
(C-Sparse?) The true covariance matrix ?? := (?? ) 1 is ?approximately sparse? along the lines
?
of Bickel and Levina [23]: for some positive constant
D for all diagonal entries, and
PpD, ?ii? ?
moreover, for some 0 ? q < 1 and c0 (p), maxi j=1 |?ij |q ? c0 (p). If q = 0, then this
condition boils down to ?? being sparse. We additionally require inf w6=02Rp
k?? wk1
kwk1
?2 .
Now we are ready to utilize Theorem 1 and derive the convergence rates for our Elem-GGM (9).
Corollary 1. Consider Gaussian graphical models (3) where the true parameter ?? has k non-zero
off-diagonal elements, and the conditions in (C-MinInf?) and (C-Sparse?) hold. Suppose that we
solve the optimization problem q
in (9) with a generalized
thresholding operator satisfying (C-Thresh)
q
10? log p0
log p0
0
and setting ? := 16(maxi ?ii )
:= a
n for p := max{n, p}. Furthermore, suppose
q n
0
b of
also that we select n := 4??12 a lognp . Then, as long as n > c3 log p0 , any optimal solution ?
(9) satisfies
b
?
??
1,off
?
8?1 a
?2
r
log p0
,
n
with probability at least 1
b
?
??
F
?
16?1 a
?2
c1 exp( c2 log p0 ).
r
k log p0
,
n
b
?
??
1,off
?
32?1 a
k
?2
r
log p0
n
We remark that the rates in Corollary 1 are asymptotically the same as those for?standard
`?1 regularq
k log p0
?
b
ized MLE estimators in (4); for instance, [13] show that |||?MLE ? |||F = O
. This is
n
remarkable given the simplicity of Elem-GGM.
5
Closed-form Estimators for Discrete Markov Random Fields
We now specialize our class of closed-form estimators (8) to the setting of discrete Markov random
fields described in Section 2.1. In this case, computing the backward mapping A? is non-trivial and
typically intractable if the graphical structure has loops [18]. Therefore, we need an approximation
of the backward map A? , for which we will leverage the tree-reweighted variational approximation
? b
discussed in Section 2.1. Consider the following map ?? := Btrw
( ), where
bst;jk
??s;j = log bs;j , and ??st;jk = ?st log
(10)
bs;j bt;k
Pn
Pn
where bs;j = n1 i=1 I[Xs,i = j] and bst;jk = n1 i=1 I[Xs,i = j]I[Xt,i = k] are the empirical
?
moments of the sufficient statistics in (6) (we define 0/0 := 1). It was shown in [20] that Btrw
(?)
6
satisfies the following property: the (pseudo)marginals computed by performing tree-reweighted
? b
variational inference with the parameters ?? := Btrw
( ) yield the sufficient statistics b. In other
?
words, the approximate backward map Btrw
computes an element in the pre-image of the approximate forward map given by tree-reweighted variational inference. Since tree-reweighted variational
?
inference approximates the true marginals well in practice, the map Btrw
(?) is thus a great candidate
for as an approximate backward map.
As an alternative to the `1 regularized approximate MLE estimators (7), we thus obtain the
?
following class of estimators using Btrw
(?) as an instance of (8):
Elem-DMRF Estimation:
(11)
minimize k?k1,E
?
? b
Btrw
( )
s. t. ?
1,E
?
n
where k ? k1,E is the maximum absolute value of edge-parameters as a dual of k ? k1,E .
? b
Note that given the empirical means of sufficient statistics, Btrw
( ) can usually be obtained easily,
without the need of explicitly specifying the log-partition function approximation B(?) in (7).
5.1
Convergence Rates for Elem-DRMF
We now derive the convergence rates of Elem-DRMF for the case where B ? (?) is selected as in
(10) following the tree reweighed approximation [20]. Let ?? be the ?true? marginals (or mean
parameters) from the true log-partition function and true canonical parameter ?? : ?? = A(?? ). We
shall require that the approximation Btrw (?) be close enough to the true A(?) in terms of backward
mapping. In addition we assume that true marginal distributions are strictly positive.
(C-LogPartition)
??
?
Btrw
(?? )
1,E
? ?.
(C-Marginal) For all s 2 V and j 2 [m], the true singleton marginal ??s;j := E?? I[Xs = j] =
P(Xs = j; ?? ) satisfies ?min < ??s;j for some strictly positive constant ?min 2 (0, 1). Similarly,
for all s, t 2 V and all j, k 2 [m], ??st;jk satisfies ?min < ??st;jk .
Now we are ready to utilize Theorem 1 to derive the convergence rates for our closed-form estimator
?
(11) when ?? has k non-zero pairwise parameters ?st
, where we recall the notatation that ?st :=
m
{?st;jk }j,k=1 is a collation of the edgewise parameters for edge (s, t). We also define k?kq,E :=
P
( s6=t k?st kq )1/q , for q 2 {1, 2, 1}.
Corollary 2. Consider discrete Markov random fields (6) when the true parameter ?? has actually
k non-zero pair-wise parameters, and the conditions in (C-LogPartition) and (C-Marginal) also
?
hold in these discrete MRFs. Suppose that we solve the optimization problem in (11) with Btrw
(?)
set as q
(10) using tree reweighed approximation. Furthermore, suppose also that we select n :=
4c2 log p
? + c1 logn p for some positive constant c1 depending only on ?min . Then, as long as n > 1?2 ,
min
there are universal positive constants (c2 , c3 ) such that any optimal solution ?b of (11) satisfies
k?b
?
? k1,E ? 2? + 2c1
r
log p b
, k?
n
with probability at least 1
6
?
p
? k2,E ? 4 k? + 4c1
c2 exp( c3 log p0 ).
r
k log p b
, k?
n
?
? k1,E ? 8k? + 8c1 k
r
log p
n
Experiments
In this section, we report a set of synthetic experiments corroborating our theoretical results on both
Gaussian and discrete graphical models.
Gaussian Graphical Models We now corroborate Corollary 1, and furthermore, compare our
estimator with the `1 regularized MLE in terms of statistical performance with respect to the
b ?? kq for q 2 {2, 1}, as well as in terms of computational performance.
parameter error k?
To generate true inverse covariance matrices ?? with a random sparsity structure, we follow the
procedure described in [25, 24]. We first generate a sparse matrix U whose non-zero entries are set
to ?1 with equal probabilities. ?? is then set to U > U and then a diagonal term is added to ensure
7
Table 1: Performance of our Elem-GM vs. state of the art QUIC algorithm [24] solving (4) under
two different regimes: (Left) (n, p) = (800, 1600), (Right) (n, p) = (5000, 10000).
Elem-GM
QUIC
K
0.01
0.02
0.05
0.1
0.5
1
2
3
4
Time(sec)
<1
<1
<1
<1
2575.5
1009
272.1
78.1
28.7
`F (off)
6.36
6.19
5.91
6
12.74
7.30
6.33
6.97
7.68
`1 (off)
0.1616
0.1880
0.1655
0.1703
0.11
0.13
0.18
0.21
0.23
FPR
0.48
0.24
0.06
0.01
0.52
0.35
0.16
0.07
0.02
TPR
0.99
0.99
0.99
0.97
1.00
0.99
0.99
0.94
0.86
Elem-GM
QUIC
K
0.05
0.1
0.5
1
2
2.5
3
3.5
Time(sec)
47.3
46.3
45.8
46.2
*
*
4.8 ? 104
2.7 ? 104
`F (off)
11.73
8.91
5.66
8.63
*
*
9.85
10.51
`1 (off)
0.1501
0.1479
0.1308
0.1111
*
*
0.1083
0.1111
FPR
0.13
0.03
0.0
0.0
*
*
0.06
0.04
TPR
1.00
1.00
1.00
0.99
*
*
1.00
0.99
Table 2: Performance of Elem-DMRF vs. the regularized MLE-based approach of [12] for structure
recovery of DRMFs.
Graph Type
Chain Graph
Grid Graph
# Parameters
128
2000
128
2000
Method
Elem-DMRF
Regularized MLE
Elem-DMRF
Regularized MLE
Elem-DMRF
Regularized MLE
Elem-DMRF
Regularized MLE
Time(sec)
0.17
7.30
21.67
4315.10
0.17
7.99
21.68
4454.44
TPR
0.87
0.81
0.79
0.75
0.97
0.84
0.80
0.77
FNR
0.01
0.01
0.12
0.21
0.01
0.02
0.12
0.18
?? is positive definite. Finally, we normalize ?? with maxpi=1 ??ii so that the maximum diagonal
entry is equal to 1. We control the number of non-zeros in U so that the number of non-zeros in the
final ?? is approximately 10p. We additionally set the number of samples n to half of the number
of variables p. Note that though the number of variables is p, the total number of entries in the
canonical parameter consisting of the covariance matrix is O(p2 ).
Table 1 summarizes the performance of our closed-form estimators in terms of computation
time,
p
b ?? k1,off and |||?
b ?? |||F,off . We fix the thresholding parameter ? = 2.5 log p/n for all
k?
p
settings, and vary the regularization parameter n = K log p/n to investigate how this regularizer
affects the final estimators. Baselines are `1 regularized MLE estimators in (4); we use QUIC
algorithms [24], which is one of the fastest way to solve (4). In the table, we show the results of the
QUIC algorithm run with a tolerance ? = 10 4 ; * indicates that the algorithm does not stop within
15 hours. In Appendix, we provide more extensive comparisons including receiver operator curves
(ROC) for these methods for settings in Table 1. As can be seen from the table and the figure, the
performance of Elem-GM estimators is both statistically competitive in terms of all types of errors
and support set recovery, while performing much better computationally than classical methods
based on `1 regularized MLE.
Discrete Graphical Models We consider two different classes of pairwise graphical models:
chain graphs and grids. For each case, the size of the alphabet is set to m = 3; the true parameter vector ?? is generated by sampling each non-zero entry from N (0, 1).
We compare Elem-DMRF with the group-sparse regularized MLE-based approach of Jalali et al.
[12], which uses group `1 /`2 regularization, where all the parameters of an edge form a group, so as
to encourage sparsity in terms of the edges, and which we solved using proximal gradient descent.
While our estimator in (11) used vanilla sparsity, we used a simple extension to the group-sparse
structured setting;
pplease see Appendix E for more details. For both methods, the tuning parameter
is set to n = c log p/n, where c is selected using cross-validation. We use 20 simulation runs
where for each run n = p/2 samples are drawn from the distribution specified by ?? .
We report true positive rates, false positive rates and timing for running each method. We note
that the timing is for running each method without counting the time spent in the cross-validation
process (Had we taken the cross-validation into account, the advantage of our method would be
even more pronounced, since the entire path of solutions can be computed via simple group-wise
thresholding operations.) The results in Table 2 show that Elem-DMRF is much faster than its
MLE-based counterpart, and yield competitive results in terms of structure recovery.
Acknowledgments E.Y and P.R. acknowledge the support of ARO via W911NF-12-1-0390 and
NSF via IIS-1149803, IIS-1320894, IIS-1447574, and DMS-1264033
8
References
[1] C. D. Manning and H. Schutze. Foundations of Statistical Natural Language Processing. MIT
Press, 1999.
[2] J.W. Woods. Markov image modeling. IEEE Transactions on Automatic Control, 23:846?850,
October 1978.
[3] M. Hassner and J. Sklansky. Markov random field models of digitized image texture. In
ICPR78, pages 538?540, 1978.
[4] G. Cross and A. Jain. Markov random field texture models. IEEE Trans. PAMI, 5:25?39, 1983.
[5] E. Ising. Beitrag zur theorie der ferromagnetismus. Zeitschrift f?ur Physik, 31:253?258, 1925.
[6] B. D. Ripley. Spatial statistics. Wiley, New York, 1981.
[7] N. Meinshausen and P. B?uhlmann. High-dimensional graphs and variable selection with the
Lasso. Annals of Statistics, 34:1436?1462, 2006.
[8] M. Yuan and Y. Lin. Model selection and estimation in the Gaussian graphical model.
Biometrika, 94(1):19?35, 2007.
[9] J. Friedman, T. Hastie, and R. Tibshirani. Sparse inverse covariance estimation with the graphical Lasso. Biostatistics, 2007.
[10] O. Bannerjee, , L. El Ghaoui, and A. d?Aspremont. Model selection through sparse maximum
likelihood estimation for multivariate Gaussian or binary data. Jour. Mach. Lear. Res., 9:485?
516, March 2008.
[11] P. Ravikumar, M. J. Wainwright, and J. Lafferty. High-dimensional ising model selection using
`1 -regularized logistic regression. Annals of Statistics, 38(3):1287?1319, 2010.
[12] A. Jalali, P. Ravikumar, V. Vasuki, and S. Sanghavi. On learning discrete graphical models
using group-sparse regularization. In Inter. Conf. on AI and Statistics (AISTATS), 14, 2011.
[13] P. Ravikumar, M. J. Wainwright, G. Raskutti, and B. Yu. High-dimensional covariance estimation by minimizing `1 -penalized log-determinant divergence. Electronic Journal of Statistics,
5:935?980, 2011.
[14] S. Boyd and L. Vandenberghe. Convex optimization. Cambridge University Press, Cambridge,
UK, 2004.
[15] R. Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical
Society, Series B, 58(1):267?288, 1996.
[16] T. Cai, W. Liu, and X. Luo. A constrained `1 minimization approach to sparse precision matrix
estimation. Journal of the American Statistical Association, 106(494):594?607, 2011.
[17] E. Yang, A. Lozano, and P. Ravikumar. Elementary estimators for high-dimensional linear
regression. In International Conference on Machine learning (ICML), 31, 2014.
[18] M. J. Wainwright and M. I. Jordan. Graphical models, exponential families and variational
inference. Foundations and Trends in Machine Learning, 1(1?2):1?305, December 2008.
[19] E. Yang and P. Ravikumar. On the use of variational inference for learning discrete graphical
models. In International Conference on Machine learning (ICML), 28, 2011.
[20] M. J. Wainwright, T. S. Jaakkola, and A. S. Willsky. Tree-reweighted belief propagation algorithms and approximate ML estimation by pseudomoment matching. In Inter. Conf. on AI and
Statistics (AISTATS), 2003.
[21] J. S. Yedidia, W. T. Freeman, and Y. Weiss. Generalized belief propagation. In NIPS 13, pages
689?695. MIT Press, 2001.
[22] A. J. Rothman, E. Levina, and J. Zhu. Generalized thresholding of large covariance matrices.
Journal of the American Statistical Association (Theory and Methods), 104:177?186, 2009.
[23] P. J. Bickel and E. Levina. Covariance regularization by thresholding. Annals of Statistics, 36
(6):2577?2604, 2008.
[24] C. J. Hsieh, M. Sustik, I. Dhillon, and P. Ravikumar. Sparse inverse covariance matrix estimation using quadratic approximation. In Neur. Info. Proc. Sys. (NIPS), 24, 2011.
[25] L. Li and K. C. Toh. An inexact interior point method for l1-regularized sparse covariance
selection. Mathematical Programming Computation, 2:291?315, 2010.
9
| 5318 |@word h:2 lognp:1 determinant:2 briefly:1 polynomial:1 norm:8 c0:2 pseudomoment:1 physik:1 simulation:3 seek:1 covariance:15 p0:9 hsieh:1 tr:1 moment:14 initial:1 liu:1 series:1 com:2 surprising:1 luo:1 toh:1 yet:1 attracted:1 written:1 partition:8 pertinent:1 v:2 stationary:1 half:1 instantiate:1 selected:2 parameterization:2 accordingly:1 fpr:2 sys:1 caveat:4 node:3 simpler:3 unbounded:1 mathematical:1 along:2 constructed:4 sii:1 c2:4 yuan:1 consists:2 specialize:1 manner:1 pairwise:9 inter:2 ra:2 indeed:1 examine:1 multi:1 freeman:1 encouraging:1 cardinality:1 provided:2 estimating:3 moreover:3 xx:2 underlying:1 bounded:1 biostatistics:1 minimizes:1 affirmative:1 developed:1 proposing:1 unified:1 guarantee:5 pseudo:1 tackle:1 exactly:1 biometrika:1 k2:3 uk:1 bst:2 control:2 enjoy:1 positive:9 understood:1 timing:2 consequence:1 zeitschrift:1 mach:1 path:1 approximately:2 pami:1 might:2 therein:1 dantzig:1 k:1 meinshausen:1 specifying:1 challenging:1 co:1 fastest:1 range:1 bi:1 statistically:1 elie:1 practical:1 unique:2 acknowledgment:1 practice:1 definite:1 procedure:2 area:1 empirical:3 universal:1 boyd:1 matching:1 word:1 pre:1 suggest:1 get:1 onto:1 interior:5 selection:12 close:2 cannot:1 operator:7 context:3 applying:1 map:18 imposed:1 center:2 logpartition:2 attention:1 convex:8 decomposable:1 gmrf:4 simplicity:1 recovery:3 estimator:49 vandenberghe:1 collated:1 s6:2 coordinate:2 annals:3 gm:5 suppose:5 programming:1 us:2 element:11 trend:1 expensive:1 particularly:2 jk:10 satisfying:2 ising:3 distributional:1 ep:1 solved:2 pradeepr:1 intuition:1 ui:2 complexity:1 solving:4 easily:1 represented:1 regularizer:2 alphabet:1 jain:1 neighborhood:3 whose:1 widely:1 valued:1 solve:6 otherwise:1 statistic:16 final:4 advantage:2 cai:1 propose:1 aro:1 product:1 j2:1 loop:1 translate:1 achieve:1 pronounced:1 normalize:1 scalability:1 convergence:8 requirement:1 spent:1 illustrate:1 derive:6 depending:1 ij:2 p2:1 strong:4 recovering:1 c:1 come:1 require:3 hassner:1 fix:1 icpr78:1 elementary:2 rothman:1 strictly:2 extension:1 hold:4 considered:1 exp:6 great:1 mapping:15 mo:8 substituting:1 bickel:2 vary:1 estimation:18 eunhyang:1 proc:1 uhlmann:1 utexas:1 successfully:1 beitrag:1 minimization:1 mit:2 gaussian:16 always:1 pn:4 shrinkage:1 jaakkola:1 corollary:6 focus:2 consistently:1 rank:2 likelihood:5 indicates:1 seamlessly:1 greatly:1 rigorous:1 baseline:1 sense:1 schutze:1 inference:7 mrfs:6 el:1 typically:3 bt:1 entire:1 ppd:1 overall:1 among:1 aforementioned:1 issue:2 denoted:1 dual:2 logn:1 spatial:2 art:4 constrained:1 mutual:1 marginal:8 field:9 equal:3 construct:1 sampling:1 broad:1 yu:1 icml:2 constitutes:1 report:2 sanghavi:1 few:3 divergence:1 sparsistency:1 consisting:1 n1:3 friedman:1 interest:1 investigate:2 highly:1 pradeep:1 chain:2 edge:11 encourage:1 injective:1 decoupled:1 unless:1 tree:9 indexed:1 desired:2 re:1 overcomplete:1 theoretical:1 minimal:2 instance:5 modeling:3 soft:3 corroborate:2 w911nf:1 entry:10 kq:3 wonder:1 characterize:1 answer:1 aclozano:1 proximal:1 synthetic:1 mles:1 st:22 density:1 jour:1 international:2 aur:1 probabilistic:2 physic:1 off:19 invertible:2 conf:2 american:2 derivative:1 leading:1 li:1 closedform:1 account:1 potential:1 singleton:1 sec:3 includes:1 satisfy:1 explicitly:1 performed:1 break:3 later:1 closed:21 observing:1 doing:1 sup:2 competitive:2 recover:1 option:1 complicated:1 ferromagnetismus:1 clime:1 minimize:8 ni:1 ggm:5 correspond:2 yield:3 confirmed:1 inexact:1 dm:1 associated:2 recovers:1 edgewise:2 boil:2 sampled:1 stop:1 popular:4 recall:1 carefully:3 actually:3 originally:1 follow:1 response:1 wei:1 though:1 generality:1 furthermore:3 just:1 hand:1 propagation:2 logistic:3 behaved:1 believe:1 true:16 counterpart:1 lozano:2 regularization:8 hence:3 nonzero:1 dhillon:1 reweighted:6 generalized:4 performs:1 l1:1 image:4 wise:10 variational:7 raskutti:1 discussed:3 belong:1 approximates:1 tpr:3 marginals:7 elem:20 association:2 significant:1 cambridge:2 ai:2 rd:2 vanilla:3 grid:2 tuning:1 similarly:1 automatic:1 language:2 had:2 ik1:1 multivariate:2 showed:2 recent:2 thresh:2 optimizing:1 belongs:1 inf:1 reweighed:2 termed:1 binary:2 watson:2 kwk1:1 der:1 devise:1 seen:2 minimum:1 additional:1 impose:1 ii:8 multiple:1 reduces:1 smooth:2 faster:2 levina:3 cross:4 long:4 lin:1 sklansky:1 ravikumar:7 mle:26 ensuring:1 scalable:1 regression:7 represent:1 c1:6 zur:1 background:1 whereas:1 separately:1 addition:1 source:1 induced:1 deficient:1 undirected:1 december:1 leveraging:1 lafferty:1 jordan:1 structural:4 near:1 yang:3 leverage:2 exceed:2 iii:1 enough:2 counting:1 variety:1 independence:2 affect:1 attracts:1 lasso:4 hastie:1 inner:1 idea:2 texas:1 det:1 whether:1 penalty:1 york:1 remark:2 involve:1 generate:2 maxpi:1 exist:1 canonical:16 nsf:1 revisit:2 sign:1 estimated:1 arising:1 correctly:2 tibshirani:2 nodewise:6 discrete:17 shall:3 group:10 key:5 ist:2 achieving:1 drawn:1 utilize:2 backward:25 bannerjee:1 graph:11 excludes:1 asymptotically:1 wood:1 minimizep:1 inverse:4 run:3 powerful:1 family:19 electronic:1 summarizes:1 scaling:4 appendix:2 def:1 bound:4 quadratic:1 constraint:6 precisely:2 encodes:1 min:5 performing:2 relatively:1 s2v:2 structured:3 neur:1 march:1 manning:1 conjugate:1 ur:1 b:3 ghaoui:1 taken:1 computationally:4 hh:2 tractable:2 sustik:1 lieu:1 available:4 operation:4 rewritten:1 prerequisite:1 yedidia:1 appropriate:1 bii:1 alternative:3 rp:15 denotes:2 running:2 ensure:2 include:1 graphical:44 k1:21 build:1 classical:3 society:1 question:2 added:1 quic:5 concentration:1 diagonal:9 jalali:2 gradient:1 polytope:8 mins2s:1 trivial:1 spanning:1 willsky:1 w6:1 relationship:1 eunho:1 minimizing:1 difficult:2 setup:1 october:1 theorie:1 info:1 trace:1 negative:1 stated:1 ized:1 unknown:1 perform:2 observation:1 markov:11 acknowledge:1 descent:2 extended:1 precise:1 digitized:1 sharp:1 arbitrary:1 ordinate:1 pair:1 specified:2 c3:3 extensive:1 established:1 hour:1 nip:2 trans:1 usually:1 regime:1 sparsity:12 program:9 including:3 max:2 royal:1 belief:2 wainwright:4 critical:1 natural:4 regularized:27 indicator:1 largescale:1 zhu:1 ready:2 aspremont:1 gmrfs:5 review:1 asymptotic:1 loss:1 ingredient:1 remarkable:1 validation:3 foundation:2 vasuki:1 sufficient:8 consistent:4 proxy:5 xp:2 thresholding:15 ibm:4 austin:1 penalized:1 surprisingly:1 absolute:1 sparse:24 departing:1 tolerance:1 overcome:1 dimension:2 boundary:1 curve:1 computes:3 forward:5 transaction:1 approximate:5 ml:1 wk1:1 corroborating:1 receiver:1 xi:2 ripley:1 continuous:1 iterative:1 table:7 additionally:2 bethe:1 domain:2 aistats:2 x1:2 roc:1 wiley:1 precision:3 sub:2 deterministically:1 exponential:15 candidate:2 lie:2 down:5 theorem:9 specific:3 xt:4 maxi:2 x:13 consist:1 exists:1 hha:1 intractable:4 false:1 texture:2 suited:1 entropy:3 simply:2 expressed:6 applies:1 mij:1 corresponds:3 satisfies:6 conditional:2 formulated:1 consequently:1 lear:1 considerable:2 fnr:1 specifically:1 called:4 total:1 select:2 highdimensional:1 support:4 ongoing:1 |
4,770 | 5,319 | Structure learning of
antiferromagnetic Ising models
Guy Bresler1
David Gamarnik2
Devavrat Shah1
Laboratory for Information and Decision Systems
Department of EECS1 and Sloan School of Management2
Massachusetts Institute of Technology
{gbresler,gamarnik,devavrat}@mit.edu
Abstract
In this paper we investigate the computational complexity of learning the
graph structure underlying a discrete undirected graphical model from i.i.d.
samples. Our first result is an unconditional computational lower bound
of (pd/2 ) for learning general graphical models on p nodes of maximum
degree d, for the class of so-called statistical algorithms recently introduced
by Feldman et al. [1]. The construction is related to the notoriously difficult
learning parities with noise problem in computational learning theory. Our
? d+2 ) runtime required by Bresler, Mossel,
lower bound suggests that the O(p
and Sly?s [2] exhaustive-search algorithm cannot be significantly improved
without restricting the class of models.
Aside from structural assumptions on the graph such as it being a tree,
hypertree, tree-like, etc., many recent papers on structure learning assume
that the model has the correlation decay property. Indeed, focusing on ferromagnetic Ising models, Bento and Montanari [3] showed that all known
low-complexity algorithms fail to learn simple graphs when the interaction
strength exceeds a number related to the correlation decay threshold. Our
second set of results gives a class of repelling (antiferromagnetic) models
that have the opposite behavior: very strong interaction allows efficient
? 2 ). We provide an algorithm whose performance inlearning in time O(p
? 2 ) and O(p
? d+2 ) depending on the strength of the
terpolates between O(p
repulsion.
1
Introduction
Graphical models have had tremendous impact in a variety of application domains. For
unstructured high-dimensional distributions, such as in social networks, biology, and finance,
an important first step is to determine which graphical model to use. In this paper we
focus on the problem of structure learning: Given access to n independent and identically
distributed samples ? (1) , . . . ? (n) from an undirected graphical model representing a discrete
random vector ? = (?1 , . . . , ?p ), the goal is to find the graph G underlying the model. Two
basic questions are 1) How many samples are required? and 2) What is the computational
complexity?
In this paper we are mostly interested in the computational complexity of structure learning.
We first consider the problem of learning a general discrete undirected graphical model of
bounded degree.
1
1.1
Learning general graphical models
Several algorithms based on exhaustively searching over possible node neighborhoods have
appeared in the last decade [4, 2, 5]. Abbeel, Koller, and Ng [4] gave algorithms for learning
general graphical models close to the true distribution in Kullback-Leibler distance. Bresler,
Mossel, and Sly [2] presented algorithms guaranteed to learn the true underlying graph.
The algorithms in both [4] and [2] perform a search over candidate neighborhoods, and for
a graph of maximum degree d, the computational complexity for recovering a graph on p
? d+2 ) (where the O
? notation hides logarithmic factors).
nodes scales as O(p
While the algorithms in [2] are guaranteed to reconstruct general models under basic
nondegeneracy conditions using an optimal number of samples n = O(d log p) (sample
complexity lower bounds were proved by Santhanam and Wainwright [6] as well as [2]), the
? d+2 ) run-time is impractically high even for constant but large graph
exponent d in the O(p
degrees. This has motivated a great deal of work on structure learning for special classes of
graphical models. But before giving up on general models, we ask the following question:
Question 1: Is it possible to learn the structure of general graphical models on p
nodes with maximum degree d using substantially less computation than pd ?
Our first result suggests that the answer to Question 1 is negative. We show an uncond
ditional computational lower bound of p 2 for the class of statistical algorithms introduced
by Feldman et al. [1]. This class of algorithms was introduced in order to understand the
apparent difficulty of the Planted Clique problem, and is based on Kearns? statistical query
model [7]. Kearns showed in his landmark paper that statistical query algorithms require
exponential computation to learn parity functions subject to classification noise, and our
hardness construction is related to this problem. Most known algorithmic approaches (including Markov chain Monte Carlo, semidefinite programming, and many others) can be
implemented as statistical algorithms, so the lower bound is fairly convincing.
We give background and prove the following theorem in Section 4.
Theorem 1.1. Statistical algorithms require at least (p 2 ) computation steps in order to
learn the structure of a general graphical models of degree d.
d
If complexity pd is to be considered intractable, what shall we consider as tractable? Writing
algorithm complexity in the form c(d)pf (d) , for high-dimensional (large p) problems the
exponent f (d) is of primary importance, and we will think of tractable algorithms as having
an f (d) that is bounded by a constant independent of d. The factor c(d) is also important,
and we will use it to compare algorithms with the same exponent f (d).
In light of Theorem 1.1, reducing computation below p (d) requires restricting the class
of models. One can either restrict the graph structure or the nature of the interactions
between variables. The seminal paper of Chow and Liu [8] makes a model restriction of
the first type, assuming that the graph is a tree; generalizations include to polytrees [9],
hypertrees [10], and others. Among the many possible assumptions of the second type,
the correlation decay property is distinguished: to the best of our knowledge all existing
low-complexity algorithms require the correlation decay property [3].
1.2
Correlation decay property
Informally, a graphical model is said to have the correlation decay property (CDP) if any
two variables ?s and ?t are asymptotically independent as the graph distance between s and
t increases. Exponential decay of correlations holds when the distance from independence
decreases exponentially fast in graph distance, and we will mean this stronger form when
referring to correlation decay. Correlation decay is known to hold for a number of pairwise
graphical models in the so-called high-temperature regime, including Ising, hard-core lattice
gas, Potts (multinomial) model, and others (see, e.g., [11, 12, 13, 14, 15, 16]).
2
Bresler, Mossel, and Sly [2] observed that it is possible to efficiently learn models with (exponential) decay of correlations, under the additional assumption that neighboring variables
have correlation bounded away from zero (as is true, e.g., for the ferromagnetic Ising model
in the high temperature regime). The algorithm they proposed for this setting pruned the
candidate set of neighbors for each node to roughly size O(d) by retaining only those variables
with sufficiently high correlations, and then within this set performed the exhaustive search
? 2 ). The
over neighborhoods mentioned before, resulting in a computational cost of dO(d) O(p
greedy algorithms of Netrapali et al. [17] and Ray et al. [18] also require the correlation decay property and perform a similar pruning step by retaining only nodes with high pairwise
correlation; they then use a different method to select the true neighborhood.
A number of papers consider the problem of reconstructing Ising models on graphs with
few short cycles, beginning with Anandkumar et al. [19]. Their results apply to the case of
Ising models on sparsely connected graphs such as the Erd?os-Renyi random graph G(p, dp ).
They additionally require the interaction parameters to be either generic or ferromagnetic.
Ferromagnetic models have the benefit that neighbors always have a non-negligible correlation because the dependencies cannot cancel, but in either case the results still require the
CDP to hold. Wu et al. [20] remove the assumption of generic parameters in [19], but again
require the CDP.
Other algorithms for structure learning are based on convex optimization, such as Ravikumar et al.?s [21] approach using regularized node-wise logistic regression. While this
algorithm does not explicitly require the CDP, Bento and Montanari [3] found that the
logistic regression algorithm of [21] provably fails to learn certain ferromagnetic Ising model
on simple graphs without correlation decay. Other convex optimization-based algorithms
such as [22, 23, 24] require similar incoherence or restricted isometry-type conditions that
are difficult to verify, but likely also require correlation decay. Since all known algorithms
for structure learning require the CDP, we ask the following question (paraphrasing Bento
and Montanari):
Question 2: Is low-complexity structure learning possible for models which do not
exhibit the CDP, on general bounded degree graphs?
Our second main result answers this question affirmatively by showing that a broad class of
repelling models on general graphs can be learned using simple algorithms, even when the
underlying model does not exhibit the CDP.
1.3
Repelling models
The antiferromagnetic Ising model has a negative interaction parameter, whereby neighboring nodes prefer to be in opposite states. Other popular antiferromagnetic models include
the Potts or coloring model, and the hard-core model.
Antiferromagnetic models have the interesting property that correlations between neighbors
can be zero due to cancellations. Thus algorithms based on pruning neighborhoods using
pairwise correlations, such as the algorithm in [2] for models with correlation decay, does not
work for anti-ferromagnetic models. To our knowledge there are no previous results that
improve on the pd computational complexity for structure learning of antiferromagnetic
models on general graphs of maximum degree d.
Our first learning algorithm, described in Section 2, is for the hard-core model.
Theorem 1.2 (Informal). It is possible to learn strongly repelling models, such as the hard? 2 ).
core model, with run-time O(p
We extend this result to weakly repelling models (equivalent to the antiferromagnetic Ising
model parameterized in a nonstandard way, see Section 3). Here ? is a repelling strength
and h is an external field.
Theorem 1.3 (Informal). Suppose ? ? (d ? ?)(h + ln 2) for an integer 0 ? ? < d. Then
? 2+? ).
it is possible to learn a repelling model with interaction ?, with run-time O(p
3
? 2 ), achievable for
The computational complexity of the algorithm interpolates between O(p
d+2
?
strongly repelling models, and O(p ), achievable for general models using exhaustive
search. The complexity depends on the repelling strength of the model, rather than structural assumptions on the graph as in [19, 20].
We remark that the strongly repelling models exhibit long-range correlations, yet the algorithmic task of graph structure learning is possible using a local procedure.
The focus of this paper is on structure learning, but the problem of parameter estimation
is equally important. It turns out that the structure learning problem is strictly more
challenging for the models we consider: once the graph is known, it is not difficult to
estimate the parameters with low computational complexity (see, e.g., [4]).
2
Learning the graph of a hard-core model
We warm up by considering the hard-core model. The analysis in this section is straightforward, but serves as an example to highlight the fact that correlation decay is not a necessary
condition for structure learning.
Given a graph G = (V, E) on |V | = p nodes, denote by I(G) ? {0, 1}p the set of independent
set indicator vectors ?, for which at least one of ?i or ?j is zero for each edge {i, j} ? E(G).
The hardcore model with fugacity ? > 0 assigns nonzero probability only to vectors in I(G),
with
?|?|
P(?) =
, ? ? I(G) .
(2.1)
Z
q
|?|
Here |?| is the number of entries of ? equal to one and Z = ??I(G) ? is the normalizing
constant called the partition function. If ? > 1 then more mass is assigned to larger
independent sets. (We use indicator vectors to define the model in order to be consistent
with the antiferromagnetic Ising model in the next section.)
Our goal is to learn the graph G = (V, E) underlying the model (2.1) given access to independent samples ? (1) , . . . , ? (n) . The following simple algorithm reconstructs G efficiently.
Algorithm 1 simpleHC(? (1) , . . . , ? (n) )
1: FOR each i, j, k:
(k)
(k)
2: IF ?i = ?j = 1, THEN S = S ? {i, j}
? = Sc
3: OUTPUT E
The idea behind the algorithm is very simple. If {i, j} belongs to the edge set E(G), then
(k)
(k)
for every sample ? (k) either ?i = 0 or ?j = 0 (or both). Thus for every i, j and k such
(k)
(k)
that ?i = ?j = 1 we can safely declare {i, j} not to be an edge. To show correctness of
the algorithm it is therefore sufficient to argue that for every non-edge {i, j} there is a high
likelihood that such an independent set ? (k) will be sampled.
Before doing this, we observe that simpleHC actually computes the maximum-likelihood
(k)
(k)
estimate for the graph G. To see this, note that an edge e = {i, j} for which ?i = ?j = 1
? since P(? (k) |G+e)
?
? Thus the ML estimate contains
for some k cannot be in G,
= 0 for any G.
a subset of those edges e which have not been ruled out by ? (1) , . . . , ? (n) . But adding any
such edge e to the graph decreases the value of the partition function in (2.1) (the sum is
over fewer independent sets), thereby increasing the likelihood of each of the samples.
The sample complexity and computational complexity of simpleHC is as follows, with proof
in the Supplement.
Theorem 2.1. Consider the hard-core model (2.1) on a graph G = (V, E) on |V | = p nodes
and with maximum degree d. The sample complexity of simpleHC is
n = O((2?)2d?2 log p) ,
4
(2.2)
i.e. with this many samples the algorithm simpleHC correctly reconstructs the graph with
probability 1 ? o(1). The computational complexity is
O(np2 ) = O((2?)2d?2 p2 log p) .
(2.3)
We next show that the sample complexity bound in Theorem 2.1 is basically tight:
Theorem 2.2 (Sample complexity lower bound). Consider the hard-core model (2.1). There
is a family of graphs on p nodes with maximum degree d such that for the probability of
successful reconstruction to approach one, the number of samples must scale as
1
p2
n=
(2?)2d log
.
d
Lemma 2.3. Suppose edge e = (i, j) ?
/ G, and let I be an independent set chosen according
to the Gibbs distribution (2.1). Then P({i, j} ? I) ? (9 ? max{1, (2?)2d?2 })?1 , ? .
The Supplementary Material contains proofs for Theorem 2.2 and Lemma 2.3.
3
Learning anti-ferromagnetic Ising models
In this section we consider the anti-ferromagnetic Ising model on a graph G = (V, E). We
parametrize the model in such a way that each configuration has probability
)
*
1
P(?) = exp H(?) , ? ? {0, 1}p ,
(3.1)
Z
where
?
?
H(?) = ??
?i ?j +
hi ? i .
(3.2)
(i,j)?E
i?V
Here ? > 0 and {hi }i?V are real-valued parameters, and we assume that |hi | ? h for all i.
Working with configurations in {0, 1}p rather than the more typical {?1, +1}p amounts to
a reparametrization (which is without loss of generality as shown for example in Appendix 1
of [25]). Setting hi = h = ln ? for all i, we recover the hard-core model with fugacity ? in
the limit ? ? ?, so we think of (3.2) as a ?soft? independent set model.
3.1
Strongly antiferromagnetic models
We start by considering the situation in which the repelling strength ? is sufficiently large
that we can modify the approach used for the hard-core model. We require some notation
to work with conditional probabilities: for each vertex b ? V , let
(i)
and
Bb = {? (i) : ?b = 1} ,
? a = 1|?b = 1) := 1 |{i ? B : ? (i) = 1}| .
P(?
a
|B|
!
"
? a = 1|?b = 1) = P(?a = 1|?b = 1). The algorithm, described next,
Of course, E P(?
? to a threshold.
determines whether each edge {a, b} is present based on comparing P
Algorithm 2 StrongRepelling
?
Input: ?, h, d, and n samples ? (1) , . . . , ? (n) ? {0, 1}p . Output: edge set E.
d h(d?1) ?2
1: Let ? = (1 + 2 e
)
! "
2: FOR each possible edge {a, b} ? V2 :
?
3: IF P? (?a = 1|?b = 1) ? (1 + e??h )?1 + ? THEN add edge (a, b) to E
?
4: OUTPUT E
Algorithm StrongRepelling obtains the following performance. The proof of Proposition 3.1 is similar to that of Theorem 2.1, replacing Lemma 2.3 by Lemma 3.2 below.
5
Proposition 3.1. Consider the antiferromagnetic Ising model (3.2) on a graph G = (V, E)
on p nodes and with maximum degree d. If
? ? d(h + ln 2) ,
then algorithm StrongRepelling has sample complexity
1
2
n = O 22d e2h(d+1) log p ,
i.e. this many samples are sufficient to reconstruct the graph with probability 1 ? o(1). The
computational complexity of StrongRepelling is
1
2
O(np2 ) = O 22d e2h(d+1) p2 log p .
When the interaction parameter ? ? d(h+ln 2) it is possible to identify edges using pairwise
statistics. The next lemma, proved in the Supplement, shows the desired separation.
Lemma 3.2. We have the following estimates:
(i) If (a, b) ?
/ E(G), then P(?a = 1|?b = 1) ?
1
1+2deg(a) eh(deg(a)+1)
(ii) Conversely, if (a, b) ? E(G), then P(?a = 1|?b = 1) ?
(ii) For any b ? V , P(?b = 1) ?
3.2
1
1+2deg(b) eh(deg(b)+1)
.
1
1+e??h
.
.
Weakly antiferromagnetic models
In this section we focus on learning weakly repelling models and show a trade-off between
computational complexity and strength of the repulsion. Recall that for strongly repelling
models our algorithm has run-time O(p2 log p), the same as for the hard-core model (infinite
repulsion).
For a subset of nodes U ? V , let G\U denote the graph obtained from G by removing nodes
in U (as well as any edges incident to nodes in U ). The following corollary is immediate
from Lemma 3.2.
Corollary 3.3. We have the conditional probability estimates for deleting subsets of nodes:
(i) If (a, b) ?
/ E(G), then for any subset of nodes U ? V \ {a, b},
PG\U (?a = 1|?b = 1) ?
1
1+2
degG\U (a) h(degG\U (a)+1)
e
.
(ii) Conversely, if (a, b) ? E(G), then for any subset of nodes U ? V \ {a, b}
PG\U (?a = 1|?b = 1) ?
1
.
1 + e??h
We can effectively remove nodes from the graph by conditioning: The family of models (3.2)
has the property that conditioning on ?i = 0 amounts to removing node i from the graph.
Fact 3.4 (Self-reducibility). Let G = (V, E), and consider the model 3.2. Then for any
subset of nodes U ? V , the probability law PG (? ? ? |?U = 0) is equal to PG\U (?V \U ? ? ).
The final ingredient is to show that we can condition by restricting attention to a subset of
the observed data, ? (1) , . . . , ? (n) , without throwing away too many samples.
Lemma 3.5. Let U ? V be a subset of nodes and denote the subset of samples with variables
(i)
?U equal to zero by AU = {? (i) : ?u = 0 for all u ? U }. Then with probability at least
h 2|U |
1 ? exp(n/2(1 + e )
) the number |AU | of such samples is at least n2 ? (1 + eh )?|U | .
We now present the algorithm. Effectively, it reduces node degree by removing nodes (which
can be done by conditioning on value zero), and then applies the strong repelling algorithm
to the residual graph.
6
Algorithm 3 WeakRepelling
?
Input: ?, h, d, and n samples ? (1) , . . . , ? (n) ? {0, 1}p . Output: edge set E.
1: Let ? = (1 + 2d eh(d?1) )?2
! "
2: FOR each possible edge (a, b) ? V2 :
3: FOR each U ? V \ {a, b} of size |U | ? ?d ? ?/(h + ln 2)?
4:
Compute P?G\U (?a = 1|?b = 1)
?
5: IF minU :|U |= P?G\U (?a = 1|?b = 1) ? (1 + e??h ) + ? THEN add edge (a, b) to E
?
6: OUTPUT E
Theorem 3.6. Let ? be a nonnegative integer strictly smaller than d, and consider the
antiferromagnetic Ising model 3.2 with
? ? (d ? ?)(h + ln 2)
on a graph G. Algorithm WeakRepelling reconstructs the graph with probability 1 ? o(1)
as p ? ? using
1
2
n = O (1 + eh )? 22d e2h(d+1) log p
i.i.d. samples, with run-time
4
!
"
?h,d (p2+? ) .
O np2+? = O
Statistical algorithms and proof of Theorem 1.1
We start by describing the statistical algorithm framework introduced by [1]. In this section
it is convenient to work with variables taking values in {?1, +1} rather than {0, 1}.
4.1
Background on statistical algorithms
Let X = {?1, +1}p denote the space of configurations and let D be a set of distributions
over X . Let F be a set of solutions (in our case, graphs) and Z : D ? 2F be a map taking
each distribution D ? D to a subset of solutions Z(D) ? F that are defined to be valid
solutions for D. In our setting, since each graphical model is identifiable, there is a single
graph Z(D) corresponding to each distribution D. For n > 0, the distributional search
problem Z over D and F using n samples is to find a valid solution f ? Z(D) given access
to n random samples from an unknown D ? D.
The class of algorithms we are interested in are called unbiased statistical algorithms, defined
by access to an unbiased oracle. Other related classes of algorithms are defined in [1], and
similar lower bounds can be derived for those as well.
Definition 4.1 (Unbiased Oracle). Let D be the true distribution. The algorithm is given
access to an oracle, which when given any function h : X ? {0, 1}, takes an independent
random sample x from D and returns h(x).
These algorithms access the sampled data only through the oracle: unbiased statistical
algorithms outsource the computation. Because the data is accessed through the oracle, it
is possible to prove unconditional lower bounds using information-theoretic methods. As
noted in the introduction, many algorithmic approaches can be implemented as statistical
algorithms.
We now define a key quantity called average correlation. The average correlation of a subset
of distributions D? ? D relative to a distribution D is denoted ?(D? , D),
> ? -= D1
1
D2
?
?(D , D) := ? 2
(4.1)
- D ? 1, D ? 1 - ,
|D |
D1 ,D2 ?D ?
D
where ?f, g?D := Ex?D [f (x)g(x)] and the ratio D1 /D represents the ratio of probability
mass functions, so (D1 /D)(x) = D1 (x)/D(x).
We quote the definition of statistical dimension with average correlation from [1], and then
state a lower bound on the number of queries needed by any statistical algorithm.
7
Definition 4.2 (Statistical dimension). Fix ? > 0, ? > 0, and search problem Z over set
of solutions F and class of distributions D over X. We consider pairs (D, DD ) consisting
of a ?reference distribution? D over X and a finite set of distributions DD ? D with the
following property: for any solution f ? F, the set Df = DD \ Z ?1 (f ) has size at least
(1 ? ?) ? |DD |. Let ?(D, DD ) be the largest integer ? so that for any subset D? ? Df with
|D? | ? |Df |/?, the average correlation is |?(D? , D)| < ? (if there is no such ? one can take
? = 0). The statistical dimension with average correlation ? and solution set bound ? is
defined to be the largest ?(D, DD ) for valid pairs (D, DD ) as described, and is denoted by
SDA(Z, ?, ?).
Theorem 4.3 ([1]). Let X be a domain and Z a search problem over a set of solutions F
and a class of distributions D over X . For ? > 0 and ? ? (0, 1), let ? = SDA(Z, ?, ?). Any
(possibly randomized) unbiased statistical algorithm that solves Z with probability ? requires
at least m calls to the Unbiased Oracle for
;
<
?(? ? ?) (? ? ?)2
m = min
,
.
2(1 ? ?)
12?
In particular, if ? ? 1/6, then any algorithm with success probability at least 2/3 requires at
least min{?/4, 1/48?} samples from the Unbiased Oracle.
In order to show that a graphical model on p nodes of maximum degree d requires
computation p (d) in this computational model, we therefore would like to show that
SDA(Z, ?, ?) = p (d) with ? = p? (d) .
4.2
Soft parities
r
For any subset S ? [p] of cardinality |S| = d, let ?S (x) = i?S xi be the parity of variables
in S. Define a probability distribution by assigning mass to x ? {?1, +1}p according to
1
pS (x) = exp(c ? ?S (x)) .
(4.2)
Z
Here c is a constant, and the partition function is
?
Z=
exp(c ? ?S (x)) = 2p?1 (ec + e?c ) .
(4.3)
x
Our
!p" family of distributions D is given by these soft parities over subsets S ? [p], and |D| =
d . The following lemma, proved in the supplementary material, computes correlations
between distributions.
Lemma 4.4. Let U denote the uniform distribution on {?1, +1}p . For S ?= T , the correlation ? pUS ? 1, pUT ? 1? is exactly equal to zero for any value of c. If S = T , the correlation
? pUS ? 1, pUS ? 1? = 1 ? (ec +e4?c )2 ? 1.
Lemma 4.5. For any set D? ? D of size at least |D|/pd/2 , the average correlation satisfies
?(D? , U ) ? dd p?d/2 .
Proof. By the preceding lemma, the only contributions to the sum (4.1) comes from choosing
?
the same set S in the sum, of which there are a fraction 1/|D
such correlation is at
! " |. Each
?
d/2
d/2 p
most one !by" Lemma 4.4, so ? ? 1/|D | ? p /|D| = p / d ? dd /pd/2 . Here we used the
estimate nk ? ( nk )k .
Proof of Theorem 1.1. Let ? = 1/6 and ? = dd p?d/2 , and consider the set of distributions
D given by soft parities as defined above. With reference distribution D = U , the uniform
distribution, Lemma 4.5 implies that SDA(Z, ?, ?) of the structure learning problem over
distribution (4.2) is at least ? = pd/2 /dd . The result follows from Theorem 4.3.
Acknowledgments
This work was supported in part by NSF grants CMMI-1335155 and CNS-1161964, and by
Army Research Office MURI Award W911NF-11-1-0036.
8
References
[1] V. Feldman, E. Grigorescu, L. Reyzin, S. Vempala, and Y. Xiao, ?Statistical algorithms and a
lower bound for detecting planted cliques,? in STOC, pp. 655?664, ACM, 2013.
[2] G. Bresler, E. Mossel, and A. Sly, ?Reconstruction of Markov random fields from samples:
Some observations and algorithms,? Approximation, Randomization and Combinatorial Optimization, pp. 343?356, 2008.
[3] J. Bento and A. Montanari, ?Which graphical models are difficult to learn?,? in NIPS, 2009.
[4] P. Abbeel, D. Koller, and A. Y. Ng, ?Learning factor graphs in polynomial time and sample
complexity,? The Journal of Machine Learning Research, vol. 7, pp. 1743?1788, 2006.
[5] I. Csisz?
ar and Z. Talata, ?Consistent estimation of the basic neighborhood of markov random
fields,? The Annals of Statistics, pp. 123?145, 2006.
[6] N. P. Santhanam and M. J. Wainwright, ?Information-theoretic limits of selecting binary
graphical models in high dimensions,? Info. Theory, IEEE Trans. on, vol. 58, no. 7, pp. 4117?
4134, 2012.
[7] M. Kearns, ?Efficient noise-tolerant learning from statistical queries,? Journal of the ACM
(JACM), vol. 45, no. 6, pp. 983?1006, 1998.
[8] C. Chow and C. Liu, ?Approximating discrete probability distributions with dependence trees,?
Information Theory, IEEE Transactions on, vol. 14, no. 3, pp. 462?467, 1968.
[9] S. Dasgupta, ?Learning polytrees,? in Proceedings of the Fifteenth conference on Uncertainty
in artificial intelligence, pp. 134?141, Morgan Kaufmann Publishers Inc., 1999.
[10] N. Srebro, ?Maximum likelihood bounded tree-width markov networks,? in Proceedings of the
Seventeenth conference on Uncertainty in artificial intelligence, pp. 504?511, Morgan Kaufmann Publishers Inc., 2001.
[11] R. L. Dobrushin, ?Prescribing a system of random variables by conditional distributions,?
Theory of Probability & Its Applications, vol. 15, no. 3, pp. 458?486, 1970.
[12] R. L. Dobrushin and S. B. Shlosman, ?Constructive criterion for the uniqueness of gibbs field,?
in Statistical physics and dynamical systems, pp. 347?370, Springer, 1985.
[13] J. Salas and A. D. Sokal, ?Absence of phase transition for antiferromagnetic potts models via
the dobrushin uniqueness theorem,? Journal of Statistical Physics, vol. 86, no. 3-4, pp. 551?579,
1997.
[14] D. Gamarnik, D. A. Goldberg, and T. Weber, ?Correlation decay in random decision networks,?
Mathematics of Operations Research, vol. 39, no. 2, pp. 229?261, 2013.
[15] D. Gamarnik and D. Katz, ?Correlation decay and deterministic fptas for counting listcolorings of a graph,? in Proceedings of the eighteenth annual ACM-SIAM symposium on
Discrete algorithms, pp. 1245?1254, Society for Industrial and Applied Mathematics, 2007.
[16] D. Weitz, ?Counting independent sets up to the tree threshold,? in Proceedings of the thirtyeighth annual ACM symposium on Theory of computing, pp. 140?149, ACM, 2006.
[17] P. Netrapalli, S. Banerjee, S. Sanghavi, and S. Shakkottai, ?Greedy learning of markov network
structure,? in 48th Allerton Conference, pp. 1295?1302, 2010.
[18] A. Ray, S. Sanghavi, and S. Shakkottai, ?Greedy learning of graphical models with small
girth,? in 50th Allerton Conference, 2012.
[19] A. Anandkumar, V. Tan, F. Huang, and A. Willsky, ?High-dimensional structure estimation
in Ising models: Local separation criterion,? Ann. of Stat., vol. 40, no. 3, pp. 1346?1375, 2012.
[20] R. Wu, R. Srikant, and J. Ni, ?Learning loosely connected Markov random fields,? Stochastic
Systems, vol. 3, no. 2, pp. 362?404, 2013.
[21] P. Ravikumar, M. Wainwright, and J. Lafferty, ?High-dimensional Ising model selection using
?1 -regularized logistic regression,? The Annals of Statistics, vol. 38, no. 3, pp. 1287?1319, 2010.
[22] S.-I. Lee, V. Ganapathi, and D. Koller, ?Efficient structure learning of markov networks using
l 1-regularization,? in Advances in neural Information processing systems, pp. 817?824, 2006.
[23] A. Jalali, C. C. Johnson, and P. D. Ravikumar, ?On learning discrete graphical models using
greedy methods.,? in NIPS, pp. 1935?1943, 2011.
[24] A. Jalali, P. Ravikumar, V. Vasuki, S. Sanghavi, and U. ECE, ?On learning discrete graphical
models using group-sparse regularization,? in Inter. Conf. on AI and Statistics (AISTATS),
vol. 14, 2011.
[25] A. Sinclair, P. Srivastava, and M. Thurley, ?Approximation algorithms for two-state antiferromagnetic spin systems on bounded degree graphs,? Journal of Statistical Physics, vol. 155,
no. 4, pp. 666?686, 2014.
9
| 5319 |@word polynomial:1 achievable:2 stronger:1 d2:2 pg:4 thereby:1 liu:2 contains:2 configuration:3 selecting:1 amp:1 existing:1 comparing:1 repelling:14 yet:1 assigning:1 must:1 partition:3 remove:2 aside:1 greedy:4 fewer:1 intelligence:2 beginning:1 core:11 short:1 detecting:1 node:25 allerton:2 accessed:1 symposium:2 prove:2 ray:2 pairwise:4 inter:1 hardness:1 indeed:1 roughly:1 behavior:1 pf:1 considering:2 increasing:1 cardinality:1 underlying:5 bounded:6 notation:2 mass:3 what:2 substantially:1 safely:1 every:3 finance:1 runtime:1 exactly:1 grant:1 before:3 negligible:1 declare:1 local:2 cdp:7 modify:1 limit:2 incoherence:1 au:2 suggests:2 challenging:1 polytrees:2 conversely:2 range:1 seventeenth:1 acknowledgment:1 procedure:1 significantly:1 convenient:1 cannot:3 close:1 selection:1 put:1 writing:1 seminal:1 restriction:1 equivalent:1 map:1 deterministic:1 eighteenth:1 straightforward:1 attention:1 convex:2 unstructured:1 assigns:1 his:1 searching:1 annals:2 construction:2 suppose:2 tan:1 programming:1 goldberg:1 sparsely:1 ising:16 distributional:1 muri:1 observed:2 ferromagnetic:8 cycle:1 connected:2 decrease:2 trade:1 mentioned:1 pd:7 complexity:24 exhaustively:1 weakly:3 tight:1 fast:1 gbresler:1 monte:1 query:4 sc:1 artificial:2 neighborhood:6 choosing:1 exhaustive:3 whose:1 apparent:1 larger:1 supplementary:2 valued:1 reconstruct:2 statistic:4 think:2 bento:4 final:1 reconstruction:2 interaction:7 neighboring:2 reyzin:1 csisz:1 p:1 depending:1 stat:1 school:1 strong:2 netrapalli:1 recovering:1 implemented:2 p2:5 come:1 implies:1 solves:1 stochastic:1 material:2 require:12 abbeel:2 generalization:1 fix:1 hypertrees:1 randomization:1 proposition:2 strictly:2 hold:3 sufficiently:2 considered:1 exp:4 great:1 minu:1 algorithmic:3 ditional:1 uniqueness:2 estimation:3 combinatorial:1 quote:1 largest:2 correctness:1 mit:1 always:1 rather:3 office:1 corollary:2 np2:3 derived:1 focus:3 potts:3 likelihood:4 industrial:1 sokal:1 repulsion:3 prescribing:1 chow:2 fptas:1 koller:3 interested:2 provably:1 classification:1 among:1 denoted:2 exponent:3 retaining:2 special:1 fairly:1 field:5 once:1 equal:4 having:1 ng:2 biology:1 represents:1 broad:1 cancel:1 others:3 sanghavi:3 salas:1 few:1 phase:1 consisting:1 cns:1 investigate:1 semidefinite:1 light:1 unconditional:2 behind:1 chain:1 edge:17 necessary:1 tree:6 loosely:1 ruled:1 desired:1 soft:4 ar:1 w911nf:1 lattice:1 cost:1 vertex:1 entry:1 subset:14 uniform:2 successful:1 johnson:1 too:1 dependency:1 answer:2 referring:1 sda:4 randomized:1 siam:1 lee:1 off:1 physic:3 again:1 reconstructs:3 huang:1 possibly:1 guy:1 sinclair:1 external:1 conf:1 return:1 ganapathi:1 inc:2 sloan:1 explicitly:1 depends:1 performed:1 doing:1 nondegeneracy:1 uncond:1 start:2 recover:1 reparametrization:1 weitz:1 contribution:1 ni:1 spin:1 kaufmann:2 efficiently:2 identify:1 basically:1 carlo:1 notoriously:1 nonstandard:1 definition:3 hardcore:1 pp:22 proof:6 sampled:2 proved:3 massachusetts:1 ask:2 popular:1 recall:1 knowledge:2 actually:1 coloring:1 focusing:1 improved:1 erd:1 done:1 strongly:5 generality:1 sly:4 correlation:34 working:1 replacing:1 o:1 banerjee:1 paraphrasing:1 logistic:3 verify:1 true:5 unbiased:7 regularization:2 assigned:1 laboratory:1 leibler:1 nonzero:1 deal:1 self:1 width:1 whereby:1 noted:1 criterion:2 theoretic:2 temperature:2 weber:1 wise:1 gamarnik:3 recently:1 multinomial:1 conditioning:3 exponentially:1 extend:1 katz:1 feldman:3 gibbs:2 ai:1 mathematics:2 cancellation:1 had:1 access:6 etc:1 add:2 pu:3 isometry:1 recent:1 showed:2 hide:1 belongs:1 certain:1 binary:1 success:1 morgan:2 additional:1 preceding:1 determine:1 ii:3 reduces:1 exceeds:1 long:1 ravikumar:4 equally:1 award:1 impact:1 basic:3 regression:3 df:3 fifteenth:1 background:2 publisher:2 subject:1 undirected:3 lafferty:1 anandkumar:2 integer:3 structural:2 call:1 counting:2 identically:1 variety:1 independence:1 gave:1 restrict:1 opposite:2 idea:1 whether:1 motivated:1 interpolates:1 shakkottai:2 remark:1 informally:1 amount:2 nsf:1 srikant:1 talata:1 correctly:1 discrete:7 shall:1 vol:12 dasgupta:1 santhanam:2 group:1 key:1 threshold:3 graph:44 asymptotically:1 fraction:1 sum:3 run:5 parameterized:1 uncertainty:2 family:3 wu:2 separation:2 decision:2 prefer:1 appendix:1 bound:12 hi:4 guaranteed:2 nonnegative:1 identifiable:1 oracle:7 strength:6 annual:2 throwing:1 min:2 pruned:1 vempala:1 department:1 according:2 smaller:1 reconstructing:1 restricted:1 grigorescu:1 ln:6 devavrat:2 turn:1 describing:1 fail:1 needed:1 tractable:2 serf:1 informal:2 parametrize:1 operation:1 apply:1 observe:1 away:2 generic:2 v2:2 distinguished:1 outsource:1 include:2 graphical:20 giving:1 approximating:1 society:1 question:7 quantity:1 planted:2 primary:1 cmmi:1 dependence:1 jalali:2 said:1 exhibit:3 dp:1 distance:4 landmark:1 argue:1 willsky:1 assuming:1 ratio:2 convincing:1 difficult:4 mostly:1 hypertree:1 stoc:1 info:1 negative:2 unknown:1 perform:2 observation:1 markov:7 finite:1 anti:3 gas:1 affirmatively:1 immediate:1 situation:1 introduced:4 david:1 pair:2 required:2 learned:1 tremendous:1 nip:2 trans:1 below:2 dynamical:1 appeared:1 regime:2 including:2 max:1 deleting:1 wainwright:3 difficulty:1 warm:1 regularized:2 eh:5 indicator:2 residual:1 representing:1 improve:1 technology:1 mossel:4 fugacity:2 reducibility:1 degg:2 relative:1 law:1 loss:1 bresler:4 highlight:1 interesting:1 srebro:1 ingredient:1 degree:14 incident:1 vasuki:1 sufficient:2 consistent:2 xiao:1 dd:11 course:1 supported:1 parity:6 last:1 understand:1 institute:1 neighbor:3 taking:2 sparse:1 distributed:1 benefit:1 dimension:4 valid:3 transition:1 computes:2 ec:2 social:1 transaction:1 bb:1 pruning:2 obtains:1 kullback:1 clique:2 ml:1 deg:4 tolerant:1 e2h:3 xi:1 search:7 decade:1 additionally:1 learn:11 nature:1 domain:2 aistats:1 main:1 montanari:4 noise:3 n2:1 fails:1 exponential:3 candidate:2 renyi:1 theorem:16 removing:3 e4:1 showing:1 decay:17 normalizing:1 intractable:1 restricting:3 adding:1 effectively:2 importance:1 supplement:2 nk:2 logarithmic:1 girth:1 likely:1 army:1 jacm:1 applies:1 springer:1 srivastava:1 antiferromagnetic:14 determines:1 satisfies:1 acm:5 conditional:3 goal:2 ann:1 absence:1 hard:11 typical:1 infinite:1 reducing:1 impractically:1 kearns:3 lemma:14 called:5 ece:1 select:1 dobrushin:3 constructive:1 d1:5 ex:1 |
4,771 | 532 | A Cortico-Cerebellar Model that Learns to
Generate Distributed Motor Commands to
Control a Kinematic Arm
N.E. Berthier S.P. Singh A.G. Barto
Department of Computer Science
University of Massachusetts
Amherst, MA 01002
.T.C. Honk
Department of Physiology
Northwestern University Medical School
Chicago, IL 60611
Abstract
A neurophysiologically-based model is presented that controls a simulated
kinematic arm during goal-directed reaches. The network generates a
quasi-feedforward motor command that is learned using training signals
generated by corrective movements. For each target, the network selects
and sets the output of a subset of pattern generators. During the movement, feedback from proprioceptors turns off the pattern generators. The
task facing individual pattern generators is to recognize when the arm
reaches the target and to turn off. A distributed representation of the motor command that resembles population vectors seen in vivo was produced
naturally by these simulations.
1
INTRODUCTION
We have recently begun to explore the properties of sensorimotor networks with
architectures inspired by the anatomy and physiology of the cerebellum and its interconnections with the red nucleus and the motor cortex (Houk 1989; Houk et al..
611
612
Berthier, Singh, Barto, and Houk
1990). It is widely accepted that these brain regions are important in the control
of limb movements (Kuypers, 1981; Ito, 1984), although relatively little attention
has been devoted to probing how the different regions might function together in
a cooperative manner. Starting from a foundation of known anatomical circuitry
and the results of microelectrode recordings from neurons in these circuits, we proposed the concept of rubrocerebellar and corticocerebellar information processing
modules that are arranged in parasagittal arrays and function as adjustable pattern
generators (APGs) capable of the storage, recall and execution of motor programs.
The aim of the present paper is to extend the APG Model to a multiple degreeof-freedom task and to investigate how the motor representation developed by the
model compares to the population vector representations seen by Georgopoulos
and coworkers (e.g., Georopoulos, 1988). A complete description of the model and
simulations reported here is contained in Berthier et al. (1991).
2
THE APG ARRAY MODEL
As shown in Figure 1 the model has three parts: a neural network that generates
control signals, a muscle model that controls joint angle, and a planar, kinematic
arm. The control network is an array of APGs that generate signals that are
fed to the limb musculature. Because here we are interested in the basic issue of
how a collection of APGs might cooperatively control multiple degree-of-freedom
movements, we use a very simplified model of the limb that ignores dynamics. The
muscles convert APG activity to changes in muscle length, which determine the
changes in the joint angles. Activation of an APG causes movement of the arm in
a direction in joint-angle space that is specific to that APG 1 , and the magnitude
of an APG's activity determines the velocity of that movement. The simultaneous
activation of selected APGs determines the arm trajectory as the superposition of
these movements. A learning rule, based on long-term depression (e.g., Ito, 1984),
adjusts the subsets of APGs that are selected as well as characteristics of their
activity in order to achieve desired movements .
Each APG consists of a positive feedback loop and a set of Purkinje cells (PCs).
The positive feedback loop is a highly simplified model of a component of a complex
cerebrocerebellar recurrent network. In the simplified model simulated here, each
APG has its own feedback loop, and the loops associated with different APGs do
not interact. When triggered by sufficiently strong activation, the neurons in these
loops fire repetitively in a self-sustaining manner. An APG's motor command is
generated through the action of its PCs which inhibit and modulate the buildup of
activity in the feedback loop. The activity of loop cells is conveyed to spinal motor
areas by rubrospinal fibers. PCs receive information that specifies and constrains
the desired movements via parallel fibers.
We hypothesize that the response of PCs to particular parallel fiber inputs is adaptively adjusted through the influence of climbing fibers that respond to corrective
movements (Houk & Barto, 1991). The APG array model assumes that climbing
fibers and PCs are aligned in a way that climbing fibers provide specialized inforITo simplify these initial simulations we ignore changes in muscle moment arms with
posture of the arm.
A Cortico-Cerebellar Model that Learns to Generate Distributed Motor Commands
Network
...........................,
APG Modules
?l?
1
Muscles
T
?
m
T
?
??
................................... .1
M
Figure 1: APG Control of Joint Angles. A collection of of APGs (adjustable pattern
generators) is connected to a simulated two degree-of-freedom, kinematic, planar
arm with antagonistic muscles at each joint. The task is to move the arm in the
plane from a central starting location to one of eight symmetrically placed targets.
Activation of an APG causes a movement of the arm that is specific to that APG,
and the magnitude of an APG's activity determines the velocity of that movement.
The simultaneous activation of selected APGs determines the arm trajectory as a
superposition of these movements.
mation to PCs. Gellman et al. (1985) showed that proprioceptive climbing fibers
are inhibited during planned movements, but the data of Gilbert and Thach (1977)
suggest that they fire during corrective movements. In the present simulations, we
assume that corrective movements are made when a movement fails to reach the
target. These corrective movements stimulate proprioceptive climbing fibers which
provides information to higher centers about the direction of the corrective movement. More detailed descriptions of APGs and relevant anatomy and physiology
can be found in Houk (1989), Houk et al. (1990), and Berthier et al. (1991).
The generation of motor commands occurs in three phases. In the first phase, we
assume that all positive feedback loops are off, and inputs provided by teleceptive
and proprioceptive parallel fibers and basket cells determine the outputs of the PCs.
We call this first phase selection. We assume that noise is present during the selection process so that individual PCs are turned off (Le., selected) probabilistic ally.
To begin the second phase, called the execution phase, loop activity is triggered by
cortical activity. Once triggered, loop activity is self-sustaining because the loop
cells have reciprocal positive connections. The triggering of loop activity causes the
motor command to be "read out." The states of the PCs in the selection phase
determine the speed and direction of the arm movement. As the movement is being performed, proprioceptive feedback and efference copy gradually depolarize the
PCs. When a large proportion of the PCs are depolarized, PC inhibition reaches a
critical value and terminates loop activity. In the third phase, the correction phase,
corrective movements trigger climbing fiber activity that alters parallel fiber-PC
connection weights.
613
614
Berthier, Singh, Barto, and Houk
A
11-: ....
w..J
I.
'.
B
[J" ....
.
:::
,
II
??
rI
'
.,: ? I.:'
. '. '.II I..
.::.
I,'
..' .1- ...
.. ?... ........,:1.
n,.??
L.!I
..? r
?
I.
I
?
I
? :.
?
~.....
.~
..'
.".:.
.1 ?
I' ?
~
c1:
?
?
u
...
~
.
?
'.
~
~"?
tI , , ~ ?
LJ
I"
'
~
'
.... ,
..
?
6
II
'0
Figure 2: A. Movement Trajectories After Training. The starting point for each
movement is the center of the workspace, and the target location is the center of
the open square. The position of the arm at each time step is shown as a dot.
Three movements are shown to each target. B. APG selection. APG selection for
movements to a given target is illustrated by a vector plot at the position of the
target. An individual APG is represented by a vector, the direction of which is
equal to the direction of movement caused by that APG in Cartesian space. The
vector length is proportional to output of the Purkinje cells during the selection
phase. The arrow points in the direction of the vector sum.
3
SIMULATIONS
We trained the APG model to control a two degree-of-freedom, kinematic, planar
arm. The task was similar to Georgopoulos (1988) and required APGs to move the
arm from a central starting point to one of eight radially symmetric, equidistant
targets. Each simulated trial started by placing the endpoint of the arm in the central starting location. The selection, execution, and correction phases of operation
were then simulated. The task facing each of the selected APGs was to turn off at
the proper time so that the movement stopped at the target.
Simulations showed that the model could learn to control movements to the eight
targets. Training typically required about 700 trials per target until the arm endpoint was consistently moved to within 1 em of the target. Figure 2 shows sample
trajectories and population vectors of APG activity. Performance never resulted
in precise movements due to the probabilistic nature of selection. Movement trajectories tended to follow straight lines in joint-angle space and were thus slightly
curved lines in the workspace. About half of the APGs in the model were used
to move to an individual target with population vectors similar to those seen by
Georgopoulos (1988). The number of APGs used for each target was dependent
on the sharpness of the climbing fiber receptive fields, with cardioid shaped receptive fields in joint-angle space giving population vectors that most resembled those
experimentally observed.
A Carrico-Cerebellar Model that Learns to Generate Distributed Motor Commands
4
ANALYSIS
In order to understand how the model worked we undertook a theoretical analysis of
its simulated behavior. Analysis indicated that the expected trajectory of a movement was a straight line in joint-angle space from the starting position to the target.
This is a special case of a mathematical result by Mussa-Ivaldi (1988). Because selection is probabilistic in the APG Array Model, trajectories in the workspace varied
from the expected trajectory. In these cases, trajectories were piecewise linear because of the asynchronous termination of APG activity. Because of the Law of
Large Numbers, the more PCs in each APG, the more closely the movement will
resemble the expected movement.
The expected population of vectors of APG activity can be shown to be cosineshaped in joint-angle space. That is, the length of the vector representing the
activity of APG m is proportional to the cosine of the angle between the direction
of action of APG m and the direction of the target in joint-angle space. The shape
of the population vectors in Cartesian space is dependent on the Jacobian of the
arm, which is a function of the arm posture.
The manner in which the outputs of PCs were set during selection leads to scaling of
movement velocity with target distance. For any given movement direction, targets
that are farther from the starting location lead to more rapid movements than closer
targets.
Updating network weights based on the expected corrective movement will, in some
cases, result in changing the weights in a way that they converge to the correct
values. However, in other cases inappropriate changes are made. In the current
simulations, we could largely avoid this problem by selecting parameter and initial
weight values so that movements were initially small in amplitude. Random initialization of the weight values sometimes led to instances from which the learning rule
could not recover.
5
DISCUSSION
In general, the present implementation of the modelled to adequate control of the
kinematic arm and mimicked the general output of nervous system seen in actual
experiments. The network implemented a spatial to temporal transformation that
transformed a target location into a time varying motor command. The model
naturally generated population vectors that were similar to those seen in vivo. Further research is needed to improve the model's robustness and to extend it to more
realistic control of a dynamical limb.
In the APG array model, APGs control arm movement in parallel so that the activity of all the modules taken together forms a distributed representation. The APG
array executes a distributed motor program because it produces a spatiotemporal
pattern of activity in the cerebrocerebellar recurrent network that is transmitted to
the spinal cord to comprise a distributed motor command.
615
616
Berthier, Singh, Barto, and Houk
5.1
PARAMETRIZED MOTOR PROGRAMS
Certain features of the APG array model relate well to the ideas about parameterized motor programs discussed by Keele (1973), Schmidt (1988), and Adams (1971,
1977). The selection phase of the APG array model provides a feasible neuronal
mechanism for preparing a parameterized motor program in advance of movement.
The execution phase is also consistent with the open-loop ideas associated with
motor programming concepts, except that, like Adams (1977), we explain the termination of the execution phase as being a consequence of proprioceptive feedback
and efference copy.
In the APG array model, the counterpart of a generalized motor program is a
set of parallel fiber weights for proprioceptive, efference copy, and target inputs.
Given these weights, a particular constellation of parallel fiber inputs signifies that
the desired endpoint of a movement is about to be reached, causing PCs to become
depolarized. Once a set of parallel fiber weights corresponding to a desired endpoint
is learned, the neuronal architecture and neurodynamics of the cerebellar network
functions in a manner that parameterizes the motor program.
Movement velocity is parameterized in the selection phase of the model's operation.
The velocity that is selected is automatically scaled so that velocity increases as the
amplitude of the movement increases. While this type of scaling is often observed
in motor performance studies, velocity can also be varied in an independent manner where velocity scaling can be applied simultaneously to all elements of a motor
program to slow down or speed up the entire movement. Although we have not addressed this issue in the present report, simulation of velocity scaling under control
of a neuromodulator can naturally be accomplished in the APG array model.
Movements terminate when the endpoint is recognized by PCs so that movement
duration is dependent on the course of the movement instead of being determined by
some internal clock because. Movement amplitude is parameterized by the weights
of the target inputs, with smaller weights corresponding to larger amplitude movements.
5.2
CORRECTIVE MOVEMENTS
We assume that the training information conveyed to the APGs is the result of crude
corrective movements stimulating proprioceptive receptors. This sensory information is conveyed to the cerebellum by climbing fibers. Learning in the APG array
model therefore requires the existence of a low-level system capable of generating
movements to spatial targets with at least a ballpark level of accuracy. Lesion (Yu
et al., 1980) and developmental studies (von Hofsten, 1982) support the existence
of a low-level system. Other evidence indicates that when limb movements are not
proceeding accurately toward their intended targets, corrective components of the
movements are generated by an unconscious, automatic control system (Goodale et
aI., 1986).
We assume that collaterals from the corticospinal and rubrospinal system that convey the motor commands to the spinal cord gate off sensory transmission through
the proprioceptive climbing fiber pathway, thus preventing sensory responses to the
initial limb movement. As the initial movement proceeds, the low-level system re-
A Corrico-Cerebellar Model that Learns to Generate Distributed Motor Commands
ceives proprioceptive feedback from the limb and feedforward information about
target location from the gaze control system. The latter information is updated as
a consequence of corrective eye movements that typically occur after an initial gaze
shift toward a visual target. Updated gaze information causes the spinal processor to generate a corrective component that is superimposed on the original motor
command (Gielen & van Gisbergen, 1990; Flash & Henis, 1991). Since climbing
fiber pathways would not be gated off by this low-level corrective process, climbing
fibers should fire to indicate the direction of the corrective movement.
We assume that the network by which climbing fiber activity is generated is specifically wired to provide appropriate training information to the APGs (Houk &
Barto, 1991). The training signal provided by a climbing fiber is specialized for the
recipient APG in that it provides directional information in joint-angle space that is
relative to the direction in which that APG moves the arm. The fact that training
information is provided in terms of joint-angle space greatly simplifies the problem
of providing errors in the correct system of reference. For example, if the network
used visual error information, the error information would have to be transformed
to joint errors.
The specialized training signals provided by the climbing fibers are determined by
the structure of the ascending network conveying proprioceptive information. This
ascending network has the same structure-but works in the opposite direction-as
the network by which the APG array influences joint movement. This is reminiscent of the error backpropagation algorithm (e.g., Rumelhart et al., 1986, Parker,
1985) where the forward and backward passes through the network in the backpropagation algorithm are accomplished by the descending and ascending networks
of the APG Array Model. This use of the ascending network to transform errors in
the workspace to errors that are relative to a particular APG's direction of action
is closely related to the use of error backpropagation for "learning with a distal
teacher" as suggested by Jordan and Rumelhart (1991).
Houk and Barto (1991) suggested that the alignment of the ascending and descending networks might come about through trophic mechanisms stimulated by
use-dependent alterations in synaptic efficacy. In the context of the present model,
this hypothesis implies that the ascending network to the inferior olive, is established first, and that the descending network by which APGs influence motoneurons
changes. We have not yet simulated this mechanism to see if it could actually generate the kind of alignment we assume in the present model.
Acknowledgements
This research was supported by ONR N00014-88-K-0339, NIMH Center Grant P50
MH48185, and a grant from the McDonnell-Pew Foundation for Cognitive Neuroscience supported by the James S. McDonnell Foundation and the Pew Charitable
Trusts.
References
Ac!ams JA (1971) A closed-loop theory of motor learning. J Motor Beh 3: 111-149
Adams J A (1977) Feedback theory of how joint receptors regulate the timing and
positioning of a limb. Psychol Rev 84: 504-523
617
618
Berthier, Singh, Barto, and Houk
Berthier NE Singh SP Barto AG Houk JC (1991) Distributed representation of
limb motor programs in arrays of adjustable pattern generators. NPB Technical
Report 3, Institute for Neuroscience, Northwestern University, Chicago IL
Flash T Henis E (1991) Arm trajectory modifications during reaching towards
visual targets. J Cognitive Neurosci 3:220-230
Gellman R Gibson AR Houk JC (1985) Inferior olivary neurons in the awake cat:
Detection of contact and passive body displacement. J Neurophys 54:40-60.
Georgopoulos A (1988) Neural integration of movement: role of motor cortex in
reaching. FASEB Journal 2:2849-2857.
Gielen CCAM Gisbergen van JAM (1990) The visual guidance of saccades and
fast aiming movements. News in Physiol Sci 5: 58-63
Gilbert PFC Thach WT (1977) Purkinje cell activity during motor learning. Brain
Res 128:309-328.
Goodale MA Pelisson D Prablanc C (1986) Large adjustments in visually guided
reaching do not depend on vision of the hand or perception of target displacement. Nature 320: 748-750
Hofsten von C (1982) Eye-hand coordination in the newborn. Dev Psycho I 18:
450-461
Houk JC (1989) Cooperative control of limb movements by the motor cortex,
brainstem and cerebellum. In: Cotterill RMJ (ed) Models of Brain Function.
Cambridge Univ Press Cambridge UK, 309-325
Houk JC Barto AG (1991) Distributed sensorimotor learning. NPB Technical
Report 1, Institute for Neuroscience, Northwestern University, Chicago IL
Houk JC Singh SP Fisher C Barto AG (1990) An adaptive sensorimotor network
inspired by the anatomy and physiology of the cerebellum. In: Miller WT Sutton RS Werbos PJ (eds) Neural Networks for Control. MIT Press Cambridge,
MA 301-348
Ito M (1984) The Cerebellum and Neural Control. Raven Press New York
Ito M (1989) Long-term depression. Annual review of Neuroscience 12: 85-102
Jordan MI Rumelhart DE (1991) Forward models: Supervised learning with a
distal teacher. Occasional Paper #40 MIT Center for Cognitive Science
Keele SW (1973) Attention and Human Performance. Goodyear Pacific Palisades,
California
Kuypers HGJM (1981) Anatomy of the descending pathways. In: Brooks VB (ed)
Handbook of Physiology Section I Volume II Part 1. American Physiological
Society Bethesda MD 597-666
Mussa-Ivaldi FA (1988) Do neurons in the motor cortex encode movement direction? An alternative hypothesis. Neurosci Lett 91:106-111
Parker DB (1985) Learning-Logic. Technical Report TR-47, Massachusetts Institute of Technology Cambridge MA
Rumelhart DE Hinton GE Williams RJ (1986) Learning internal representations
by error propagation. In: Rumelhart DE McClelland JL (eds) Parallel Distributed Processing. Explorations in the Microstructure of Cognition, Vol. 1:
Foundations. Bradford Books/MIT Press Cambridge MA
Schmidt RA (1988) Motor Control and Motor Learning. Human Kinetics Champaign, Illinois
| 532 |@word trial:2 proportion:1 open:2 termination:2 simulation:8 r:1 tr:1 moment:1 ivaldi:2 initial:5 efficacy:1 selecting:1 current:1 neurophys:1 activation:5 yet:1 reminiscent:1 olive:1 physiol:1 realistic:1 chicago:3 berthier:8 shape:1 motor:35 hypothesize:1 plot:1 half:1 selected:6 nervous:1 plane:1 undertook:1 reciprocal:1 farther:1 proprioceptor:1 provides:3 location:6 mathematical:1 become:1 consists:1 pathway:3 manner:5 ra:1 expected:5 rapid:1 behavior:1 brain:3 inspired:2 automatically:1 little:1 actual:1 inappropriate:1 provided:4 begin:1 circuit:1 ballpark:1 kind:1 developed:1 ag:3 transformation:1 temporal:1 ti:1 olivary:1 scaled:1 uk:1 control:20 medical:1 grant:2 positive:4 timing:1 consequence:2 aiming:1 receptor:2 sutton:1 might:3 initialization:1 resembles:1 sustaining:2 corticospinal:1 directed:1 backpropagation:3 displacement:2 area:1 gibson:1 physiology:5 kuypers:2 suggest:1 selection:12 storage:1 context:1 influence:3 descending:4 gilbert:2 center:5 williams:1 attention:2 starting:7 duration:1 sharpness:1 rule:2 adjusts:1 array:15 population:8 antagonistic:1 updated:2 target:29 trigger:1 unconscious:1 programming:1 hypothesis:2 velocity:9 element:1 rumelhart:5 updating:1 werbos:1 cooperative:2 observed:2 role:1 module:3 region:2 cord:2 connected:1 news:1 hofsten:2 movement:64 inhibit:1 developmental:1 nimh:1 constrains:1 goodale:2 dynamic:1 trophic:1 trained:1 singh:7 depend:1 joint:15 represented:1 fiber:22 corrective:15 cat:1 univ:1 fast:1 widely:1 efference:3 larger:1 interconnection:1 transform:1 triggered:3 causing:1 aligned:1 loop:15 relevant:1 turned:1 achieve:1 description:2 moved:1 transmission:1 produce:1 generating:1 adam:3 wired:1 recurrent:2 ac:1 school:1 strong:1 implemented:1 resemble:1 indicate:1 come:1 implies:1 direction:14 guided:1 anatomy:4 closely:2 honk:1 correct:2 exploration:1 brainstem:1 human:2 gellman:2 jam:1 ja:1 microstructure:1 adjusted:1 cooperatively:1 correction:2 kinetics:1 sufficiently:1 houk:16 visually:1 cognition:1 circuitry:1 superposition:2 coordination:1 mit:3 mation:1 aim:1 reaching:3 avoid:1 varying:1 command:13 barto:11 newborn:1 encode:1 consistently:1 indicates:1 superimposed:1 greatly:1 am:1 dependent:4 lj:1 typically:2 entire:1 initially:1 psycho:1 quasi:1 transformed:2 selects:1 interested:1 microelectrode:1 issue:2 spatial:2 special:1 integration:1 equal:1 apgs:17 once:2 never:1 field:2 shaped:1 comprise:1 preparing:1 placing:1 yu:1 report:4 simplify:1 inhibited:1 piecewise:1 npb:2 simultaneously:1 recognize:1 resulted:1 individual:4 phase:14 mussa:2 intended:1 fire:3 freedom:4 detection:1 investigate:1 kinematic:6 highly:1 cotterill:1 alignment:2 pc:17 devoted:1 capable:2 closer:1 collateral:1 desired:4 re:2 guidance:1 theoretical:1 stopped:1 instance:1 purkinje:3 planned:1 ar:1 dev:1 signifies:1 subset:2 reported:1 teacher:2 spatiotemporal:1 adaptively:1 amherst:1 workspace:4 probabilistic:3 off:7 gaze:3 together:2 von:2 central:3 cognitive:3 book:1 american:1 de:3 alteration:1 jc:5 caused:1 performed:1 closed:1 red:1 reached:1 recover:1 parallel:9 vivo:2 il:3 square:1 accuracy:1 characteristic:1 largely:1 miller:1 conveying:1 climbing:14 directional:1 modelled:1 accurately:1 produced:1 trajectory:10 straight:2 executes:1 processor:1 simultaneous:2 explain:1 reach:4 tended:1 basket:1 synaptic:1 ed:4 sensorimotor:3 james:1 naturally:3 associated:2 mi:1 radially:1 massachusetts:2 begun:1 recall:1 amplitude:4 actually:1 higher:1 supervised:1 follow:1 planar:3 response:2 henis:2 arranged:1 until:1 clock:1 hand:2 ally:1 trust:1 propagation:1 indicated:1 stimulate:1 depolarize:1 concept:2 counterpart:1 read:1 symmetric:1 proprioceptive:10 illustrated:1 distal:2 cerebellum:5 during:9 self:2 inferior:2 cosine:1 generalized:1 complete:1 p50:1 passive:1 recently:1 specialized:3 spinal:4 endpoint:5 volume:1 jl:1 extend:2 discussed:1 rubrospinal:2 cambridge:5 ai:1 pew:2 automatic:1 illinois:1 dot:1 cortex:4 inhibition:1 own:1 showed:2 certain:1 n00014:1 onr:1 accomplished:2 muscle:6 seen:5 transmitted:1 motoneuron:1 recognized:1 determine:3 converge:1 coworkers:1 signal:5 ii:4 multiple:2 rj:1 champaign:1 positioning:1 technical:3 repetitively:1 long:2 basic:1 vision:1 cerebellar:5 sometimes:1 cell:6 c1:1 receive:1 rmj:1 addressed:1 depolarized:2 pass:1 recording:1 db:1 jordan:2 call:1 symmetrically:1 feedforward:2 equidistant:1 architecture:2 triggering:1 opposite:1 idea:2 parameterizes:1 simplifies:1 shift:1 buildup:1 york:1 cause:4 action:3 depression:2 adequate:1 detailed:1 mcclelland:1 generate:7 specifies:1 alters:1 neuroscience:4 per:1 anatomical:1 vol:1 changing:1 pj:1 backward:1 convert:1 musculature:1 sum:1 angle:12 parameterized:4 respond:1 thach:2 scaling:4 vb:1 apg:39 annual:1 activity:20 occur:1 worked:1 georgopoulos:4 ri:1 awake:1 generates:2 speed:2 relatively:1 department:2 pacific:1 mcdonnell:2 terminates:1 slightly:1 em:1 cardioid:1 smaller:1 bethesda:1 rev:1 modification:1 gradually:1 taken:1 turn:3 mechanism:3 needed:1 ge:1 fed:1 ascending:6 operation:2 eight:3 limb:10 occasional:1 palisade:1 appropriate:1 regulate:1 mimicked:1 robustness:1 schmidt:2 gate:1 alternative:1 existence:2 original:1 recipient:1 assumes:1 sw:1 giving:1 society:1 contact:1 move:4 posture:2 occurs:1 receptive:2 fa:1 md:1 distance:1 simulated:7 sci:1 parametrized:1 toward:2 length:3 providing:1 relate:1 implementation:1 proper:1 adjustable:3 gated:1 gisbergen:2 neuron:4 curved:1 hinton:1 precise:1 beh:1 varied:2 required:2 connection:2 california:1 learned:2 established:1 brook:1 suggested:2 proceeds:1 dynamical:1 pattern:7 perception:1 program:9 critical:1 arm:24 representing:1 improve:1 technology:1 eye:2 ne:1 started:1 psychol:1 review:1 acknowledgement:1 relative:2 law:1 neurophysiologically:1 northwestern:3 generation:1 proportional:2 facing:2 generator:6 foundation:4 nucleus:1 degree:3 conveyed:3 consistent:1 charitable:1 parasagittal:1 neuromodulator:1 course:1 placed:1 supported:2 copy:3 asynchronous:1 cortico:2 understand:1 institute:3 distributed:11 van:2 feedback:10 lett:1 cortical:1 ignores:1 sensory:3 made:2 collection:2 preventing:1 simplified:3 forward:2 adaptive:1 ignore:1 keele:2 logic:1 handbook:1 neurodynamics:1 stimulated:1 learn:1 nature:2 terminate:1 interact:1 complex:1 pfc:1 sp:2 neurosci:2 arrow:1 noise:1 lesion:1 convey:1 body:1 neuronal:2 parker:2 slow:1 probing:1 fails:1 position:3 crude:1 third:1 jacobian:1 ito:4 learns:4 down:1 specific:2 resembled:1 constellation:1 physiological:1 evidence:1 raven:1 magnitude:2 execution:5 faseb:1 cartesian:2 led:1 gielen:2 explore:1 visual:4 contained:1 adjustment:1 saccade:1 determines:4 ma:5 stimulating:1 modulate:1 goal:1 flash:2 towards:1 fisher:1 feasible:1 change:5 experimentally:1 determined:2 except:1 specifically:1 wt:2 called:1 bradford:1 accepted:1 internal:2 support:1 latter:1 |
4,772 | 5,320 | On Sparse Gaussian Chain Graph Models
Seyoung Kim
Lane Center for Computational Biology
Carnegie Mellon University
[email protected]
Calvin McCarter
Machine Learning Department
Carnegie Mellon University
[email protected]
Abstract
In this paper, we address the problem of learning the structure of Gaussian chain
graph models in a high-dimensional space. Chain graph models are generalizations of undirected and directed graphical models that contain a mixed set of directed and undirected edges. While the problem of sparse structure learning has
been studied extensively for Gaussian graphical models and more recently for
conditional Gaussian graphical models (CGGMs), there has been little previous
work on the structure recovery of Gaussian chain graph models. We consider linear regression models and a re-parameterization of the linear regression models
using CGGMs as building blocks of chain graph models. We argue that when the
goal is to recover model structures, there are many advantages of using CGGMs
as chain component models over linear regression models, including convexity of
the optimization problem, computational efficiency, recovery of structured sparsity, and ability to leverage the model structure for semi-supervised learning. We
demonstrate our approach on simulated and genomic datasets.
1
Introduction
Probabilistic graphical models have been extensively studied as a powerful tool for modeling a set
of conditional independencies in a probability distribution [12]. In this paper, we are concerned with
a class of graphical models, called chain graph models, that has been proposed as a generalization of
undirected graphical models and directed acyclic graphical models [4, 9, 14]. Chain graph models
are defined over chain graphs that contain a mixed set of directed and undirected edges but no
partially directed cycles.
In particular, we study the problem of learning the structure of Gaussian chain graph models in a
high-dimensional setting. While the problem of learning sparse structures from high-dimensional
data has been studied extensively for other related models such as Gaussian graphical models
(GGMs) [8] and more recently conditional Gaussian graphical models (CGGMs) [17, 20], to our
knowledge, there is little previous work that addresses this problem for Gaussian chain graph models. Even with a known chain graph structure, current methods for parameter estimation are hindered
by the presence of multiple locally optimal solutions [1, 7, 21].
Since the seminal work on conditional random fields (CRFs) [13], a general recipe for constructing
chain graph models [12] has been given as using CRFs as building blocks for the model. We employ
this construction for Gaussian chain graph models and propose to use the recently-introduced sparse
CGGMs [17, 20] as a Gaussian equivalent of general CRFs. When the goal is to learn the model
structure, we show that this construction is superior to the popular alternative approach of using
linear regression as component models. Some of the key advantages of our approach are due to the
fact that the sparse Gaussian chain graph models inherit the desirable properties of sparse CGGM
such as convexity of the optimization problem and structured output prediction. In fact, our work is
the first to introduce a joint estimation procedure for both the graph structure and parameters as a
convex optimization problem, given the groups of variables for chain components. Another advan1
xj
3 - xj
2
xj
3
xj
2
xj
3 ..2
. xj
xj
3 - xj
2
...
..
R
xj
1
xj
3
xj
2
xj
3 ..2
. xj
.....
..
.
.
.
. R1
xj
xj
xj
xj
xj
xj
1
1
4
1
4 . - xj
4 - xj
1
(a)
(b)
(c)
(d)
(e)
(f)
Figure 1: Illustration of chain graph models. (a) A chain graph with two components, {x1 , x2 } and
{x3 }. (b) The moralized graph of the chain graph in (a). (c) After inference in the chain graph
in (a), inferred indirect dependencies are shown as the dotted line. (d) A chain graph with three
components, {x1 , x2 }, {x3 }, and {x4 }. (e) The moralized graph of the chain graph in (d). (f) After
inference in the chain graph in (d), inferred indirect dependencies are shown as the dotted lines.
tage of our approach is the ability to model a functional mapping from multiple related variables to
other multiple related variables in a more natural way via moralization in chain graphs than other
approaches that rely on complex penalty functions for inducing structured sparsity [11, 15].
Our work on sparse Gaussian chain graphs is motivated by problems in integrative genomic data
analyses [6, 18]. While sparse GGMs have been extremely popular for learning networks from
datasets of single modality such as gene-expression levels [8], we propose that sparse Gaussian chain
graph models with CGGM components can be used to learn a cascade of networks by integrating
multiple types of genomic data in a single statistical analysis. We show that our approach can
reveal the module structures as well as the functional mapping between modules in different types
of genomic data effectively. Furthermore, as the cost of collecting each data type differs, we show
that semi-supervised learning can be used to make effective use of both fully-observed and partiallyobserved data.
2
Sparse Gaussian Chain Graph Models
We consider a chain graph model for a probability distribution over J random variables x =
{x1 , . . . , xJ }. The chain graph model assumes that the random variables are partitioned into C
chain components {x1 , . . . , xC }, the ? th component having size |? |. In addition, it assumes a partially directed graph structure, where edges between variables within each chain component are
undirected and edges across two chain components are directed. Given this chain graph structure,
the joint probability distribution factorizes as follows:
p(x) =
C
Y
p(x? |xpa(? ) ),
? =1
where xpa(? ) is the set of variables that are parents of one or more variables in x? . Each factor
p(x? |xpa(? ) ) models the conditional distribution of the chain component variables x? given xpa(? ) .
This model can also be viewed as being constructed with CRFs for p(x? |xpa(? ) )?s [13].
The conditional independence properties of undirected and directed graphical models have been
extended to chain graph models [9, 14]. This can be easily seen by first constructing a moralized
graph, where undirected edges are added between any pairs of nodes in xpa(? ) for each chain component ? and all the directed edges are converted into undirected edges (Figure 1). Then, subsets of
variables xa and xb are conditionally independent given xc , if xa and xb are separated by xc in the
moralized graph. This conditional independence criterion for a chain graph is called c-separation
and generalizes d-separation for Bayesian networks [12].
In this paper, we focus on Gaussian chain graph models, where both p(x) and p(x? |xpa(? ) )?s are
Gaussian distributed. Below, we review linear regression models and CGGMs as chain component
models, and introduce our approach for learning chain graph model structures.
2.1
Sparse Linear Regression as Chain Component Model
As the specific functional form of p(x? |xpa(? ) ) in Gaussian chain graphs models, a linear regression
model with multivariate responses has been widely considered [2, 3, 7]:
p(x? |xpa(? ) ) = N (B? xpa(? ) , ??1
? ),
|? |?|pa(? )|
(1)
where B? ? R
is the matrix of regression coefficients and ?? is the |? | ? |? | inverse
covariance matrix that models correlated noise. Then, the non-zero elements in B? indicate the
2
presence of directed edges from xpa(? ) to x? , and the non-zero elements in ?? correspond to the
undirected edges among the variables in x? . When the graph structure is known, an iterative procedure has been proposed to estimate the model parameters, but it converges only to one of many
locally-optimal solutions [7].
When the chain component model has the form of Eq. (1), in order to jointly estimate the sparse
graph structure and the parameters, we adopt sparse multivariate regression with covariance estimation (MRCE) [16] for each chain component and solve the following optimization problem:
min
C
X
tr((X? ?Xpa(? ) BT? )?? (X? ?Xpa(? ) BT? )T )?N log |?? | +?
? =1
C
X
||B? ||1 + ?
? =1
C
X
||?? ||1 ,
? =1
where X? ? RN ?|?| is a dataset for N samples, || ? ||1 is the sparsity-inducing L1 penalty, and ?
and ? are the regularization parameters that control the amount of sparsity in the parameters. As in
MRCE [16], the problem above is not convex, but only bi-convex.
2.2
Sparse Conditional Gaussian Graphical Model as Chain Component Model
As an alternative model for p(x? |xpa(? ) ) in Gaussian chain graph models, a re-parameterization of
the linear regression model in Eq. (1) with natural parameters has been considered [14]. This model
also has been called a CGGM [17] or Gaussian CRF [20] due to its equivalence to a CRF. A CGGM
for p(x? |xpa(? ) ) takes the standard form of undirected graphical models as a log-linear model:
1
p(x? |xpa(? ) ) = exp ? xT? ?? x? ? xT? ??,pa(? ) xpa(? ) /A(xpa(? ) ),
(2)
2
where ?? ? R|? |?|? | and ??,pa(? ) ? R|? |?|pa(? )| are the parameters for the feature weights between
pairs of variables within x? and between pairs of variables across x? and xpa(? ) , respectively, and
A(xpa(? ) ) is the normalization constant. The non-zero elements of ?? and ??,pa(? ) indicate edges
among the variables in x? and between x? and xpa(? ) , respectively.
The linear regression model in Eq. (1) can be viewed as the result of performing inference in the
probabilistic graphical model given by the CGGM in Eq. (2). This relationship between the two
models can be seen by re-writing Eq. (2) in the form of a Gaussian distribution:
?1
p(x? |xpa(? ) ) = N (???1
? ??,pa(? ) xpa(? ) , ?? ),
where marginalization in a CGGM involves computing B? xpa(? ) =
a linear regression model parameterized by B? .
???1
? ??,pa(? ) xpa(? )
(3)
to obtain
In order to estimate the graph structure and parameters for Gaussian chain graph models with CGGMs as chain component models, we adopt the procedure for learning a sparse CGGM [17, 20] and
minimize the negative log-likelihood of data along with sparsity-inducing L1 penalty:
min ?L(X; ?) + ?
C
X
||??,pa(? ) ||1 + ?
? =1
C
X
||?? ||1 ,
? =1
where ? = {?? , ??,pa(? ) , ? = 1, . . . , C} and L(X; ?) is the data log-likelihood for dataset X ?
RN ?J for N samples. Unlike MRCE, the optimization problem for a sparse CGGM is convex,
and efficient algorithms have been developed to find the globally-optimal solution with substantially
lower computation time than that for MRCE [17, 20].
While maximum likelihood estimation leads to the equivalent parameter estimates for CGGMs and
linear regression models via the transformation B? = ???1
? ??,pa(? ) , imposing a sparsity constraint on each model leads to different estimates for the sparsity pattern of the parameters and the
model structure [17]. The graph structure of a sparse CGGM directly encodes the probabilistic dependencies among the variables, whereas the sparsity pattern of B? = ???1
? ??,pa(? ) obtained after
marginalization can be interpreted as indirect influence of covariates xpa(? ) on responses x? . As illustrated in Figures 1(c) and 1(f), the CGGM parameters ??,pa(? ) (directed edges with solid line)
can be interpreted as direct dependencies between pairs of variables across x? and xpa(? ) , whereas
B? = ???1
? ??,pa(? ) obtained from inference can be viewed as indirect and inferred dependencies
(directed edges with dotted line).
3
We argue in this paper that when the goal is to learn the model structure, performing the estimation
with CGGMs for chain component models can lead to a more meaningful representation of the
underlying structure in data than imposing a sparsity constraint on linear regresssion models. Then
the corresponding linear regression model can be inferred via marginalization. This approach also
inherits many of the advantages of sparse CGGMs such as convexity of optimization problem.
2.3
Markov Properties and Chain Component Models
When a CGGM is used as the component model, the overall chain graph model is known to have
Lauritzen-Wermuth-Frydenberg (LWF) Markov properties [9]. The LWF Markov properties also
correspond to the standard probabilistic independencies in more general chain graphs constructed
by using CRFs as building blocks [12].
Many previous works have noted that LWF Markov properties do not hold for the chain graph models with linear regression models [2, 3]. The alternative Markov properties (AMP) were therefore
introduced as the set of probabilistic independencies associated with chain graph models with linear
regression component models [2, 3]. It has been shown that the LWF and AMP Markov properties are equivalent only for chain graph structures that do not contain the graph in Figure 1(a) as a
subgraph [2, 3]. For example, according to the LWF Markov property, in the chain graph model in
Figure 1(a), x1 ? x3 |x2 as x1 and x3 are separated by x2 in the moralized graph in Figure 1(b).
However, the corresponding AMP Markov property implies a different probabilistic independence
relationship, x1 ? x3 . In the model in Figure 1(d), according to the LWF Markov property, we have
x1 ? x3 |{x2 , x4 }, whereas the AMP Markov property gives x1 ? x3 |x4 .
We observe that when using sparse CGGMs as chain component models, we estimate a model with
the LWF Markov properties and perform marginalization in this model to obtain a model with linearregression chain components that can be interpreted with the AMP Markov properties.
3
Sparse Two-Layer Gaussian Chain Graph Models for Structured Sparsity
Another advantage of using CGGMs as chain component models instead of linear regression is
that the moralized graph, which is used to define the LWF Markov properties, can be leveraged to
discover the underlying structure in a correlated functional mapping from multiple inputs to multiple
outputs. In this section, we show that a sparse two-layer Gaussian chain graph model with CGGM
components can be used to learn structured sparsity. The key idea behind our approach is that
while inference in CGGMs within the chain graph model can reveal the shared sparsity patterns for
multiple related outputs, a moralization of the chain graph can reveal those for multiple inputs.
Statistical methods for learning models with structured sparsity were extensively studied in the literature of multi-task learning, where the goal is to find input features that influence multiple related
outputs simultaneously [5, 11, 15]. Most of the previous works assumed the output structure to be
known a priori. Then, they constructed complex penalty functions that leverage this known output structure, in order to induce structured sparsity pattern in the estimated parameters in linear
regression models. In contrast, a sparse CGGM was proposed as an approach for performing a joint
estimation of the output structure and structured sparsity for multi-task learning. As was discussed
in Section 2.2, once the CGGM structure is estimated, the inputs relevant for multiple related outputs
could be revealed via probabilistic inference in the graphical model.
While sparse CGGMs focused on leveraging the output structure for improved predictions, another
aspect of learning structured sparsity is to consider the input structure to discover multiple related
inputs jointly influencing an output. As CGGM is a discriminative model that does not model the
input distribution, it is unable to capture input relatedness directly, although discriminative models
in general are known to improve prediction accuracy. We address this limitation of CGGMs by
embedding CGGMs within a chain graph and examining the moralized graph.
We set up a two-layer Gaussian chain graph model for inputs x and outputs y as follows:
1
1
p(y, x) = p(y|x)p(x) = exp(? yT ?yy y ? xT ?xy y)/A1 (x)
exp(? xT ?xx x)/A2 ,
2
2
where a CGGM is used for p(y|x) and a GGM for p(x), and A1 (x) and A2 are normalization constants. As the full model factorizes into two factors p(y|x) and p(x) with distinct sets of parameters,
4
a sparse graph structure and parameters can be learned by using the optimization methods for sparse
CGGM [20] and sparse GGM [8, 10].
The estimated Gaussian chain graph model leads to a GGM over both the inputs and outputs, which
reveals the structure of the moralized graph:
?1 !
?yy
?Txy
p(y, x) = N 0,
.
T
?xy ?xx + ?xy ??1
yy ?xy
In the above GGM, we notice that the graph structure over inputs x consists of two components,
one for ?xx describing the conditional dependencies within the input variables and another for
T
?xy ??1
yy ?xy that reflects the results of moralization in the chain graph. If the graph ?yy contains
T
connected components, the operation ?xy ??1
yy ?xy for moralization induces edges among those
inputs influencing the outputs in each connected component.
Our approach is illustrated in Figure 2.
yl
yl
yl
yl
yl
yl
yl
yl
yl
yl
1
2
3
4
5
1
2
3
4
5
Given the model in Figure 2(a), Figure
.
.
. ... . . . . .1...
.
.. 3
Y
..
..
..3
.
.
.
.
. .A
. .
2(b) illustrates the inferred structured
I
K.A....
AKA
.. .
.... .
.. ..
.
sparsity for a functional mapping from
xl
xl
xl
xl
xl
xl
xl
xl
xl
xl
xl
xl
1
2
3
4
5
6
1
2
3
4
5
6
multiple inputs to multiple outputs. In
(a)
(b)
Figure 2(b), the dotted edges correspond
to inferred indirect dependencies intro- Figure 2: Illustration of sparse two-layer Gaussian chain
duced via marginalization in the CGGM graphs with CGGMs. (a) A two-layer Gaussian chain
p(y|x), which reveals how each input graph. (b) The results of performing inference and moralis influencing multiple related outputs. ization in (a). The dotted edges correspond to indirect deOn the other hand, the additional edges pendencies inferred by inference. The edges among xj ?s
among xj ?s have been introduced by represent the dependencies introduced by moralization.
T
moralization ?xy ??1
yy ?xy for multiple inputs jointly influencing each output. Combining the results of marginalization and moralization, the two connected components in Figure 2(b) represent
the functional mapping from {x1 , x2 } to {y1 , y2 } and from {x3 , x4 , x5 } to {y3 , y4 , y5 }, respectively.
4
Sparse Multi-layer Gaussian Chain Graph Models
In this section, we extend the two-layer Gaussian chain graph model from the previous section into
a multi-layer model to model data that are naturally organized into multiple layers. Our approach is
motivated by problems in integrative genomic data analysis. In order to study the genetic architecture of complex diseases, data are often collected for multiple data types, such as genotypes, gene
expressions, and phenotypes for a population of individuals [6, 18]. The primary goal of such studies
is to identify the genotype features that influence gene expressions, which in turn influence phenotypes. In such problems, data can be naturally organized into multiple layers, where the influence of
features in each layer propagates to the next layer in sequence. In addition, it is well-known that the
expressions of genes within the same functional module are correlated and influenced by the common genotype features and that the coordinated expressions of gene modules affect multiple related
phenotypes jointly. These underlying structures in the genomic data can be potentially revealed by
inference and moralization in sparse Gaussian chain graph models with CGGM components.
In addition, we explore the use of semi-supervised learning, where the top and bottom layer data
are fully observed but the middle-layer data are collected only for a subset of samples. In our
application, genotype data and phenotype data are relatively easy to collect from patients? blood
samples and from observations. However, gene-expression data collection is more challenging, as
invasive procedure such as surgery or biopsy is required to obtain tissue samples.
4.1
Models
Given variables, x = {x1 , . . . , xJ }, y = {y1 , . . . , yK }, and z = {z1 , . . . , zL }, at each of the three
layers, we set up a three-layer Gaussian chain graph model as follows:
p(z, y|x) = p(z|y)p(y|x)
1 T
1 T
T
T
= exp(? z ?zz z ? y ?yz z)/C2 (y) exp(? y ?yy y ? x ?xy y)/C1 (x) , (4)
2
2
5
where C1 (x) and C2 (y) are the normalization constants. In our application, x, y, and z correspond
to genotypes, gene-expression levels, and phenotypes, respectively. As the focus of such studies
lies on discovering how the genotypic variability influences gene expressions and phenotypes rather
than the structure in genotype features, we do not model p(x) directly.
Given the estimated sparse model for Eq. (4), structured sparsity pattern can be recovered via
T
?1 T
inference and moralization. Computing Bxy = ???1
yy ?xy and Byz = ??zz ?yz corresponds
to performing inference to reveal how multiple related yk ?s in ?yy (or zl ?s in ?zz ) are jointly
influenced by a common set of relevant xj ?s (or yk ?s). On the other hand, the effects of moralization
can be seen from the joint distribution p(z, y|x) derived from Eq. (4):
T
?1
p(z, y|x) = N (???1
(zz,yy) ?(yz,xy) x, ?(zz,yy) ),
?zz
?Tyz
where ?(yz,xy) = (0J?L , ?xy ) and ?(zz,yy) =
. ?(zz,yy) corT
?yz ?yy + ?yz ??1
zz ?yz
responds to the undirected graphical model over z and y conditional on x after moralization.
4.2
Semi-supervised Learning
Given a dataset D = {Do , Dh }, where Do = {Xo , Yo , Zo } for the fully-observed data and Dh =
{Xh , Zh } for the samples with missing gene-expression levels, for semi-supervised learning, we
adopt an EM algorithm that iteratively maximizes the expected log-likelihood of complete data:
L(Do ; ?) + E L(Dh , Yh ; ?) ,
combined with L1 -regularization, where L(Do ; ?) is the data log-likelihood with respect to the
model in Eq. (4) and the expectation is taken with respect to:
p(y|z, x) = N (?y|x,z , ?y|x,z ),
T ?1
.
?y|x,z = ??y|x,z (?yz z + ?Txy x) and ?y|x,z = (?yy + ?yz ??1
zz ?yz )
5
Results
In this section, we empirically demonstrate that CGGMs are more effective components for sparse
Gaussian chain graph models than linear regression for various tasks, using synthetic and real-world
genomic datasets. We used the sparse three-layer structure for p(z, y|x) in all our experiments.
5.1
Simulation Study
In simulation study, we considered two scenarios for true models, CGGM-based and linearregression-based Gaussian chain graph models. We evaluated the performance in terms of graph
structure recovery and prediction accuracy in both supervised and semi-supervised settings.
In order to simulate data, we assumed the problem size of J=500, K=100, and L=50 for x, y, and
z, respectively, and generated samples from known true models. Since we do not model p(x), we
used an arbitrary choice of multinomial distribution to generate samples for x. The true parameters
for CGGM-based simulation were set as follows. We set the graph structure in ?yy to a randomlygenerated scale-free network with a community structure [19] with six communities. The edge
weights were drawn randomly from a uniform distribution [0.8, 1.2]. We then set ?yy to the graph
Laplacian of this network plus small positive values along the diagonal so that ?yy is positive
definite. We generated ?zz using a similar strategy, assuming four communities. ?xy was set to
a sparse random matrix, where 0.4% of the elements have non-zero values drawn from a uniform
distribution [-1.2,-0.8]. ?yz was generated using a similar strategy, with a sparsity level of 0.5%. We
set the sparsity pattern of ?yz so that it roughly respects the functional mapping from communities
in y to communities in z. Specifically, after reordering the variables in y and z by performing
hierarchical clustering on each of the two networks ?yy and ?zz , the non-zero elements were
selected randomly around the diagonal of ?yz .
We set the true parameters for the linear-regression-based models using the same strategy as the
CGGM-based simulation above for ?yy and ?zz . We set Bxy so that 50% of the variables in x
have non-zero influence on five randomly chosen variables in y in one randomly chosen community
in ?yy . We set Byz in a similar manner, assuming 80% of the variables in y are relevant to eight
randomly-chosen variables in z from a randomly-chosen community in ?zz .
6
1
0.8
0.6
0.4
CG?semi
CG
LR?semi
LR
0.5
Recall
0.2
0
0
0.6
0.4
0.2
1
0.6
0.4
0.2
0
0
0.5
Recall
(a)
1
Precision
1
0.8
Precision
1
0.8
Precision
1
0.8
Precision
Precision
1
0.8
0.6
0.4
0.2
0
0
0.5
Recall
(b)
1
0.6
0.4
0.2
0
0
0.5
Recall
(c)
1
0
0
(d)
0.5
Recall
1
(e)
Figure 4: Precision/recall curves for graph structure recovery in CGGM-based simulation study. (a)
?yy , (b) ?zz , (c) Bxy , (d) Byz , and (e) ?xy . (CG: CGGM-based models with supervised learning,
CG-semi: CG with semi-supervised learning, LR: linear-regression-based models with supervised
learning, LR-semi: LR with semi-supervised learning.)
1.2
5
3
2
0.6
1
0.8
0.4
1
LR
0.8
0.6
0.4
0.2
CG?semi CG LR?semi
test err
0.6
test err
0.8
1.2
1
4
test err
test err
1
CG?semi CG LR?semi
CG?semi CG LR?semi
LR
0.4
CG?semi CG LR?semi
LR
LR
(a)
(b)
(c)
(d)
Figure 5: Prediction errors in CGGM-based simulation study. The same estimated models in Figure
4 were used to predict (a) y given x, z, (b) z given x, (c) y given x, and (d) z given y.
0.4
0.2
0
0
1
1
0.8
0.8
0.6
0.4
0.2
0.5
Recall
1
0
0
Precision
Precision
0.6
1
0.8
Precision
CG?semi
CG
LR?semi
LR
Precision
1
0.8
0.6
0.4
0.2
0.5
Recall
0
0
1
0.6
0.4
0.2
0.5
Recall
1
0
0
0.5
Recall
1
(a)
(b)
(c)
(d)
Figure 6: Performance for graph structure recovery in linear-regression-based simulation study.
Precision/recall curves are shown for (a) ?yy , (b) ?zz , (c) Bxy , and (d) Byz .
Each dataset consisted of 600 samples, of which 400
and 200 samples were used as training and test sets.
To select the regularization parameters, we estimated
a model using 300 samples, evaluated prediction errors on the other 100 samples in the training set, and
selected the values with the lowest prediction errors.
We used the optimization methods in [20] for CGGMbased models and the MRCE procedure [16] for linearregression-based models.
(c)
(a)
(b)
(d)
(e)
Figure 3 illustrates how the model with CGGM chain Figure 3: Illustration of the structured sparcomponents can be used to discover the structured sity recovered by the model with CGGM
sparsity via inference and moralization. In each panel, components, simulated dataset. (a) ?zz .
T
black and bright pixels correspond to zero and non- (b) Byz = ???1
zz ?yz shows the effects of
zero values, respectively. While Figure 3(a) shows marginalization (white vertical bars). The
how variables in z are related in ?zz , Figure 3(b) effects of moralization are shown in (c)
T
?1 T
shows Byz = ???1
zz ?yz obtained via marginaliza- ?yy + ?yz ?zz ?yz , and its decomposiT
tion within the CGGM p(z|y), where functional map- tion into (d) ?yy and (e) ?yz ??1
zz ?yz .
pings from variables in y to multiple related variables
in z can be seen as white vertical bars. In Figure 3(c), the effects of moralization ?yy +
T
?1 T
?yz ??1
zz ?yz are shown, which further decomposes into ?yy (Figure 3(d)) and ?yz ?zz ?yz
(Figure 3(e)). The additional edges among variables in y in Figure 3(e) correspond to the edges
introduced via moralization and show the groupings of the variables y as the block structure along
the diagonal. By examining Figures 3(b) and 3(e), we can infer a functional mapping from modules
in y to modules in z.
In order to systematically compare the performance of the two types of models, we examined the
average performance over 30 randomly-generated datasets. We considered both supervised and
semi-supervised settings. Assuming that 200 samples out of the total 400 training samples were
7
30
40
1
0.5
CG?semi CG LR?semi
30
20
10
LR
CG?semi CG LR?semi
test err
1.5
2
test err
test err
test err
2
1.5
1
20
10
0.5
LR
CG?semi CG LR?semi
LR
0
CG?semi CG LR?semi
LR
(a)
(b)
(c)
(d)
Figure 7: Prediction errors in linear-regression-based simulation study. The same estimated models
in Figure 6 were used to predict (a) y given x, z, (b) z given x, (c) y given x, and (d) z given y.
missing data for y, for supervised learning, we used only those samples with complete data; for
semi-supervised learning, we used all samples, including partially-observed cases.
The precision/recall curves for recovering the true graph structures are shown in Figure 4, using
datasets simulated from the true models with CGGM components. Each curve was obtained as an
average over 30 different datasets. We observe that in both supervised and semi-supervised settings, the models with CGGM components outperform the ones with linear regression components.
In addition, the performance of the CGGM-based models improves significantly, when using the
partially-observed data in addition to the fully-observed samples (the curve for CG-semi in Figure 4), compared to using only the fully-observed samples (the curve for CG in Figure 4). This
improvement from using partially-observed data is substantially smaller for the linear-regressionbased models. The average prediction errors from the same set of estimated models in Figure 4 are
shown in Figure 5. The CGGM-based models outperform in all prediction tasks, because they can
leverage the underlying structure in the data and estimate models more effectively.
For the simulation scenario using the linear-regression-based true models, we show the results for
precision/recall curves and prediction errors in Figures 6 and 7, respectively. We find that even
though the data were generated from chain graph models with linear regression components, the
CGGM-based methods perform as well as or better than the other models.
5.2
Integrative Genomic Data Analysis
We applied the two types of three-layer chain graph Table 1: Prediction errors, mouse diabetes data
models to single-nucleotide-polymorphism (SNP),
gene-expression, and phenotype data from the pancreTask CG-semi CG LR-semi LR
atic islets study for diabetic mice [18]. We selected
y | x, z 0.9070 0.9996 1.0958 0.9671
200 islet gene-expression traits after performing hierz | x 1.0661 1.0585 1.0505 1.0614
archical clustering to find several gene modules. Our
y | x 0.8989 0.9382 0.9332 0.9103
dataset also included 1000 SNPs and 100 pancreatic
z | y 1.0712 1.0861 1.1095 1.0765
islet cell phenotypes. Of the total 506 samples, we
used 406 as training set, of which 100 were held out as a validation set to select regularization
parameters, and used the remaining 100 samples as test set to evaluate prediction accuracies. We
considered both supervised and semi-supervised settings, assuming gene expressions are missing
for 150 mice. In supervised learning, only those samples without missing gene expressions were
used.
As can be seen from the prediction errors in Table 1, the models with CGGM chain components are
more accurate in various prediction tasks. In addition, the CGGM-based models can more effectively
leverage the samples with partially-observed data than linear-regression-based models.
6
Conclusions
In this paper, we addressed the problem of learning the structure of Gaussian chain graph models
in a high-dimensional space. We argued that when the goal is to recover the model structure, using
sparse CGGMs as chain component models has many advantages such as recovery of structured
sparsity, computational efficiency, globally-optimal solutions for parameter estimates, and superior
performance in semi-supervised learning.
Acknowledgements
This material is based upon work supported by an NSF CAREER Award No. MCB-1149885, Sloan
Research Fellowship, and Okawa Foundation Research Grant.
8
References
[1] F. Abegaz and E. Wit. Sparse time series chain graphical models for reconstructing genetic
networks. Biostatistics, pages 586?599, 2013.
[2] S. Andersson, D. Madigan, and D. Perlman. An alternative Markov property for chain graphs.
In Proceedings of the 12th Conference on Uncertainty in Artificial Intelligence, pages 40?48.
Morgan Kaufmann, 1996.
[3] S. Andersson, D. Madigan, and D. Perlman. Alternative Markov properties for chain graphs.
Scandinavian Journal of Statistics, 28:33?85, 2001.
[4] W. Buntine. Chain graphs for learning. In Proceedings of the 11th Conference on Uncertainty
in Artificial Intelligence, pages 46?54. Morgan Kaufmann, 1995.
[5] X. Chen, X. Shi, X. Xu, Z. Wang, R. Mills, C. Lee, and J. Xu. A two-graph guided multi-task
lasso approach for eQTL mapping. In Proceedings of the 15th International Conference on
Artificial Intelligence and Statistics (AISTATS), volume 16. JMLR W&CP, 2012.
[6] Y. Chen, J. Zhu, P.K. Lum, X. Yang, S. Pinto, D.J. MacNeil, C. Zhang, J. Lamb, S. Edwards,
S.K. Sieberts, et al. Variations in DNA elucidate molecular networks that cause disease. Nature, 452(27):429?35, 2008.
[7] M. Drton and M. Eichler. Maximum likelihood estimation in Gaussian chain graph models
under the alternative Markov property. Scandinavian Journal of Statistics, 33:247?57, 2006.
[8] J. Friedman, T. Hastie, and R. Tibshirani. Sparse inverse covariance estimation with the graphical lasso. Biostatistics, 9(3):432?41, 2008.
[9] M. Frydenberg. The chain graph Markov property. Scandinavian Journal of Statistics, 17:
333?53, 1990.
[10] C.J. Hsieh, M. Sustik, I. Dhillon, and P. Ravikumar. Sparse inverse covariance matrix estimation using quadratic approximation. In Advances in Neural Information Processing Systems
(NIPS) 24, 2011.
[11] L. Jacob, G. Obozinski, and J. Vert. Group lasso with overlap and graph lasso. In Proceedings
of the 26th International Conference on Machine Learning, 2009.
[12] D. Koller and N. Friedman. Probabilistic Graphical Models: Principles and Techniques. The
MIT Press, 2009.
[13] J. Lafferty, A. McCallum, and F. Pereira. Conditional random fields: probabilistic models for
segmenting and labeling sequence data. In Proceedings of the 18th International Conference
on Machine Learning, 2001.
[14] S.L. Lauritzen and N. Wermuth. Graphical models for associations between variables, some
of which are qualitative and some quantitative. The Annals of Statistics, 17(1):31?57, 1989.
[15] G. Obozinski, M.J. Wainwright, and M.J. Jordan. High-dimensional union support recovery in
multivariate regression. In Advances in Neural Information Processing Systems 21, 2008.
[16] A. Rothman, E. Levina, and J. Zhu. Sparse multivariate regression with covariance estimation.
Journal of Computational and Graphical Statistics, 19(4):947?962, 2010.
[17] K.A. Sohn and S. Kim. Joint estimation of structured sparsity and output structure in multipleoutput regression via inverse-covariance regularization. In Proceedings of the 15th International Conference on Artificial Intelligence and Statistics (AISTATS), volume 16. JMLR
W&CP, 2012.
[18] Z. Tu, M.P. Keller, C. Zhang, M.E. Rabaglia, D.M. Greenawalt, X. Yang, I.M. Wang, H. Dai,
M.D. Bruss, P.Y. Lum, Y.P. Zhou, D.M. Kemp, C. Kendziorski, B.S. Yandell, A.D. Attie, E.E.
Schadt, and J. Zhu. Integrative analysis of a cross-loci regulation network identifies app as a
gene regulating insulin secretion from pancreatic islets. PLoS Genetics, 8(12):e1003107, 2012.
[19] J. Wu, Z. Gao, and H. Sun. Cascade and breakdown in scale-free networks with community
structure. Physical Review, 74:066111, 2006.
[20] M. Wytock and J.Z. Kolter. Sparse Gaussian conditional random fields: algorithms, theory,
and application to energy forecasting. In Proceedings of the 30th International Conference on
Machine Learning, volume 28. JMLR W&CP, 2013.
[21] J. Yin and H. Li. A sparse conditional Gaussian graphical model for analysis of genetical
genomics data. The annals of applied statistics, 5(4):2630, 2011.
9
| 5320 |@word middle:1 integrative:4 pancreatic:2 simulation:9 covariance:6 hsieh:1 jacob:1 tr:1 solid:1 contains:1 series:1 genetic:2 amp:5 cort:1 err:8 current:1 recovered:2 intelligence:4 discovering:1 selected:3 parameterization:2 mccallum:1 lr:25 node:1 zhang:2 five:1 along:3 constructed:3 direct:1 c2:2 qualitative:1 consists:1 manner:1 introduce:2 expected:1 roughly:1 multi:5 globally:2 little:2 discover:3 underlying:4 xx:3 maximizes:1 panel:1 biostatistics:2 lowest:1 interpreted:3 substantially:2 developed:1 transformation:1 quantitative:1 y3:1 collecting:1 control:1 zl:2 grant:1 segmenting:1 positive:2 influencing:4 black:1 plus:1 studied:4 examined:1 equivalence:1 collect:1 challenging:1 bi:1 directed:12 perlman:2 union:1 block:4 definite:1 differs:1 x3:8 procedure:5 cascade:2 significantly:1 vert:1 integrating:1 induce:1 madigan:2 influence:7 seminal:1 writing:1 yandell:1 equivalent:3 map:1 center:1 crfs:5 yt:1 missing:4 shi:1 txy:2 keller:1 convex:4 focused:1 wit:1 recovery:7 bruss:1 embedding:1 population:1 variation:1 annals:2 construction:2 elucidate:1 diabetes:1 pa:13 element:5 breakdown:1 observed:9 bottom:1 module:7 wermuth:2 wang:2 capture:1 linearregression:3 cycle:1 connected:3 sun:1 plo:1 yk:3 disease:2 byz:6 convexity:3 covariates:1 upon:1 efficiency:2 easily:1 joint:5 indirect:6 various:2 zo:1 separated:2 distinct:1 effective:2 artificial:4 labeling:1 widely:1 solve:1 ability:2 statistic:8 insulin:1 jointly:5 advantage:5 sequence:2 propose:2 tu:1 relevant:3 combining:1 subgraph:1 inducing:3 recipe:1 parent:1 r1:1 converges:1 sity:1 lauritzen:2 eq:8 edward:1 recovering:1 c:1 involves:1 indicate:2 implies:1 guided:1 biopsy:1 material:1 argued:1 polymorphism:1 generalization:2 rothman:1 hold:1 around:1 considered:5 exp:5 mapping:8 predict:2 adopt:3 a2:2 estimation:11 eqtl:1 xpa:27 tool:1 reflects:1 mit:1 genomic:8 gaussian:39 rather:1 zhou:1 factorizes:2 derived:1 focus:2 inherits:1 yo:1 improvement:1 likelihood:6 aka:1 contrast:1 cg:27 kim:2 inference:12 bt:2 koller:1 pixel:1 overall:1 among:7 priori:1 field:3 once:1 having:1 zz:24 biology:1 x4:4 employ:1 randomly:7 simultaneously:1 individual:1 friedman:2 mrce:5 drton:1 regulating:1 genotype:6 behind:1 held:1 xb:2 chain:87 secretion:1 accurate:1 edge:20 xy:17 nucleotide:1 re:3 modeling:1 moralization:15 cost:1 subset:2 uniform:2 examining:2 cggm:35 buntine:1 dependency:8 synthetic:1 combined:1 international:5 probabilistic:9 yl:10 lee:1 mouse:3 leveraged:1 wytock:1 li:1 converted:1 coefficient:1 coordinated:1 kolter:1 sloan:1 tion:2 recover:2 minimize:1 bright:1 ggm:4 accuracy:3 kaufmann:2 correspond:7 identify:1 bayesian:1 bxy:4 tissue:1 app:1 ping:1 influenced:2 energy:1 invasive:1 naturally:2 associated:1 dataset:6 popular:2 recall:13 knowledge:1 improves:1 organized:2 islet:4 supervised:21 response:2 improved:1 evaluated:2 though:1 furthermore:1 xa:2 hand:2 reveal:4 building:3 effect:4 contain:3 y2:1 true:7 ization:1 consisted:1 regularization:5 iteratively:1 dhillon:1 illustrated:2 white:2 conditionally:1 x5:1 noted:1 criterion:1 crf:2 demonstrate:2 complete:2 l1:3 cp:3 snp:2 recently:3 superior:2 common:2 functional:10 multinomial:1 empirically:1 physical:1 eichler:1 volume:3 discussed:1 extend:1 association:1 trait:1 mellon:2 imposing:2 sssykim:1 scandinavian:3 multivariate:4 scenario:2 seen:5 morgan:2 additional:2 dai:1 semi:38 full:1 desirable:1 multiple:21 infer:1 levina:1 cross:1 molecular:1 award:1 ravikumar:1 a1:2 laplacian:1 prediction:15 regression:30 calvin:1 patient:1 cmu:2 expectation:1 normalization:3 represent:2 cell:1 c1:2 addition:6 whereas:3 fellowship:1 addressed:1 modality:1 unlike:1 undirected:11 leveraging:1 lafferty:1 jordan:1 leverage:4 presence:2 revealed:2 yang:2 easy:1 concerned:1 xj:26 independence:3 marginalization:7 affect:1 architecture:1 lasso:4 hastie:1 hindered:1 idea:1 okawa:1 motivated:2 expression:13 six:1 forecasting:1 penalty:4 cause:1 amount:1 extensively:4 locally:2 induces:1 sohn:1 dna:1 generate:1 outperform:2 nsf:1 notice:1 dotted:5 estimated:8 tibshirani:1 yy:28 carnegie:2 group:2 independency:3 key:2 four:1 regresssion:1 blood:1 drawn:2 graph:91 genetical:1 inverse:4 parameterized:1 powerful:1 uncertainty:2 lamb:1 wu:1 separation:2 frydenberg:2 layer:18 quadratic:1 constraint:2 x2:6 encodes:1 lane:1 aspect:1 simulate:1 extremely:1 min:2 performing:7 multipleoutput:1 relatively:1 department:1 structured:15 according:2 across:3 smaller:1 em:1 reconstructing:1 partitioned:1 xo:1 taken:1 describing:1 turn:1 locus:1 sustik:1 generalizes:1 operation:1 eight:1 observe:2 hierarchical:1 alternative:6 assumes:2 top:1 clustering:2 remaining:1 graphical:21 xc:3 yz:23 surgery:1 added:1 intro:1 strategy:3 primary:1 responds:1 diagonal:3 unable:1 simulated:3 argue:2 tage:1 y5:1 collected:2 kemp:1 assuming:4 relationship:2 illustration:3 y4:1 regulation:1 potentially:1 negative:1 perform:2 vertical:2 observation:1 datasets:6 markov:17 extended:1 variability:1 y1:2 rn:2 arbitrary:1 duced:1 community:8 inferred:7 introduced:5 pair:4 required:1 z1:1 learned:1 nip:1 address:3 bar:2 below:1 pattern:6 sparsity:23 genotypic:1 including:2 wainwright:1 overlap:1 natural:2 rely:1 zhu:3 lwf:8 improve:1 identifies:1 lum:2 genomics:1 review:2 literature:1 acknowledgement:1 zh:1 fully:5 reordering:1 mixed:2 limitation:1 acyclic:1 validation:1 foundation:1 propagates:1 principle:1 systematically:1 genetics:1 supported:1 regressionbased:1 free:2 sparse:40 distributed:1 curve:7 world:1 collection:1 relatedness:1 gene:15 reveals:2 assumed:2 discriminative:2 iterative:1 diabetic:1 decomposes:1 table:2 learn:4 nature:1 career:1 complex:3 constructing:2 inherit:1 aistats:2 noise:1 decomposit:1 x1:11 xu:2 precision:13 pereira:1 xh:1 xl:12 lie:1 jmlr:3 yh:1 moralized:8 specific:1 xt:4 grouping:1 effectively:3 schadt:1 illustrates:2 chen:2 phenotype:8 mill:1 yin:1 explore:1 gao:1 partially:6 pinto:1 corresponds:1 dh:3 obozinski:2 conditional:13 goal:6 viewed:3 seyoung:1 shared:1 included:1 specifically:1 called:3 total:2 andersson:2 meaningful:1 select:2 ggms:2 support:1 evaluate:1 mcb:1 correlated:3 |
4,773 | 5,321 | Provable Submodular Minimization using
Wolfe?s Algorithm
Deeparnab Chakrabarty?
Prateek Jain?
Pravesh Kothari?
Abstract
Owing to several applications in large scale learning and vision problems, fast
submodular function minimization (SFM) has become a critical problem. Theoretically, unconstrained SFM can be performed in polynomial time [10, 11]. However,
these algorithms are typically not practical. In 1976, Wolfe [21] proposed an
algorithm to find the minimum Euclidean norm point in a polytope, and in 1980,
Fujishige [3] showed how Wolfe?s algorithm can be used for SFM. For general
submodular functions, this Fujishige-Wolfe minimum norm algorithm seems to
have the best empirical performance.
Despite its good practical performance, very little is known about Wolfe?s minimum
norm algorithm theoretically. To our knowledge, the only result is an exponential
time analysis due to Wolfe [21] himself. In this paper we give a maiden convergence
analysis of Wolfe?s algorithm. We prove that in t iterations, Wolfe?s algorithm
returns an O(1/t)-approximate solution to the min-norm point on any polytope. We
also prove a robust version of Fujishige?s theorem which shows that an O(1/n2 )approximate solution to the min-norm point on the base polytope implies exact
submodular minimization. As a corollary, we get the first pseudo-polynomial time
guarantee for the Fujishige-Wolfe minimum norm algorithm for unconstrained
submodular function minimization.
1
Introduction
An integer-valued1 function f : 2X ? Z defined over subsets of some finite ground set X of n
elements is submodular if it satisfies the following diminishing marginal returns property: for every
S ? T ? X and i ? X \ T , f (S ? {i}) ? f (S) ? f (T ? {i}) ? f (T ). Submodularity arises
naturally in several applications such as image segmentation [17], sensor placement [18], etc. where
minimizing an arbitrary submodular function is an important primitive.
In submodular function minimization (SFM), we assume access to an evaluation oracle for f which
for any subset S ? X returns the value f (S). We denote the time taken by the oracle to answer a
single query as EO. The objective is to find a set T ? X satisfying f (T ) ? f (S) for every S ? X.
In 1981, Grotschel, Lovasz and Schrijver [8] demonstrated the first polynomial time algorithm for
SFM using the ellipsoid algorithm. This algorithm, however, is practically infeasible due to the
running time and the numerical issues in implementing the ellipsoid algorithm. In 2001, Schrijver [19]
and Iwata et al. [9] independently designed combinatorial polynomial time algorithms for SFM.
Currently, the best algorithm is by Iwata and Orlin [11] with a running time of O(n5 EO + n6 ).
However, from a practical stand point, none of the provably polynomial time algorithms exhibit good
performance on instances of SFM encountered in practice (see ?4). This, along with the widespread
applicability of SFM in machine learning, has inspired a large body of work on practically fast
procedures (see [1] for a survey). But most of these procedures focus either on special submodular
?
Microsoft Research, 9 Lavelle Road, Bangalore 560001.
University of Texas at Austin (Part of the work done while interning at Microsoft Research)
1
One can assume any function is integer valued after suitable scaling.
?
1
functions such as decomposable functions [16, 20] or on constrained SFM problems [13, 12, 15,
14].
Fujishige-Wolfe?s Algorithm for SFM: For any submodular function f , the base polytope Bf of f
is defined as follows:
Bf = {x ? Rn : x(A) ? f (A), ?A ? X, and x(X) = f (X)},
(1)
P
n
where x(A) := i?A xi and xi is the i-th coordinate of x ? R . Fujishige [3] showed that if one can
obtain the minimum norm point on the base polytope, then one can solve SFM. Finding the minimum
norm point, however, is a non-trivial problem; at present, to our knowledge, the only polynomial
time algorithm known is via the ellipsoid method. Wolfe [21] described an iterative procedure to find
minimum norm points in polytopes as long as linear functions could be (efficiently) minimized over
them. Although the base polytope has exponentially many constraints, a simple greedy algorithm
can minimize any linear function over it. Therefore using Wolfe?s procedure on the base polytope
coupled with Fujishige?s theorem becomes a natural approach to SFM. This was suggested as early
as 1984 in Fujishige [4] and is now called the Fujishige-Wolfe algorithm for SFM.
This approach towards SFM was revitalized in 2006 when Fujishige and Isotani [6, 7] announced
encouraging computational results regarding the minimum norm point algorithm. In particular, this
algorithm significantly out-performed all known provably polynomial time algorithms. Theoretically,
however, little is known regarding the convergence of Wolfe?s procedure except for the finite, but
exponential, running time Wolfe himself proved. Nor is the situation any better for its application
on the base polytope. Given the practical success, we believe this is an important, and intriguing,
theoretical challenge.
In this work, we make some progress towards analyzing the Fujishige-Wolfe method for SFM and, in
fact, Wolfe?s algorithm in general. In particular, we prove the following two results:
? We prove (in Theorem 4) that for any polytope B, Wolfe?s algorithm converges to an ?approximate solution, in O(1/?) steps. More precisely, in O(nQ2 /?) iterations, Wolfe?s
algorithm returns a point kxk22 ? kx? k22 + ?, where Q = maxp?B kpk2 .
? We prove (in Theorem 5) a robust version of a theorem by Fujishige [3] relating min-norm
points on the base polytope to SFM. In particular, we prove that an approximate min-norm
point solution provides an approximate solution to SFM as well. More precisely, if x
satisfies kxk22 ? z T x + ?2 for all z ? Bf , then, f (Sx ) ? minS f (S) + 2n?, where Sx can
be constructed efficiently using x.
Together, these two results gives us our main result which is a pseudopolynomial bound on the
running time of the Fujishige-Wolfe algorithm for submodular function minimization.
Theorem 1. (Main Result.) Fix a submodular function f : 2X ? Z. The FujishigeWolfe algorithm returns the minimizer of f in O((n5 EO + n7 )F 2 ) time where F :=
maxni=1 (|f ({i})|, |f ([n]) ? f ([n] \ i)|).
Our analysis suggests that the Fujishige-Wolfe?s algorithm is dependent on F and has worse dependence on n than the Iwata-Orlin [11] algorithm. To verify this, we conducted empirical study on
several standard SFM problems. However, for the considered benchmark functions, running time of
Fujishige-Wolfe?s algorithm seemed to be independent of F and exhibited better dependence on n
than the Iwata-Orlin algorithm. This is described in ?4.
2
Preliminaries: Submodular Functions and Wolfe?s Algorithm
2.1 Submodular Functions and SFM
Given a ground set X on n elements, without loss of generality we think of it as the first n integers
[n] := {1, 2, . . . , n}. f be a submodular function. Since submodularity is translation invariant,
we assume f (?) = 0. For a submodular function f , we write Bf ? Rn for the associated base
polyhedron of f defined in (1). Given x ? Rn , one can find the minimum value of q > x over
q ? Bf in O(n log n + nEO) time using the following greedy algorithm: Renumber indices such
that x1 ? ? ? ? ? xn . Set qi? = f ([i]) ? f ([i ? 1]). Then, it can be proved that q ? ? Bf and is the
minimizer of the x> q for q ? Bf .
The connection between the SFM problem and the base polytope was first established in the following
minimax theorem of Edmonds [2].
2
Theorem 2 (Edmonds [2]). Given any submodular function f with f (?) = 0, we have
!
X
min f (S) = max
xi
S?[n]
x?Bf
i:xi <0
The following theorem of Fujishige [3] shows the connection between finding the minimum norm
point in the base polytope Bf of a submodular function f and the problem of SFM on input f . This
forms the basis of Wolfe?s algorithm. In ?3.2, we prove a robust version of this theorem.
Theorem 3 (Fujishige?s Theorem [3]). Let f : 2[n] ? Z be a submodular function and let Bf be the
associated base polyhedron. Let x? be the optimal solution to minx?Bf ||x||. Define S = {i | x?i <
0}. Then, f (S) ? f (T ) for every T ? [n].
2.2
Wolfe?s Algorithm for Minimum Norm Point of a polytope.
We now present Wolfe?s algorithm for computing the minimum-norm point in an arbitrary polytope
B ? Rn . We assume a linear optimization oracle (LO) which takes input a vector x ? Rn and
outputs a vector q ? arg minp?B x> p.
We
some definitions. The affine hull of a finite set S ? Rn is aff(S) = {y | y =
P start by recalling
P
?
?
z,
?
z?S z
z?S z = 1}. The affine minimizer of S is defined as y = arg minz?aff(S) ||z||2 ,
and y satisfies the following affine minimizer property: for any v ? aff(S), v > y = ||y||2 . The
procedure AffineMinimizer(S) returns (y, ?) where y is the affine minimizer and ? = (?s )s?S is
the set of coefficients expressing y as an affine combination of points in S. This procedure can be
naively implemented in O(|S|3 + n|S|2 ) as follows. Let B be the n ? |S| matrix where each column
in a point in S. Then ? = (B > B)?1 1/1> (B > B)?1 1 and y = B?.
Algorithm 1 Wolfe?s Algorithm
P
1. Let q be an arbitrary vertex of B. Initialize x ? q. We always maintain x = i?S ?i qi as a
convex combination of a subset S of vertices of B. Initialize S = {q} and ?1 = 1.
2. WHILE(true): (MAJOR CYCLE)
(a) q := LO(x).
// Linear Optimization: q ? arg minp?B x> p.
(b) IF ||x||2 ? x> q + ?2 THEN break.
// Termination Condition. Output x.
(c) S := S ? {q}.
(d) WHILE(true): (MINOR CYCLE)
i. (y, ?) = AffineMinimizer(S).
//y = arg minz?aff(S) ||z||.
ii. IF ?i ? 0 for all i THEN break.
//If y ? conv(S), then end minor loop.
iii. ELSE
// If y ?
/ conv(S), then update x to the intersection of the boundary of conv(S) and the segment joining y and
previous x. Delete points from S which are not required to describe the new x as a convex combination.
P
? := mini:?i <0 ?i /(?i ? ?i )
// Recall, x =
i ?i qi .
Update x ? ?y + (1 ? ?)x.
// By definition of ?, the new x lies in conv(S).
Update ?i ? ??i + (1 ? ?)?i .
//This sets the coefficients of the new x
S = {i : ?i > 0}.
// Delete points which have ?i = 0. This deletes at least one point.
(e) Update x ? y.
// After the minor loop terminates, x is updated to be the affine minimizer of the current set S.
3. RETURN x.
When ? = 0, the algorithm on termination (if it terminates) returns the minimum norm point in B
since ||x||2 ? x> x? ? ||x|| ? ||x? ||. For completeness, we sketch Wolfe?s argument in [21] of finite
termination. Note that |S| ? n always; otherwise the affine minimizer is 0 which either terminates
the program or starts a minor cycle which decrements |S|. Thus, the number of minor cycles in a
major cycle ? n, and it suffices to bound the number of major cycles. Each major cycle is associated
with a set S whose affine minimizer, which is the current x, lies in the convex hull of S. Wolfe calls
such sets corrals. Next, we show that ||x|| strictly decreases across iterations (major or minor cycle)
of the algorithm, which proves that no corral repeats,
thus bounding the number of major cycles by
the number of corrals. The latter is at most N
,
where
N is the number of vertices of B.
n
Consider iteration j which starts with xj and ends with xj+1 . Let Sj be the set S at the beginning
of iteration j. If the iteration is a major cycle, then xj+1 is the affine minimizer of Sj ? {qj }
3
2
where qj = LO(xj ). Since x>
j qj < ||xj || (the algorithm doesn?t terminate in iteration j) and
>
2
xj+1 qj = ||xj+1 || (affine minimizer property), we get xj 6= xj+1 , and so ||xj+1 || < ||xj || (since
the affine minimizer is unique). If the iteration is a minor cycle, then xj+1 = ?xj + (1 ? ?)yj , where
yj is the affine minimizer of Sj and ? < 1. Since ||yj || < ||xj || (yj 6= xj since yj ?
/ conv(Sj )), we
get ||xj+1 || < ||xj ||.
3
Analysis
Our refined analysis of Wolfe?s algorithm is encapsulated in the following theorem.
Theorem 4. Let B be an arbitrary polytope such that the maximum Euclidean norm of any vertex of
B is at most Q. After O(nQ2 /?2 ) iterations, Wolfe?s algorithm returns a point x ? B which satisfies
||x||2 ? x> q + ?2 , for all points q ? B. In particular, this implies ||x||2 ? ||x? ||2 + 2?2 .
The above theorem shows that Wolfe?s algorithm converges to the minimum norm point at an 1/t-rate.
We stress that the above is for any polytope. To apply this to SFM, we prove the following robust
version of Fujishige?s theorem connecting the minimum norm point in the base polytope and the set
minimizing the submodular function value.
Theorem 5. Fix a submodular function f with base polytope Bf . Let x ? Bf be such that ||x||2 ?
x> q + ?2 for all q ? Bf . Renumber indices such that x1 ? ? ? ? ? xn . Let S = {1, 2, . . . , k},where
k is smallest index satisfying (C1) xk+1 ? 0 and (C2) xk+1 ? xk ? ?/n. Then, f (S) ? f (T ) + 2n?
1
for any subset T ? S. In particular, if ? = 4n
and f is integer-valued, then S is a minimizer.
Theorem 4 and Theorem 5 implies our main theorem.
Theorem 1. (Main Result.) Fix a submodular function f : 2X ? Z. The FujishigeWolfe algorithm returns the minimizer of f in O((n5 EO + n7 )F 2 ) time where F :=
maxni=1 (|f ({i})|, |f ([n]) ? f ([n] \ i)|).
Proof. The vertices of Bf are well understood: for every permutation ? of [n], we have a vertex
with x?(i) = f ({?(1), . . . , ?(i)}) ? f ({?(1), . . . , ?(i ? 1)}). By submodularity of f , we get
for all i, |xi | ? F . Therefore, for any point x ? Bf , ||x||2 ? nF 2 . Choose ? = 1/4n. From
Theorem 4 we know that if we run O(n4 F 2 ) iterations of Wolfe, we will get a point x ? Bf such
that ||x||2 ? x> q + ?2 for all q ? Bf . Theorem 5 implies this solves the SFM problem. The running
time for each iteration is dominated by the time for the subroutine to compute the affine minimizer of
S which is at most O(n3 ), and the linear optimization oracle. For Bf , LO(x) can be implemented in
O(n log n + nEO) time. This proves the theorem.
We prove Theorem 4 and Theorem 5 in ?3.1 and ?3.2, respectively.
3.1
Analysis of Wolfe?s Min-norm Point Algorithm
The stumbling block in the analysis of Wolfe?s algorithm is the interspersing of major and minor
cycles which oscillates the size of S preventing it from being a good measure of progress. Instead, in
our analysis, we use the norm of x as the measure of progress. Already we have seen that ||x|| strictly
decreases. It would be nice to quantify how much the decrease is, say, across one major cycle. This,
at present, is out of our reach even for major cycles which contain two or more minor cycles in them.
However, we can prove significant drop in norm in major cycles which have at most one minor cycle
in them. We call such major cycles good. The next easy, but very useful, observation is the following:
one cannot have too many bad major cycles without having too many good major cycles.
Lemma 1. In any consecutive 3n + 1 iterations, there exists at least one good major cycle.
Proof. Consider a run of r iterations where all major cycles are bad, and therefore contain ? 2
minor cycles. Say there are k major cycles and r ? k minor cycles, and so r ? k ? 2k implying
r ? 3k. Let SI be the set S at the start of these iterations and SF be the set at the end. We have
|SF | ? |SI | + k ? (r ? k) ? |SI | + 2k ? r ? n ? 3r . Therefore, r ? 3n, since |SF | ? 0.
Before proceeding, we introduce some notation.
Definition 1. Given a point x ? B, let us denote err(x) := ||x||2 ? ||x? ||2 . Given a point x and
q, let ?(x, q) := ||x||2 ? x> q and let ?(x) := maxq?B ?(x, q) = ||x||2 ? minq?B x> q. Observe
that ?(x) ? err(x)/2 since ?(x) ? ||x||2 ? x> x? ? (||x||2 ? ||x? ||2 )/2.
4
We now use t to index all good major cycles. Let xt be the point x at the beginning of the t-th
good major cycle. The next theorem shows that the norm significantly drops across good major
cycles.
Theorem 6. For t iterating over good major cycles, err(xt ) ? err(xt+1 ) ? ?2 (xt )/8Q2 .
We now complete the proof of Theorem 4 using Theorem 6.
Proof of Theorem 4. Using Theorem 6, we get that err(xt ) ? err(xt+1 ) ? err(xt )2 /32Q2 since
?(x) ? err(x)/2 for all x. We claim that in t? ? 64Q2 /?2 good major cycles, we reach xt with
err(xt? ) ? ?2 . To see this rewrite as follows:
err(xt )
, for all t.
err(xt+1 ) ? err(xt ) 1 ?
32Q2
Now let e0 := err(x0 ). Define t0 , t1 , . . . such that for all k ? 1 we have err(xt ) > e0 /2k for
t ? [tk?1 , tk ). That is, tk is thefirst time t at which err(xt ) ? e0 /2k . Note that for t ? [tk?1 , tk ),
we have err(xt+1 ) ? err(xt ) 1 ? 32Qe02 2k . This implies in 32Q2 2k /e0 time units after tk?1 , we
will have err(xt ) ? err(xtk?1 )/2; we have used the fact that (1 ? ?)1/? < 1/2 when ? < 1/32.
That is, tk ? tk?1 + 32Q2 2k /e0 . We are interested in t? = tK where 2K = e0 /?2 . We get
2
t? ? 32Q
1 + 2 + ? ? ? + 2K ? 64Q2 2K /e0 = 64Q2 /?2 .
e0
Next, we claim that in t?? < t? + t0 good major cycles, where t0 = 8Q2 /?2 , we obtain an xt??
with ?(xt?? ) ? ?2 . This is because, if not, then, using Theorem 6, in each of the good major
cycles t? + 1, t? + 2, . . . t? + t0 , err(x) falls additively by > ?4 /8Q2 and thus err(xt? +t0 ) <
err(xt? ) ? ?2 ? 0, which is a contradiction. Therefore, in O(Q2 /?2 ) good major cycles, the
algorithm obtains an x = xt?? with ?(x) ? ?2 , proving Theorem 4.
The rest of this subsection is dedicated to proving Theorem 6.
Proof of Theorem 6: We start off with a simple geometric lemma.
Lemma 2. Let S be a subset of Rn and suppose y is the minimum norm point of aff(S). Let x and
q be arbitrary points in aff(S). Then,
||x||2 ? ||y||2 ?
?(x, q)2
4Q2
(2)
where Q is an upper bound on ||x||, ||q||.
Proof. Since y is the minimum norm point in aff(S), we have x> y = q > y = ||y||2 . In particular,
||x ? y||2 = ||x||2 ? ||y||2 . Therefore,
?(x, q) = kxk2 ? xT q = kxk2 ? x> y + y > q ? xT q = (y ? x)T (q ? x) ? ky ? xk ? kq ? xk
? ky ? xk(kxk + kqk) ? 2Qky ? xk,
where the first inequality is Cauchy-Schwartz and the second is triangle inequality. Lemma now
follows by taking square of the above expression and by observing that ky ? xk2 = kxk2 ? kyk2 .
The above lemma takes case of major cycles with no minor cycles in them.
Lemma 3 (Progress in Major Cycle with no Minor Cycles). Let t be the index of a good major cycle
with no minor cycles. Then err(xt ) ? err(xt+1 ) ? ?2 (xt )/4Q2 .
Proof. Let St be the set S at start of the tth good major cycle, and let qt be the point minimizing x>
t q.
Let S = St ? qt and let y be the minimum norm point in aff(S). Since there are no minor cycles,
y ? conv(S). Abuse notation and let xt+1 = y be the iterate at the call of the next major cycle (and
not the next good major cycle). Since the norm monotonically decreases, it suffices to prove the
lemma statement for this xt+1 . Now apply Lemma 2 with x = xt and q = qt and S = St ? qt . We
have that err(xt ) ? err(xt+1 ) = ||xt ||2 ? ||y||2 ? ?(xt , qt )2 /4Q2 = ?(xt )2 /4Q2 .
Now we have to argue about major cycles with exactly one minor cycle. The next observation is a
useful structural result.
5
Lemma 4 (New Vertex Survives a Minor Cycle.). Consider any (not necessarily good) major
cycle. Let xt , St , qt be the parameters at the beginning of this cycle, and let xt+1 , St+1 , qt+1 be the
parameters at the beginning of the next major cycle. Then, qt ? St+1 .
Proof. Clearly St+1 ? St ? qt since qt is added and then maybe minor cycles remove some points
from S. Suppose qt ?
/ St+1 . Well, then St+1 ? St . But xt+1 is the affine minimizer of St+1 and xt
is the affine minimizer of St . Since St is the larger set, we get ||xt || ? ||xt+1 ||. This contradicts the
strict decrease in the norm.
Lemma 5 (Progress in an iteration with exactly one minor cyvle). Suppose the tth good major cycle
has exactly one minor cycle. Then, err(xt ) ? err(xt+1 ) ? ?(xt )2 /8Q2 .
Proof. Let xt , St , qt be the parameters at the beginning of the tth good major cycle. Let y be the
affine minimizer of St ?qt . Since there is one minor cycle, y ?
/ conv(St ?qt ). Let z = ?xt +(1??)y
be the intermediate x, that is, point in the line segment [xt , y] which lies in conv(St ? qt ). Let S 0 be
the set after the single minor cycle is run. Since there is just one minor cycle, we get xt+1 (abusing
notation once again since the next major cycle maynot be good) is the affine minimizer of S 0 .
Let A , ||xt ||2 ? ||y||2 . From Lemma 2, and using qt is the minimizer of x>
t q over all q, we have:
A = ||xt ||2 ? ||y||2 ? ?2 (xt )/4Q2
(3)
Recall, z = ?xt + (1 ? ?)y for some ? ? [0, 1]. Since y is the min-norm point of aff(St ? qt ), and
xt ? St , we get ||z||2 = ?2 ||xt ||2 + (1 ? ?2 )||y||2 . this yields:
||xt ||2 ? ||z||2 = (1 ? ?2 ) ||xt ||2 ? ||y||2 = (1 ? ?2 )A
(4)
Further, recall that S 0 is the set after the only minor cycle in the tth iteration is run and thus, from
Lemma 4, qt ? S 0 . z ? conv(S 0 ) by definition. And since there is only one minor cycle, xt+1 is the
affine minimizer of S 0 . We can apply Lemma 2 with z, qt and xt+1 , to get
?2 (z, qt )
(5)
||z||2 ? ||xt+1 ||2 ?
4Q2
Now we lower bound ?2 (z, qt ). By definition of z, we have:
>
>
2
z > qt = ?x>
t qt + (1 ? ?)y qt = ?xt qt + (1 ? ?)||y||
>
2
where the last equality follows since y qt = ||y|| (since qt ? St ? qt and y is affine minimizer of
St ? qt ). This gives
?(z, qt )
=
||z||2 ? z > qt
=
=
2
?2 ||xt ||2 + (1 ? ?2 )||y||2 ? ?x>
t qt + (1 ? ?)||y||
2
2
?(||xt ||2 ? x>
t qt ) ? ?(1 ? ?) ||xt || ? ||y||
=
? (?(xt ) ? (1 ? ?)A)
(6)
From (4),(5), and (6), we get
2
?2 (?(xt ) ? (1 ? ?)A)
(7)
4Q2
We need to show that the RHS is at least ?(xt )2 /8Q2 . Intuitively, if ? is small (close to 0), the
first term implies this using (3), and if ? is large (close to 1), then the second term implies this. The
following paragraph formalizes this intuition for any ?.
errt ? errt+1 ? (1 ? ?2 )A +
Now, if (1 ? ?2 )A > ?(xt )2 /8Q2 , we are done. Therefore, we assume (1 ? ?2 )A ? ?(xt )2 /8Q2 .
In this case, using the fact that ?(xt ) ? ||xt ||2 + ||xt ||||qt || ? 2Q2 , we get that
?(xt )
(1 ? ?)A ? (1 ? ?2 )A ? ?(xt ) ?
? ?(xt )/4
8Q2
Substituting in (7), and using (3), we get
9?2 ?(xt )2
?(xt )2
(1 ? ?2 )?(xt )2
errt ? errt+1 ?
+
?
(8)
2
2
4Q
64Q
8Q2
This completes the proof of the lemma.
Lemma 3 and Lemma 5 complete the proof of Theorem 6.
6
3.2 A Robust version of Fujishige?s Theorem
In this section we prove Theorem 5 which we restate below.
Theorem 5. Fix a submodular function f with base polytope Bf . Let x ? Bf be such that ||x||2 ?
x> q + ?2 for all q ? Bf . Renumber indices such that x1 ? ? ? ? ? xn . Let S = {1, 2, . . . , k},where
k is smallest index satisfying (C1) xk+1 ? 0 and (C2) xk+1 ? xk ? ?/n. Then, f (S) ? f (T ) + 2n?
1
for any subset T ? S. In particular, if ? = 4n
and f is integer-valued, then S is a minimizer.
Before proving the theorem, note that setting ? = 0 gives Fujishige?s theorem Theorem 3.
Proof. We claim that the following inequality holds. Below, [i] := {1, . . . , i}.
n?1
X
(xi+1 ? xi ) ? (f ([i]) ? x([i])) ? ?2
(9)
i=1
P
We prove this shortly. Let S and k be as defined in the theorem statement. Note that i?S:xi ?0 xi ?
n?, since (C2) doesn?t hold for any index i < k with xi ? 0.PFurthermore, since xk+1 ? xk ? ?/n,
we get using (9), f (S) ? x(S) ? n?. Therefore, f (S) ? i?S:xi <0 xi + 2n? which implies the
theorem due to Theorem 2.
Now we prove (9). Let z ? Bf be the point which minimizes z > x. By the Greedy algorithm
described in Section 2.1, we know that zi = f ([i]) ? f ([i ? 1]). Next, we write x in a different basis
Pn?1
as follows: x = i=1 (xi ? xi+1 )1[i] + xn 1[n] . Here 1[i] is used as the shorthand for the vector
which has 1?s in the first i coordinates and 0s everywhere else. Taking dot product with (x ? z), we
get
||x||2 ? x> z = (x ? z)> x =
n?1
X
(xi ? xi+1 ) x> 1[i] ? z > 1[i] + xn x> 1[n] ? z > 1[n] (10)
i=1
Since zi = f ([i]) ? f ([i ? 1]), we get x> 1[i] ? z > 1[i] is x([i]) ? f ([i]). Therefore the RHS of (10)
is the LHS of (9). The LHS of (10), by the assumption of the theorem, is at most ?2 implying (9).
4
Discussion and Conclusions
(a)
(b)
(c)
Figure 1: Running time comparision of Iwata-Orlin?s (IO) method [11] vs Wolfe?s method. (a):
s-t mincut function, (b) Iwata?s 3 groups function [16]. (c): Total number of iterations required by
Wolfe?s method for solving s-t mincut with increasing F
We have shown that the Fujishige-Wolfe algorithm solves SFM in O((n5 EO + n7 )F 2 ) time, where
F is the maximum change in the value of the function on addition or deletion of an element. Although
this is the first pseudopolynomial time analysis of the algorithm, we believe there is room for
improvement and hope our work triggers more interest.
Note that our anlaysis of the Fujishige-Wolfe algorithm is weaker than the best known method in
terms of time complexity (IO method by [11]) on two counts: a) dependence on n, b) dependence
on F . In contrast, we found this algorithm significantly outperforming the IO algorithm empirically
? we show two plots here. In Figure 1 (a), we run both on Erdos-Renyi graphs with p = 0.8 and
randomly chosen s, t nodes. In Figure 1 (b), we run both on the Iwata group functions [16] with 3
groups. Perhaps more interestingly, in Figure 1 (c), we ran the Fujishige-Wolfe algorithm on the
simple path graph where s, t were the end points, and changed the capacities on the edges of the
graph which changed the parameter F . As can be seen, the number of iterations of the algorithm
remains constant even for exponentially increasing F .
7
References
[1] Francis Bach. Convex analysis and optimization with submodular functions: a tutorial. CoRR,
abs/1010.4207, 2010. 1
[2] Jack Edmonds. Matroids, submodular functions and certain polyhedra. Combinatorial Structures
and Their Applications, pages 69?87, 1970. 2, 3
[3] Satoru Fujishige. Lexicographieally optimal base of a polymatroid with respect to a weight
vector. Math. Oper. Res., 5:186?196, 1980. 1, 2, 3
[4] Satoru Fujishige. Submodular systems and related topics. Math. Programming Study, 1984. 2
[5] Satoru Fujishige. Submodular functions and optimization. Elsevier, 2005.
[6] Satoru Fujishige, Takumi Hayashi, and Shigueo Isotani. The minimum-norm-point algorithm
applied to submodular function minimization and linear programming. 2006. 2
[7] Satoru Fujishige and Shigueo Isotani. A submodular function minimization algorithm based on
the minimum-norm base. Pacific Journal of Optimization, 7:3, 2011. 2
[8] Martin Gr?otschel, L?aszl?o Lov?asz, and Alexander Schrijver. The ellipsoid method and its
consequences in combinatorial optimization. Combinatorica, 1(2):169?197, 1981. 1
[9] Satoru Iwata, Lisa Fleischer, and Satoru Fujishige. A combinatorial, strongly polynomial-time
algorithm for minimizing submodular functions. In STOC, pages 97?106, 2000. 1
[10] Satoru Iwata, Lisa Fleischer, and Satoru Fujishige. A combinatorial strongly polynomial
algorithm for minimizing submodular functions. J. ACM, 48(4):761?777, 2001. 1
[11] Satoru Iwata and James B. Orlin. A simple combinatorial algorithm for submodular function
minimization. In SODA, pages 1230?1237, 2009. 1, 2, 7
[12] Rishabh Iyer, Stefanie Jegelka, and Jeff Bilmes. Curvature and optimal algorithms for learning
and minimizing submodular functions. CoRR, abs/1311.2110, 2013. 1
[13] Rishabh Iyer, Stefanie Jegelka, and Jeff Bilmes. Fast semidifferential-based submodular function
optimization. In ICML (3), pages 855?863, 2013. 1
[14] Rishabh K. Iyer and Jeff A. Bilmes. Submodular optimization with submodular cover and
submodular knapsack constraints. In NIPS, pages 2436?2444, 2013. 1
[15] Stefanie Jegelka, Francis Bach, and Suvrit Sra. Reflection methods for user-friendly submodular
optimization. In NIPS, pages 1313?1321, 2013. 1
[16] Stefanie Jegelka, Hui Lin, and Jeff A. Bilmes. On fast approximate submodular minimization.
In NIPS, pages 460?468, 2011. 1, 7
[17] Pushmeet Kohli and Philip H. S. Torr. Dynamic graph cuts and their applications in computer
vision. In Computer Vision: Detection, Recognition and Reconstruction, pages 51?108. 2010. 1
[18] Andreas Krause, Ajit Paul Singh, and Carlos Guestrin. Near-optimal sensor placements in
gaussian processes: Theory, efficient algorithms and empirical studies. Journal of Machine
Learning Research, 9:235?284, 2008. 1
[19] Alexander Schrijver. A combinatorial algorithm minimizing submodular functions in strongly
polynomial time. J. Comb. Theory, Ser. B, 80(2):346?355, 2000. 1
[20] Peter Stobbe and Andreas Krause. Efficient minimization of decomposable submodular functions. In NIPS, pages 2208?2216, 2010. 1
[21] Phillip Wolfe. Finding the nearest point in a polytope. Math. Programming, 11:128 ? 149, 1976.
1, 2, 3
8
| 5321 |@word kohli:1 version:5 polynomial:10 norm:31 seems:1 bf:23 semidifferential:1 termination:3 corral:3 additively:1 interestingly:1 err:28 current:2 si:3 intriguing:1 numerical:1 remove:1 designed:1 drop:2 update:4 plot:1 v:1 implying:2 greedy:3 xk:12 beginning:5 provides:1 completeness:1 node:1 math:3 along:1 constructed:1 c2:3 become:1 prove:14 shorthand:1 paragraph:1 comb:1 introduce:1 x0:1 theoretically:3 lov:1 nor:1 inspired:1 little:2 encouraging:1 increasing:2 becomes:1 grotschel:1 conv:9 notation:3 prateek:1 minimizes:1 q2:25 finding:3 guarantee:1 pseudo:1 formalizes:1 every:4 nf:1 friendly:1 oscillates:1 exactly:3 schwartz:1 ser:1 unit:1 before:2 t1:1 understood:1 io:3 consequence:1 despite:1 joining:1 analyzing:1 path:1 abuse:1 suggests:1 practical:4 unique:1 yj:5 practice:1 block:1 procedure:7 empirical:3 significantly:3 road:1 get:17 cannot:1 close:2 satoru:10 demonstrated:1 primitive:1 independently:1 convex:4 survey:1 minq:1 decomposable:2 contradiction:1 proving:3 coordinate:2 updated:1 suppose:3 trigger:1 user:1 exact:1 programming:3 wolfe:41 element:3 satisfying:3 recognition:1 cut:1 aszl:1 cycle:59 decrease:5 ran:1 intuition:1 complexity:1 dynamic:1 singh:1 rewrite:1 segment:2 solving:1 basis:2 triangle:1 jain:1 fast:4 describe:1 query:1 refined:1 whose:1 maxni:2 larger:1 valued:3 solve:1 say:2 otherwise:1 maxp:1 think:1 reconstruction:1 product:1 loop:2 ky:3 convergence:2 converges:2 tk:9 nearest:1 qt:34 minor:26 progress:5 solves:2 implemented:2 implies:8 quantify:1 submodularity:3 restate:1 owing:1 hull:2 implementing:1 fix:4 suffices:2 preliminary:1 strictly:2 hold:2 practically:2 considered:1 ground:2 claim:3 substituting:1 major:37 early:1 smallest:2 xk2:1 consecutive:1 encapsulated:1 pravesh:1 combinatorial:7 currently:1 survives:1 minimization:11 lovasz:1 hope:1 clearly:1 sensor:2 always:2 gaussian:1 pn:1 corollary:1 focus:1 improvement:1 polyhedron:3 contrast:1 elsevier:1 dependent:1 typically:1 diminishing:1 subroutine:1 interested:1 provably:2 issue:1 arg:4 constrained:1 special:1 initialize:2 marginal:1 once:1 having:1 icml:1 minimized:1 bangalore:1 randomly:1 microsoft:2 maintain:1 recalling:1 ab:2 detection:1 interest:1 evaluation:1 rishabh:3 edge:1 lh:2 euclidean:2 re:1 e0:8 theoretical:1 delete:2 instance:1 column:1 cover:1 applicability:1 vertex:7 subset:6 kq:1 conducted:1 gr:1 too:2 interning:1 answer:1 st:22 off:1 together:1 connecting:1 again:1 choose:1 worse:1 return:10 oper:1 coefficient:2 performed:2 break:2 observing:1 francis:2 start:6 carlos:1 orlin:5 minimize:1 square:1 efficiently:2 yield:1 none:1 bilmes:4 reach:2 stobbe:1 definition:5 chakrabarty:1 james:1 naturally:1 associated:3 proof:12 proved:2 recall:3 knowledge:2 subsection:1 segmentation:1 nq2:2 done:2 strongly:3 generality:1 just:1 sketch:1 widespread:1 abusing:1 perhaps:1 believe:2 phillip:1 k22:1 verify:1 true:2 contain:2 equality:1 kyk2:1 stress:1 complete:2 dedicated:1 reflection:1 neo:2 image:1 jack:1 polymatroid:1 empirically:1 exponentially:2 relating:1 expressing:1 significant:1 unconstrained:2 submodular:41 dot:1 access:1 etc:1 base:16 curvature:1 showed:2 certain:1 suvrit:1 inequality:3 stumbling:1 success:1 outperforming:1 seen:2 minimum:20 guestrin:1 eo:5 monotonically:1 ii:1 bach:2 long:1 lin:1 qi:3 n5:4 vision:3 himself:2 iteration:18 c1:2 addition:1 krause:2 else:2 completes:1 revitalized:1 rest:1 exhibited:1 asz:1 strict:1 fujishige:30 n7:3 integer:5 call:3 structural:1 near:1 intermediate:1 iii:1 easy:1 iterate:1 xj:17 zi:2 andreas:2 regarding:2 texas:1 qj:4 t0:5 fleischer:2 expression:1 peter:1 useful:2 iterating:1 maybe:1 tth:4 tutorial:1 edmonds:3 write:2 group:3 deletes:1 kqk:1 maiden:1 graph:4 run:6 everywhere:1 soda:1 scaling:1 sfm:24 announced:1 bound:4 encountered:1 oracle:4 comparision:1 placement:2 constraint:2 precisely:2 kpk2:1 aff:9 n3:1 dominated:1 argument:1 min:8 xtk:1 martin:1 pacific:1 combination:3 terminates:3 across:3 contradicts:1 n4:1 pseudopolynomial:2 intuitively:1 invariant:1 taken:1 remains:1 count:1 know:2 end:4 apply:3 observe:1 shortly:1 knapsack:1 running:7 shigueo:2 mincut:2 prof:2 objective:1 already:1 added:1 valued1:1 dependence:4 exhibit:1 minx:1 otschel:1 capacity:1 philip:1 errt:4 topic:1 polytope:20 argue:1 cauchy:1 trivial:1 provable:1 index:8 ellipsoid:4 mini:1 minimizing:7 statement:2 stoc:1 upper:1 kothari:1 observation:2 benchmark:1 finite:4 situation:1 rn:7 ajit:1 arbitrary:5 required:2 connection:2 deletion:1 polytopes:1 established:1 maxq:1 nip:4 suggested:1 below:2 challenge:1 program:1 max:1 critical:1 suitable:1 natural:1 minimax:1 kxk22:2 stefanie:4 n6:1 coupled:1 nice:1 geometric:1 loss:1 permutation:1 affine:19 jegelka:4 minp:2 translation:1 austin:1 lo:4 changed:2 repeat:1 last:1 infeasible:1 weaker:1 lisa:2 fall:1 taking:2 matroids:1 boundary:1 xn:5 stand:1 seemed:1 doesn:2 preventing:1 pushmeet:1 sj:4 approximate:6 obtains:1 erdos:1 xi:16 iterative:1 terminate:1 robust:5 sra:1 necessarily:1 main:4 decrement:1 bounding:1 rh:2 paul:1 n2:1 body:1 x1:3 exponential:2 sf:3 lie:3 kxk2:3 minz:2 renumber:3 renyi:1 theorem:47 bad:2 xt:77 naively:1 exists:1 corr:2 hui:1 iyer:3 anlaysis:1 kx:1 sx:2 intersection:1 kxk:1 deeparnab:1 hayashi:1 iwata:10 satisfies:4 minimizer:23 acm:1 fujishigewolfe:2 towards:2 room:1 jeff:4 change:1 isotani:3 except:1 torr:1 lemma:16 called:1 total:1 schrijver:4 combinatorica:1 latter:1 arises:1 alexander:2 |
4,774 | 5,322 | A Differential Equation for Modeling Nesterov?s
Accelerated Gradient Method: Theory and Insights
Weijie Su1
2
Stephen Boyd2
Emmanuel J. Cand`es1,3
1
Department of Statistics, Stanford University, Stanford, CA 94305
Department of Electrical Engineering, Stanford University, Stanford, CA 94305
3
Department of Mathematics, Stanford University, Stanford, CA 94305
{wjsu, boyd, candes}@stanford.edu
Abstract
We derive a second-order ordinary differential equation (ODE), which is the limit
of Nesterov?s accelerated gradient method. This ODE exhibits approximate equivalence to Nesterov?s scheme and thus can serve as a tool for analysis. We show that
the continuous time ODE allows for a better understanding of Nesterov?s scheme.
As a byproduct, we obtain a family of schemes with similar convergence rates.
The ODE interpretation also suggests restarting Nesterov?s scheme leading to an
algorithm, which can be rigorously proven to converge at a linear rate whenever
the objective is strongly convex.
1
Introduction
As data sets and problems are ever increasing in size, accelerating first-order methods is both of
practical and theoretical interest. Perhaps the earliest first-order method for minimizing a convex
function f is the gradient method, which dates back to Euler and Lagrange. Thirty years ago, in a
seminar paper [11] Nesterov proposed an accelerated gradient method, which may take the following
form: starting with x0 and y0 = x0 , inductively define
xk = yk?1 ? s?f (yk?1 )
k?1
yk = x k +
(xk ? xk?1 ).
k+2
(1.1)
For a fixed step size s = 1/L, where L is the Lipschitz constant of ?f , this scheme exhibits the
convergence rate
Lkx ? x? k2
0
.
f (xk ) ? f ? ? O
k2
?
?
?
Above, x is any minimizer of f and f = f (x ). It is well-known that this rate is optimal among
all methods having only information about the gradient of f at consecutive iterates [12]. This is in
contrast to vanilla gradient descent methods, which can only achieve a rate of O(1/k) [17]. This
improvement relies on the introduction of the momentum term xk ? xk?1 as well as the particularly
tuned coefficient (k ? 1)/(k + 2) ? 1 ? 3/k. Since the introduction of Nesterov?s scheme, there
has been much work on the development of first-order accelerated methods, see [12, 13, 14, 1, 2] for
example, and [19] for a unified analysis of these ideas.
In a different direction, there is a long history relating ordinary differential equations (ODE) to optimization, see [6, 4, 8, 18] for references. The connection between ODEs and numerical optimization
is often established via taking step sizes to be very small so that the trajectory or solution path converges to a curve modeled by an ODE. The conciseness and well-established theory of ODEs provide
deeper insights into optimization, which has led to many interesting findings [5, 7, 16].
1
In this work, we derive a second-order ordinary differential equation, which is the exact limit of
Nesterov?s scheme by taking small step sizes in (1.1). This ODE reads
? + 3 X? + ?f (X) = 0
(1.2)
X
t
?
for t > 0, with initial conditions X(0) = x0 , X(0)
= 0; here, x0 is the starting point in Nesterov?s
?
? = d2 X/dt2 denotes the
scheme, X denotes the time derivative or velocity dX/dt and similarly X
?
acceleration. The time parameter in this ODE is related to the step size in (1.1) via t ? k s. Case
studies are provided to demonstrate that the homogeneous and conceptually simpler ODE can serve
as a tool for analyzing and generalizing Nesterov?s scheme. To the best of our knowledge, this work
is the first to model Nesterov?s scheme or its variants by ODEs.
We denote by FL the class of convex functions f with L?Lipschitz continuous gradients defined on
Rn , i.e., f is convex, continuously differentiable, and obeys
k?f (x) ? ?f (y)k ? Lkx ? yk
for any x, y ? Rn , where k ? k is the standard Euclidean norm and L > 0 is the Lipschitz constant
throughout this paper. Next, S? denotes the class of ??strongly convex functions f on Rn with
continuous gradients, i.e., f is continuously differentiable and f (x) ? ?kxk2 /2 is convex. Last, we
set S?,L = FL ? S? .
2
Derivation of the ODE
Assume f ? FL for L > 0. Combining the two equations of (1.1) and applying a rescaling give
xk+1 ? xk
k ? 1 xk ? xk?1 ?
?
?
(2.1)
=
? s?f (yk ).
k+2
s
s
?
Introduce the ansatz xk ? X(k s) for some smooth curve X(t)
? defined for t ? 0. For fixed t,
as the step size s goes to zero, X(t) ? xt/?s = xk and X(t + s) ? x(t+?s)/?s = xk+1 with
?
k = t/ s. With these approximations, we get Taylor expansions:
?
?
1 ? ?
?
(xk+1 ? xk )/ s = X(t)
+ X(t)
s + o( s)
2
?
?
1 ? ?
?
s + o( s)
(xk ? xk?1 )/ s = X(t) ? X(t)
2
?
?
?
s?f (yk ) = s?f (X(t)) + o( s),
where in the last equality we use yk ? X(t) = o(1). Thus (2.1) can be written as
?
1 ? ?
?
s + o( s)
X(t)
+ X(t)
2
?
? ?
?
3 s ?
1 ? ?
= 1?
s + o( s) ? s?f (X(t)) + o( s).
X(t) ? X(t)
t
2
?
By comparing the coefficients of s in (2.2), we obtain
(2.2)
? + 3 X? + ?f (X) = 0
X
t
?
for t > 0. The first initial condition is X(0) = x0 . Taking k = 1 in (2.1) yields (x2 ? x1 )/ s =
?
?
= 0 (vanishing initial
? s?f (y1 ) = o(1). Hence, the second initial condition is simply X(0)
velocity). In the formulation of [1] (see also [20]), the momentum coefficient (k ? 1)/(k + 2) is
?1
? 1), where ?k are iteratively defined as
replaced by ?k (?k?1
p
?k4 + 4?k2 ? ?k2
?k+1 =
(2.3)
2
?1
starting from ?0 = 1. A bit of analysis reveals that ?k (?k?1
? 1) asymptotically equals 1 ? 3/k +
2
O(1/k ), thus leading to the same ODE as (1.1).
2
Classical results in ODE theory do not directly imply the existence or uniqueness of the solution to
this ODE because the coefficient 3/t is singular at t = 0. In addition, ?f is typically not analytic
at x0 , which leads to the inapplicability of the power series method for studying singular ODEs.
Nevertheless, the ODE is well posed: the strategy we employ for showing this constructs a series of
ODEs approximating (1.2) and then chooses a convergent subsequence by some compactness arguments such as the Arzel?a-Ascoli theorem. A proof of this theorem can be found in the supplementary
material for this paper.
Theorem 2.1. For any f ? F? , ?L>0 FL and any x0 ? Rn , the ODE (1.2) with initial conditions
?
X(0) = x0 , X(0)
= 0 has a unique global solution X ? C 2 ((0, ?); Rn ) ? C 1 ([0, ?); Rn ).
3
Equivalence between the ODE and Nesterov?s scheme
We study the stable step size allowed for numerically solving the ODE in the presence of accumulated errors. The finite difference approximation of (1.2) by the forward Euler method is
X(t + ?t) ? 2X(t) + X(t ? ?t) 3 X(t) ? X(t ? ?t)
+
+ ?f (X(t)) = 0,
(3.1)
?t2
t
?t
which is equivalent to
3?t
3?t
X(t) ? ?t2 ?f (X(t)) ? 1 ?
X(t ? ?t).
X(t + ?t) = 2 ?
t
t
Assuming that f is sufficiently smooth, for small perturbations ?x, ?f (x + ?x) ? ?f (x) +
?2 f (x)?x, where ?2 f (x) is the Hessian of f evaluated at x. Identifying k = t/?t, the characteristic equation of this finite difference scheme is approximately
3?t
3?t
det ?2 ? 2 ? ?t2 ?2 f ?
?+1?
= 0.
(3.2)
t
t
The numerical stability of (3.1) with respect to accumulated errors is equivalent to this: all the roots
2
2
of (3.2) lie in the unit circle
? [9]. When ? f LIn (i.e., LIn ? ? f is positive semidefinite), if
?t/t small and ?t? < 2/ L, we see that all the roots of (3.2) lie in the unit circle. On the other
hand, if ?t > 2/ L, (3.2) can possibly have a root ? outside the unit circle, causing numerical
instability. Under our identification s = ?t2 , a step
? size of s = 1/L in Nesterov?s scheme (1.1) is
approximately equivalent to a step size of ?t = 1/ L in the forward Euler method, which is stable
for numerically integrating (3.1).
As a comparison, note that the corresponding ODE for gradient descent with updates xk+1 = xk ?
s?f (xk ), is
?
X(t)
+ ?f (X(t)) = 0,
whose finite difference scheme has the characteristic equation det(? ? (1 ? ?t?2 f )) = 0. Thus,
to guarantee ?In 1 ? ?t?2 f In in worst case analysis,
? one can only choose ?t ? 2/L for a
fixed step size, which is much smaller than the step size 2/ L for (3.1) when ?f is very variable,
i.e., L is large.
Next, we exhibit approximate equivalence between the ODE and Nesterov?s scheme in terms of
convergence rates. We first recall the original result from [11].
Theorem 3.1 (Nesterov). For any f ? FL , the sequence {xk } in (1.1) with step size s ? 1/L obeys
f (xk ) ? f ? ?
2kx0 ? x? k2
.
s(k + 1)2
Our first result indicates that the trajectory of ODE (1.2) closely resembles the sequence {xk } in
terms of the convergence rate to a minimizer x? .
Theorem 3.2. For any f ? F? , let X(t) be the unique global solution to (1.2) with initial condi?
tions X(0) = x0 , X(0)
= 0. For any t > 0,
f (X(t)) ? f ? ?
3
2kx0 ? x? k2
.
t2
Proof of Theorem 3.2. Consider the energy functional defined as
t
E(t) , t2 (f (X(t)) ? f ? ) + 2kX + X? ? x? k2 ,
2
whose time derivative is
?
? + 4hX + t X? ? x? , 3 X? + t Xi.
(3.3)
E? = 2t(f (X) ? f ? ) + t2 h?f, Xi
2
2
2
? + tX/2
? with ?t?f (X)/2, (3.3) gives
Substituting 3X/2
t
E? = 2t(f (X) ? f ? ) + 4hX ? x? , ? ?f (X)i = 2t(f (X) ? f ? ) ? 2thX ? x? , ?f (X)i ? 0,
2
where the inequality follows from the convexity of f . Hence by monotonicity of E and non?
negativity of 2kX + tX/2
? x? k2 , the gap obeys f (X(t)) ? f ? ? E(t)/t2 ? E(0)/t2 =
? 2 2
2kx0 ? x k /t .
4
A family of generalized Nesterov?s schemes
In this section we show how to exploit the power of the ODE for deriving variants of Nesterov?s
scheme. One would be interested in studying the ODE (1.2) with the number 3 appearing in the
? replaced by a general constant r as in
coefficient of X/t
? + r X? + ?f (X) = 0, X(0) = x0 , X(0)
?
X
= 0.
(4.1)
t
Using arguments similar to those in the proof of Theorem 2.1, this new ODE is guaranteed to assume
a unique global solution for any f ? F? .
4.1
Continuous optimization
To begin with, we consider a modified energy functional defined as
2
2t2
t
? ? x?
.
E(t) =
X(t)
(f (X(t)) ? f ? ) + (r ? 1)
X(t)
+
r?1
r?1
? = ?t?f (X), the time derivative E? is equal to
Since rX? + tX
2t2
4t
? + 2hX + t X? ? x? , rX? + tXi
?
(f (X) ? f ? ) +
h?f, Xi
r?1
r?1
r?1
4t
=
(f (X) ? f ? ) ? 2thX ? x? , ?f (X)i. (4.2)
r?1
A consequence of (4.2) is this:
Theorem 4.1. Suppose r > 3 and let X be the unique solution to (4.1) for some f ? F? . Then X
obeys
(r ? 1)2 kx0 ? x? k2
f (X(t)) ? f ? ?
2t2
and
Z ?
(r ? 1)2 kx0 ? x? k2
.
t(f (X(t)) ? f ? )dt ?
2(r ? 3)
0
Proof of Theorem 4.1. By (4.2), the derivative dE/dt equals
2(r ? 3)t
2(r ? 3)t
(f (X) ? f ? ) ? ?
(f (X) ? f ? ), (4.3)
r?1
r?1
where the inequality follows from the convexity of f . Since f (X) ? f ? , (4.3) implies that E is
non-increasing. Hence
2t(f (X) ? f ? ) ? 2thX ? x? , ?f (X)i ?
2t2
(f (X(t)) ? f ? ) ? E(t) ? E(0) = (r ? 1)kx0 ? x? k2 ,
r?1
4
yielding the first inequality of the theorem as desired. To complete the proof, by (4.2) it follows that
Z ?
Z ?
dE
2(r ? 3)t
?
(f (X) ? f )dt ? ?
dt = E(0) ? E(?) ? (r ? 1)kx0 ? x? k2 ,
r?1
dt
0
0
as desired for establishing the second inequality.
We now demonstrate faster convergence rates under the assumption of strong convexity. Given a
strongly convex function f , consider a new energy functional defined as
2
2t ?
(2r ? 3)2 t
?
3
?
?
E(t) = t (f (X(t)) ? f ) +
X(t) + 2r ? 3 X(t) ? x
.
8
? gives
As in Theorem 4.1, a more refined study of the derivative of E(t)
Theorem 4.2. For any f ? S?,L (Rn ), the unique solution X to (4.1) with r ? 9/2 obeys
5
f (X(t)) ? f ? ?
Cr 2 kx0 ? x? k2
?
t3 ?
for any t > 0 and a universal constant C > 1/2.
The restriction r ? 9/2 is an artifact required in the proof. We believe that this theorem should be
valid as long as r ? 3. For example, the solution to (4.1) with f (x) = kxk2 /2 is
X(t) =
2
r?1
2
?((r + 1)/2)J(r?1)/2 (t)
x0 ,
(4.4)
r?1
t 2
where J(r?1)/2 (?) is the first kind Bessel function of order (r?1)/2. For large t, this Bessel function
p
obeys J(r?1)/2 (t) = 2/(?t)(cos(t ? (r ? 1)?/4 ? ?/4) + O(1/t)). Hence,
f (X(t)) ? f ? . kx0 ? x? k2 /tr ,
in which the inequality fails if 1/tr is replaced by any higher order rate. For general strongly convex
functions, such refinement, if possible, might require a construction of a more sophisticated energy
functional and careful analysis. We leave this problem for future research.
4.2
Composite optimization
Inspired by Theorem 4.2, it is tempting to obtain such analogies for the discrete Nesterov?s scheme
as well. Following the formulation of [1], we consider the composite minimization:
minimize
n
x?R
f (x) = g(x) + h(x),
where g ? FL for some L > 0 and h is convex on Rn with possible extended value ?. Define the
proximal subgradient
h
i
x ? argminz kz ? (x ? s?g(x))k2 /(2s) + h(z)
.
Gs (x) ,
s
Parametrizing by a constant r, we propose a generalized Nesterov?s scheme,
xk = yk?1 ? sGs (yk?1 )
(4.5)
k?1
(xk ? xk?1 ),
yk = x k +
k+r?1
starting from y0 = x0 . The discrete analog of Theorem 4.1 is below, whose proof is also deferred to
the supplementary materials as well.
Theorem 4.3. The sequence {xk } given by (4.5) with r > 3 and 0 < s ? 1/L obeys
f (xk ) ? f ? ?
and
?
X
(r ? 1)2 kx0 ? x? k2
2s(k + r ? 2)2
(k + r ? 1)(f (xk ) ? f ? ) ?
k=1
5
(r ? 1)2 kx0 ? x? k2
.
2s(r ? 3)
The idea behind the proof is the same as that employed for Theorem 4.1; here, however, the energy
functional is defined as
E(k) = 2s(k + r ? 2)2 (f (xk ) ? f ? )/(r ? 1) + k(k + r ? 1)yk ? kxk ? (r ? 1)x? k2 /(r ? 1).
The first inequality in Theorem 4.3 suggests that the generalized Nesterov?s scheme still achieves
O(1/k 2 ) convergence rate. However, if the error bound satisfies
c
f (xk? ) ? f ? ? ?2
k
for some c > 0 and a dense subsequence {k ? }, i.e., |{k ? } ? {1, . . . , m}| ? ?m for any positive
integer m and some ? > 0, then the second inequality of the theorem is violated. Hence, the second
inequality is not trivial because it implies the error bound is in some sense O(1/k 2 ) suboptimal.
In closing, we would like to point out this new scheme is equivalent to setting ?k = (r?1)/(k+r?1)
?1
and letting ?k (?k?1
? 1) replace the momentum coefficient (k ? 1)/(k + r ? 1). Then, the equal
sign ? = ? in (2.3) has to be replaced by ? ? ?. In examining the proof of Theorem 1(b) in [20], we
can get an alternative proof of Theorem 4.3 by allowing (2.3), which appears in Eq. (36) in [20], to
be an inequality.
5
Accelerating to linear convergence by restarting
Although an O(1/k 3 ) convergence rate is guaranteed for generalized Nesterov?s schemes (4.5), the
example (4.4) provides evidence that O(1/poly(k)) is the best rate achievable under strong convexity. In contrast, the vanilla gradient method achieves linear convergence
O((1 ? ?/L)k ) and
p
[12] proposed a first-order method with a convergence rate of O((1 ? ?/L)k ), which, however,
requires knowledge of the condition number ?/L. While it is relatively easy to bound the Lipschitz
constant L by the use of backtracking [3, 19], estimating the strong convexity parameter ?, if not
impossible, is very challenging. Among many approaches to gain acceleration via adaptively estimating ?/L, [15] proposes a restarting procedure for Nesterov?s scheme in which (1.1) is restarted
with x0 = y0 := xk whenever ?f (yk )T (xk+1 ? xk ) > 0. In the language of ODEs, this gradi? negative along the trajectory. Although it has been
ent based restarting essentially keeps h?f, Xi
empirically observed that this method significantly boosts convergence, there is no general theory
characterizing the convergence rate.
In this section, we propose a new restarting scheme we call the speed restarting scheme. The underlying motivation is to maintain a relatively high velocity X? along the trajectory. Throughout this
section we assume f ? S?,L for some 0 < ? ? L.
?
Definition 5.1. For ODE (1.2) with X(0) = x0 , X(0)
= 0, let
2
?
dkX(u)k
> 0}
T = T (f, x0 ) = sup{t > 0 : ?u ? (0, t),
du
be the speed restarting time.
? decreases. The definition itself does not imply that
In words, T is the first time the velocity kXk
0 < T < ?, which is proven in the supplementary materials. Indeed, f (X(t)) is a decreasing
function before time T ; for t ? T ,
? 2
df (X(t))
? = ? 3 kXk
? 2 ? 1 dkXk ? 0.
= h?f (X), Xi
dt
t
2 dt
The speed restarted ODE is thus
? + 3 X(t)
?
X(t)
+ ?f (X(t)) = 0,
(5.1)
tsr
? Xi
? = 0 and between two consecutive restarts, tsr grows just
where tsr is set to zero whenever hX,
as t. That is, tsr = t ? ? , where ? is the latest restart time. In particular, tsr = 0 at t = 0. The
theorem below guarantees linear convergence of the solution to (5.1). This is a new result in the
literature [15, 10].
Theorem 5.2. There exists positive constants c1 and c2 , which only depend on the condition number
L/?, such that for any f ? S?,L , we have
c1 Lkx0 ? x? k2 ?c2 t?L
.
e
f (X sr (t)) ? f (x? ) ?
2
6
5.1
Numerical examples
Below we present a discrete analog to the restarted scheme. There, kmin is introduced to avoid
having consecutive restarts that are too close. To compare the performance of the restarted scheme
with the original (1.1), we conduct four simulation studies, including both smooth and non-smooth
objective functions. Note that the computational costs of the restarted and non-restarted schemes are
the same.
Algorithm 1 Speed Restarting Nesterov?s Scheme
input: x0 ? Rn , y0 = x0 , x?1 = x0 , 0 < s ? 1/L, kmax ? N+ and kmin ? N+
j?1
for k = 1 to kmax do
1
xk ? argminx ( 2s
kx ? yk?1 + s?g(yk?1 )k2 + h(x))
j?1
yk ? xk + j+2 (xk ? xk?1 )
if kxk ? xk?1 k < kxk?1 ? xk?2 k and j ? kmin then
j?1
else
j ?j+1
end if
end for
Quadratic. f (x) = 21 xT Ax + bT x is a strongly convex function, in which A is a 500 ? 500 random
positive definite matrix and b a random vector. The eigenvalues of A are between 0.001 and 1. The
vector b is generated as i. i. d. Gaussian random variables with mean 0 and variance 25.
Log-sum-exp.
f (x) = ? log
m
hX
i=1
i
exp((aTi x ? bi )/?) ,
where n = 50, m = 200, ? = 20. The matrix A = {aij } is a random matrix with i. i. d. standard
Gaussian entries, and b = {bi } has i. i. d. Gaussian entries with mean 0 and variance 2. This function
is not strongly convex.
Matrix completion. f (X) = 21 kXobs ? Mobs k2F + ?kXk? , in which the ground truth M is a
rank-5 random matrix of size 300 ? 300. The regularization parameter is set to ? = 0.05. The 5
singular values of M are 1, . . . , 5. The observed set is independently sampled among the 300 ? 300
entries so that 10% of the entries are actually observed.
Lasso in ?1 ?constrained form with large sparse design. f = 12 kAx ? bk2 s.t. kxk1 ? ?, where
A is a 5000 ? 50000 random sparse matrix with nonzero probability 0.5% for each entry and b is
generated as b = Ax0 + z. The nonzero entries of A independently follow the Gaussian distribution
with mean 0 and variance 1/25. The signal x0 is a vector with 250 nonzeros and z is i. i. d. standard
Gaussian noise. The parameter ? is set to kx0 k1 .
In these examples, kmin is set to be 10 and the step sizes are fixed to be 1/L. If the objective is in
composite form, the Lipschitz bound applies to the smooth part. Figures 1(a), 1(b), 1(c) and 1(d)
present the performance of the speed restarting scheme, the gradient restarting scheme proposed in
[15], the original Nesterov?s scheme and the proximal gradient method. The objective functions
include strongly convex, non-strongly convex and non-smooth functions, violating the assumptions
in Theorem 5.2. Among all the examples, it is interesting to note that both speed restarting scheme
empirically exhibit linear convergence by significantly reducing bumps in the objective values. This
leaves us an open problem of whether there exists provable linear convergence rate for the gradient
restarting scheme as in Theorem 5.2. It is also worth pointing that compared with gradient restarting,
the speed restarting scheme empirically exhibits more stable linear convergence rate.
6
Discussion
This paper introduces a second-order ODE and accompanying tools for characterizing Nesterov?s
accelerated gradient method. This ODE is applied to study variants of Nesterov?s scheme. Our
7
8
2
10
10
srN
grN
oN
PG
6
10
srN
grN
oN
PG
0
10
4
10
?2
10
2
10
?4
f ? f*
f ? f*
10
0
10
?6
10
?2
10
?8
10
?4
10
?10
10
?6
10
?12
0
200
400
600
800
1000
1200
10
1400
0
500
1000
iterations
1500
iterations
(a) min 12 xT Ax + bx
(b) min ? log(
2
Pm
i=1
(aT
i x?bi )/?
e
)
5
10
10
srN
grN
oN
PG
0
10
srN
grN
oN
PG
?2
10
0
10
?4
f ? f*
f?f
*
10
?6
10
?8
?5
10
10
?10
10
?12
10
?10
0
20
40
(c) min
60
80
1
kXobs
2
?
100
iterations
120
Mobs k2F
140
160
180
10
200
0
200
400
600
800
1000
1200
1400
iterations
+ ?kXk?
(d) min
1
kAx
2
? bk
2
s.t. kxk1 ? C
Figure 1: Numerical performance of speed restarting (srN), gradient restarting (grN) proposed in
[15], the original Nesterov?s scheme (oN) and the proximal gradient (PG)
approach suggests (1) a large family of generalized Nesterov?s schemes that are all guaranteed to
converge at the rate 1/k 2 , and (2) a restarted scheme provably achieving a linear convergence rate
whenever f is strongly convex.
In this paper, we often utilize ideas from continuous-time ODEs, and then apply these ideas to
discrete schemes. The translation, however, involves parameter tuning and tedious calculations.
This is the reason why a general theory mapping properties of ODEs into corresponding properties
for discrete updates would be a welcome advance. Indeed, this would allow researchers to only
study the simpler and more user-friendly ODEs.
7
Acknowledgements
We would like to thank Carlos Sing-Long and Zhou Fan for helpful discussions about parts of this
paper, and anonymous reviewers for their insightful comments and suggestions.
8
References
[1] A. Beck and M. Teboulle. A fast iterative shrinkage-thresholding algorithm for linear inverse
problems. SIAM J. Imaging Sci., 2(1):183?202, 2009.
[2] S. Becker, J. Bobin, and E. J. Cand`es. NESTA: a fast and accurate first-order method for sparse
recovery. SIAM Journal on Imaging Sciences, 4(1):1?39, 2011.
[3] S. Becker, E. J. Cand`es, and M. Grant. Templates for convex cone problems with applications
to sparse signal recovery. Mathematical Programming Computation, 3(3):165?218, 2011.
[4] A. Bloch (Editor). Hamiltonian and gradient flows, algorithms, and control, volume 3. American Mathematical Soc., 1994.
[5] F. H. Branin. Widely convergent method for finding multiple solutions of simultaneous nonlinear equations. IBM Journal of Research and Development, 16(5):504?522, 1972.
[6] A. A. Brown and M. C. Bartholomew-Biggs. Some effective methods for unconstrained optimization based on the solution of systems of ordinary differential equations. Journal of
Optimization Theory and Applications, 62(2):211?224, 1989.
[7] R. Hauser and J. Nedic. The continuous Newton?Raphson method can look ahead. SIAM
Journal on Optimization, 15(3):915?925, 2005.
[8] U. Helmke and J. Moore. Optimization and dynamical systems. Proceedings of the IEEE,
84(6):907, 1996.
[9] J. J. Leader. Numerical Analysis and Scientific Computation. Pearson Addison Wesley, 2004.
[10] R. Monteiro, C. Ortiz, and B. Svaiter. An adaptive accelerated first-order method for convex
optimization, 2012.
[11] Y. Nesterov. A method of solving a convex programming problem with convergence rate
O(1/k 2 ). In Soviet Mathematics Doklady, volume 27, pages 372?376, 1983.
[12] Y. Nesterov. Introductory lectures on convex optimization: A basic course, volume 87 of
Applied Optimization. Kluwer Academic Publishers, Boston, MA, 2004.
[13] Y. Nesterov. Smooth minimization of non-smooth functions. Mathematical programming,
103(1):127?152, 2005.
[14] Y. Nesterov. Gradient methods for minimizing composite objective function. CORE Discussion Papers, 2007.
[15] B. O?Donoghue and E. J. Cand`es. Adaptive restart for accelerated gradient schemes. Found.
Comput. Math., 2013.
[16] Y.-G. Ou. A nonmonotone ODE-based method for unconstrained optimization. International
Journal of Computer Mathematics, (ahead-of-print):1?21, 2014.
[17] R. T. Rockafellar. Convex analysis. Princeton Landmarks in Mathematics. Princeton University
Press, Princeton, NJ, 1997. Reprint of the 1970 original, Princeton Paperbacks.
[18] J. Schropp and I. Singer. A dynamical systems approach to constrained minimization. Numerical functional analysis and optimization, 21(3-4):537?551, 2000.
[19] P. Tseng. On accelerated proximal gradient methods for convex-concave optimization. submitted to SIAM J. 2008.
[20] P. Tseng. Approximation accuracy, gradient methods, and error bound for structured convex
optimization. Mathematical Programming, 125(2):263?295, 2010.
9
| 5322 |@word achievable:1 norm:1 tedious:1 open:1 d2:1 simulation:1 pg:5 tr:2 initial:6 series:2 nesta:1 tuned:1 ati:1 kx0:12 nonmonotone:1 comparing:1 dx:1 written:1 numerical:7 analytic:1 update:2 leaf:1 xk:40 vanishing:1 hamiltonian:1 core:1 iterates:1 provides:1 math:1 simpler:2 mathematical:4 along:2 c2:2 branin:1 differential:5 introductory:1 introduce:1 bobin:1 x0:19 indeed:2 cand:4 inspired:1 decreasing:1 dkx:1 increasing:2 provided:1 begin:1 estimating:2 underlying:1 kind:1 unified:1 finding:2 nj:1 guarantee:2 friendly:1 concave:1 doklady:1 k2:20 control:1 unit:3 grant:1 positive:4 before:1 engineering:1 limit:2 consequence:1 analyzing:1 establishing:1 path:1 approximately:2 might:1 resembles:1 equivalence:3 suggests:3 challenging:1 co:1 bi:3 obeys:7 practical:1 thirty:1 unique:5 definite:1 procedure:1 universal:1 significantly:2 composite:4 boyd:1 word:1 integrating:1 get:2 tsr:5 close:1 kmax:2 applying:1 instability:1 impossible:1 restriction:1 equivalent:4 reviewer:1 go:1 latest:1 starting:4 independently:2 convex:21 identifying:1 recovery:2 insight:2 deriving:1 stability:1 construction:1 suppose:1 user:1 exact:1 programming:4 homogeneous:1 velocity:4 particularly:1 observed:3 kxk1:2 electrical:1 worst:1 decrease:1 yk:15 convexity:5 inductively:1 nesterov:32 rigorously:1 dt2:1 depend:1 solving:2 serve:2 biggs:1 tx:3 soviet:1 derivation:1 fast:2 effective:1 outside:1 refined:1 pearson:1 whose:3 stanford:7 posed:1 supplementary:3 widely:1 statistic:1 itself:1 sequence:3 differentiable:2 eigenvalue:1 propose:2 causing:1 combining:1 date:1 achieve:1 ent:1 convergence:18 converges:1 leave:1 tions:1 derive:2 completion:1 eq:1 strong:3 soc:1 involves:1 implies:2 direction:1 closely:1 material:3 require:1 hx:5 anonymous:1 weijie:1 accompanying:1 sufficiently:1 ground:1 exp:2 mapping:1 bump:1 pointing:1 substituting:1 achieves:2 consecutive:3 uniqueness:1 tool:3 minimization:3 gaussian:5 modified:1 avoid:1 cr:1 zhou:1 shrinkage:1 es1:1 earliest:1 ax:2 improvement:1 rank:1 indicates:1 contrast:2 sense:1 helpful:1 accumulated:2 typically:1 bt:1 compactness:1 interested:1 provably:1 monteiro:1 among:4 development:2 proposes:1 constrained:2 equal:4 construct:1 having:2 look:1 k2f:2 future:1 t2:13 employ:1 inapplicability:1 beck:1 replaced:4 argminx:1 maintain:1 ortiz:1 interest:1 deferred:1 introduces:1 semidefinite:1 yielding:1 behind:1 bloch:1 accurate:1 byproduct:1 conduct:1 euclidean:1 taylor:1 srn:5 circle:3 desired:2 theoretical:1 modeling:1 teboulle:1 ax0:1 ordinary:4 cost:1 entry:6 euler:3 examining:1 too:1 hauser:1 proximal:4 svaiter:1 chooses:1 adaptively:1 international:1 siam:4 ansatz:1 continuously:2 choose:1 possibly:1 american:1 derivative:5 leading:2 rescaling:1 bx:1 de:2 coefficient:6 rockafellar:1 root:3 sup:1 carlos:1 candes:1 minimize:1 accuracy:1 variance:3 characteristic:2 yield:1 t3:1 conceptually:1 identification:1 grn:5 trajectory:4 rx:2 worth:1 researcher:1 ago:1 history:1 submitted:1 simultaneous:1 whenever:4 definition:2 energy:5 proof:10 conciseness:1 gain:1 sampled:1 recall:1 knowledge:2 ou:1 sophisticated:1 actually:1 back:1 appears:1 wesley:1 higher:1 dt:8 violating:1 restarts:2 follow:1 formulation:2 evaluated:1 strongly:9 just:1 hand:1 nonlinear:1 artifact:1 perhaps:1 scientific:1 grows:1 believe:1 mob:2 brown:1 equality:1 hence:5 regularization:1 read:1 iteratively:1 nonzero:2 moore:1 gradi:1 generalized:5 complete:1 demonstrate:2 txi:1 functional:6 empirically:3 volume:3 analog:2 interpretation:1 relating:1 numerically:2 kluwer:1 tuning:1 vanilla:2 unconstrained:2 mathematics:4 similarly:1 pm:1 closing:1 bartholomew:1 language:1 stable:3 lkx:2 inequality:9 employed:1 converge:2 bessel:2 tempting:1 stephen:1 signal:2 multiple:1 nonzeros:1 smooth:8 faster:1 academic:1 calculation:1 long:3 lin:2 raphson:1 kax:2 variant:3 basic:1 essentially:1 df:1 iteration:4 c1:2 condi:1 addition:1 kmin:4 ode:37 else:1 singular:3 publisher:1 sr:1 comment:1 flow:1 integer:1 call:1 presence:1 easy:1 lasso:1 suboptimal:1 idea:4 donoghue:1 det:2 whether:1 accelerating:2 becker:2 hessian:1 welcome:1 argminz:1 sign:1 discrete:5 four:1 nevertheless:1 achieving:1 k4:1 utilize:1 imaging:2 asymptotically:1 subgradient:1 year:1 sum:1 cone:1 inverse:1 family:3 throughout:2 bit:1 fl:6 bound:5 guaranteed:3 convergent:2 fan:1 quadratic:1 g:1 ahead:2 x2:1 speed:8 argument:2 min:4 relatively:2 department:3 structured:1 helmke:1 smaller:1 y0:4 equation:9 singer:1 addison:1 letting:1 end:2 studying:2 apply:1 appearing:1 alternative:1 existence:1 original:5 denotes:3 include:1 newton:1 exploit:1 emmanuel:1 k1:1 approximating:1 classical:1 objective:6 print:1 strategy:1 exhibit:5 gradient:22 thank:1 paperback:1 sci:1 restart:2 landmark:1 tseng:2 trivial:1 reason:1 provable:1 assuming:1 modeled:1 minimizing:2 negative:1 design:1 allowing:1 ascoli:1 sing:1 finite:3 descent:2 parametrizing:1 extended:1 ever:1 y1:1 rn:9 perturbation:1 introduced:1 bk:1 required:1 connection:1 established:2 boost:1 below:3 dynamical:2 including:1 power:2 nedic:1 scheme:41 imply:2 reprint:1 negativity:1 understanding:1 sg:1 literature:1 acknowledgement:1 lecture:1 interesting:2 suggestion:1 proven:2 analogy:1 thresholding:1 bk2:1 editor:1 translation:1 ibm:1 course:1 last:2 aij:1 allow:1 deeper:1 template:1 taking:3 characterizing:2 sparse:4 curve:2 valid:1 kz:1 forward:2 refinement:1 adaptive:2 approximate:2 restarting:16 keep:1 monotonicity:1 global:3 reveals:1 xi:6 leader:1 subsequence:2 continuous:6 iterative:1 why:1 ca:3 expansion:1 du:1 poly:1 dense:1 motivation:1 noise:1 allowed:1 x1:1 seminar:1 momentum:3 fails:1 comput:1 lie:2 kxk2:2 theorem:25 xt:3 showing:1 insightful:1 evidence:1 exists:2 kx:3 gap:1 boston:1 generalizing:1 led:1 backtracking:1 simply:1 lagrange:1 kxk:7 restarted:7 applies:1 minimizer:2 satisfies:1 relies:1 truth:1 ma:1 acceleration:2 careful:1 lipschitz:5 replace:1 reducing:1 e:3 accelerated:8 violated:1 princeton:4 |
4,775 | 5,323 | Learning Distributed Representations for Structured
Output Prediction
Vivek Srikumar?
University of Utah
[email protected]
Christopher D. Manning
Stanford University
[email protected]
Abstract
In recent years, distributed representations of inputs have led to performance gains
in many applications by allowing statistical information to be shared across inputs. However, the predicted outputs (labels, and more generally structures) are
still treated as discrete objects even though outputs are often not discrete units
of meaning. In this paper, we present a new formulation for structured prediction where we represent individual labels in a structure as dense vectors and allow
semantically similar labels to share parameters. We extend this representation
to larger structures by defining compositionality using tensor products to give a
natural generalization of standard structured prediction approaches. We define a
learning objective for jointly learning the model parameters and the label vectors
and propose an alternating minimization algorithm for learning. We show that
our formulation outperforms structural SVM baselines in two tasks: multiclass
document classification and part-of-speech tagging.
1
Introduction
In recent years, many computer vision and natural language processing (NLP) tasks have benefited
from the use of dense representations of inputs by allowing superficially different inputs to be related
to one another [26, 9, 7, 4]. For example, even though words are not discrete units of meaning, traditional NLP models use indicator features for words. This forces learning algorithms to learn separate
parameters for orthographically distinct but conceptually similar words. In contrast, dense vector
representations allow sharing of statistical signal across words, leading to better generalization.
Many NLP and vision problems are structured prediction problems. The output may be an atomic
label (tasks like document classification) or a composition of atomic labels to form combinatorial
objects like sequences (e.g. part-of-speech tagging), labeled trees (e.g. parsing) or more complex
graphs (e.g. image segmentation). Despite both the successes of distributed representations for
inputs and the clear similarities over the output space, it is still usual to handle outputs as discrete
objects. But are structures, and the labels that constitute them, really discrete units of meaning?
Consider, for example, the popular 20 Newsgroups dataset [13] which presents the multiclass
classification problem of identifying a newsgroup label given the text of a posting. Labels
include comp.os.mswindows.misc, sci.electronics, comp.sys.mac.hardware,
rec.autos and rec.motorcycles. The usual strategy is to train a classifier that uses separate
weights for each label. However, the labels themselves have meaning that is independent of the training data. From the label, we can see that comp.os.mswindows.misc, sci.electronics
and comp.sys.mac.hardware are semantically closer to each other than the other two. A similar argument can be made for not just atomic labels but their compositions too. For example, a
part-of-speech tagging system trained as a sequence model might have to learn separate parameters
?
This work was done when the author was at Stanford University.
1
for the JJ?NNS and JJR?NN transitions even though both encode a transition from an adjective to
a noun. Here, the similarity of the transitions can be inferred from the similarity of its components.
In this paper, we propose a new formulation for structured output learning called D ISTRO (DIStributed STRuctred Output), which accounts for the fact that labels are not atomic units of meaning.
We model label meaning by representing individual labels as real valued vectors. Doing so allows us
to capture similarities between labels. To allow for arbitrary structures, we define compositionality
of labels as tensor products of the label vectors corresponding to its sub-structures. We show that
doing so gives us a natural extension of standard structured output learning approaches, which can
be seen as special cases with one-hot label vectors.
We define a learning objective that seeks to jointly learn the model parameters along with the label
representations and propose an alternating algorithm for minimizing the objective for structured
hinge loss. We evaluate our approach on two tasks which have semantically rich labels: multiclass
classification on the newsgroup data and part-of-speech tagging for English and Basque. In all cases,
we show that D ISTRO outperforms the structural SVM baselines.
1.1
Related Work
This paper considers the problem of using distributed representations for arbitrary structures and is
related to recent work in deep learning and structured learning. Recent unsupervised representation
learning research has focused on the problem of embedding inputs in vector spaces [26, 9, 16, 7].
There has been some work [22] on modeling semantic compositionality in NLP, but the models do
not easily generalize to arbitrary structures. In particular, it is not easy to extend these approaches
to use advances in knowledge-driven learning and inference that standard structured learning and
prediction algorithms enable.
Standard learning approaches for structured output allow for modeling arbitrarily complex structures
(subject to inference difficulties) and structural SVMs [25] or conditional random fields [12] are
commonly used. However, the output itself is treated as a discrete object and similarities between
outputs are not modeled. For multiclass classification, the idea of classifying to a label set that
follow a known hierarchy has been explored [6], but such a taxonomy is not always available.
The idea of distributed representations for outputs has been discussed in the connectionist literature
since the eighties [11, 21, 20]. In recent years, we have seen several lines of research that address the
problem in the context of multiclass classification by framing feature learning as matrix factorization
or sparse encoding [23, 1, 3]. As in this paper, the goal has often explicitly been to discover shared
characteristics between the classes [2]. Indeed, the inference formulation we propose is very similar
to inference in these lines of work. Also related is recent research in the NLP community that
explores the use of tensor decompositions for higher order feature combinations [14]. The primary
novelty in this paper is that in addition to representing atomic labels in a distributed manner, we
model their compositions in a natural fashion to generalize standard structured prediction.
2
Preliminaries and Notation
In this section, we give a very brief overview of structured prediction with the goal of introducing
notation and terminology for the next sections. We represent inputs to the structured prediction
problem (such as, sentences, documents or images) by x ? X and output structures (such as labels
or trees) by y ? Y. We define the feature function ? : X ? Y ? <n that captures the relationship
between the input x and the structure y as an n dimensional vector. A linear model scores the
structure y with a weight vector w ? <n as wT ?(x, y). We predict the output for an input x as
arg maxy wT ?(x, y). This problem of inference is a combinatorial optimization problem.
We will use the structures in Figure 1 as running examples. In the case of multiclass classification,
the output y is one of a finite set of labels (Figure 1, left). For more complex structures, the feature
vector is decomposed over the parts of the structure. For example, the usual representation of a
first-order linear sequence model (Figure 1, middle) decomposes the sequence into emissions and
transitions and the features decompose over these [8]. In this case, each emission is associated with
one label and a transition is associated with an ordered pair of labels.
2
Compositional part
Label yp = (y0 , y1 )
y
y0
y1
y2
y0
y1
y2
Atomic part
Label yp = (y)
x
Atomic part
Label yp = (y0 )
Multiclass classification
x
Compositional part
Label yp = (y0 , y1 , y2 )
Sequence labeling. The emissions are
atomic and the transitions are compositional.
x
A purely compositional part
Figure 1: Three examples of structures. In all cases, x represents the input and the y?s denote the outputs to
be predicted. Here, each square represents a part as defined in the text and circles represent random variables
for inputs and outputs (as in factor graphs). The left figure shows multiclass classification, which has an atomic
part associated with exactly one label. The middle figure shows a first-order sequence labeling task that has both
atomic parts (emissions) and compositional ones (transitions). The right figure shows a purely compositional
part where all outputs interact. The feature functions for these structures are shown at the end of Section 3.1.
In the general case, we denote the parts (or equivalently, factors in a factor graph) in the structure
for input x by ?x . Each part p ? ?x is associated with a list of discrete labels, denoted by yp =
(yp0 , yp1 , ? ? ? ). Note that the size of the list yp is a modeling choice; for example, transition parts in
the first-order Markov model correspond to two consecutive labels, as shown in Figure 1.
We denote the set of labels in the problem as L = {l1 , l2 , ? ? ? , lM } (e.g. the set of part-of-speech
tags). All the elements of the part labels yp are members of this set. For notational convenience,
we denote the first element of the list yp by yp (without boldface) and the rest by yp1: . In the rest of
the paper, we will refer to a part associated with a single label as atomic and all other parts where
yp has more than one element as compositional. In Figure 1, we see examples of a purely atomic
structure (multiclass classification), a purely compositional structure (right) and a structure that is a
mix of the two (first order sequence, middle).
The decomposition of the structure decomposes the feature function over the parts as
X
?(x, y) =
?p (x, yp ) .
(1)
p??x
The scoring function wT ?(x, y) also decomposes along this sum. Standard definitions of structured prediction models leave the definition of the part-specific feature function ?p to be problem
dependent. We will focus on this aspect in Section 3 to define our model.
With definitions of a scoring function and inference, we can state the learning objective. Given a
collection of N training examples of the form (xi , yi ), training is the following regularized risk
minimization problem:
?
1 X
minn wT w +
L(xi , yi ; w).
(2)
w?< 2
N i
Here, L represents a loss function such as the hinge loss (for structural SVMs) or the log loss (for
conditional random fields) and penalizes model errors.The hyper-parameter ? trades off between
generalization and accuracy.
3
Distributed Representations for Structured Output
As mentioned in Section 2, the choice of the feature function ?p for a part p is left to be problem
specific. The objective is to capture the correlations between the relevant attributes of the input x
and the output labels yp . Typically, this is done by conjoining the labels yp with a user-defined
feature vector ?p (x) that is dependent only on the input.
3
When applied to atomic parts (e.g. multiclass classification), conjoining the label with the input features effectively allocates a different portion of the weight vector for each label. For compositional
parts (e.g. transitions in sequence models), this ensures that each combination of labels is associated
with a different portion of the weight vector. The implicit assumption in this design is that labels and
label combinations are distinct units of meaning and hence do not share any parameters across them.
In this paper, we posit that in most naturally occurring problems and their associated labels, this
assumption is not true. In fact, labels often encode rich semantic information with varying degrees
of similarities to each other. Because structures are composed of atomic labels, the same applies to
structures too.
From Section 2, we see that for the purpose of inference, structures are completely defined by
their feature vectors, which are decomposed along the atomic and compositional parts that form the
structure. Thus, our goal is to develop a feature representation for labeled parts that exploits label
similarity. More explicitly, our desiderata are:
1. First, we need to be able to represent labeled atomic parts using a feature representation
that accounts for relatedness of labels in such a way that statistical strength (i.e. weights)
can be shared across different labels.
2. Second, we need an operator that can construct compositional parts to build larger structures so that the above property can be extended to arbitrary structured output.
3.1
The D ISTRO model
In order to assign a notion of relatedness between labels, we associate a d dimensional unit vector
al to each label l ? L. We will refer to the d ? M matrix comprising of all the M label vectors as
A, the label matrix.
We can define the feature vectors for parts, and thus entire structures, using these label vectors. To
do so, we define the notion of a feature tensor function for a part p that has been labeled with a list
of m labels yp . The feature tensor function is a function ?p that maps the input x and the label list
yp associated with the part to a tensor of order m + 1. The tensor captures the relationships between
the input and all the m labels associated with it. We recursively define the feature tensor function
using the label vectors as:
alyp ? ?p (x),
p is atomic,
(3)
?p (x, yp , A) =
1:
alyp ? ?p x, yp , A , p is compositional.
Here, the symbol ? denotes the tensor product operation. Unrolling the recursion in this definition
shows that the feature tensor function for a part is the tensor product of the vectors for all the labels
associated with that part and the feature vector associated with the input for the part. For an input x
and a structure y, we use the feature tensor function to define its feature representation as
X
?A (x, y) =
vec (?p (x, yp , A))
(4)
p??x
Here, vec(?) denotes the vectorization operator that converts a tensor into a vector by stacking its
elements. Figure 2 shows an example of the process of building the feature vector for a part that is
labeled with two labels. With this definition of the feature vector, we can use the standard approach
to score structures using a weight vector as wT ?A (x, y).
In our running examples from Figure 1, we have the following definitions of feature functions for
each of the cases:
1. Purely atomic part, multiclass classification (left): Denote the feature vector associated
with x as ?. For an atomic part, the definition of the feature tensor function in Equation (3)
T
effectively produces a d ?
|?| matrix aly ? . Thus the feature vector for the structure y is
T
?A (x, y) = vec aly ? . For this case, the score for an input x being assigned a label y
can be explicitly be written as the following summation:
wT ?A (x, y) =
|?|
d X
X
i=0 j=0
4
wdj+i aly ,i ?j
vec (
?
?
)?
vec (
al1 ? <d
al2 ? <d
?p (x) ? <N
)
d?d?N
Feature tensor
?
Feature vector ? <d
2
N
Figure 2: This figure summarizes feature vector generation for a compositional part labeled with two labels
l1 and l2 . Each label is associated with a d dimensional label vector and the feature vector for the input is N
dimensional. Vectorizing the feature tensor produces a final feature vector that is a d2 N -dimensional vector.
2. Purely compositional part (right): For a compositional part, the feature tensor function
produces a tensor whose elements effectively enumerate every possible combination of
elements of input vector ?p (x) and the associated label vectors.
So, the feature vector for
the structure is ?A (x, y) = vec aly0 ? aly1 ? aly2 ? ?p (x) .
3. First order sequence (middle): This structure presents a combination of atomic and compositional parts. Suppose we denote the input emission features by ?E,i for the ith label
and the input features corresponding to the transition1 from yi to yi+1 by ?T,i . With this
notation, we can define the feature vector for the structure as
X
X
?A (x, y) =
vec alyi ? ?E,i +
vec alyi ? alyi+1 ? ?T,i .
i
3.2
i
Discussion
Connection to standard structured prediction For a part p, a traditional structured model conjoins all its associated labels to the input feature vector to get the feature vector for that assignment
of the labels. According to the definition of Equation (3), we propose that these label conjunctions
should be replaced with a tensor product, which generalizes the standard method. Indeed, if the
labels are represented via one-hot vectors, then we would recover standard structured prediction
where each label (or group of labels) is associated with a separate section of the weight vector. For
example, for multiclass classification, if each label is associated with a separate one-hot vector, then
the feature tensor for a given label will be a matrix where exactly one column is the input feature
vector ?p (x) and all other entries are zero. This argument also extends to compositional parts.
Dimensionality of label vectors If labels are represented by one-hot vectors, the dimensionality
of the label vectors will be M , the number of labels in the problem. However, in D ISTRO, in addition
to letting the label vectors be any unit vector, we can also allow them to exist in a lower dimensional
space. This presents us with a decision with regard to the dimensionality d.
The choice of d is important for two reasons. First, it determines the number of parameters in the
model. If a part is associated with m labels, recall that the feature tensor function produces a m + 1
order tensor formed by taking the tensor product of the m label vectors and the input features. That
is, the feature vector for the part is a dm |?p (x)| dimensional vector. (See 2 for an illustration.)
Smaller d thus leads to smaller weight vectors. Second, if the dimensionality of the label vectors
is lower, it encourages more weights to be shared across labels. Indeed, for purely atomic and
compositional parts if the labels are represented by M dimensional vectors, we can show that for
any weight vector that scores these labels via the feature representation defined in Equation (4), there
is another weight vector that assigns the same scores using one-hot weight vectors.
4
Learning Weights and Label Vectors
In this section, we will address the problem of learning the weight vectors w and the label vectors
A from data. We are given a training set with N examples of the form (xi , yi ). The goal of learning
1
In a linear sequence model defined as a CRF or a structural SVM, these transition input features can simply
be an indicator that selects a specific portion of the weight vector.
5
is to minimize regularized risk over the training set. This leads to a training objective similar to
that of structural SVMs or conditional random fields (Equation (2)). However, there are two key
differences. First, the feature vectors for structures are not fixed as in structural SVMs or CRFs but
are functions of the label vectors. Second, the minimization is over not just the weight vectors, but
also over the label vectors that require regularization.
In order to encourage the labels to share weights, we propose to impose a rank penalty over the
label matrix A in the learning objective. Since the rank minimization problem is known to be
computationally intractable in general [27], we use the well known nuclear norm surrogate to replace
the rank [10]. This gives us the learning objective defined as f below:
?1 T
1 X
f (w, A) =
w w + ?2 ||A||? +
L(xi , yi ; w, A)
(5)
2
N i
Here, the ||A||? is the nuclear norm of A, defined as the sum of the singular values of the matrix.
Compared to the objective in Equation (2), the loss function L is also dependent of the label matrix
via the new definition of the features. In this paper, we instantiate the loss using the structured hinge
loss [25]. That is, we define L to be
L(xi , yi ; w, A) = max wT ?A (xi , y) + ?(y, yi ) ? wT ?A (xi , yi )
(6)
y
Here, ? is the Hamming distance. This defines the D ISTRO extension of the structural SVM.
The goal of learning is to minimize the objective function f in terms of both its parameters w and
A, where each column of A is restricted to be a unit vector by definition. However, the objective
is not longer jointly convex in both w and A because of the product terms in the definition of the
feature tensor.
We use an alternating minimization algorithm for solving the optimization problem (Algorithm 1).
If the label matrix A is fixed, then so are the feature representations of structures (from Equation
(4)). Thus, for a fixed A (lines 2 and 5), the problem of minimizing f (w, A) with respect to only w
is identical to the learning problem of structural SVMs. Since gradient computation and inference
do not change from the usual setting, we can solve this minimization over w using stochastic subgradient descent (SGD). For fixed weight vectors (line 4), we implemented stochastic sub-gradient
descent using the proximal gradient method [18] for solving for A. The supplementary material
gives further details about the steps of the algorithm.
Algorithm 1 Learning algorithm by alternating minimization.
The goal is to solve
minw,A f (w, A). The input to the problem is a training set of examples consisting of pairs of
labeled inputs (xi , yi ) and T , the number of iterations.
1:
2:
3:
4:
5:
6:
7:
Initialize A0 randomly
Initialize w0 = minw f (w, A0 )
for t = 1, ? ? ? , T do
At ? minA f (wt?1 , A)
wt+1 ? minw f (w, At )
end for
return (wT +1 , AT )
Even though the objective function is not jointly convex in w and A, in our experiments (Section
5), we found that in all but one trial, the non-convexity of the objective did not affect performance.
Because the feature functions are multilinear in w and A, multiple equivalent solutions can exist
(from the perspective of the score assigned to structures) and the eventual point of convergence is
dependent on the initialization.
For regularizing the label matrix, we also experimented with the Frobenius norm and found that not
only does the nuclear norm have an intuitive explanation (rank minimization) but also performed
better. Furthermore, the proximal method itself does not add significantly to the training time because the label matrix is small. In practice, training time is affected by the density of the label
vectors and sparser vectors correspond to faster training because the sparsity can be used to speed
up dot product computation. Prediction is as fast as inference in standard models, however, because
the only change is in feature computation via the vectorization operator, which can be performed
efficiently.
6
5
Experiments
We demonstrate the effectiveness of D ISTRO on two tasks ? document classification (purely atomic
structures) and part-of-speech (POS) tagging (both atomic and compositional structures). In both
cases, we compare to structural SVMs ? i.e. the case of one-hot label vectors ? as the baseline.
We selected the hyper-parameters for all experiments by cross validation. We ran the alternating
algorithm for 5 epochs for all cases with 5 epochs of SGD for both the weight and label vectors.
We allowed the baseline to run for 25 epochs over the data. For the proposed method, we ran all
the experiments five times with different random initializations for the label vectors and report the
average accuracy. Even though the objective is not convex, we found that the learning algorithm
converged quickly in almost all trials. When it did not, the objective value on the training set at the
end of each alternating SGD step in the algorithm was a good indicator for ill-behaved initializations.
This allowed us to discard bad initializations during training.
5.1
Atomic structures: Multiclass Classification
Our first application is the problem of document classification with the 20 Newsgroups Dataset [13].
This dataset is collection of about 20,000 newsgroup posts partitioned roughly evenly among 20
newsgroups. The task is to predict the newsgroup label given the post. As observed in Section 1,
some newsgroups are more closely related to each other than others.
We used the ?bydate? version of the data with tokens as features. Table 1 reports the performance of
the baseline and variants of D ISTRO for newsgroup classification. The top part of the table compares
the baseline to our method and we see that modeling the label semantics gives us a 2.6% increase in
accuracy. In a second experiment (Table 1, bottom), we studied the effect of explicitly reducing the
label vector dimensionality. We see that even with 15 dimensional vectors, we can outperform the
baseline and the performance of the baseline is almost matched with 10 dimensional vectors. Recall
that the size of the weight vector increases with increasing label vector dimensionality (see Figure
2). This motivates a preference for smaller label vectors.
Algorithm
Structured SVM
D ISTRO
D ISTRO
D ISTRO
Label Matrix Rank Average accuracy (%)
20
81.4
19
84.0
Reduced dimensionality setting
15
83.1
10
80.9
Table 1: Results on 20 newsgroup classification. The top part of the table compares the baseline against the
full D ISTRO model. The bottom part shows the performance of two versions of D ISTRO where the dimensionality of the label vectors is fixed. Even with 10-dimensional vectors, we can almost match the baseline.
5.2
Compositional Structures: Sequence classification
We evaluated D ISTRO for English and Basque POS tagging using first-order sequence models.
English POS tagging has been long studied using the Penn Treebank data [15]. We used the standard
train-test split [8, 24] ? we trained on sections 0-18 of the Treebank and report performance on
sections 22-24. The data is labeled with 45 POS labels. Some labels are semantically close to each
other because they express variations of a base part-of-speech tag. For example, the labels NN, NNS,
NNP and NNPS indicate singular and plural versions of common and proper nouns
We used the Basque data from the CoNLL 2007 shared task [17] for training the Basque POS tagger.
This data comes from the 3LB Treebank. There are 64 fine grained parts of speech. Interestingly,
the labels themselves have a structure. For example, the labels IZE and ADJ indicate a noun and
an adjective respectively. However, Basque can take internal noun ellipsis inside noun-forms, which
are represented with tags like IZE IZEELI and ADJ IZEELI to indicate nouns and adjectives
with internal ellipses.
In both languages, many labels and transitions between labels are semantically close to each other.
This observation has led, for example, to the development of the universal part-of-speech tag set
7
[19]. Clearly, the labels should not be treated as independent units of meaning and the model should
be allowed to take advantage of the dependencies between labels.
Language
English
Basque
Algorithm
Structured SVM
D ISTRO
D ISTRO
Structured SVM
D ISTRO
Label Matrix Rank
45
5
20
64
58
Average accuracy (%)
96.2
95.1
96.7
91.5
92.4
Table 2: Results on part-of-speech tagging. The top part of the table shows results on English, where we see
a 0.5% gain in accuracy. The bottom part shows Basque results where we see a nearly 1% improvement.
For both languages, we extracted the following emission features: indicators for the words, their
prefixes and suffixes of length 3, the previous and next words and the word shape according to the
Stanford NLP pipeline2,3 . Table 2 presents the results for the two languages. We evaluate using the
average accuracy over all tags. In the English case, we found that the performance plateaued for any
label matrix with rank greater than 20 and we see an improvement of 0.5% accuracy. For Basque,
we see an improvement of 0.9% over the baseline.
Note that unlike the atomic case, the learning objective for the first order Markov model is not even
bilinear in the weights and the label vectors. However, in practice, we found that this did not cause
any problems. In all but one run, the test performance remained consistently higher than the baseline.
Moreover, the outlier converged to a much higher objective value; it could easily be identified. As
an analysis experiment, we initialized the model with one-hot vectors (i.e. the baseline) and found
that this gives us similar improvements as reported in the table.
6
Conclusion
We have presented a new model for structured output prediction called Distributed Structured Output
(D ISTRO). Our model is motivated by two observations. First, distributed representations for inputs
have led to performance gains by uncovering shared characteristics across inputs. Second, often,
structures are composed of semantically rich labels and sub-structures. Just like inputs, similarities
between components of structures can be exploited for better performance. To take advantage of
similarities among structures, we have proposed to represent labels by real-valued vectors and model
compositionality using tensor products between the label vectors. This not only lets semantically
similar labels share parameters, but also allows construction of complex structured output that can
take advantage of similarities across its component parts.
We have defined the objective function for learning with D ISTRO and presented a learning algorithm
that jointly learns the label vectors along with the weights using alternating minimization. We
presented an evaluation of our approach for two tasks ? document classification, which is an instance
of multiclass classification, and part-of-speech tagging for English and Basque, modeled as firstorder sequence models. Our experiments show that allowing the labels to be represented by realvalued vectors improves performance over the corresponding structural SVM baselines.
Acknowledgments
We thank the anonymous reviewers for their valuable comments. Stanford University gratefully acknowledges the support of the Defense Advanced Research Projects Agency (DARPA) Deep Exploration and Filtering of Text (DEFT) Program under Air Force Research Laboratory (AFRL) contract
no. FA8750-13-2-0040. Any opinions, findings, and conclusion or recommendations expressed in
this material are those of the authors and do not necessarily reflect the view of the DARPA, AFRL,
or the US government.
2
http://nlp.stanford.edu/software/corenlp.shtml
Note that our POS systems are not state-of-the-art implementations, which typically use second order
Markov models with additional features and specialized handling of unknown words. However, surprisingly,
for Basque, even the baseline gives better accuracy than the second order TnT tagger[5, 19].
3
8
References
[1] J. Abernethy, F. Bach, T. Evgeniou, and J. Vert. Low-rank matrix factorization with attributes. arXiv
preprint cs/0611124, 2006.
[2] Y. Amit, M. Fink, N. Srebro, and S. Ullman. Uncovering shared structures in multiclass classification. In
International Conference on Machine Learning, 2007.
[3] A. Argyriou, T. Evgeniou, and M. Pontil. Multi-task feature learning. Advances in Neural Information
Processing Systems, 2007.
[4] Y. Bengio, A. Courville, and P. Vincent. Representation learning: A review and new perspectives. IEEE
Transactions on Pattern Analysis and Machine Intelligence, 2013.
[5] T. Brants. TnT: a statistical part-of-speech tagger. In Conference on Applied Natural Language Processing, 2000.
[6] N. Cesa-Bianchi, C. Gentile, and L. Zaniboni. Hierarchical classification: combining bayes with svm. In
International Conference on Machine learning, 2006.
[7] A. Coates, A. Ng, and H. Lee. An analysis of single-layer networks in unsupervised feature learning. In
International Conference on Artificial Intelligence and Statistics, 2011.
[8] M. Collins. Discriminative training methods for hidden Markov models: Theory and experiments with
perceptron algorithms. In Conference on Empirical Methods in Natural Language Processing, 2002.
[9] R. Collobert, J. Weston, L. Bottou, M. Karlen, K. Kavukcuoglu, and P. Kuksa. Natural language processing (almost) from scratch. Journal for Machine Learning Research, 12, 2011.
[10] M. Fazel, H. Hindi, and S. Boyd. Rank minimization and applications in system theory. In Proceedings
of the American Control Conference, volume 4, 2004.
[11] G. E. Hinton. Representing part-whole hierarchies in connectionist networks. In Annual Conference of
the Cognitive Science Society, 1988.
[12] J. Lafferty, A. McCallum, and F. Pereira. Conditional random fields: Probabilistic models for segmenting
and labeling sequence data. In Machine Learning, 2001.
[13] K. Lang. Newsweeder: Learning to filter netnews. In International Conference on Machine Learning,
1995.
[14] T. Lei, Y. Xin, Y. Zhang, R. Barzilay, and T. Jaakkola. Low-rank tensors for scoring dependency structures. In Annual Meeting of the Association for Computational Linguistics, 2014.
[15] M. Marcus, G. Kim, M. Marcinkiewicz, R. MacIntyre, A. Bies, M. Ferguson, K. Katz, and B. Schasberger.
The Penn Treebank: Annotating Predicate Argument Structure. In Workshop on Human Language Technology, 1994.
[16] T. Mikolov, K. Chen, G. Corrado, and J. Dean. Efficient estimation of word representations in vector
space. arXiv preprint arXiv:1301.3781, 2013.
[17] J. Nivre, J. Hall, S. K?ubler, R. McDonald, J. Nilsson, S. Riedel, and D. Yuret. The CoNLL 2007 shared
task on dependency parsing. In CoNLL shared task session of EMNLP-CoNLL, 2007.
[18] N. Parikh and S. Boyd. Proximal algorithms. Foundations and Trends in optimization, 1(3), 2013.
[19] S. Petrov, D. Das, and R. McDonald. A universal part-of-speech tagset. arXiv preprint arXiv:1104.2086,
2011.
[20] T. A Plate. Holographic reduced representations. IEEE Transactions on Neural Networks, 6(3), 1995.
[21] P. Smolensky. Tensor product variable binding and the representation of symbolic structures in connectionist systems. Artificial intelligence, 46(1), 1990.
[22] R. Socher, B. Huval, C. Manning, and A. Ng. Semantic Compositionality Through Recursive MatrixVector Spaces. In Empirical Methods in Natural Language Processing, 2012.
[23] N. Srebro, J. Rennie, and T. Jaakkola. Maximum-margin matrix factorization. In Advances in Neural
Information Processing Systems, 2004.
[24] K. Toutanova, D. Klein, C. Manning, and Y. Singer. Feature-rich part-of-speech tagging with a cyclic
dependency network. In Conference of the North American Chapter of the Association for Computational
Linguistics on Human Language Technology, 2003.
[25] I. Tsochantaridis, T. Joachims, T. Hofmann, and Y. Altun. Large margin methods for structured and
interdependent output variables. Journal for Machine Learning Research, 2005.
[26] J. Turian, L. Ratinov, and Y. Bengio. Word Representations: A Simple and General Method for SemiSupervised Learning. In Annual Meeting of the Association for Computational Linguistics, 2010.
[27] L. Vandenberghe and S. Boyd. Semidefinite programming. SIAM review, 38(1), 1996.
9
| 5323 |@word trial:2 version:3 middle:4 norm:4 d2:1 seek:1 decomposition:2 sgd:3 recursively:1 electronics:2 cyclic:1 score:6 document:6 interestingly:1 prefix:1 fa8750:1 outperforms:2 adj:2 lang:1 written:1 parsing:2 shape:1 hofmann:1 intelligence:3 instantiate:1 selected:1 mccallum:1 sys:2 ith:1 preference:1 nnp:1 zhang:1 five:1 tagger:3 along:4 bydate:1 inside:1 manner:1 tagging:10 indeed:3 kuksa:1 themselves:2 roughly:1 multi:1 decomposed:2 unrolling:1 increasing:1 project:1 discover:1 notation:3 matched:1 moreover:1 finding:1 every:1 firstorder:1 fink:1 exactly:2 classifier:1 control:1 unit:9 penn:2 segmenting:1 despite:1 encoding:1 bilinear:1 might:1 initialization:4 studied:2 factorization:3 fazel:1 acknowledgment:1 atomic:25 practice:2 recursive:1 pontil:1 universal:2 empirical:2 significantly:1 vert:1 boyd:3 word:10 symbolic:1 get:1 convenience:1 close:2 tsochantaridis:1 operator:3 altun:1 context:1 risk:2 equivalent:1 map:1 reviewer:1 dean:1 crfs:1 convex:3 focused:1 identifying:1 assigns:1 yuret:1 nuclear:3 vandenberghe:1 deft:1 embedding:1 handle:1 notion:2 variation:1 hierarchy:2 suppose:1 construction:1 user:1 programming:1 us:1 associate:1 element:6 trend:1 rec:2 srikumar:1 labeled:8 observed:1 bottom:3 preprint:3 capture:4 ensures:1 trade:1 valuable:1 ran:2 mentioned:1 agency:1 convexity:1 trained:2 solving:2 purely:8 completely:1 easily:2 po:6 darpa:2 represented:5 chapter:1 train:2 distinct:2 fast:1 artificial:2 labeling:3 hyper:2 netnews:1 abernethy:1 whose:1 stanford:6 larger:2 valued:2 solve:2 supplementary:1 annotating:1 rennie:1 statistic:1 jointly:5 itself:2 final:1 sequence:14 advantage:3 propose:6 product:10 relevant:1 motorcycle:1 combining:1 intuitive:1 frobenius:1 convergence:1 produce:4 leave:1 object:4 develop:1 barzilay:1 implemented:1 c:3 predicted:2 indicate:3 come:1 posit:1 closely:1 attribute:2 filter:1 stochastic:2 exploration:1 human:2 enable:1 opinion:1 material:2 require:1 government:1 assign:1 generalization:3 really:1 preliminary:1 decompose:1 anonymous:1 marcinkiewicz:1 multilinear:1 summation:1 extension:2 hall:1 predict:2 lm:1 al1:1 consecutive:1 purpose:1 estimation:1 label:126 combinatorial:2 minimization:10 clearly:1 always:1 varying:1 shtml:1 jaakkola:2 conjunction:1 encode:2 emission:6 focus:1 joachim:1 notational:1 improvement:4 rank:10 consistently:1 ubler:1 contrast:1 baseline:15 kim:1 inference:9 dependent:4 suffix:1 nn:2 ferguson:1 typically:2 entire:1 a0:2 hidden:1 comprising:1 selects:1 semantics:1 arg:1 classification:23 ill:1 among:2 denoted:1 uncovering:2 development:1 noun:6 special:1 initialize:2 art:1 field:4 construct:1 evgeniou:2 ng:2 identical:1 represents:3 unsupervised:2 nearly:1 connectionist:3 report:3 others:1 eighty:1 randomly:1 composed:2 individual:2 replaced:1 consisting:1 evaluation:1 semidefinite:1 closer:1 encourage:1 allocates:1 minw:3 tree:2 plateaued:1 penalizes:1 circle:1 initialized:1 instance:1 column:2 modeling:4 assignment:1 stacking:1 mac:2 introducing:1 entry:1 holographic:1 predicate:1 too:2 reported:1 dependency:4 proximal:3 nns:2 density:1 explores:1 international:4 siam:1 contract:1 off:1 lee:1 probabilistic:1 corenlp:1 quickly:1 reflect:1 cesa:1 emnlp:1 cognitive:1 american:2 leading:1 return:1 yp:18 ullman:1 account:2 huval:1 north:1 explicitly:4 collobert:1 performed:2 view:1 doing:2 portion:3 recover:1 bayes:1 minimize:2 air:1 square:1 accuracy:9 formed:1 characteristic:2 efficiently:1 correspond:2 conceptually:1 generalize:2 vincent:1 kavukcuoglu:1 comp:4 converged:2 sharing:1 definition:11 against:1 petrov:1 dm:1 naturally:1 associated:18 hamming:1 gain:3 dataset:3 popular:1 yp1:2 knowledge:1 recall:2 dimensionality:8 improves:1 segmentation:1 afrl:2 higher:3 nivre:1 follow:1 formulation:4 done:2 though:5 evaluated:1 furthermore:1 just:3 implicit:1 correlation:1 christopher:1 o:2 defines:1 behaved:1 lei:1 semisupervised:1 utah:2 building:1 effect:1 y2:3 true:1 ize:2 hence:1 assigned:2 regularization:1 alternating:7 laboratory:1 misc:2 semantic:3 vivek:1 during:1 encourages:1 mina:1 plate:1 crf:1 demonstrate:1 mcdonald:2 l1:2 meaning:8 image:2 regularizing:1 parikh:1 common:1 specialized:1 al2:1 overview:1 volume:1 extend:2 discussed:1 association:3 katz:1 refer:2 composition:3 vec:8 session:1 language:11 gratefully:1 dot:1 similarity:10 longer:1 add:1 base:1 recent:6 perspective:2 driven:1 discard:1 zaniboni:1 success:1 arbitrarily:1 meeting:2 yi:10 exploited:1 scoring:3 seen:2 matrixvector:1 greater:1 additional:1 impose:1 gentile:1 novelty:1 corrado:1 signal:1 multiple:1 mix:1 full:1 karlen:1 faster:1 match:1 cross:1 long:1 bach:1 post:2 ellipsis:2 prediction:13 desideratum:1 variant:1 vision:2 arxiv:5 iteration:1 represent:5 addition:2 fine:1 singular:2 rest:2 unlike:1 comment:1 subject:1 member:1 lafferty:1 effectiveness:1 structural:11 split:1 easy:1 bengio:2 newsgroups:4 affect:1 identified:1 idea:2 multiclass:15 orthographically:1 motivated:1 defense:1 penalty:1 speech:14 cause:1 constitute:1 jj:1 compositional:20 deep:2 enumerate:1 generally:1 clear:1 hardware:2 svms:6 reduced:2 conjoining:2 http:1 outperform:1 exist:2 macintyre:1 coates:1 klein:1 discrete:7 affected:1 express:1 group:1 key:1 terminology:1 graph:3 subgradient:1 year:3 mswindows:2 sum:2 convert:1 run:2 ratinov:1 extends:1 almost:4 decision:1 summarizes:1 conll:4 layer:1 courville:1 annual:3 strength:1 riedel:1 software:1 tag:5 aspect:1 speed:1 argument:3 mikolov:1 structured:27 according:2 combination:5 manning:4 across:7 smaller:3 y0:5 partitioned:1 nilsson:1 maxy:1 outlier:1 restricted:1 computationally:1 equation:6 singer:1 letting:1 end:3 available:1 operation:1 generalizes:1 hierarchical:1 denotes:2 running:2 nlp:7 include:1 top:3 linguistics:3 hinge:3 exploit:1 brant:1 build:1 amit:1 society:1 tensor:27 objective:18 strategy:1 primary:1 usual:4 traditional:2 surrogate:1 gradient:3 distance:1 separate:5 thank:1 sci:2 w0:1 evenly:1 considers:1 reason:1 boldface:1 marcus:1 length:1 modeled:2 relationship:2 minn:1 illustration:1 minimizing:2 equivalently:1 taxonomy:1 design:1 implementation:1 motivates:1 proper:1 unknown:1 allowing:3 bianchi:1 observation:2 markov:4 finite:1 descent:2 defining:1 extended:1 hinton:1 y1:4 tnt:2 arbitrary:4 lb:1 community:1 aly:3 inferred:1 compositionality:5 pair:2 sentence:1 connection:1 framing:1 address:2 able:1 below:1 pattern:1 smolensky:1 sparsity:1 program:1 adjective:3 max:1 explanation:1 hot:7 difficulty:1 treated:3 natural:8 force:2 indicator:4 regularized:2 recursion:1 advanced:1 yp0:1 representing:3 hindi:1 technology:2 brief:1 realvalued:1 acknowledges:1 auto:1 text:3 epoch:3 literature:1 l2:2 vectorizing:1 review:2 interdependent:1 loss:7 generation:1 filtering:1 srebro:2 validation:1 foundation:1 degree:1 treebank:4 classifying:1 share:4 token:1 surprisingly:1 english:7 allow:5 perceptron:1 taking:1 sparse:1 distributed:10 regard:1 superficially:1 transition:11 rich:4 author:2 made:1 commonly:1 collection:2 transaction:2 relatedness:2 xi:8 discriminative:1 vectorization:2 decomposes:3 table:9 learn:3 basque:10 interact:1 bottou:1 complex:4 necessarily:1 da:1 did:3 dense:3 whole:1 turian:1 plural:1 allowed:3 benefited:1 fashion:1 sub:3 pereira:1 conjoins:1 posting:1 learns:1 grained:1 remained:1 bad:1 specific:3 symbol:1 explored:1 list:5 svm:9 experimented:1 intractable:1 workshop:1 socher:1 toutanova:1 effectively:3 occurring:1 margin:2 sparser:1 chen:1 led:3 simply:1 expressed:1 ordered:1 newsweeder:1 recommendation:1 applies:1 binding:1 nnps:1 determines:1 extracted:1 weston:1 conditional:4 goal:6 eventual:1 shared:9 replace:1 change:2 reducing:1 semantically:7 wt:11 called:2 xin:1 newsgroup:6 internal:2 support:1 collins:1 evaluate:2 tagset:1 argyriou:1 scratch:1 handling:1 |
4,776 | 5,324 | Predictive Entropy Search for Efficient Global
Optimization of Black-box Functions
Jos?e Miguel Hern?andez-Lobato
[email protected]
University of Cambridge
Matthew W. Hoffman
[email protected]
University of Cambridge
Zoubin Ghahramani
[email protected]
University of Cambridge
Abstract
We propose a novel information-theoretic approach for Bayesian optimization
called Predictive Entropy Search (PES). At each iteration, PES selects the next
evaluation point that maximizes the expected information gained with respect to
the global maximum. PES codifies this intractable acquisition function in terms
of the expected reduction in the differential entropy of the predictive distribution. This reformulation allows PES to obtain approximations that are both more
accurate and efficient than other alternatives such as Entropy Search (ES). Furthermore, PES can easily perform a fully Bayesian treatment of the model hyperparameters while ES cannot. We evaluate PES in both synthetic and realworld applications, including optimization problems in machine learning, finance,
biotechnology, and robotics. We show that the increased accuracy of PES leads to
significant gains in optimization performance.
1
Introduction
Bayesian optimization techniques form a successful approach for optimizing black-box functions
[5]. The goal of these methods is to find the global maximizer of a nonlinear and generally nonconvex function f whose derivatives are unavailable. Furthermore, the evaluations of f are usually
corrupted by noise and the process that queries f can be computationally or economically very
expensive. To address these challenges, Bayesian optimization devotes additional effort to modeling the unknown function f and its behavior. These additional computations aim to minimize the
number of evaluations that are needed to find the global optima.
Optimization problems are widespread in science and engineering and as a result so are Bayesian
approaches to this problem. Bayesian optimization has successfully been used in robotics to adjust
the parameters of a robot?s controller to maximize gait speed and smoothness [16] as well as parameter tuning for computer graphics [6]. Another example application in drug discovery is to find the
chemical derivative of a particular molecule that best treats a given disease [20]. Finally, Bayesian
optimization can also be used to find optimal hyper-parameter values for statistical [29] and machine
learning techniques [24].
As described above, we are interested in finding the global maximizer x? = arg maxx?X f (x) of
a function f over some bounded domain, typically X ? Rd . We assume that f (x) can only be
evaluated via queries to a black-box that provides noisy outputs of the form yi ? N (f (xi ), ? 2 ).
We note, however, that our framework can be extended to other non-Gaussian likelihoods. In this
setting, we describe a sequential search algorithm that, after n iterations, proposes to evaluate f at
some location xn+1 . To make this decision the algorithm conditions on all previous observations
eN
Dn = {(x1 , y1 ), . . . , (xn , yn )}. After N iterations the algorithm makes a final recommendation x
for the global maximizer of the latent function f .
We take a Bayesian approach to the problem described above and use a probabilistic model for the
eN . In this work we use a zero-mean Gaussian
latent function f to guide the search and to select x
1
Algorithm 2 PES acquisition function
Input: a black-box with unknown mean f
1: for n = 1, . . . , N do
2:
select xn = arg maxx?X ?n?1 (x)
3:
query the black-box at xn to obtain yn
4:
augment data Dn = Dn?1 ? {(xn , yn )}
5: end for
eN = arg maxx?X ?N (x)
6: return x
Input: a candidate x; data Dn
1: sample M hyperparameter values {? (i) }
2: for i = 1, . . . , M do
3:
sample f (i) ? p(f |Dn , ?, ? (i) )
(i)
4:
set x? ? arg maxx?X f (i) (x)
(i)
(i)
e (i) , v
e (i)
5:
compute m0 , V0 and m
(i)
(i)
(i)
6:
compute vn (x) and vn (x|x? )
7: end for
8: return ?n (x) as in (10)
precomputed
Algorithm 1 Generic Bayesian optimization
process (GP) prior for f [22]. This prior is specified by a positive-definite kernel function k(x, x0 ).
Given any finite collection of points {x1 , . . . , xn }, the values of f at these points are jointly zeromean Gaussian with covariance matrix Kn , where [Kn ]ij = k(xi , xj ). For the Gaussian likelihood
described above, the vector of concatenated observations yn is also jointly Gaussian with zero-mean.
Therefore, at any location x, the latent function f (x) conditioned on past observations Dn is then
Gaussian with marginal mean ?n (x) and variance vn (x) given by
?n (x) = kn (x)T (Kn + ? 2 I)?1 yn ,
vn (x) = k(x, x) ? kn (x)T (Kn + ? 2 I)?1 kn (x) ,
(1)
where kn (x) is a vector of cross-covariance terms between x and {x1 , . . . , xn }.
Bayesian optimization techniques use the above predictive distribution p(f (x)|Dn ) to guide the
search for the global maximizer x? . In particular, p(f (x)|Dn ) is used during the computation of an
acquisition function ?n (x) that is optimized at each iteration to determine the next evaluation location xn+1 . This process is shown in Algorithm 1. Intuitively, the acquisition function ?n (x) should
be high in areas where the maxima is most likely to lie given the current data. However, ?n (x)
eN is a
should also encourage exploration of the search space to guarantee that the recommendation x
global optimum of f , not just a global optimum of the posterior mean. Several acquisition functions
have been proposed in the literature. Some examples are the probability of improvement (PI) [14],
the expected improvement (EI) [19, 13] or upper confidence bounds (UCB) [26]. Alternatively, one
can combine several of these acquisition functions [10].
The acquisition functions described above are based on probabilistic measures of improvement (PI
an EI) or on optimistic estimates of the latent function (UCB) which implicitly trade off between
exploiting the posterior mean and exploring based on the uncertainty. An alternate approach, introduced by [28], proposes maximizing the expected posterior information gain about the global
maximizer x? evaluated over a grid in the input space. A similar strategy was later employed by [9]
which although it requires no such grid, instead relies on a difficult-to-evaluate approximation. In
Section 2 we derive a rearrangement of this information-based acquisition function which leads to a
more straightforward approximation that we call Predictive Entropy Search (PES). In Section 3 we
show empirically that our approximation is more accurate than that of [9]. We evaluate this claim on
both synthetic and real-world problems and further show that this leads to real gains in performance.
2
Predictive entropy search
We propose to follow the information-theoretic method for active data collection described in [17].
We are interested in maximizing information about the location x? of the global maximum, whose
posterior distribution is p(x? |Dn ). Our current information about x? can be measured in terms
of the negative differential entropy of p(x? |Dn ). Therefore, our strategy is to select xn+1 which
maximizes the expected reduction in this quantity. The corresponding acquisition function is
?n (x) = H[p(x? |Dn )] ? Ep(y|Dn ,x) [H[p(x? |Dn ? {(x, y)})]] ,
(2)
R
where H[p(x)] = ? p(x) log p(x)dx represents the differential entropy of its argument and the
expectation above is taken with respect to the posterior predictive distribution of y given x. The
exact evaluation of (2) is infeasible in practice. The main difficulties are i) p(x? |Dn ? {(x, y)})
must be computed for many different values of x and y during the optimization of (2) and ii) the
entropy computations themselves are not analytical. In practice, a direct evaluation of (2) is only
2
possible after performing many approximations [9]. To avoid this, we follow the approach described
in [11] by noting that (2) can be equivalently written as the mutual information between x? and y
given Dn . Since the mutual information is a symmetric function, ?n (x) can be rewritten as
?n (x) = H[p(y|Dn , x)] ? Ep(x? |Dn ) [H[p(y|Dn , x, x? )]] ,
(3)
where p(y|Dn , x, x? ) is the posterior predictive distribution for y given the observed data Dn and
the location of the global maximizer of f . Intuitively, conditioning on the location x? pushes the
posterior predictions up in locations around x? and down in regions away from x? . Note that, unlike
the previous formulation, this objective is based on the entropies of predictive distributions, which
are analytic or can be easily approximated, rather than on the entropies of distributions on x? whose
approximation is more challenging.
The first term in (3) can be computed analytically using the posterior marginals for f (x) in (1), that
is, H[p(y|Dn , x)] = 0.5 log[2?e (vn (x) + ? 2 )], where we add ? 2 to vn (x) because y is obtained
by adding Gaussian noise with variance ? 2 to f (x). The second term, on the other hand, must be
(i)
approximated. We first approximate the expectation in (3) by averaging over samples x? drawn
approximately from p(x? |Dn ). For each of these samples, we then approximate the corresponding
(i)
entropy function H[p(y|Dn , x, x? )] using expectation propagation [18]. The code for all these
operations is publicly available at http://jmhl.org.
2.1
Sampling from the posterior over global maxima
In this section we show how to approximately sample from the conditional distribution of the global
maximizer x? given the observed data Dn , that is,
p(x? |Dn ) = p f (x? ) = max f (x)Dn .
(4)
x?X
If the domain X is restricted to some finite set of m points, the latent function f takes the form
of an m-dimensional
vector f . The probability that the ith element of f is optimal can then be
R
Q
written as p(f |Dn ) j?m I[fi ? fj ] df . This suggests the following generative process: i) draw
a sample from the posterior distribution p(f |Dn ) and ii) return the index of the maximum element
in the sampled vector. This process is known as Thompson sampling or probability matching when
used as an arm-selection strategy in multi-armed bandits [8]. This same approach could be used for
sampling the maximizer over a continuous domain X . At first glance this would require constructing
an infinite-dimensional object representing the function f . To avoid this, one could sequentially
construct f while it is being optimized. However, evaluating such an f would ultimately have cost
O(m3 ) where m is the number of function evaluations necessary to find the optimum. Instead,
we propose to sample and optimize an analytic approximation to f . We will briefly derive this
approximation below, but more detail is given in Appendix A.
Given a shift-invariant kernel k, Bochner?s theorem [4] asserts the existence of its Fourier dual s(w),
which is equal to the spectral density of k. Letting p(w) = s(w)/? be the associated normalized
density, we can write the kernel as the expectation
T
0
k(x, x0 ) = ? Ep(w) [e?iw (x?x ) ] = 2? Ep(w,b) [cos(wT x + b) cos(wT x0 + b)] ,
(5)
p
where b ? U[0, 2?]. Let ?(x) = 2?/m cos(Wx + b) denote an m-dimensional feature mapping
where W and b consist of m stacked samples from p(w, b). The kernel k can then be approximated
by the inner product of these features, k(x, x0 ) ? ?(x)T ?(x0 ). This approach was used by [21]
as an approximation method in the context of kernel methods. The feature mapping ?(x) allows
us to approximate the Gaussian process prior for f with a linear model f (x) = ?(x)T ? where
? ? N (0, I) is a standard Gaussian. By conditioning on Dn , the posterior for ? is also multivariate
Gaussian, ?|Dn ? N (A?1 ?T yn , ? 2 A?1 ) where A = ?T ? + ? 2 I and ?T = [?(x1 ) . . . ?(xn )].
Let ?(i) and ? (i) be a random set of features and the corresponding posterior weights sampled both
according to the generative process given above. They can then be used to construct the function
f (i) (x) = ?(i) (x)T ? (i) , which is an approximate posterior sample of f ?albeit one with a finite
(i)
parameterization. We can then maximize this function to obtain x? = arg maxx?X f (i) (x), which
is approximately distributed according to p(x? |Dn ). Note that for early iterations when n < m, we
can efficiently sample ? (i) with cost O(n2 m) using the method described in Appendix B.2 of [23].
This allows us to use a large number of features in ?(i) (x).
3
2.2
Approximating the predictive entropy
We now show how to approximate H[p(y|D
R n , x, x? )] in (3). Note that we can write the argument to
H in this expression as p(y|Dn , x, x? ) = p(y|f (x))p(f (x)|Dn , x? ) df (x). Here p(f (x)|Dn , x? )
is the posterior distribution on f (x) given Dn and the location x? of the global maximizer of f .
When the likelihood p(y|f (x)) is Gaussian, we have that p(f (x)|Dn ) is analytically tractable since it
is the predictive distribution of a Gaussian process. However, by further conditioning on the location
x? of the global maximizer we are introducing additional constraints, namely that f (z) ? f (x? )
for all z ? X . These constraints make p(f (x)|Dn , x? ) intractable. To circumvent this difficulty, we
instead use the following simplified constraints:
C1. x? is a local maximum. This is achieved by letting ?f (x? ) = 0 and ensuring that
?2 f (x? ) is negative definite. We further assume that the non-diagonal elements of
?2 f (x? ), denoted by upper[?2 f (x? )], are known, for example they could all be zero.
This simplifies the negative-definite constraint. We denote by C1.1 the constraint given by
?f (x? ) = 0 and upper[?2 f (x? )] = 0. We denote by C1.2 the constraint that forces the
elements of diag[?2 f (x? )] to be negative.
C2. f (x? ) is larger than past observations. We also assume that f (x? ) ? f (xi ) for all
i ? n. However, we only observe f (xi ) noisily via yi . To avoid making inference on these
latent function values, we approximate the above hard constraints with the soft constraint
f (x? ) > ymax + , where ? N (0, ? 2 ) and ymax is the largest yi seen so far.
C3. f (x) is smaller than f (x? ). This simplified constraint only conditions on the given x
rather than requiring f (x? ) ? f (z) for all z ? X .
We incorporate these simplified constraints into p(f (x)|Dn ) to approximate p(f (x)|Dn , x? ). This
is achieved by multiplying p(f (x)|Dn ) with specific factors that encode the above constraints. In
what follows we briefly show how to construct these factors; more detail is given in Appendix B.
Consider the latent variable z = [f (x? ); diag[?2 f (x? )]]. To incorporate constraint C1.1 we
can condition on the data and on the ?observations? given by the constraints ?f (x? ) = 0 and
upper[?2 f (x? )] = 0. Since f is distributed according to a GP, the joint distribution between z and
these observations is multivariate Gaussian. The covariance between the noisy observations yn and
the extra noise-free derivative observations can be easily computed [25]. The resulting conditional
distribution is also multivariate Gaussian with mean m0 and covariance V0 . These computations
are similar to those performed in (1). Constraints C1.2 and C2 can then be incorporated by writing
hQ
i
d
2
p(z|Dn , C1, C2) ? ??2 (f (x? ) ? ymax )
(6)
i=1 I [? f (x? )]ii ? 0 N (z|m0 , V0 ) ,
where ??2 is the cdf of a zero-mean Gaussian distribution with variance ? 2 . The first new factor in
this expression guarantees that f (x? ) > ymax + , where we have marginalized out, and the second
set of factors guarantees that the entries in diag[?2 f (x? )] are negative.
Later integrals that make use of p(z|Dn , C1, C2), however, will not admit a closed-form expression. As a result we compute a Gaussian approximation q(z) to this distribution using Expectation
Propagation (EP) [18]. The resulting algorithm is similar to the implementation of EP for binary
classification with Gaussian processes [22]. EP approximates each non-Gaussian factor in (6) with
a Gaussian factor whose mean and variance are m
e i and vei , respectively. The EP approximation can
Qd+1
e i , vei )]N (z|m0 , V0 ). Note that these computations have so
then be written as q(z) ? [ i=1 N (zi |m
e v
e } once and store them for later use, where
far not depended on x, so we can compute {m0 , V0 , m,
e = (m
e = (?
m
? 1, . . . , m
? d+1 ) and v
v1 , . . . , v?d+1 ).
We will now describe how to compute the predictive variance of some latent function value f (x)
given these constraints. Let f = [f (x); f (x? )] be a vector given by the concatenation of the values
of the latent function at x and x? . The joint distribution between f , z, the evaluations yn collected
so far and the derivative ?observations? ?f (x? ) = 0 and upper[?2 f (x? )] = 0 is multivariate
Gaussian. Using q(z), we then obtain the following approximation:
R
p(f |Dn , C1, C2) ? p(f |z, Dn , C1.1) q(z) dz = N (f |mf , Vf ) .
(7)
Implicitly we are assuming above that f depends on our observations and constraint C1.1, but is
independent of C1.2 and C2 given z. The computations necessary to obtain mf and Vf are similar
4
to those used above and in (1). The required quantities are similar to the ones used by EP to make
predictions in the Gaussian process binary classifier [22]. We can then incorporate C3 by multiplying
N (f |mf , Vf ) with a factor that guarantees f (x) < f (x? ). The predictive distribution for f (x) given
Dn and all the constraints can be approximated as
R
p(f (x)|Dn , C1, C2, C3) ? Z ?1 I(f1 < f2 ) N (f |mf , Vf ) df2 ,
(8)
where Z is a normalization constant. The variance of the right hand size of (8) is given by
vn (x|x? ) = [Vf ]1,1 ? v ?1 ?(? + ?){[Vf ]1,1 ? [Vf ]1,2 }2 ,
(9)
?
T
T
where v = [?1, 1] Vf [?1, 1], ? = m/ v, m = [?1, 1] mf , ? = ?(?)/?(?), and ?(?) and
?(?) are the standard Gaussian density function and cdf, respectively. By further approximating (8) by a Gaussian distribution with the same mean and variance we can write the entropy as
H[p(y|Dn , x, x? )] ? 0.5 log[2?e(vn (x|x? ) + ? 2 )].
The computation of (9) can be numerically unstable when s is very close to zero. This occurs when
[Vf ]1,1 is very similar to [Vf ]1,2 . To avoid these numerical problems, we multiply [Vf ]1,2 by the
largest 0 ? ? ? 1 that guarantees that s > 10?10 . This can be understood as slightly reducing
the amount of dependence between f (x) and f (x? ) when x is very close to x? . Finally, fixing
upper[?2 f (x? )] to be zero can also produce poor predictions when the actual f does not satisfy
this constraint. To avoid this, we instead fix this quantity to upper[?2 f (i) (x? )], where f (i) is the
(i)
ith sample function optimized in Section 2.1 to sample x? .
2.3
Hyperparameter learning and the PES acquisition function
We now show how the previous approximations are integrated to compute the acquisition function
used by predictive entropy search (PES). This acquisition function performs a formal treatment of the
hyperparameters. Let ? denote a vector of hyperparameters which includes any kernel parameters
as well as the noise variance ? 2 . Let p(?|Dn ) ? p(?) p(Dn |?) denote the posterior distribution
over these parameters where p(?) is a hyperprior and p(Dn |?) is the GP marginal likelihood. For a
fully Bayesian treatment of ? we must marginalize the acquisition function (3) with respect to this
posterior. The corresponding integral has no analytic expression and must be approximated using
Monte Carlo. This approach is also taken in [24].
(i)
We draw M samples {? (i) } from p(?|Dn ) using slice sampling [27]. Let x? denote a sampled
(i)
global maximizer drawn from p(x? |Dn , ? (i) ) as described in Section 2.1. Furthermore, let vn (x)
(i)
(i)
and vn (x|x? ) denote the predictive variances computed as described in Section 2.2 when the
model hyperparameters are fixed to ? (i) . We then write the marginalized acquisition function as
o
PM n
(i)
(i)
(i)
1
2
2
?n (x) = M
(10)
i=1 0.5 log[vn (x) + ? ] ? 0.5 log[vn (x|x? ) + ? ] .
Note that PES is effectively marginalizing the original acquisition function (2) over p(?|Dn ). This
is a significant advantage with respect to other methods that optimize the same information-theoretic
acquisition function but do not marginalize over the hyper-parameters. For example, the approach
of [9] approximates (2) only for fixed ?. The resulting approximation is computationally very expensive and recomputing it to average over multiple samples from p(?|Dn ) is infeasible in practice.
Algorithm 2 shows pseudo-code for computing the PES acquisition function. Note that most of the
computations necessary for evaluating (10) can be done independently of the input x, as noted in the
pseudo-code. This initial cost is dominated by a matrix inversion necessary to pre-compute V for
each hyperparameter sample. The resulting complexity is O[M (n+d+d(d?1)/2)3 ]. This cost can
be reduced to O[M (n + d)3 ] by ignoring the derivative observations imposed on upper[?2 f (x? )]
by constraint C1.1. Nevertheless, in the problems that we consider d is very small (less than 20).
After these precomputations are done, the evaluation of (10) is O[M (n + d + d(d ? 1)/2)].
3
Experiments
In our experiments, wePuse Gaussian process priors for f with squared-exponential kernels
k(x, x0 ) = ? 2 exp{?0.5 i (xi ? x0i )2 /`2i }. The corresponding spectral density is zero-mean Gaus2
sian with covariance given by diag([`?2
i ]) and normalizing constant ? = ? . The model hyperpa2
rameters are {?, `1 , . . . , `d , ? }. We use broad, uninformative Gamma hyperpriors.
5
1
0.0
0.35
x
x
0.30
0.06
0.06
0.25
x
x
x
x 0.0
0.25
x
0.05
0.25
0.05
0.20
0
6
x
x
0.2
0.2
x
0.05
3
0.0
x
5
0.0
2
0.0
0.2
x
x
0.07
0.04
0.03
0.1
x
0.02
0.25
0.0
5
0.01
0.2
x
0.2
5
0.25
0.2
0.0
2
x
x
x
x
0.15
0.03
x
x
0.25
0.01
x
0.1
x
0.25
0.04
0.30
0.0
4
0.3
0.35
0.02
0.03
0.15
x
0.1
0.10
x
0.2
5
0.05
x
0.03
x
0.01
x
0.06
0.03
0.02
0.04
0.25
x
0.05
0.05
0.05
0.04
0.20
x
0.01
0.03
0.00
0.00
Figure 1: Comparison of different estimates of the objective function ?n (x) given by (2). Left, ground truth
obtained by the rejection sampling method RS. Middle, approximation produced by the ES method. Right,
approximation produced by the proposed PES method. These plots show that the PES objective is much more
similar to the RS ground truth than the ES objective.
First, we analyze the accuracy of PES in the task of approximating the differential entropy (2).
We compare the PES approximation (10), with the approximation used by the entropy search (ES)
method [9]. We also compare with the ground truth for (2) obtained using a rejection sampling (RS)
algorithm based on (3). For this experiment we generate the data Dn using an objective function f
sampled from the Gaussian process prior as in [9]. The domain X of f is fixed to be [0, 1]2 and data
are generated using ? 2 = 1, ? 2 = 10?6 , and `2i = 0.1. To compute (10) we avoid sampling the
hyperparameters and use the known values directly. We further fix M = 200 and m = 1000.
The ground truth rejection sampling scheme works as follows. First, X is discretized using a uniform
grid. The expectation with respect to p(x? |Dn ) in (3) is then approximated using sampling. For this,
we sample x? by evaluating a random sample from p(f |Dn ) on each grid cell and then selecting the
cell with highest value. Given x? , we then approximate H[p(y|Dn , x, x? )] by rejection sampling.
We draw samples from p(f |Dn ) and reject those whose corresponding grid cell with highest value is
not x? . Finally, we approximate H[p(y|Dn , x, x? )] by first, adding zero-mean Gaussian noise with
variance ? 2 to the the evaluations at x of the functions not rejected during the previous step and
second, we estimate the differential entropy of the resulting samples using kernels [1].
Figure 1 shows the objective functions produced by RS, ES and PES for a particular Dn with 10
measurements whose locations are selected uniformly at random in [0, 1]2 . The locations of the
collected measurements are displayed with an ?x? in the plots. The particular objective function
used to generate the measurements in Dn is displayed in the left part of Figure 2. The plots in
Figure 1 show that the PES approximation to (2) is more similar to the ground truth given by RS
than the approximation produced by ES. In this figure we also see a discrepancy between RS and
PES at locations near x = (0.572, 0.687). This difference is an artifact of the discretization used in
RS. By zooming in and drawing many more samples we would see the same behavior in both plots.
We now evaluate the performance of PES in the task of finding the optimum of synthetic black-box
objective functions. For this, we reproduce the within-model comparison experiment described in
[9]. In this experiment we optimize objective functions defined in the 2-dimensional unit domain
X = [0, 1]2 . Each objective function is generated by first sampling 1024 function values from the
GP prior assumed by PES, using the same ? 2 , `i and ? 2 as in the previous experiment. The objective
function is then given by the resulting GP posterior mean. We generated a total of 1000 objective
functions by following this procedure. The left plot in Figure 2 shows an example function.
In these experiments we compared the performance of PES with that of ES [9] and expected improvement (EI) [13], a widely used acquisition function in the Bayesian optimization literature. We
again assume that the optimal hyper-parameter values are known to all methods. Predictive performance is then measured in terms of the immediate regret (IR) |f (e
xn ) ? f (x? )|, where x? is the
en is the recommendation of each algorithm had we
known location of the global maximum and x
stopped at step n?for all methods this is given by the maximizer of the posterior mean. The right
plot in Figure 2 shows the decimal logarithm of the median of the IR obtained by each method
across the 1000 different objective functions. Confidence bands equal to one standard deviation are
obtained using the bootstrap method. Note that while averaging these results is also interesting,
corresponding to the expected performance averaged over the prior, here we report the median IR
6
1.5
0
0.5
2
? ?
?
1.5
2
?0.5
0.5
1
x
1
x
x
Log10 Median IR
x
Results on Synthetic Cost Functions
1
0.5
0.5
?1
1
x
?0.5
?1
0
?0.5
x
?2
0
x
x
?0.5
?1
?1
0
?1
?1.5
0.5
1
? ?
?
?
?
?
?
?
?
?
?
?3.5
?
?
Methods
?
?
EI
?
?
?
?
?
? ?
?
? ?
? ?
?
?
?
1
?
0.5
0
0
?
?
PES
?
?5.5
?2
0
0
1
0
ES
?
? ?
?
?
?0.5
0
?
?
?2.5
1.5
?0.5
x
?0
.5
?1
?
? ?
?4.5
0.5
x
?1.5
? ?
? ? ?
? ? ?
?
?
?0.5
?1.5
?
?
?
? ?
? ?
? ?
? ?
?
?
?
? ? ?
? ? ?
? ? ?
? ? ?
? ? ?
? ? ? ?
? ?
? ? ? ?
? ? ?
?
?
? ? ?
? ?
? ?
?
?
?
?
?
?
?
?
? ?
? ?
?
? ?
? ?
?
? ? ?
? ?
?
? ?
? ?
? ?
? ? ? ?
10
20
30
40
Number of Function Evaluations
50
Figure 2: Left, example of objective functions f . Right, median of the immediate regret (IR) for the methods
PES, ES and EI in the experiments with synthetic objective functions.
Results on Cosines Cost Function
Results on Branin Cost Function
?
?
Results on Hartmann Cost Function
? ? ?
?
?
? ?
?
? ?
?
? ? ? ?
?
?
?
? ?
?
?
? ?
?
? ?
? ?
? ?
?
? ? ? ?
?
?
?
?
?
?
? ?
?
?
?
?
?
?
?
? ?
?
?
? ?
?
?
?
?
?
?
?
?
? ?
?
? ?
?
?
?
?
?
?
?
? ?
?
?
?
?
?
?
?
?
? ?
?
? ?
?
? ? ?
?
?
?
? ?
?
?
?
? ? ? ? ?
?
? ?
? ? ?
?
?
?
?
?
? ?
?
?
?
?
?
? ?
?
?
? ? ?
? ? ?
? ? ?
? ? ? ?
?
?
?
?
?
?
?0.6
?
?
?
Log10 Median IR
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?1.9
?
?
?
?
?
?
?
?
?
?1.6
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?2.6
?
?
?
?
?
?
?
?
?
?
?
?
PES?NB
?
?
?3.6
?
?
?
20
?1.7
? ?
?
?
? ?
? ? ?
?
?
? ?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?3.9
?
?
?
?
?
10
?
?
?
?
?
PES
?
? ?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
ES
?
?
?
?
EI
?
?
?
?
?0.7
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
0
?
?
?
?
?
?
?
?
?
Methods
?
?
?
?
?
?
?2.9
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?0.9
?
?
?
?
?
30
Number of Function Evaluations
?
? ? ?
?
?
?4.6
?
0
10
20
30
Number of Function Evaluations
?2.7
0
10
20
30
? ? ?
? ?
? ? ?
? ? ?
?
40
50
Number of Function Evaluations
Figure 3: Median of the immediate regret (IR) for the methods EI, ES, PES and PES-NB in the experiments
with well-known synthetic benchmark functions.
because the empirical distribution of IR values is very heavy-tailed. In this case, the median is more
representative of the exact location of the bulk of the data. These results indicate that the best method
in this setting is PES, which significantly outperforms ES and EI. The plot also shows that in this
case ES is significantly better than EI.
We perform another series of experiments in which we optimize well-known synthetic benchmark
functions including a mixture of cosines [2] and Branin-Hoo (both functions defined in [0, 1]2 ) as
well as the Hartmann-6 (defined in [0, 1]6 ) [15]. In all instances, we fix the measurement noise
to ? 2 = 10?3 . For both PES and EI we marginalize the hyperparameters ? using the approach
described in Section 2.3. ES, by contrast, cannot average its approximation of (2) over the posterior
on ?. Instead, ES works by fixing ? to an estimate of its posterior mean (obtained using slice
sampling) [27]. To evaluate the gains produced by the fully Bayesian treatment of ? in PES, we also
compare with a version of PES (PES-NB) which performs the same non-Bayesian (NB) treatment
of ? as ES. In PES-NB we use a single fixed hyperparameter as in previous sections with value
given by the posterior mean of ?. All the methods are initialized with three random measurements
collected using latin hypercube sampling [5].
The plots in Figure 3 show the median IR obtained by each method on each function across 250
random initializations. Overall, PES is better than PES-NB and ES. Furthermore, PES-NB is also
significantly better than ES in most of the cases. These results show that the fully Bayesian treatment
of ? in PES is advantageous and that PES can produce better approximations than ES. Note that
PES performs better than EI in the Branin and cosines functions, while EI is significantly better on
the Hartmann problem. This appears to be due to the fact that entropy-based strategies explore more
aggressively which in higher-dimensional spaces takes more iterations. The Hartmann problem,
however, is a relatively simple problem and as a result the comparatively more greedy behavior of
EI does not result in significant adverse consequences. Note that the synthetic functions optimized
in the previous experiment were much more multimodal that the ones considered here.
3.1
Experiments with real-world functions
We finally optimize different real-world cost functions. The first one (NNet) returns the predictive
accuracy of a neural network on a random train/test partition of the Boston Housing dataset [3].
7
Hydrogen
NNet Cost
Portfolio
Methods
? ?
0.6
EI
Log10 Median IR
?
?
?
?0.9
?
ES
?
PES
? ?
?
?
?
?
? ? ? ?
?
?
?
?
PES?NB
?
? ?
?
?
?
?
?
?
? ?
?
?
? ?
?
? ? ? ?
? ? ? ?
? ?
? ? ? ? ?
? ?
?
?
? ?
? ?
?
?
?
?
?
? ?
?
?
?
?
? ?
?
?
?
?
?
?
?
?
? ?
? ? ? ?
?
? ?
? ?
? ?
? ?
?
?
?1.4
? ?
? ?
?
?
?
?
?
?
? ?
? ?
?
? ?
? ?
?
? ?
? ? ?
?
10
20
30
40
Function Evaluations
0
10
20
30
40
?
?
?
?
Function Evaluations
?
?
?
?
? ?
?
?
?
?
?
10
20
?
?
?
?
?
? ?
?
?
?
?
? ? ? ?
?1.9
?
?
? ? ? ? ? ?
?
?
?
?
?
?
? ?
?
? ? ?
? ? ?
? ?
? ? ?
?
? ?
? ? ?
?
?
? ?
? ? ?
? ?
? ?
? ?
? ?
? ?
?
?
? ?
? ? ?
?
? ? ?
?
?
?
?
?
? ?
? ?
? ?
? ? ? ? ? ? ? ?
0
?
?
? ?
?
?
? ?
?
? ? ? ? ?
?
? ? ?
?
? ? ? ?
?
? ? ?
?
?
? ?
? ?
? ?
?
?
? ? ?
? ? ? ?
? ? ?
? ?
? ?
? ? ?
? ? ? ?
? ?
?
? ?
? ?
?
?
?
? ? ? ? ? ?
? ? ?
?
?
?
?
? ?
? ? ? ? ? ?
?
? ? ?
?
?
? ?
?
? ? ? ?
?
?
?
?
?
?
?
?
?
?
?
? ? ?
? ?
? ? ?
? ?
?
? ?
?
?
? ?
?
? ?
? ? ?
?
?
?
? ?
?
?
?
?
?
? ?
?
? ?
?
? ?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
? ?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
? ?
? ?
? ?
?0.1
0
?
?0.9
?
?
?
?
?
?
?
?
? ?
?
? ?
? ?
?
?
?
?
? ?
? ?
?
?
?
?
?
?
? ?
? ?
?
?
?
?
?
? ?
?
?
?
?
?
?
?
? ?
?
? ?
?
?
? ?
?
?
? ? ?
?
?
?
?
? ?
?
?
? ?
?
?
?
?
?
?
?
? ? ?
? ?
?
?
?
? ?
?
?
? ?
?
?
?
?
?
?
?
?
? ?
?
?
?0.4
Walker B
? ?
? ? ? ?
? ? ? ?
? ?
? ?
? ?
? ?
?
? ?
? ? ?
?
? ? ? ? ? ?
?
?
?
?
? ? ?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
?
? ?
?
?
?
?
?
Walker A
? ? ?
?
30
40
Function Evaluations
?
?
?
?
?
?
?
?
?1.9
?
?
?0.3
0
10
20
30
Function Evaluations
0
10
20
30
Function Evaluations
Figure 4: Median of the immediate regret (IR) for the methods PES, PES-NB, ES and EI in the experiments
with non-analytic real-world cost functions.
The variables to optimize are the weight-decay parameter and the number of training iterations for
the neural network. The second function (Hydrogen) returns the amount of hydrogen production of
a particular bacteria in terms of the PH and Nitrogen levels of the growth medium [7]. The third
one (Portfolio) returns the ratio of the mean and the standard deviation (the Sharpe ratio) of the
1-year ahead returns generated by simulations from a multivariate time-series model that is adjusted
to the daily returns of stocks AXP, BA and HD. The time-series model is formed by univariate
GARCH models connected with a Student?s t copula [12]. These three functions (NNet, Hydrogen
and Portfolio) have as domain [0, 1]2 . Furthermore, in these examples, the ground truth function
that we want to optimize is unknown and is only available through noisy measurements. To obtain
a ground truth, we approximate each cost function as the predictive distribution of a GP that is
adjusted to data sampled from the original function (1000 uniform samples for NNet and Portfolio
and all the available data for Hydrogen [7]). Finally, we also consider another real-world function
that returns the walking speed of a bipedal robot [30]. This function is defined in [0, 1]8 and its
inputs are the parameters of the robot?s controller. In this case the ground truth function is noiseless
and can be exactly evaluated through expensive numerical simulation. We consider two versions of
this problem (Walker A) with zero-mean, additive noise of ? = 0.01 and (Walker B) with ? = 0.1.
Figure 4 shows the median IR values obtained by each method on each function across 250 random
initializations, except in Hydrogen where we used 500 due to its higher level of noise. Overall, PES,
ES and PES-NB perform similarly in NNet, Hydrogen and Portfolio. EI performs rather poorly in
these first three functions. This method seems to make excessively greedy decisions and fails to
explore the search space enough. This strategy seems to be advantageous in Walker A, where EI
obtains the best results. By contrast, PES, ES and PES-NB tend to explore more in this latter dataset.
This leads to worse results than those of EI. Nevertheless, PES is significantly better than PES-NB
and ES in both Walker datasets and better than EI in the noisier Walker B. In this case, the fully
Bayesian treatment of hyper-parameters performed by PES produces improvements in performance.
4
Conclusions
We have proposed a novel information-theoretic approach for Bayesian optimization. Our method,
predictive entropy search (PES), greedily maximizes the amount of one-step information on the location x? of the global maximum using its posterior differential entropy. Since this objective function
is intractable, PES approximates the original objective using a reparameterization that measures
entropy in the posterior predictive distribution of the function evaluations. PES produces more accurate approximations than Entropy Search (ES), a method based on the original, non-transformed
acquisition function. Furthermore, PES can easily marginalize its approximation with respect to
the posterior distribution of its hyper-parameters, while ES cannot. Experiments with synthetic and
real-world functions show that PES often outperforms ES in terms of immediate regret. In these experiments, we also observe that PES often produces better results than expected improvement (EI),
a popular heuristic for Bayesian optimization. EI often seems to make excessively greedy decisions, while PES tends to explore more. As a result, EI seems to perform better for simple objective
functions while often getting stuck with noisier objectives or for functions with many modes.
Acknowledgements J.M.H.L acknowledges support from the Rafael del Pino Foundation.
8
References
[1] I. Ahmad and P.-E. Lin. A nonparametric estimation of the entropy for absolutely continuous distributions.
IEEE Transactions on Information Theory, 22(3):372?375, 1976.
[2] B. S. Anderson, A. W. Moore, and D. Cohn. A nonparametric approach to noisy and costly optimization.
In ICML, pages 17?24, 2000.
[3] K. Bache and M. Lichman. UCI machine learning repository, 2013.
[4] S. Bochner. Lectures on Fourier integrals. Princeton University Press, 1959.
[5] E. Brochu, V. M. Cora, and N. de Freitas. A tutorial on Bayesian optimization of expensive cost functions,
with application to active user modeling and hierarchical reinforcement learning. Technical Report UBC
TR-2009-23 and arXiv:1012.2599v1, Dept. of Computer Science, University of British Columbia, 2009.
[6] E. Brochu, N. de Freitas, and A. Ghosh. Active preference learning with discrete choice data. In NIPS,
pages 409?416, 2007.
[7] E. H. Burrows, W.-K. Wong, X. Fern, F. W. R. Chaplen, and R. L. Ely. Optimization of ph and nitrogen
for enhanced hydrogen production by synechocystis sp. pcc 6803 via statistical and machine learning
methods. Biotechnology Progress, 25(4):1009?1017, 2009.
[8] O. Chapelle and L. Li. An empirical evaluation of Thompson sampling. In NIPS, pages 2249?2257, 2011.
[9] P. Hennig and C. J. Schuler. Entropy search for information-efficient global optimization. Journal of
Machine Learning Research, 13, 2012.
[10] M. W. Hoffman, E. Brochu, and N. de Freitas. Portfolio allocation for Bayesian optimization. In UAI,
pages 327?336, 2011.
[11] N. Houlsby, J. M. Hern?andez-Lobato, F. Huszar, and Z. Ghahramani. Collaborative Gaussian processes
for preference learning. In NIPS, pages 2096?2104, 2012.
[12] E. Jondeau and M. Rockinger. The copula-GARCH model of conditional dependencies: An international
stock market application. Journal of international money and finance, 25(5):827?853, 2006.
[13] D. R. Jones, M. Schonlau, and W. J. Welch. Efficient global optimization of expensive black-box functions. Journal of Global optimization, 13(4):455?492, 1998.
[14] H. Kushner. A new method of locating the maximum of an arbitrary multipeak curve in the presence of
noise. Journal of Basic Engineering, 86, 1964.
[15] D. Lizotte. Practical Bayesian Optimization. PhD thesis, University of Alberta, Canada, 2008.
[16] D. Lizotte, T. Wang, M. Bowling, and D. Schuurmans. Automatic gait optimization with Gaussian process
regression. In IJCAI, pages 944?949, 2007.
[17] D. J. MacKay. Information-based objective functions for active data selection. Neural Computation,
4(4):590?604, 1992.
[18] T. P. Minka. A family of algorithms for approximate Bayesian inference. PhD thesis, Massachusetts
Institute of Technology, 2001.
?
[19] J. Mo?ckus, V. Tiesis, and A. Zilinskas.
The application of Bayesian methods for seeking the extremum.
In L. Dixon and G. Szego, editors, Toward Global Optimization, volume 2. Elsevier, 1978.
[20] D. M. Negoescu, P. I. Frazier, and W. B. Powell. The knowledge-gradient algorithm for sequencing
experiments in drug discovery. INFORMS Journal on Computing, 23(3):346?363, 2011.
[21] A. Rahimi and B. Recht. Random features for large-scale kernel machines. In NIPS, pages 1177?1184,
2007.
[22] C. E. Rasmussen and C. K. Williams. Gaussian processes for machine learning. The MIT Press, 2006.
[23] M. W. Seeger. Bayesian inference and optimal design for the sparse linear model. Journal of Machine
Learning Research, 9:759?813, 2008.
[24] J. Snoek, H. Larochelle, and R. P. Adams. Practical Bayesian optimization of machine learning algorithms. In NIPS, pages 2960?2968, 2012.
[25] E. Solak, R. Murray-Smith, W. E. Leithead, D. J. Leith, and C. E. Rasmussen. Derivative observations in
Gaussian process models of dynamic systems. In NIPS, pages 1057?1064, 2003.
[26] N. Srinivas, A. Krause, S. M. Kakade, and M. Seeger. Gaussian process optimization in the bandit setting:
No regret and experimental design. In ICML, pages 1015?1022, 2010.
[27] J. Vanhatalo, J. Riihim?aki, J. Hartikainen, P. Jyl?anki, V. Tolvanen, and A. Vehtari. Bayesian modeling
with Gaussian processes using the matlab toolbox GPstuff (v3.3). CoRR, abs/1206.5754, 2012.
[28] J. Villemonteix, E. Vazquez, and E. Walter. An informational approach to the global optimization of
expensive-to-evaluate functions. Journal of Global Optimization, 44(4):509?534, 2009.
[29] Z. Wang, S. Mohamed, and N. de Freitas. Adaptive Hamiltonian and Riemann Monte Carlo samplers. In
ICML, 2013.
[30] E. Westervelt and J. Grizzle. Feedback Control of Dynamic Bipedal Robot Locomotion. Control and
Automation Series. CRC PressINC, 2007.
9
| 5324 |@word economically:1 middle:1 version:2 briefly:2 inversion:1 advantageous:2 seems:4 repository:1 pcc:1 zilinskas:1 r:7 simulation:2 vanhatalo:1 eng:1 covariance:5 tr:1 reduction:2 initial:1 series:4 lichman:1 selecting:1 past:2 outperforms:2 freitas:4 current:2 discretization:1 riihim:1 dx:1 must:4 written:3 numerical:2 partition:1 wx:1 additive:1 analytic:4 plot:8 generative:2 selected:1 greedy:3 parameterization:1 ith:2 smith:1 hamiltonian:1 provides:1 location:15 preference:2 org:1 branin:3 dn:62 c2:7 direct:1 differential:6 combine:1 x0:6 snoek:1 market:1 expected:8 behavior:3 themselves:1 multi:1 discretized:1 informational:1 riemann:1 alberta:1 actual:1 armed:1 bounded:1 maximizes:3 medium:1 what:1 finding:2 ghosh:1 extremum:1 guarantee:5 pseudo:2 growth:1 finance:2 exactly:1 classifier:1 uk:3 control:2 unit:1 anki:1 yn:8 positive:1 engineering:2 local:1 treat:1 understood:1 depended:1 consequence:1 tends:1 leithead:1 leith:1 approximately:3 black:7 initialization:2 precomputations:1 suggests:1 challenging:1 co:3 averaged:1 practical:2 practice:3 regret:6 definite:3 bootstrap:1 procedure:1 powell:1 area:1 empirical:2 drug:2 maxx:5 reject:1 significantly:5 matching:1 confidence:2 pre:1 zoubin:2 cannot:3 close:2 selection:2 marginalize:4 nb:12 context:1 writing:1 wong:1 optimize:7 imposed:1 dz:1 lobato:2 maximizing:2 straightforward:1 williams:1 independently:1 thompson:2 welch:1 schonlau:1 hd:1 jmh233:1 reparameterization:1 enhanced:1 user:1 exact:2 locomotion:1 jmhl:1 element:4 expensive:6 approximated:6 walking:1 bache:1 ep:9 observed:2 wang:2 region:1 connected:1 trade:1 highest:2 ahmad:1 disease:1 vehtari:1 complexity:1 cam:3 dynamic:2 ultimately:1 predictive:20 f2:1 easily:4 joint:2 multimodal:1 stock:2 stacked:1 train:1 walter:1 describe:2 monte:2 query:3 hyper:5 multipeak:1 whose:6 heuristic:1 larger:1 widely:1 drawing:1 gp:6 jointly:2 noisy:4 final:1 housing:1 advantage:1 analytical:1 propose:3 gait:2 product:1 uci:1 poorly:1 ymax:4 asserts:1 getting:1 exploiting:1 ijcai:1 optimum:5 produce:5 adam:1 object:1 derive:2 informs:1 ac:3 fixing:2 miguel:1 measured:2 x0i:1 ij:1 progress:1 indicate:1 larochelle:1 qd:1 exploration:1 crc:1 require:1 f1:1 andez:2 fix:3 adjusted:2 exploring:1 hartikainen:1 around:1 considered:1 ground:8 exp:1 mapping:2 mo:1 claim:1 matthew:1 m0:5 early:1 estimation:1 tiesis:1 iw:1 largest:2 successfully:1 hoffman:2 cora:1 mit:1 gaussian:32 aim:1 rather:3 avoid:6 encode:1 improvement:6 frazier:1 sequencing:1 likelihood:4 contrast:2 seeger:2 greedily:1 lizotte:2 elsevier:1 inference:3 typically:1 integrated:1 bandit:2 reproduce:1 transformed:1 selects:1 interested:2 arg:5 dual:1 classification:1 hartmann:4 augment:1 denoted:1 overall:2 proposes:2 copula:2 mutual:2 marginal:2 equal:2 construct:3 once:1 mackay:1 sampling:14 represents:1 broad:1 jones:1 icml:3 discrepancy:1 report:2 gamma:1 ab:1 rearrangement:1 multiply:1 nnet:5 evaluation:21 adjust:1 mixture:1 bipedal:2 sharpe:1 accurate:3 integral:3 encourage:1 necessary:4 bacteria:1 daily:1 hyperprior:1 logarithm:1 initialized:1 stopped:1 increased:1 recomputing:1 modeling:3 soft:1 jyl:1 instance:1 cost:13 introducing:1 deviation:2 entry:1 uniform:2 successful:1 graphic:1 kn:8 dependency:1 corrupted:1 synthetic:9 recht:1 density:4 international:2 szego:1 probabilistic:2 off:1 jos:1 squared:1 again:1 thesis:2 worse:1 admit:1 derivative:6 return:9 li:1 vei:2 de:4 student:1 includes:1 automation:1 dixon:1 devotes:1 satisfy:1 ely:1 depends:1 later:3 performed:2 closed:1 optimistic:1 analyze:1 houlsby:1 collaborative:1 minimize:1 formed:1 publicly:1 accuracy:3 ir:12 variance:10 efficiently:1 bayesian:26 produced:5 fern:1 carlo:2 multiplying:2 vazquez:1 acquisition:19 villemonteix:1 nitrogen:2 minka:1 mohamed:1 associated:1 gain:4 sampled:5 dataset:2 treatment:7 popular:1 massachusetts:1 knowledge:1 brochu:3 appears:1 higher:2 axp:1 follow:2 formulation:1 evaluated:3 box:7 done:2 zeromean:1 furthermore:6 just:1 rejected:1 anderson:1 df2:1 hand:2 ei:22 cohn:1 nonlinear:1 maximizer:12 propagation:2 glance:1 widespread:1 del:1 mode:1 artifact:1 excessively:2 normalized:1 requiring:1 analytically:2 chemical:1 aggressively:1 symmetric:1 moore:1 during:3 bowling:1 aki:1 noted:1 cosine:3 theoretic:4 performs:4 fj:1 novel:2 fi:1 empirically:1 conditioning:3 volume:1 approximates:3 marginals:1 numerically:1 significant:3 measurement:6 pino:1 cambridge:3 smoothness:1 tuning:1 rd:1 grid:5 pm:1 similarly:1 automatic:1 had:1 portfolio:6 chapelle:1 robot:4 money:1 v0:5 add:1 posterior:24 multivariate:5 grizzle:1 noisily:1 optimizing:1 store:1 nonconvex:1 binary:2 yi:3 garch:2 seen:1 additional:3 employed:1 determine:1 maximize:2 bochner:2 v3:1 ii:3 multiple:1 rahimi:1 technical:1 cross:1 lin:1 ensuring:1 prediction:3 basic:1 regression:1 controller:2 noiseless:1 expectation:6 df:2 arxiv:1 iteration:7 kernel:9 normalization:1 robotics:2 achieved:2 c1:13 cell:3 uninformative:1 want:1 krause:1 median:11 walker:7 extra:1 unlike:1 tend:1 call:1 near:1 noting:1 presence:1 latin:1 enough:1 xj:1 zi:1 inner:1 simplifies:1 shift:1 expression:4 effort:1 locating:1 biotechnology:2 matlab:1 generally:1 amount:3 nonparametric:2 band:1 ph:2 reduced:1 http:1 generate:2 tutorial:1 bulk:1 write:4 hyperparameter:4 discrete:1 hennig:1 reformulation:1 nevertheless:2 drawn:2 v1:2 year:1 realworld:1 uncertainty:1 family:1 vn:12 draw:3 decision:3 appendix:3 vf:11 huszar:1 bound:1 ahead:1 constraint:19 westervelt:1 dominated:1 fourier:2 speed:2 argument:2 performing:1 relatively:1 according:3 alternate:1 poor:1 hoo:1 smaller:1 slightly:1 across:3 kakade:1 making:1 ckus:1 intuitively:2 restricted:1 invariant:1 taken:2 computationally:2 hern:2 precomputed:1 synechocystis:1 needed:1 letting:2 tractable:1 end:2 available:3 operation:1 rewritten:1 observe:2 hyperpriors:1 away:1 generic:1 spectral:2 hierarchical:1 alternative:1 existence:1 original:4 kushner:1 marginalized:2 log10:3 concatenated:1 ghahramani:2 murray:1 approximating:3 hypercube:1 comparatively:1 seeking:1 objective:20 quantity:3 codifies:1 occurs:1 strategy:5 costly:1 dependence:1 diagonal:1 gradient:1 hq:1 zooming:1 concatenation:1 evaluate:7 collected:3 unstable:1 toward:1 assuming:1 code:3 index:1 decimal:1 ratio:2 equivalently:1 difficult:1 negative:5 ba:1 implementation:1 design:2 unknown:3 perform:4 upper:8 observation:12 datasets:1 benchmark:2 finite:3 displayed:2 immediate:5 extended:1 incorporated:1 y1:1 arbitrary:1 canada:1 introduced:1 namely:1 required:1 specified:1 c3:3 optimized:4 toolbox:1 nip:6 address:1 usually:1 below:1 challenge:1 including:2 max:1 difficulty:2 force:1 circumvent:1 sian:1 arm:1 representing:1 scheme:1 technology:1 acknowledges:1 columbia:1 prior:7 literature:2 discovery:2 acknowledgement:1 chaplen:1 marginalizing:1 fully:5 lecture:1 interesting:1 rameters:1 allocation:1 foundation:1 editor:1 pi:2 heavy:1 production:2 free:1 rasmussen:2 infeasible:2 guide:2 formal:1 institute:1 sparse:1 distributed:2 slice:2 curve:1 feedback:1 xn:11 world:6 evaluating:3 stuck:1 collection:2 reinforcement:1 adaptive:1 simplified:3 far:3 transaction:1 approximate:11 obtains:1 implicitly:2 rafael:1 global:25 active:4 sequentially:1 uai:1 assumed:1 xi:5 alternatively:1 search:15 latent:9 continuous:2 hydrogen:8 tailed:1 schuler:1 molecule:1 ignoring:1 solak:1 unavailable:1 schuurmans:1 constructing:1 domain:6 diag:4 sp:1 main:1 noise:9 hyperparameters:6 n2:1 x1:4 representative:1 en:5 fails:1 exponential:1 candidate:1 lie:1 pe:59 burrow:1 third:1 down:1 theorem:1 british:1 specific:1 decay:1 normalizing:1 intractable:3 consist:1 albeit:1 sequential:1 adding:2 gained:1 effectively:1 corr:1 phd:2 conditioned:1 push:1 mf:5 rejection:4 entropy:25 boston:1 likely:1 explore:4 univariate:1 gaus2:1 recommendation:3 ubc:1 truth:8 relies:1 cdf:2 conditional:3 goal:1 hard:1 adverse:1 infinite:1 except:1 reducing:1 uniformly:1 averaging:2 wt:2 sampler:1 called:1 total:1 e:28 experimental:1 m3:1 ucb:2 select:3 support:1 latter:1 noisier:2 absolutely:1 incorporate:3 dept:1 princeton:1 srinivas:1 |
4,777 | 5,325 | Submodular meets Structured: Finding Diverse
Subsets in Exponentially-Large Structured Item Sets
Adarsh Prasad
UT Austin
[email protected]
Stefanie Jegelka
UC Berkeley
[email protected]
Dhruv Batra
Virginia Tech
[email protected]
Abstract
To cope with the high level of ambiguity faced in domains such as Computer
Vision or Natural Language processing, robust prediction methods often search
for a diverse set of high-quality candidate solutions or proposals. In structured
prediction problems, this becomes a daunting task, as the solution space (image
labelings, sentence parses, etc.) is exponentially large. We study greedy algorithms for finding a diverse subset of solutions in structured-output spaces by
drawing new connections between submodular functions over combinatorial item
sets and High-Order Potentials (HOPs) studied for graphical models. Specifically,
we show via examples that when marginal gains of submodular diversity functions
allow structured representations, this enables efficient (sub-linear time) approximate maximization by reducing the greedy augmentation step to inference in a
factor graph with appropriately constructed HOPs. We discuss benefits, tradeoffs, and show that our constructions lead to significantly better proposals.
1
Introduction
Many problems in Computer Vision, Natural Language Processing and Computational Biology involve mappings from an input space X to an exponentially large space Y of structured outputs.
For instance, Y may be the space of all segmentations of an image with n pixels, each of which
may take L labels, so |Y| = Ln . Formulations such as Conditional Random Fields (CRFs) [24],
Max-Margin Markov Networks (M3 N) [31], and Structured Support Vector Machines (SSVMs) [32]
have successfully provided principled ways of scoring all solutions y ? Y and predicting the single
highest scoring or maximum a posteriori (MAP) configuration, by exploiting the factorization of a
structured output into its constituent ?parts?.
In a number of scenarios, the posterior P(y|x) has several modes due to ambiguities, and we seek
not only a single best prediction but a set of good predictions:
(1) Interactive Machine Learning. Systems like Google Translate (for machine translation) or
Photoshop (for interactive image segmentation) solve structured prediction problems that are often
ambiguous ("what did the user really mean?"). Generating a small set of relevant candidate solutions
for the user to select from can greatly improve the results.
(2) M-Best hypotheses in cascades. Machine learning algorithms are often cascaded, with the
output of one model being fed into another [33]. Hence, at the initial stages it is not necessary to
make a single perfect prediction. We rather seek a set of plausible predictions that are subsequently
re-ranked, combined or processed by a more sophisticated mechanism.
In both scenarios, we ideally want a small set of M plausible (i.e., high scoring) but non-redundant
(i.e., diverse) structured-outputs to hedge our bets.
Submodular Maximization and Diversity. The task of searching for a diverse high-quality subset of items from a ground set V has been well-studied in information retrieval [5], sensor placement [22], document summarization [26], viral marketing [17], and robotics [10]. Across these
domains, submodularity has emerged as an a fundamental and practical concept ? a property of
functions for measuring diversity of a subset of items. Specifically, a set function F : 2V ? R is
submodular if its marginal gains, F (a|S) ? F (S ?a)?F (S) are decreasing, i.e. F (a|S) ? F (a|T )
1
argmax F (a | S) ?
a?V
(a) Image
(b) All segmentations: |V | = Ln
r(y)
+
d(y | S)
(c) Structured Representation.
Figure 1: (a) input image; (b) space of all possible object segmentations / labelings (each item is a segmentation); (c) we convert the problem of finding the item with the highest marginal gain F (a|S) to a MAP inference
problem in a factor graph over base variables y with an appropriately defined HOP.
for all S ? T and a ?
/ T . In addition, if F is monotone, i.e., F (S) ? F (T ), ?S ? T , then a simple
greedy algorithm (that in each iteration t adds to the current set S t the item with the largest marginal
gain F (a|S t )) achieves an approximation factor of (1 ? 1e ) [27]. This result has had significant
practical impact [21]. Unfortunately, if the number of items |V | is exponentially large, then even a
single linear scan for greedy augmentation is infeasible.
In this work, we study conditions under which it is feasible to greedily maximize a submodular
function over an exponentially large ground set V = {v1 , . . . , vN } whose elements are combinatorial objects, i.e., labelings of a base set of n variables y = {y1 , y2 , . . . , yn }. For instance, in image
segmentation, the base variables yi are pixel labels, and each item a ? V is a particular labeling of
the pixels. Or, if each base variable ye indicates the presence or absence of an edge e in a graph,
then each item may represent a spanning tree or a maximal matching. Our goal is to find a set of
M plausible and diverse configurations efficiently, i.e. in time sub-linear in |V | (ideally scaling as
a low-order polynomial in log |V |). We will assume F (?) to be monotone submodular, nonnegative
and normalized (F (?) = 0), and base our study on the greedy algorithm. As a running example, we
focus on pixel labeling, where each base variable takes values in a set [L] = {1, . . . , L} of labels.
Contributions. Our principal contribution is a conceptual one. We observe that marginal gains of a
number of submodular functions allow structured representations, and this enables efficient greedy
maximization over exponentially large ground sets ? by reducing the greedy augmentation step to
a MAP inference query in a discrete factor graph augmented with a suitably constructed HighOrder Potential (HOP). Thus, our work draws new connections between two seemingly disparate
but highly related areas in machine learning ? submodular maximization and inference in graphical
models with structured HOPs. As specific examples, we construct submodular functions for three
different, task-dependent definitions of diversity, and provide reductions to three different HOPs for
which efficient inference techniques have already been developed. Moreover, we present a generic
recipe for constructing such submodular functions, which may be ?plugged? with efficient HOPs
discovered in future work. Our empirical contribution is an efficient algorithm for producing a set of
image segmentations with significantly higher oracle accuracy1 than previous works. The algorithm
is general enough to transfer to other applications. Fig. 1 shows an overview of our approach.
Related work: generating multiple solutions. Determinental Point Processesare an elegant probabilistic model over sets of items with a preference for diversity. Its generalization to a structured
setting [23] assumes a tree-structured model, an assumption that we do not make. Guzman-Rivera et
al. [14, 15] learn a set of M models, each producing one solution, to form the set of solutions. Their
approach requires access to the learning sub-routine and repeated re-training of the models, which
is not always possible, as it may be expensive or proprietary. We assume to be given a single (pretrained) model from which we must generate multiple diverse, good solutions. Perhaps the closest
to our setting are recent techniques for finding diverse M -best solutions [2, 28] or modes [7, 8]
in graphical models. While [7] and [8] are inapplicable since they are restricted to chain and tree
graphs, we compare to other baselines in Section 3.2 and 4.
1.1 Preliminaries and Notation
We select from a ground set V of N items. Each item is a labeling y = {y1 , y2 , . . . , yn }
of n base variables. For clarity, we use non-bold letters a ? V for items, and boldface letters y for base set configurations. Uppercase letters refer to functions over the ground set items
F (a|A), R(a|A), D(a|A), and lowercase letters to functions over base variables f (y), r(y), d(y).
1
The accuracy of the most accurate segmentation in the set.
2
Formally, there is a bijection ? : V 7? [L]m that maps items a ? V to their representation as base
variable labelings y = ?(a). For notational simplicity, we often use y ? S to mean ??1 (y) ? S,
i.e. the item corresponding to the labeling y is present in the set S ? V . We write ` ? y if the label
` is used in y, i.e. ?j s.t. yj = `. For a set c ? [n], we use yc to denote the tuple {yi | i ? c}.
Our goal to find an ordered set or list of items S ? V that maximizes a scoring function F . Lists
generalize the notation of sets, and allow for reasoning of item order and repetitions. More details
about list vs set prediction can be found in [29, 10].
Scoring Function. We trade off the relevance and diversity of list S ? V via a scoring function
F : 2V ? R of the form
F (S) = R(S) + ?D(S),
(1)
P
where R(S) = a?S R(a) is a modular nonnegative relevance function that aggregates the quality
of all items in the list; D(S) is a monotone normalized submodular function that measure the diversity of items in S; and ? ? 0 is a trade-off parameter. Similar objective functions were used e.g. in
[26]. They are reminiscent of the general paradigm in machine learning of combining a loss function that measures quality (e.g. training error) and a regularization term that encourages desirable
properties (e.g. smoothness, sparsity, or ?diversity?).
Submodular Maximization. We aim to find a list S that maximizes F (S) subject to a cardinality
constraint |S| ? M . For monotone submodular F , this may be done via a greedy algorithm that
starts out with S 0 = ?, and iteratively adds the next best item:
S t = S t?1 ? at ,
at ? argmaxa?V F (a | S t?1 ).
(2)
The final solution S M is within a factor of (1 ? 1e ) of the optimal solution S ? : F (S M ) ? (1 ?
1
?
e )F (S ) [27]. The computational bottleneck is that in each iteration, we must find the item with
the largest marginal gain. Clearly, if |V | has exponential size, we cannot touch each item even once.
Instead, we propose ?augmentation sub-routines? that exploit the structure of V and maximize the
marginal gain by solving an optimization problem over the base variables.
2
Marginal Gains in Configuration Space
To solve the greedy augmentation step via optimization over y, we transfer the marginal gain from
the world of items to the world of base variables and derive functions on y from F :
(3)
F (??1 (y) | S) = R(??1 (y)) +? D(??1 (y) | S) .
{z
} | {z }
|
{z
}
|
f (y|S)
r(y)
d(y|S)
Maximizing F (a|S) now means maximizing f (y|S) for y = ?(a). This can be a hard combinatorial
optimization problem in general. However, as we will see, there is a broad class of useful functions
F for which f inherits exploitable structure, and argmaxy f (y|S) can be solved efficiently, exactly
or at least approximately.
Relevance Function. We use a structured relevance function R(a) that is the score of a factor graph
defined over the base variables y. Let G = (V, E) be a graph defined over {y1 , y2 , . . . , yn }, i.e.
V = [n], E ? V2 . Let C = {C | C ? V} be a set of cliques in the graph, and let ?C : [L]|C| 7? R
be the log-potential functions (or factors)
for these cliques. The quality of an item a = ??1 (y)
P
is then given by R(a) = r(y)
P = C?C ?C (y
PC ). For instance, with only node and edge factors,
this quality becomes r(y) = p?V ?p (yp ) + (p,q)?E ?pq (yp , yq ). In this model, finding the single
highest quality item corresponds to maximum a posteriori (MAP) inference in the factor graph.
Although we refer to terms with probabilistic interpretations such as ?MAP?, we treat our relevance
function P
as output of an energy-based model [25] such as a Structured SVM [32]. For instance,
r(y) = C?C ?C (yC ) = w| ?(y) for parameters w and feature vector ?(y). Moreover, we assume that the relevance function r(y) is nonnegative2 . This assumption ensures that F (?) is monotone. If F is non-monotone, algorithms other than the greedy are needed [4, 12]. We leave this
generalization for future work. In most application domains the relevance function is learned from
data and thus our positivity assumption is not restrictive ? one can simply learn a positive relevance
function. For instance, in SSVMs, the relevance weights are learnt to maximize the margin between
the correct labeling and all incorrect ones. We show in the supplement that SSVM parameters that
assign nonnegative scores to all labelings achieve exactly the same hinge loss (and thus the same
generalization error) as without the nonnegativity constraint.
2
Strictly speaking, this condition is sufficient but not necessary. We only need nonnegative marginal gains.
3
Figure 2: Diversity via groups:
(a) groups defined by the presence of labels (i.e. #groups
= L); (b) groups defined by
Hamming balls around each
item/labeling (i.e. #groups =
Ln ). In each case, diversity is
measured by how many groups
are covered by a new item. See
text for details.
(a) Label Groups
3
(b) Hamming Ball Groups
Structured Diversity Functions
We next discuss a general recipe for constructing monotone submodular diversity functions D(S),
and for reducing their marginal gains to structured representations over the base variables
d(y|S).
S
Our scheme relies on constructing groups Gi that cover the ground set, i.e. V = i Gi . These
groups will be defined by task-dependent characteristics ? for instance, in image segmentation, G`
can be the set of all segmentations that contain label `. The groups can be overlapping. For instance,
if a segmentation y contains pixels labeled ?grass? and ?cow?, then y ? Ggrass and y ? Gcow .
Group Coverage: Count Diversity. Given V and a set of groups {Gi }, we measure the diversity
of a list S in terms of its group coverage, i.e., the number of groups covered jointly by items in S:
D(S) = i | Gi ? S 6= ? ,
(4)
where we define Gi ? S as the intersection of Gi with the set of unique items in S. It is easy to show
that this function is monotone submodular. If G` is the group of all segmentations that contain label
`, then the diversity measure of a list of segmentations S is the number of object labels that appear
in any a ? S. The marginal gain is the number of new groups covered by a:
D(a | S) = i | a ? Gi and S ? Gi = ? .
(5)
Thus, the greedy algorithm will try to find an item/segmentation that belongs to as many as yet
unused groups as possible.
Group Coverage: General Diversity. More generally, instead of simply counting the number of
groups covered by S, we can use a more refined decay
X
D(S) =
h Gi ? S .
(6)
i
where h is any nonnegative nondecreasing concave scalar function. This is a sum of submodular
functions and hence submodular.
Eqn. (4) is a special case of Eqn. (6) with h(y) = min{1, y}.
?
Other possibilities are ?, or log(1 + ?). For this general definition of diversity, the marginal gain is
X
D(a | S) =
h 1 + Gi ? S ? h Gi ? S .
(7)
i:Gi 3a
Since h is concave, the gain h 1 + Gi ? S ? h Gi ? S decreases as S becomes larger. Thus,
the marginal gain of an item a is proportional to how rare each group Gi 3 a is in the list S.
In each step of the greedy algorithm, we maximize r(y) + ?d(y|S). We already established a
structured representation of r(y) via a factor graph on y. In the next few subsections, we specify
three example definitions of groups Gi that instantiate three diversity functions D(S). For each
D(S), we show how the marginal gains D(a|S) can be expressed as a specific High-Order Potential
(HOP) d(y|S) in the factor graph over y. These HOPs are known to be efficiently optimizable, and
hence we can solve the augmentation step efficiently. Table 1 summarizes these connections.
Diversity and Parsimony. If the groups Gi are overlapping, some y can belong to many groups
simultaneously. While such a y may offer an immediate large gain in diversity, in many applications
it is more natural to seek a small list of complementary labelings rather than having all labels occur
in the same y. For instance, in image segmentation with groups defined by label presence (Sec. 3.1),
natural scenes are unlikely to contain many labels at the same time. Instead, the labels should
be spread across the selected labelings y ? S. Hence, we include a parsimony factor p(y) that
biases towards simpler labelings y. This term is a modular function and does not affect the diversity
functions directly. We next outline some example instantiations of the functions (4) and (6).
4
Groups (Gi )
Higher Order Potentials
Section 3.1
Labels
Label Cost
Supplement Label Transitions Co-operative Cuts
Section 3.2
Hamming Balls
Cardinality Potentials
Table 1: Different diversity functions and corresponding HOPs.
3.1 Diversity of Labels
For the first example, let G` be the set of all labelings y containing the label `, i.e. y ? G` if and only
if yj = ` for some j ? [n]. Such a diversity function arises in multi-class image segmentation ? if the
highest scoring segmentation contains ?sky? and ?grass?, then we would like to add complementary
segmentations that contain an unused class label, say ?sheep? or ?cow?.
Structured Representation of Marginal Gains. The marginal gain for this diversity function turns
out to be a HOP called label cost [9]. It penalizes each label that occurs in a previous segmentation.
Let lcountS (`) be the number of segmentations in S that contain label `. In the simplest case of
coverage diversity (4), the marginal gain provides a constant reward for every as yet unseen label `:
X
d(y | S) = ` | y ? G` , S ? G` = ? =
1.
(8)
`?y,lcountS (`)=0
For the general group coverage diversity (6), the gain becomes
d(y|S) =
X
X
h 1 + G` ? S ? h G` ? S =
h 1 + lcountS (`) ? h lcountS (`) .
`:G` 3y
`?y
Thus, d(y|S) rewards the presence of a label ` in y by an amount proportional to howP
rare ` is in
the segmentations already chosen in S. The parsimony factor in this setting is p(y) = `?y c(`).
In the simplest case, c(`) = ?1, i.e. we are charged a constant for every label used in y.
With this type of diversity (and parsimony terms), the greedy augmentation step is equivalent to
performing MAP inference in a factor graph augmented with label reward HOPs: argmaxy r(y) +
?(d(y | S) + p(y)). Delong et al. [9] show how to perform approximate MAP inference with such
label costs via an extension to the standard ?-expansion [3] algorithm.
Label Transitions. Label Diversity can be extended to reward not just the presence of previously
unseen labels, but also the presence of previously unseen label transitions (e.g., a person in front
of a car or a person in front of a house). Formally, we define one group G`,`0 per label pair `, `0 ,
and y ? G`,`0 if it contains two adjacent variables yi , yj with labels yi = `, yj = `0 . This diversity
function rewards the presence of a label pair (`, `0 ) by an amount proportional to how rare this pair
is in the segmentations that are part of S. For such functions, the marginal gain d(y|S) becomes
a HOP called cooperative cuts [16]. The inference algorithm in [19] gives a fully polynomial-time
approximation scheme for any nondecreasing, nonnegative h, and the exact gain maximizer for the
count function h(y) = min{1, y}. Further details may be found in the supplement.
3.2 Diversity via Hamming Balls
The label diversity function simply rewarded the presence of a label `, irrespective of which or how
many variables yi were assigned
that label. The next diversity function rewards a large Hamming disPn
tance Ham(y1 , y2 ) = i=1 [[yi1 6= yi2 ]] between configurations (where [[?]] is the Iverson bracket.)
Let Bk (y) denote the k-radius Hamming ball centered at y, i.e. B(y) = {y0 | Ham(y0 , y) ? k}.
The previous section constructed one group per label `. Now, we construct one group Gy for each
configuration y, which is the k-radius Hamming ball centered at y, i.e. Gy = Bk (y).
Structured Representation of Marginal Gains. For this diversity, the marginal gain d(y|S) becomes a HOP called cardinality potential [30]. For count group coverage, this becomes
d(y|S) = y0 | Gy0 ? (S ? y) 6= ? ? y0 | Gy0 ? S 6= ?
(9a)
[
[
h [
i
=
Bk (y0 ) ?
Bk (y0 ) = Bk (y) ? Bk (y) ?
Bk (y0 ) ,
(9b)
y0 ?S?y
y0 ?S
0
y0 ?S
i.e., the marginal gain of adding y is the number of new configurations y covered by the Hamming
ball centered at y. Since the size of the intersection of Bk (y) with a union of Hamming balls does
not have a straightforward structured representation, we maximize a lower bound on d(y|S) instead:
X
Bk (y) ? Bk (y0 )
(10)
d(y | S) ? dlb (y | S) ? Bk (y) ?
y0 ?S
5
This lower bound dlb (y|S) overcounts the intersection in Eqn. (9b) by summing the intersections
with each Bk (y0 ) separately. We can also interpret this lower bound as clipping the series arising
from the inclusion-exclusion principle to the first-order terms. Importantly, (10) depends on y only
via its Hamming distance to y0 . This is a cardinality potential that depends only on the number of
variables yi assigned to a particular label. Specifically, ignoring constant terms, the lower bound can
be
factors (one for each previous solution y0 ? S): dlb (y|S) =
P written as a summation of cardinality
b
0
0
0
y0 ?S ?y (y), where ?y (y) = |S| ? Iy (y), b is a constant (size of a k-radius Hamming ball), and
Iy0 (y) is the number of points in the intersection of k-radius Hamming balls centered at y0 and y.
With this approximation, the greedy step means performing MAP inference in a factor graph augmented with cardinality potentials: argmaxy r(y) + ?dlb (y|S). This may be solved via messagepassing, and all outgoing messages from cardinality factors can be computed in O(n log n) time
[30]. While this algorithm does not offer any approximation guarantees, it performs well in practice. A subtle point to note is that dlb (y|S) is always decreasing w.r.t. |S| but may become negative
due to over-counting. We can fix this by clamping dlb (y|S) to be greater than 0, but in our experiments this was unnecessary ? the greedy algorithm never chose a set where dlb (y|S) was negative.
Comparison to DivMBest. The greedy algorithm for Hamming diversity is similar in spirit to the
recent work of Batra et al. [2], who also proposed a greedy algorithm (DivMBest) for finding diverse
MAP solutions in graphical models. They did not provide any justification for greedy, and our
formulation sheds some light on their work. Similar to our approach, at
Peach greedy step, DivMBest
involves maximizing a diversity-augmented score: argmaxy r(y)+? y0 ?S ?y0 (y). However, their
Pn
diversity function grows linearly with the Hamming distance, ?y0 (y) = Ham(y0 , y) = i=1 [[yi0 6=
yi ]]. Linear diversity rewards are not robust, and tend to over-reward diversity. Our formulation uses
b
a robust diversity function ?y0 (y) = |S|
? Iy0 (y) that saturates as y moves far away from y0 .
In our experiments, we make the saturation behavior smoothly tunable via a parameter ?: Iy0 (y) =
0
e?? Ham(y ,y) . A larger ? corresponds to Hamming balls of smaller radius, and can be set to optimize
performance on validation data. We found this to work better than directly tuning the radius k.
4
Experiments
We apply our greedy maximization algorithms to two image segmentation problems: (1) interactive
binary segmentation (object cutout) (Section 4.1); (2) category-level object segmentation on the
PASCAL VOC 2012 dataset [11] (Section 4.2). We compare all methods by their respective oracle
accuracies, i.e. the accuracy of the most accurate segmentation in the set of M diverse segmentations
returned by that method. For a small value of M ? 5 to 10, a high oracle accuracy indicates that
the algorithm has achieved high recall and has identified a good pool of candidate solutions for
further processing in a cascaded pipeline. In both experiments, the label ?background? is typically
expected to appear somewhere in the image, and thus does not play a role in the label cost/transition
diversity functions. Furthermore, in binary segmentation there is only one non-background label.
Thus, we report results with Hamming diversity only (label cost and label transition diversities are
not applicable). For the multi-class segmentation experiments, we report experiments with all three.
Baselines. We compare our proposed methods against DivMBest [2], which greedily produces
diverse segmentation by explicitly adding a linear Hamming distance term to the factor graph. Each
Hamming term isP
decomposable along the variables yi and simply modifies the node potentials
0
? i ) = ?(yi )+?
?(y
y0 ?S [[yi 6= yi ]]. DivMBest has been shown to outperform techniques such as MBest-MAP [34, 1], which produce high scoring solutions without a focus on diversity, and samplingbased techniques, which produce diverse solutions without a focus on the relevance term [2]. Hence,
we do not include those methods here. We also report results for combining different diversity
functions via two operators: (?), where we generate the top M
k solutions for each of k diversity
functions and then concatenate these lists; and (?), where we linearly combine diversity functions
(with coefficients chosen by k-D grid search) and generate M solutions using the combined diversity.
4.1 Interactive segmentation
In interactive foreground-background segmentation, the user provides partial labels via scribbles.
One way to minimize interactions is for the system to provide a set of candidate segmentations for
the user to choose from. We replicate the experimental setup of [2], who curated 100 images from
the PASCAL VOC 2012 dataset, and manually provided scribbles on objects contained in them.
For each image, the relevance model r(y) is a 2-label pairwise CRF, with a node term for each
6
Label Cost (LC)
min{1,
?}
p
(?)
log(1 + ?)
MAP
M=5
M=15
42.35
42.35
42.35
45.43
45.72
46.28
45.58
50.01
50.39
Hamming Ball (HB)
DivMBest
HB
MAP
M=5
M=15
43.43
43.43
51.21
51.71
52.90
55.32
? Combined Diversity
HB ? LC ? LT
DivMBest ? HB ? LC ? LT
M=15
M=16
56.97
-
57.39
Label Transition (LT)
min{1,
?}
p
(?)
log(1 + ?)
MAP
M=5
M=15
42.35
42.35
42.35
44.26
45.43
45.92
44.78
46.21
46.89
? Combined Diversity
M=15
DivMBest ? HB
DivMBest ? LC ? LT
55.89
53.47
Table 2: PASCAL VOC 2012 val oracle accuracies for different diversity functions.
superpixel in the image and an edge term for each adjacent pair of superpixels. At each superpixel,
we extract colour and texture features. We train a Transductive SVM from the partial supervision
provided by the user scribbles. The node potentials are derived from the scores of these TSVMs. The
edge potentials are contrast-sensitive Potts. Fifty of the images were used for tuning the diversity
parameters ?, ?, and the other 50 for reporting oracle accuracies. The 2-label contrast-sensitive Potts
model results in a supermodular relevance function r(y), which can be efficiently maximized via
graph cuts [20]. The Hamming ball diversity dlb (y|S) is a collection of cardinality factors, which
we optimize with the Cyborg implementation [30].
Results. For each of the 50 test images in our dataset we generated the single best y1 and 5 additional solutions {y2 , . . . , y6 } using each method. Table 3 shows the average oracle accuracies for
DivMBest, Hamming ball diversity, and their two combinations. We can see that the combinations
slightly outperform both approaches.
DivMBest
Hamming Ball
DivMBest?Hamming Ball
DivMBest?Hamming Ball
MAP
M=2
M=6
91.57
91.57
-
93.16
93.95
-
95.02
94.86
95.16
95.14
Table 3: Interactive segmentation: oracle pixel accuracies averaged over 50 test images
4.2
Category level Segmentation
In category-level object segmentation, we label each pixel with one of 20 object categories or background. We construct a multi-label pairwise CRF on superpixels. Our node potentials are outputs of
category-specific regressors trained by [6], and our edge potentials are multi-label Potts. Inference
in the presence of diversity terms is performed with the implementations of Delong et al. [9] for
label costs, Tarlow et al. [30] for Hamming ball diversity, and Boykov et al. [3] for label transitions.
Figure 3: Qualitative Results:
each row shows the original image, ground-truth segmentation
(GT) from PASCAL, the singlebest segmentation y1 , and oracle
segmentation from the M = 15
segmentations (excluding y1 ) for
different definitions of diversity.
Hamming typically performs the
best. In certain situations (row3),
label transitions help since the
single-best segmentation y1 included a rare pair of labels (dogcat boundary).
Results. We evaluate all methods on the PASCAL VOC 2012 data [11], consisting of train, val
and test partitions with about 1450 images each. We train the regressors of [6] on train, and
report oracle accuracies of different methods on val (we cannot report oracle results on test since
those annotations are not publicly available). Diversity parameters (?, ?) are chosen by performing cross-val on val. The standard PASCAL accuracy is the corpus-level intersection-over-union
measure, averaged over all categories. For both label cost and transition, we try 3 different concave
7
p
functions h(?) = min{1, ?}, (?) and log(1 + ?). Table 2 shows the results.3 Hamming ball diversity performs the best, followed by DivMBest, and label cost/transitions are worse here. We found
that while worst on average, label transition diversity helps in an interesting scenario ? when the
first best segmentation y1 includes a pair of rare or mutually confusing labels (say dog-cat). Fig. 3
shows an example, and more illustrations are provided in the supplement. In these cases, searching
for a different label transition produces a better segmentation. Finally, we note that lists produced
with combined diversity significantly outperform any single method (including DivMBest).
5
Discussion and Conclusion
In this paper, we study greedy algorithms for maximizing scoring functions that promote diverse
sets of combinatorial configurations. This problem arises naturally in domains such as Computer
Vision, Natural Language Processing, or Computational Biology, where we want to search for a set
of diverse high-quality solutions in a structured output space.
The diversity functions we propose are monotone submodular functions by construction. Thus, if
r(y) + p(y) ? 0 for all y, then the entire scoring function F is monotone submodular. We showed
that r(y) can simply be learned to be positive. The greedy algorithm for maximizing monotone
submodular functions has proved useful in moderately-sized unstructured spaces. To the best of our
knowledge, this is the first generalization to exponentially large structured output spaces. In particular, our contribution lies in reducing the greedy augmentation step to inference with structured,
efficiently solvable HOPs. This insight makes new connections between submodular optimization
and work on inference in graphical models. We now address some questions.
Can we sample? One question that may be posed is how random sampling would perform for large
ground sets V . Unfortunately, the expected value of a random sample of M elements can be much
worse than the optimal value F (S ? ), especially if N is large. Lemma 1 is proved in the supplement.
Lemma 1. Let S ? V be a sample of size M taken uniformly at random. There exist monotone
submodular functions where E[F (S)] ? M
N max|S|=M F (S).
Guarantees? If F is nonnegative, monotone submodular, then using an exact HOP inference algorithm will clearly result in an approximation factor of 1 ? 1/e. But many HOP inference procedures
are approximate. Lemma 2 formalizes how approximate inference affects the approximation bounds.
Lemma 2. Let F ? 0 be monotone submodular. If each step of the greedy algorithm uses an
approximate marginal gain maximizer bt+1 with F (bt+1 | S t ) ? ? maxa?V F (a | S t ) ? t+1 , then
PM
F (S M ) ? (1 ? e1? ) max|S|?M F (S) ? i=1 t .
Parts of Lemma 2 have been observed in previous work [13, 29]; we show the combination in the
supplement. If F is monotone but not nonnegative,
then Lemma 2 can be extended to a relative error
P
F (S M )?Fmin
1
i i
bound F (S ? )?Fmin ? (1 ? e? ) ? F (S ? )?Fmin that refers to Fmin = minS F (S) and the optimal
solution S ? . While stating these results, we add that further additive approximation losses occur if
the approximation bound for inference is computed on a shifted or reflected function (positive scores
vs positive energies). We pose theoretical improvements as an open question for future work. That
said, our experiments convincingly show that the algorithms perform very well in practice, even
when there are no guarantees (as with Hamming Ball diversity).
Generalization. In addition to the three specific examples in Section 3, our constructions generalize
to the broad HOP class of upper-envelope potentials [18]. The details are provided in the supplement.
Acknowledgements. We thank Xiao Lin for his help. The majority of this work was done while AP was an
intern at Virginia Tech. AP and DB were partially supported by the National Science Foundation under Grant
No. IIS-1353694 and IIS-1350553, the Army Research Office YIP Award W911NF-14-1-0180, and the Office
of Naval Research Award N00014-14-1-0679, awarded to DB. SJ was supported by gifts from Amazon Web
Services, Google, SAP, The Thomas and Stacey Siebel Foundation, Apple, C3Energy, Cisco, Cloudera, EMC,
Ericsson, Facebook, GameOnTalis, Guavus, HP, Huawei, Intel, Microsoft, NetApp, Pivotal, Splunk, Virdata,
VMware, WANdisco, and Yahoo!.
References
[1] D. Batra. An Efficient Message-Passing Algorithm for the M-Best MAP Problem. In UAI, 2012. 6
3
MAP accuracies in Table 2 are different because of two different approximate MAP solvers: LabelCost/Transition use alpha-expansion and HammingBall/DivMBest use message-passing.
8
[2] D. Batra, P. Yadollahpour, A. Guzman-Rivera, and G. Shakhnarovich. Diverse M-Best Solutions in
Markov Random Fields. In ECCV, 2012. 2, 6
[3] Y. Boykov, O. Veksler, and R. Zabih. Efficient approximate energy minimization via graph cuts. PAMI,
20(12):1222?1239, 2001. 5, 7
[4] N. Buchbinder, M. Feldman, J. Naor, and R. Schwartz. A tight (1/2) linear-time approximation to unconstrained submodular maximization. In FOCS, 2012. 3
[5] J. Carbonell and J. Goldstein. The use of MMR, diversity-based reranking for reordering documents
and producing summaries. In Proc. 21st annual international ACM SIGIR conference on Research and
Development in Information Retrieval, SIGIR ?98, pages 335?336, 1998. 1
[6] J. Carreira, R. Caseiro, J. Batista, and C. Sminchisescu. Semantic segmentation with second-order pooling. In ECCV, pages 430?443, 2012. 7
[7] C. Chen, V. Kolmogorov, Y. Zhu, D. Metaxas, and C. H. Lampert. Computing the m most probable modes
of a graphical model. In AISTATS, 2013. 2
[8] C. Chen, H. Liu, D. Metaxas, and T. Zhao. Mode estimation for high dimensional discrete tree graphical
models. In NIPS, 2014. 2
[9] A. Delong, A. Osokin, H. N. Isack, and Y. Boykov. Fast approximate energy minimization with label
costs. In CVPR, pages 2173?2180, 2010. 5, 7
[10] D. Dey, T. Liu, M. Hebert, and J. A. Bagnell. Contextual sequence prediction with application to control
library optimization. In Robotics Science and Systems (RSS), 2012. 1, 3
[11] M. Everingham, L. V. Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The PASCAL Visual Object
Classes Challenge 2012 (VOC2012). 6, 7
[12] U. Feige, V. S. Mirrokni, and J. Vondrak. Maximizing non-monotone submodular functions. In FOCS,
2007. ISBN 0-7695-3010-9. 3
[13] P. Goundan and A. Schulz. Revisiting the greedy approach to submodular set function maximization.
Manuscript, 2009. 8
[14] A. Guzman-Rivera, D. Batra, and P. Kohli. Multiple Choice Learning: Learning to Produce Multiple
Structured Outputs. In Proc. NIPS, 2012. 2
[15] A. Guzman-Rivera, P. Kohli, D. Batra, and R. Rutenbar. Efficiently enforcing diversity in multi-output
structured prediction. In AISTATS, 2014. 2
[16] S. Jegelka and J. Bilmes. Submodularity beyond submodular energies: Coupling edges in graph cuts. In
CVPR, 2011. 5
[17] D. Kempe, J. Kleinberg, and E. Tardos. Maximizing the spread of influence through a social network. In
ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD), 2003. 1
[18] P. Kohli and M. P. Kumar. Energy minimization for linear envelope MRFs. In CVPR, 2010. 8
[19] P. Kohli, A. Osokin, and S. Jegelka. A principled deep random field model for image segmentation. In
CVPR, 2013. 5
[20] V. Kolmogorov and R. Zabih. What energy functions can be minimized via graph cuts? PAMI, 26(2):
147?159, 2004. 7
[21] A. Krause and S. Jegelka. Submodularity in machine learning: New directions. ICML Tutorial, 2013. 2
[22] A. Krause, A. Singh, and C. Guestrin. Near-optimal sensor placements in Gaussian processes: Theory,
efficient algorithms and empirical studies. JMLR, 9:235?284, 2008. 1
[23] A. Kulesza and B. Taskar. Structured determinantal point processes. In Proc. NIPS, 2010. 2
[24] J. D. Lafferty, A. McCallum, and F. C. N. Pereira. Conditional random fields: Probabilistic models for
segmenting and labeling sequence data. In ICML, 2001. 1
[25] Y. LeCun, S. Chopra, R. Hadsell, M. Ranzato, and F. Huang. A tutorial on energy-based learning. In
Predicting Structured Data. MIT Press, 2006. 3
[26] H. Lin and J. Bilmes. A class of submodular functions for document summarization. In ACL, 2011. 1, 3
[27] G. Nemhauser, L. Wolsey, and M. Fisher. An analysis of approximations for maximizing submodular set
functions. Mathematical Programming, 14(1):265?294, 1978. 2, 3
[28] D. Park and D. Ramanan. N-best maximal decoders for part models. In ICCV, 2011. 2
[29] M. Streeter and D. Golovin. An online algorithm for maximizing submodular functions. In NIPS, 2008.
3, 8
[30] D. Tarlow, I. E. Givoni, and R. S. Zemel. HOP-MAP: Efficient message passing with high order potentials.
In AISTATS, pages 812?819, 2010. 5, 6, 7
[31] B. Taskar, C. Guestrin, and D. Koller. Max-Margin Markov networks. In NIPS, 2003. 1
[32] I. Tsochantaridis, T. Joachims, T. Hofmann, and Y. Altun. Large margin methods for structured and
interdependent output variables. JMLR, 6:1453?1484, 2005. 1, 3
[33] P. Viola and M. J. Jones. Robust real-time face detection. Int. J. Comput. Vision, 57(2):137?154, May
2004. ISSN 0920-5691. 1
[34] C. Yanover and Y. Weiss. Finding the m most probable configurations using loopy belief propagation. In
NIPS, 2003. 6
9
| 5325 |@word kohli:4 polynomial:2 replicate:1 yi0:1 suitably:1 everingham:1 open:1 seek:3 prasad:1 r:1 rivera:4 reduction:1 initial:1 configuration:9 contains:3 score:5 series:1 siebel:1 liu:2 document:3 batista:1 current:1 contextual:1 yet:2 must:2 reminiscent:1 written:1 determinantal:1 concatenate:1 partition:1 additive:1 kdd:1 hofmann:1 enables:2 v:2 grass:2 greedy:25 instantiate:1 selected:1 item:33 reranking:1 samplingbased:1 mccallum:1 yi1:1 tarlow:2 provides:2 bijection:1 node:5 preference:1 simpler:1 mathematical:1 along:1 constructed:3 iverson:1 become:1 incorrect:1 qualitative:1 naor:1 focs:2 combine:1 pairwise:2 expected:2 behavior:1 multi:5 photoshop:1 voc:4 decreasing:2 cardinality:8 solver:1 becomes:7 provided:5 gift:1 moreover:2 notation:2 maximizes:2 what:2 parsimony:4 maxa:1 developed:1 finding:7 guarantee:3 formalizes:1 berkeley:2 sky:1 every:2 concave:3 interactive:6 shed:1 exactly:2 schwartz:1 control:1 ramanan:1 grant:1 yn:3 producing:3 appear:2 segmenting:1 positive:4 service:1 treat:1 guavus:1 meet:1 approximately:1 ap:2 pami:2 chose:1 acl:1 studied:2 co:1 factorization:1 averaged:2 practical:2 unique:1 lecun:1 yj:4 union:2 practice:2 procedure:1 area:1 empirical:2 mbest:1 significantly:3 cascade:1 matching:1 cloudera:1 refers:1 argmaxa:1 altun:1 cannot:2 tsochantaridis:1 operator:1 influence:1 optimize:2 equivalent:1 map:19 charged:1 crfs:1 maximizing:9 straightforward:1 modifies:1 williams:1 sigir:2 hadsell:1 simplicity:1 decomposable:1 unstructured:1 amazon:1 insight:1 importantly:1 his:1 searching:2 cutout:1 justification:1 tardos:1 construction:3 play:1 user:5 exact:2 programming:1 us:2 givoni:1 hypothesis:1 superpixel:2 element:2 expensive:1 curated:1 cut:6 labeled:1 cooperative:1 observed:1 role:1 taskar:2 solved:2 worst:1 revisiting:1 ensures:1 ranzato:1 trade:2 highest:4 decrease:1 principled:2 ham:4 moderately:1 reward:8 ideally:2 highorder:1 trained:1 singh:1 solving:1 shakhnarovich:1 tight:1 inapplicable:1 isp:1 cat:1 kolmogorov:2 train:4 fast:1 query:1 zemel:1 labeling:7 aggregate:1 refined:1 whose:1 emerged:1 modular:2 solve:3 plausible:3 larger:2 drawing:1 say:2 posed:1 cvpr:4 gi:18 unseen:3 transductive:1 jointly:1 nondecreasing:2 final:1 seemingly:1 online:1 sequence:2 isbn:1 propose:2 interaction:1 maximal:2 relevant:1 combining:2 translate:1 achieve:1 constituent:1 recipe:2 exploiting:1 produce:5 generating:2 perfect:1 leave:1 object:9 help:3 derive:1 coupling:1 stating:1 pose:1 measured:1 coverage:6 c:1 involves:1 direction:1 submodularity:3 radius:6 correct:1 subsequently:1 centered:4 assign:1 fix:1 generalization:5 really:1 preliminary:1 probable:2 summation:1 strictly:1 extension:1 around:1 dhruv:1 ground:8 mapping:1 achieves:1 estimation:1 proc:3 applicable:1 combinatorial:4 label:62 utexas:1 sensitive:2 largest:2 repetition:1 successfully:1 minimization:3 mit:1 clearly:2 sensor:2 always:2 gaussian:1 aim:1 rather:2 pn:1 bet:1 office:2 derived:1 focus:3 inherits:1 naval:1 notational:1 potts:3 joachim:1 indicates:2 improvement:1 superpixels:2 tech:2 greatly:1 sigkdd:1 greedily:2 baseline:2 contrast:2 posteriori:2 inference:17 dependent:2 huawei:1 lowercase:1 mrfs:1 entire:1 unlikely:1 typically:2 bt:2 koller:1 labelings:9 schulz:1 pixel:7 pascal:7 yahoo:1 development:1 delong:3 special:1 kempe:1 yip:1 uc:1 marginal:23 field:4 once:1 having:1 never:1 sampling:1 hop:19 manually:1 construct:3 broad:2 biology:2 icml:2 y6:1 park:1 promote:1 foreground:1 future:3 minimized:1 guzman:4 report:5 isack:1 few:1 vmware:1 simultaneously:1 national:1 argmax:1 consisting:1 microsoft:1 detection:1 message:4 highly:1 possibility:1 mining:1 sheep:1 argmaxy:4 bracket:1 pc:1 uppercase:1 light:1 chain:1 accurate:2 edge:6 tuple:1 partial:2 necessary:2 respective:1 tree:4 peach:1 plugged:1 penalizes:1 re:2 theoretical:1 instance:8 cover:1 w911nf:1 measuring:1 maximization:8 clipping:1 cost:10 loopy:1 subset:4 rare:5 veksler:1 virginia:2 front:2 eec:1 learnt:1 combined:5 divmbest:16 person:2 st:1 fundamental:1 international:1 caseiro:1 probabilistic:3 off:2 pool:1 emc:1 iy:1 augmentation:8 ambiguity:2 cisco:1 containing:1 choose:1 huang:1 positivity:1 worse:2 zhao:1 yp:2 dlb:8 potential:16 diversity:65 gy:2 bold:1 sec:1 includes:1 coefficient:1 int:1 explicitly:1 depends:2 performed:1 try:2 start:1 annotation:1 contribution:4 minimize:1 publicly:1 accuracy:11 characteristic:1 efficiently:7 who:2 maximized:1 generalize:2 metaxas:2 produced:1 bilmes:2 apple:1 facebook:1 definition:4 against:1 energy:8 naturally:1 hamming:28 gain:27 sap:1 tunable:1 dataset:3 proved:2 recall:1 subsection:1 ut:1 car:1 knowledge:2 segmentation:45 subtle:1 routine:2 sophisticated:1 goldstein:1 manuscript:1 higher:2 supermodular:1 voc2012:1 reflected:1 specify:1 zisserman:1 daunting:1 wei:1 formulation:3 done:2 dey:1 furthermore:1 marketing:1 stage:1 just:1 eqn:3 web:1 touch:1 overlapping:2 propagation:1 google:2 maximizer:2 mode:4 quality:8 perhaps:1 grows:1 ye:1 concept:1 y2:5 normalized:2 contain:5 hence:5 regularization:1 assigned:2 iteratively:1 semantic:1 adjacent:2 encourages:1 ambiguous:1 outline:1 crf:2 performs:3 reasoning:1 image:21 virdata:1 boykov:3 netapp:1 viral:1 overview:1 exponentially:7 belong:1 interpretation:1 interpret:1 significant:1 refer:2 tsvms:1 feldman:1 smoothness:1 tuning:2 unconstrained:1 grid:1 pm:1 hp:1 inclusion:1 submodular:32 language:3 had:1 pq:1 stacey:1 access:1 supervision:1 etc:1 base:14 add:4 gt:1 posterior:1 closest:1 recent:2 exclusion:1 showed:1 belongs:1 awarded:1 rewarded:1 scenario:3 buchbinder:1 certain:1 n00014:1 binary:2 vt:1 yi:11 scoring:10 guestrin:2 greater:1 additional:1 maximize:5 redundant:1 paradigm:1 ii:2 multiple:4 desirable:1 offer:2 cross:1 retrieval:2 lin:2 e1:1 award:2 impact:1 prediction:10 vision:4 iteration:2 represent:1 robotics:2 achieved:1 proposal:2 addition:2 want:2 separately:1 operative:1 adarsh:2 background:4 winn:1 krause:2 appropriately:2 ssvms:2 fifty:1 envelope:2 subject:1 tend:1 elegant:1 db:2 pooling:1 lafferty:1 spirit:1 near:1 presence:9 unused:2 counting:2 chopra:1 enough:1 easy:1 hb:5 affect:2 identified:1 cow:2 tradeoff:1 bottleneck:1 colour:1 returned:1 c3energy:1 speaking:1 passing:3 proprietary:1 deep:1 ssvm:1 useful:2 generally:1 covered:5 involve:1 dbatra:1 amount:2 zabih:2 processed:1 category:6 simplest:2 generate:3 outperform:3 exist:1 tutorial:2 shifted:1 arising:1 per:2 diverse:15 discrete:2 write:1 group:31 clarity:1 yadollahpour:1 v1:1 graph:18 monotone:16 convert:1 sum:1 letter:4 reporting:1 vn:1 draw:1 summarizes:1 scaling:1 confusing:1 bound:7 followed:1 nonnegative:8 oracle:10 annual:1 jones:1 occur:2 placement:2 constraint:2 scene:1 fmin:4 vondrak:1 kleinberg:1 min:6 kumar:1 performing:3 structured:31 ball:20 combination:3 across:2 smaller:1 slightly:1 y0:24 feige:1 restricted:1 iccv:1 pipeline:1 taken:1 ln:3 mutually:1 previously:2 discus:2 count:3 mechanism:1 turn:1 needed:1 fed:1 optimizable:1 available:1 apply:1 observe:1 v2:1 generic:1 away:1 original:1 thomas:1 assumes:1 running:1 include:2 top:1 graphical:7 hinge:1 somewhere:1 exploit:1 restrictive:1 especially:1 objective:1 move:1 already:3 question:3 occurs:1 mirrokni:1 bagnell:1 said:1 nemhauser:1 distance:3 thank:1 majority:1 decoder:1 carbonell:1 spanning:1 boldface:1 enforcing:1 issn:1 illustration:1 setup:1 unfortunately:2 negative:2 disparate:1 implementation:2 summarization:2 perform:3 upper:1 markov:3 immediate:1 situation:1 extended:2 saturates:1 excluding:1 viola:1 y1:9 discovered:1 gameontalis:1 bk:12 pair:6 dog:1 rutenbar:1 sentence:1 connection:4 learned:2 established:1 nip:6 address:1 beyond:1 yc:2 kulesza:1 sparsity:1 challenge:1 convincingly:1 saturation:1 max:4 tance:1 including:1 belief:1 gool:1 natural:5 ranked:1 predicting:2 cascaded:2 solvable:1 yanover:1 zhu:1 scheme:2 improve:1 yq:1 library:1 irrespective:1 stefanie:1 extract:1 gy0:2 faced:1 text:1 interdependent:1 acknowledgement:1 discovery:1 val:5 relative:1 loss:3 par:1 fully:1 reordering:1 interesting:1 wolsey:1 proportional:3 validation:1 foundation:2 jegelka:4 sufficient:1 xiao:1 principle:1 translation:1 austin:1 row:1 eccv:2 summary:1 supported:2 hebert:1 infeasible:1 bias:1 allow:3 face:1 benefit:1 boundary:1 world:2 transition:13 collection:1 regressors:2 osokin:2 far:1 cope:1 scribble:3 splunk:1 sj:1 approximate:8 alpha:1 mmr:1 social:1 clique:2 instantiation:1 uai:1 conceptual:1 summing:1 unnecessary:1 corpus:1 search:3 streeter:1 table:7 learn:2 transfer:2 robust:4 golovin:1 messagepassing:1 ignoring:1 sminchisescu:1 expansion:2 constructing:3 domain:4 did:2 aistats:3 spread:2 yi2:1 linearly:2 lampert:1 repeated:1 complementary:2 pivotal:1 exploitable:1 augmented:4 fig:2 intel:1 lc:4 sub:4 nonnegativity:1 pereira:1 exponential:1 comput:1 candidate:4 house:1 lie:1 jmlr:2 specific:4 list:12 decay:1 svm:2 ericsson:1 adding:2 supplement:7 texture:1 margin:4 clamping:1 chen:2 smoothly:1 intersection:6 lt:4 simply:5 army:1 intern:1 visual:1 expressed:1 ordered:1 contained:1 partially:1 scalar:1 pretrained:1 corresponds:2 truth:1 relies:1 acm:2 hedge:1 conditional:2 goal:2 sized:1 towards:1 absence:1 feasible:1 hard:1 fisher:1 included:1 specifically:3 carreira:1 reducing:4 uniformly:1 principal:1 lemma:6 batra:6 called:3 experimental:1 m3:1 stefje:1 select:2 formally:2 support:1 scan:1 arises:2 relevance:12 evaluate:1 outgoing:1 |
4,778 | 5,326 | The Noisy Power Method:
A Meta Algorithm with Applications
Moritz Hardt?
IBM Research Almaden
Eric Price?
IBM Research Almaden
Abstract
We provide a new robust convergence analysis of the well-known power method for
computing the dominant singular vectors of a matrix that we call the noisy power
method. Our result characterizes the convergence behavior of the algorithm when
a significant amount noise is introduced after each matrix-vector multiplication.
The noisy power method can be seen as a meta-algorithm that has recently found a
number of important applications in a broad range of machine learning problems
including alternating minimization for matrix completion, streaming principal
component analysis (PCA), and privacy-preserving spectral analysis. Our general
analysis subsumes several existing ad-hoc convergence bounds and resolves a
number of open problems in multiple applications:
Streaming PCA. A recent work of Mitliagkas et al. (NIPS 2013) gives a spaceefficient algorithm for PCA in a streaming model where samples are drawn from a
gaussian spiked covariance model. We give a simpler and more general analysis that
applies to arbitrary distributions confirming experimental evidence of Mitliagkas
et al. Moreover, even in the spiked covariance model our result gives quantitative
improvements in a natural parameter regime. It is also notably simpler and follows
easily from our general convergence analysis of the noisy power method together
with a matrix Chernoff bound.
Private PCA. We provide the first nearly-linear time algorithm for the problem
of differentially private principal component analysis that achieves nearly tight
worst-case error bounds. Complementing our worst-case bounds, we show that the
error dependence of our algorithm on the matrix dimension can be replaced by an
essentially tight dependence on the coherence of the matrix. This result resolves the
main problem left open by Hardt and Roth (STOC 2013). The coherence is always
bounded by the matrix dimension but often substantially smaller thus leading to
strong average-case improvements over the optimal worst-case bound.
1
Introduction
Computing the dominant singular vectors of a matrix is one of the most important algorithmic
tasks underlying many applications including low-rank approximation, PCA, spectral clustering,
dimensionality reduction, matrix completion and topic modeling. The classical problem is wellunderstood, but many recent applications in machine learning face the fundamental problem of
approximately finding singular vectors in the presence of noise. Noise can enter the computation
through a variety of sources including sampling error, missing entries, adversarial corruptions and
privacy constraints. It is desirable to have one robust method for handling a variety of cases without
the need for ad-hoc analyses. In this paper we consider the noisy power method, a fast general purpose
method for computing the dominant singular vectors of a matrix when the target matrix can only be
accessed through inaccurate matrix-vector products.
?
?
Email: [email protected]
Email: [email protected]
1
Figure 1 describes the method when the target matrix A is a symmetric d ? d matrix?a generalization
to asymmetric matrices is straightforward. The algorithm starts from an initial matrix X0 ? Rd?p
and iteratively attempts to perform the update rule X` ? AX` . However, each such matrix product is
followed by a possibly adversarially and adaptively chosen perturbation G` leading to the update rule
X` ? AX` + G` . It will be convenient though not necessary to maintain that X` has orthonormal
columns which can be achieved through a QR-factorization after each update.
Input: Symmetric matrix A ? Rd?d , number of iterations L, dimension p
1. Choose X0 ? Rd?p .
2. For ` = 1 to L:
(a) Y` ? AX`?1 + G` where G` ? Rd?p is some perturbation
(b) Let Y` = X` R` be a QR-factorization of Y`
Output: Matrix XL
Figure 1: Noisy Power Method (NPM)
The noisy power method is a meta algorithm that when instantiated with different settings of G`
and X0 adapts to a variety of applications. In fact, there have been a number of recent surprising
applications of the noisy power method:
1. Jain et al. [JNS13, Har14] observe that the update rule of the well-known alternating least
squares heuristic for matrix completion can be considered as an instance of NPM. This lead
to the first provable convergence bounds for this important heuristic.
2. Mitgliakas et al. [MCJ13] observe that NPM applies to a streaming model of principal
component analysis (PCA) where it leads to a space-efficient and practical algorithm for
PCA in settings where the covariance matrix is too large to process directly.
3. Hardt and Roth [HR13] consider the power method in the context of privacy-preserving
PCA where noise is added to achieve differential privacy.
In each setting there has so far only been an ad-hoc analysis of the noisy power method. In the first
setting, only local convergence is argued, that is, X0 has to be cleverly chosen. In the second setting,
the analysis only holds for the spiked covariance model of PCA. In the third application, only the
case p = 1 was considered.
In this work we give a completely general analysis of the noisy power method that overcomes
limitations of previous analyses. Our result characterizes the global convergence properties of the
algorithm in terms of the noise G` and the initial subspace X0 . We then consider the important
case where X0 is a randomly chosen orthonormal basis. This case is rather delicate since the initial
correlation between a random matrix X0 and the target subspace is vanishing in the dimension d for
small p. Another important feature of the analysis is that it shows how X` converges towards the first
k 6 p singular vectors. Choosing p to be larger than the target dimension leads to a quantitatively
stronger result. Theorem 2.3 formally states our convergence bound. Here we highlight one useful
corollary to illustrate our more general result.
Corollary 1.1. Let k 6 p. Let U ? Rd?k represent the top k singular vectors of A and let
?1 > ? ? ? > ?n > 0 denote its singular values. Suppose X0 is an orthonormal basis of a random
p-dimensional subspace. Further suppose that at every step of NPM we have
5kG` k 6 ?(?k ? ?k+1 ) and
5kU > G` k 6 (?k ? ?k+1 )
? ?
p? k?1
?
? d
??(d)
for some fixed parameter ? and ? < 1/2. Then with all but ? ??(p+1?k) +
probability,
e
there
?k
exists an L = O( ?k ??k+1 log(d? /?)) so that after L steps we have that (I ? XL XL> )U
6 ?.
The corollary shows that the algorithm converges in the strong sense that the entire spectral norm
of U up to an ? error is contained in the space spanned by XL . To achieve this the result places two
assumptions on the magnitude of the noise. The total spectral norm of G` must be bounded by ?
times the separation between ?k and ?k+1 . This dependence on the singular value separation arises
even in the classical perturbation theory of Davis-Kahan [DK70]. The second condition is specific to
the power method and requires that the noise term is proportionally smaller when projected onto the
space spanned by the top k singular vectors. This condition ensures that the correlation between X`
2
and U that is initially very small is not destroyed by the noise addition step. If the noise term has
some sphericalpproperties (e.g. a Gaussian matrix), we expect the projection onto U to be smaller
by a factor of k/d, since the space U is k-dimensional. In the case where p = k + ?(k) this is
precisely what the condition requires. When p = k the requirement is stronger by a factor of k. This
phenomenon stems from the fact that the smallest singular value of a random p ? k gaussian matrix
behaves differently in the square and the rectangular case.
We demonstrate the usefulness of our convergence bound with several novel results in some of the
aforementioned applications.
1.1
Application to memory-efficient streaming PCA
In the streaming PCA setting we receive a stream of samples z1 , z2 , . . . zn ? Rd drawn i.i.d. from
an unknown distribution D over Rd . Our goal is to compute the dominant k eigenvectors of the
covariance matrix A = Ez?D zz > . The challenge is to do this in space linear in the output size,
namely O(kd). Recently, Mitgliakas et al. [MCJ13] gave an algorithm for this problem based on the
noisy power method. We analyze the same algorithm, which we restate here and call SPM:
Input: Stream of samples z1 , z2 , . . . , zn ? Rd , iterations L, dimension p
1. Let X0 ? Rd?p be a random orthonormal basis. Let T = bm/Lc
2. For ` = 1 to L:
P`T
(a) Compute Y` = A` X`?1 where A` = i=(`?1)T +1 zi zi>
(b) Let Y` = X` R` be a QR-factorization of Y`
Output: Matrix XL
Figure 2: Streaming Power Method (SPM)
The algorithm can be executed in space O(pd) since the update step can compute the d ? p matrix
A` X`?1 incrementally without explicitly computing A` . The algorithm maps to our setting by
defining G` = (A` ? A)X`?1 . With this notation Y` = AX`?1 + G` . We can apply Corollary 1.1
directly once we have suitable bounds on kG` k and kU > G` k.
The result of [MCJ13] is specific to the spiked covariance model. The spiked covariance model
is defined by an orthonormal basis U ? Rd?k and a diagonal matrix ? ? Rk?k with diagonal
entries ?1 > ?2 > ? ? ? > ?k > 0. The distribution D(U, ?) is defined as the normal distribution
N(0, (U ?2 U > + ? 2 Idd?d )). Without loss of generality we can scale the examples
such that ?1
= 1.
One corollary of our result shows that the algorithm outputs XL such that
(I ? XL XL> )U
6 ?
with probability 9/10 provided p = k + ?(k) and the number of samples satisfies
6
? +1
n=?
kd .
?2 ?6k
Previously, the same bound1 was known with a quadratic dependence on k in the case where p = k.
Here we can strengthen the bound by increasing p slightly.
While we can get some improvements even in the spiked covariance model, our result is substantially
more general and applies to any distribution. The sample complexity bound we get varies according
to a technical parameter of the distribution. Roughly speaking, we get a near linear sample complexity
if the distribution is either ?round? (as in the spiked covariance setting) or is very well approximated
by a k dimensional subspace. To illustrate the latter condition, we have the following result without
making any assumptions other than scaling the distribution:
Corollary 1.2. Let D be any distribution scaled so that Pr {kzk > t} 6 exp(?t) for every t > 1.
Let U represent the top k eigenvectors of the covariance matrix E zz > and ?1 > ? ? ? > ?d > 0 its
eigenvalues. Then,
SPM invoked
with p = k + ?(k) outputs a matrix XL such with probability
(I ? XL X > )U
6 ? provided SPM receives n samples where n satisfies n =
9/10
we
have
L
?k
? 2
O
?
d
.
3
? k(?k ??k+1 )
1
That the bound stated in [MCJ13] has a ? 6 dependence is not completely obvious. There is a O(? 4 ) in the
numerator and log((? 2 + 0.75?2k )/(? 2 + 0.5?2k )) in the denominator which simplifies to O(1/? 2 ) for constant
?k and ? 2 > 1.
3
The corollary establishes a sample complexity that?s linear in d provided that the spectrum decays
quickly, as is common in applications. For example, if the spectrum follows a power law so that
? 2c+2 d/?2 ).
?j ? j ?c for a constant c > 1/2, the bound becomes n = O(k
1.2
Application to privacy-preserving spectral analysis
Many applications of singular vector computation are plagued by the fact that the underlying matrix
contains sensitive information about individuals. A successful paradigm in privacy-preserving data
analysis rests on the notion of differential privacy which requires all access to the data set to be
randomized in such a way that the presence or absence of a single data item is hidden. The notion of
data item varies and could either refer to a single entry, a single row, or a rank-1 matrix of bounded
norm. More formally, Differential Privacy requires that the output distribution of the algorithm
changes only slightly with the addition or deletion of a single data item. This requirement often
necessitates the introduction of significant levels of noise that make the computation of various
objectives challenging. Differentially private singular vector computation has been studied actively
since the work of Blum et al. [BDMN05]. There are two main objectives. The first is computational
efficiency. The second objective is to minimize the amount of error that the algorithm introduces.
In this work, we give a fast algorithm for differentially private singular vector computation based
on the noisy power method that leads to nearly optimal bounds in a number of settings that were
considered in previous work. The algorithm is described in Figure 3. It?s a simple instance of NPM
in which each noise matrix G` is a gaussian random matrix scaled so that the algorithm achieves
(?, ?)-differential privacy (as formally defined in Definition E.1). It is easy to see that the algorithm
can be implemented in time nearly linear in the number of nonzero entries of the input matrix (input
sparsity). This will later lead to strong improvements in running time compared with several previous
works.
Input: Symmetric A ? Rd?d , L, p, privacy parameters ?, ? > 0
1. Let pX0 be a random orthonormal basis and put ?
??1 4pL log(1/?)
2. For ` = 1 to L:
(a) Y` ? AX`?1 + G` where G` ? N(0, kX`?1 k2? ? 2 )d?p .
(b) Compute the QR-factorization Y` = X` R`
Output: Matrix XL
=
Figure 3: Private Power Method (PPM). Here kXk? = maxij |Xij |.
We first state a general purpose analysis of PPM that follows from Corollary 1.1.
Theorem 1.3. Let k 6 p. Let U ? Rd?k represent the top k singular vectors of A and let
?1 > ? ? ? > ?d > 0 denote its singular values. Then, PPM satisfies (?, ?)-differential privacy and
?k
log(d)) iterations we have with probability 9/10 that
after L = O( ?k ??
k+1
(I ? XL XL> )U
6 O
!
?
?
p
? max kX` k? d log L
?
??
.
?k ? ?k+1
p? k?1
When p = k + ?(k) the trailing factor becomes a constant. If p = k it creates a factor k overhead.
In the worst-case we can always bound kX` k? by 1 since X` is an orthonormal basis. However, in
principle we could hope that a much better bound holds provided that the target subspace U has small
coordinates. Hardt and Roth [HR12, HR13] suggested a way to accomplish a stronger bound by
considering a notion of coherence of A, denoted as ?(A). Informally, the coherence is a well-studied
parameter that varies between 1 and n, but is often observed to be small. Intuitively, the coherence
measures the correlation between the singular vectors of the matrix with the standard basis. Low
coherence means that the singular vectors have small coordinates in the standard basis. Many results
on matrix completion and robust PCA crucially rely on the assumption that the underlying matrix
has low coherence [CR09, CT10, CLMW11] (though the notion of coherence here will be somewhat
different).
4
Theorem 1.4. Under the assumptions of Theorem 1.3, we have the conclusion
!
p
?
p
(I ? XL XL> )U
6 O ? ?(A) log d log L ? ?
?
.
?k ? ?k+1
p? k?1
Hardt
? and Roth proved this result for the case where p = 1. The extension to p > 1 lost a factor
of d in general and therefore gave no improvement over Theorem 1.3. Our result resolves the
main problem left open in their work. The strength of Theorem 1.4 is that the bound is essentially
dimension-free under a natural assumption on the matrix and never worse than our worst-case result.
It is also known that in general the dependence on d achieved in Theorem 1.3 is best possible in the
worst case (see discussion in [HR13]) so that further progress requires making stronger assumptions.
Coherence is a natural such assumption.
The proof of Theorem 1.4 proceeds by showing that each
p
iterate X` satisfies kX` k? 6 O( ?(A) log(d)/d) and applying Theorem 1.3. To do this we exploit
a non-trivial symmetry of the algorithm that we discuss in Section E.3.
Other variants of differential privacy. Our discussion above applied to (?, ?)-differential privacy
under changing a single entry of the matrix. Several works consider other variants of differential
privacy. It is generally easy to adapt the power method to these settings by changing the noise
distribution or its scaling. To illustrate this aspect, we consider the problem of privacy-preserving
principal component analysis as recently studied by [CSS12, KT13]. Both works consider an
algorithm called exponential mechanism. The first work gives a heuristic implementation that may
not converge, while the second work gives a provably polynomial time algorithm though the running
time is more than cubic. Our algorithm gives strong improvements in running time
? while giving
? k) factor. We
nearly optimal accuracy guarantees as it matches a lower bound of [KT13] up to a O(
also improve the error dependence on k by
?polynomial factors compared to previous work. Moreover,
we get an accuracy improvement of O( d) for the case of (?, ?)-differential privacy, while these
previous works only apply to (?, 0)-differential privacy. Section E.2 provides formal statements.
1.3
Related Work
Numerical Analysis. One might expect that a suitable analysis of the noisy power method would
have appeared in the numerical analysis literature. However, we are not aware of a reference and
there are a number of points to consider. First, our noise model is adaptive thus setting it apart from
the classical perturbation theory of the singular vector decomposition [DK70]. Second, we think
of the perturbation at each step as large making it conceptually different from floating point errors.
Third, research in numerical analysis over the past decades has largely focused on faster Krylov
subspace methods. There is some theory of inexact Krylov methods, e.g., [SS07] that captures the
effect of noisy matrix-vector products in this context. Related to our work are also results on the
perturbation stability of the QR-factorization since those could be used to obtain convergence bounds
for subspace iteration. Such bounds, however, must depend on the condition number of the matrix
that the QR-factorization is applied to. See Chapter 19.9 in [Hig02] and the references therein for
background. Our proof strategy avoids this particular dependence on the condition number.
Streaming PCA. PCA in the streaming model is related to a host of well-studied problems that we
cannot survey completely here. We refer to [ACLS12, MCJ13] for a thorough discussion of prior
work. Not mentioned therein is a recent work on incremental PCA [BDF13] that leads to space
efficient algorithms computing the top singular vector; however, it?s not clear how to extend their
results to computing multiple singular vectors.
Privacy. There has been much work on differentially private spectral analysis starting with Blum
et al. [BDMN05] who used an algorithm known as Randomized Response which adds a single
noise matrix N either to the input matrix A or the covariance matrix AA> . This approach appears
in a number of papers, e.g. [MM09]. While often easy to analyze it has the disadvantage that it
converts sparse matrices to dense matrices and is often impractical on large data sets. Chaudhuri
et al. [CSS12] and Kapralov-Talwar [KT13] use the so-called exponential mechanism to sample
approximate eigenvectors of the matrix. The sampling is done using a heuristic approach without
convergence polynomial time convergence guarantees in the first case and using a polynomial time
algorithm in the second. Both papers achieve a tight dependence on the matrix dimension d (though
5
the dependence on k is suboptimal in general). Most closely related to our work are the results of
Hardt and Roth [HR13, HR12] that introduced matrix coherence as a way to circumvent existing
worst-case lower bounds on the error. They also analyzed a natural noisy variant of power iteration
for the case of computing the dominant eigenvector of A. When multiple eigenvectors are needed,
their algorithm uses the well-known deflation technique. However, this step loses
p control of the
coherence of the original matrix and hence results in suboptimal bounds. In fact, a rank(A) factor
is lost.
1.4
Open Questions
We believe Corollary 1.1 to be a fairly precise characterization of the convergence of the noisy power
method to the top k singular vectors when p = k. The main flaw is that the noise tolerance depends
on the eigengap ?k ? ?k+1 , which could be very small. We have some conjectures for results that do
not depend on this eigengap.
First, when p > k, we think that Corollary 1.1 might hold using the gap ?k ? ?p+1 instead of
?k ? ?k+1 . Unfortunately, our proof technique relies on the principal angle decreasing at each step,
which does not necessarily hold with the larger level of noise. Nevertheless we expect the principal
angle to decrease fairly fast on average, so that XL will contain a subspace very close to U . We are
actually unaware of this sort of result even in the noiseless setting.
Conjecture 1.5. Let X0 be a random p-dimensional basis for p > k. Suppose at every step we have
?
?
p? k?1
?
100kG` k 6 ?(?k ? ?p+1 ) and 100kU T G` k 6
d
?k
Then with high probability, after L = O( ?k ??
log(d/?)) iterations we have
p+1
k(I ? XL XL> )U k 6 ?.
The second way of dealing with a small eigengap would be to relax our goal. Corollary 1.1 is quite
stringent in that it requires XL to approximate the top k singular vectors U , which gets harder when
the eigengap approaches zero and the kth through p + 1st singular vectors are nearly indistinguishable.
A relaxed goal would be for XL to spectrally approximate A, that is
k(I ? XL XL> )Ak 6 ?k+1 + ?.
(1)
This weaker goal is known to be achievable in the noiseless setting without any eigengap at all.
In particular, [?] shows that (1) happens after L = O( ?k+1
? log n) steps in the noiseless setting. A
plausible extension to the noisy setting would be:
Conjecture 1.6. Let X0 be a random 2k-dimensional basis. Suppose at every step we have
p
kG` k 6 ? and kU T G` k 6 ? k/d
Then with high probability, after L = O( ?k+1
? log d) iterations we have that
k(I ? XL XL> )Ak 6 ?k+1 + O(?).
1.5
Organization
All proofs can be found in the supplementary material. In the remaining space, we limit ourselves to
a more detailed discussion of our convergence analysis and the application to streaming PCA. The
entire section on privacy is in the supplementary materials in Section E.
2
Convergence of the noisy power method
Figure 1 presents our basic algorithm that we analyze in this section. An important tool in our analysis
are principal angles, which are useful in analyzing the convergence behavior of numerical eigenvalue
methods. Roughly speaking, we will show that the tangent of the k-th principal angle between X and
the top k eigenvectors of A decreases as ?k+1 /?k in each iteration of the noisy power method.
6
Definition 2.1 (Principal angles). Let X and Y be subspaces of Rd of dimension at least k. The
principal angles 0 6 ?1 6 ? ? ? 6 ?k between X and Y and associated principal vectors x1 , . . . , xk
and y1 , . . . , yk are defined recursively via
hx, yi
?i (X , Y) = min arccos
: x ? X , y ? Y, x ? xj , y ? yj for all j < i
kxk2 kyk2
and xi , yi are the x and y that give this value. For matrices X and Y , we use ?k (X, Y ) to denote the
kth principal angle between their ranges.
2.1
Convergence argument
Fix parameters 1 6 k 6 p 6 d. In this section we consider a symmetric d ? d matrix A with singular
values ?1 > ?2 > ? ? ? > ?d . We let U ? Rd?k contain the first k eigenvectors of A. Our main
lemma shows that tan ?k (U, X) decreases multiplicatively in each step.
Lemma 2.2. Let U contain the largest k eigenvectors of a symmetric matrix A ? Rd?d , and let
X ? Rd?p for p > k. Let G ? Rd?p satisfy
4kU > Gk 6 (?k ? ?k+1 ) cos ?k (U, X)
4kGk 6 (?k ? ?k+1 )?.
for some ? < 1. Then
tan ?k (U, AX + G) 6 max ?, max ?,
?k+1
?k
1/4 !
!
tan ?k (U, X) .
We can inductively apply the previous lemma to get the following general convergence result.
Theorem 2.3. Let U represent the top k eigenvectors of the matrix A and ? = 1 ? ?k+1 /?k . Suppose
that the initial subspace X0 and noise G` is such that
5kU > G` k 6 (?k ? ?k+1 ) cos ?k (U, X0 )
5kG` k 6 ?(?k ? ?k+1 )
at every stage `, for some ? < 1/2. Then there exists an L . ?1 log tan ?k ?(U,X0 ) such that for all
` > L we have tan ?(U, XL ) 6 ?.
2.2
Random initialization
The next lemma essentially follows from bounds on the smallest singular value of gaussian random
matrices [RV09].
Lemma 2.4. For an arbitrary orthonormal U and random subspace X, we have
?
d
?
tan ?k (U, X) 6 ? ?
p? k?1
with all but ? ??(p+1?k) + e??(d) probability.
With this lemma we can prove the corollary that we stated in the introduction.
Proof of Corollary 1.1. By Lemma 2.4, with the desired probability ?
we have tan ?k (U, X0 ) 6
?
?
? ?d
.
p? k?1
?
p? k?1
?
.
2?? d
>
XL XL )U k
Hence cos ?k (U, X0 ) > 1/(1 + tan ?k (U, X0 )) >
ply Theorem 2.3 to get that tan ?k (U, XL ) 6 ?. Then k(I ?
tan ?k (U, XL ) 6 ?.
7
Rescale ? and ap= sin ?k (U, XL ) 6
3
Memory efficient streaming PCA
In the streaming PCA setting we receive a stream of samples z1 , z2 , ? ? ? ? Rd . Each sample is drawn
i.i.d. from an unknown distribution D over Rd . Our goal is to compute the dominant k eigenvectors
of the covariance matrix A = Ez?D zz > . The challenge is to do this with small space, so we cannot
store the d2 entries of the sample covariance matrix. We would like to use O(dk) space, which is
necessary even to store the output.
The streaming power method (Figure 2, introduced by [MCJ13]) is a natural algorithm that performs
streaming PCA with O(dk) space. The question that arises is how many samples it requires to
achieve a given level of accuracy, for various distributions D. Using our general analysis of the noisy
power method, we show that the streaming power method requires fewer samples and applies to more
distributions than was previously known. We analyze a broad class of distributions:
Definition 3.1. A distribution D over Rd is (B, p)-round if for
n every p-dimensional
oprojection P and
p
all t > 1 we have Prz?D {kzk > t} 6 exp(?t) and Prz?D kP zk > t ? Bp/d 6 exp(?t) .
The first condition just corresponds to a normalization of the samples drawn from D. Assuming the
first condition holds, the second condition always holds with B = d/p. For this reason our analysis
in principle applies to any distribution, but the sample complexity will depend quadratically on B.
Let us illustrate this definition through the example of the spiked covariance model studied
by [MCJ13]. The spiked covariance model is defined by an orthonormal basis U ? Rd?k and a
diagonal matrix ? ? Rk?k with diagonal entries ?1 > ?2 > ? ? ? > ?k > 0. The distribution D(U,
P ?)
is defined as the normal distribution N(0, (U ?2 U > + ? 2 Idd?d )/D) where D = ?(d? 2 + i ?2i )
is a normalization factor chosen so that the distribution satisfies the norm bound. Note that the the
i-th eigenvalue of the covariance matrix is ?i = (?2i + ? 2 )/D for 1 6 i 6 k and ?i = ? 2 /D for
i > k. We show in Lemma D.2 that the spiked covariance model D(U, ?) is indeed (B, p)-round for
?21 +? 2
B = O( tr(?)/d+?
2 ), which is constant for ? & ?1 . We have the following main theorem.
Theorem 3.2. Let D be a (B, p)-round distribution over Rd with covariance matrix A whose
eigenvalues are ?1 > ?2 > ? ? ? > ?d > 0. Let U ? Rd?k be an orthonormal basis for the
eigenvectors corresponding to the first k eigenvalues of A. Then, the streaming power method SPM
returns an orthonormal basis X ? Rd?p such that tan ?(U, X) 6 ? with probability 9/10 provided
that SPM receives n samples from D for some n satisfying
B 2 ?k k log2 d
?
n6O
?2 (?k ? ?k+1 )3 d
if p = k + ?(k). More generally, for all p > k one can get the slightly stronger result
!
?
?
2
2
2
Bp?
max{1/?
,
Bp/(
p
?
k
?
1)
}
log
d
k
?
n6O
.
(?k ? ?k+1 )3 d
Instantiating with the spiked covariance model gives the following:
Corollary 3.3. In the spiked covariance model D(U, ?) the conclusion of Theorem 3.2 holds for
p = 2k with
2
(?1 + ? 2 )2 (?2k + ? 2 )
?
n=O
dk .
?2 ?6k
? ?6 +1
When ?1 = O(1) and ?k = ?(1) this becomes n = O
?
dk
.
?2
We can apply Theorem 3.2 to all distributions that have exponentially concentrated norm by setting
B = d/p. This gives the following result.
Corollary 3.4. Let D be any distribution scaled such that Prz?D [kzk > t] 6 exp(?t) for all t > 1.
Then the conclusion of Theorem 3.2 holds for p = 2k with
?k
?
n=O
?d .
?2 k(?k ? ?k+1 )3
If the eigenvalues follow a power law, ?j ? j ?c for a constant c > 1/2, this gives an n =
? 2c+2 d/?2 ) bound on the sample complexity.
O(k
8
References
[ACLS12]
Raman Arora, Andrew Cotter, Karen Livescu, and Nathan Srebro. Stochastic optimization for pca and pls. In Communication, Control, and Computing (Allerton), 2012 50th
Annual Allerton Conference on, pages 861?868. IEEE, 2012.
[BDF13]
Akshay Balsubramani, Sanjoy Dasgupta, and Yoav Freund. The fast convergence of
incremental PCA. In Proc. 27th Neural Information Processing Systems (NIPS), pages
3174?3182, 2013.
[BDMN05] Avrim Blum, Cynthia Dwork, Frank McSherry, and Kobbi Nissim. Practical privacy:
the SuLQ framework. In Proc. 24th PODS, pages 128?138. ACM, 2005.
[CLMW11] Emmanuel J. Cand?s, Xiaodong Li, Yi Ma, and John Wright. Robust principal component analysis? J. ACM, 58(3):11, 2011.
[CR09]
Emmanuel J. Cand?s and Benjamin Recht. Exact matrix completion via convex optimization. Foundations of Computional Mathematics, 9:717?772, December 2009.
[CSS12]
Kamalika Chaudhuri, Anand Sarwate, and Kaushik Sinha. Near-optimal differentially
private principal components. In Proc. 26th Neural Information Processing Systems
(NIPS), 2012.
Emmanuel J. Cand?s and Terence Tao. The power of convex relaxation: near-optimal
matrix completion. IEEE Transactions on Information Theory, 56(5):2053?2080, 2010.
[CT10]
[DK70]
Chandler Davis and W. M. Kahan. The rotation of eigenvectors by a perturbation. iii.
SIAM J. Numer. Anal., 7:1?46, 1970.
[Har14]
Moritz Hardt. Understanding alternating minimization for matrix completion. In Proc.
55th Foundations of Computer Science (FOCS). IEEE, 2014.
[Hig02]
Nicholas J. Higham. Accuracy and Stability of Numerical Algorithms. Society for
Industrial and Applied Mathematics, 2002.
[HR12]
Moritz Hardt and Aaron Roth. Beating randomized response on incoherent matrices.
In Proc. 44th Symposium on Theory of Computing (STOC), pages 1255?1268. ACM,
2012.
[HR13]
Moritz Hardt and Aaron Roth. Beyond worst-case analysis in private singular vector
computation. In Proc. 45th Symposium on Theory of Computing (STOC). ACM, 2013.
[JNS13]
Prateek Jain, Praneeth Netrapalli, and Sujay Sanghavi. Low-rank matrix completion
using alternating minimization. In Proc. 45th Symposium on Theory of Computing
(STOC), pages 665?674. ACM, 2013.
[KT13]
Michael Kapralov and Kunal Talwar. On differentially private low rank approximation.
In Proc. 24rd Symposium on Discrete Algorithms (SODA). ACM-SIAM, 2013.
[MCJ13]
Ioannis Mitliagkas, Constantine Caramanis, and Prateek Jain. Memory limited, streaming PCA. In Proc. 27th Neural Information Processing Systems (NIPS), pages 2886?
2894, 2013.
[MM09]
Frank McSherry and Ilya Mironov. Differentially private recommender systems: building privacy into the net. In Proc. 15th KDD, pages 627?636. ACM, 2009.
[RV09]
Mark Rudelson and Roman Vershynin. Smallest singular value of a random rectangular
matrix. Communications on Pure and Applied Mathematics, 62(12):1707?1739, 2009.
[SS07]
Valeria Simoncini and Daniel B. Szyld. Recent computational developments in krylov
subspace methods for linear systems. Numerical Linear Algebra With Applications,
14:1?59, 2007.
9
| 5326 |@word kgk:1 private:10 polynomial:4 stronger:5 norm:5 achievable:1 open:4 d2:1 crucially:1 covariance:20 decomposition:1 tr:1 harder:1 recursively:1 reduction:1 initial:4 contains:1 daniel:1 past:1 existing:2 com:1 z2:3 surprising:1 jns13:2 must:2 john:1 numerical:6 confirming:1 kdd:1 update:5 fewer:1 item:3 complementing:1 xk:1 vanishing:1 provides:1 characterization:1 allerton:2 simpler:2 accessed:1 differential:10 symposium:4 focs:1 prove:1 overhead:1 privacy:21 x0:17 notably:1 indeed:1 roughly:2 cand:3 behavior:2 decreasing:1 resolve:3 considering:1 increasing:1 becomes:3 provided:5 moreover:2 bounded:3 underlying:3 notation:1 what:1 kg:5 prateek:2 substantially:2 eigenvector:1 spectrally:1 finding:1 impractical:1 guarantee:2 quantitative:1 every:6 thorough:1 scaled:3 k2:1 control:2 local:1 limit:1 ak:2 analyzing:1 approximately:1 ap:1 might:2 therein:2 studied:5 initialization:1 challenging:1 co:3 factorization:6 limited:1 range:2 practical:2 yj:1 lost:2 convenient:1 projection:1 get:8 onto:2 cannot:2 close:1 put:1 context:2 applying:1 map:1 roth:7 missing:1 straightforward:1 starting:1 pod:1 convex:2 rectangular:2 focused:1 survey:1 pure:1 mironov:1 rule:3 orthonormal:11 spanned:2 stability:2 notion:4 coordinate:2 target:5 suppose:5 tan:11 strengthen:1 exact:1 us:1 livescu:1 kunal:1 approximated:1 satisfying:1 asymmetric:1 observed:1 capture:1 worst:8 ensures:1 decrease:3 yk:1 mentioned:1 benjamin:1 pd:1 complexity:5 inductively:1 ppm:3 depend:3 tight:3 algebra:1 creates:1 eric:1 efficiency:1 completely:3 basis:13 necessitates:1 easily:1 differently:1 various:2 chapter:1 caramanis:1 instantiated:1 fast:4 jain:3 kp:1 choosing:1 quite:1 heuristic:4 larger:2 plausible:1 supplementary:2 whose:1 relax:1 kahan:2 think:2 noisy:20 hoc:3 eigenvalue:6 net:1 product:3 chaudhuri:2 achieve:4 adapts:1 differentially:7 qr:6 convergence:19 requirement:2 incremental:2 converges:2 simoncini:1 illustrate:4 andrew:1 completion:8 rescale:1 progress:1 strong:4 netrapalli:1 implemented:1 c:1 restate:1 closely:1 stochastic:1 chandler:1 stringent:1 material:2 argued:1 hx:1 fix:1 generalization:1 har14:2 extension:2 pl:1 hold:8 considered:3 wright:1 normal:2 exp:4 plagued:1 algorithmic:1 trailing:1 achieves:2 smallest:3 purpose:2 proc:10 utexas:1 sensitive:1 largest:1 establishes:1 tool:1 cotter:1 minimization:3 hope:1 gaussian:5 always:3 rather:1 corollary:15 ax:6 improvement:7 rank:5 industrial:1 adversarial:1 sense:1 flaw:1 streaming:17 inaccurate:1 entire:2 sulq:1 initially:1 hidden:1 tao:1 provably:1 aforementioned:1 almaden:2 denoted:1 development:1 arccos:1 fairly:2 once:1 never:1 aware:1 sampling:2 chernoff:1 zz:3 adversarially:1 broad:2 nearly:6 sanghavi:1 quantitatively:1 roman:1 randomly:1 individual:1 floating:1 replaced:1 ourselves:1 maintain:1 delicate:1 attempt:1 organization:1 dwork:1 numer:1 introduces:1 analyzed:1 mcsherry:2 necessary:2 desired:1 sinha:1 instance:2 column:1 modeling:1 disadvantage:1 zn:2 yoav:1 entry:7 usefulness:1 successful:1 too:1 varies:3 accomplish:1 vershynin:1 adaptively:1 st:1 recht:1 fundamental:1 randomized:3 siam:2 terence:1 michael:1 together:1 quickly:1 ilya:1 choose:1 possibly:1 worse:1 leading:2 return:1 kobbi:1 actively:1 li:1 ioannis:1 subsumes:1 satisfy:1 explicitly:1 valeria:1 ad:3 stream:3 depends:1 later:1 analyze:4 characterizes:2 px0:1 start:1 kapralov:2 sort:1 spaceefficient:1 minimize:1 square:2 accuracy:4 largely:1 who:1 conceptually:1 corruption:1 email:2 definition:4 inexact:1 obvious:1 proof:5 associated:1 proved:1 hardt:9 dimensionality:1 actually:1 appears:1 follow:1 response:2 done:1 though:4 generality:1 just:1 stage:1 correlation:3 receives:2 incrementally:1 spm:6 believe:1 xiaodong:1 effect:1 building:1 contain:3 hence:2 moritz:4 alternating:4 symmetric:5 iteratively:1 nonzero:1 round:4 indistinguishable:1 numerator:1 kyk2:1 sin:1 kaushik:1 davis:2 demonstrate:1 performs:1 invoked:1 novel:1 recently:3 common:1 rotation:1 behaves:1 exponentially:1 sarwate:1 extend:1 significant:2 refer:2 enter:1 rd:25 sujay:1 mathematics:3 access:1 add:1 dominant:6 recent:5 constantine:1 apart:1 store:2 meta:3 yi:3 seen:1 preserving:5 somewhat:1 relaxed:1 converge:1 paradigm:1 wellunderstood:1 multiple:3 desirable:1 stem:1 technical:1 match:1 adapt:1 faster:1 host:1 instantiating:1 variant:3 basic:1 denominator:1 essentially:3 noiseless:3 iteration:8 represent:4 normalization:2 achieved:2 cr09:2 receive:2 addition:2 background:1 singular:27 source:1 rest:1 december:1 anand:1 idd:2 call:2 near:3 presence:2 iii:1 easy:3 destroyed:1 variety:3 iterate:1 xj:1 gave:2 zi:2 suboptimal:2 simplifies:1 praneeth:1 pca:22 eigengap:5 karen:1 speaking:2 useful:2 generally:2 proportionally:1 eigenvectors:11 informally:1 clear:1 detailed:1 amount:2 concentrated:1 xij:1 discrete:1 dasgupta:1 nevertheless:1 blum:3 drawn:4 changing:2 relaxation:1 convert:1 talwar:2 angle:7 soda:1 place:1 separation:2 raman:1 coherence:11 scaling:2 bound:26 followed:1 quadratic:1 annual:1 strength:1 constraint:1 precisely:1 bp:3 dk70:3 aspect:1 nathan:1 argument:1 min:1 conjecture:3 according:1 kd:2 cleverly:1 smaller:3 describes:1 slightly:3 making:3 happens:1 intuitively:1 spiked:12 pr:1 previously:2 discus:1 mechanism:2 deflation:1 needed:1 apply:4 observe:2 ct10:2 balsubramani:1 spectral:6 nicholas:1 original:1 top:9 clustering:1 running:3 remaining:1 rudelson:1 log2:1 exploit:1 giving:1 emmanuel:3 classical:3 society:1 objective:3 added:1 question:2 strategy:1 dependence:10 diagonal:4 kth:2 subspace:12 topic:1 nissim:1 trivial:1 reason:1 provable:1 assuming:1 multiplicatively:1 executed:1 unfortunately:1 stoc:4 statement:1 frank:2 gk:1 stated:2 implementation:1 mcj13:8 anal:1 unknown:2 perform:1 recommender:1 defining:1 communication:2 precise:1 y1:1 perturbation:7 arbitrary:2 introduced:3 namely:1 z1:3 quadratically:1 deletion:1 nip:4 beyond:1 suggested:1 proceeds:1 krylov:3 computional:1 beating:1 regime:1 sparsity:1 challenge:2 appeared:1 including:3 memory:3 max:4 maxij:1 power:30 suitable:2 natural:5 rely:1 circumvent:1 improve:1 arora:1 incoherent:1 prior:1 literature:1 understanding:1 tangent:1 multiplication:1 law:2 freund:1 loss:1 expect:3 highlight:1 limitation:1 srebro:1 foundation:2 szyld:1 principle:2 ibm:3 row:1 free:1 clmw11:2 formal:1 weaker:1 face:1 akshay:1 sparse:1 tolerance:1 kzk:3 dimension:9 avoids:1 unaware:1 adaptive:1 projected:1 bm:1 mm09:2 far:1 transaction:1 approximate:3 bdmn05:3 overcomes:1 dealing:1 global:1 xi:1 spectrum:2 decade:1 ku:6 zk:1 robust:4 symmetry:1 necessarily:1 main:6 dense:1 noise:17 x1:1 cubic:1 lc:1 exponential:2 xl:30 kxk2:1 ply:1 third:2 theorem:16 rk:2 specific:2 showing:1 cynthia:1 decay:1 dk:4 evidence:1 exists:2 avrim:1 kamalika:1 higham:1 mitliagkas:3 magnitude:1 kx:4 gap:1 npm:5 ez:2 kxk:1 contained:1 pls:1 applies:5 aa:1 corresponds:1 loses:1 satisfies:5 relies:1 acm:7 ma:1 goal:5 towards:1 price:1 absence:1 change:1 principal:14 lemma:8 total:1 called:2 sanjoy:1 experimental:1 aaron:2 formally:3 mark:1 latter:1 arises:2 phenomenon:1 handling:1 |
4,779 | 5,327 | Two-Layer Feature Reduction for Sparse-Group
Lasso via Decomposition of Convex Sets
Jie Wang, Jieping Ye
Computer Science and Engineering
Arizona State University, Tempe, AZ 85287
{jie.wang.ustc, jieping.ye}@asu.edu
Abstract
Sparse-Group Lasso (SGL) has been shown to be a powerful regression technique for simultaneously discovering group and within-group sparse patterns by
using a combination of the `1 and `2 norms. However, in large-scale applications,
the complexity of the regularizers entails great computational challenges. In this
paper, we propose a novel two-layer feature reduction method (TLFre) for SGL
via a decomposition of its dual feasible set. The two-layer reduction is able to
quickly identify the inactive groups and the inactive features, respectively, which
are guaranteed to be absent from the sparse representation and can be removed
from the optimization. Existing feature reduction methods are only applicable
for sparse models with one sparsity-inducing regularizer. To our best knowledge,
TLFre is the first one that is capable of dealing with multiple sparsity-inducing
regularizers. Moreover, TLFre has a very low computational cost and can be integrated with any existing solvers. Experiments on both synthetic and real data sets
show that TLFre improves the efficiency of SGL by orders of magnitude.
1
Introduction
Sparse-Group Lasso (SGL) [5, 16] is a powerful regression technique in identifying important
groups and features simultaneously. To yield sparsity at both group and individual feature levels,
SGL combines the Lasso [18] and group Lasso [28] penalties. In recent years, SGL has found great
success in a wide range of applications, including but not limited to machine learning [20, 27], signal processing [17], bioinformatics [14] etc. Many research efforts have been devoted to developing
efficient solvers for SGL [5, 16, 10, 21]. However, when the feature dimension is extremely high,
the complexity of the SGL regularizers imposes great computational challenges. Therefore, there
is an increasingly urgent need for nontraditional techniques to address the challenges posed by the
massive volume of the data sources.
Recently, El Ghaoui et al. [4] proposed a promising feature reduction method, called SAFE screening, to screen out the so-called inactive features, which have zero coefficients in the solution, from
the optimization. Thus, the size of the data matrix needed for the training phase can be significantly
reduced, which may lead to substantial improvement in the efficiency of solving sparse models.
Inspired by SAFE, various exact and heuristic feature screening methods have been proposed for
many sparse models such as Lasso [25, 11, 19, 26], group Lasso [25, 22, 19], etc. It is worthwhile
to mention that the discarded features by exact feature screening methods such as SAFE [4], DOME
[26] and EDPP [25] are guaranteed to have zero coefficients in the solution. However, heuristic feature screening methods like Strong Rule [19] may mistakenly discard features which have nonzero
coefficients in the solution. More recently, the idea of exact feature screening has been extended
to exact sample screening, which screens out the nonsupport vectors in SVM [13, 23] and LAD
[23]. As a promising data reduction tool, exact feature/sample screening would be of great practical
importance because they can effectively reduce the data size without sacrificing the optimality [12].
1
However, all of the existing feature/sample screening methods are only applicable for the sparse
models with one sparsity-inducing regularizer. In this paper, we propose an exact two-layer feature
screening method, called TLFre, for the SGL problem. The two-layer reduction is able to quickly
identify the inactive groups and the inactive features, respectively, which are guaranteed to have zero
coefficients in the solution. To the best of our knowledge, TLFre is the first screening method which
is capable of dealing with multiple sparsity-inducing regularizers.
We note that most of the existing exact feature screening methods involve an estimation of the dual
optimal solution. The difficulty in developing screening methods for sparse models with multiple
sparsity-inducing regularizers like SGL is that the dual feasible set is the sum of simple convex
sets. Thus, to determine the feasibility of a given point, we need to know if it is decomposable with
respect to the summands, which is itself a nontrivial problem (see Section 2). One of our major
contributions is that we derive an elegant decomposition method of any dual feasible solutions of
SGL via the framework of Fenchel?s duality (see Section 3). Based on the Fenchel?s dual problem
of SGL, we motivate TLFre by an in-depth exploration of its geometric properties and the optimality
conditions. We derive the set of the regularization parameter values corresponding to zero solutions.
To develop TLFre, we need to estimate the upper bounds involving the dual optimal solution. To this
end, we first give an accurate estimation of the dual optimal solution via the normal cones. Then,
we formulate the estimation of the upper bounds via nonconvex optimization problems. We show
that these nonconvex problems admit closed form solutions. Experiments on both synthetic and real
data sets demonstrate that the speedup gained by TLFre in solving SGL can be orders of magnitude.
All proofs are provided in the long version of this paper [24].
Notation: Let k ? k1 , k ? k and k ? k? be the `1 , `2 and `? norms, respectively. Denote by B1n , B n , and
n
the unit `1 , `2 , and `? norm balls in Rn (we omit the superscript if it is clear from the context).
B?
For a set C, let int C be its interior. If C is closed and convex, we define the projection operator as
PC (w) := argminu?C kw ? uk. We denote by IC (?) the indicator function of C, which is 0 on C and
? elsewhere. Let ?0 (Rn ) be the class of proper closed convex functions on Rn . For f ? ?0 (Rn ),
let ?f be its subdifferential. The domain of f is the set dom f := {w : f (w) < ?}. For w ? Rn ,
let [w]i be its ith component. For ? ? R, let sgn(?) = sign(?) if ? 6= 0, and sgn(0) = 0. We define
sign([w]i ), if [w]i 6= 0;
SGN(w) = s ? Rn : [s]i ?
[?1, 1],
if [w]i = 0.
We denote by ?+ = max(?, 0). Then, the shrinkage operator S? (w) : Rn ? Rn with ? ? 0 is
[S? (w)]i = (|[w]i | ? ?)+ sgn([w]i ), i = 1, . . . , n.
2
(1)
Basics and Motivation
In this section, we briefly review some basics of SGL. Let y ? RN be the response vector and
X ? RN ?p be the matrix of features. With the group information available, the SGL problem [5] is
2
XG
XG ?
1
minp
y
?
X
?
ng k?g k + ?2 k?k1 ,
(2)
g g
+ ?1
g=1
g=1
??R 2
where ng is the number of features in the g th group, Xg ? RN ?ng denotes the predictors in that
group with the corresponding coefficient vector ?g , and ?1 , ?2 are positive regularization parameters. Without loss of generality, let ?1 = ?? and ?2 = ? with ? > 0. Then, problem (2) becomes:
2
X
XG
G ?
1
min
y?
Xg ? g
+ ? ?
ng k?g k + k?k1 .
(3)
g=1
g=1
??Rp 2
By the Lagrangian multipliers method [24], the dual problem of SGL is
n
o
2
?
sup 12 kyk2 ? 12
y? ? ?
: XTg ? ? Dg? := ? ng B + B? , g = 1, . . . , G .
(4)
?
It is well-known that the dual feasible set of Lasso is the intersection of closed half spaces (thus a
polytope); for group Lasso, the dual feasible set is the intersection of ellipsoids. Surprisingly, the
geometric properties of these dual feasible sets play fundamentally important roles in most of the
existing screening methods for sparse models with one sparsity-inducing regularizer [23, 11, 25, 4].
When we incorporate multiple sparse-inducing regularizers to the sparse models, problem (4) indicates that the dual feasible set can be much more complicated. Although (4) provides a geometric
2
description of the dual feasible set of SGL, it is not suitable for further analysis. Notice that, even
the feasibility of a given point ? is not easy to determine, since it is nontrivial to tell if XTg ? can
?
be decomposed into b1 + b2 with b1 ? ? ng B and b2 ? B? . Therefore, to develop screening
methods for SGL, it is desirable to gain deeper understanding of the sum of simple convex sets.
In the next section, we analyze the dual feasible set of SGL in depth via the Fenchel?s Duality
Theorem. We show that for each XTg ? ? Dg? , Fenchel?s duality naturally leads to an explicit decom?
position XTg ? = b1 + b2 , with one belonging to ? ng B and the other one belonging to B? . This
lays the foundation of the proposed screening method for SGL.
3
The Fenchel?s Dual Problem of SGL
In Section 3.1, we derive the Fenchel?s dual of SGL via Fenchel?s Duality Theorem. We then
motivate TLFre and sketch our approach in Section 3.2. In Section 3.3, we discuss the geometric
properties of the Fenchel?s dual of SGL and derive the set of (?, ?) leading to zero solutions.
3.1 The Fenchel?s Dual of SGL via Fenchel?s Duality Theorem
To derive the Fenchel?s dual problem of SGL, we need the Fenchel?s Duality Theorem as stated in
Theorem 1. The conjugate of f ? ?0 (Rn ) is the function f ? ? ?0 (Rn ) defined by
f ? (z) = supw hw, zi ? f (w).
Theorem 1. [Fenchel?s Duality Theorem] Let f ? ?0 (RN ), ? ? ?0 (Rp ), and T (?) = y ? X?
be an affine mapping from Rp to RN . Let p? , d? ? [??, ?] be primal and dual values defined,
respectively, by the Fenchel problems:
p? = inf ??Rp f (y ? X?) + ??(?); d? = sup??RN ?f ? (??) ? ??? (XT ?) + ?hy, ?i.
One has p? ? d? . If, furthermore, f and ? satisfy the condition 0 ? int (dom f ? y + Xdom ?),
then the equality holds, i.e., p? = d? , and the supreme is attained in the dual problem if finite.
We omit the proof of Theorem 1 since it is a slight modification of Theorem 3.3.5 in [2].
Let f (w) = 21 kwk2 , and ??(?) be the second term in (3). Then, SGL can be written as
min? f (y ? X?) + ??(?).
To derive the Fenchel?s dual problem of SGL, Theorem 1 implies that we need to find f ? and ?? . It
is well-known that f ? (z) = 21 kzk2 . Therefore, we only need to find ?? , where the concept infimal
convolution is needed. Let h, g ? ?0 (Rn ). The infimal convolution of h and g is defined by
(hg)(?) = inf ? h(?) + g(? ? ?),
and it is exact at a point ? if there exists a ? ? (?) such that (hg)(?) = h(? ? (?)) + g(? ? ? ? (?)).
hg is exact if it is exact at every point of its domain, in which case it is denoted by h g.
PG ?
?
Lemma 2. Let ??
1 (?) = ?
g=1 ng k?g k, ?2 (?) = k?k1 and ?(?) = ?1 (?) + ?2 (?). More?
over, let Cg? = ? ng B ? Rng , g = 1, . . . , G. Then, the following hold:
PG
PG
?
?
(i): (??
(?2 )? (?) = g=1 IB? (?g ),
1 ) (?) =
g=1 ICg (?g ) ,
PG
? ?P
(? )
?
(ii): ?? (?) = ((??
(?2 )? ) (?) = g=1 IB g ??Bn?g g ,
1)
where ?g ? Rng is the sub-vector of ? corresponding to the g th group.
Note that PB? (?g ) admits a closed form solution, i.e., [PB? (?g )]i = sgn ([?g ]i ) min (|[?g ]i | , 1).
Combining Theorem 1 and Lemma 2, the Fenchel?s dual of SGL can be derived as follows.
Theorem 3. For the SGL problem in (3), the following hold:
(i): The Fenchel?s dual of SGL is given by:
?
inf 12 k y? ? ?k2 ? 12 kyk2 :
XTg ? ? PB? (XTg ?)
? ? ng , g = 1, . . . , G .
?
(5)
(ii): Let ? ? (?, ?) and ?? (?, ?) be the optimal solutions of problems (3) and (5), respectively. Then,
??? (?, ?) =y ? X? ? (?, ?),
?
XTg ?? (?, ?) ?? ng ?k?g? (?, ?)k + ?k?g? (?, ?)k1 , g = 1, . . . , G.
3
(6)
(7)
Remark 1. We note that the shrinkage operator can also be expressed by
S? (w) = w ? P?B? (w), ? ? 0.
(8)
Therefore, problem (5) can be written more compactly as
?
inf 21 k y? ? ?k2 ? 12 kyk2 :
S1 (XTg ?)
? ? ng , g = 1, . . . , G .
(9)
?
Remark 2. Eq. (6) and Eq. (7) can be obtained by the Fenchel-Young inequality [2, 24]. They
are the so-called KKT conditions [3] and can also be obtained by the Lagrangian multiplier method
[24]. Moreover, for the SGL problem, its Lagrangian dual in (4) and Fenchel?s dual in (5) are indeed
equivalent to each other [24].
Remark 3. An appealing advantage of the Fenchel?s dual in (5) is that we have a natural decomposition of all points ?g ? Dg? : ?g = PB? (?g ) + S1 (?g )) with PB? (?g ) ? B? and S1 (?g ) ? Cg? . As a
result, this leads to a convenient way to determine the feasibility of any dual variable ? by checking
if S1 (XTg ?) ? Cg? , g = 1, . . . , G.
3.2 Motivation of the Two-Layer Screening Rules
We motive the two-layer screening rules via the KKT condition in Eq. (7). As implied by the name,
there are two layers in our method. The first layer aims to identify the inactive groups, and the
second layer is designed to detect the inactive features for the remaining groups.
by Eq. (7), we have the following cases by noting ?kwk1 = SGN(w) and
o
(n
w
,
if w 6= 0,
kwk
?kwk =
{u : kuk ? 1}, if w = 0.
Case 1. If ?g? (?, ?) 6= 0, we have
( ? [? ? (?,?)]
i
? ng k?g? (?,?)k + sign([?g? (?, ?)]i ), if [?g? (?, ?)]i 6= 0,
T ?
g
[Xg ? (?, ?)]i ?
[?1, 1],
if [?g? (?, ?)]i = 0.
(10)
In view of Eq. (10), we can see that
?
?
? ? (?1 ,?2 )
(a): S1 (XTg ?? (?, ?)) = ? ng k?g? (?1 ,?2 )k and kS1 (XTg ?? (?, ?))k = ? ng ,
g
(b): If [XTg ?? (?, ?]i ? 1 then [?g? (?, ?)]i = 0.
Case 2. If
?g? (?, ?)
(11)
(12)
= 0, we have
?
[XTg ?? (?, ?)]i ? ? ng [ug ]i + [?1, 1], kug k ? 1.
(13)
The first layer (group-level) of TLFre From (11) in Case 1, we have
S1 (XTg ?? (?, ?))
< ??ng ? ?g? (?, ?) = 0.
(R1)
Clearly, (R1) can be used to identify the inactive groups and thus a group-level screening rule.
The second layer (feature-level) of TLFre Let xgi be the ith column of Xg .
[XTg ?? (?, ?)]i = xTgi ?? (?, ?). In view of (12) and (13), we can see that
T ?
xg ? (?, ?) ? 1 ? [?g? (?, ?)]i = 0.
i
We have
(R2)
Different from (R1), (R2) detects the inactive features and thus it is a feature-level screening rule.
However, we cannot directly apply (R1) and (R2) to identify the inactive groups/features because
both need to know ?? (?, ?). Inspired by the SAFE rules [4], we can first estimate a region ?
containing ?? (?, ?). Let XTg ? = {XTg ? : ? ? ?}. Then, (R1) and (R2) can be relaxed as follows:
?
(R1? )
sup?g kS1 (?g )k : ?g ? ?g ? XTg ? < ? ng ? ?g? (?, ?) = 0,
T
sup? xg ? : ? ? ? ? 1 ? [?g? (?, ?)]i = 0.
(R2? )
i
Inspired by (R1? ) and (R2? ), we develop TLFre via the following three steps:
Step 1. Given ? and ?, we estimate a region ? that contains ?? (?, ?).
Step 2. We solve for the supreme values in (R1? ) and (R2? ).
Step 3. By plugging in the supreme values from Step 2, (R1? ) and (R2? ) result in the desired
two-layer screening rules for SGL.
4
3.3
The Set of Parameter Values Leading to Zero Solution
?
For notational convenience, let Fg? = {? : kS1 (XTg ?)k ? ? ng }, g = 1, . . . , G; and thus the
?
feasible set of the Fenchel?s dual of SGL is F = ?g=1,...,G Fg? . In view of problem (5) [or (9)],
we can see that ?? (?, ?) is the projection of y/? on F ? , i.e., ?? (?, ?) = PF ? (y/?). Thus, if
y/? ? F ? , we have ?? (?, ?) = y/?. Moreover, by (R1), we can see that ? ? (?, ?) = 0 if y/? is an
interior point of F ? . Indeed, we have the following stronger result.
?
T
Theorem 4. For the SGL problem, let ??
max = maxg {?g : S1 (Xg y/?g ) = ? ng }. Then,
y
y
?
?
?
?
? ? F ? ? (?, ?) = ? ? ? (?, ?) = 0 ? ? ? ?max .
?g in the definition of ??
max admits a closed form solution [24]. Theorem 4 implies that the optimal
solution ? ? (?, ?) is 0 as long as y/? ? F ? . This geometric property also leads to an explicit
characterization of the set of (?1 , ?2 ) such that the corresponding solution of problem (2) is 0. We
denote by ??? (?1 , ?2 ) the optimal solution of problem (2).
Corollary 5. For the SGL problem in (2), let ?max
(?2 ) = maxg ?1ng kS?2 (XTg y)k. Then,
1
(i): ??? (?1 , ?2 ) = 0 ? ?1 ? ?max (?2 ).
1
(ii): If ?1 ? ?max
:= maxg
1
4
?1 kXT
g yk
ng
or ?2 ? ?max
:= kXT yk? , then ??? (?1 , ?2 ) = 0.
2
The Two-Layer Screening Rules for SGL
We follow the three steps in Section 3.2 to develop TLFre. In Section 4.1, we give an accurate
estimation of ?? (?, ?) via normal cones [15]. Then, we compute the supreme values in (R1? ) and
(R2? ) by solving nonconvex problems in Section 4.2. We present the TLFre rules in Section 4.3.
4.1 Estimation of the Dual Optimal Solution
Because of the geometric property of the dual problem in (5), i.e., ?? (?, ?) = PF ? (y/?), we have
a very useful characterization of the dual optimal solution via the so-called normal cones [15].
Definition 1. [15] For a closed convex set C ? Rn and a point w ? C, the normal cone to C at w is
NC (w) = {v : hv, w0 ? wi ? 0, ?w0 ? C}.
(14)
? ?).
? ?) is known if ?
? = ?? . Thus, we can estimate ?? (?, ?) in terms of ?? (?,
By Theorem 4, ?? (?,
max
?
for
?
(?,
?)
to
be
estimated.
Due to the same reason, we only consider the cases with ? < ??
max
Remark 4. In many applications, the parameter values that perform the best are usually unknown.
To determine appropriate parameter values, commonly used approaches such as cross validation
and stability selection involve solving SGL many times over a grip of parameter values. Thus, given
{?(i) }Ii=1 and ?(1) ? ? ? ? ? ?(J ) , we can fix the value of ? each time and solve SGL by varying the
value of ?. We repeat the process until we solve SGL for all of the parameter values.
? ?) is known with ?
? ? ?? . Let ?g ,
Theorem 6. For the SGL problem in (3), suppose that ?? (?,
max
?
g = 1, . . . , G, be defined by Theorem 4. For any ? ? (0, ?), we define
? ?
?
? < ?? ,
if ?
max
? = y/? ? ? (?, ?),
where X? = argmaxXg ?g ,
n? (?)
T
?
?
X? S1 (X? y/?max ), if ? = ??
max ,
? = y ? ?? (?,
? ?),
v? (?, ?)
?
? ? = v? (?, ?)
? ?
v? (?, ?)
?
?
hv? (?,?),n
? (?)i
?
n? (?).
? 2
kn? (?)k
Then, the following hold:
? ? NF ? (?? (?,
? ?)),
(i): n? (?)
?
? ?) + 1 v? (?, ?))k
?
?
(ii): k?? (?, ?) ? (?? (?,
? 21 kv?
(?, ?)k.
2 ?
? Theorem 6 shows that ?? (?, ?)
? = ?? (?,
? ?) + 1 v? (?, ?).
For notational convenience, let o? (?, ?)
2 ?
?
? centered at o? (?, ?).
?
lies inside the ball of radius 21 kv?
(?, ?)k
4.2 Solving for the supreme values via Nonconvex Optimization
We solve the optimization problems in (R1? ) and (R2? ). To simplify notations, let
? ? 1 kv? (?, ?)k},
?
? = {? : k? ? o? (?, ?)k
?
2
? ? 1 kv? (?, ?)kkX
?
?g = ?g : k?g ? XT o? (?, ?)k
g k2 , g = 1, . . . , G.
g
2
5
?
(15)
(16)
Theorem 6 indicates that ?? (?, ?) ? ?. Moreover, we can see that XTg ? ? ?g , g = 1, . . . , G. To
develop the TLFre rule by (R1? ) and (R2? ), we need to solve the following optimization problems:
? ?) = sup {kS1 (?g )k : ?g ? ?g }, g = 1, . . . , G,
s?g (?, ?;
(17)
?g
?
T
?
t (?, ?; ?) = sup {|x ?| : ? ? ?}, i = 1, . . . , ng , g = 1, . . . , G.
(18)
?
gi
gi
Solving problem (17) We consider the following equivalent problem of (17):
1
1
? ?) 2 = sup
kS1 (?g )k2 : ?g ? ?g .
s? (?, ?;
(19)
We can see that the objective function of problem (19) is continuously differentiable and the feasible
set is a ball. Thus, (19) is a nonconvex problem because we need to maximize a convex function
subject to a convex set. We derive the closed form solutions of problems (17) and (19) as follows.
?
? r = 1 kv? (?, ?)kkX
?
Theorem 7. For problems (17) and (19), let c = XTg o? (?, ?),
g k2 and ?g be
?
2
the set of the optimal solutions.
(i) Suppose that c ?
/ B? , i.e., kck? > 1. Let u = rS1 (c)/kS1 (c)k. Then,
? ?) = kS1 (c)k + r and ?? = {c + u}.
s?g (?, ?;
(20)
g
(ii) Suppose that c is a boundary point of B? , i.e., kck? = 1. Then,
? ?) = r and ?? = {c + u : u ? NB (c), kuk = r} .
s?g (?, ?;
(21)
g
?
?
?
(iii) Suppose that c ? int B? , i.e., kck? < 1. Let i ? I = {i : |[c]i | = kck? }. Then,
? ?) = (kck? + r ? 1) ,
s?g (?, ?;
(22)
+
?
if ?g ? B? ,
??g ,
?
?
?
?
?
?g = {c + r ? sgn([c]i )ei : i ? I } , if ?g 6? B? and c 6= 0,
?
{r ? ei? , ?r ? ei? : i? ? I ? } ,
if ?g 6? B? and c = 0,
2
g
?g
2
where ei is the ith standard basis vector.
Solving problem in (18) Problem (18) can be solved directly via the Cauchy-Schwarz inequality.
? + 1 kv? (?, ?)kkx
? ?) = |xT o? (?, ?)|
?
Theorem 8. For problem (18), we have t?gi (?, ?;
gi k.
gi
?
2
4.3 The Proposed Two-Layer Screening Rules
To develop the two-layer screening rules for SGL, we only need to plug the supreme values
? 2 ; ?1 ) in (R1? ) and (R2? ). We present the TLFre rule as follows.
? 2 ; ?1 ) and t? (?2 , ?
s?g (?2 , ?
gi
Theorem 9. For the SGL problem in (3), suppose that we are given ? and a sequence of parameter
(0)
values ??
> ?(1) > . . . > ?(J ) . For each integer 0 ? j < J , we assume that ? ? (?(j) , ?)
max = ?
? (j)
? (j+1)
(?
, ?(j) ) and s?g (?(j+1) , ?(j) ; ?) be given by Eq. (6), Theorems 6
is known. Let ? (? , ?), v?
and 7, respectively. Then, for g = 1, . . . , G, the following holds
?
s?g (?(j+1) , ?(j) ; ?) < ? ng ? ?g? (?(j+1) , ?) = 0.
(L1 )
For the g?th group that does not pass the rule in (L1 ), we have [?g?? (?(j+1) , ?)]i = 0 if
T y?X? ? (?(j) ,?) 1 ? (j+1) (j) 1 ? (j+1) (j)
+
v
(?
,
?
)
, ? )kkxg?i k ? 1.
xg?i
+ 2 kv? (?
(j)
?
2
?
(L2 )
(L1 ) and (L2 ) are the first layer and second layer screening rules of TLFre, respectively.
5
Experiments
We evaluate TLFre on both synthetic and real data sets. To measure the performance of TLFre, we
compute the rejection ratios of (L1 ) and (L2 ), respectively. Specifically, let m be the number of
features that have 0 coefficients in the solution, G be the index set of groups that are discarded by
(L1 ) and p be the number of inactive
features that are detected by (L2 ). The rejection ratios of (L1 )
P
ng
and (L2 ) are defined by r1 = g?G
and r2 = |p|
m
m , respectively. Moreover, we report the speedup
gained by TLFre, i.e., the ratio of the running time of solver without screening to the running time
of solver with TLFre. The solver used in this paper is from SLEP [9].
To determine appropriate values of ? and ? by cross validation or stability selection, we can run
TLFre with as many parameter values as we need. Given a data set, for illustrative purposes only,
we select seven values of ? from {tan(?) : ? = 5? , 15? , 30? , 45? , 60? , 75? , 85? }. Then, for each
value of ?, we run TLFre along a sequence of 100 values of ? equally spaced on the logarithmic
scale of ?/??
max from 1 to 0.01. Thus, 700 pairs of parameter values of (?, ?) are sampled in total.
6
0.3
0.1
200
400
?2
600
0.01 0.02 0.04
800
Rejection Ratio
Rejection Ratio
0.1 0.2 0.4
?/??max
1
0.7
0.5
0.3
0.1
0.1 0.2 0.4
?/??max
1
0.7
0.5
0.3
0.1 0.2 0.4
?/??max
1
?
(e) ? = tan(45 )
1
0.5
0.3
0.01 0.02 0.04
0.7
0.5
0.3
0.1 0.2 0.4
?/??max
1
1
0.9
0.7
0.5
0.3
0.1
0.1 0.2 0.4
?/??max
1
0.01 0.02 0.04
?
(f) ? = tan(60 )
0.7
(d) ? = tan(30? )
1
0.9
0.01 0.02 0.04
1
0.9
0.1
0.1 0.2 0.4
?/??max
0.1
0.01 0.02 0.04
?
0.3
(c) ? = tan(15? )
1
0.9
0.1
0.01 0.02 0.04
0.5
0.01 0.02 0.04
(b) ? = tan(5? )
(a)
1
0.9
0.7
0.1
Rejection Ratio
0
0
0.5
Rejection Ratio
100
0.7
1
0.9
Rejection Ratio
200
1
0.9
Rejection Ratio
?1
300
Rejection Ratio
?max
1 (?2 )
? = tan(5? )
? = tan(15? )
? = tan(30? )
? = tan(45? )
? = tan(60? )
? = tan(75? )
? = tan(85? )
400
0.1 0.2 0.4
?/??max
1
?
(g) ? = tan(75 )
(h) ? = tan(85 )
Figure 1: Rejection ratios of TLFre on the Synthetic 1 data set.
0.5
0.3
0.1
500
?2
0.01 0.02 0.04
1000
Rejection Ratio
Rejection Ratio
0.7
0.5
0.3
0.1
0.01 0.02 0.04
0.1 0.2 0.4
?/??max
1
?
(e) ? = tan(45 )
1
0.3
0.7
0.5
0.3
1
0.1 0.2 0.4
?/??max
1
0.7
0.5
0.3
0.01 0.02 0.04
?
0.5
0.3
0.01 0.02 0.04
0.1 0.2 0.4
?/??max
1
1
0.9
0.7
0.5
0.3
0.1
0.1 0.2 0.4
?/??max
?
(f) ? = tan(60 )
0.7
(d) ? = tan(30? )
1
0.9
0.1
0.01 0.02 0.04
1
0.9
0.1
0.1 0.2 0.4
?/??max
(c) ? = tan(15? )
1
0.9
0.1
0.1 0.2 0.4
?/??max
0.5
0.01 0.02 0.04
(b) ? = tan(5? )
(a)
1
0.9
0.7
0.1
Rejection Ratio
0
0
Rejection Ratio
0.7
1
0.9
Rejection Ratio
200
1
0.9
Rejection Ratio
?1
400
Rejection Ratio
?max
1 (?2 )
? = tan(5? )
? = tan(15? )
? = tan(30? )
? = tan(45? )
? = tan(60? )
? = tan(75? )
? = tan(85? )
600
(g) ? = tan(75 )
1
0.01 0.02 0.04
0.1 0.2 0.4
?/??max
1
?
(h) ? = tan(85 )
Figure 2: Rejection ratios of TLFre on the Synthetic 2 data set.
5.1 Simulation Studies
We perform experiments on two synthetic data sets that are commonly used in the literature [19, 29].
The true model is y = X? ? + 0.01, ? N (0, 1). We generate two data sets with 250 ? 10000
entries: Synthetic 1 and Synthetic 2. We randomly break the 10000 features into 1000 groups. For
Synthetic 1, the entries of the data matrix X are i.i.d. standard Gaussian with pairwise correlation
zero, i.e., corr(xi , xi ) = 0. For Synthetic 2, the entries of the data matrix X are drawn from i.i.d.
standard Gaussian with pairwise correlation 0.5|i?j| , i.e., corr(xi , xj ) = 0.5|i?j| . To construct ? ? ,
we first randomly select ?1 percent of groups. Then, for each selected group, we randomly select ?2
percent of features. The selected components of ? ? are populated from a standard Gaussian and the
remaining ones are set to 0. We set ?1 = ?2 = 10 for Synthetic 1 and ?1 = ?2 = 20 for Synthetic 2.
The figures in the upper left corner of Fig. 1 and Fig. 2 show the plots of ?max
(?2 ) (see Corollary
1
5) and the sampled parameter values of ? and ? (recall that ?1 = ?? and ?2 = ?). For the other
figures, the blue and red regions represent the rejection ratios of (L1 ) and (L2 ), respectively. We
can see that TLFre is very effective in discarding inactive groups/features; that is, more than 90%
of inactive features can be detected. Moreover, we can observe that the first layer screening (L1 )
becomes more effective with a larger ?. Intuitively, this is because the group Lasso penalty plays a
more important role in enforcing the sparsity with a larger value of ? (recall that ?1 = ??). The top
and middle parts of Table 1 indicate that the speedup gained by TLFre is very significant (up to 30
times) and TLFre is very efficient. Compared to the running time of the solver without screening,
the running time of TLFre is negligible. The running time of TLFre includes that of computing
kXg k2 , g = 1, . . . , G, which can be efficiently computed by the power method [6]. Indeed, this can
be shared for TLFre with different parameter values.
5.2 Experiments on Real Data Set
We perform experiments on the Alzheimer?s Disease Neuroimaging Initiative (ADNI) data set
(http://adni.loni.usc.edu/). The data matrix consists of 747 samples with 426040 single
7
Table 1: Running time (in seconds) for solving SGL along a sequence of 100 tuning parameter
values of ? equally spaced on the logarithmic scale of ?/??
max from 1.0 to 0.01 by (a): the solver [9]
without screening; (b): the solver combined with TLFre. The top and middle parts report the results
of TLFre on Synthetic 1 and Synthetic 2. The bottom part reports the results of TLFre on the ADNI
data set with the GMV data as response.
tan(5? ) tan(15? ) tan(30? ) tan(45? ) tan(60? ) tan(75? ) tan(85? )
?
solver
TLFre
Synthetic 1
TLFre+solver
speedup
298.36
0.77
10.26
29.09
301.74
0.78
12.47
24.19
308.69
0.79
15.73
19.63
307.71
0.79
17.69
17.40
311.33
0.81
19.71
15.79
307.53
0.79
21.95
14.01
291.24
0.77
22.53
12.93
solver
TLFre
Synthetic 2
TLFre+solver
speedup
294.64
0.79
11.05
26.66
294.92
0.80
12.89
22.88
297.29
0.80
16.08
18.49
297.50
0.81
18.90
15.74
297.59
0.81
20.45
14.55
295.51
0.81
21.58
13.69
292.24
0.82
22.80
12.82
30838.29
64.96
386.80
79.73
31096.10
65.00
402.72
77.22
30850.78
64.89
391.63
78.78
30728.27
65.17
385.98
79.61
30572.35
65.05
382.62
79.90
0.5
0.3
0.1
50
100
?2
150
1
0.7
0.5
0.3
0.1
0.1 0.2 0.4
?/??max
(e) ? = tan(45? )
1
0.3
1
0.7
0.5
0.3
0.1 0.2 0.4
?/??max
1
(f) ? = tan(60? )
0.3
0.1 0.2 0.4
?/??max
1
(d) ? = tan(30 )
1
0.9
0.7
0.5
0.3
0.01 0.02 0.04
0.5
?
0.1
0.01 0.02 0.04
0.7
0.01 0.02 0.04
(c) ? = tan(15 )
1
0.9
1
0.9
0.1
0.1 0.2 0.4
?/??max
?
0.1
0.01 0.02 0.04
0.5
0.01 0.02 0.04
(b) ? = tan(5 )
Rejection Ratio
Rejection Ratio
0.1 0.2 0.4
?/??max
?
(a)
0.7
0.1
0.01 0.02 0.04
1
0.9
1
0.9
Rejection Ratio
0.7
Rejection Ratio
0
0
1
0.9
Rejection Ratio
50
Rejection Ratio
?max
1 (?2 )
? = tan(5? )
? = tan(15? )
? = tan(30? )
? = tan(45? )
? = tan(60? )
? = tan(75? )
? = tan(85? )
100
?1
solver
30652.56 30755.63
TLFre
64.08
64.56
TLFre+solver 372.04
383.17
speedup
82.39
80.27
Rejection Ratio
ADNI+GMV
1
0.9
0.7
0.5
0.3
0.1
0.1 0.2 0.4
?/??max
(g) ? = tan(75? )
1
0.01 0.02 0.04
0.1 0.2 0.4
?/??max
1
(h) ? = tan(85? )
Figure 3: Rejection ratios of TLFre on the ADNI data set with grey matter volume as response.
nucleotide polymorphisms (SNPs), which are divided into 94765 groups. The response vector is the
grey matter volume (GMV).
The figure in the upper left corner of Fig. 3 shows the plots of ?max
(?2 ) (see Corollary 5) and the
1
sampled parameter values of ? and ?. The other figures present the rejection ratios of (L1 ) and (L2 )
by blue and red regions, respectively. We can see that almost all of the inactive groups/features are
discarded by TLFre. The rejection ratios of r1 + r2 are very close to 1 in all cases. The bottom part
of Table 1 shows that TLFre leads to a very significant speedup (about 80 times). In other words, the
solver without screening needs about eight and a half hours to solve the 100 SGL problems for each
value of ?. However, combined with TLFre, the solver needs only six to eight minutes. Moreover,
we can observe that the computational cost of TLFre is negligible compared to that of the solver
without screening. This demonstrates the efficiency of TLFre.
6
Conclusion
In this paper, we propose a novel feature reduction method for SGL via decomposition of convex
sets. We also derive the set of parameter values that lead to zero solutions of SGL. To the best
of our knowledge, TLFre is the first method which is applicable to sparse models with multiple
sparsity-inducing regularizers. More importantly, the proposed approach provides novel framework
for developing screening methods for complex sparse models with multiple sparsity-inducing regularizers, e.g., `1 SVM that performs both sample and feature selection, fused Lasso and tree Lasso
with more than two regularizers. Experiments on both synthetic and real data sets demonstrate the
effectiveness and efficiency of TLFre. We plan to generalize the idea of TLFre to `1 SVM, fused
Lasso and tree Lasso, which are expected to consist of multiple layers of screening.
8
References
[1] H. H. Bauschke and P. L. Combettes. Convex Analysis and Monotone Operator Theory in Hilbert Spaces.
Springer, 2011.
[2] J. Borwein and A. Lewis. Convex Analysis and Nonlinear Optimization, Second Edition. Canadian
Mathematical Society, 2006.
[3] S. Boyd and L. Vandenberghe. Convex Optimization. Cambridge University Press, 2004.
[4] L. El Ghaoui, V. Viallon, and T. Rabbani. Safe feature elimination in sparse supervised learning. Pacific
Journal of Optimization, 8:667?698, 2012.
[5] J. Friedman, T. Hastie, and R. Tibshirani. A note on the group lasso and a sparse group lasso. arXiv:1001.0736.
[6] N. Halko, P. Martinsson, and J. Tropp. Finding structure with randomness: Probabilistic algorithms for
constructing approximate matrix decompositions. SIAM Review, 53:217?288, 2011.
[7] J.-B. Hiriart-Urruty. From convex optimization to nonconvex optimization. necessary and sufficient conditions for global optimality. In Nonsmooth optimization and related topics. Springer, 1988.
[8] J.-B. Hiriart-Urruty. A note on the Legendre-Fenchel transform of convex composite functions. In Nonsmooth Mechanics and Analysis. Springer, 2006.
[9] J. Liu, S. Ji, and J. Ye. SLEP: Sparse Learning with Efficient Projections. Arizona State University, 2009.
[10] J. Liu and J. Ye. Moreau-Yosida regularization for grouped tree structure learning. In Advances in neural
information processing systems, 2010.
[11] J. Liu, Z. Zhao, J. Wang, and J. Ye. Safe screening with variational inequalities and its application to
lasso. In International Conference on Machine Learning, 2014.
[12] K. Ogawa, Y. Suzuki, S. Suzumura, and I. Takeuchi. Safe sample screening for Support Vector Machine.
arXiv:1401.6740, 2014.
[13] K. Ogawa, Y. Suzuki, and I. Takeuchi. Safe screening of non-support vectors in pathwise SVM computation. In ICML, 2013.
[14] J. Peng, J. Zhu, A. Bergamaschi, W. Han, D. Noh, J. Pollack, and P. Wang. Regularized multivariate
regression for indentifying master predictors with application to integrative genomics study of breast
cancer. The Annals of Appliced Statistics, 4:53?77, 2010.
[15] A. Ruszczy?nski. Nonlinear Optimization. Princeton University Press, 2006.
[16] N. Simon, J. Friedman., T. Hastie., and R. Tibshirani. A Sparse-Group Lasso. Journal of Computational
and Graphical Statistics, 22:231?245, 2013.
[17] P. Sprechmann, I. Ram??rez, G. Sapiro., and Y. Eldar. C-HiLasso: a collaborative hierarchical sparse
modeling framework. IEEE Transactions on Signal Processing, 59:4183?4198, 2011.
[18] R. Tibshirani. Regression shringkage and selection via the lasso. Journal of the Royal Statistical Society
Series B, 58:267?288, 1996.
[19] R. Tibshirani, J. Bien, J. Friedman, T. Hastie, N. Simon, J. Taylor, and R. Tibshirani. Strong rules for
discarding predictors in lasso-type problems. Journal of the Royal Statistical Society Series B, 74:245?
266, 2012.
[20] M. Vidyasagar. Machine learning methods in the cocomputation biology of cancer. In Proceedings of the
Royal Society A, 2014.
[21] M. Vincent and N. Hansen. Sparse group lasso and high dimensional multinomial classification. Computational Statistics and Data Analysis, 71:771?786, 2014.
[22] J. Wang, J. Jun, and J. Ye. Efficient mixed-norm regularization: Algorithms and safe screening methods.
arXiv:1307.4156v1.
[23] J. Wang, P. Wonka, and J. Ye. Scaling svm and least absolute deviations via exact data reduction. In
International Conference on Machine Learning, 2014.
[24] J. Wang and J. Ye. Two-Layer feature reduction for sparse-group lasso via decomposition of convex sets.
arXiv:1410.4210v1, 2014.
[25] J. Wang, J. Zhou, P. Wonka, and J. Ye. Lasso screening rules via dual polytope projection. In Advances
in neural information processing systems, 2013.
[26] Z. J. Xiang and P. J. Ramadge. Fast lasso screening tests based on correlations. In IEEE ICASSP, 2012.
[27] D. Yogatama and N. Smith. Linguistic structured sparsity in text categorization. In Proceedings of the
Annual Meeting of the Association for Computational Linguistics, 2014.
[28] M. Yuan and Y. Lin. Model selection and estimation in regression with grouped variables. Journal of the
Royal Statistical Society Series B, 68:49?67, 2006.
[29] H. Zou and T. Hastie. Regularization and variable selection via the elastic net. Journal of the Royal
Statistical Society Series B, 67:301?320, 2005.
9
| 5327 |@word briefly:1 version:1 middle:2 norm:4 stronger:1 grey:2 integrative:1 simulation:1 bn:1 decomposition:7 pg:4 mention:1 reduction:10 liu:3 contains:1 series:4 existing:5 written:2 designed:1 plot:2 half:2 asu:1 discovering:1 selected:2 ith:3 smith:1 provides:2 characterization:2 rabbani:1 mathematical:1 along:2 initiative:1 yuan:1 consists:1 combine:1 inside:1 pairwise:2 peng:1 expected:1 indeed:3 mechanic:1 inspired:3 detects:1 decomposed:1 pf:2 solver:17 becomes:2 provided:1 moreover:7 notation:2 finding:1 sapiro:1 every:1 nf:1 k2:6 demonstrates:1 uk:1 unit:1 omit:2 positive:1 negligible:2 engineering:1 tempe:1 k:1 ramadge:1 limited:1 range:1 practical:1 significantly:1 composite:1 boyd:1 projection:4 convenient:1 word:1 cannot:1 interior:2 selection:6 operator:4 close:1 nb:1 context:1 convenience:2 equivalent:2 lagrangian:3 jieping:2 convex:15 formulate:1 xgi:1 decomposable:1 identifying:1 rule:17 importantly:1 vandenberghe:1 stability:2 annals:1 play:2 suppose:5 massive:1 exact:11 tan:50 lay:1 bottom:2 role:2 wang:8 hv:2 solved:1 region:4 removed:1 yk:2 substantial:1 disease:1 yosida:1 complexity:2 dom:2 motivate:2 solving:8 bergamaschi:1 efficiency:4 basis:1 compactly:1 icassp:1 various:1 regularizer:3 sgl:46 fast:1 effective:2 detected:2 tell:1 heuristic:2 posed:1 solve:6 larger:2 statistic:3 gi:6 transform:1 itself:1 superscript:1 advantage:1 kxt:2 differentiable:1 sequence:3 net:1 propose:3 hiriart:2 supreme:6 combining:1 shringkage:1 description:1 inducing:9 kv:7 az:1 ogawa:2 r1:16 categorization:1 derive:8 develop:6 eq:6 strong:2 implies:2 indicate:1 nontraditional:1 safe:9 radius:1 exploration:1 centered:1 sgn:7 elimination:1 fix:1 polymorphism:1 hold:5 ic:1 normal:4 great:4 mapping:1 major:1 purpose:1 estimation:6 applicable:3 hansen:1 schwarz:1 grouped:2 tool:1 decom:1 clearly:1 gaussian:3 aim:1 zhou:1 shrinkage:2 varying:1 corollary:3 linguistic:1 derived:1 improvement:1 notational:2 indicates:2 cg:3 detect:1 el:2 integrated:1 noh:1 dual:33 supw:1 motive:1 denoted:1 eldar:1 classification:1 plan:1 construct:1 ng:25 biology:1 kw:1 icml:1 report:3 nonsmooth:2 fundamentally:1 simplify:1 tlfre:50 randomly:3 dg:3 simultaneously:2 individual:1 usc:1 phase:1 friedman:3 screening:38 pc:1 primal:1 regularizers:9 devoted:1 accurate:2 capable:2 necessary:1 nucleotide:1 tree:3 taylor:1 desired:1 sacrificing:1 pollack:1 fenchel:22 column:1 modeling:1 cost:2 deviation:1 entry:3 predictor:3 slep:2 bauschke:1 kn:1 synthetic:17 combined:2 nski:1 international:2 siam:1 probabilistic:1 quickly:2 continuously:1 fused:2 borwein:1 containing:1 admit:1 corner:2 zhao:1 leading:2 b2:3 includes:1 coefficient:6 int:3 matter:2 satisfy:1 kzk2:1 view:3 break:1 closed:8 analyze:1 sup:7 kwk:2 red:2 complicated:1 simon:2 contribution:1 kxg:1 collaborative:1 takeuchi:2 efficiently:1 yield:1 identify:5 spaced:2 generalize:1 vincent:1 randomness:1 icg:1 definition:2 naturally:1 proof:2 gain:1 sampled:3 recall:2 knowledge:3 improves:1 hilbert:1 attained:1 supervised:1 follow:1 response:4 loni:1 generality:1 furthermore:1 until:1 correlation:3 sketch:1 mistakenly:1 ei:4 tropp:1 nonlinear:2 name:1 ye:9 concept:1 true:1 multiplier:2 regularization:5 equality:1 nonzero:1 edpp:1 kyk2:3 illustrative:1 demonstrate:2 performs:1 l1:9 percent:2 snp:1 variational:1 novel:3 recently:2 ug:1 multinomial:1 ji:1 volume:3 association:1 slight:1 martinsson:1 lad:1 kwk2:1 significant:2 cambridge:1 tuning:1 populated:1 entail:1 han:1 etc:2 summands:1 multivariate:1 recent:1 inf:4 discard:1 nonconvex:6 inequality:3 success:1 kwk1:1 meeting:1 relaxed:1 determine:5 maximize:1 signal:2 ii:6 multiple:7 desirable:1 adni:5 plug:1 cross:2 long:2 lin:1 divided:1 equally:2 plugging:1 feasibility:3 involving:1 regression:5 basic:2 breast:1 arxiv:4 represent:1 subdifferential:1 source:1 subject:1 elegant:1 rs1:1 effectiveness:1 integer:1 alzheimer:1 noting:1 canadian:1 iii:1 easy:1 xj:1 zi:1 hastie:4 lasso:24 reduce:1 idea:2 absent:1 inactive:14 six:1 effort:1 penalty:2 remark:4 jie:2 useful:1 clear:1 involve:2 grip:1 reduced:1 generate:1 http:1 notice:1 sign:3 estimated:1 tibshirani:5 blue:2 kck:5 group:36 pb:5 drawn:1 kuk:2 viallon:1 v1:2 ram:1 monotone:1 year:1 sum:2 cone:4 run:2 powerful:2 master:1 almost:1 infimal:2 scaling:1 bound:2 layer:21 guaranteed:3 arizona:2 annual:1 nontrivial:2 hy:1 optimality:3 extremely:1 min:3 speedup:7 pacific:1 developing:3 structured:1 combination:1 ball:3 belonging:2 conjugate:1 legendre:1 increasingly:1 b1n:1 wi:1 urgent:1 appealing:1 modification:1 s1:8 ks1:7 intuitively:1 ghaoui:2 yogatama:1 discus:1 needed:2 know:2 urruty:2 sprechmann:1 end:1 available:1 apply:1 observe:2 worthwhile:1 eight:2 appropriate:2 hierarchical:1 rp:4 denotes:1 remaining:2 running:6 top:2 linguistics:1 graphical:1 k1:5 society:6 implied:1 objective:1 ruszczy:1 gmv:3 w0:2 seven:1 polytope:2 topic:1 cauchy:1 reason:1 enforcing:1 index:1 ellipsoid:1 ratio:30 nc:1 neuroimaging:1 wonka:2 stated:1 proper:1 unknown:1 perform:3 upper:4 rng:2 convolution:2 discarded:3 finite:1 extended:1 rn:18 pair:1 kkx:3 hour:1 address:1 able:2 usually:1 pattern:1 sparsity:11 challenge:3 bien:1 including:1 max:43 royal:5 power:1 suitable:1 vidyasagar:1 difficulty:1 natural:1 regularized:1 indicator:1 zhu:1 xg:11 jun:1 genomics:1 text:1 review:2 geometric:6 understanding:1 checking:1 l2:7 literature:1 xiang:1 loss:1 mixed:1 kug:1 validation:2 foundation:1 affine:1 sufficient:1 imposes:1 minp:1 cancer:2 elsewhere:1 surprisingly:1 repeat:1 deeper:1 wide:1 absolute:1 sparse:22 fg:2 moreau:1 boundary:1 dimension:1 depth:2 commonly:2 suzuki:2 dome:1 transaction:1 approximate:1 dealing:2 global:1 kkt:2 b1:3 xi:3 maxg:3 table:3 promising:2 elastic:1 complex:1 zou:1 constructing:1 domain:2 motivation:2 edition:1 fig:3 screen:2 combettes:1 sub:1 position:1 explicit:2 lie:1 ib:2 hw:1 young:1 rez:1 theorem:23 minute:1 xt:3 discarding:2 r2:14 svm:5 admits:2 exists:1 consist:1 effectively:1 gained:3 importance:1 corr:2 magnitude:2 rejection:29 intersection:2 logarithmic:2 halko:1 expressed:1 pathwise:1 springer:3 lewis:1 shared:1 feasible:11 specifically:1 lemma:2 argminu:1 called:5 total:1 pas:1 duality:7 select:3 support:2 ustc:1 bioinformatics:1 incorporate:1 evaluate:1 princeton:1 |
4,780 | 5,328 | Median Selection Subset Aggregation for Parallel
Inference
Xiangyu Wang
Dept. of Statistical Science
Duke University
[email protected]
Peichao Peng
Statistics Department
University of Pennsylvania
[email protected]
David B. Dunson
Dept. of Statistical Science
Duke University
[email protected]
Abstract
For massive data sets, efficient computation commonly relies on distributed algorithms that store and process subsets of the data on different machines, minimizing
communication costs. Our focus is on regression and classification problems involving many features. A variety of distributed algorithms have been proposed
in this context, but challenges arise in defining an algorithm with low communication, theoretical guarantees and excellent practical performance in general settings. We propose a MEdian Selection Subset AGgregation Estimator (message)
algorithm, which attempts to solve these problems. The algorithm applies feature
selection in parallel for each subset using Lasso or another method, calculates the
?median? feature inclusion index, estimates coefficients for the selected features in
parallel for each subset, and then averages these estimates. The algorithm is simple, involves very minimal communication, scales efficiently in both sample and
feature size, and has theoretical guarantees. In particular, we show model selection
consistency and coefficient estimation efficiency. Extensive experiments show excellent performance in variable selection, estimation, prediction, and computation
time relative to usual competitors.
1
Introduction
The explosion in both size and velocity of data has brought new challenges to the design of statistical
algorithms. Parallel inference is a promising approach for solving large scale problems. The typical
procedure for parallelization partitions the full data into multiple subsets, stores subsets on different
machines, and then processes subsets simultaneously. Processing on subsets in parallel can lead to
two types of computational gains. The first reduces time for calculations within each iteration of
optimization or sampling algorithms via faster operations; for example, in conducting linear algebra
involved in calculating likelihoods or gradients [1?7]. Although such approaches can lead to substantial reductions in computational bottlenecks for big data, the amount of gain is limited by the
need to communicate across computers at each iteration. It is well known that communication costs
are a major factor driving the efficiency of distributed algorithms, so that it is of critical importance
to limit communication. This motivates the second type of approach, which conducts computations
completely independently on the different subsets, and then combines the results to obtain the final
output. This limits communication to the final combining step, and may lead to simpler and much
faster algorithms. However, a major issue is how to design algorithms that are close to communication free, which can preserve or even improve the statistical accuracy relative to (much slower)
algorithms applied to the entire data set simultaneously. We focus on addressing this challenge in
this article.
There is a recent flurry of research in both Bayesian and frequentist settings focusing on the second
approach [8?14]. Particularly relevant to our approach is the literature on methods for combining
point estimators obtained in parallel for different subsets [8, 9, 13]. Mann et al. [9] suggest using
1
averaging for combining subset estimators, and Zhang et al. [8] prove that such estimators will
achieve the same error rate as the ones obtained from the full set if the number of subsets m is well
chosen. Minsker [13] utilizes the geometric median to combine the estimators, showing robustness
and sharp concentration inequalities. These methods function well in certain scenarios, but might
not be broadly useful. In practice, inference for regression and classification typically contains two
important components: One is variable or feature selection and the other is parameter estimation.
Current combining methods are not designed to produce good results for both tasks.
To obtain a simple and computationally efficient parallel algorithm for feature selection and coefficient estimation, we propose a new combining method, referred to as message. The detailed
algorithm will be fully described in the next section. There are related methods, which were proposed with the very different goal of combining results from different imputed data sets in missing
data contexts [15]. However, these methods are primarily motivated for imputation aggregation, do
not improve computational time, and lack theoretical guarantees. Another related approach is the
bootstrap Lasso (Bolasso) [16], which runs Lasso independently for multiple bootstrap samples, and
then intersects the results to obtain the final model. Asymptotic properties are provided under fixed
number of features (p fixed) and the computational burden is not improved over applying Lasso to
the full data set. Our message algorithm has strong justification in leading to excellent convergence
properties in both feature selection and prediction, while being simple to implement and computationally highly efficient.
The article is organized as follows. In section 2, we describe message in detail. In section 3, we
provide theoretical justifications and show that message can produce better results than full data inferences under certain scenarios. Section 4 evaluates the performance of message via extensive numerical experiments. Section 5 contains a discussion of possible generalizations of the new method
to broader families of models and online learning. All proofs are provided in the supplementary
materials.
2
Parallelized framework
Consider the linear model which has n observations and p features,
Y = X? + ?,
where Y is an n ? 1 response vector, X is an n ? p matrix of features and ? is the observation error,
which is assumed to have mean zero and variance ? 2 . The fundamental idea for communication
efficient parallel inference is to partition the data set into m subsets, each of which contains a small
portion of the data n/m. Separate analysis on each subset will then be carried out and the result will
be aggregated to produce the final output.
As mentioned in the previous section, regression problems usually consist of two stages: feature
selection and parameter estimation. For linear models, there is a rich literature on feature selection
and we only consider two approaches. The risk inflation criterion (RIC), or more generally, the generalized information criterion (GIC) is an l0 -based feature selection technique for high dimensional
data [17?20]. GIC attempts to solve the following optimization problem,
? ? = arg
M
min
M ?{1,2,??? ,p}
kY ? XM ?M k22 + ?|M |? 2
(1)
for some well chosen ?. For ? = 2(log p + log log p) it corresponds to RIC [18], for ? =
(2 log p + log n) it corresponds to extended BIC [19] and ? = log n reduces to the usual BIC.
Konishi and Kitagawa [18] prove the consistency of GIC for high dimensional data under some
regularity conditions.
Lasso [21] is an l1 based feature selection technique, which solves the following problem
1
?? = arg min kY ? X?k22 + ?k?k1
? n
(2)
for some well chosen ?. Lasso transfers the original NP hard l0 -based optimization to a problem
that can be solved in polynomial time. Zhao and Yu [22] prove the selection consistency of Lasso
under the Irrepresentable condition. Based on the model selected by either GIC or Lasso, we could
then apply the ordinary least square (OLS) estimator to find the coefficients.
2
As briefly discussed in the introduction, averaging and median aggregation approaches possess different advantages but also suffer from certain drawbacks. To carefully adapt these features to regression and classification, we propose the median selection subset aggregation (message) algorithm,
which is motivated as follows.
Averaging of sparse regression models leads to an inflated number of features having non-zero coefficients, and hence is not appropriate for model aggregation when feature selection is of interest.
When conducting Bayesian variable selection, the median probability model has been recommended
as selecting the single model that produces the best approximation to model-averaged predictions
under some simplifying assumptions [23]. The median probability model includes those features
having inclusion probabilities greater than 1/2. We can apply this notion to subset-based inference
by including features that are included in a majority of the subset-specific analyses, leading to select(i)
(i)
ing the ?median model?. Let ? (i) = (?1 , ? ? ? , ?p ) denote a vector of feature inclusion indicators
(i)
for the ith subset, with ?j = 1 if feature j is included so that the coefficient ?j on this feature is
(i)
non-zero, with ?j = 0 otherwise. The inclusion indicator vector for the median model M? can be
obtained by
? = arg min
??{0,1}p
m
X
i=1
k? ? ? (i) k1 ,
or equivalently,
(i)
?j = median{?j , i = 1, 2, ? ? ? , m} for j = 1, 2, ? ? ? , p.
If we apply Lasso or GIC to the full data set, in the presence of heavy-tailed observation errors, the
estimated feature inclusion indicator vector will converge to the true inclusion vector at a polynomial
rate. It is shown in the next section that the convergence rate of the inclusion vector for the median
model can be improved to be exponential, leading to substantial gains in not only computational
time but also feature selection performance. The intuition for this gain is that in the heavy-tailed
case, a proportion of the subsets will contain outliers having a sizable influence on feature selection.
By taking the median, we obtain a central model that is not so influenced by these outliers, and hence
can concentrate more rapidly. As large data sets typically contain outliers and data contamination,
this is a substantial practical advantage in terms of performance even putting aside the computational gain. After feature selection, we obtain estimates of the coefficients for each selected feature
by averaging the coefficient estimates from each subset, following the spirit of [8]. The message
algorithm (described in Algorithm 1) only requires each machine to pass the feature indicators to
a central computer, which (essentially instantaneously) calculates the median model, passes back
the corresponding indicator vector to the individual computers, which then pass back coefficient
estimates for averaging. The communication costs are negligible.
3
Theory
In this section, we provide theoretical justification for the message algorithm in the linear model
case. The theory is easily generalized to a much wider range of models and estimation techniques,
as will be discussed in the last section.
Throughout the paper we will assume X = (x1 , ? ? ? , xp ) is an n ? p feature matrix, s = |S| is the
number of non-zero coefficients and ?(A) is the eigenvalue for matrix A. Before we proceed to the
theorems, we enumerate several conditions that are required for establishing the theory. We assume
there exist constants V1 , V2 > 0 such that
A.1 Consistency condition for estimation.
?
?
1 T
n xi xi ? V1 for
?min ( n1 XST XS )
i = 1, 2, ? ? ? , p
? V2
A.2 Conditions on ?, |S| and ?
? E(?2k ) < ? for some k > 0
? s = |S| ? c1 n? for some 0 ? ? < 1
3
Algorithm 1 Message algorithm
Initialization:
1: Input (Y, X), n, p, and m
2:
# n is the sample size, p is the number of features and m is the number of subsets
3: Randomly partition (Y, X) into m subsets (Y (i) , X (i) ) and distribute them on m machines.
Iteration:
4: for i = 1 to m do
5:
? (i) = minM? loss(Y (i) , X (i) ) # ? (i) is the estimated model via Lasso or GIC
6: # Gather all subset models ? (i) to obtain the median model M?
7: for j = 1 to p do
(i)
8:
?j = median{?j , i = 1, 2, ? ? ? , m}
9: # Redistribute the estimated model M? to all subsets
10: for i = 1 to m do
(i)T
(i)
(i)T (i)
# Estimate the coefficients
11:
? (i) = (X? X? )?1 X? Y?
(i)
12: # Gather
Pm all subset estimations ?
13: ?? = i=1 ? (i) /m
14:
? ?
15: return ?,
1??
? mini?S |?i | ? c2 n? 2 for some 0 < ? ? 1
A.3 (Lasso) The strong irrepresentable condition.
? Assuming XS and XS c are the features having non-zero and zero coefficients, respectively, there exists some positive constant vector ? such that
|XSTc XS (XST XS )?1 sign(?S )| < 1 ? ?
A.4 (Generalized information criterion, GIC) The sparse Riesz condition.
? There exist constants ? ? 0 and c > 0 such that ? > cn?? , where
? = inf ?min (X?T X? /n)
|?|?|S|
A.1 is the usual consistency condition for regression. A.2 restricts the behaviors of the three key
terms and is crucial for model selection. These are both usual assumptions. See [19,20,22]. A.3 and
A.4 are specific conditions for model selection consistency for Lasso/GIC. As noted in [22], A.3 is
almost sufficient and necessary for sign consistency. A.4 could be relaxed slightly as shown in [19],
but for simplicity we rely on this version. To ameliorate possible concerns on how realistic these
conditions are, we provide further justifications via Theorem 3 and 4 in the supplementary material.
Theorem 1. (GIC) Assume each subset satisfies A.1, A.2 and A.4, and p ? n? for some ? < k(? ?
?), where ? = max{?/k, 2?}. If ? < ? , 2? < ? and ? in (1) are chosen so that ? = c0 /? 2 (n/m)? ??
?1
for some c0 < cc2 /2, then there exists some constant C0 such that for n ? (2C0 p)(k? ?k?) and
?1
?1
m = ?(4C0 )?(k? ?k?) ? n/p(k? ?k?) ?, the selected model M? follows,
n1??/(k? ?k?)
P (M? = MS ) ? 1 ? exp ?
,
24(4C0 )(k? ?k?)
(i)T
and defining C0? = mini ?min (X?
(i)
X? /ni ), the mean square error follows,
?1
n1??/(k? ?k?)
? 2 V2 s
??1
??1 2
2
+ exp ?
.
(1
+
2C
sV
)k?k
+
C
?
Ek?? ? ?k22 ?
1
2
0
0
n
24(4C0 )(k? ?k?)
Theorem 2. (Lasso) Assume each subset satisfies A.1, A.2 and A.3, and p ? n? for some ? <
??? +1
k(? ? ?). If ? < ? and ? in (2) are chosen so that ? = c0 (n/m) 2 for some c0 < c1 V2 /c2 ,
?1
?1
then there exists some constant C0 such that for n ? (2C0 p)(k? ?k?) and m = ?(4C0 )(k? ?k?) ?
?1
n/p(k? ?k?) ?, the selected model M? follows
n1??/(k? ?k?)
P (M? = MS ) ? 1 ? exp ?
,
24(4C0 )(k? ?k?)
4
and with the same C0? defined in Theorem 1, we have
? 2 V2?1 s
n1??/(k? ?k?)
??1
??1 2
2
2
?
(1 + 2C0 sV1 )k?k2 + C0 ? .
Ek? ? ?k2 ?
+ exp ?
n
24(4C0 )(k? ?k?)
The above two theorems boost the model consistency property from the original polynomial rate
[20, 22] to an exponential rate for heavy-tailed errors. In addition, the mean square error, as shown
in the above equation, preserves almost the same convergence rate as if the full data is employed
and the true model is known. Therefore, we expect a similar or better performance of message with
a significantly lower computation load. Detailed comparisons are demonstrated in Section 4.
4
Experiments
This section assesses the performance of the message algorithm via extensive examples, comparing
the results to
? Full data inference. (denoted as ?full data?)
? Subset averaging. Partition and average the estimates obtained on all subsets. (denoted as
?averaging?)
? Subset median. Partition and take the marginal median of the estimates obtained on all
subsets (denoted as ?median?)
? Bolasso. Run Lasso on multiple bootstrap samples and intersect to select model. Then
estimate the coefficients based on the selected model. (denoted as ?Bolasso?)
The Lasso part of all algorithms will be implemented by the ?glmnet? package [24]. (We did not
use ADMM [25] for Lasso as its actual performance might suffer from certain drawbacks [6] and is
reported to be slower than ?glmnet? [26])
4.1
Synthetic data sets
We use the linear model and the logistic model for (p; s) = (1000; 3) or (10,000; 3) with different
sample size n and different partition number m to evaluate the performance. The feature vector is
drawn from a multivariate normal distribution with correlation ? = 0 or 0.5. Coefficients ? are
chosen as,
?
?i ? (?1)ber(0.4) (8 log n/ n + |N (0, 1)|), i ? S
Since GIC is intractable to implement (NP hard), we combine it with Lasso for variable selection:
Implement Lasso for a set of different ??s and determine the optimal one via GIC. The concrete
setup of models are as follows,
Case 1 Linear model with ? ? N (0, 22 ).
Case 2 Linear model with ? ? t(0, df = 3).
Case 3 Logistic model.
For p = 1, 000, we simulate 200 data sets for each case, and vary the sample size from 2000 to
10,000. For each case, the subset size is fixed to 400, so the number of subsets will be changing
? probability of selecting the
from 5 to 25. In the experiment, we record the mean square error for ?,
true model and computational time, and plot them in Fig 1 - 6. For p = 10,000, we simulate 50
data sets for each case, and let the sample size range from 20,000 to 50,000 with subset size fixed to
2000. Results for p = 10,000 are provided in supplementary materials.
It is clear that message had excellent performance in all of the simulation cases, with low MSE,
high probability of selecting the true model, and low computational time. The other subset-based
methods we considered had similar computational times and also had computational burdens that
effectively did not increase with sample size, while the full data analysis and bootstrap Lasso approach both were substantially slower than the subset methods, with the gap increasing linearly in
sample size. In terms of MSE, the averaging and median approaches both had dramatically worse
5
2000
4000
6000
8000
10000
0.5
1.0
1.5
median
fullset
average
message
bolasso
0.0
0.0
0.0
0.2
seconds
2.0
1.0
0.8
0.6
prob
median
fullset
average
message
bolasso
0.4
0.2
0.3
median
fullset
average
message
bolasso
0.1
value
Computational time
2.5
Probability to select the true model
0.4
Mean square error
2000
4000
Sample size n
6000
8000
10000
2000
4000
Sample size n
6000
8000
10000
Sample size n
Figure 1: Results for case 1 with ? = 0.
Probability to select the true model
2000
4000
6000
8000
10000
3.0
2.0
1.0
median
fullset
average
message
bolasso
0.0
0.2
0.0
0.0
seconds
0.8
0.6
prob
median
fullset
average
message
bolasso
0.4
0.4
0.6
median
fullset
average
message
bolasso
0.2
value
Computational time
1.0
0.8
Mean square error
2000
4000
Sample size n
6000
8000
10000
2000
4000
Sample size n
6000
8000
10000
Sample size n
Figure 2: Results for case 1 with ? = 0.5.
4000
6000
8000
10000
2.5
2.0
1.5
seconds
0.5
0.0
0.0
0.00
2000
median
fullset
average
message
bolasso
1.0
1.0
0.6
0.8
Computational time
0.2
0.04
median
fullset
average
message
bolasso
0.4
prob
0.10
0.06
0.08
Probability to select the true model
median
fullset
average
message
bolasso
0.02
value
Mean square error
2000
4000
Sample size n
6000
8000
10000
2000
4000
Sample size n
6000
8000
10000
Sample size n
Figure 3: Results for case 2 with ? = 0.
2000
4000
6000
Sample size n
8000
10000
2.5
2.0
0.5
1.0
1.5
median
fullset
average
message
bolasso
0.0
seconds
1.0
0.6
0.0
prob
0.4
median
fullset
average
message
bolasso
0.2
0.20
0.00
0.10
value
median
fullset
average
message
bolasso
Computational time
0.8
Probability to select the true model
0.30
Mean square error
2000
4000
6000
8000
10000
Sample size n
Figure 4: Results for case 2 with ? = 0.5.
6
2000
4000
6000
Sample size n
8000
10000
Probability to select the true model
6
median
fullset
average
message
bolasso
4
seconds
0.6
median
fullset
average
message
bolasso
0.4
prob
0.8
8
6
3
4
5
median
fullset
average
message
bolasso
2000
4000
6000
8000
10000
0
0
0.0
1
0.2
2
2
value
Computational time
1.0
Mean square error
2000
4000
Sample size n
6000
8000
10000
2000
4000
Sample size n
6000
8000
10000
Sample size n
Figure 5: Results for case 3 with ? = 0.
Probability to select the true model
Computational time
12
10
8
median
fullset
average
message
bolasso
6
seconds
0.6
prob
median
fullset
average
message
bolasso
2000
4000
6000
8000
Sample size n
10000
2
0
0
0.0
2
0.2
4
4
value
6
median
fullset
average
message
bolasso
0.4
8
0.8
10
1.0
Mean square error
2000
4000
6000
8000
10000
Sample size n
2000
4000
6000
8000
10000
Sample size n
Figure 6: Results for case 3 with ? = 0.5.
performance than message in every case, while bootstrap Lasso was competitive (MSEs were same
order of magnitude with message ranging from effectively identical to having a small but significant
advantage), with both message and bootstrap Lasso clearly outperforming the full data approach. In
terms of feature selection performance, averaging had by far the worst performance, followed by the
full data approach, which was substantially worse than bootstrap Lasso, median and message, with
no clear winner among these three methods. Overall message clearly had by far the best combination
of low MSE, accurate model selection and fast computation.
4.2
Individual household electric power consumption
This data set contains measurements of electric power consumption for every household with a
one-minute sampling rate [27]. The data have been collected over a period of almost 4 years and
contain 2,075,259 measurements. There are 8 predictors, which are converted to 74 predictors due
to re-coding of the categorical variables (date and time). We use the first 2,000,000 samples as the
training set and the remaining 75,259 for testing the prediction accuracy. The data are partitioned
into 200 subsets for parallel inference. We plot the prediction accuracy (mean square error for test
samples) against time for full data, message, averaging and median method in Fig 7. Bolasso is
excluded as it did not produce meaningful results within the time span.
To illustrate details of the performance, we split the time line into two parts: the early stage shows
how all algorithms adapt to a low prediction error and a later stage captures more subtle performance
of faster algorithms (full set inference excluded due to the scale). It can be seen that message
dominates other algorithms in both speed and accuracy.
4.3
HIGGS classification
The HIGGS data have been produced using Monte Carlo simulations from a particle physics model
[28]. They contain 27 predictors that are of interest to physicists wanting to distinguish between two
classes of particles. The sample size is 11,000,000. We use the first 10,000,000 samples for training
a logistic model and the rest to test the classification accuracy. The training set is partitioned into
1,000 subsets for parallel inference. The classification accuracy (probability of correctly predicting
the class of test samples) against computational time is plotted in Fig 8 (Bolasso excluded for the
same reason as above).
7
Mean prediction error (later stage)
0.0
0.0016
value
message
median
average
0.0020
0.4
message
median
average
fullset
0.2
value
0.6
0.0024
0.8
Mean prediction error (earlier stage)
0.060
0.065
0.070
0.075
0.080
0.084
0.086
Time (sec)
0.088
0.090
0.092
0.094
Time (sec)
Figure 7: Results for power consumption data.
0.55
0.60
message
median
average
fullset
0.50
value
0.65
Mean prediction accuracy
0.1
0.2
0.3
0.4
0.5
0.6
0.7
Time (sec)
Figure 8: Results for HIGGS classification.
Message adapts to the prediction bound quickly. Although the classification results are not as good
as the benchmarks listed in [28] (due to the choice of a simple parametric logistic model), our new
algorithm achieves the best performance subject to the constraints of the model class.
5
Discussion and conclusion
In this paper, we proposed a flexible and efficient message algorithm for regression and classification with feature selection. Message essentially eliminates the computational burden attributable to
communication among machines, and is as efficient as other simple subset aggregation methods. By
selecting the median model, message can achieve better accuracy even than feature selection on the
full data, resulting in an improvement also in MSE performance. Extensive simulation experiments
show outstanding performance relative to competitors in terms of computation, feature selection and
prediction.
Although the theory described in Section 3 is mainly concerned with linear models, the algorithm
is applicable in fairly wide situations. Geometric median is a topological concept, which allows the
median model to be obtained in any normed model space. The properties of the median model result
from independence of the subsets and weak consistency on each subset. Once these two conditions
are satisfied, the property shown in Section 3 can be transferred to essentially any model space. The
follow-up averaging step has been proven to be consistent for all M estimators with a proper choice
of the partition number [8].
References
[1] Gonzalo Mateos, Juan Andr?es Bazerque, and Georgios B Giannakis. Distributed sparse linear
regression. Signal Processing, IEEE Transactions on, 58(10):5262?5276, 2010.
[2] Joseph K Bradley, Aapo Kyrola, Danny Bickson, and Carlos Guestrin. Parallel coordinate
descent for l1-regularized loss minimization. arXiv preprint arXiv:1105.5379, 2011.
[3] Chad Scherrer, Ambuj Tewari, Mahantesh Halappanavar, and David Haglin. Feature clustering
for accelerating parallel coordinate descent. In NIPS, pages 28?36, 2012.
8
[4] Alexander Smola and Shravan Narayanamurthy. An architecture for parallel topic models.
Proceedings of the VLDB Endowment, 3(1-2):703?710, 2010.
[5] Feng Yan, Ningyi Xu, and Yuan Qi. Parallel inference for latent dirichlet allocation on graphics
processing units. In NIPS, volume 9, pages 2134?2142, 2009.
[6] Zhimin Peng, Ming Yan, and Wotao Yin. Parallel and distributed sparse optimization. preprint,
2013.
[7] Ofer Dekel, Ran Gilad-Bachrach, Ohad Shamir, and Lin Xiao. Optimal distributed online
prediction using mini-batches. The Journal of Machine Learning Research, 13:165?202, 2012.
[8] Yuchen Zhang, John C Duchi, and Martin J Wainwright. Communication-efficient algorithms
for statistical optimization. In NIPS, volume 4, pages 5?1, 2012.
[9] Gideon Mann, Ryan T McDonald, Mehryar Mohri, Nathan Silberman, and Dan Walker. Efficient large-scale distributed training of conditional maximum entropy models. In NIPS, volume 22, pages 1231?1239, 2009.
[10] Steven L Scott, Alexander W Blocker, Fernando V Bonassi, Hugh A Chipman, Edward I
George, and Robert E McCulloch. Bayes and big data: The consensus monte carlo algorithm.
In EFaBBayes 250 conference, volume 16, 2013.
[11] Willie Neiswanger, Chong Wang, and Eric Xing. Asymptotically exact, embarrassingly parallel MCMC. arXiv preprint arXiv:1311.4780, 2013.
[12] Xiangyu Wang and David B Dunson. Parallelizing MCMC via weierstrass sampler. arXiv
preprint arXiv:1312.4605, 2013.
[13] Stanislav Minsker. Geometric median and robust estimation in banach spaces. arXiv preprint
arXiv:1308.1334, 2013.
[14] Stanislav Minsker, Sanvesh Srivastava, Lizhen Lin, and David B Dunson. Robust and scalable
bayes via a median of subset posterior measures. arXiv preprint arXiv:1403.2660, 2014.
[15] Angela M Wood, Ian R White, and Patrick Royston. How should variable selection be performed with multiply imputed data? Statistics in medicine, 27(17):3227?3246, 2008.
[16] Francis R Bach. Bolasso: model consistent lasso estimation through the bootstrap. In Proceedings of the 25th international conference on Machine learning, pages 33?40. ACM, 2008.
[17] Dean P Foster and Edward I George. The risk inflation criterion for multiple regression. The
Annals of Statistics, pages 1947?1975, 1994.
[18] Sadanori Konishi and Genshiro Kitagawa. Generalised information criteria in model selection.
Biometrika, 83(4):875?890, 1996.
[19] Jiahua Chen and Zehua Chen. Extended bayesian information criteria for model selection with
large model spaces. Biometrika, 95(3):759?771, 2008.
[20] Yongdai Kim, Sunghoon Kwon, and Hosik Choi. Consistent model selection criteria on high
dimensions. The Journal of Machine Learning Research, 98888(1):1037?1057, 2012.
[21] Robert Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal
Statistical Society. Series B (Methodological), pages 267?288, 1996.
[22] Peng Zhao and Bin Yu. On model selection consistency of lasso. The Journal of Machine
Learning Research, 7:2541?2563, 2006.
[23] Maria Maddalena Barbieri and James O Berger. Optimal predictive model selection. Annals
of Statistics, pages 870?897, 2004.
[24] Jerome Friedman, Trevor Hastie, and Rob Tibshirani. Regularization paths for generalized
linear models via coordinate descent. Journal of statistical software, 33(1):1, 2010.
[25] Stephen Boyd, Neal Parikh, Eric Chu, Borja Peleato, and Jonathan Eckstein. Distributed optimization and statistical learning via the alternating direction method of multipliers. FoundaR in Machine Learning, 3(1):1?122, 2011.
tions and Trends
[26] Xingguo Li, Tuo Zhao, Xiaoming Yuan, and Han Liu. An R Package flare for high dimensional
linear regression and precision matrix estimation, 2013.
[27] K. Bache and M. Lichman. UCI machine learning repository, 2013.
[28] Pierre Baldi, Peter Sadowski, and Daniel Whiteson. Deep learning in high-energy physics:
Improving the search for exotic particles. arXiv preprint arXiv:1402.4735, 2014.
9
| 5328 |@word repository:1 version:1 briefly:1 polynomial:3 proportion:1 c0:18 dekel:1 vldb:1 simulation:3 simplifying:1 reduction:1 liu:1 contains:4 series:1 selecting:4 lichman:1 daniel:1 bradley:1 current:1 com:1 comparing:1 danny:1 chu:1 john:1 numerical:1 partition:7 realistic:1 designed:1 plot:2 bickson:1 aside:1 selected:6 flare:1 ith:1 record:1 weierstrass:1 simpler:1 zhang:2 c2:2 yuan:2 prove:3 combine:3 dan:1 baldi:1 peng:3 behavior:1 ming:1 actual:1 increasing:1 provided:3 exotic:1 mcculloch:1 substantially:2 guarantee:3 every:2 biometrika:2 k2:2 unit:1 before:1 negligible:1 positive:1 generalised:1 limit:2 physicist:1 minsker:3 establishing:1 barbieri:1 path:1 might:2 initialization:1 limited:1 range:2 ms:1 averaged:1 practical:2 testing:1 practice:1 implement:3 bootstrap:8 procedure:1 intersect:1 yan:2 significantly:1 boyd:1 suggest:1 close:1 selection:34 irrepresentable:2 context:2 applying:1 risk:2 influence:1 dean:1 demonstrated:1 missing:1 independently:2 normed:1 bachrach:1 simplicity:1 estimator:7 konishi:2 notion:1 coordinate:3 justification:4 annals:2 shamir:1 massive:1 exact:1 duke:4 velocity:1 trend:1 particularly:1 bache:1 xw56:1 steven:1 preprint:7 wang:3 solved:1 worst:1 capture:1 contamination:1 ran:1 substantial:3 mentioned:1 intuition:1 flurry:1 solving:1 algebra:1 predictive:1 efficiency:2 eric:2 completely:1 easily:1 intersects:1 fast:1 describe:1 monte:2 supplementary:3 solve:2 otherwise:1 statistic:4 final:4 online:2 advantage:3 eigenvalue:1 propose:3 relevant:1 combining:6 uci:1 rapidly:1 date:1 achieve:2 adapts:1 ky:2 convergence:3 regularity:1 produce:5 wider:1 illustrate:1 tions:1 stat:2 solves:1 sizable:1 implemented:1 strong:2 involves:1 edward:2 riesz:1 inflated:1 concentrate:1 direction:1 drawback:2 material:3 mann:2 bin:1 generalization:1 ryan:1 kitagawa:2 inflation:2 considered:1 normal:1 exp:4 driving:1 major:2 vary:1 early:1 achieves:1 estimation:11 applicable:1 instantaneously:1 minimization:1 brought:1 clearly:2 shrinkage:1 broader:1 l0:2 focus:2 improvement:1 methodological:1 kyrola:1 likelihood:1 mainly:1 maria:1 kim:1 inference:11 entire:1 typically:2 issue:1 classification:9 arg:3 among:2 denoted:4 overall:1 yahoo:1 flexible:1 scherrer:1 fairly:1 marginal:1 once:1 having:5 sampling:2 identical:1 yu:2 np:2 primarily:1 kwon:1 randomly:1 simultaneously:2 preserve:2 individual:2 n1:5 attempt:2 friedman:1 interest:2 message:45 highly:1 multiply:1 chong:1 redistribute:1 halappanavar:1 accurate:1 explosion:1 necessary:1 haglin:1 ohad:1 conduct:1 yuchen:1 re:1 plotted:1 zhimin:1 theoretical:5 minimal:1 earlier:1 ordinary:1 cost:3 addressing:1 subset:43 predictor:3 graphic:1 reported:1 sv:1 synthetic:1 sv1:1 fundamental:1 international:1 hugh:1 physic:2 quickly:1 concrete:1 central:2 satisfied:1 juan:1 worse:2 ek:2 zhao:3 leading:3 return:1 li:1 distribute:1 converted:1 coding:1 sec:3 includes:1 coefficient:14 later:2 higgs:3 performed:1 shravan:1 francis:1 portion:1 competitive:1 aggregation:7 carlos:1 parallel:16 bayes:2 xing:1 ass:1 square:11 ni:1 accuracy:8 variance:1 conducting:2 efficiently:1 ningyi:1 weak:1 bayesian:3 produced:1 carlo:2 minm:1 influenced:1 trevor:1 mahantesh:1 competitor:2 evaluates:1 against:2 energy:1 involved:1 james:1 proof:1 gain:5 gic:11 organized:1 subtle:1 embarrassingly:1 carefully:1 back:2 focusing:1 follow:1 response:1 improved:2 stage:5 smola:1 correlation:1 jerome:1 chad:1 lack:1 bonassi:1 logistic:4 k22:3 contain:4 true:10 concept:1 multiplier:1 willie:1 hence:2 regularization:1 excluded:3 alternating:1 neal:1 white:1 noted:1 criterion:7 generalized:4 m:2 mcdonald:1 duchi:1 l1:2 ranging:1 parikh:1 ols:1 winner:1 volume:4 banach:1 discussed:2 lizhen:1 significant:1 measurement:2 consistency:10 pm:1 narayanamurthy:1 inclusion:7 particle:3 gonzalo:1 had:6 han:1 patrick:1 multivariate:1 posterior:1 recent:1 inf:1 scenario:2 store:2 certain:4 inequality:1 outperforming:1 seen:1 guestrin:1 greater:1 relaxed:1 george:2 employed:1 parallelized:1 xiangyu:2 fernando:1 aggregated:1 converge:1 signal:1 recommended:1 determine:1 period:1 full:14 multiple:4 reduces:2 stephen:1 ing:1 borja:1 faster:3 adapt:2 calculation:1 bach:1 dept:2 lin:2 calculates:2 prediction:12 involving:1 regression:11 aapo:1 qi:1 essentially:3 scalable:1 df:1 arxiv:12 iteration:3 gilad:1 c1:2 addition:1 xst:2 median:49 walker:1 crucial:1 parallelization:1 rest:1 eliminates:1 posse:1 pass:1 subject:1 spirit:1 chipman:1 presence:1 split:1 concerned:1 variety:1 independence:1 bic:2 pennsylvania:1 lasso:25 architecture:1 hastie:1 idea:1 cn:1 bottleneck:1 motivated:2 accelerating:1 suffer:2 peter:1 proceed:1 deep:1 enumerate:1 useful:1 generally:1 detailed:2 clear:2 dramatically:1 listed:1 tewari:1 amount:1 imputed:2 exist:2 restricts:1 andr:1 sign:2 estimated:3 correctly:1 tibshirani:2 broadly:1 bolasso:24 putting:1 key:1 drawn:1 imputation:1 changing:1 v1:2 asymptotically:1 blocker:1 year:1 wood:1 run:2 package:2 prob:6 communicate:1 ameliorate:1 family:1 throughout:1 almost:3 utilizes:1 ric:2 bound:1 followed:1 distinguish:1 topological:1 constraint:1 software:1 nathan:1 simulate:2 speed:1 min:6 span:1 martin:1 xingguo:1 xiaoming:1 transferred:1 department:1 combination:1 across:1 slightly:1 giannakis:1 partitioned:2 joseph:1 rob:1 outlier:3 computationally:2 equation:1 neiswanger:1 ofer:1 operation:1 apply:3 v2:5 appropriate:1 pierre:1 frequentist:1 batch:1 robustness:1 slower:3 original:2 angela:1 remaining:1 clustering:1 dirichlet:1 household:2 calculating:1 medicine:1 k1:2 society:1 silberman:1 feng:1 parametric:1 concentration:1 sanvesh:1 usual:4 gradient:1 separate:1 majority:1 consumption:3 topic:1 collected:1 consensus:1 reason:1 assuming:1 index:1 berger:1 mini:3 minimizing:1 equivalently:1 setup:1 dunson:4 robert:2 design:2 motivates:1 proper:1 wotao:1 observation:3 benchmark:1 descent:3 defining:2 extended:2 communication:11 situation:1 sharp:1 cc2:1 parallelizing:1 peleato:1 tuo:1 david:4 eckstein:1 required:1 extensive:4 boost:1 nip:4 usually:1 xm:1 scott:1 gideon:1 challenge:3 ambuj:1 including:1 max:1 royal:1 wainwright:1 power:3 critical:1 rely:1 regularized:1 predicting:1 indicator:5 wanting:1 improve:2 mateos:1 carried:1 categorical:1 literature:2 geometric:3 relative:3 asymptotic:1 georgios:1 fully:1 loss:2 expect:1 stanislav:2 allocation:1 proven:1 foundar:1 gather:2 sufficient:1 xp:1 consistent:3 article:2 xiao:1 foster:1 heavy:3 endowment:1 efabbayes:1 mohri:1 last:1 free:1 ber:1 wide:1 taking:1 sparse:4 distributed:8 dimension:1 rich:1 commonly:1 far:2 transaction:1 assumed:1 xi:2 search:1 latent:1 tailed:3 promising:1 transfer:1 robust:2 improving:1 whiteson:1 mse:4 excellent:4 mehryar:1 electric:2 did:3 linearly:1 big:2 arise:1 x1:1 xu:1 fig:3 referred:1 attributable:1 zehua:1 precision:1 exponential:2 ian:1 theorem:6 minute:1 choi:1 load:1 specific:2 sadowski:1 showing:1 x:5 concern:1 dominates:1 burden:3 consist:1 exists:3 intractable:1 effectively:2 importance:1 magnitude:1 gap:1 jiahua:1 chen:2 entropy:1 yin:1 bazerque:1 glmnet:2 applies:1 corresponds:2 satisfies:2 relies:1 acm:1 conditional:1 goal:1 admm:1 hard:2 included:2 typical:1 averaging:11 sampler:1 pas:2 e:1 meaningful:1 select:8 jonathan:1 alexander:2 outstanding:1 evaluate:1 mcmc:2 srivastava:1 |
4,781 | 5,329 | Asymmetric LSH (ALSH) for Sublinear Time
Maximum Inner Product Search (MIPS)
Ping Li
Department of Statistics and Biostatistics
Department of Computer Science
Rutgers University
Piscataway, NJ 08854, USA
[email protected]
Anshumali Shrivastava
Department of Computer Science
Computing and Information Science
Cornell University
Ithaca, NY 14853, USA
[email protected]
Abstract
We present the first provably sublinear time hashing algorithm for approximate
Maximum Inner Product Search (MIPS). Searching with (un-normalized) inner
product as the underlying similarity measure is a known difficult problem and
finding hashing schemes for MIPS was considered hard. While the existing Locality Sensitive Hashing (LSH) framework is insufficient for solving MIPS, in this
paper we extend the LSH framework to allow asymmetric hashing schemes. Our
proposal is based on a key observation that the problem of finding maximum inner products, after independent asymmetric transformations, can be converted into
the problem of approximate near neighbor search in classical settings. This key
observation makes efficient sublinear hashing scheme for MIPS possible. Under
the extended asymmetric LSH (ALSH) framework, this paper provides an example of explicit construction of provably fast hashing scheme for MIPS. Our proposed algorithm is simple and easy to implement. The proposed hashing scheme
leads to significant computational savings over the two popular conventional LSH
schemes: (i) Sign Random Projection (SRP) and (ii) hashing based on p-stable
distributions for L2 norm (L2LSH), in the collaborative filtering task of item recommendations on Netflix and Movielens (10M) datasets.
1
Introduction and Motivation
The focus of this paper is on the problem of Maximum Inner Product Search (MIPS). In this problem,
we are given a giant data vector collection S of size N , where S ? RD , and a given query point
q ? RD . We are interested in searching for p ? S which maximizes (or approximately maximizes)
the inner product q T p. Formally, we are interested in efficiently computing
p = arg max q T x
x?S
(1)
The MIPS problem is related to near neighbor search (NNS), which instead requires computing
p = arg min ??q ? x??22 = arg min(??x??22 ? 2q T x)
x?S
x?S
(2)
These two problems are equivalent if the norm of every element x ? S is constant. Note that the
value of the norm ??q??2 has no effect as it is a constant and does not change the identity of arg max
or arg min. There are many scenarios in which MIPS arises naturally at places where the norms of
the elements in S have significant variations [13] and cannot be controlled, e.g., (i) recommender
system, (ii) large-scale object detection with DPM, and (iii) multi-class label prediction.
Recommender systems: Recommender systems are often based on collaborative filtering which
relies on past behavior of users, e.g., past purchases and ratings. Latent factor modeling based on
matrix factorization [14] is a popular approach for solving collaborative filtering. In a typical matrix
factorization model, a user i is associated with a latent user characteristic vector ui , and similarly,
an item j is associated with a latent item characteristic vector vj . The rating ri,j of item j by user i
is modeled as the inner product between the corresponding characteristic vectors.
1
In this setting, given a user i and the corresponding learned latent vector ui finding the right item j,
to recommend to this user, involves computing
j = arg max
ri,j ? = arg max
uTi vj ?
?
?
j
j
(3)
which is an instance of the standard MIPS problem. It should be noted that we do not have control
over the norm of the learned vector, i.e., ?vj ?2 , which often has a wide range in practice [13].
If there are N items to recommend, solving (3) requires computing N inner products. Recommendation systems are typically deployed in on-line application over web where the number N is huge.
A brute force linear scan over all items, for computing arg max, would be prohibitively expensive.
Large-scale object detection with DPM: Deformable Part Model (DPM) based representation of
images is the state-of-the-art in object detection tasks [8]. In DPM model, firstly a set of part filters
are learned from the training dataset. During detection, these learned filter activations over various
patches of the test image are used to score the test image. The activation of a filter on an image patch
is an inner product between them. Typically, the number of possible filters are large (e.g., millions)
and so scoring the test image is costly. Recently, it was shown that scoring based only on filters with
high activations performs well in practice [7]. Identifying those filters having high activations on a
given image patch requires computing top inner products. Consequently, an efficient solution to the
MIPS problem will benefit large scale object detections based on DPM.
Multi-class (and/or multi-label) prediction: The models for multi-class SVM (or logistic regression) learn a weight vector wi for each of the class label i. After the weights are learned, given a
new test data vector xtest , predicting its class label is basically an MIPS problem:
ytest = arg max xTtest wi
i?L
(4)
where L is the set of possible class labels. Note that the norms of the vectors ?wi ?2 are not constant.
The size, ?L?, of the set of class labels differs in applications. Classifying with large number of possible class labels is common in multi-label learning and fine grained object classification, for instance,
prediction task with ?L? = 100, 000 [7]. Computing such high-dimensional vector multiplications for
predicting the class label of a single instance can be expensive in, e.g., user-facing applications.
1.1
The Need for Hashing Inner Products
Solving the MIPS problem can have significant practical impact. [19, 13] proposed solutions based
on tree data structure combined with branch and bound space partitioning technique similar to k-d
trees [9]. Later, the same method was generalized for general max kernel search [5], where the runtime guarantees, like other space partitioning methods, are heavily dependent on the dimensionality
and the expansion constants. In fact, it is well-known that techniques based on space partitioning
(such as k-d trees) suffer from the curse of dimensionality. For example, [24] showed that techniques
based on space partitioning degrade to linear search, even for dimensions as small as 10 or 20.
Locality Sensitive Hashing (LSH) [12] based randomized techniques are common and successful
in industrial practice for efficiently solving NNS (near neighbor search). Unlike space partitioning
techniques, both the running time as well as the accuracy guarantee of LSH based NNS are in a way
independent of the dimensionality of the data. This makes LSH suitable for large scale processing
system dealing with ultra-high dimensional datasets which are common in modern applications.
Furthermore, LSH based schemes are massively parallelizable, which makes them ideal for modern
?Big? datasets. The prime focus of this paper will be on efficient hashing based algorithms for
MIPS, which do not suffer from the curse of dimensionality.
1.2
Our Contributions
We develop Asymmetric LSH (ALSH), an extended LSH scheme for efficiently solving the approximate MIPS problem. Finding hashing based algorithms for MIPS was considered hard [19, 13]. We
formally show that, under the current framework of LSH, there cannot exist any LSH for solving
MIPS. Despite this negative result, we show that it is possible to relax the current LSH framework to
allow asymmetric hash functions which can efficiently solve MIPS. This generalization comes with
no extra cost and the ALSH framework inherits all the theoretical guarantees of LSH.
Our construction of asymmetric LSH is based on an interesting fact that the original MIPS problem,
after asymmetric transformations, reduces to the problem of approximate near neighbor search in
2
classical settings. Based on this key observation, we provide an example of explicit construction of
asymmetric hash function, leading to the first provably sublinear query time hashing algorithm for
approximate similarity search with (un-normalized) inner product as the similarity. The new ALSH
framework is of independent theoretical interest. We report other explicit constructions in [22, 21].
We also provide experimental evaluations on the task of recommending top-ranked items with collaborative filtering, on Netflix and Movielens (10M) datasets. The evaluations not only support our
theoretical findings but also quantify the obtained benefit of the proposed scheme, in a useful task.
2
2.1
Background
Locality Sensitive Hashing (LSH)
A commonly adopted formalism for approximate near-neighbor search is the following:
Definition: (c-Approximate Near Neighbor or c-NN) Given a set of points in a D-dimensional space
RD , and parameters S0 > 0, ? > 0, construct a data structure which, given any query point q, does
the following with probability 1 ? ?: if there exists an S0 -near neighbor of q in P , it reports some
cS0 -near neighbor of q in P .
In the definition, the S0 -near neighbor of point q is a point p with Sim(q, p) ? S0 , where Sim is the
similarity of interest. Popular techniques for c-NN are often based on Locality Sensitive Hashing
(LSH) [12], which is a family of functions with the nice property that more similar objects in the
domain of these functions have a higher probability of colliding in the range space than less similar
ones. In formal terms, consider H a family of hash functions mapping RD to a set I.
Definition: (Locality Sensitive Hashing (LSH)) A family H is called (S0 , cS0 , p1 , p2 )-sensitive if,
for any two point x, y ? RD , h chosen uniformly from H satisfies the following:
? if Sim(x, y) ? S0 then P rH (h(x) = h(y)) ? p1
? if Sim(x, y) ? cS0 then P rH (h(x) = h(y)) ? p2
For efficient approximate nearest neighbor search, p1 > p2 and c < 1 is needed.
Fact 1 [12]: Given a family of (S0 , cS0 , p1 , p2 ) -sensitive hash functions, one can construct a data
log p1
structure for c-NN with O(n? log n) query time and space O(n1+? ), where ? = log
< 1.
p2
2.2
LSH for L2 Distance (L2LSH)
[6] presented a novel LSH family for all Lp (p ? (0, 2]) distances. In particular, when p = 2, this
scheme provides an LSH family for L2 distances. Formally, given a fixed (real) number r, we choose
a random vector a with each component generated from i.i.d. normal, i.e., ai ? N (0, 1), and a scalar
b generated uniformly at random from [0, r]. The hash function is defined as:
hL2
a,b (x) = ?
aT x + b
?
r
(5)
where ?? is the floor operation. The collision probability under this scheme can be shown to be
2
2
L2
P r(hL2
(1 ? e?(r/d) /2 ) (6)
Fr (d) = 1 ? 2?(?r/d) ? ?
a,b (x) = ha,b (y)) = Fr (d);
2?(r/d)
where ?(x) = ??? ?12? e? 2 dx is the cumulative density function (cdf) of standard normal distribution and d = ??x ? y??2 is the Euclidean distance between the vectors x and y. This collision
probability Fr (d) is a monotonically decreasing function of the distance d and hence hL2
a,b is an LSH
for L2 distances. This ?
scheme is also the part of LSH package [1]. Here r is a parameter. As argued
previously, ??x ? y??2 = (??x??22 + ??y??22 ? 2xT y) is not monotonic in the inner product xT y unless the
given data has a constant norm. Hence, hL2
a,b is not suitable for MIPS.
x
x2
The recent work on coding for random projections [16] showed that L2LSH can be improved when
the data are normalized for building large-scale linear classifiers as well as near neighbor search [17].
In particular, [17] showed that 1-bit coding (i.e., sign random projections (SRP) [10, 3]) or 2-bit
coding are often better compared to using more bits. It is known that SRP is designed for retrieving
T
y
. Again, ordering under this similarity can be very
with cosine similarity: Sim(x, y) = ??x??x2 ??y??
2
different from the ordering of inner product and hence SRP is also unsuitable for solving MIPS.
3
3
3.1
Hashing for MIPS
A Negative Result
We first show that, under the current LSH framework, it is impossible to obtain a locality sensitive
hashing scheme for MIPS. In [19, 13], the authors also argued that finding locality sensitive hashing
for inner products could be hard, but to the best of our knowledge we have not seen a formal proof.
Theorem 1 There cannot exist any LSH family for MIPS.
Proof: Suppose there exists such hash function h. For un-normalized inner products the self similarity of a point x with itself is Sim(x, x) = xT x = ??x??22 and there may exist another points y, such that
Sim(x, y) = y T x > ??x??22 + C, for any constant C. Under any single randomized hash function h,
the collision probability of the event {h(x) = h(x)} is always 1. So if h is an LSH for inner product
then the event {h(x) = h(y)} should have higher probability compared to the event {h(x) = h(x)},
since we can always choose y with Sim(x, y) = S0 + ? > S0 and cS0 > Sim(x, x) ?S0 and c < 1.
This is not possible because the probability cannot be greater than 1. This completes the proof. ?
3.2
Our Proposal: Asymmetric LSH (ALSH)
The basic idea of LSH is probabilistic bucketing and it is more general than the requirement of
having a single hash function h. The classical LSH algorithms use the same hash function h for both
the preprocessing step and the query step. One assigns buckets in the hash table to all the candidates
x ? S using h, then uses the same h on the query q to identify relevant buckets. The only requirement
for the proof of Fact 1, to work is that the collision probability of the event {h(q) = h(x)} increases
with the similarity Sim(q, x). The theory [11] behind LSH still works if we use hash function h1
for preprocessing x ? S and a different hash function h2 for querying, as long as the probability of
the event {h2 (q) = h1 (x)} increases with Sim(q, x), and there exist p1 and p2 with the required
property. The traditional LSH definition does not allow this asymmetry but it is not a required
condition in the proof. For this reason, we can relax the definition of c-NN without losing runtime
guarantees. [20] used a related (asymmetric) idea for solving 3-way similarity search.
We first define a modified locality sensitive hashing in a form which will be useful later.
Definition: (Asymmetric Locality Sensitive Hashing (ALSH)) A family H, along with the two
?
?
vector functions Q ? RD ? RD (Query Transformation) and P ? RD ? RD (Preprocessing
Transformation), is called (S0 , cS0 , p1 , p2 )-sensitive if, for a given c-NN instance with query q and
any x in the collection S, the hash function h chosen uniformly from H satisfies the following:
? if Sim(q, x) ? S0 then P rH (h(Q(q))) = h(P (x))) ? p1
? if Sim(q, x) ? cS0 then P rH (h(Q(q)) = h(P (x))) ? p2
When Q(x) = P (x) = x, we recover the vanilla LSH definition with h(.) as the required hash
function. Coming back to the problem of MIPS, if Q and P are different, the event {h(Q(x)) =
h(P (x))} will not have probability equal to 1 in general. Thus, Q ? P can counter the fact that self
similarity is not highest with inner products. We just need the probability of the new collision event
{h(Q(q)) = h(P (y))} to satisfy the conditions in the definition of c-NN for Sim(q, y) = q T y. Note
that the query transformation Q is only applied on the query and the pre-processing transformation
P is applied to x ? S while creating hash tables. It is this asymmetry which will allow us to solve
MIPS efficiently. In Section 3.3, we explicitly show a construction (and hence the existence) of
asymmetric locality sensitive hash function for solving MIPS. The source of randomization h for
both q and x ? S is the same. Formally, it is not difficult to show a result analogous to Fact 1.
Theorem 2 Given a family of hash function H and the associated query and preprocessing transformations P and Q, which is (S0 , cS0 , p1 , p2 ) -sensitive, one can construct a data structure for
log p1
c-NN with O(n? log n) query time and space O(n1+? ), where ? = log
.
p2
3.3
From MIPS to Near Neighbor Search (NNS)
Without loss of any generality, let U < 1 be a number such that ??xi ??2 ? U < 1, ?xi ? S. If this is
not the case then define a scaling transformation,
U
? x;
M = maxxi ?S ??xi ??2 ;
(7)
S(x) =
M
4
Note that we are allowed one time preprocessing and asymmetry, S is the part of asymmetric transformation. For simplicity of arguments, let us assume that ??q??2 = 1, the arg max is anyway independent of the norm of the query. Later we show in Section 3.6 that it can be easily removed.
We are now ready to describe the key step in our algorithm. First, we define two vector transformations P ? RD ? RD+m and Q ? RD ? RD+m as follows:
m
P (x) = [x; ??x??22 ; ??x??42 ; ....; ??x??22 ];
Q(x) = [x; 1/2; 1/2; ....; 1/2],
(8)
i
where [;] is the concatenation. P (x) appends m scalers of the form ??x??22 at the end of the vector x,
while Q(x) simply appends m ?1/2? to the end of the vector x. By observing that
m
m+1
1
Q(q)T P (xi ) = q T xi + (??xi ??22 + ??xi ??42 + ... + ??xi ??22 ); ??P (xi )??22 = ??xi ??22 + ??xi ??42 + ... + ??xi ??22
2
we obtain the following key equality:
??Q(q) ? P (xi )??22 = (1 + m/4) ? 2q T xi + ??xi ??22
m+1
(9)
m+1
Since ??xi ??2 ? U < 1, ??xi ??2
? 0, at the tower rate (exponential to exponential). The term
(1 + m/4) is a fixed constant. As long as m is not too small (e.g., m ? 3 would suffice), we have
arg max q T x ? arg min ??Q(q) ? P (x)??2
x?S
(10)
x?S
This gives us the connection between solving un-normalized MIPS and approximate near neighbor
search. Transformations P and Q, when norms are less than 1, provide correction to the L2 distance
??Q(q) ? P (xi )??2 making it rank correlate with the (un-normalized) inner product. This works only
m+1
after shrinking the norms, as norms greater than 1 will instead blow the term ??xi ??22 .
3.4
Fast Algorithms for MIPS
Eq. (10) shows that MIPS reduces to the standard approximate near neighbor search problem which
m+1
m+1
can be efficiently solved. As the error term ??xi ??22
< U2
goes to zero at a tower rate, it quickly
becomes negligible for any practical purposes. In fact, from theoretical perspective, since we are
interested in guarantees for c-approximate solutions, this additional error can be absorbed in the
approximation parameter c. Formally, we can state the following theorem.
Theorem 3 Given a c-approximate instance of MIPS, i.e., Sim(q, x) = q T x, and a query q such
that ??q??2 = 1 along with a collection S having ??x??2 ? U < 1 ?x ? S. Let P and Q be the vector
transformations defined in (8). We have the following two conditions for hash function hL2
a,b (5)
?
m+1
T
L2
L2
2
)
1) if q x ? S0 then P r[ha,b (Q(q)) = ha,b (P (x))] ? Fr ( 1 + m/4 ? 2S0 + U
?
L2
(
)
2) if q T x ? cS0 then P r[hL2
(Q(q))
=
h
(P
(x))]
?
F
1
+
m/4
?
2cS
r
0
a,b
a,b
where the function Fr is defined in (6).
?
?
Thus, we have obtained p1 = Fr ( (1 + m/4) ? 2S0 + U 2m+1 ) and p2 = Fr ( (1 + m/4) ? 2cS0 ).
Applying Theorem 2, we can construct data structures with worst case O(n? log n) query time
guarantees for c-approximate MIPS, where
?
log Fr ( 1 + m/4 ? 2S0 + U 2m+1 )
(11)
?=
?
log Fr ( 1 + m/4 ? 2cS0 )
We need p1 > p2 in order for ? < 1. This requires us to have ?2S0 + U 2
m+1
U2
2S0
m+1
< ?2cS0 , which boils
m+1
U2
2S0
down to the condition c < 1 ?
. Note that
can be made arbitrarily close to zero with the
appropriate value of m. For any given c < 1, there always exist U < 1 and m such that ? < 1. This
way, we obtain a sublinear query time algorithm for MIPS.
We also have one more parameter r for the hash function ha,b . Recall the definition of Fr in Eq. (6):
2
2
Fr (d) = 1 ? 2?(?r/d) ? ?2?(r/d)
(1 ? e?(r/d) /2 ). Thus, given a c-approximate MIPS instance, ?
5
1
1
0.9
0.9
S0 = 0.5U
0.6
0.7
0.7
0.6
0.6
0.7
0.7
0.6
0.8
0.5
0.8
0.5
0.4
0.3
1
S0 = 0.5U
0.8
?
?*
0.8
m=3,U=0.83, r=2.5
0.4
S0 = 0.9U
0.8
0.6
0.4
0.2
0.3
1
0
c
S0 = 0.9U
0.8
0.6
0.4
0.2
0
c
?
Figure 1: Left panel: Optimal values of ? with respect to approximation ratio c for different S0 .
The optimization of Eq. (14) was conducted by a grid search over parameters r, U and m, given S0
and c. Right Panel: ? values (dashed curves) for m = 3, U = 0.83 and r = 2.5. The solid curves are
?? values. See more details about parameter recommendations in arXiv:1405.5869.
is a function of 3 parameters: U , m, r. The algorithm with the best query time chooses U , m and r,
which minimizes the value of ?. For convenience, we define
?
m+1
log Fr ( 1 + m/4 ? 2S0 + U 2m+1 )
U2
?
?
? = min
s.t.
< 1 ? c, m ? N+ , 0 < U < 1. (12)
U,m,r
2S
0
log Fr ( 1 + m/4 ? 2cS0 )
See Figure 1 for the plots of ?? . With this best value of ?, we can state our main result in Theorem 4.
Theorem 4 (Approximate MIPS is Efficient) For the problem of c-approximate MIPS with ??q??2 =
?
?
1, one can construct a data structure having O(n? log n) query time and space O(n1+? ), where
?
? < 1 is the solution to constraint optimization (14).
3.5
Practical Recommendation of Parameters
Just like in the typical LSH framework, the value of ?? in Theorem 4 depends on the c-approximate
instance we aim to solve, which requires knowing the similarity threshold S0 and the approximation
ratio c. Since, ??q??2 = 1 and ??x??2 ? U < 1, ?x ? S, we have q t x ? U . A reasonable choice of the
threshold S0 is to choose a high fraction of U, for example, S0 = 0.9U or S0 = 0.8U .
The computation of ?? and the optimal values of corresponding parameters can be conducted via a
grid search over the possible values of U , m and r. We compute ?? in Figure 1 (Left Panel). For
convenience, we recommend m = 3, U = 0.83, and r = 2.5. With this choice of the parameters,
Figure 1 (Right Panel) shows that the ? values using these parameters are very close to ?? values.
3.6
Removing the Condition ??q??2 = 1
Changing norms of the query does not affect the arg maxx?C q T x. Thus in practice for retrieving topranked items, normalizing the query should not affect the performance. But for theoretical purposes,
we want the runtime guarantee to be independent of ??q??2 . We are interested in the c-approximate
instance which being a threshold based approximation changes if the query is normalized.
Previously, transformations P and Q were precisely meant to remove the dependency on the norms
of x. Realizing the fact that we are allowed asymmetry, we can use the same idea to get rid of the
norm of q. Let M be the upper bound on all the norms or the radius of the space as defined in
Eq (7). Let the transformation S ? RD ? RD be the ones defined in Eq (7). Define asymmetric
transformations P ? ? RD ? RD+2m and Q? ? RD ? RD+2m as
P ? (x) = [x; ??x??22 ; ??x??42 ; ....; ??x??22 ; 1/2; ...1/2]; Q? (x) = [x; 1/2; ....; 1/2; ??x??22 ; ??x??42 ; ....; ??x??22 ],
m
m
Given the query q and data point x, our new asymmetric transformations are Q? (S(q)) and
P ? (S(x)) respectively. We observe that
??Q? (S(q)) ? P ? (S(x))??22 =
m+1
Both ??S(x)??22
m+1
, ??S(q)??22
m+1
? U2
m+1
m+1
U2
m
+ ??S(x)??22
+ ??S(q)??22
? 2q t x ? ( 2 )
2
M
(13)
? 0. Using exactly same arguments as before, we obtain
6
Theorem 5 (Unconditional Approximate MIPS is Efficient) For the problem of c-approximate
?
MIPS in a bounded space, one can construct a data structure having O(n?u log n) query time and
?
space O(n1+?u ), where ??u < 1 is the solution to constraint optimization (14).
?
m+1
U2
2m+1 )
(
m/2 ? 2S0 ( M
log
F
r
2 ) + 2U
U (2 ?2) M 2
?
< 1 ? c,
(14)
s.t.
?u =
min
?
2
0<U <1,m?N,r
S0
log F ( m/2 ? 2cS ( U ))
r
0
M2
Again, for any c-approximate MIPS instance, with S0 and c, we can always choose m big enough
such that ??u < 1. The theoretical guarantee only depends on the radius of the space M .
3.7
A Generic Recipe for Constructing Asymmetric LSHs
We are allowed any asymmetric transformation on x and q. This gives us a lot of flexibility to construct ALSH for new similarities S that we are interested in. The generic idea is to take a particular
similarity Sim(x, q) for which we know an existing LSH or ALSH. Then we construct transformations P and Q such Sim(P (x), Q(q)) is monotonic in the similarity S that we are interested in.
The other observation that makes it easier to construct P and Q is that LSH based guarantees are
independent of dimensions, thus we can expand the dimensions like we did for P and Q.
This paper focuses on using L2LSH to convert near neighbor search of L2 distance into an ALSH
(i.e., L2-ALSH) for MIPS. We can devise new ALSHs for MIPS using other similarities and hash
functions. For instance, utilizing sign random projections (SRP), the known LSH for correlations,
we can construct different P and Q leading to a better ALSH (i.e., Sign-ALSH) for MIPS [22]. We
are aware another work [18] which performs very similarly to Sign-ALSH. Utilizing minwise hashing [2, 15], which is the LSH for resemblance and is known to outperform SRP in sparse data [23],
we can construct an even better ALSH (i.e., MinHash-ALSH) for MIPS over binary data [21].
4
Evaluations
Datasets. We evaluate the proposed ALSH scheme for the MIPS problem on two popular collaborative filtering datasets on the task of item recommendations: (i) Movielens(10M), and (ii) Netflix.
Each dataset forms a sparse user-item matrix R, where the value of R(i, j) indicates the rating
of user i for movie j. Given the user-item ratings matrix R, we follow the standard PureSVD procedure [4] to generate user and item latent vectors. This procedure generates latent vectors ui for
each user i and vector vj for each item j, in some chosen fixed dimension f . The PureSVD method
returns top-ranked items based on the inner products uTi vj , ?j. Despite its simplicity, PureSVD
outperforms other popular recommendation algorithms [4]. Following [4], we use the same choices
for the latent dimension f , i.e., f = 150 for Movielens and f = 300 for Netflix.
4.1
Ranking Experiment for Hash Code Quality Evaluations
We are interested in knowing, how the two hash functions correlate with the top-10 inner products.
For this task, given a user i and its corresponding user vector ui , we compute the top-10 gold
standard items based on the actual inner products uTi vj , ?j. We then compute K different hash
codes of the vector ui and all the item vectors vj s. For every item vj , we compute the number of
times its hash values matches (or collides) with the hash values of query which is user ui , i.e., we
compute M atchesj = ?K
t=1 1(ht (ui ) = ht (vj )), based on which we rank all the items.
Figure 2 reports the precision-recall curves in our ranking experiments for top-10 items, for comparing our proposed method with two baseline methods: the original L2LSH and the original sign
random projections (SRP). These results confirm the substantial advantage of our proposed method.
4.2 LSH Bucketing Experiment
We implemented the standard (K, L)-parameterized (where L is number of hash tables) bucketing
algorithm [1] for retrieving top-50 items based on PureSVD procedure using the proposed ALSH
hash function and the two baselines: SRP and L2LSH. We plot the recall vs the mean ratio of inner
product required to achieve that recall. The ratio being computed relative to the number of inner
products required in a brute force linear scan. In order to remove the effect of algorithm parameters
(K, L) on the evaluations, we report the result from the best performing K and L chosen from
K ? {5, 6, ..., 30} and L ? {1, 2, ..., 200} for each query. We use m = 3, U = 0.83, and r = 2.5 for
7
Top 10, K = 16
5
0
0
20
40
60
Recall (%)
10
NetFlix
6
4
Top 10, K = 16
2
0
0
20
40
60
Recall (%)
80
100
Top 10, K = 64
10
0
0
100
Proposed
L2LSH
SRP
8
Precision (%)
80
20
20
40
60
Recall (%)
20
80
15
10
Top 10, K = 64
5
0
0
Movielens
Proposed
L2LSH
SRP
40
Top 10, K = 256
20
0
0
100
Proposed
L2LSH
SRP
NetFlix
60
Precision (%)
10
Proposed
L2LSH
SRP
Movielens
20
40
60
Recall (%)
80
50
40
Precision (%)
Precision (%)
Movielens
30
Precision (%)
Proposed
L2LSH
SRP
Precision (%)
15
100
Proposed
L2LSH
SRP
NetFlix
30
Top 10, K = 256
20
10
20
40
60
Recall (%)
80
100
0
0
20
40
60
Recall (%)
80
100
Figure 2: Ranking. Precision-Recall curves (higher is better), of retrieving top-10 items, with the
number of hashes K ? {16, 64, 256}. The proposed algorithm (solid, red if color is available) significantly outperforms L2LSH. We fix the parameters m = 3, U = 0.83, and r = 2.5 for our proposed
method and we present the results of L2LSH for all r values in {1, 1.5, 2, 2.5, 3, 3.5, 4, 4.5, 5}.
0.8
1
Proposed
Top 50
SRP
Movielens
L2LSH
Fraction Multiplications
Fraction Multiplications
1
0.6
0.4
0.2
0
0
0.2
0.4
0.6
Recall
0.8
0.8
Top 50
Netflix
0.6
0.4
0.2
0
0
1
Proposed
SRP
L2LSH
0.2
0.4
0.6
Recall
0.8
1
Figure 3: Bucketing. Mean number of inner products per query, relative to a linear scan, evaluated by different hashing schemes at different recall levels, for generating top-50 recommendations
(Lower is better). The results corresponding to the best performing K and L (for a wide range of K
and L) at a given recall value, separately for all the three hashing schemes, are shown.
our hashing scheme. For L2LSH, we observe that using r = 4 usually performs well and so we show
results for r = 4. The results are summarized in Figure 3, confirming that the proposed ALSH leads
to significant savings compared to baseline hash functions.
5
Conclusion
MIPS (maximum inner product search) naturally arises in numerous practical scenarios, e.g., collaborative filtering. This problem is challenging and, prior to our work, there existed no provably
sublinear time hashing algorithms for MIPS. Also, the existing framework of classical LSH (locality
sensitive hashing) is not sufficient for solving MIPS. In this study, we develop ALSH (asymmetric
LSH), which generalizes the existing LSH framework by applying (appropriately chosen) asymmetric transformations to the input query vector and the data vectors in the repository. We present an
implementation of ALSH by proposing a novel transformation which converts the original inner
products into L2 distances in the transformed space. We demonstrate, both theoretically and empirically, that this implementation of ALSH provides provably efficient as well as practical solution
to MIPS. Other explicit constructions of ALSH, for example, ALSH through cosine similarity, or
ALSH through resemblance (for binary data), will be presented in followup technical reports.
Acknowledgments
The research is partially supported by NSF-DMS-1444124, NSF-III-1360971, NSF-Bigdata1419210, ONR-N00014-13-1-0764, and AFOSR-FA9550-13-1-0137. We appreciate the constructive comments from the program committees of KDD 2014 and NIPS 2014. Shrivastava would also
like to thank Thorsten Joachims and the Class of CS6784 (Spring 2014) for valuable feedbacks.
8
References
[1] A. Andoni and P. Indyk. E2lsh: Exact euclidean locality sensitive hashing. Technical report,
2004.
[2] A. Z. Broder. On the resemblance and containment of documents. In the Compression and
Complexity of Sequences, pages 21?29, Positano, Italy, 1997.
[3] M. S. Charikar. Similarity estimation techniques from rounding algorithms. In STOC, pages
380?388, Montreal, Quebec, Canada, 2002.
[4] P. Cremonesi, Y. Koren, and R. Turrin. Performance of recommender algorithms on topn recommendation tasks. In Proceedings of the fourth ACM conference on Recommender
systems, pages 39?46. ACM, 2010.
[5] R. R. Curtin, A. G. Gray, and P. Ram. Fast exact max-kernel search. In SDM, pages 1?9, 2013.
[6] M. Datar, N. Immorlica, P. Indyk, and V. S. Mirrokn. Locality-sensitive hashing scheme based
on p-stable distributions. In SCG, pages 253 ? 262, Brooklyn, NY, 2004.
[7] T. Dean, M. A. Ruzon, M. Segal, J. Shlens, S. Vijayanarasimhan, and J. Yagnik. Fast, accurate
detection of 100,000 object classes on a single machine. In Computer Vision and Pattern
Recognition (CVPR), 2013 IEEE Conference on, pages 1814?1821. IEEE, 2013.
[8] P. F. Felzenszwalb, R. B. Girshick, D. McAllester, and D. Ramanan. Object detection with
discriminatively trained part-based models. Pattern Analysis and Machine Intelligence, IEEE
Transactions on, 32(9):1627?1645, 2010.
[9] J. H. Friedman and J. W. Tukey. A projection pursuit algorithm for exploratory data analysis.
IEEE Transactions on Computers, 23(9):881?890, 1974.
[10] M. X. Goemans and D. P. Williamson. Improved approximation algorithms for maximum cut
and satisfiability problems using semidefinite programming. Journal of ACM, 42(6):1115?
1145, 1995.
[11] S. Har-Peled, P. Indyk, and R. Motwani. Approximate nearest neighbor: Towards removing
the curse of dimensionality. Theory of Computing, 8(14):321?350, 2012.
[12] P. Indyk and R. Motwani. Approximate nearest neighbors: Towards removing the curse of
dimensionality. In STOC, pages 604?613, Dallas, TX, 1998.
[13] N. Koenigstein, P. Ram, and Y. Shavitt. Efficient retrieval of recommendations in a matrix
factorization framework. In CIKM, pages 535?544, 2012.
[14] Y. Koren, R. Bell, and C. Volinsky. Matrix factorization techniques for recommender systems.
[15] P. Li and A. C. K?onig. Theory and applications b-bit minwise hashing. Commun. ACM, 2011.
[16] P. Li, M. Mitzenmacher, and A. Shrivastava. Coding for random projections. In ICML, 2014.
[17] P. Li, M. Mitzenmacher, and A. Shrivastava. Coding for random projections and approximate
near neighbor search. Technical report, arXiv:1403.8144, 2014.
[18] B. Neyshabur and N. Srebro. A simpler and better lsh for maximum inner product search
(mips). Technical report, arXiv:1410.5518, 2014.
[19] P. Ram and A. G. Gray. Maximum inner-product search using cone trees. In KDD, pages
931?939, 2012.
[20] A. Shrivastava and P. Li. Beyond pairwise: Provably fast algorithms for approximate k-way
similarity search. In NIPS, Lake Tahoe, NV, 2013.
[21] A. Shrivastava and P. Li. Asymmetric minwise hashing. Technical report, 2014.
[22] A. Shrivastava and P. Li.
arXiv:1410.5410, 2014.
An improved scheme for asymmetric lsh.
Technical report,
[23] A. Shrivastava and P. Li. In defense of minhash over simhash. In AISTATS, 2014.
[24] R. Weber, H.-J. Schek, and S. Blott. A quantitative analysis and performance study for
similarity-search methods in high-dimensional spaces. In VLDB, pages 194?205, 1998.
9
| 5329 |@word repository:1 compression:1 norm:15 cs0:13 vldb:1 scg:1 xtest:1 solid:2 score:1 document:1 past:2 existing:4 outperforms:2 current:3 comparing:1 activation:4 dx:1 kdd:2 confirming:1 remove:2 designed:1 plot:2 hash:29 v:1 intelligence:1 item:22 realizing:1 fa9550:1 provides:3 tahoe:1 firstly:1 simpler:1 along:2 retrieving:4 schek:1 pairwise:1 theoretically:1 behavior:1 p1:12 multi:5 decreasing:1 actual:1 curse:4 becomes:1 underlying:1 suffice:1 maximizes:2 biostatistics:1 panel:4 bounded:1 minimizes:1 proposing:1 finding:6 transformation:20 giant:1 nj:1 guarantee:9 quantitative:1 every:2 runtime:3 exactly:1 prohibitively:1 classifier:1 control:1 brute:2 partitioning:5 ramanan:1 onig:1 before:1 negligible:1 dallas:1 despite:2 datar:1 approximately:1 challenging:1 factorization:4 range:3 practical:5 acknowledgment:1 practice:4 implement:1 differs:1 procedure:3 maxx:1 significantly:1 bell:1 projection:8 pre:1 get:1 cannot:4 close:2 convenience:2 cremonesi:1 vijayanarasimhan:1 impossible:1 applying:2 conventional:1 equivalent:1 dean:1 go:1 simplicity:2 identifying:1 assigns:1 m2:1 utilizing:2 shlens:1 searching:2 anyway:1 variation:1 exploratory:1 analogous:1 construction:6 suppose:1 heavily:1 user:15 exact:2 losing:1 programming:1 us:1 element:2 expensive:2 recognition:1 asymmetric:22 cut:1 solved:1 worst:1 ordering:2 counter:1 highest:1 removed:1 valuable:1 substantial:1 ui:7 complexity:1 peled:1 trained:1 solving:12 easily:1 various:1 tx:1 fast:5 describe:1 query:27 solve:3 cvpr:1 relax:2 statistic:1 itself:1 indyk:4 advantage:1 sequence:1 sdm:1 product:29 coming:1 fr:13 e2lsh:1 relevant:1 flexibility:1 achieve:1 deformable:1 gold:1 recipe:1 motwani:2 requirement:2 asymmetry:4 lshs:1 generating:1 object:8 koenigstein:1 develop:2 montreal:1 stat:1 nearest:3 eq:5 sim:17 p2:12 implemented:1 c:3 involves:1 come:1 quantify:1 radius:2 filter:6 anshu:1 mcallester:1 argued:2 fix:1 generalization:1 randomization:1 ultra:1 correction:1 considered:2 normal:2 mapping:1 purpose:2 estimation:1 label:9 sensitive:17 anshumali:1 always:4 aim:1 modified:1 cornell:2 focus:3 inherits:1 joachim:1 rank:2 indicates:1 industrial:1 baseline:3 dependent:1 nn:7 typically:2 expand:1 transformed:1 interested:7 provably:6 arg:13 classification:1 art:1 equal:1 construct:11 saving:2 having:5 aware:1 icml:1 purchase:1 report:10 recommend:3 modern:2 n1:4 friedman:1 detection:7 huge:1 interest:2 evaluation:5 semidefinite:1 unconditional:1 behind:1 har:1 accurate:1 unless:1 tree:4 euclidean:2 girshick:1 theoretical:6 instance:10 formalism:1 modeling:1 cost:1 successful:1 rounding:1 conducted:2 too:1 dependency:1 nns:4 combined:1 chooses:1 density:1 broder:1 randomized:2 probabilistic:1 quickly:1 again:2 choose:4 creating:1 leading:2 return:1 li:8 converted:1 segal:1 blow:1 coding:5 summarized:1 satisfy:1 explicitly:1 ranking:3 depends:2 later:3 h1:2 lot:1 observing:1 tukey:1 red:1 netflix:8 recover:1 collaborative:6 contribution:1 accuracy:1 characteristic:3 efficiently:6 identify:1 basically:1 ping:1 parallelizable:1 definition:9 volinsky:1 dm:1 naturally:2 associated:3 proof:5 boil:1 dataset:2 popular:5 appends:2 recall:15 knowledge:1 color:1 dimensionality:6 satisfiability:1 back:1 hashing:31 higher:3 follow:1 improved:3 evaluated:1 mitzenmacher:2 generality:1 furthermore:1 just:2 correlation:1 web:1 logistic:1 quality:1 gray:2 resemblance:3 usa:2 effect:2 building:1 normalized:7 hence:4 equality:1 turrin:1 during:1 self:2 noted:1 cosine:2 generalized:1 hl2:6 demonstrate:1 performs:3 image:6 weber:1 novel:2 recently:1 common:3 empirically:1 million:1 extend:1 significant:4 ai:1 rd:19 vanilla:1 grid:2 similarly:2 lsh:44 stable:2 similarity:19 showed:3 recent:1 perspective:1 italy:1 commun:1 prime:1 scenario:2 massively:1 n00014:1 topn:1 binary:2 arbitrarily:1 onr:1 yagnik:1 devise:1 scoring:2 seen:1 greater:2 additional:1 floor:1 monotonically:1 dashed:1 ii:3 branch:1 reduces:2 technical:6 match:1 long:2 retrieval:1 controlled:1 impact:1 prediction:3 regression:1 basic:1 vision:1 rutgers:2 arxiv:4 kernel:2 proposal:2 background:1 want:1 fine:1 separately:1 completes:1 source:1 ithaca:1 extra:1 collides:1 unlike:1 appropriately:1 comment:1 nv:1 dpm:5 quebec:1 near:15 ideal:1 iii:2 easy:1 mips:50 enough:1 minhash:2 affect:2 followup:1 inner:29 idea:4 knowing:2 defense:1 suffer:2 useful:2 collision:5 generate:1 outperform:1 exist:5 nsf:3 sign:6 cikm:1 per:1 key:5 threshold:3 alsh:25 changing:1 ht:2 ram:3 fraction:3 convert:2 cone:1 package:1 parameterized:1 fourth:1 place:1 family:9 reasonable:1 uti:3 patch:3 lake:1 scaling:1 bit:4 bound:2 koren:2 existed:1 scaler:1 constraint:2 precisely:1 ri:2 colliding:1 x2:2 generates:1 argument:2 min:6 spring:1 performing:2 department:3 charikar:1 piscataway:1 bucketing:4 wi:3 lp:1 making:1 bigdata1419210:1 thorsten:1 bucket:2 previously:2 committee:1 needed:1 know:1 end:2 adopted:1 available:1 operation:1 generalizes:1 pursuit:1 neyshabur:1 observe:2 appropriate:1 generic:2 ruzon:1 existence:1 original:4 top:17 running:1 unsuitable:1 classical:4 appreciate:1 pingli:1 costly:1 traditional:1 distance:9 thank:1 concatenation:1 degrade:1 tower:2 reason:1 code:2 modeled:1 insufficient:1 ratio:4 difficult:2 stoc:2 negative:2 implementation:2 recommender:6 upper:1 observation:4 datasets:6 extended:2 canada:1 rating:4 shavitt:1 required:5 connection:1 learned:5 nip:2 brooklyn:1 beyond:1 usually:1 pattern:2 program:1 max:10 suitable:2 event:7 ranked:2 force:2 predicting:2 scheme:19 movie:1 numerous:1 ready:1 nice:1 prior:1 l2:12 multiplication:3 relative:2 afosr:1 loss:1 discriminatively:1 sublinear:6 interesting:1 filtering:6 querying:1 facing:1 srebro:1 h2:2 sufficient:1 s0:34 classifying:1 supported:1 formal:2 allow:4 ytest:1 neighbor:18 wide:2 felzenszwalb:1 simhash:1 sparse:2 benefit:2 curve:4 dimension:5 feedback:1 cumulative:1 author:1 collection:3 commonly:1 preprocessing:5 made:1 curtin:1 correlate:2 transaction:2 approximate:25 dealing:1 confirm:1 rid:1 containment:1 recommending:1 xi:20 search:27 un:5 latent:7 table:3 learn:1 shrivastava:8 expansion:1 williamson:1 constructing:1 domain:1 vj:9 did:1 aistats:1 main:1 rh:4 motivation:1 big:2 positano:1 srp:16 allowed:3 deployed:1 ny:2 shrinking:1 precision:8 explicit:4 topranked:1 exponential:2 candidate:1 grained:1 maxxi:1 theorem:9 down:1 removing:3 xt:3 svm:1 normalizing:1 exists:2 andoni:1 easier:1 locality:13 simply:1 absorbed:1 partially:1 scalar:1 recommendation:9 u2:7 monotonic:2 blott:1 satisfies:2 relies:1 acm:4 cdf:1 identity:1 consequently:1 towards:2 hard:3 change:2 movielens:8 typical:2 uniformly:3 called:2 goemans:1 experimental:1 formally:5 immorlica:1 support:1 arises:2 scan:3 meant:1 minwise:3 constructive:1 evaluate:1 |
4,782 | 533 | Adaptive Elastic Models for Hand-Printed
Character Recognition
Geoffrey E. Hinton, Christopher K. I. Williams and Michael D. Revow
Department of Computer Science, University of Toronto
Toronto, Ontario, Canada M5S lA4
Abstract
Hand-printed digits can be modeled as splines that are governed by about
8 control points . For each known digit. the control points have preferred
., home" locations, and deformations of the digit are generated by moving
the control points away from their home locations. Images of digits can be
produced by placing Gaussian ink generators uniformly along the spline.
Real images can be recognized by finding the digit model most likely to
have generated the data. For each digit model we use an elastic matching
algorithm to minimize an energy function that includes both the deformation energy of the digit model and the log probability that the model
would generate the inked pixels in the image. The model with the lowest
total energy wins. If a uniform noise process is included in the model of
image generation, some of the inked pixels can be rejected as noise as a
digit model is fitting a poorly segmented image. The digit models learn
by modifying the home locations of the control points.
1
Introduction
Given good bottom-up segmentation and normalization, feedforward neural networks are an efficient way to recognize digits in zip codes. (Ie Cun et al., 1990).
However. in some cases. it is not possible to correctly segment and normalize the
digits without using knowledge of their shapes, so to achieve close to human performance on images of whole zip codes it will be necessary to use models of shapes
to influence the segmentation and normalization of the digits. One way of doing
this is to use a large cooperative network that simultaneously segments, normalizes
and recognizes all of the digit.s in a zip code. A first step in this direct.ion is to take
a poorly segmented image of a single digit and to explain the image properly in
terms of an appropriately normalized, deformed digit model plus noise. The ability
of t.he model to reject some parts of the image as noise is the first step towards
model-driven segmentation.
512
Adaptive Elastic Models for Hand)Printed Character Recognition
2
Elastic models
One technique for recognizing a digit is to perform an elastic match with many
different exemplars of each known digit-class and to pick the class of the nearest
neighbor. Unfortunately this requires a large number of elastic matches, each of
which is expensive . By using one elastic model to capture all the variations of a given
digit we greatly reduce the number of elastic matches required . Burr (1981a, 1981b)
has investigated several types of elastic model and elastic matching procedure. We
describe a different kind of elastic model that is based on splines. Each elastic
model contains parameters that define an ideal shape and also define a deformation
energy for departures from this ideal. These parameters are initially set by hand
but can be improved by learning. They are an efficient way to represent the many
possible instances of a given digit .
Each digit is modelled by a deformable spline whose shape is determined by the
positions of 8 control points . Every point on the spline is a weighted average of
four control points, with the weighting coefficients changing smoothly as we move
along the spline. 1 To generate an ideal example of a digit we put the 8 control
points at their home locations for that model. To deform the digit we move the
control points away from their home locations. Currently we assume that, for each
model, the control points have independent, radial Gaussian distributions about
their home locations. So the negative log probability of a deformation (its energy)
is proportional to the sum of the squares of the departures of the control points
from their home locations.
The deformation energy function only penalizes shape deformations . Translation,
rotation, dilation , elongation, and shear do not change the shape of an object so we
want the deformation energy to be invariant under these affine transformations. We
achieve this by giving each model its own "object-based frame". Its deformation
energy is computed relative to this frame, not in image coordinates. When we fit
the model to data, we repeatedly recompute the best affine transformation between
the object-based frame and the image (see section 4). The repeated recomputation
of the affine transform during the model fit means that the shape of the digit is
influencing the normalization.
Although we will use our digit models for recognizing images, it helps to start by
considering how we would use them for generating images . The generative model is
an elaboration of the probabilistic interpretation of the elastic net given by Durbin,
Szeliski & Yuille (1989) . Given a particular spline, we space a number of "beads"
uniformly along the spline. Each bead defines the center of a Gaussian ink generator.
The number of beads on the spline and the variance of the ink generators can easily
be changed without changing the spline itself.
To generate a noisy image of a particular digit class, run the following procedure:
? Pick an affine transformation from the model's intrinsic reference frame to the
image frame (i .e. pick a size, position, orientation, slant and elongation for the
digit) .
1 In computing the weighting coefficients we use a cubic B-spline and we treat the first
and last control points as if they were doubled.
513
514
Hinton, Williams, and Revow
? Pick a defo~mation of the mo.d~l (i.e. ~~ve the control !)Qi~ts awa1 from their
home locatIOns). The probabIlIty of pIckmg a deformatIOn IS ~ e- de.Jornl
? Repeat many times:
Either (with probability 1I"noi.H) add a randomly positioned noise pixel
Or pick a bead at random and generate a pixel from the Gaussian
distribution defined by the bead.
3
Recognizing isolated digits
We recognize an image by finding which model is most likely to have generated it.
Each possible model is fitted to the image and the one that has the lowest cost fit is
the winner. The cost of a fit is the negative log probability of generating the image
gi ven the model.
- log
J
P(I) P( image
I 1)
dI
(1 )
IE model
instances
We can approximate this by just considering the best fitting model instance 2 and
ignoring the fact that the model should not generate ink where there is no ink in
the image: 3
E
= A EdeJorm
-
L
logP(pixel
I best
model instance)
(2)
inked
pixels
The probability of an inked pixel is the sum of the probabilities of all the possible
ways of generating it from the mixture of Gaussian beads or the uniform noise field.
P(i)
= 1I"noi.H + 1I"model
N
B
(3)
where N is the total number of pixels, B is the number of beads, 11" is a mlxmg
proportion', and Pb( i) is the probability density of pixel i under Gaussian bead b.
4
The search procedure for fitting a model to an image
Every Gaussian bead in a model has the same variance. When fitting data, we start
with a big variance and gradually reduce it as in the elastic net algorithm of Durbin
and Willshaw (1987) . Each iteration of the elastic matching algorithm involves
three steps:
21n effect, we are assuming that the integral in equation 1 can be approximated by the
height of the highest peak, and so we are ignoring variations between models in the width
of the peak or the number of peaks.
3If the inked pixels are rare, poor models sin mainly by not inking those pixels that
should be inked rather than by inking those pixels that should not be inked.
Adaptive Elastic Models for Hand) Printed Character Recognition
? Given the current locations of the Gaussians, compute the responsibility that
each Gaussian has for each inked pixel. This is just the probability of generating
the pixel from that Gaussian, normalized by the total probability of generating
the pixel.
? Assuming that the responsibilities remain fixed, as in the EM algorithm of
Dempster, Laird and Rubin (1977), we invert a 16 x 16 matrix to find the
image locations for the 8 control points at which the forces pulling the control
points towards their home locations are balanced by the forces exerted on the
control points by the inked pixels. These forces come via the forces that the
inked pixels exert on the Gaussian beads.
? Given the new image locations of the control points, we recompute the affine
transformation from the object-based frame to the image frame. We choose
the affine transformation that minimizes the sum of the squared distances, in
object-based coordinates, between the control points and their home locations.
The residual squared differences determine the deformation energy.
Some stages in the fitting of a model to data are shown in Fig. 1. This search
technique avoids nearly all local minima when fitting models to isolated digits. But
if we get a high deformation energy in the best fitting model, we can try alternative
starting configurations for the models.
5
Learning the digit models
We can do discriminative learning by adjusting the home positions and variances
of the control points to minimize the objective function
c =-
L
e-Ecorrect
10gp(cor7'ect digit),
p(correct digit) =
training
cases
=-----~"""
Lall digits e-Ed'Y'1
(4)
For a model parameter such as the x coordinate of the home location of one of the
control points we need oC /
in order to do gradient descent learning. Equation
4 allows us to express oC /
in terms of oE /
but there is a subtle problem:
Changing a parameter of an elastic model causes a simple change in the energy
of the configuration that the model previously settled to, but the model no longer
settles to that configuration. So it appears that we need to consider how the energy
is affected by the change in the configuration. Fortunately, derivatives are simple at
an energy minimum because small changes in the configuration make no change in
the energy (to first order). Thus the inner loop settling leads to simple derivatives
for the outer loop learning, as in the Boltzmann machine (Hinton, 1989).
ax
ax
6
ax
Results on the hand-filtered dataset
We are trying out the scheme out on a relatively simple task - we have a model of
a two and a model of a three, and we want the two model to win on "two" images,
and the three model to win on "three" images.
We have tried many variations of the character models, the preprocessing, the initial
affine transformations of the models, the annealing schedule for the variances, the
515
516
Hinton, Williams, and Revow
c:=
(b)
(a)
,,--,.
,
"",ue:;~,. .~~4P
(c)
(d)
Figure 1: The sequellce> (n) 1.0 (d) shows SOIIlC stages or rHf.illg a model :~ 1.0 SOllie
daf.1\.. The grey circles I?e>presellf. the heads Oil the splille, alld t.he> m,dius or t.he rircl(~
represents t.he standard deviation or t.he Gaussian. (a.) shows the illitia.1 conliglll'atioll, with eight beads equally spaced along the spline. 111 (b) and (c) the va.riallce
is 11I'ogl:es~ively decrca.~ed and t.he Humber or heads is incrf~ased. The ri lIal ra \lsi IIg
GO beads is showlI in (d). We use about. three iterat.ions al. each or nve variallces
on our "annealing schedule". III this example, we used 1Tnoiu = 0.3 which lIIa.kes it.
cheaper to explain the extrft,nCOliS 1I0ise pixels and the flourishes 011 t.he cllds or t.11!~
:~ as noise rather lhall deformillg t.he llIodel to briug Gallssiall heads cI()s(~ t.o t.hese
pixels.
Adaptive Elastic Models for Hand)Printed Character Recognition
mixing proportion of the noise, and the relative importance of deformation energy
versus data-fit energy.
Our current best performance is 10 errors (1.6%) on a test set of 304 two's and 304
three's. We reject cases if the best-fitting model is highly deformed, but on this
test set the deformation energy never reached the rejection criterion. The training
set has 418 cases, and we have a validation set of 200 cases to tell us when to stop
learning. Figure 2 shows the effect of learning on the models. The initial affine
transform is defined by the minimal vertical rectangle around the data.
(BEFORE)
[0
lTI
IAFTER
[i]
I
lTI
Figure 2: The two and three models before and after learning. The control points
are labelled 1 through 8. We used maximum likelihood learning in which each digit
model is trained only on instances of that digit. After each pass through all those
instances, the home location of each control point (in the object-based frame) is
redefined to be the average location of the control point in the final fits of the model
of the digit to the instances of the digit. Most of the improvement in performance
occurred after the fist pass, and after five updates of the home locations of the
control points, performance on the validation set started to decrease. Similar results
were obtained with discriminative training. We could also update the variance of
each control point to be its variance in the final fits, though we did not adapt the
variances in this simulation.
517
518
Hinton, Williams, and Revow
The images are preprocessed to eliminate variations due to stroke-width and paper
and ink intensities. First, we use a standard local thresholding algorithm to make
a binary decision for each pixel. Then we pick out the five largest connected components (hopefully digits). We put a box around each component , then thin all the
data in the box . If we ourselves cannot recognize the resulting image we eliminate
it from the data set. The training, validation and test data is all from the training
portion of the United States Postal Service Handwritten ZIP Code Database (1987)
which was made available by the USPS Office of Advanced Technology.
7
Discussion
Before we tried using splines to model digits, we used models that consisted of a
fixed number of Gaussian beads with elastic energy constraints operating between
neighboring beads. To constrain the curvature we used energy terms that involved
triples of beads. With this type of energy function, we had great difficulty using
a single model to capture topologically different instances of a digit. For example,
when the central loop of a 3 changes to a cusp and then to an open bend, the sign
of the curvature reverses. With a spline model it is easy to model these topological
variants by small changes in the relative vertical locations of the central two control
points (see figure 2). This advantage of spline models is pointed out by (Edelman,
Ullman and Flash, 1990) who use a different kind of spline that they fit to character
data by directly locating candidate knot points in the image.
Spline models also make it easy to increase the number of Gaussian beads as their
variance is decreased. This coarse-to-fine strategy is much more efficient than using
a large number of beads at all variances, but it is much harder to implement if the
deformation energy explicitly depends on particular bead locations, since changiug
the number of beads then requires a new function for the deformation energy.
In determining where on the spline to place the Gaussian beads, we initially used
a fixed set of blending coefficients for each bead . These coefficients are the weight:s
used to specify the bead location as a weighted center of gravity of the loca.l-iollS of
4 control points. Unfortunately this yields too few beads in portions of a digit such
as a long tail of a 2 which are governed by just a few control points. Performance
was much improved by spacing the beads uniformly along the curve.
By using spline models, we build in a lot of prior knowledge about wha.t characters
look like, so we can describe the shape of a character using only a small number
of parameters (16 coordinates and 8 variances). This means that the learning is
exploring a much smaller space than a conventional feed-forward network. Also,
because the parameters are easy to interpret, we can start with fairly good initial
models of the characters. So learning only requires a few updates of the parameters.
Obvious extensions of the deformation energy function include using elliptical Gaussians for the distributions of the control points, or using full covariance matrices for
neighboring pairs of control points. Another obvious modification is to use elliptical rather than circular Gaussians for the beads . If strokes curve gently relative to
their thickness, the distribution of ink can be modelled much better using elliptical
Gaussians. However, an ellipse takes about twice as many operations to fit and is
not helpful in regions of sharp curvature. Our simulations suggest that, on average,
two circular beads are more flexible than one elliptical bead.
Adaptive Elastic Models for Hand) Printed Character Recognition
Currently we do not impose any penalty on extremely sheared or elongated affine
transformations, though this would probably improve performance. Having an explicit representation of the affine transformation of each digit should prove very
helpful for recognizing multiple digits, since it will allow us to impose a penalty on
differences in the affine transformations of neighboring digits.
Presegmented images of single digits contain many different kinds of noise that
cannot be eliminated by simple bottom-up operations. These include descenders,
underlines, and bits of other digits; corrections; dirt in recycled paper; smudges and
misplaced postal franks. To really understand the image we probably need to model
a wide variety of structured noise. We are currently experimenting with one simple
way of incorporating noise models. After each digit model has been used to segment
a noisy image into one digit instance plus noise, we try to fit more complicated noise
models to the residual noise. A good fit greatly decreases the cost of that noise and
hence improves this interpretation of the image. We intend to handle flourishes on
the ends of characters in this way rather than using more elaborate digit models
that include optional flourishes.
One of our main motivations in developing elastic models is the belief that a strong
prior model should make learning easier, should reduce confident errors, and should
allow top-down segmentation . Although we have shown that elastic spline models can be quite effective, we have not yet demonstrated that they are superior to
feedforward nets and there is a serious weakness of our approach: Elastic matching is slow. Fitting the models to the data takes much more computation than a
feedforward net. So in the same number of cycles, a feedforward net can try many
alternative bottom-up segmentations and normalizations and select the overall segmentation that leads to the most recognizable digit string.
Acknowledgements
This research was funded by Apple and by the Ontario Information Technology Research
Centre. We thank Allan Jepson and Richard Durbin for suggesting spline models.
References
Burr, D. J. (1981a) . A dynamic model for image registration. Comput. Gmphics image
Process., 15:102-112.
Burr, D. J. (1981b). Elastic matching of line drawings. IEEE Trans. Pattern Analysis and
Machine Intelligence, 3(6):708-713.
Dempster, A. P., Laird, N. M., and Rubin, D. B. (1977). Maximum likelihood from
incomplete data via the EM algorithm . Proc. Roy. Stat. Soc., B-39:1-38 .
Durbin, R., Szeliski, R., and Yuille, A. L. (1989). An analysis of the elastic net approach
to the travelling salesman problem. Neural Computation, 1:348-358.
Durbin, R. and Willshaw, D. (1987). An analogue approach to the travelling salesman
problem. Nature, 326:689-691.
Edelman, S., Ullman, S., and Flash, T. (1990). Reading cursive handwriting by alignment
of letter prototypes. Internat. Journal of Comput. Vision, 5(3):303-33l.
Hinton, G. E. (1989). Deterministic Boltzmann learning performs steepest descent in
weight-space. Neural Computation, 1:143-150.
Ie Cun, Y., Boser, B., Denker, J., Henderson, D., Howard, R., Hubbard, W., and Jackel, L.
(1990). Handwritten digit recognition with a back-propagation network. In Advances
in Neural Information Processing Systems 2, pages 396-404. Morgan Kaufmann.
519
PART IX
CONTROL
AND PLANNING
| 533 |@word deformed:2 proportion:2 underline:1 open:1 grey:1 simulation:2 tried:2 covariance:1 pick:6 harder:1 initial:3 configuration:5 contains:1 united:1 current:2 elliptical:4 yet:1 shape:8 update:3 generative:1 intelligence:1 steepest:1 filtered:1 recompute:2 coarse:1 postal:2 toronto:2 location:20 five:2 height:1 along:5 direct:1 ect:1 edelman:2 prove:1 fitting:9 burr:3 recognizable:1 allan:1 ra:1 planning:1 considering:2 lowest:2 kind:3 minimizes:1 string:1 finding:2 transformation:9 every:2 gravity:1 willshaw:2 control:30 misplaced:1 before:3 service:1 influencing:1 local:2 treat:1 plus:2 exert:1 twice:1 implement:1 digit:50 procedure:3 reject:2 printed:6 matching:5 radial:1 suggest:1 doubled:1 get:1 close:1 cannot:2 bend:1 put:2 influence:1 conventional:1 elongated:1 demonstrated:1 center:2 deterministic:1 williams:4 go:1 starting:1 handle:1 variation:4 coordinate:4 roy:1 recognition:6 expensive:1 approximated:1 cooperative:1 database:1 bottom:3 capture:2 region:1 connected:1 cycle:1 oe:1 decrease:2 highest:1 noi:2 balanced:1 dempster:2 dynamic:1 hese:1 trained:1 segment:3 yuille:2 iig:1 usps:1 easily:1 describe:2 effective:1 tell:1 whose:1 quite:1 drawing:1 ability:1 gi:1 gp:1 transform:2 itself:1 noisy:2 la4:1 laird:2 final:2 advantage:1 net:6 neighboring:3 loop:3 ogl:1 poorly:2 ontario:2 achieve:2 deformable:1 mixing:1 normalize:1 rhf:1 generating:5 object:6 help:1 stat:1 exemplar:1 nearest:1 strong:1 soc:1 involves:1 come:1 revers:1 correct:1 modifying:1 human:1 cusp:1 settle:1 really:1 blending:1 exploring:1 extension:1 correction:1 around:2 great:1 mo:1 proc:1 currently:3 jackel:1 hubbard:1 largest:1 weighted:2 gaussian:14 mation:1 rather:4 office:1 ax:3 properly:1 improvement:1 likelihood:2 mainly:1 experimenting:1 greatly:2 helpful:2 eliminate:2 initially:2 pixel:20 overall:1 orientation:1 flexible:1 loca:1 fairly:1 field:1 exerted:1 never:1 having:1 elongation:2 eliminated:1 placing:1 ven:1 represents:1 look:1 nearly:1 thin:1 presegmented:1 spline:21 descender:1 serious:1 few:3 richard:1 randomly:1 simultaneously:1 recognize:3 ve:1 cheaper:1 ourselves:1 highly:1 circular:2 alignment:1 weakness:1 henderson:1 mixture:1 integral:1 ively:1 necessary:1 incomplete:1 illg:1 penalizes:1 circle:1 deformation:16 isolated:2 minimal:1 fitted:1 instance:9 logp:1 cost:3 deviation:1 rare:1 uniform:2 recognizing:4 too:1 thickness:1 confident:1 density:1 peak:3 ie:3 probabilistic:1 michael:1 recomputation:1 squared:2 settled:1 central:2 choose:1 derivative:2 ullman:2 deform:1 suggesting:1 de:1 includes:1 coefficient:4 explicitly:1 depends:1 try:3 lot:1 responsibility:2 doing:1 reached:1 start:3 portion:2 complicated:1 alld:1 minimize:2 square:1 variance:11 who:1 kaufmann:1 spaced:1 yield:1 flourish:3 modelled:2 handwritten:2 produced:1 knot:1 apple:1 m5s:1 stroke:2 explain:2 ed:2 energy:23 involved:1 obvious:2 di:1 handwriting:1 stop:1 dataset:1 adjusting:1 knowledge:2 improves:1 segmentation:6 subtle:1 schedule:2 positioned:1 back:1 appears:1 feed:1 specify:1 improved:2 though:2 box:2 rejected:1 just:3 stage:2 hand:8 christopher:1 hopefully:1 propagation:1 defines:1 pulling:1 oil:1 effect:2 normalized:2 consisted:1 contain:1 hence:1 sin:1 during:1 width:2 ue:1 oc:2 inked:10 criterion:1 trying:1 performs:1 image:35 dirt:1 superior:1 rotation:1 shear:1 winner:1 gently:1 tail:1 he:8 interpretation:2 occurred:1 interpret:1 slant:1 pointed:1 centre:1 had:1 funded:1 moving:1 longer:1 operating:1 internat:1 add:1 curvature:3 own:1 driven:1 binary:1 morgan:1 minimum:2 fortunately:1 impose:2 zip:4 recognized:1 determine:1 fist:1 full:1 multiple:1 segmented:2 match:3 adapt:1 long:1 elaboration:1 equally:1 va:1 qi:1 iterat:1 variant:1 vision:1 iteration:1 normalization:4 represent:1 invert:1 ion:2 want:2 fine:1 spacing:1 annealing:2 decreased:1 appropriately:1 probably:2 ideal:3 feedforward:4 iii:1 easy:3 variety:1 fit:11 ased:1 reduce:3 inner:1 prototype:1 penalty:2 locating:1 cause:1 repeatedly:1 cursive:1 generate:5 lsi:1 sign:1 correctly:1 affected:1 express:1 four:1 pb:1 changing:3 kes:1 preprocessed:1 lti:2 registration:1 rectangle:1 sum:3 run:1 letter:1 topologically:1 place:1 home:14 decision:1 bit:1 nve:1 topological:1 durbin:5 lall:1 constraint:1 constrain:1 ri:1 extremely:1 relatively:1 department:1 structured:1 developing:1 poor:1 remain:1 smaller:1 em:2 character:11 cun:2 modification:1 invariant:1 gradually:1 equation:2 previously:1 end:1 travelling:2 salesman:2 available:1 gaussians:4 operation:2 eight:1 denker:1 away:2 alternative:2 top:1 include:3 recognizes:1 giving:1 build:1 ellipse:1 ink:7 move:2 objective:1 intend:1 strategy:1 gradient:1 win:3 distance:1 thank:1 outer:1 assuming:2 code:4 modeled:1 unfortunately:2 frank:1 negative:2 boltzmann:2 redefined:1 perform:1 vertical:2 howard:1 descent:2 t:1 optional:1 hinton:6 head:3 frame:8 sharp:1 canada:1 intensity:1 pair:1 required:1 boser:1 trans:1 pattern:1 departure:2 reading:1 belief:1 analogue:1 difficulty:1 force:4 settling:1 residual:2 advanced:1 scheme:1 improve:1 technology:2 started:1 prior:2 acknowledgement:1 determining:1 relative:4 generation:1 proportional:1 geoffrey:1 versus:1 generator:3 validation:3 triple:1 affine:11 rubin:2 thresholding:1 daf:1 translation:1 normalizes:1 changed:1 repeat:1 last:1 allow:2 understand:1 szeliski:2 neighbor:1 wide:1 curve:2 avoids:1 forward:1 made:1 adaptive:5 preprocessing:1 approximate:1 iolls:1 preferred:1 discriminative:2 bead:27 search:2 dilation:1 learn:1 nature:1 elastic:25 ignoring:2 recycled:1 investigated:1 jepson:1 did:1 main:1 whole:1 noise:15 big:1 motivation:1 repeated:1 fig:1 elaborate:1 cubic:1 slow:1 position:3 explicit:1 comput:2 candidate:1 governed:2 weighting:2 ix:1 down:1 intrinsic:1 incorporating:1 importance:1 ci:1 sheared:1 easier:1 rejection:1 smoothly:1 likely:2 flash:2 towards:2 labelled:1 revow:4 change:7 included:1 determined:1 uniformly:3 total:3 pas:2 e:1 select:1 |
4,783 | 5,330 | A Latent Source Model for
Online Collaborative Filtering
Guy Bresler
George H. Chen
Devavrat Shah
Department of Electrical Engineering and Computer Science
Massachusetts Institute of Technology
Cambridge, MA 02139
{gbresler,georgehc,devavrat}@mit.edu
Abstract
Despite the prevalence of collaborative filtering in recommendation systems, there
has been little theoretical development on why and how well it works, especially
in the ?online? setting, where items are recommended to users over time. We
address this theoretical gap by introducing a model for online recommendation
systems, cast item recommendation under the model as a learning problem, and
analyze the performance of a cosine-similarity collaborative filtering method. In
our model, each of n users either likes or dislikes each of m items. We assume
there to be k types of users, and all the users of a given type share a common
string of probabilities determining the chance of liking each item. At each time
step, we recommend an item to each user, where a key distinction from related
bandit literature is that once a user consumes an item (e.g., watches a movie),
then that item cannot be recommended to the same user again. The goal is to
maximize the number of likable items recommended to users over time. Our main
result establishes that after nearly log(km) initial learning time steps, a simple
collaborative filtering algorithm achieves essentially optimal performance without
knowing k. The algorithm has an exploitation step that uses cosine similarity and
two types of exploration steps, one to explore the space of items (standard in the
literature) and the other to explore similarity between users (novel to this work).
1
Introduction
Recommendation systems have become ubiquitous in our lives, helping us filter the vast expanse of
information we encounter into small selections tailored to our personal tastes. Prominent examples
include Amazon recommending items to buy, Netflix recommending movies, and LinkedIn recommending jobs. In practice, recommendations are often made via collaborative filtering, which boils
down to recommending an item to a user by considering items that other similar or ?nearby? users
liked. Collaborative filtering has been used extensively for decades now including in the GroupLens
news recommendation system [20], Amazon?s item recommendation system [17], the Netflix Prize
winning algorithm by BellKor?s Pragmatic Chaos [16, 24, 19], and a recent song recommendation
system [1] that won the Million Song Dataset Challenge [6].
Most such systems operate in the ?online? setting, where items are constantly recommended to
users over time. In many scenarios, it does not make sense to recommend an item that is already
consumed. For example, once Alice watches a movie, there?s little point to recommending the same
movie to her again, at least not immediately, and one could argue that recommending unwatched
movies and already watched movies could be handled as separate cases. Finally, what matters is
whether a likable item is recommended to a user rather than an unlikable one. In short, a good
online recommendation system should recommend different likable items continually over time.
1
Despite the success of collaborative filtering, there has been little theoretical development to justify
its effectiveness in the online setting. We address this theoretical gap with our two main contributions
in this paper. First, we frame online recommendation as a learning problem that fuses the lines of
work on sleeping bandits and clustered bandits. We impose the constraint that once an item is
consumed by a user, the system can?t recommend the item to the same user again. Our second
main contribution is to analyze a cosine-similarity collaborative filtering algorithm. The key insight
is our inclusion of two types of exploration in the algorithm: (1) the standard random exploration
for probing the space of items, and (2) a novel ?joint? exploration for finding different user types.
Under our learning problem setup, after nearly log(km) initial time steps, the proposed algorithm
achieves near-optimal performance relative to an oracle algorithm that recommends all likable items
first. The nearly logarithmic dependence is a result of using the two different exploration types. We
note that the algorithm does not know k.
Outline. We present our model and learning problem for online recommendation systems in Section
2, provide a collaborative filtering algorithm and its performance guarantee in Section 3, and give
the proof idea for the performance guarantee in Section 4. An overview of experimental results is
given in Section 5. We discuss our work in the context of prior work in Section 6.
2
A Model and Learning Problem for Online Recommendations
We consider a system with n users and m items. At each time step, each user is recommended an
item that she or he hasn?t consumed yet, upon which, for simplicity, we assume that the user immediately consumes the item and rates it +1 (like) or ?1 (dislike).1 The reward earned by the recommendation system up to any time step is the total number of liked items that have been recommended
so far across all users. Formally, index time by t ? {1, 2, . . . }, and users by u ? [n] , {1, . . . , n}.
(t)
Let ?ut ? [m] , {1, . . . , m} be the item recommended to user u at time t. Let Yui ? {?1, 0, +1}
be the rating provided by user u for item i up to and including time t, where 0 indicates that no rating
has been given yet. A reasonable objective is to maximize the expected reward r(T ) up to time T :
r(T ) ,
T X
n
X
(T )
E[Yu?
]=
ut
t=1 u=1
m X
n
X
(T )
E[Yui ].
i=1 u=1
The ratings are noisy: the latent item preferences for user u are represented by a length-m vector
pu ? [0, 1]m , where user u likes item i with probability pui , independently across items. For a user
u, we say that item i is likable if pui > 1/2 and unlikable if pui < 1/2. To maximize the expected
reward r(T ) , clearly likable items for the user should be recommended before unlikable ones.
In this paper, we focus on recommending likable items. Thus, instead of maximizing the expected
reward r(T ) , we aim to maximize the expected number of likable items recommended up to time T :
(T )
r+ ,
T X
n
X
E[Xut ] ,
(1)
t=1 u=1
where Xut is the indicator random variable for whether the item recommended to user u at time t is
(T )
likable, i.e., Xut = +1 if pu?ut > 1/2 and Xut = 0 otherwise. Maximizing r(T ) and r+ differ
since the former asks that we prioritize items according to their probability of being liked.
Recommending likable items for a user in an arbitrary order is sufficient for many real recommendation systems such as for movies and music. For example, we suspect that users wouldn?t actually
prefer to listen to music starting from the songs that their user type would like with highest probability to the ones their user type would like with lowest probability; instead, each user would listen to
songs that she or he finds likable, ordered such that there is sufficient diversity in the playlist to keep
the user experience interesting. We target the modest goal of merely recommending likable items,
in any order. Of course, if all likable items have the same probability of being liked and similarly
(T )
for all unlikable items, then maximizing r(T ) and r+ are equivalent.
1
In practice, a user could ignore the recommendation. To keep our exposition simple, however, we stick to
this setting that resembles song recommendation systems like Pandora that per user continually recommends
a single item at a time. For example, if a user rates a song as ?thumbs down? then we assign a rating of ?1
(dislike), and any other action corresponds to +1 (like).
2
The fundamental challenge is that to learn about a user?s preference for an item, we need the user to
rate (and thus consume) the item. But then we cannot recommend that item to the user again! Thus,
the only way to learn about a user?s preferences is through collaboration, or inferring from other
users? ratings. Broadly, such inference is possible if the users preferences are somehow related.
In this paper, we assume a simple structure for shared user preferences. We posit that there are
k < n different types of users, where users of the same type have identical item preference vectors.
The number of types k represents the heterogeneity in the population. For ease of exposition, in this
paper we assume that a user belongs to each user type with probability 1/k. We refer to this model
as a latent source model, where each user type corresponds to a latent source of users. We remark
that there is evidence suggesting real movie recommendation data to be well modeled by clustering
of both users and items [21]. Our model only assumes clustering over users.
Our problem setup relates to some versions of the multi-armed bandit problem. A fundamental
difference between our setup and that of the standard stochastic multi-armed bandit problem [23, 8]
is that the latter allows each item to be recommended an infinite number of times. Thus, the solution
concept for the stochastic multi-armed bandit problem is to determine the best item (arm) and keep
choosing it [3]. This observation applies also to ?clustered bandits? [9], which like our work seeks to
capture collaboration between users. On the other hand, sleeping bandits [15] allow for the available
items at each time step to vary, but the analysis is worst-case in terms of which items are available
over time. In our setup, the sequence of items that are available is not adversarial. Our model
combines the collaborative aspect of clustered bandits with dynamic item availability from sleeping
bandits, where we impose a strict structure on how items become unavailable.
3
A Collaborative Filtering Algorithm and Its Performance Guarantee
This section presents our algorithm C OLLABORATIVE -G REEDY and its accompanying theoretical
performance guarantee. The algorithm is syntactically similar to the ?-greedy algorithm for multiarmed bandits [22], which explores items with probability ? and otherwise greedily chooses the best
item seen so far based on a plurality vote. In our algorithm, the greedy choice, or exploitation, uses
the standard cosine-similarity measure. The exploration, on the other hand, is split into two types, a
standard item exploration in which a user is recommended an item that she or he hasn?t consumed
yet uniformly at random, and a joint exploration in which all users are asked to provide a rating for
the next item in a shared, randomly chosen sequence of items. Let?s fill in the details.
Algorithm. At each time step t, either all the users are asked to explore, or an item is recommended
to each user by choosing the item with the highest score for that user. The pseudocode is described
in Algorithm 1. There are two types of exploration: random exploration, which is for exploring the
space of items, and joint exploration, which helps to learn about similarity between users. For a
pre-specified rate ? ? (0, 4/7], we set the probability of random exploration to be ?R (n) = 1/n?
Algorithm 1: C OLLABORATIVE -G REEDY
Input: Parameters ? ? [0, 1], ? ? (0, 4/7].
Select a random ordering ? of the items [m]. Define
?R (n) =
1
,
n?
and
?J (t) =
1
.
t?
for time step t = 1, 2, . . . , T do
With prob. ?R (n): (random exploration) for each user, recommend a random item that the
user has not rated.
With prob. ?J (t): (joint exploration) for each user, recommend the first item in ? that the user
has not rated.
With prob. 1 ? ?J (t) ? ?R (n): (exploitation) for each user u, recommend an item j that the
(t)
user has not rated and that maximizes score peuj , which depends on threshold ?.
end
3
(decaying with the number of users), and the probability of joint exploration to be ?J (t) = 1/t?
(decaying with time).2
(t)
(t)
Next, we define user u?s score peui for item i at time t. Recall that we observe Yui = {?1, 0, +1}
as user u?s rating for item i up to time t, where 0 indicates that no rating has been given yet. We
define
?P
(t)
?
(t)
? v?Neu(t) 1{Yvi = +1} if P
P
(t)
eu(t) 1{Yvi 6= 0} > 0,
(t)
v?N
peui ,
eu(t) 1{Yvi 6= 0}
v?N
?
?
1/2
otherwise,
where the neighborhood of user u is given by
eu(t) , {v ? [n] : hYeu(t) , Yev(t) i ? ?|supp(Yeu(t) ) ? supp(Yev(t) )|},
N
(t)
and Yeu consists of the revealed ratings of user u restricted to items that have been jointly explored.
In other words,
(
(t)
Yui if item i is jointly explored by time t,
(t)
e
Yui =
0
otherwise.
The neighborhoods are defined precisely by cosine similarity with respect to jointed explored items.
(t)
(t)
(t)
(t)
To see this, for users u and v with revealed ratings Yeu and Yev , let ?uv , supp(Yeu )?supp(Yev )
(t)
(t)
be the support overlap of Yeu and Yev , and let h?, ?i?uv be the dot product restricted to entries in
?uv . Then
(t)
(t)
(t)
(t)
hYeu , Yev i?uv
hYeu , Yev i
q
,
=q
|?uv |
(t)
(t)
(t)
(t)
hYeu , Yeu i?uv hYev , Yev i?uv
(t)
(t)
which is the cosine similarity of revealed rating vectors Yeu and Yev restricted to the overlap of
their supports. Thus, users u and v are neighbors if and only if their cosine similarity is at least ?.
Theoretical performance guarantee. We now state our main result on the proposed collaborative
filtering algorithm?s performance with respect to the objective stated in equation (1). We begin with
two reasonable, and seemingly necessary, conditions under which our the results will be established.
A1 No ?-ambiguous items. There exists some constant ? > 0 such that
|pui ? 1/2| ? ?
for all users u and items i. (Smaller ? corresponds to more noise.)
A2 ?-incoherence. There exist a constant ? ? [0, 1) such that if users u and v are of different
types, then their item preference vectors pu and pv satisfy
1
h2pu ? 1, 2pv ? 1i ? 4??2 ,
m
where 1 is the all ones vector. Note that a different way to write the left-hand side is
1
E[ m
hYu? , Yv? i], where Yu? and Yv? are fully-revealed rating vectors of users u and v, and
the expectation is over the random ratings of items.
The first condition is a low noise condition to ensure that with a finite number of samples, we can
correctly classify each item as either likable or unlikable. The incoherence condition asks that the
different user types are well-separated so that cosine similarity can tease apart the users of different
types over time. We provide some examples after the statement of the main theorem that suggest the
incoherence condition to be reasonable, allowing E[hYu? , Yv? i] to scale as ?(m) rather than o(m).
We assume that the number of users satisfies n = O(mC ) for some constant C > 1. This is without
loss of generality since otherwise, we can randomly divide the n users into separate population
2
For ease of presentation, we set the two explorations to have the same decay rate ?, but our proof easily
extends to encompass different decay rates for the two exploration types. Furthermore, the constant 4/7 ? ?
is not special. It could be different and only affects another constant in our proof.
4
pools, each of size O(mC ) and run the recommendation algorithm independently for each pool to
achieve the same overall performance guarantee.
Finally, we define ?, the minimum proportion of likable items for any user (and thus any user type):
Pm
1{pui > 1/2}
? , min i=1
.
m
u?[n]
Theorem 1. Let ? ? (0, 1) be some pre-specified tolerance. Take as input to C OLLABORATIVE G REEDY ? = 2?2 (1 + ?) where ? ? [0, 1) is as defined in A2, and ? ? (0, 4/7]. Under the latent
source model and assumptions A1 and A2, if the number of users n = O(mC ) satisfies
1 4 1/?
n = ? km log +
,
?
?
then for any Tlearn ? T ? ?m, the expected proportion of likable items recommended by
C OLLABORATIVE -G REEDY up until time T satisfies
(T )
r+
Tlearn
? 1?
(1 ? ?),
Tn
T
where
Tlearn = ?
log km
??
?4 (1 ? ?)2
1/(1??)
+
4 1/?
?
.
Theorem 1 says that there are Tlearn initial time steps for which the algorithm may be giving poor
recommendations. Afterward, for Tlearn < T < ?m, the algorithm becomes near-optimal, recommending a fraction of likable items 1?? close to what an optimal oracle algorithm (that recommends
all likable items first) would achieve. Then for time horizon T > ?m, we can no longer guarantee
that there are likable items left to recommend. Indeed, if the user types each have the same fraction
of likable items, then even an oracle recommender would use up the ?m likable items by this time.
Meanwhile, to give a sense of how long the learning period Tlearn is, note that when ? = 1/2, we
have Tlearn scaling as log2 (km), and if we choose ? close to 0, then Tlearn becomes nearly log(km).
In summary, after Tlearn initial time steps, the simple algorithm proposed is essentially optimal.
To gain intuition for incoherence condition A2, we calculate the parameter ? for three examples.
Example 1. Consider when there is no noise, i.e., ? = 21 . Then users? ratings are deterministic
given their user type. Produce k vectors of probabilities by drawing m independent Bernoulli( 12 )
random variables (0 or 1 with probability 21 each) for each user type. For any item i and pair of
?
users u and v of different types, Yui
? Yvi? is a Rademacher random variable (?1 with probability 12
each), and thus the inner product of two user rating vectors is equal to the sum of m Rademacher
q
log m
random variables. Standard concentration inequalities show that one may take ? = ?
to
m
satisfy ?-incoherence with probability 1 ? 1/poly(m).
Example 2. We expand on the previous example by choosing an arbitrary ? > 0 and making all
latent source probability vectors have entries equal to 21 ? ? with probability 12 each. As before let
?
user u and v are from different type. Now E[Yui
? Yvi? ] = ( 12 + ?)2 + ( 12 ? ?)2 ? 2( 41 ? ?2 ) = 4?2
1
?
?
2
if pui = pvi and E[Yui ? Yvi ] = 2( 4 ? ? ) ? ( 12 + ?)2 ? ( 21 ? ?)2 = ?4?2 if pui = 1 ? pvi .
The value of the inner product E[hYu? , Yv? i] is again equal to the sum of m Rademacher
random
q
log m
2
suffices to
variables, but this time scaled by 4? . For similar reasons as before, ? = ?
m
satisfy ?-incoherence with probability 1 ? 1/poly(m).
Example 3. Continuing with the previous example, now suppose each entry is 12 +? with probability
? ? (0, 1/2) and 12 ? ? with probability 1 ? ?. Then for two users u and v of different types,
?
2
2
pui = pvi with probability ?2 + (1 ? ?)2 . This implies that E[hYu? , Yq
v i] = 4m? (1 ? 2?) .
log m
Again, using standard concentration, this shows that ? = (1 ? 2?)2 + ?
suffices to satisfy
m
?-incoherence with probability 1 ? 1/poly(m).
5
4
Proof of Theorem 1
Recall that Xut is the indicator random variable for whether the item ?ut recommended to user u
at time t is likable, i.e., pu?ut > 1/2. Given assumption A1, this is equivalent to the event that
pu?ut ? 12 + ?. The expected proportion of likable items is
(T )
T
T
n
n
r+
1 XX
1 XX
=
E[Xut ] =
P(Xut = 1).
Tn
T n t=1 u=1
T n t=1 u=1
Our proof focuses on lower-bounding P(Xut = 1). The key idea is to condition on what we call the
?good neighborhood? event Egood (u, t):
n
n
neighbors from the same user type (?good neighbors?),
Egood (u, t) = at time t, user u has ?
5k
o
?tn1??
and ?
neighbors from other user types (?bad neighbors?) .
10km
This good neighborhood event will enable us to argue that after an initial learning time, with high
probability there are at most ? as many ratings from bad neighbors as there are from good neighbors.
The proof of Theorem 1 consists of two parts. The first part uses joint exploration to show that after
a sufficient amount of time, the good neighborhood event Egood (u, t) holds with high probability.
Lemma 1. For user u, after
1/(1??)
2 log(10kmn? /?)
t?
?4 (1 ? ?)2
time steps,
?4 (1 ? ?)2 t1??
n
? 12 exp ?
.
P(Egood (u, t)) ? 1 ? exp ?
8k
20
In the above lower bound, the first exponentially decaying term could be thought of as the penalty
for not having enough users in the system from the k user types, and the second decaying term could
be thought of as the penalty for not yet clustering the users correctly.
The second part of our proof to Theorem 1 shows that, with high probability, the good neighborhoods
have, through random exploration, accurately estimated the probability of liking each item. Thus,
we correctly classify each item as likable or not with high probability, which leads to a lower bound
on P(Xut = 1).
Lemma 2. For user u at time t, if the good neighborhood event Egood (u, t) holds and t ? ?m, then
?2 tn1??
1
1
P(Xut = 1) ? 1 ? 2m exp ?
? ?? ?.
40km
t
n
Here, the first exponentially decaying term could be thought of as the cost of not classifying items
correctly as likable or unlikable, and the last two decaying terms together could be thought of as the
cost of exploration (we explore with probability ?J (t) + ?R (n) = 1/t? + 1/n? ).
We defer the proofs of Lemmas 1 and 2 to the supplementary material. Combining these lemmas
and choosing appropriate constraints on the numbers of users and items, we produce the following
lemma.
Lemma 3. Let ? ? (0, 1) be some pre-specified tolerance. If the number of users n and items m
satisfy
n
4 4 1/? o
,
n ? max 8k log ,
? ?
1/(1??)
1/(1??)
2 log(10kmn? /?)
20 log(96/?)
4 1/?
,
,
?m ? t ? max
,
4
2
4
2
? (1 ? ?)
? (1 ? ?)
?
16m
40km
log
nt1?? ?
,
?2
?
then P(Xut = 1) ? 1 ? ?.
6
Proof. With the above conditions on n and t satisfied, we combine Lemmas 1 and 2 to obtain
?4 (1 ? ?)2 t1??
?2 tn1??
n
P(Xut = 1) ? 1 ? exp ?
? 12 exp ?
? 2m exp ?
8k
20
40km
1
?
1
?
?
?
?
? ? ? ? ? 1 ? ? ? ? ? = 1 ? ?.
t
n
4 8 8 4 4
Theorem 1 follows as a corollary to Lemma 3. As previously mentioned, without loss of generality,
we take n = O(mC ). Then with number of users n satisfying
1 4 1/?
,
O(mC ) = n = ? km log +
?
?
and for any time step t satisfying
?m ? t ? ?
log km
??
?4 (1 ? ?)2
1/(1??)
+
4 1/?
?
, Tlearn ,
we simultaneously meet all of the conditions of Lemma 3. Note that the upper bound on number of
users n appears since without it, Tlearn would depend on n (observe that in Lemma 3, we ask that t
be greater than a quantity that depends on n). Provided that the time horizon satisfies T ? ?m, then
n
n
T
T
(T )
r+
1 X X
(T ? Tlearn )(1 ? ?)
1 X X
P(Xut = 1) ?
(1 ? ?) =
?
,
Tn
Tn
Tn
T
u=1
u=1
t=Tlearn
t=Tlearn
yielding the theorem statement.
5
Experimental Results
We provide only a summary of our experimental results here, deferring full details to the supplementary material. We simulate an online recommendation system based on movie ratings from the
Movielens10m and Netflix datasets, each of which provides a sparsely filled user-by-movie rating
matrix with ratings out of 5 stars. Unfortunately, existing collaborative filtering datasets such as the
two we consider don?t offer the interactivity of a real online recommendation system, nor do they
allow us to reveal the rating for an item that a user didn?t actually rate. For simulating an online system, the former issue can be dealt with by simply revealing entries in the user-by-item rating matrix
over time. We address the latter issue by only considering a dense ?top users vs. top items? subset
of each dataset. In particular, we consider only the ?top? users who have rated the most number of
items, and the ?top? items that have received the most number of ratings. While this dense part of the
dataset is unrepresentative of the rest of the dataset, it does allow us to use actual ratings provided
by users without synthesizing any ratings. A rigorous validation would require an implementation
of an actual interactive online recommendation system, which is beyond the scope of our paper.
First, we validate that our latent source model is reasonable for the dense parts of the two datasets
we consider by looking for clustering behavior across users. We find that the dense top users vs. top
movies matrices do in fact exhibit clustering behavior of users and also movies, as shown in Figure
1(a). The clustering was found via Bayesian clustered tensor factorization, which was previously
shown to model real movie ratings data well [21].
Next, we demonstrate our algorithm C OLLABORATIVE -G REEDY on the two simulated online movie
recommendation systems, showing that it outperforms two existing recommendation algorithms
Popularity Amongst Friends (PAF) [4] and a method by Deshpande and Montanari (DM) [12]. Following the experimental setup of [4], we quantize a rating of 4 stars or more to be +1 (likable), and
a rating less than 4 stars to be ?1 (unlikable). While we look at a dense subset of each dataset, there
are still missing entries. If a user u hasn?t rated item j in the dataset, then we set the corresponding
true rating to 0, meaning that in our simulation, upon recommending item j to user u, we receive
0 reward, but we still mark that user u has consumed item j; thus, item j can no longer be recommended to user u. For both Movielens10m and Netflix datasets, we consider the top n = 200 users
and the top m = 500 movies. For Movielens10m, the resulting user-by-rating matrix has 80.7%
nonzero entries. For Netflix, the resulting matrix has 86.0% nonzero entries. For an algorithm that
7
Average cumulative reward
0
50
100
150
200
0
100
200
300
400
60
50
40
30
20
Collaborative-Greedy
Popularity Amongst Friends
Deshpande-Montanari
10
0
?10
0
100
200
300
400
500
Time
500
(a)
(b)
Figure 1: Movielens10m dataset: (a) Top users by top movies matrix with rows and columns reordered to show clustering of users and items. (b) Average cumulative rewards over time.
recommends item ?ut to user u at time t, we measure the algorithm?s average cumulative reward up
PT Pn
(T )
to time T as n1 t=1 u=1 Yu?ut , where we average over users. For all four methods, we recommend items until we reach time T = 500, i.e., we make movie recommendations until each user has
seen all m = 500 movies. We disallow the matrix completion step for DM to see the users that we
actually test on, but we allow it to see the the same items as what is in the simulated online recommendation system in order to compute these items? feature vectors (using the rest of the users in the
dataset). Furthermore, when a rating is revealed, we provide DM both the thresholded rating and the
non-thresholded rating, the latter of which DM uses to estimate user feature vectors over time. We
discuss choice of algorithm parameters in the supplementary material. In short, parameters ? and
? of our algorithm are chosen based on training data, whereas we allow the other algorithms to use
whichever parameters give the best results on the test data. Despite giving the two competing algorithms this advantage, C OLLABORATIVE -G REEDY outperforms the two, as shown in Figure 1(b).
Results on the Netflix dataset are similar.
6
Discussion and Related Work
This paper proposes a model for online recommendation systems under which we can analyze the
performance of recommendation algorithms. We theoretical justify when a cosine-similarity collaborative filtering method works well, with a key insight of using two exploration types.
The closest related work is by Biau et al. [7], who study the asymptotic consistency of a cosinesimilarity nearest-neighbor collaborative filtering method. Their goal is to predict the rating of the
next unseen item. Barman and Dabeer [4] study the performance of an algorithm called Popularity Amongst Friends, examining its ability to predict binary ratings in an asymptotic informationtheoretic setting. In contrast, we seek to understand the finite-time performance of such systems.
Dabeer [11] uses a model similar to ours and studies online collaborative filtering with a moving
horizon cost in the limit of small noise using an algorithm that knows the numbers of user types and
item types. We do not model different item types, our algorithm is oblivious to the number of user
types, and our performance metric is different. Another related work is by Deshpande and Montanari [12], who study online recommendations as a linear bandit problem; their method, however,
does not actually use any collaboration beyond a pre-processing step in which offline collaborative
filtering (specifically matrix completion) is solved to compute feature vectors for items.
Our work also relates to the problem of learning mixture distributions (c.f., [10, 18, 5, 2]), where
one observes samples from a mixture distribution and the goal is to learn the mixture components
and weights. Existing results assume that one has access to the entire high-dimensional sample
or that the samples are produced in an exogenous manner (not chosen by the algorithm). Neither
assumption holds in our setting, as we only see each user?s revealed ratings thus far and not the user?s
entire preference vector, and the recommendation algorithm affects which samples are observed (by
choosing which item ratings are revealed for each user). These two aspects make our setting more
challenging than the standard setting for learning mixture distributions. However, our goal is more
modest. Rather than learning the k item preference vectors, we settle for classifying them as likable
or unlikable. Despite this, we suspect having two types of exploration to be useful in general for
efficiently learning mixture distributions in the active learning setting.
Acknowledgements. This work was supported in part by NSF grant CNS-1161964 and by Army
Research Office MURI Award W911NF-11-1-0036. GHC was supported by an NDSEG fellowship.
8
References
[1] Fabio Aiolli. A preliminary study on a recommender system for the million songs dataset challenge. In
Proceedings of the Italian Information Retrieval Workshop, pages 73?83, 2013.
[2] Anima Anandkumar, Rong Ge, Daniel Hsu, Sham M. Kakade, and Matus Telgarsky. Tensor decompositions for learning latent variable models, 2012. arXiv:1210.7559.
[3] Peter Auer, Nicol`o Cesa-Bianchi, and Paul Fischer. Finite-time analysis of the multiarmed bandit problem.
Machine Learning, 47(2-3):235?256, May 2002.
[4] Kishor Barman and Onkar Dabeer. Analysis of a collaborative filter based on popularity amongst neighbors. IEEE Transactions on Information Theory, 58(12):7110?7134, 2012.
[5] Mikhail Belkin and Kaushik Sinha. Polynomial learning of distribution families. In Foundations of
Computer Science (FOCS), 2010 51st Annual IEEE Symposium on, pages 103?112. IEEE, 2010.
[6] Thierry Bertin-Mahieux, Daniel P.W. Ellis, Brian Whitman, and Paul Lamere. The million song dataset.
In Proceedings of the 12th International Conference on Music Information Retrieval (ISMIR 2011), 2011.
[7] G?erard Biau, Beno??t Cadre, and Laurent Rouvi`ere. Statistical analysis of k-nearest neighbor collaborative
recommendation. The Annals of Statistics, 38(3):1568?1592, 2010.
[8] S?ebastien Bubeck and Nicol`o Cesa-Bianchi. Regret analysis of stochastic and nonstochastic multi-armed
bandit problems. Foundations and Trends in Machine Learning, 5(1):1?122, 2012.
[9] Loc Bui, Ramesh Johari, and Shie Mannor. Clustered bandits, 2012. arXiv:1206.4169.
[10] Kamalika Chaudhuri and Satish Rao. Learning mixtures of product distributions using correlations and
independence. In Conference on Learning Theory, pages 9?20, 2008.
[11] Onkar Dabeer. Adaptive collaborating filtering: The low noise regime. In IEEE International Symposium
on Information Theory, pages 1197?1201, 2013.
[12] Yash Deshpande and Andrea Montanari. Linear bandits in high dimension and recommendation systems,
2013. arXiv:1301.1722.
[13] Roger B. Grosse, Ruslan Salakhutdinov, William T. Freeman, and Joshua B. Tenenbaum. Exploiting
compositionality to explore a large space of model structures. In Uncertainty in Artificial Intelligence,
pages 306?315, 2012.
[14] Wassily Hoeffding. Probability inequalities for sums of bounded random variables. Journal of the American statistical association, 58(301):13?30, 1963.
[15] Robert Kleinberg, Alexandru Niculescu-Mizil, and Yogeshwer Sharma. Regret bounds for sleeping experts and bandits. Machine Learning, 80(2-3):245?272, 2010.
[16] Yehuda Koren. The BellKor solution to the Netflix grand prize. http://www.netflixprize.com/
assets/GrandPrize2009_BPC_BellKor.pdf, August 2009.
[17] Greg Linden, Brent Smith, and Jeremy York. Amazon.com recommendations: item-to-item collaborative
filtering. IEEE Internet Computing, 7(1):76?80, 2003.
[18] Ankur Moitra and Gregory Valiant. Settling the polynomial learnability of mixtures of gaussians. Proceedings of the 51st Annual IEEE Symposium on Foundations of Computer Science, 2010.
[19] Martin Piotte and Martin Chabbert. The pragmatic theory solution to the netflix grand prize. http://
www.netflixprize.com/assets/GrandPrize2009_BPC_PragmaticTheory.pdf, August 2009.
[20] Paul Resnick, Neophytos Iacovou, Mitesh Suchak, Peter Bergstrom, and John Riedl. Grouplens: An
open architecture for collaborative filtering of netnews. In Proceedings of the 1994 ACM Conference on
Computer Supported Cooperative Work, CSCW ?94, pages 175?186, New York, NY, USA, 1994. ACM.
[21] Ilya Sutskever, Ruslan Salakhutdinov, and Joshua B. Tenenbaum. Modelling relational data using
bayesian clustered tensor factorization. In NIPS, pages 1821?1828, 2009.
[22] Richard S. Sutton and Andrew G. Barto. Reinforcement Learning: An Introduction. MIT Press, Cambridge, MA, 1998.
[23] William R. Thompson. On the Likelihood that one Unknown Probability Exceeds Another in View of the
Evidence of Two Samples. Biometrika, 25:285?294, 1933.
[24] Andreas T?oscher and Michael Jahrer. The bigchaos solution to the netflix grand prize. http://www.
netflixprize.com/assets/GrandPrize2009_BPC_BigChaos.pdf, September 2009.
9
| 5330 |@word exploitation:3 version:1 polynomial:2 proportion:3 open:1 km:12 seek:2 simulation:1 decomposition:1 asks:2 initial:5 loc:1 score:3 pandora:1 daniel:2 movielens10m:4 ours:1 outperforms:2 existing:3 com:4 yet:5 nt1:1 john:1 v:2 greedy:3 intelligence:1 item:113 prize:4 short:2 smith:1 provides:1 mannor:1 preference:9 mahieux:1 become:2 symposium:3 focs:1 consists:2 wassily:1 combine:2 manner:1 indeed:1 expected:6 andrea:1 behavior:2 nor:1 multi:4 salakhutdinov:2 freeman:1 little:3 armed:4 actual:2 considering:2 becomes:2 provided:3 begin:1 xx:2 bounded:1 maximizes:1 didn:1 lowest:1 what:4 string:1 finding:1 guarantee:7 interactive:1 biometrika:1 tlearn:14 scaled:1 stick:1 grant:1 continually:2 before:3 t1:2 engineering:1 limit:1 despite:4 sutton:1 meet:1 incoherence:7 laurent:1 bergstrom:1 resembles:1 ankur:1 challenging:1 alice:1 ease:2 factorization:2 pui:8 practice:2 regret:2 yehuda:1 prevalence:1 cadre:1 thought:4 revealing:1 pre:4 word:1 suggest:1 cannot:2 close:2 selection:1 lamere:1 context:1 www:3 equivalent:2 deterministic:1 missing:1 maximizing:3 starting:1 independently:2 thompson:1 amazon:3 simplicity:1 immediately:2 insight:2 fill:1 population:2 beno:1 linkedin:1 annals:1 target:1 suppose:1 pt:1 user:131 us:5 trend:1 satisfying:2 egood:5 sparsely:1 muri:1 cooperative:1 observed:1 resnick:1 electrical:1 capture:1 worst:1 calculate:1 solved:1 news:1 earned:1 ordering:1 eu:3 highest:2 consumes:2 observes:1 mentioned:1 intuition:1 reward:8 asked:2 dynamic:1 personal:1 depend:1 bellkor:2 reordered:1 upon:2 whitman:1 easily:1 joint:6 represented:1 yash:1 separated:1 gbresler:1 artificial:1 netnews:1 choosing:5 neighborhood:7 netflixprize:3 yogeshwer:1 supplementary:3 say:2 consume:1 otherwise:5 drawing:1 expanse:1 ability:1 fischer:1 unseen:1 statistic:1 jointly:2 noisy:1 online:18 seemingly:1 sequence:2 advantage:1 product:4 combining:1 chaudhuri:1 achieve:2 validate:1 exploiting:1 sutskever:1 rademacher:3 produce:2 liked:4 telgarsky:1 help:1 friend:3 completion:2 andrew:1 nearest:2 thierry:1 received:1 job:1 implies:1 differ:1 posit:1 alexandru:1 filter:2 stochastic:3 exploration:22 enable:1 settle:1 material:3 require:1 assign:1 suffices:2 clustered:6 plurality:1 preliminary:1 brian:1 exploring:1 helping:1 rong:1 accompanying:1 hold:3 exp:6 scope:1 predict:2 matus:1 achieves:2 vary:1 a2:4 ruslan:2 grouplens:2 ere:1 establishes:1 mit:2 clearly:1 aim:1 rather:3 pn:1 barto:1 office:1 corollary:1 focus:2 she:3 bernoulli:1 indicates:2 modelling:1 likelihood:1 contrast:1 adversarial:1 greedily:1 rigorous:1 sense:2 inference:1 niculescu:1 entire:2 her:1 italian:1 bandit:17 expand:1 playlist:1 overall:1 issue:2 development:2 proposes:1 special:1 equal:3 once:3 having:2 identical:1 represents:1 yu:3 look:1 nearly:4 recommend:10 richard:1 oblivious:1 belkin:1 randomly:2 simultaneously:1 cns:1 n1:1 william:2 mixture:7 yielding:1 yui:8 necessary:1 experience:1 modest:2 filled:1 divide:1 continuing:1 theoretical:7 sinha:1 classify:2 column:1 elli:1 rao:1 suchak:1 w911nf:1 cost:3 introducing:1 entry:7 subset:2 examining:1 satish:1 learnability:1 gregory:1 chooses:1 st:2 fundamental:2 explores:1 international:2 grand:3 pool:2 michael:1 together:1 ilya:1 again:6 satisfied:1 ndseg:1 cesa:2 choose:1 prioritize:1 hoeffding:1 moitra:1 guy:1 brent:1 american:1 expert:1 supp:4 suggesting:1 jeremy:1 diversity:1 star:3 availability:1 matter:1 satisfy:5 hasn:3 depends:2 view:1 exogenous:1 analyze:3 johari:1 netflix:9 decaying:6 yv:4 defer:1 collaborative:22 contribution:2 greg:1 who:3 efficiently:1 biau:2 dealt:1 bayesian:2 thumb:1 accurately:1 produced:1 mc:5 asset:3 anima:1 reach:1 neu:1 iacovou:1 deshpande:4 dm:4 proof:9 boil:1 gain:1 hsu:1 xut:13 dataset:11 massachusetts:1 ask:1 recall:2 ut:8 listen:2 ubiquitous:1 actually:4 auer:1 jahrer:1 appears:1 generality:2 furthermore:2 roger:1 until:3 correlation:1 hand:3 somehow:1 reveal:1 usa:1 concept:1 true:1 former:2 nonzero:2 tn1:3 kaushik:1 ambiguous:1 cosine:9 won:1 prominent:1 pdf:3 outline:1 demonstrate:1 tn:5 syntactically:1 meaning:1 chaos:1 novel:2 common:1 pseudocode:1 overview:1 exponentially:2 million:3 association:1 he:3 refer:1 multiarmed:2 cambridge:2 uv:7 consistency:1 pm:1 similarly:1 inclusion:1 dabeer:4 dot:1 moving:1 access:1 similarity:11 longer:2 pu:5 closest:1 recent:1 belongs:1 apart:1 scenario:1 inequality:2 binary:1 success:1 life:1 joshua:2 seen:2 minimum:1 george:1 greater:1 impose:2 sharma:1 determine:1 maximize:4 recommended:17 period:1 relates:2 liking:2 encompass:1 full:1 sham:1 exceeds:1 offer:1 long:1 retrieval:2 award:1 a1:3 watched:1 essentially:2 expectation:1 metric:1 arxiv:3 tailored:1 sleeping:4 receive:1 whereas:1 fellowship:1 source:6 operate:1 rest:2 strict:1 suspect:2 likable:27 shie:1 effectiveness:1 call:1 anandkumar:1 near:2 revealed:7 split:1 recommends:4 enough:1 affect:2 independence:1 nonstochastic:1 competing:1 architecture:1 inner:2 idea:2 andreas:1 knowing:1 consumed:5 whether:3 handled:1 penalty:2 song:8 peter:2 york:2 action:1 remark:1 useful:1 amount:1 extensively:1 tenenbaum:2 http:3 exist:1 nsf:1 estimated:1 per:1 correctly:4 popularity:4 broadly:1 write:1 key:4 four:1 threshold:1 neither:1 thresholded:2 vast:1 fuse:1 merely:1 fraction:2 sum:3 run:1 prob:3 uncertainty:1 extends:1 family:1 reasonable:4 ismir:1 prefer:1 jointed:1 scaling:1 bound:4 internet:1 koren:1 oracle:3 annual:2 constraint:2 precisely:1 nearby:1 pvi:3 aspect:2 simulate:1 kleinberg:1 min:1 martin:2 department:1 according:1 poor:1 riedl:1 across:3 smaller:1 kakade:1 deferring:1 making:1 restricted:3 equation:1 previously:2 devavrat:2 discus:2 know:2 ge:1 whichever:1 end:1 available:3 gaussians:1 observe:2 appropriate:1 simulating:1 encounter:1 shah:1 assumes:1 clustering:7 include:1 ensure:1 top:10 log2:1 music:3 giving:2 especially:1 tensor:3 objective:2 already:2 quantity:1 concentration:2 dependence:1 exhibit:1 amongst:4 september:1 fabio:1 separate:2 simulated:2 argue:2 reason:1 length:1 index:1 modeled:1 setup:5 unfortunately:1 robert:1 statement:2 stated:1 synthesizing:1 implementation:1 ebastien:1 unknown:1 allowing:1 recommender:2 upper:1 observation:1 bianchi:2 datasets:4 finite:3 ramesh:1 mitesh:1 yev:9 heterogeneity:1 relational:1 looking:1 frame:1 arbitrary:2 august:2 rating:36 compositionality:1 cast:1 pair:1 specified:3 distinction:1 established:1 nip:1 address:3 beyond:2 regime:1 challenge:3 including:2 max:2 overlap:2 event:5 settling:1 indicator:2 arm:1 mizil:1 movie:18 technology:1 rated:5 yq:1 barman:2 prior:1 literature:2 taste:1 acknowledgement:1 dislike:3 nicol:2 determining:1 relative:1 asymptotic:2 fully:1 bresler:1 loss:2 interesting:1 interactivity:1 filtering:19 afterward:1 bertin:1 validation:1 foundation:3 sufficient:3 classifying:2 share:1 collaboration:3 row:1 course:1 summary:2 supported:3 last:1 tease:1 offline:1 yvi:6 allow:5 side:1 disallow:1 institute:1 neighbor:10 understand:1 mikhail:1 tolerance:2 dimension:1 cumulative:3 made:1 adaptive:1 wouldn:1 reinforcement:1 far:3 transaction:1 ignore:1 informationtheoretic:1 bui:1 keep:3 active:1 buy:1 recommending:11 don:1 latent:8 decade:1 why:1 learn:4 unavailable:1 quantize:1 poly:3 meanwhile:1 main:5 dense:5 montanari:4 bounding:1 noise:5 paul:3 kmn:2 aiolli:1 grosse:1 ny:1 probing:1 inferring:1 pv:2 winning:1 down:2 theorem:8 bad:2 showing:1 explored:3 decay:2 linden:1 evidence:2 exists:1 workshop:1 kamalika:1 valiant:1 onkar:2 horizon:3 chen:1 gap:2 reedy:6 logarithmic:1 simply:1 explore:5 army:1 bubeck:1 erard:1 cscw:1 ordered:1 watch:2 recommendation:33 applies:1 corresponds:3 collaborating:1 chance:1 constantly:1 satisfies:4 ma:2 acm:2 goal:5 presentation:1 exposition:2 unrepresentative:1 shared:2 infinite:1 specifically:1 uniformly:1 justify:2 lemma:10 total:1 called:1 experimental:4 vote:1 pragmatic:2 formally:1 select:1 support:2 mark:1 latter:3 |
4,784 | 5,331 | Clustering from Labels and Time-Varying Graphs
Shiau Hong Lim
National University of Singapore
[email protected]
Yudong Chen
EECS, University of California, Berkeley
[email protected]
Huan Xu
National University of Singapore
[email protected]
Abstract
We present a general framework for graph clustering where a label is observed to
each pair of nodes. This allows a very rich encoding of various types of pairwise
interactions between nodes. We propose a new tractable approach to this problem
based on maximum likelihood estimator and convex optimization. We analyze our
algorithm under a general generative model, and provide both necessary and sufficient conditions for successful recovery of the underlying clusters. Our theoretical
results cover and subsume a wide range of existing graph clustering results including planted partition, weighted clustering and partially observed graphs. Furthermore, the result is applicable to novel settings including time-varying graphs such
that new insights can be gained on solving these problems. Our theoretical findings are further supported by empirical results on both synthetic and real data.
1
Introduction
In the standard formulation of graph clustering, we are given an unweighted graph and seek a partitioning of the nodes into disjoint groups such that members of the same group are more densely
connected than those in different groups. Here, the presence of an edge represents some sort of
affinity or similarity between the nodes, and the absence of an edge represents the lack thereof.
In many applications, from chemical interactions to social networks, the interactions between nodes
are much richer than a simple ?edge? or ?non-edge?. Such extra information may be used to improve
the clustering quality. We may represent each type of interaction by a label. One simple setting of
this type is weighted graphs, where instead of a 0-1 graph, we have edge weights representing
the strength of the pairwise interaction. In this case the observed label between each pair is a
real number. In a more general setting, the label need not be a number. For example, on social
networks like Facebook, the label between two persons may be ?they are friends?, ?they went to
different schools?, ?they liked 21 common pages?, or the concatenation of these. In such cases
different labels carry different information about the underlying community structure. Standard
approaches convert these pairwise interactions into a simple edge/non-edge, and then apply standard
clustering algorithms, which might lose much of the information. Even in the case of a standard
weighted/unweighted graph, it is not immediately clear how the graph should be used. For example,
should the absence of an edge be interpreted as a neutral observation carrying no information, or as
a negative observation which indicates dissimilarity between the two nodes?
We emphasize that the forms of labels can be very general. In particular, a label can take the form
of a time series, i.e., the record of time varying interaction such as ?A and B messaged each other
on June 1st, 4th, 15th and 21st?, or ?they used to be friends, but they stop seeing each other since
2012?. Thus, the labeled graph model is an immediate tool for analyzing time-varying graphs.
1
In this paper, we present a new and principled approach for graph clustering that is directly based on
pairwise labels. We assume that between each pair of nodes i and j, a label Lij is observed which
is an element of a label set L. The set L may be discrete or continuous, and need not have any
structure. The standard graph model corresponds to a binary label set L = {edge, non-edge}, and
a weighted graph corresponds to L = R. Given the observed labels L = (Lij ) ? Ln?n , the goal
is to partition the n nodes into disjoint clusters. Our approach is based on finding a partition that
optimizes a weighted objective appropriately constructed from the observed labels. This leads to a
combinatorial optimization problem, and our algorithm uses its convex relaxation.
To systematically evaluate clustering performance, we consider a generalization of the stochastic
block model [1] and the planted partition model [2]. Our model assumes that the observed labels
are generated based on an underlying set of ground truth clusters, where pairs from the same cluster generate labels using a distribution ? over L and pairs from different clusters use a different
distribution ?. The standard stochastic block model corresponds to the case where ? and ? are twopoint distributions with ?(edge) = p and ?(edge) = q. We provide theoretical guarantees for our
algorithm under this generalized model.
Our results cover a wide range of existing clustering settings?with equal or stronger theoretical
guarantees?including the standard stochastic block model, partially observed graphs and weighted
graphs. Perhaps surprisingly, our framework allows us to handle new classes of problems that are not
a priori obvious to be a special case of our model, including the clustering of time-varying graphs.
1.1
Related work
The planted partition model/stochastic block model [1, 2] are standard models for studying graph
clustering. Variants of the models cover partially observed graphs [3, 4] and weighted graphs [5, 6].
All these models are special cases of ours. Various algorithms have been proposed and analyzed
under these models, such as spectral clustering [7, 8, 1], convex optimization approaches [9, 10, 11]
and tensor decomposition methods [12]. Ours is based on convex optimization; we build upon and
extend the approach in [13], which is designed for clustering unweighted graphs whose edges have
different levels of uncertainty, a special case of our problem (cf. Section 4.2 for details).
Most related to our setting is the labelled stochastic block model proposed in [14] and [15]. A
main difference in their model is that they assume each observation is a two-step process: first
an edge/non-edge is observed; if it is an edge then a label is associated with it. In our model
all observations are in the form of labels; in particular, an edge or no-edge is also a label. This
covers their setting as a special case. Our model is therefore more general and natural?as a result
our theory covers a broad class of subproblems including time-varying graphs. Moreover, their
analysis is mainly restricted to the two-cluster setting with edge probabilities on the order of ?(1/n),
while we allow for an arbitrary number of clusters and a wide range of edge/label distributions.
In addition, we consider the setting where the distributions of the labels are not precisely known.
Algorithmically, they use belief propagation [14] and spectral methods [15].
Clustering time-varying graphs has been studied in various context; see [16, 17, 18, 19, 20] and
the references therein. Most existing algorithms use heuristics and lack theoretical analysis. Our
approach is based on a generative model and has provable performance guarantees.
2
Problem setup and algorithms
We assume n nodes are partitioned into r disjoint clusters of size at least K, which are unknown and
considered as the ground truth. For each pair (i, j) of nodes, a label Lij ? L is observed, where L is
the set of all possible labels.1 These labels are generated independently across pairs according to the
distributions ? and ?. In particular, the probability of observing the label Lij is ?(Lij ) if i and j are
in the same cluster, and ?(Lij ) otherwise. The goal is to recover the ground truth clusters given the
labels. Let L = (Lij ) ? Ln?n be the matrix of observed labels. We represent the true clusters by
an n ? n cluster matrix Y ? , where Yij? = 1 if nodes i and j belong to the same cluster and Yij? = 0
otherwise (we use the convention Yii? = 1 for all i). The problem is therefore to find Y ? given L.
1
Note that L does not have to be finite. Although some of the results are presented for finite L, they can be
easily adapted to the other cases, for instance, by replacing summation with integration.
2
We take an optimization approach to this problem. To motivate our algorithm, first consider the
case of clustering a weighted graph, where all labels are real numbers. Positive weights indicate
in-cluster interaction while negative weights indicate cross-cluster interaction. A natural approach
is to cluster the nodes in a way that maximizes the total weight inside the clusters (this is equivalent
to correlation clustering
P [21]). Mathematically, this is to find a clustering, represented by a cluster
matrix Y , such that i,j Lij Yij is maximized. For the case of general labels, we pick a weight
function w : L 7? R, which assigns a number Wij = w(Lij ) to each label, and then solve the
following max-weight problem:
max hW, Y i s.t. Y is a cluster matrix;
Y
(1)
P
here hW, Y i := ij Wij Yij is the standard trace inner product. Note that this effectively converts
the problem of clustering from labels into a weighted clustering problem.
The program (1) is non-convex due to the constraint. Our algorithm is based on a convex relaxation
of (1), using the now well-known fact that a cluster matrix is a block-diagonal 0-1 matrix and thus
has nuclear norm2 equal to n [22, 3, 23]. This leads to the following convex optimization problem:
max
Y
s.t.
hW, Y i
(2)
kY k? ? n; 0 ? Yij ? 1, ?(i, j).
We say that this program succeeds if it has a unique optimal solution equal to the true cluster matrix
Y ? . We note that a related approach is considered in [13], which is discussed in section 4.
One has the freedom of choosing the weight function w. Intuitively, w should assign w(Lij ) > 0
to a label Lij with ?(Lij ) > ?(Lij ), so the program (2) is encouraged to place i and j in the same
cluster, the more likely possibility; similarly we should have w(Lij ) < 0 if ?(Lij ) < ?(Lij ). A
good weight function should further reflect the information in ? and ?. Our theoretical results in
section 3 characterize the performance of the program (2) for any given weight function; building
on this, we further derive the optimal choice for the weight function.
3
Theoretical results
In this section, we provide theoretical analysis for the performance of the convex program (2) under
the probabilistic model described in section 2. The proofs are given in the supplementary materials.
Our main result is a general theorem that gives sufficient conditions for (2) to recover the true cluster
matrix Y ? . The conditions are stated in terms of the label distribution ? and ?,Pthe minimum size
of the true clusters K, and any given weight function w. Define E? w :=
l?L w(l)?(l) and
P
Var? w := l?L [w(l) ? E? w]2 ?(l); E? w and Var? w are defined similarly.
Theorem 1 (Main). Suppose b is any number that satisfies |w(l)| ? b, ?l ? L almost surely. There
exists a universal constant c > 0 such that if
?
?
b log n + K log n Var? w
?E? w ? c
,
(3)
K p
?
b log n + n log n max(Var? w, Var? w)
E? w ? c
,
(4)
K
then Y ? is the unique solution to (2) with probability at least 1 ? n?10 . 3
The theorem holds for any given weight function w. In the next two subsections, we show how to
choose w optimally, and then address the case where w deviates from the optimal choice.
3.1
Optimal weights
A good candidate for the weight function w can be derived from the maximum likelihood estimator (MLE) of Y ? . Given the observed labels L, the log-likelihood of the true cluster matrix taking
2
The nuclear norm of a matrix is defined as the sum of its singular values. A cluster matrix is positive
semidefinite so its nuclear norm is equal to its trace.
3
In all our results, the choice n?10 is arbitrary. In particular, the constant c scales linearly with the exponent.
3
the value Y is
log Pr(L|Y ? = Y ) =
X
log ?(Lij )Yij ?(Lij )1?Yij = hW, Y i + c
i,j
where c is independent of Y and W is given by the weight function w(l) = wMLE (l) := log ?(l)
?(l) .
MLE
The MLE thus corresponds to using the log-likelihood ratio w
(?) as the weight function. The
following theorem is a consequence of Theorem 1 and characterizes the performance of using the
MLE weights. In the sequel, we use D(?k?) to denote the KL divergence between two distributions.
MLE
Theorem 2 (MLE). Suppose
w is used, and b and ? are any numbers which satisfy with
?(l)
D(?||?) ? ?D(?||?) and log ?(l) ? b, ?l ? L. There exists a universal constant c > 0 such
that Y ? is the unique solution to (2) with probability at least 1 ? n?10 if
log n
D(?||?) ? c(b + 2)
,
K
n log n
.
D(?||?) ? c(? + 1)(b + 2)
K2
Moreover, we always have D(?||?) ? (2b + 3)D(?||?), so we can take ? = 2b + 3.
(5)
(6)
Note that the theorem has the intuitive interpretation that the in/cross-cluster label distributions ? and
? should be sufficiently different, measured by their KL divergence. Using a classical result in information theory [24], we may replace the KL divergences with a quantity that is often easier to work
with, as summarized below. The LHS of (7) is sometimes called the triangle discrimination [24].
Corollary 1 (MLE 2). Suppose wMLE is used, and b, ? are defined as in Theorem 2. There exists a
universal constant c such that Y ? is the unique solution to (2) with probability at least 1 ? n?10 if
X (?(l) ? ?(l))2
n log n
? c(? + 1)(b + 2)
.
(7)
?(l) + ?(l)
K2
l?L
We may take ? = 2b + 3.
The MLE weight wMLE turns out to be near-optimal, at least in the two-cluster case, in the sense that
no other weight function (in fact, no other algorithm) has significantly better performance. This is
shown by establishing a necessary condition for any algorithm to recover Y ? . Here, an algorithm is
a measurable function Y? that maps the data L to a clustering (represented by a cluster matrix).
Theorem 3 (Converse). The following holds for some universal constants c, c0 > 0. Suppose K =
n
0
2 , and b defined in Theorem 2 satisfies b ? c . If
X (?(l) ? ?(l))2
c log n
?
,
(8)
?(l) + ?(l)
n
l?L
then inf Y? supY ? P(Y? =
6 Y ? ) ? 12 , where the supremum is over all possible cluster matrices.
Under the assumption of Theorem 3, the conditions (7) and (8) match up to a constant factor.
Remark. The MLE weight |wMLE (l)| becomes large if ?(l) = o(?(l)) or ?(l) = o(?(l)), i.e., when
the in-cluster probability is negligible compared to the cross-cluster one (or the other way around).
It can be shown that in this case the MLE weight is actually order-wise better than a bounded weight
function. We give this result in the supplementary material due to space constraints.
3.2
Monotonicity
We sometimes do not know the exact true distributions ? and ? to compute wMLE . Instead, we might
compute the weight using the log likelihood ratios of some ?incorrect? distribution ?
? and ??. Our
algorithm has a nice monotonicity property: as long as the divergence of the true ? and ? is larger
than that of ?
? and ?? (hence an ?easier? problem), then the problem should still have the same, if not
better probability of success, even though the wrong weights are used.
We say that (?, ?) is more divergent then (?
?, ??) if, for each l ? L, we have that either
?(l)
?(l)
?
?(l)
?(l)
?(l)
??(l)
?
?
? 1 or
?
?
? 1.
?(l)
??(l)
??(l)
?(l)
?
?(l)
?
?(l)
4
(l)
Theorem 4 (Monotonicity). Suppose we use the weight function w(l) = log ????(l)
, ?l, while the
actual label distributions are ? and ?. If the conditions in Theorem 2 or Corollary 1 hold with ?, ?
replaced by ?
?, ??, and (?, ?) is more divergent than (?
?, ??), then with probability at least 1 ? n?10
Y ? is the unique solution to (2).
This result suggests that one way to choose the weight function is by using the log-likelihood ratio
based on a ?conservative? estimate (i.e., a less divergent one) of the true label distribution pair.
3.3
Using inaccurate weights
In the previous subsection we consider using a conservative log-likelihood ratio as the weight. We
now consider a more general weight function w which need not be conservative, but is only required
to be not too far from the true log-likelihood ratio wMLE . Let
?(l)
?(l) := w(l) ? wMLE (l) = w(l) ? log
?(l)
P
P
be the error for each label l ? L. Accordingly, let ?? := l?L ?(l)?(l) and ?? := l?L ?(l)?(l)
be the average errors with respect to ? and ?. Note that ?? and ?? can be either positive or negative.
The following characterizes the performance of using such a w.
Theorem 5 (Inaccurate Weights). Let b and ? be defined as in Theorem 2. If the weight w satisfies
?(l)
, ?l ? L, |?? | ? ?D(?||?), |?? | ? ?D(?||?)
|w(l)| ? ? log
?(l)
for some ? < 1 and ? > 0. Then Y ? is unique solution to (2) with probability at least 1 ? n?10 if
?2
?2
log n
n log n
D(?||?) ? c
and D(?||?) ? c
(b + 2)
(? + 1)(b + 2)
.
(1 ? ?)2
K
(1 ? ?)2
K2
Therefore, as long as the errors ?? and ?? in w are not too large, the condition for recovery will be
order-wise similar to that in Theorem 2 for using the MLE weight. The numbers ? and ? measure
the amount of inaccuracy in w w.r.t. wMLE . The last two conditions in Theorem 5 thus quantify the
relation between the inaccuracy in w and the price we need to pay for using such a weight.
4
Consequences and applications
We apply the general results in the last section to different special cases. In sections 4.1 and 4.2, we
consider two simple settings and show that two immediate corollaries of our main theorems recover,
and in fact improve upon, existing results. In sections 4.3 and 4.4, we turn to the more complicated
setting of clustering time-varying graphs and derive several novel results.
4.1
Clustering a Gaussian matrix with partial observations
Analogous to the planted partition model for unweighted graphs, the bi-clustering [5] or submatrixlocalization [6, 23] problem concerns with weighted graph whose adjacency matrix has Gaussian
entries. We consider a generalization of this problem where some of the entries are unobserved.
n?n
Specifically, we observe a matrix L ? (R ? {?})
, which has r submatrices of size K ? K with
disjoint row and column support, such that Lij =? (meaning unobserved) with probability 1 ? s and
otherwise Lij ? N (uij , 1). Here the means of the Gaussians satisfy: uij = u
? if (i, j) is inside the
? > u ? 0. Clustering is equivalent to locating these
submatrices and uij = u if outside, where u
submatrices with elevated mean, given the large Gaussian matrix L with partial observations.4
This is a special case of our labeled framework with L = R ? {?}. Computing the log-likelihood
MLE
ratios for two Gaussians, we obtain wMLE (Lij ) = 0 if Lij =?,
(Lij ) ? Lij ? (?
u + u)/2
? and w
otherwise. This problem is interesting only when u
? ? u . log n (otherwise simple element-wise
thresholding [5, 6] finds the submatrices), which we assume to hold. Clearly D (?k?) = D (?k?) =
1
u ? u)2 . The following can be proved using our main theorems (proof in the appendix).
4 s(?
4
Here for simplicity we consider the clustering setting instead of bi-clustering. The latter setting corresponds
to rectangular L and submatrices. Extending our results to this setting is relatively straightforward.
5
Corollary 2 (Gaussian Graphs). Under the above setting, Y ? is the unique solution to (2) with
weights w = wMLE with probability at least 1 ? 2n?10 provided
2
s (?
u ? u) ? c
n log3 n
.
K2
In the fully observed case, this recovers the results in [23, 5, 6] up to log factors. Our results are
more general as we allow for partial observations, which is not considered in previous work.
4.2
Planted Partition with non-uniform edge densities
The work in [13] considers a variant of the planted partition model with non-uniform edge densities,
where each pair (i, j) has an edge with probability 1 ? uij > 1/2 if they are in the same cluster, and
with probability uij < 1/2 otherwise. The number uij can be considered as a measure of the level of
uncertainty in the observation between i and j, and is known or can be estimated in applications like
cloud-clustering. They show that using the knowledge of {uij } improves clustering performance,
and such a setting covers clustering of partially observed graphs that is considered in [11, 3, 4].
Here we consider a more general setting that does not require the in/cross-cluster edge density to be
symmetric around 12 . Suppose each pair (i, j) is associated with two numbers pij and qij , such that
if i and j are in the same cluster (different clusters, resp.), then there is an edge with probability pij
(qij , resp.); we know pij and qij but not which of them is the probability that generates the edge.
The values of pij and qij are generated i.i.d. randomly as (pij , qij ) ? D by some distribution D on
[0, 1] ? [0, 1]. The goal is to find the clusters given the graph adjacency matrix A, (pij ) and (qij ).
This model is a special case of our labeled framework. The labels have the form Lij =
(Aij , pij , qij ) ? L = {0, 1} ? [0, 1] ? [0, 1], generated by the distributions
pD(p, q),
l = (1, p, q)
qD(p, q),
l = (1, p, q)
?(l) =
?(l) =
(1 ? p)D(p, q), l = (0, p, q)
(1 ? q)D(p, q), l = (0, p, q).
p
1?p
ij
ij
+(1?Aij ) log 1?qij
. It turns out it is more
The MLE weight has the form wMLE (Lij ) = Aij log qij
convenient to use a conservative weight in which we replace pij and qij with p?ij = 34 pij + 14 qij and
q?ij = 14 pij + 34 qij . Applying Theorem 4 and Corollary 1, we immediately obtain the following.
Corollary 3 (Non-uniform Density). Program (2) recovers Y ? with probability at least 1 ? n?10 if
n log n
(pij ? qij )2
, ?(i.j).
?c
ED
pij (1 ? qij )
K2
Here ED is the expectation w.r.t. the distribution D, and LHS above is in fact independent of (i, j).
Corollary 3 improves upon existing results for several settings.
? Clustering partially observed graphs. Suppose D is such that pij = p and qij = q with probability s, and pij = qij otherwise, where p > q. This extends the standard planted partition model:
each pair is unobserved with probability 1 ? s. For this setting we require
s(p ? q)2
n log n
&
.
p(1 ? q)
K2
When s = 1. this matches the best existing bounds for standard planted partition [9, 12] up to a
log factor. For the partial observation setting with s ? 1, the work in [4] gives a similar bound
under the additional assumption p > 0.5 > q, which is not required by our result. For general
p and q, the best existing bound is given in [3, 9], which replaces unobserved entries with 0 and
s(p?q)2
log n
requires the condition p(1?sq)
& nK
. Our result is tighter when p and q are close to 1.
2
? Planted partition with non-uniformity. The model and algorithm in [13] isa special case
of ours
with symmetric densities pij ? 1 ? qij , for which we recover their result ED (1?2qij )2 & nlogn
K2 .
Corollary 3 is more general as it removes the symmetry assumption.
6
4.3
Clustering time-varying multiple-snapshot graphs
Standard graph clustering concerns with clustering on a single, static graph. We now consider a
setting where the graph can be time-varying. Specifically, we assume that for each time interval
t = 1, 2, . . . , T , we observed a snapshot of the graph L(t) ? Ln?n . We assume each snapshot is
generated by the distributions ? and ?, independent of other snapshots.
We can map this problem into our original labeled framework, by considering the whole time se? ij := (L(1) , . . . , L(T ) ) observed at the pair (i, j) as a single label. In this case the label
quence of L
ij
ij
set is thus the set of all possible sequences, i.e., L? = (L)T , and the label distributions are (with a
? ij ) = ?(L(1) ) . . . ?(L(T ) ), with ?(?) given similarly. The MLE weight
slight abuse of notation) ?(L
ij
ij
(normalized by T ) is thus the average log-likelihood ratio:
w
MLE
(1)
(T )
(t)
T
?(Lij ) . . . ?(Lij )
?(Lij )
1
1X
?
(Lij ) = log
log
=
.
(1)
(T )
(t)
T
T t=1
?(Lij ) . . . ?(Lij )
?(Lij )
? ij ) is the average of T independent random variables, its variance scales with
Since wMLE (L
Applying Theorem 1, with almost identical proof as in Theorem 2 we obtain the following:
1
T
.
Corollary 4 (Independent Snapshots). Suppose | log ?(l)
?(l) | ? b, ?l ? L and D(?||?) ? ?D(?||?).
?
The program (2) with MLE weights given recovers Y with probability at least 1 ? n?10 provided
log n
,
K
n log n
n log n o
D(?||?) ? c(b + 2) max
, (? + 1)
.
K
T K2
D(?||?) ? c(b + 2)
(9)
(10)
Setting T = 1 recovers Theorem 2. When the second term in (10) dominates, the corollary says that
the problem becomes easier if we observe more snapshots, with the tradeoff quantified precisely.
4.4
Markov sequence of snapshots
We now consider the more general and useful setting where the snapshots form a Markov chain. For
simplicity we assume that the Markov chain is time-invariant and has a unique stationary distribution
(t)
which is also the initial distribution. Therefore, the observations Lij at each (i, j) are generated by
first drawing a label from the stationary distribution ?
? (or ??) at t = 1, then applying a one-step
transition to obtain the label at each subsequent t. In particular, given the previously observed label
l, let the intra-cluster and inter-cluster conditional distributions be ?(?|l) and ?(?|l). We assume that
the Markov chains with respect to both ? and ? are geometrically ergodic such that for any ? ? 1,
and label-pair L(1) , L(? +1) ,
| Pr? (L(? +1) |L(1) ) ? ?
?(L(? +1) )| ? ?? ?
and
| Pr? (L(? +1) |L(1) ) ? ??(L(? +1) )| ? ?? ?
for some constants ? ? 1 and ? < 1 that only depend on ? and ?. Let Dl (?||?)
Pbe the KL-divergence
between ?(?|l) and ?(?|l); Dl (?||?) is similarly defined. Let E?? Dl (?||?) = l?L ?
?(l)Dl (?||?) and
similarly for E?? Dl (?||?). As in the previous subsection, we use the average log-likelihood ratio as
the weight. Define ? = (1??) minl?{??(l),?? (l)} . Applying Theorem 1 gives the following corollary.
See sections H?I in the supplementary material for the proof and additional discussion.
Corollary
Snapshots).
Under the above setting, suppose for each label-pair (l, l0 ),
5 (Markov
?
? (l)
?(l0 |l)
? ||?
?) ? ?D(?
?||?
? ) and E?? Dl (?||?) ? ?E?? Dl (?||?). The
log ??(l) ? b, log ?(l0 |l) ? b, D(?
program (2) with MLE weights recovers Y ? with probability at least 1 ? n?10 provided
1
1
log n
D(?
? ||?
?) + 1 ?
E?? Dl (?||?) ? c(b + 2)
T
T
K
n log n
1
1
n log n o
D(?
?||?
?) + 1 ?
E?? Dl (?||?) ? c(b + 2) max
, (? + 1)?
.
T
T
K
T K2
7
(11)
(12)
As an illuminating example, consider the case where ?
? ? ??, i.e., the marginal distributions for
individual snapshots are identical or very close. It means that the information is contained in the
change of labels, but not in the individual labels, as made evident in the LHSs of (11) and (12).
In this case, it is necessary to use the temporal information in order to perform clustering. Such
information would be lost if we disregard the ordering of the snapshots, for example, by aggregating
or averaging the snapshots then apply a single-snapshot clustering algorithm. This highlights an
essential difference between clustering time-varying graphs and static graphs.
5
Empirical results
To solve the convex program (2), we follow [13, 9] and adapt the ADMM algorithm by [25]. We
perform 100 trials for each experiment, and report the success rate, i.e., the fraction of trials where
the ground-truth clustering is fully recovered. Error bars show 95% confidence interval. Additional
empirical results are provided in the supplementary material.
We first test the planted partition model with partial observations under the challenging sparse (p
and q close to 0) and dense settings (p and q close to 1); cf. section 4.2. Figures 1 and 2 show the
results for n = 1000 with 4 equal-size clusters. In both cases, each pair is observed with probability
0.5. For comparison, we include results for the MLE weights as well as the linear weights (based on
linear approximation of the log-likelihood ratio), uniform weights and a imputation scheme where
all unobserved entries are assumed to be ?no-edge?.
q = 0.02, s = 0.5
14 weeks
p = 0.98, s = 0.5
1
1
0.9
0.6
0.4
0.2
MLE
linear
uniform
no partial
0.6
0.4
0.2
0
0.1
0.2
0.3
p?q
0.4
Figure 1: Sparse graphs
0.5
accuracy
0.8
MLE
linear
uniform
no partial
success rate
success rate
0.8
0.85
0.8
0.75
0
0.1
0.2
0.3
p?q
0.4
Figure 2: Dense graphs
0.5
0.7
0
Markov
independent
aggregate
0.005
0.01
0.015
0.02
fraction of data used in estimation
0.025
Figure 3: Reality Mining dataset
2
Corollary 3 predicts more success as the ratio s(p?q)
p(1?q) gets larger. All else being the same, distributions with small ? (sparse) are ?easier? to solve. Both predictions are consistent with the empirical
results in Figs. 1 and 2. The results also show that the MLE weights outperform the other weights.
For real data, we use the Reality Mining dataset [26], which contains individuals from two main
groups, the MIT Media Lab and the Sloan Business School, which we use as the ground-truth
clusters. The dataset records when two individuals interact, i.e., become proximal of each other or
make a phone call, over a 9-month period. We choose a window of 14 weeks (the Fall semester)
where most individuals have non-empty interaction data. These consist of 85 individuals with 25 of
them from Sloan. We represent the data as a time-varying graph with 14 snapshots (one per week)
and two labels?an ?edge? if a pair of individuals interact within the week, and ?no-edge? otherwise.
We compare three models: Markov sequence, independent snapshots, and the aggregate (union)
graphs. In each trial, the in/cross-cluster distributions are estimated from a fraction of randomly
selected pairwise interaction data. The vertical axis in Figure 3 shows the fraction of pairs whose
cluster relationship are correctly identified. From the results, we infer that the interactions between
individuals are likely not independent across time, and are better captured by the Markov model.
Acknowledgments
S.H. Lim and H. Xu were supported by the Ministry of Education of Singapore through AcRF Tier
Two grant R-265-000-443-112. Y. Chen was supported by NSF grant CIF-31712-23800 and ONR
MURI grant N00014-11-1-0688.
8
References
[1] K. Rohe, S. Chatterjee, and B. Yu. Spectral clustering and the high-dimensional stochastic
block model. Annals of Statistics, 39:1878?1915, 2011.
[2] A. Condon and R. M. Karp. Algorithms for graph partitioning on the planted partition model.
Random Structures and Algorithms, 18(2):116?140, 2001.
[3] S. Oymak and B. Hassibi. Finding dense clusters via low rank + sparse decomposition.
arXiv:1104.5186v1, 2011.
[4] Y. Chen, A. Jalali, S. Sanghavi, and H. Xu. Clustering partially observed graphs via convex
optimization. Journal of Machine Learning Research, 15:2213?2238, June 2014.
[5] Sivaraman Balakrishnan, Mladen Kolar, Alessandro Rinaldo, Aarti Singh, and Larry Wasserman. Statistical and computational tradeoffs in biclustering. In NIPS Workshop on Computational Trade-offs in Statistical Learning, 2011.
[6] Mladen Kolar, Sivaraman Balakrishnan, Alessandro Rinaldo, and Aarti Singh. Minimax localization of structural information in large noisy matrices. In NIPS, pages 909?917, 2011.
[7] F. McSherry. Spectral partitioning of random graphs. In FOCS, pages 529?537, 2001.
[8] K. Chaudhuri, F. Chung, and A. Tsiatas. Spectral clustering of graphs with general degrees in
the extended planted partition model. COLT, 2012.
[9] Y. Chen, S. Sanghavi, and H. Xu. Clustering sparse graphs. In NIPS 2012., 2012.
[10] B. Ames and S. Vavasis. Nuclear norm minimization for the planted clique and biclique problems. Mathematical Programming, 129(1):69?89, 2011.
[11] C. Mathieu and W. Schudy. Correlation clustering with noisy input. In SODA, page 712, 2010.
[12] Anima Anandkumar, Rong Ge, Daniel Hsu, and Sham M Kakade. A tensor spectral approach
to learning mixed membership community models. arXiv preprint arXiv:1302.2684, 2013.
[13] Y. Chen, S. H. Lim, and H. Xu. Weighted graph clustering with non-uniform uncertainties. In
ICML, 2014.
[14] Simon Heimlicher, Marc Lelarge, and Laurent Massouli?e. Community detection in the labelled
stochastic block model. In NIPS Workshop on Algorithmic and Statistical Approaches for
Large Social Networks, 2012.
[15] Marc Lelarge, Laurent Massouli?e, and Jiaming Xu. Reconstruction in the Labeled Stochastic
Block Model. In IEEE Information Theory Workshop, Seville, Spain, September 2013.
[16] S. Fortunato. Community detection in graphs. Physics Reports, 486(3-5):75?174, 2010.
[17] Jimeng Sun, Christos Faloutsos, Spiros Papadimitriou, and Philip S. Yu. Graphscope:
parameter-free mining of large time-evolving graphs. In ACM KDD, 2007.
[18] D. Chakrabarti, R. Kumar, and A. Tomkins. Evolutionary clustering. In ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 554?560. ACM, 2006.
[19] Vikas Kawadia and Sameet Sreenivasan. Sequential detection of temporal communities by
estrangement confinement. Scientific Reports, 2, 2012.
[20] N.P. Nguyen, T.N. Dinh, Y. Xuan, and M.T. Thai. Adaptive algorithms for detecting community structure in dynamic social networks. In INFOCOM, pages 2282?2290. IEEE, 2011.
[21] N. Bansal, A. Blum, and S. Chawla. Correlation clustering. Machine Learning, 56(1), 2004.
[22] A. Jalali, Y. Chen, S. Sanghavi, and H. Xu. Clustering partially observed graphs via convex
optimization. In ICML, 2011.
[23] Brendan P.W. Ames. Guaranteed clustering and biclustering via semidefinite programming.
Mathematical Programming, pages 1?37, 2013.
[24] F. Topsoe. Some inequalities for information divergence and related measures of discrimination. IEEE Transactions on Information Theory, 46(4):1602?1609, Jul 2000.
[25] Z. Lin, M. Chen, L. Wu, and Y. Ma. The augmented lagrange multiplier method for exact
recovery of corrupted low-rank matrices. Technical Report UILU-ENG-09-2215, UIUC, 2009.
[26] Nathan Eagle and Alex (Sandy) Pentland. Reality mining: Sensing complex social systems.
Personal Ubiquitous Comput., 10(4):255?268, March 2006.
[27] Alexandre B. Tsybakov. Introduction to Nonparametric Estimation. Springer, 2009.
9
| 5331 |@word trial:3 stronger:1 norm:3 c0:1 seek:1 condon:1 decomposition:2 eng:1 pick:1 carry:1 initial:1 mpexuh:1 series:1 contains:1 daniel:1 ours:3 existing:7 recovered:1 subsequent:1 partition:14 kdd:1 remove:1 designed:1 discrimination:2 stationary:2 generative:2 selected:1 accordingly:1 record:2 detecting:1 node:12 ames:2 semester:1 mathematical:2 constructed:1 become:1 chakrabarti:1 incorrect:1 qij:18 focs:1 inside:2 pairwise:5 inter:1 uiuc:1 actual:1 window:1 considering:1 becomes:2 provided:4 spain:1 underlying:3 moreover:2 maximizes:1 bounded:1 notation:1 medium:1 interpreted:1 finding:3 unobserved:5 guarantee:3 temporal:2 berkeley:2 k2:9 wrong:1 partitioning:3 converse:1 grant:3 positive:3 negligible:1 aggregating:1 consequence:2 encoding:1 analyzing:1 establishing:1 laurent:2 abuse:1 might:2 therein:1 studied:1 quantified:1 suggests:1 challenging:1 schudy:1 range:3 bi:2 unique:8 acknowledgment:1 lost:1 block:9 union:1 sq:1 nlogn:1 empirical:4 universal:4 submatrices:5 significantly:1 evolving:1 convenient:1 confidence:1 seeing:1 get:1 close:4 context:1 applying:4 equivalent:2 measurable:1 map:2 straightforward:1 independently:1 convex:11 rectangular:1 ergodic:1 simplicity:2 recovery:3 immediately:2 assigns:1 wasserman:1 estimator:2 insight:1 nuclear:4 handle:1 analogous:1 resp:2 annals:1 suppose:9 exact:2 programming:3 us:1 element:2 predicts:1 labeled:5 muri:1 observed:22 cloud:1 preprint:1 connected:1 sun:1 went:1 ordering:1 trade:1 principled:1 alessandro:2 pd:1 thai:1 dynamic:1 personal:1 motivate:1 carrying:1 solving:1 uniformity:1 depend:1 singh:2 upon:3 localization:1 triangle:1 easily:1 various:3 represented:2 aggregate:2 choosing:1 outside:1 whose:3 richer:1 heuristic:1 solve:3 supplementary:4 say:3 larger:2 otherwise:8 drawing:1 statistic:1 noisy:2 sequence:3 propose:1 reconstruction:1 interaction:12 product:1 yii:1 pthe:1 chaudhuri:1 intuitive:1 ky:1 shiau:1 cluster:44 empty:1 extending:1 xuan:1 liked:1 derive:2 friend:2 measured:1 ij:12 school:2 indicate:2 convention:1 quantify:1 qd:1 stochastic:8 larry:1 material:4 adjacency:2 education:1 require:2 assign:1 generalization:2 tighter:1 summation:1 mathematically:1 yij:7 rong:1 hold:4 sufficiently:1 considered:5 ground:5 around:2 algorithmic:1 week:4 sandy:1 aarti:2 estimation:2 applicable:1 lose:1 label:51 combinatorial:1 sivaraman:2 topsoe:1 tool:1 weighted:11 minimization:1 mit:1 clearly:1 offs:1 always:1 gaussian:4 varying:12 karp:1 corollary:13 derived:1 l0:3 june:2 quence:1 rank:2 likelihood:12 indicates:1 mainly:1 sigkdd:1 brendan:1 sense:1 membership:1 inaccurate:2 relation:1 uij:7 wij:2 colt:1 priori:1 exponent:1 special:8 integration:1 marginal:1 equal:5 encouraged:1 identical:2 represents:2 broad:1 yu:2 icml:2 papadimitriou:1 report:4 sanghavi:3 randomly:2 national:2 densely:1 divergence:6 individual:8 replaced:1 freedom:1 detection:3 biclique:1 possibility:1 mining:5 intra:1 analyzed:1 semidefinite:2 mcsherry:1 chain:3 edge:29 partial:7 necessary:3 lh:3 huan:1 theoretical:8 instance:1 column:1 cover:6 uilu:1 neutral:1 entry:4 uniform:7 successful:1 too:2 characterize:1 optimally:1 eec:2 corrupted:1 proximal:1 synthetic:1 person:1 st:2 density:5 oymak:1 international:1 sequel:1 probabilistic:1 physic:1 pbe:1 reflect:1 choose:3 chung:1 summarized:1 satisfy:2 sloan:2 lab:1 infocom:1 analyze:1 observing:1 characterizes:2 sort:1 recover:5 complicated:1 jul:1 simon:1 accuracy:1 variance:1 maximized:1 anima:1 ed:3 facebook:1 lelarge:2 thereof:1 obvious:1 associated:2 proof:4 recovers:5 static:2 stop:1 hsu:1 proved:1 dataset:3 lim:3 subsection:3 knowledge:2 improves:2 ubiquitous:1 actually:1 alexandre:1 follow:1 formulation:1 though:1 furthermore:1 correlation:3 tsiatas:1 replacing:1 propagation:1 lack:2 acrf:1 quality:1 perhaps:1 scientific:1 building:1 normalized:1 true:9 multiplier:1 twopoint:1 hence:1 chemical:1 symmetric:2 hong:1 generalized:1 bansal:1 evident:1 meaning:1 wise:3 novel:2 common:1 extend:1 belong:1 discussed:1 interpretation:1 elevated:1 slight:1 dinh:1 similarly:5 similarity:1 optimizes:1 inf:1 phone:1 n00014:1 inequality:1 binary:1 success:5 onr:1 captured:1 minimum:1 additional:3 ministry:1 surely:1 minl:1 period:1 multiple:1 sham:1 infer:1 technical:1 match:2 adapt:1 cross:5 long:2 lin:1 mle:21 prediction:1 variant:2 expectation:1 arxiv:3 represent:3 sometimes:2 jiaming:1 addition:1 interval:2 else:1 singular:1 appropriately:1 extra:1 norm2:1 member:1 balakrishnan:2 call:1 anandkumar:1 structural:1 near:1 presence:1 identified:1 confinement:1 inner:1 tradeoff:2 locating:1 cif:1 remark:1 useful:1 clear:1 se:1 amount:1 nonparametric:1 tsybakov:1 generate:1 vavasis:1 outperform:1 nsf:1 singapore:3 estimated:2 disjoint:4 algorithmically:1 per:1 correctly:1 discrete:1 group:4 blum:1 imputation:1 v1:1 graph:53 relaxation:2 geometrically:1 fraction:4 convert:2 sum:1 uncertainty:3 soda:1 massouli:2 place:1 almost:2 extends:1 wu:1 appendix:1 bound:3 pay:1 guaranteed:1 replaces:1 eagle:1 strength:1 adapted:1 precisely:2 constraint:2 alex:1 generates:1 nathan:1 kumar:1 relatively:1 according:1 march:1 across:2 partitioned:1 kakade:1 intuitively:1 restricted:1 pr:3 invariant:1 tier:1 ln:3 previously:1 turn:3 know:2 ge:1 tractable:1 studying:1 gaussians:2 apply:3 observe:2 spectral:6 chawla:1 faloutsos:1 vikas:1 original:1 assumes:1 clustering:48 cf:2 include:1 tomkins:1 build:1 classical:1 tensor:2 objective:1 quantity:1 planted:13 diagonal:1 jalali:2 september:1 affinity:1 evolutionary:1 concatenation:1 philip:1 considers:1 provable:1 relationship:1 ratio:10 kolar:2 setup:1 subproblems:1 trace:2 negative:3 stated:1 fortunato:1 unknown:1 perform:2 vertical:1 observation:11 snapshot:15 markov:8 finite:2 mladen:2 pentland:1 immediate:2 subsume:1 extended:1 jimeng:1 arbitrary:2 community:6 pair:17 required:2 kl:4 california:1 nu:2 inaccuracy:2 nip:4 address:1 bar:1 below:1 program:9 including:5 max:6 belief:1 natural:2 business:1 representing:1 scheme:1 improve:2 minimax:1 mathieu:1 axis:1 lij:34 deviate:1 nice:1 sg:2 discovery:1 fully:2 messaged:1 highlight:1 mixed:1 interesting:1 var:5 illuminating:1 supy:1 degree:1 sufficient:2 pij:15 consistent:1 thresholding:1 systematically:1 row:1 supported:3 surprisingly:1 last:2 free:1 aij:3 allow:2 wide:3 fall:1 taking:1 sparse:5 yudong:2 transition:1 rich:1 unweighted:4 made:1 adaptive:1 nguyen:1 far:1 social:5 log3:1 transaction:1 emphasize:1 supremum:1 monotonicity:3 clique:1 assumed:1 continuous:1 reality:3 symmetry:1 interact:2 complex:1 marc:2 main:6 dense:3 linearly:1 whole:1 xu:7 augmented:1 fig:1 hassibi:1 christos:1 comput:1 candidate:1 hw:4 theorem:24 rohe:1 sensing:1 divergent:3 concern:2 dominates:1 exists:3 dl:9 essential:1 consist:1 workshop:3 effectively:1 gained:1 sequential:1 dissimilarity:1 chatterjee:1 nk:1 chen:8 easier:4 likely:2 rinaldo:2 lagrange:1 contained:1 partially:7 biclustering:2 springer:1 corresponds:5 truth:5 satisfies:3 acm:3 ma:1 conditional:1 goal:3 month:1 labelled:2 replace:2 absence:2 price:1 change:1 admm:1 specifically:2 averaging:1 conservative:4 total:1 called:1 disregard:1 succeeds:1 support:1 latter:1 evaluate:1 |
4,785 | 5,332 | Discrete Graph Hashing
?
Wei Liu? Cun Mu? Sanjiv Kumar Shih-Fu Chang?
IBM T. J. Watson Research Center ? Columbia University Google Research
[email protected] [email protected]
[email protected] [email protected]
Abstract
Hashing has emerged as a popular technique for fast nearest neighbor search in gigantic databases. In particular, learning based hashing has received considerable
attention due to its appealing storage and search ef?ciency. However, the performance of most unsupervised learning based hashing methods deteriorates rapidly
as the hash code length increases. We argue that the degraded performance is due
to inferior optimization procedures used to achieve discrete binary codes. This
paper presents a graph-based unsupervised hashing model to preserve the neighborhood structure of massive data in a discrete code space. We cast the graph
hashing problem into a discrete optimization framework which directly learns the
binary codes. A tractable alternating maximization algorithm is then proposed to
explicitly deal with the discrete constraints, yielding high-quality codes to well
capture the local neighborhoods. Extensive experiments performed on four large
datasets with up to one million samples show that our discrete optimization based
graph hashing method obtains superior search accuracy over state-of-the-art unsupervised hashing methods, especially for longer codes.
1
Introduction
During the past few years, hashing has become a popular tool for tackling a variety of large-scale
computer vision and machine learning problems including object detection [6], object recognition
[35], image retrieval [22], linear classi?er training [19], active learning [24], kernel matrix approximation [34], multi-task learning [36], etc. In these problems, hashing is exploited to map similar
data points to adjacent binary hash codes, thereby accelerating similarity search via highly ef?cient
Hamming distances in the code space. In practice, hashing with short codes, say about one hundred
bits per sample, can lead to signi?cant gains in both storage and computation. This scenario is called
Compact Hashing in the literature, which is the focus of this paper.
Early endeavors in hashing concentrated on using random permutations or projections to construct
randomized hash functions. The well-known representatives include Min-wise Hashing (MinHash)
[3] and Locality-Sensitive Hashing (LSH) [2]. MinHash estimates the Jaccard set similarity and is
improved by b-bit MinHash [18]. LSH can accommodate a variety of distance or similarity metrics
such as p distances for p ? (0, 2], cosine similarity [4], and kernel similarity [17]. Due to randomized hashing, one needs more bits per hash table to achieve high precision. This typically reduces
recall, and multiple hash tables are thus required to achieve satisfactory accuracy of retrieved nearest
neighbors. The overall number of hash bits used in an application can easily run into thousands.
Beyond the data-independent randomized hashing schemes, a recent trend in machine learning is to
develop data-dependent hashing techniques that learn a set of compact hash codes using a training
set. Binary codes have been popular in this scenario for their simplicity and ef?ciency in computation. The compact hashing scheme can accomplish almost constant-time nearest neighbor search,
after encoding the whole dataset to short binary codes and then aggregating them into a hash table.
Additionally, compact hashing is particularly bene?cial to storing massive-scale data. For example, saving one hundred million samples each with 100 binary bits costs less than 1.5 GB, which
1
can easily ?t in memory. To create effective compact codes, several methods have been proposed.
These include the unsupervised methods, e.g., Iterative Quantization [9], Isotropic Hashing [14],
Spectral Hashing [38, 37], and Anchor Graph Hashing [23], the semi-supervised methods, e.g.,
Weakly-Supervised Hashing [25], and the supervised methods, e.g., Semantic Hashing [30], Binary
Reconstruction Embeddings [16], Minimal Loss Hashing [27], Kernel-based Supervised Hashing
[22], Hamming Distance Metric Learning [28], and Column Generation Hashing [20].
This paper focuses on the problem of unsupervised learning of compact hash codes. Here we argue
that most unsupervised hashing methods suffer from inadequate search performance, particularly
low recall, when applied to learn relatively longer codes (say around 100 bits) in order to achieve
higher precision. The main reason is that the discrete (binary) constraints which should be imposed
on the codes during learning itself have not been treated adequately. Most existing methods either
neglect the discrete constraints like PCA Hashing and Isotropic Hashing, or discard the constraints
to solve the relaxed optimizations and afterwards round the continuous solutions to obtain the binary codes like Spectral Hashing and Anchor Graph Hashing. Crucially, we ?nd that the hashing
performance of the codes obtained by such relaxation + rounding schemes deteriorates rapidly when
the code length increases (see Fig. 2). Till now, very few approaches work directly in the discrete
code space. Parameter-Sensitive Hashing [31] and Binary Reconstruction Embeddings (BRE) learn
the parameters of prede?ned hash functions by progressively tuning the codes generated by such
functions; Iterative Quantization (ITQ) iteratively learns the codes by explicitly imposing the binary
constraints. While ITQ and BRE work in the discrete space to generate the hash codes, they do not
capture the local neighborhoods of raw data in the code space well. ITQ targets at minimizing the
quantization error between the codes and the PCA-reduced data. BRE trains the Hamming distances
to mimic the 2 distances among a limited number of sampled data points, but could not incorporate
the entire dataset into training due to its expensive optimization procedure.
In this paper, we leverage the concept of Anchor Graphs [21] to capture the neighborhood structure
inherent in a given massive dataset, and then formulate a graph-based hashing model over the whole
dataset. This model hinges on a novel discrete optimization procedure to achieve nearly balanced
and uncorrelated hash bits, where the binary constraints are explicitly imposed and handled. To
tackle the discrete optimization in a computationally tractable manner, we propose an alternating
maximization algorithm which consists of solving two interesting subproblems. For brevity, we call
the proposed discrete optimization based graph hashing method as Discrete Graph Hashing (DGH).
Through extensive experiments carried out on four benchmark datasets with size up to one million,
we show that DGH consistently obtains higher search accuracy than state-of-the-art unsupervised
hashing methods, especially when relatively longer codes are learned.
2
Discrete Graph Hashing
First we de?ne a few main notations used throughout this paper: sgn(x) denotes the sign function
which returns 1 for x > 0 and ?1 otherwise; In denotes the n?n identity matrix; 1 denotes a vector
with all 1 elements; 0 denotes a vector or matrix of all 0 elements; diag(c) represents a diagonal
matrix with elements of vector c being its diagonal entries; tr(?), ? F , ? 1 , and ?, ? express
matrix trace norm, matrix Frobenius norm, 1 norm, and inner-product operator, respectively.
Anchor Graphs. In the discrete graph hashing model, we need to choose a neighborhood graph that
can easily scale to massive data points. For simplicity and ef?ciency, we choose Anchor Graphs [21],
which involve no special indexing scheme but still have linear construction time in the number of
data points. An anchor graph uses a small set of m points (called anchors), U = {uj ? Rd }m
j=1 , to
d n
approximate the neighborhood structure underlying the input dataset X = {xi ? R }i=1 . Af?nities
(or similarities) of all n data points are computed with respect to these m anchors in linear time
O(dmn) where m n. The true af?nity matrix Ao ? Rn?n is then approximated by using these
af?nities.
Speci?cally, an anchor graph leverages a nonlinear data-to-anchor mapping (Rd ? Rm ) z(x) =
2
2
1)
m)
?1 exp(? D (x,u
), ? ? ? , ?m exp(? D (x,u
) /M , where ?j ? {1, 0} and ?j = 1 if and only
t
t
if anchor uj is one of s m closest anchors of x in U according to some distance function
m
D 2 (x,uj )
)
D() (e.g., 2 distance), t > 0 is the bandwidth parameter, and M =
j=1 ?j exp(?
t
leading to z(x)1 = 1. Then, the anchor graph builds a data-to-anchor af?nity matrix Z =
2
z(x1 ), ? ? ? , z(xn ) ? Rn?m that is highly sparse. Finally, the anchor graph gives a data-to-data
af?nity matrix as A = Z??1 Z ? Rn?n where ? = diag(Z 1) ? Rm?m . Such an af?nity
matrix empirically approximates the true af?nity matrix Ao , and has two nice characteristics: 1)
A is a low-rank positive semide?nite (PSD) matrix with rank at most m, so the anchor graph does
not need to compute it explicitly but instead keeps its low-rank form and only saves Z and ? in
memory; 2) A has unit row and column sums, so the resulting graph Laplacian is L = In ? A. The
two characteristics permit convenient and ef?cient matrix manipulations upon A, as shown later on.
We also de?ne an anchor graph af?nity function as A(x, x ) = z (x)??1 z(x ) in which (x, x )
is any pair of points in Rd .
Learning Model. The purpose of unsupervised hashing is to learn to map each data point xi to an
r-bit binary hash code b(xi ) ? {1, ?1}r given a training dataset X = {xi }ni=1 . For simplicity, let
us denote b(xi ) as bi , and the corresponding code matrix as B = [b1 , ? ? ? , bn ] ? {1, ?1}n?r . The
standard graph-based hashing framework, proposed by [38], aims to learn the hash codes such that
the neighbors in the input space have small Hamming distances in the code space. This is formulated
as:
n
1
min
bi ? bj 2 Aoij = tr B Lo B , s.t. B ? {?1}n?r , 1 B = 0, B B = nIr , (1)
B
2 i,j=1
where Lo is the graph Laplacian based on the true af?nity matrix Ao1 . The constraint 1 B = 0
is imposed to maximize the information from each hash bit, which occurs when each bit leads to
a balanced partitioning of the dataset X . Another constraint B B = nIr makes r bits mutually
uncorrelated to minimize the redundancy among these bits. Problem (1) is NP-hard, and Weiss et al.
[38] therefore solved a relaxed problem by dropping the discrete (binary) constraint B ? {?1}n?r
and making a simplifying assumption of data being distributed uniformly.
We leverage the anchor graph to replace Lo by the anchor graph Laplacian L = In ? A. Hence, the
objective in Eq. (1) can be rewritten as a maximization problem:
max tr B AB , s.t. B ? {1, ?1}n?r , 1 B = 0, B B = nIr .
(2)
B
In [23], the solution to this problem is obtained via spectral relaxation [33] in which B is relaxed
to be a matrix of reals followed by a thresholding step (threshold is 0) that brings the ?nal discrete
B. Unfortunately, this procedure may result in poor codes due to ampli?cation of the error caused
by the relaxation as the code length r increases. To this end, we propose to directly solve the binary
codes B without resorting to such error-prone relaxations.
Let us de?ne a set ? = Y ? Rn?r |1 Y = 0, Y Y = nIr }. Then we formulate a more general
graph hashing framework which softens the last two hard constraints in Eq. (2) as:
?
(3)
max tr B AB ? dist2 (B, ?), s.t. B ? {1, ?1}n?r ,
B
2
where dist(B, ?) = minY?? B ? YF measures the distance from any matrix B to the set ?, and
? ? 0 is a tuning parameter. If problem (2) is feasible, we can enforce dist(B, ?) = 0 in Eq. (3) by
imposing a very large ?, thereby turning problem (3) into problem (2). However, in Eq. (3) we allow
a certaindiscrepancy between B and ? (controlled by ?), which makes problem (3) more ?exible.
Since tr B B) = tr Y Y) = nr, problem (3) can be equivalently transformed to the following
problem:
max Q(B, Y) := tr B AB + ?tr B Y ,
B,Y
(4)
s.t. B ? {1, ?1}n?r , Y ? Rn?r , 1 Y = 0, Y Y = nIr .
We call the code learning model formulated in Eq. (4) as Discrete Graph Hashing (DGH). Because
concurrently imposing B ? {?1}n?r and B ? ? will make graph hashing computationally intractable, DGH does not pursue the latter constraint but penalizes the distance from the target code
matrix B to ?. Different from the previous graph hashing methods which discard the discrete constraint B ? {?1}n?r to obtain continuously relaxed B, our DGH model enforces this constraint to
directly achieve discrete B. As a result, DGH yields nearly balanced and uncorrelated binary bits. In
Section 3, we will propose a computationally tractable optimization algorithm to solve this discrete
programming problem in Eq. (4).
1
The spectral hashing method in [38] did not compute the true af?nity matrix Ao because of the scalability
issue, but instead used a complete graph built over 1D PCA embeddings.
3
Algorithm 1 Signed Gradient Method (SGM) for B-Subproblem
Input: B(0) ? {1, ?1}n?r and Y ? ?.
j := 0; repeat B(j+1) := sgn C 2AB(j) + ?Y, B(j) , j := j + 1, until B(j) converges.
Output: B = B(j) .
Out-of-Sample Hashing. Since a hashing scheme should be able to generate the hash code for any
data point q ? Rd beyond the points in the training set X , here we address the out-of-sample extension of the DGH model. Similar to the objective in Eq. (1), we minimize the Hamming distances
between a novel data point q and its neighbors (revealed by the af?nity function A) in X as
n
1
b(q) ? b?i 2 A(q, xi ) = arg max
b(q) ? arg min r
b(q), (B? ) Z??1 z(q) ,
r
b(q)?{?1} 2
b(q)?{?1}
i=1
where B? = [b?1 , ? ? ? , b?n ] is the solution of problem (4). After pre-computing a matrix
W =
?1
?
r?m
?
(B ) Z? ? R
in the training phase, one can compute the hash code b (q) = sgn Wz(q)
for any novel data point q very ef?ciently.
3
Alternating Maximization
The graph hashing problem in Eq. (4) is essentially a nonlinear mixed-integer program involving
both discrete variables in B and continuous variables in Y. It turns out that problem (4) is generally
NP-hard and also dif?cult to approximate. In speci?c, since the Max-Cut problem is a special case
of problem (4) when ? = 0 and r = 1, there exists no polynomial-time algorithm which can achieve
the global optimum, or even an approximate solution with its objective value beyond 16/17 of the
global maximum unless P = NP [11]. To this end, we propose a tractable alternating maximization
algorithm to optimize problem (4), leading to good hash codes which are demonstrated to exhibit
superior search performance through extensive experiments conducted in Section 5.
The proposed algorithm proceeds by alternately solving the B-subproblem
max
f (B) := tr B AB + ?tr Y B
B?{?1}n?r
and the Y-subproblem
max
Y?Rn?r
tr B Y ,
s.t. 1 Y = 0, Y Y = nIr .
(5)
(6)
In what follows, we propose an iterative ascent procedure called Signed Gradient Method for subproblem (5) and derive a closed-form optimal solution to subproblem (6). As we can show, our
alternating algorithm is provably convergent. Schemes for choosing good initializations are also
discussed. Due to the space limit, all the proofs of lemmas, theorems and propositions presented in
this section are placed in the supplemental material.
3.1
B-Subproblem
We tackle subproblem (5) with a simple iterative ascent procedure described in Algorithm 1. In the
j-th iteration, we de?ne a local function f?j (B) that linearizes f (B) at the point B(j) , and employ
f?j (B) as a surrogate of f (B) for discrete optimization. Given B(j) , the next discrete point is
derived as B(j+1) ? arg maxB?{?1}n?r f?j (B) := f B(j) + ?f B(j) , B ? B(j) . Note that
(j)
(j+1)
since ?f B
may include zero entries, multiple
could exist. To avoid this
solutions for B
x, x = 0
ambiguity, we introduce the function C(x, y) =
to specify the following update:
y, x = 0
B(j+1) := sgn C ?f B(j) , B(j) = sgn C 2AB(j) + ?Y, B(j) ,
(7)
in which
C is applied in an element-wise manner, and no update is carried out to the entries where
?f B(j) vanishes.
Due to the PSD property of the matrix A, f is a convex function and thus f (B) ? f?j (B) for any B.
Taking advantage of the fact f B(j+1) ? f?j B(j+1) ? f?j B(j) ? f B(j) , Lemma 1 ensures
that both the sequence of cost values f (B(j) ) and the sequence of iterates B(j) converge.
4
Algorithm 2 Discrete Graph Hashing (DGH)
Input: B0 ? {1, ?1}n?r and Y0 ? ?.
k := 0;
repeat Bk+1 := SGM(Bk , Yk ), Yk+1 ? ?(JBk+1 ), k := k + 1, until Q(Bk , Yk ) converges.
Output: B? = Bk , Y? = Yk .
Lemma 1. If B(j) is the sequence of iterates produced by Algorithm 1, then f B(j+1) ?
(j)
f B
holds for any integer j ? 0, and both f (B(j) ) and B(j) converge.
Our idea of optimizing a proxy function f?j (B) can be considered as a special case of majorization
methodology exploited in the ?eld of optimization. The majorization method typically deals with a
generic constrained optimization problem: min g(x), s.t. x ? F, where g : Rn ? R is a continuous function and F ? Rn is a compact set. The majorization method starts with a feasible point
x0 ? F, and then proceeds by setting xj+1 as a minimizer of g?j (x) over F, where g?j satisfying
g?j (xj ) = g(xj ) and g?j (x) ? g(x) ?x ? F is called a majorization function of g at xj . In speci?c,
in our scenario, problem (5) is equivalent to minB?{?1}n?r ?f (B), and the linear surrogate ?f?j
is a majorization function of ?f at point B(j) . The majorization method was ?rst systematically
introduced by [5] to deal with multidimensional scaling problems, although the EM algorithm [7],
proposed at the same time, also falls into the framework of majorization methodology. Since then,
the majorization method has played an important role in various statistics problems such as multidimensional data analysis [12], hyperparameter learning [8], conditional random ?elds and latent
likelihoods [13], and so on.
Y-Subproblem
3.2
An analytical solution to subproblem (6) can be obtained with the aid of a centering matrix J = In ?
r
1
k=1 ?k uk vk ,
n 11 . Write the singular value decomposition (SVD) of JB as JB = U?V =
where r ? r is the rank of JB, ?1 , ? ? ? , ?r are the positive singular values, and U = [u1 , ? ? ? , ur ]
and V = [v1 , ? ? ? , vr ] contain the left- and right-singular vectors, respectively. Then, by employing
? ? Rn?(r?r ) and V
? ? Rr?(r?r )
a Gram-Schmidt process, one can easily construct matrices U
?
2
?
?
?
?
?
such that U U = Ir?r , [U 1] U = 0, and V V = Ir?r , V V = 0 . Now we are ready to
characterize a closed-form solution of the Y-subproblem by Lemma 2.
?
?
? is an optimal solution to the Y-subproblem in Eq. (6).
Lemma 2. Y = n[U U][V
V]
?
?
?
V]
For notational convenience, we de?ne the set of all matrices in the form of n[U U][V
as ?(JB). Lemma 2 reveals that any matrix in ?(JB) is an optimal solution to subproblem (6).
In practice, to compute such an optimal Y , we perform the
eigendecomposition over the small
2
?
0
?
? ?, and
? , which gives V, V,
r ? r matrix B JB to have B JB = [V V]
[V V]
0 0
? is initially set to a random matrix followed
immediately leads to U = JBV??1 . The matrix U
by the aforementioned Gram-Schmidt orthogonalization. It can be seen that Y is uniquely optimal
when r = r (i.e., JB is full column rank).
3.3
DGH Algorithm
The proposed alternating maximization algorithm, also referred to as Discrete Graph Hashing
(DGH), for solving the raw problem in Eq. (4) is summarized in Algorithm 2, in which we introduce
SGM(?, ?) to represent the functionality of Algorithm 1. The convergence of Algorithm 2 is guaranteed by Theorem 1, whose proof is based on the nature of the proposed alternating maximization
procedure that always generates a monotonically non-decreasing and bounded sequence.
Theorem 1. If (Bk , Yk ) is the sequence generated
by Algorithm
2, then Q(Bk+1 , Yk+1 ) ?
Q(Bk , Yk ) holds for any integer k ? 0, and Q(Bk , Yk ) converges starting with any feasible
initial point (B0 , Y0 ).
Initialization. Since the DGH algorithm deals with discrete and non-convex optimization, a good
choice of an initial point (B0 , Y0 ) is vital. Here we suggest two different initial points which are
both feasible to problem (4).
2
? and V
? are nothing but 0.
Note that when r = r, U
5
m
Let us perform the eigendecomposition over A to obtain A = P?P = k=1 ?k pk p
k , where
?1 , ? ? ? , ?m are the eigenvalues arranged in a non-increasing order, and p1 , ? ? ? , pm are the corresponding normalized eigenvectors.
We write ? = diag(?1 , ? ? ? , ?m )?
and P = [p1 , ? ? ? , pm
]. Note
?
that ?1 = 1 and p1 = 1/ n. The ?rst initialization used is Y0 = nH, B0 = sgn(H) , where
H = [p2 , ? ? ? , pr+1 ] ? Rn?r . The initial codes B0 were used as the ?nal codes by [23].
Alternatively,
? Y0 can be allowed to consist of orthonormal columns within the column space of H,
i.e., Y0 = nHR subject to some orthogonal matrix R ? Rr?r . We can obtain R along with B0
by solving a new discrete optimization problem:
(8)
max tr R H AB0 , s.t. R ? Rr?r , RR = Ir , B0 ? {1, ?1}n?r ,
R,B0
which is motivated by the proposition below.
Proposition 1. For any orthogonal matrix R ? Rr?r and any binary matrix B ? {1, ?1}n?r , we
1
have tr B AB ? tr2 R H AB .
r
Proposition 1 implies that the optimization in Eq. (8) can be interpreted as to maximize a lower
bound of tr B AB which is the ?rst term of the objective Q(B, Y) in the original problem
(4). We still exploit an alternating maximization procedure to solve problem (8).
AH =
Noticing
? where ?
? = diag(?2 , ? ? ? , ?r+1 ), the objective in Eq. (8) is equal to tr R ?H
? B0 ). The
H?
? j ,
alternating procedure starts with R0 = Ir , and then makes the simple updates Bj0 := sgn H?R
? jV
? j, V
? j ? Rr?r stem from the full SVD U
? j?
? for j = 0, 1, 2, ? ? ? , where U
? jV
? of
Rj+1 := U
j
j
j
? B . When convergence is reached, we obtain the optimized rotation R that yields
the matrix ?H
0
?
?
the second initialization Y0 = nHR, B0 = sgn(H?R)
.
Empirically, we ?nd that the second initialization typically gives a better objective value Q(B0 , Y0 )
at the start than the ?rst one, as it aims to maximize the lower bound of the ?rst term in the objective
Q. We also observe that the second initialization often results in a higher objective value Q(B? , Y? )
at convergence (Figs. 1-2 in the supplemental material show convergence curves of Q starting from
the two initial points). We call DGH using the ?rst and second initializations as DGH-I and DGH-R,
respectively. Regarding the convergence property, we would like to point out that since the DGH algorithm (Algorithm 2) works on a mixed-integer objective, it is hard to quantify the convergence to a
local optimum of the objective function Q. Nevertheless, this does not affect the performance of our
algorithm in practice. In our experiments in Section 5, we consistently ?nd a convergent sequence
{(Bk , Yk )} arriving at a good objective value when started with the suggested initializations.
4
Discussions
Here we analyze space and time complexities of DGH-I/DGH-R. The space complexity is O (d +
s + r)n in the training stage and O(rn) for storing hash codes in the test stage for DGH-I/DGH-R.
and the whole DGH
Let TB and TG be the budget iteration numbers of optimizing the B-subproblem
problem, respectively.
Then,
the
training
time
complexity
of
DGH-I
is
O
dmn
+ m2 n + (mTB +
2
sTB + r)rTG n , and the training time complexity of DGH-R is O dmn + m n + (mTB + sTB +
r)rTG n + r2 TR n , where TR is the budget iteration number for seeking the initial point via Eq. (8).
Note that the time for ?nding anchors and building the anchor graph is O(dmn) which is included
in the above training time. Their test time (referring to encoding a query to an r-bit code) is both
O(dm + sr). In our experiments, we ?x m, s, TB , TG , TR to constants independent of the dataset
size n, and make r ? 128. Thus, DGH-I/DGH-R enjoy linear training time and constant test time.
It is worth mentioning again that the low-rank PSD property of the anchor graph af?nity matrix A is
advantageous for training DGH, permitting ef?cient matrix computations in O(n) time, such as the
eigendecomposition of A (encountered in initializations) and multiplying A with B (encountered
in solving the B-subproblem with Algorithm 1).
It is interesting to point out that DGH falls into the asymmetric hashing category [26] in the sense
that hash codes are generated differently for samples within the dataset and queries outside the
dataset. Unlike most existing hashing techniques, DGH directly solves the hash codes B? of the
training samples via the proposed discrete optimization in Eq. (4) without relying on any explicit or
prede?ned hash functions. On the other hand, the hash code
for any
query q is induced from the
solved codes B? , leading to a hash function b? (q) = sgn Wz(q) parameterized by the matrix
6
(b) Hash lookup success rate @ SUN397
Success rate
(a) Hash lookup success rate @ CIFAR?10
(c) Hash lookup success rate @ YouTube Faces
1
1
1
0.9
0.9
0.9
0.9
0.8
0.8
0.8
0.8
0.7
0.7
0.6
0.6
0.5
0.4
0.3
0.2
0.1
0
LSH
KLSH
ITQ
IsoH
SH
IMH
1?AGH
2?AGH
BRE
DGH?I
DGH?R
8 1216
24
LSH
KLSH
ITQ
IsoH
SH
IMH
1?AGH
2?AGH
BRE
DGH?I
DGH?R
0.4
0.3
0.2
0.1
48
# bits
64
96
0
8 1216
24
0.6
0.5
0.4
0.3
0.2
32
48
# bits
64
96
0.1
(d) Hash lookup success rate @ Tiny?1M
LSH
KLSH
ITQ
IsoH
SH
IMH
1?AGH
2?AGH
BRE
DGH?I
DGH?R
0.7
0.7
0.5
32
1
0.6
LSH
KLSH
ITQ
IsoH
SH
IMH
1?AGH
2?AGH
BRE
DGH?I
DGH?R
1216 24 32
0.5
0.4
0.3
0.2
0.1
48
64
# bits
96
128
0
1216 24 32
48
64
# bits
96
128
Figure 1: Hash lookup success rates for different hashing techniques. DGH tends to achieve nearly
100% success rates even for longer code lengths.
F?measure within Hamming radius 2
0.3
(a) Hash lookup F?measure @ CIFAR?10
0.2
LSH
KLSH
ITQ
IsoH
SH
IMH
1?AGH
2?AGH
BRE
DGH?I
DGH?R
0.25
0.2
0.15
(c) Hash lookup F?measure @ YouTube Faces
(b) Hash lookup F?measure @ SUN397
0.8
0.18
0.16
(d) Hash lookup F?measure @ Tiny?1M
0.2
0.6
0.14
0.5
0.12
0.1
0.15
0.4
0.08
0.1
0.1
0.3
0.06
0.2
0.04
0.05
8 1216
24
32
48
# bits
64
96
0
0.05
0.1
0.02
0
0.25
0.7
8 1216
24
32
48
# bits
64
96
0
1216 24 32
48
64
# bits
96
128
0
1216 24 32
48
64
# bits
96
128
Figure 2: Mean F-measures of hash lookup within Hamming radius 2 for different techniques. DGH
tends to retain good recall even for longer codes, leading to much higher F-measures than the others.
W which was computed using B? . While the hashing mechanisms for producing B? and b? (q) are
distinct, they are tightly coupled and prone to be adaptive to speci?c datasets. The ?exibility of the
asymmetric hashing nature of DGH is validated through the experiments shown in the next section.
5
Experiments
We conduct large-scale similarity search experiments on four benchmark datasets: CIFAR-10 [15],
SUN397 [40], YouTube Faces [39], and Tiny-1M. CIFAR-10 is a labeled subset of the 80 Million
Tiny Images dataset [35], which consists of 60K images from ten object categories with each image
represented by a 512-dimensional GIST feature vector [29]. SUN397 contains about 108K images
from 397 scene categories, where each image is represented by a 1,600-dimensional feature vector
extracted by PCA from 12,288-dimensional Deep Convolutional Activation Features [10]. The raw
YouTube Faces dataset contains 1,595 different people, from which we choose 340 people such that
each one has at least 500 images to form a subset of 370,319 face images, and represent each face
image as a 1,770-dimensional LBP feature vector [1]. Tiny-1M is one million subset of the 80M
tiny images, where each image is represented by a 384-dimensional GIST vector. In CIFAR-10, 100
images are sampled uniformly randomly from each object category to form a separate test (query)
set of 1K images; in SUN397, 100 images are sampled uniformly randomly from each of the 18
largest scene categories to form a test set of 1.8K images; in YouTube Faces, the test set includes
3.8K face images which are evenly sampled from the 38 people each containing more than 2K faces;
in Tiny-1M, a separate subset of 5K images randomly sampled from the 80M images is used as the
test set. In the ?rst three datasets, groundtruth neighbors are de?ned based on whether two samples
share the same class label; in Tiny-1M which does not have full annotations, we de?ne groundtruth
neighbors for a given query as the samples among the top 2% 2 distances from the query in the 1M
training set, so each query has 20K groundtruth neighbors.
We evaluate twelve unsupervised hashing methods including: two randomized methods LSH [2] and
Kernelized LSH (KLSH) [17], two linear projection based methods Iterative Quantization (ITQ) [9]
and Isotropic Hashing (IsoH) [14], two spectral methods Spectral Hashing (SH) [38] and its weighted version MDSH [37], one manifold based method Inductive Manifold Hashing (IMH) [32], two
existing graph-based methods One-Layer Anchor Graph Hashing (1-AGH) and Two-Layer Anchor
Graph Hashing (2-AGH) [23], one distance preservation method Binary Reconstruction Embeddings (BRE) [16] (unsupervised version), and our proposed discrete optimization based methods
DGH-I and DGH-R. We use the publicly available codes of the competing methods, and follow the
conventional parameter settings therein. In particular, we use the Gaussian kernel and 300 randomly
sampled exemplars (anchors) to run KLSH; IMH, 1-AGH, 2-AGH, DGH-I and DGH-R also use
m = 300 anchors (obtained by K-means clustering with 5 iterations) for fair comparison. This
choice of m gives a good trade-off between hashing speed and performance. For 1-AGH, 2-AGH,
DGH-I and DGH-R that all use anchor graphs, we adopt the same construction parameters s, t on
each dataset (s = 3 and t is tuned following AGH), and 2 distance as D(?). For BRE, we uniformly
7
Table 1: Hamming ranking performance on YouTube Faces and Tiny-1M. r denotes the number of
hash bits used in the hashing methods. All training and test times are in seconds.
Method
2 Scan
YouTube Faces
Mean Precision / Top-2K
TrainTime
TestTime
r = 48
r = 96
r = 128
r = 128
r = 128
0.7591
?
LSH
KLSH
ITQ
IsoH
SH
MDSH
IMH
1-AGH
2-AGH
BRE
DGH-I
DGH-R
0.0830
0.3982
0.7017
0.6093
0.5897
0.6110
0.3150
0.7138
0.6727
0.5564
0.7086
0.7245
0.1005
0.5210
0.7493
0.6962
0.6655
0.6752
0.3641
0.7571
0.7377
0.6238
0.7644
0.7672
0.1061
0.5871
0.7562
0.7058
0.6736
0.6795
0.3889
0.7646
0.7521
0.6483
0.7750
0.7805
6.4
16.1
169.0
73.6
108.9
118.8
92.1
84.1
94.7
10372.0
402.6
408.9
1.8?10?5
4.8?10?5
1.8?10?5
1.8?10?5
2.0?10?4
4.9?10?5
2.3?10?5
2.1?10?5
3.5?10?5
9.0?10?5
2.1?10?5
2.1?10?5
Tiny-1M
Mean Precision / Top-20K
TrainTime
TestTime
r = 48
r = 96
r = 128
r = 128
r = 128
1
?
0.1155
0.3054
0.3925
0.3896
0.1857
0.3312
0.2257
0.4061
0.3925
0.3943
0.4045
0.4208
0.1324
0.4105
0.4726
0.4816
0.1923
0.3878
0.2497
0.4117
0.4099
0.4836
0.4865
0.5006
0.1766
0.4705
0.5052
0.5161
0.2079
0.3955
0.2557
0.4107
0.4152
0.5218
0.5178
0.5358
6.1
20.7
297.3
13.5
61.4
193.6
139.3
141.4
272.5
8419.0
1769.4
2793.4
1.0?10?5
4.6?10?5
1.0?10?5
1.0?10?5
1.6?10?4
2.8?10?5
2.7?10?5
3.4?10?5
4.7?10?5
8.8?10?5
3.3?10?5
3.3?10?5
randomly sample 1K, and 2K training samples to train the distance preservations on CIFAR-10 &
SUN397, and YouTube Faces & Tiny-1M, respectively. For DGH-I and DGH-R, we set the penalty
parameter ? to the same value in [0.1, 5] on each dataset, and ?x TR = 100, TB = 300, TG = 20.
We employ two widely used search procedures hash lookup and Hamming ranking with 8 to 128
hash bits for evaluations. The Hamming ranking procedure ranks the dataset samples according to
their Hamming distances to a given query, while the hash lookup procedure ?nds all the points within
a certain Hamming radius away from the query. Since hash lookup can be achieved in constant time
by using a single hash table, it is the main focus of this work. We carry out hash lookup within a
Hamming ball of radius 2 centered on each query, and report the search recall and F-measure which
are averaged over all queries for each dataset. Note that if table lookup fails to ?nd any neighbors
within a given radius for a query, we call it a failed query and assign it zero recall and F-measure. To
quantify the failed queries, we report the hash lookup success rate which gives the proportion of the
queries for which at least one neighbor is retrieved. For Hamming ranking, mean average precision
(MAP) and mean precision of top-retrieved samples are computed.
The hash lookup results are shown in Figs. 1-2. DGH-I/DGH-R achieve the highest (close to 100%)
hash lookup success rates, and DGH-I is slightly better than DGH-R. The reason is that the asymmetric hashing scheme exploited by DGH-I/DGH-R poses a tight linkage to connect queries and
database samples, providing a more adaptive out-of-sample extension than the traditional symmetric hashing schemes used by the competing methods. Also, DGH-R achieves the highest F-measure
except on CIFAR-10, where DGH-I is highest while DGH-R is the second. The F-measures of
KLSH, IsoH, SH and BRE deteriorate quickly and are with very poor values (< 0.05) when r ? 48
due to poor recall3 . Although IMH achieves nice hash lookup succuss rates, its F-measures are
much lower than DGH-I/DGH-R due to lower precision. MDSH produces the same hash bits as SH,
so is not included in the hash lookup experiments. DGH-I/DGH-R employ the proposed discrete
optimization to yield high-quality codes that preserve the local neighborhood of each data point
within a small Hamming ball, so obtain much higher search accuracy in F-measure and recall than
SH, 1-AGH and 2-AGH which rely on relaxed optimizations and degrade drastically when r ? 48.
Finally, we report the Hamming ranking results in Table 1 and the table in the sup-material, which
clearly show the superiority of DGH-R over the competing methods in MAP and mean precision;
on the ?rst three datasets, DGH-R even outperforms exhaustive 2 scan. The training time of DGHI/DGH-R is acceptable and faster than BRE, and their test time (i.e., coding time since hash lookup
time is small enough to be ignored) is comparable with 1-AGH.
6 Conclusion
This paper investigated a pervasive problem of not enforcing the discrete constraints in optimization pertaining to most existing hashing methods. Instead of resorting to error-prone continuous
relaxations, we introduced a novel discrete optimization technique that learns the binary hash codes
directly. To achieve this, we proposed a tractable alternating maximization algorithm which solves
two interesting subproblems and provably converges. When working with a neighborhood graph,
the proposed method yields high-quality codes to well preserve the neighborhood structure inherent
in the data. Extensive experimental results on four large datasets up to one million showed that our
discrete optimization based graph hashing technique is highly competitive.
3
The recall results are shown in Fig. 3 of the supplemental material, which indicate that DGH-I achieves the
highest recall except on YouTube Faces, where DGH-R is highest while DGH-I is the second.
8
References
[1] T. Ahonen, A. Hadid, and M. Pietikainen. Face description with local binary patterns: Application to face recognition. TPAMI,
28(12):2037?2041, 2006.
[2] A. Andoni and P. Indyk. Near-optimal hashing algorithms for approximate nearest neighbor in high dimensions. Communications of the
ACM, 51(1):117?122, 2008.
[3] A. Z. Broder, M. Charikar, A. M. Frieze, and M. Mitzenmacher. Min-wise independent permutations. In Proc. STOC, 1998.
[4] M. Charikar. Similarity estimation techniques from rounding algorithms. In Proc. STOC, 2002.
[5] J. de Leeuw. Applications of convex analysis to multidimensinal scaling. Recent Developments in Statistics, pages 133?146, 1977.
[6] T. Dean, M. A. Ruzon, M. Segal, J. Shlens, S. Vijayanarasimhan, and J. Yagnik. Fast, accurate detection of 100,000 object classes on a
single machine. In Proc. CVPR, 2013.
[7] A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum likelihood from incomplete data via the em algorithm. Journal of the Royal
Statistical Society, Series B, 39(1):1?38, 1977.
[8] C.-S. Foo, C. B. Do, and A. Y. Ng. A majorization-minimization algorithm for (multiple) hyperparameter learning. In Proc. ICML, 2009.
[9] Y. Gong, S. Lazebnik, A. Gordo, and F. Perronnin. Iterative quantization: A procrustean approach to learning binary codes for large-scale
image retrieval. TPAMI, 35(12):2916?2929, 2013.
[10] Y. Gong, L. Wang, R. Guo, and S. Lazebnik. Multi-scale orderless pooling of deep convolutional activation features. In Proc. ECCV,
2014.
[11] J. Hastad. Some optimal inapproximability results. Journal of the ACM, 48(4):798?859, 2001.
[12] W. J. Heiser. Convergent computation by iterative majorization: theory and applications in multidimensional data analysis. Recent
advances in descriptive multivariate analysis, pages 157?189, 1995.
[13] T. Jebara and A. Choromanska. Majorization for crfs and latent likelihoods. In NIPS 25, 2012.
[14] W. Kong and W.-J. Li. Isotropic hashing. In NIPS 25, 2012.
[15] A. Krizhevsky. Learning multiple layers of features from tiny images. Technical report, 2009.
[16] B. Kulis and T. Darrell. Learning to hash with binary reconstructive embeddings. In NIPS 22, 2009.
[17] B. Kulis and K. Grauman. Kernelized locality-sensitive hashing. TPAMI, 34(6):1092?1104, 2012.
[18] P. Li and A. C. Konig. Theory and applications of b-bit minwise hashing. Communications of the ACM, 54(8):101?109, 2011.
[19] P. Li, A. Shrivastava, J. Moore, and A. C. Konig. Hashing algorithms for large-scale learning. In NIPS 24, 2011.
[20] X. Li, G. Lin, C. Shen, A. van den Hengel, and A. R. Dick. Learning hash functions using column generation. In Proc. ICML, 2013.
[21] W. Liu, J. He, and S.-F. Chang. Large graph construction for scalable semi-supervised learning. In Proc. ICML, 2010.
[22] W. Liu, J. Wang, R. Ji, Y.-G. Jiang, and S.-F. Chang. Supervised hashing with kernels. In Proc. CVPR, 2012.
[23] W. Liu, J. Wang, S. Kumar, and S.-F. Chang. Hashing with graphs. In Proc. ICML, 2011.
[24] W. Liu, J. Wang, Y. Mu, S. Kumar, and S.-F. Chang. Compact hyperplane hashing with bilinear functions. In Proc. ICML, 2012.
[25] Y. Mu, J. Shen, and S. Yan. Weakly-supervised hashing in kernel space. In Proc. CVPR, 2010.
[26] B. Neyshabur, P. Yadollahpour, Y. Makarychev, R. Salakhutdinov, and N. Srebro. The power of asymmetry in binary hashing. In NIPS
26, 2013.
[27] M. Norouzi and D. J. Fleet. Minimal loss hashing for compact binary codes. In Proc. ICML, 2011.
[28] M. Norouzi, D. J. Fleet, and R. Salakhudinov. Hamming distance metric learning. In NIPS 25, 2012.
[29] A. Oliva and A. Torralba. Modeling the shape of the scene: a holistic representation of the spatial envelope. IJCV, 42(3):145?175, 2001.
[30] R. Salakhutdinov and G. Hinton. Semantic hashing. International Journal of Approximate Reasoning, 50(7):969?978, 2009.
[31] G. Shakhnarovich, P. Viola, and T. Darrell. Fast pose estimation with parameter-sensitive hashing. In Proc. ICCV, 2003.
[32] F. Shen, C. Shen, Q. Shi, A. van den Hengel, and Z. Tang. Inductive hashing on manifolds. In Proc. CVPR, 2013.
[33] J. Shi and J. Malik. Normalized cuts and image segmentation. TPAMI, 22(8):888?905, 2000.
[34] Q. Shi, J. Petterson, G. Dror, J. Langford, A. Smola, and S. V. N. Vishwanathan. Hash kernels for structured data. JMLR, 10:2615?2637,
2009.
[35] A. Torralba, R. Fergus, and W. T. Freeman. 80 million tiny images: a large dataset for non-parametric object and scene recognition.
TPAMI, 30(11):1958?1970, 2008.
[36] K. Q. Weinberger, A. Dasgupta, J. Langford, A. J. Smola, and J. Attenberg. Feature hashing for large scale multitask learning. In Proc.
ICML, 2009.
[37] Y. Weiss, R. Fergus, and A. Torralba. Multidimensional spectral hashing. In Proc. ECCV, 2012.
[38] Y. Weiss, A. Torralba, and R. Fergus. Spectral hashing. In NIPS 21, 2008.
[39] L. Wolf, T. Hassner, and I. Maoz. Face recognition in unconstrained videos with matched background similarity. In Proc. CVPR, 2011.
[40] J. Xiao, J. Hays, K. A. Ehinger, A. Oliva, and A. Torralba. Sun database: Large-scale scene recognition from abbey to zoo. In Proc.
CVPR, 2010.
9
| 5332 |@word multitask:1 kong:1 kulis:2 version:2 polynomial:1 norm:3 advantageous:1 nd:5 proportion:1 heiser:1 crucially:1 bn:1 simplifying:1 decomposition:1 eld:2 thereby:2 tr:19 accommodate:1 carry:1 initial:6 liu:5 contains:2 series:1 tuned:1 leeuw:1 past:1 existing:4 outperforms:1 com:2 activation:2 tackling:1 sanjiv:1 cant:1 shape:1 gist:2 progressively:1 update:3 hash:52 cult:1 isotropic:4 short:2 iterates:2 along:1 become:1 consists:2 ijcv:1 softens:1 introduce:2 manner:2 deteriorate:1 x0:1 p1:3 dist:2 multi:2 salakhutdinov:2 relying:1 decreasing:1 freeman:1 increasing:1 notation:1 underlying:1 bounded:1 matched:1 what:1 interpreted:1 pursue:1 dror:1 supplemental:3 cial:1 multidimensional:4 tackle:2 grauman:1 rm:2 uk:1 partitioning:1 unit:1 enjoy:1 gigantic:1 producing:1 superiority:1 positive:2 local:6 aggregating:1 tends:2 limit:1 bilinear:1 encoding:2 jiang:1 signed:2 initialization:9 therein:1 dif:1 mentioning:1 limited:1 bi:2 averaged:1 enforces:1 practice:3 procedure:12 nite:1 dmn:4 yan:1 projection:2 convenient:1 pre:1 suggest:1 convenience:1 close:1 operator:1 storage:2 vijayanarasimhan:1 optimize:1 equivalent:1 imposed:3 sanjivk:1 shi:3 map:4 center:1 demonstrated:1 conventional:1 attention:1 starting:2 convex:3 agh:22 formulate:2 shen:4 simplicity:3 immediately:1 m2:1 orthonormal:1 shlens:1 target:2 construction:3 massive:4 programming:1 us:1 trend:1 element:4 recognition:5 particularly:2 expensive:1 approximated:1 satisfying:1 asymmetric:3 cut:2 database:3 labeled:1 role:1 subproblem:14 solved:2 capture:3 wang:4 thousand:1 ensures:1 sun:1 trade:1 highest:5 yk:9 balanced:3 sgm:3 vanishes:1 mu:3 complexity:4 miny:1 dempster:1 weakly:2 solving:5 tight:1 shakhnarovich:1 upon:1 easily:4 differently:1 various:1 represented:3 train:2 distinct:1 fast:3 effective:1 reconstructive:1 pertaining:1 query:16 neighborhood:9 choosing:1 outside:1 exhaustive:1 whose:1 emerged:1 widely:1 solve:4 cvpr:6 say:2 otherwise:1 statistic:2 itself:1 laird:1 indyk:1 semide:1 advantage:1 sequence:6 rr:6 analytical:1 eigenvalue:1 tpami:5 reconstruction:3 propose:5 descriptive:1 product:1 rapidly:2 holistic:1 till:1 nity:10 achieve:10 maoz:1 description:1 frobenius:1 scalability:1 dist2:1 rst:8 convergence:6 konig:2 optimum:2 darrell:2 asymmetry:1 produce:1 converges:4 object:6 derive:1 develop:1 gong:2 pose:2 exemplar:1 nearest:4 b0:11 received:1 eq:14 solves:2 p2:1 signi:1 itq:10 implies:1 quantify:2 indicate:1 radius:5 functionality:1 centered:1 sgn:9 prede:2 material:4 hassner:1 assign:1 ao:3 proposition:4 nities:2 extension:2 hold:2 around:1 considered:1 exp:3 mapping:1 bj:1 makarychev:1 gordo:1 klsh:9 early:1 adopt:1 achieves:3 salakhudinov:1 purpose:1 abbey:1 estimation:2 proc:18 torralba:5 label:1 sensitive:4 largest:1 create:1 tool:1 weighted:1 minimization:1 concurrently:1 clearly:1 always:1 gaussian:1 aim:2 avoid:1 pervasive:1 derived:1 focus:3 validated:1 notational:1 vk:1 consistently:2 rank:7 likelihood:3 sense:1 dependent:1 perronnin:1 typically:3 entire:1 initially:1 kernelized:2 transformed:1 choromanska:1 provably:2 overall:1 among:3 issue:1 arg:3 aforementioned:1 development:1 art:2 special:3 constrained:1 spatial:1 equal:1 construct:2 saving:1 ng:1 bj0:1 represents:1 unsupervised:10 nearly:3 icml:7 mimic:1 discrepancy:1 report:4 jb:8 np:3 inherent:2 few:3 employ:3 others:1 randomly:5 frieze:1 preserve:3 tightly:1 petterson:1 phase:1 psd:3 ab:9 detection:2 highly:3 evaluation:1 sh:10 yielding:1 accurate:1 fu:1 orthogonal:2 unless:1 conduct:1 incomplete:1 penalizes:1 minimal:2 column:6 modeling:1 hastad:1 maximization:9 tg:3 cost:2 entry:3 subset:4 hundred:2 krizhevsky:1 rounding:2 inadequate:1 conducted:1 characterize:1 connect:1 accomplish:1 referring:1 twelve:1 randomized:4 broder:1 international:1 retain:1 off:1 tr2:1 continuously:1 quickly:1 again:1 ambiguity:1 containing:1 choose:3 leading:4 return:1 li:4 segal:1 de:8 lookup:21 summarized:1 coding:1 includes:1 explicitly:4 caused:1 ranking:5 performed:1 later:1 closed:2 analyze:1 sup:1 reached:1 start:3 competitive:1 annotation:1 majorization:11 minimize:2 ni:1 degraded:1 accuracy:4 ir:4 characteristic:2 convolutional:2 publicly:1 yield:4 dean:1 raw:3 norouzi:2 produced:1 zoo:1 multiplying:1 worth:1 cation:1 ah:1 mtb:2 centering:1 dm:1 testtime:2 proof:2 hamming:17 gain:1 sampled:6 dataset:17 ahonen:1 popular:3 recall:8 segmentation:1 bre:13 hashing:87 higher:5 supervised:7 follow:1 methodology:2 specify:1 wei:4 improved:1 arranged:1 mitzenmacher:1 stage:2 smola:2 until:2 langford:2 hand:1 working:1 nonlinear:2 google:2 brings:1 quality:3 building:1 concept:1 true:4 contain:1 normalized:2 adequately:1 hence:1 inductive:2 alternating:10 symmetric:1 iteratively:1 satisfactory:1 moore:1 semantic:2 deal:4 adjacent:1 round:1 during:2 uniquely:1 inferior:1 cosine:1 procrustean:1 complete:1 orthogonalization:1 reasoning:1 image:22 wise:3 lazebnik:2 ef:7 novel:4 superior:2 rotation:1 empirically:2 ji:1 nh:1 million:7 discussed:1 he:1 approximates:1 imposing:3 tuning:2 rd:4 resorting:2 pm:2 unconstrained:1 lsh:10 longer:5 similarity:9 etc:1 closest:1 multivariate:1 recent:3 showed:1 retrieved:3 optimizing:2 discard:2 scenario:3 manipulation:1 certain:2 hay:1 binary:24 watson:1 success:9 yagnik:1 exploited:3 seen:1 relaxed:5 speci:4 r0:1 converge:2 maximize:3 monotonically:1 preservation:2 semi:2 afterwards:1 multiple:4 full:3 reduces:1 stem:1 rj:1 technical:1 faster:1 af:12 retrieval:2 cifar:7 lin:1 permitting:1 laplacian:3 controlled:1 involving:1 scalable:1 oliva:2 vision:1 metric:3 essentially:1 iteration:4 kernel:7 represent:2 achieved:1 lbp:1 background:1 singular:3 envelope:1 unlike:1 minb:1 sr:1 ascent:2 subject:1 induced:1 pooling:1 stb:2 linearizes:1 call:4 ciently:1 near:1 ee:1 leverage:3 integer:4 revealed:1 embeddings:5 minhash:3 maxb:1 variety:2 xj:4 vital:1 affect:1 enough:1 bandwidth:1 competing:3 inner:1 idea:1 regarding:1 fleet:2 whether:1 motivated:1 pca:4 handled:1 gb:1 accelerating:1 linkage:1 penalty:1 suffer:1 deep:2 ignored:1 generally:1 involve:1 eigenvectors:1 ten:1 concentrated:1 category:5 reduced:1 generate:2 exist:1 sign:1 deteriorates:2 per:2 discrete:35 hyperparameter:2 dropping:1 write:2 dasgupta:1 express:1 redundancy:1 shih:1 four:4 threshold:1 nevertheless:1 jv:2 yadollahpour:1 nal:2 ampli:1 v1:1 graph:44 relaxation:5 year:1 sum:1 run:2 noticing:1 parameterized:1 almost:1 throughout:1 groundtruth:3 acceptable:1 jaccard:1 scaling:2 comparable:1 bit:26 bound:2 layer:3 followed:2 played:1 convergent:3 guaranteed:1 encountered:2 constraint:14 vishwanathan:1 scene:5 generates:1 u1:1 speed:1 min:5 kumar:3 relatively:2 ned:3 charikar:2 structured:1 according:2 ball:2 poor:3 slightly:1 em:2 y0:8 ur:1 appealing:1 cun:1 making:1 den:2 iccv:1 indexing:1 pr:1 computationally:3 mutually:1 turn:1 mechanism:1 tractable:5 end:2 available:1 rewritten:1 permit:1 neyshabur:1 observe:1 away:1 spectral:8 enforce:1 generic:1 ruzon:1 attenberg:1 save:1 schmidt:2 rtg:2 weinberger:1 original:1 denotes:5 top:4 include:3 clustering:1 hinge:1 neglect:1 cally:1 exploit:1 especially:2 uj:3 build:1 society:1 seeking:1 objective:11 malik:1 occurs:1 parametric:1 diagonal:2 nr:1 surrogate:2 exhibit:1 gradient:2 traditional:1 distance:18 separate:2 evenly:1 degrade:1 manifold:3 argue:2 reason:2 enforcing:1 length:4 code:54 providing:1 minimizing:1 dick:1 equivalently:1 unfortunately:1 stoc:2 subproblems:2 trace:1 perform:2 datasets:7 benchmark:2 viola:1 hinton:1 communication:2 rn:11 jebara:1 bk:9 introduced:2 cast:1 required:1 bene:1 extensive:4 pair:1 optimized:1 learned:1 alternately:1 nip:7 address:1 beyond:3 able:1 proceeds:2 below:1 suggested:1 pattern:1 program:1 tb:3 built:1 including:2 memory:2 video:1 sfchang:1 max:8 wz:2 royal:1 power:1 treated:1 rely:1 turning:1 scheme:8 ne:6 nding:1 started:1 carried:2 ready:1 coupled:1 columbia:3 nir:6 nice:2 literature:1 loss:2 permutation:2 mixed:2 generation:2 interesting:3 srebro:1 eigendecomposition:3 proxy:1 rubin:1 thresholding:1 xiao:1 systematically:1 tiny:13 storing:2 uncorrelated:3 ibm:2 eccv:2 row:1 lo:3 prone:3 share:1 repeat:2 last:1 placed:1 arriving:1 drastically:1 allow:1 neighbor:11 fall:2 taking:1 face:16 sparse:1 orderless:1 distributed:1 van:2 curve:1 dimension:1 xn:1 gram:2 hengel:2 adaptive:2 employing:1 approximate:5 obtains:2 compact:9 keep:1 global:2 active:1 reveals:1 anchor:27 b1:1 xi:6 fergus:3 alternatively:1 search:12 iterative:7 continuous:4 latent:2 table:8 additionally:1 learn:5 nature:2 shrivastava:1 investigated:1 diag:4 did:1 pk:1 main:3 whole:3 nothing:1 allowed:1 fair:1 x1:1 fig:4 representative:1 cient:3 referred:1 crfs:1 ehinger:1 aid:1 vr:1 precision:8 fails:1 foo:1 explicit:1 ciency:3 jmlr:1 learns:3 tang:1 theorem:3 exible:1 er:1 r2:1 intractable:1 exists:1 quantization:5 consist:1 andoni:1 budget:2 locality:2 failed:2 inapproximability:1 chang:5 wolf:1 minimizer:1 extracted:1 acm:3 conditional:1 identity:1 endeavor:1 formulated:2 dgh:69 sun397:6 replace:1 considerable:1 hard:4 feasible:4 included:2 youtube:9 except:2 uniformly:4 hyperplane:1 classi:1 lemma:6 called:4 pietikainen:1 svd:2 experimental:1 people:3 guo:1 latter:1 scan:2 brevity:1 minwise:1 incorporate:1 evaluate:1 exibility:1 |
4,786 | 5,333 | Large-scale L-BFGS using MapReduce
Weizhu Chen, Zhenghao Wang, Jingren Zhou
Microsoft
{wzchen,zhwang,jrzhou}@microsoft.com
Abstract
L-BFGS has been applied as an effective parameter estimation method for various
machine learning algorithms since 1980s. With an increasing demand to deal
with massive instances and variables, it is important to scale up and parallelize
L-BFGS effectively in a distributed system. In this paper, we study the problem
of parallelizing the L-BFGS algorithm in large clusters of tens of thousands of
shared-nothing commodity machines. First, we show that a naive implementation
of L-BFGS using Map-Reduce requires either a significant amount of memory or a
large number of map-reduce steps with negative performance impact. Second, we
propose a new L-BFGS algorithm, called Vector-free L-BFGS, which avoids the
expensive dot product operations in the two loop recursion and greatly improves
computation efficiency with a great degree of parallelism. The algorithm scales
very well and enables a variety of machine learning algorithms to handle a massive
number of variables over large datasets. We prove the mathematical equivalence
of the new Vector-free L-BFGS and demonstrate its excellent performance and
scalability using real-world machine learning problems with billions of variables
in production clusters.
1
Introduction
In the big data era, many applications require solving optimization problems with billions of variables on a huge amount of training data. Problems of this scale are more common nowadays, such
as Ads CTR prediction[1] and deep neural network[2]. The other trend is the wide adoption of mapreduce [3] environments built with commodity hardware. Those large-scale optimization problems
are often expected to be solved in a map-reduce environment where big data are stored.
When a problem is with huge number of variables, it can be solved efficiently only if the storage and
computation cost are maintained effectively. Among a diverse collection of large-scale optimization
methods, Limited-memory BFGS (L-BFGS)[4] is one of the frequently used optimization methods
in practice[5]. In this paper, we study the L-BFGS implementation for billion-variable scale problems in a map-reduce environment. The original L-BFGS algorithm and its update procedure were
proposed in 1980s. A lot of popular optimization software packages implement it as a fundamental building block. Approaches to apply it in a problem with up to millions of variables are well
studied and implemented in various optimization packages [6]. However, studies about how to scale
L-BFGS into billions of variables are still in their very early stages. For such a massive scale, the
parameters, their gradients, and the associated L-BFGS historical states are not only too large to
be stored in the memory of a single computation node, but also create too huge computation complexity for a processor or multicores to conquer it within reasonable time. Therefore, it is critical
to explore an effective decomposition over both examples and models via distributed learning. Yet,
to our knowledge, there is still very limited work to explore billion-variable scale L-BFGS. This
directly leads to the consequence that very little work can scale various machine learning algorithms
up to billion-variable scale using L-BFGS on map-reduce.
1
In this paper, we start by carefully studying the implementation of L-BFGS in map-reduce environment. We examine two typical L-BFGS implementations in map-reduce and present their scaling
obstacles. Particularly, given a problem with d variables and m historical states to approximate Hessian [5], traditional implementation[6][5], either need to store 2md variables in memory or need to
perform 2m map-reduce steps per iteration. This clearly creates huge overhead for the problem with
billions of variables and prevents a scalable implementation in map-reduce.
To conquer these limitations, we reexamine the original L-BFGS algorithm and propose a new LBFGS update procedure, called Vector-free L-BFGS (VL-BFGS), which is specifically devised for
distributed learning with huge number of variables. In particular, we replace the original L-BFGS
update procedure depending on vector operations, as known as two-loop recursion, by a new procedure only relying on scalar operations. The new two-loop recursion in VL-BFGS is mathematically
equivalent to the original algorithm but independent on the number of variable. Meanwhile, it reduces the memory requirement from O(md) to O(m2 ) where d could be billion-scale but m is often
less than 10. Alternatively, it only require 3 map-reduce steps compared to 2m map-reduce steps in
another naive implementation.
This new algorithm enables the implementation of a collection of machine learning algorithms to
scale to billion variables in a map-reduce environment. We demonstrate its scalability and advantage
over other approaches designed for large scale problems with billions of variables, and share our
experience after deploying it into an industrial cluster with tens of thousands of machines.
2
Related Work
L-BFGS [4][7] is a quasi-newton method based on the BFGS [8][9] update procedure, while maintaining a compact approximation of Hessian with modest storage requirement. Traditional implementation of L-BFGS follows [6] or [5] using the compact two-loop recursion update procedure.
Although it has been applied in the industry to solve various optimization problems for decades,
recent work, such as [10][11], continue to demonstrate its reliability and effectiveness over other
optimization methods. In contrast to our work, theirs implemented L-BFGS on a single machine
while we focus on the L-BFGS implementation in a distributed environment.
In the context of distributed learning, there recently have been extensive research break-through.
GraphLab [12] built a parallel distributed framework for graph computation. [13] introduced a
framework to parallelize various machine learning algorithms in a multi-core environment. [14] applied the ADMM technique into distributed learning. [15] proposed a delayed version of distributed
online learning. General distributed learning techniques closer to our work are the approaches based
on parallel gradient calculation followed by a centralized algorithm ([7][16][17]). Different from our
work, theirs built on fully connected environment such as MPI while we focus on the map-reduce
environment with loose connection. Their centralized algorithm is often the bottleneck of the whole
procedure and limits the scalability of the algorithm. For example, [17] clearly stated that it is impractical for their L-BFGS algorithm to run their large dataset due to huge memory consumption in
the centralized algorithm although L-BFGS has been shown to be an excellent candidate for their
problem. Moreover, the closest to our work lies in applying L-BFGS in the map-reduce-like environment, such as [18][2]. They are solving large-scale problems in a map-reduce adapted environment
using L-BFGS. [18] run L-BFGS on a map-reduce plus AllReduce environment to demonstrate the
power of large-scale learning with map-reduce. Although it has been shown to scale up to billion
of data instances with trillion entries in their data matrix, the number of variables in their problem
is only about 16 million due to the constraints in centralized computation of L-BFGS direction. [2]
used L-BFGS to solve the deep learning problem. It introduced the parameter servers to split a
global model into multiple partitions and store each partition separately. Despite their successes,
from the algorithmic point of view, their two-loop recursion update procedure is still highly dependent on the number of variable. Compared with these work, our proposed two-loop recursion
updating procedure is independent on the number of variables and with much better parallelism.
Furthermore, the proposed algorithm can run on pure map-reduce environment while previous work
[2] and [18] require special components such as AllReduce or parameter servers. In addition, it is
straightforward for previous work, such as [2][18][17], to leverage our proposal to scale up their
problem into another order of magnitude in terms of number of variables.
2
3
L-BFGS Algorithm
Given an optimization problem with d variables, BFGS requires to store a dense d by d matrix to
approximate the inverse Hessian, where L-BFGS only need to store a few vectors of length d to
approximate the Hessian implicitly. Let us denote f as the objective function, g as the gradient and
? as the dot product between two vectors. L-BFGS maintains the historical states of previous m
(generally m = 10) updates of current position x and its gradient g = ?f (x).
In L-BFGS algorithm, the historical states are represented as the last m updates of form sk =
xk+1 ? xk and yk = gk+1 ? gk where sk represents the position difference and yk represents the
gradient difference in iteration k. Each of them is a vector of length d. All of these 2m vector
with the original gradient gk will be used to calculate a new direction in line 3 of Algorithm 1.
Algorithm 1: L-BFGS Algorithm Outline
Input: starting point x0 , integer history size m > 0, k=1;
Output: the position x with a minimal objective function
1 while no converge do
2
Calculate gradient ?f (xk ) at position xk ;
3
Compute direction pk using Algorithm 2 ;
4
Compute xk+1 = xk + ?k pk where ?k is chosen to satisfy Wolfe conditions;
5
if k > m then
6
Discard vector pair sk?m , yk?m from memory storage;;
7
end
8
Update sk = xk+1 ? xk , yk = ?f (xk+1 ) ? ?f (xk ), k = k + 1 ;
9 end
Algorithm 2: L-BFGS two-loop recursion
Input: ?f (xk ), si , yi where i = k ? m, ..., k ? 1
Output: new direction p
1 p = ??f (xk ) ;
2 for i ? k ? 1 to k ? m do
3
?i ? ssii?y?pi ;
4
p = p ? ?i ? yi ;
5 end
s
?y
6 p = ( yk?1 ?yk?1 )p
k?1 k?1
7 for i ? k ? m to k ? 1 do
8
? = syii?y?pi ;
9
p = p + (?i ? ?) ? si ;
10 end
The core update procedure in Algorithm 1 is the line 3 to calculate a new direction pk using s and
y with current gradient ?f (xk ). The most common approach for this calculation is the two-loop
recursion in Algorithm 2[5][6]. It initializes the direction p with gradient and continues to update it
using historical states y and s. More information about two-loop recursion could be found from [5].
4
A Map-Reduce Implementation
The main procedure in Algorithm 1 lies in Line 2, 3 and 4. The calculation of gradient in Line 2
can be straightforwardly parallelized by dividing the data into multiple partitions. In the map-reduce
environment, we can use one map step to calculate the partial gradient for partial data and one reduce
to aggregate them into a global gradient vector. The verification of the Wolfe condition only depends
on the calculation of the objective function following the line search procedure[5]. So thus Line 4
can also be easily parallelized following the same approach as Line 2. Therefore, the challenge in
the L-BFGS algorithm is Line 3. In other words,the difficulties come from the calculation of the
two-loop recursion, as shown in Algorithm 2.
3
4.1
Centralized Update
The simplest implementation for Algorithm 2 may be to run it in a single processor. We can easily
perform this in a singleton reduce. However, the challenge is that Algorithm 2 requires 2m + 1
vectors and each of them has a length of d. This could be feasible when d is in million scale. Nevertheless, when d is in billion scale, either the storage or the computation cost becomes a significant
challenge and makes it impractical to implement it in map-reduce. Given the Ads CTR prediction
task [1] as an example, there are more than 1 billion of features. If we set m = 10 in a linear model,
it will produce 21 ? 1 = 21 billion variables. Even if we compactly use a single-precision floating
point to represent a variable, it requires 84 GB memory to store the historical states and gradient.
For a map-reduce cluster built from commodity hardware and shared with other applications, this is
generally unfeasible nowadays. For example, for the cluster into which we deployed the L-BFGS,
its maximal memory limitation for a map-reduce step is 6 GB.
4.2
Distributed Update
Due to the storage limitation in centralized update, an alternative is to store s and y into multiple
partitions without overlap and use a map-reduce step to calculate every dot product, such as si ?p and
si ? yi in Line 3 of Algorithm 2. Yet, if each dot product within the for-loop in Algorithm 2 requires a
map-reduce step to perform the calculation, this will result in at least 2m map-reduce steps in a twoloop recursion. If we call Algorithm 2 for N times(iterations) in Algorithm 1, it will lead to 2mN
map-reduce steps. For example, if m = 10 and N = 100, this will produce 2000 map-reduce steps
in a map-reduce job. Unfortunately, each map-reduce step will bring significant overhead due to the
scheduling cost and application launching cost. For a job with thousands of map-reduce steps, both
these cost often dominate the overall running time and make the useful computational time spent
in algorithmic vector operations negligible. Moreover, given our current production cluster as an
example, a job with such a huge number of map-reduce step is too large for execution. It will trigger
a compilation timeout error before becoming too complicated for an execution engine to execute it.
5
Vector-free L-BFGS
For the reasons mentioned, a feasible two-loop recursion procedure has to limit both the memory
consumption and the number of map-reduce steps per iteration. To strictly limit the memory consumption in Algorithm 2, we can not store the 2m + 1 vectors with length d in memory unless d is
only up to million scale. To comply with the allowable map-reduce steps per iteration, it is neither
practical to perform map-reduce steps within the for-loop in Algorithm 2. Both of these assumptions
motivate us to carefully re-examine Algorithm 2 and lead to the proposed algorithm in this section.
5.1
Basic Idea
Before illustrating the new procedure, let us describe following three observations in Algorithm 2
that guide the design of the new procedure in Algorithm 3:
1. All inputs are invariable during Algorithm 2.
2. All operations applied on p are linear with respect to the inputs. In other words, p could be
formalized as a linear combination of the inputs although its coefficients are unknown.
3. The core numeric operation is the dot product between two vectors.
Observation 1 and 2 motivate us to formalize the inputs as (2m + 1) invariable base vectors.
b1 = sk?m , b2 = sk?m+1 , ..., bm = sk?1
(1)
bm+1 = yk?m , bm+2 = yk?m+1 , ..., b2m = yk?1
(2)
b2m+1 = ?f (xi )
(3)
So thus we can represent p as a linear combination of bi . Assume ? as the scalar coefficients in this
linear combination, we can write p as:
2m+1
X
p=
?k bk
(4)
k=1
4
Since bk are the inputs and invariants during the two-loop recursion, if we can calculate the coefficients ?k , we can proceed to calculate the direction p.
Following observation 3 with an re-examination of Algorithm 2, we classify the dot product operations into two categories in terms of whether p is involved in the calculation. For the first category
only involving the dot product between the inputs (si , yi ), a straightforward intuition is to precompute their dot products to produce a scalar, so as to replace each dot product with a scalar in
the two-loop recursion. However, the second category of dot products involving p can not follow
this same procedure. Because the direction p is ever-changing during the for loop, any dot products
involving p can not be settled or pre-computed. Fortunately, thanks to the linear decomposition of
p in observation 2 and Eqn.4, we can decompose any dot product involving p into a summation of
dot products with its based vectors and corresponding coefficients. This new elegant mathematical
procedure only happens after we formalize p as the linear combination of the base vectors.
5.2
The VL-BFGS Algorithm
We present the algorithmic procedure in Algorithm 3. Let us denote the results of dot products
between every two base vectors as a scalar matrix of (2m + 1) ? (2m + 1) scalars. The proposed
VL-BFGS algorithm only takes it as the input. Similar as the original L-BFGS algorithm, it has a
two-loop recursion, but all the operations are only dependent on scalar operations. In Line 1-2, it
assigns the initial values for ?i . This is equivalent to Line 1 in Algorithm 2 to use opposite direction
of gradient as the initial direction. The original calculation of ?i in Line 6 relies on the direction
vector p. It is worth noting that p is variable within the first loop in which ? is updated. So thus we
can not pre-compute any dot product involving p. However, as mentioned earlier and according to
observation 2 and Eqn.4, we can formalize bj ? p as a summation from a list of dot products between
base vectors and corresponding coefficients, as shown in Line 6 of Algorithm 3. Meanwhile, since all
base vectors are invariable, their dot products can be pre-computed and replaced with scalars,which
then multiply the ever-changing ?l . But these are only scalar operations and they are extremely
efficient. Line 7 continues to update scalar coefficient ?m+j , which is equivalent to update the
direction p with respect to the base vector bm+j or corresponding yj . This whole procedure is the
same when we apply it to Line 14 and 15. With the new formalization of p in Eqn.4 and the
Algorithm 3: Vector-free L-BFGS two-loop recursion
Input: (2m + 1) ? (2m + 1) dot product matrix between bi
Output: The coefficients ?i where i = 1, 2, ...2m + 1
1 for i ? 1 to 2m + 1 do
2
?i = i ? 2m ? 0 : ?1
3 end
4 for i = k ? 1 to k ? m do
5
j = i ? (k ? m) + 1 ;
6
7
8
9
10
11
12
13
b ?p
?i ? ssii?y?pi = bj ?bjm+j =
?m+j = ?m+j ? ?i ;
end
for i ? 1 to 2m + 1 do
m ?b2m
?i = ( bb2m
?b2m )?i
end
for i ? k ? m to k ? 1 do
j = i ? (k ? m) + 1 ;
b
?p
P2m+1
P2m+1
? b
?l bl ?bj
bj ?bm+j
l=1
l m+j
l=1
? = bjm+j
?bm+j =
bj ?bm+j
15
?j = ?j + (?i ? ?)
16 end
14
?bl
;
;
invariability of yi and si during Algorithm 2, Line 4 in Algorithm 2 updating with yi (equivalent to
bm+j ) is mathematically equivalent to Line 7 in Algorithm 3, so as Line 9 in Algorithm 2 and Line
15 in Algorithm 3. For other lines between these two algorithms, it is easy to infer their equivalence
with the consideration of Eqn.1-4. Thus, Algorithm 3 is mathematically equivalent to Algorithm 2.
5
5.3
Complexity Analysis and Comparison
Using the dot product matrix of scalars as the input, the calculation in Algorithm 3 is substantially
efficient, since all the calculation is based on scalars. Altogether, it only requires 8m2 multiplications
between scalars in the two for-loops. This is tiny compared to any vector operation involving billionscale of variables. Thus, it is not necessary to parallelize Algorithm 3 in implementation.
To integrate Algorithm 3 as the core step in Algorithm 1, there are two extra steps we need to
perform before and after it. One is to calculate the dot product matrix between the (2m + 1) base
vectors. Because all base vectors have the same dimension d, we can partition them using the
same way and use one map-reduce step to calculate the dot product matrix. This computation is
greatly parallelizable and intrinsically suitable for map-reduce. Even without the consideration of
parallization, a first glance tells us it may require about 4m2 dot products. However, since all the
si and yi except the first ones are unchanged in a new iteration, we can save the tiny dot product
matrix and reuse most entries across iterations. With the consideration of the commutative law of
multiplication since si ? yj ? yj ? si , each new iteration only need to calculate 6m new dot products
which involve new sk , yk and gk . Thus, the complexity is only 6md and this calculation is fully
parallel in map-reduce, with each partition only calculating a small portion of 6md multiplications.
The other and the final step is to calculate the new direction p based on ?i and the base vectors. The
complexity is another 2md multiplications, which means the overall complexity of the algorithm
is 8md multiplications. Since the overall ? is just a tiny vector with 2m + 1 dimensions, we can
join it with all the other base vectors, and then use the same approach as dot product calculation to
produce the final direction p using Eqn.4. A single map-reduce step is sufficient for this final step.
Altogether, without considering the gradient calculation which is same to all algorithms, VL-BFGS
only require 3 map-reduce steps for one iteration in the update.
For the centralized update approach in section 4.1, it also requires 6md multiplications in each
two loop recursion. In addition to being a centralized approach, as we analyzed above, it requires
(2m + 1) ? d memory storage. This clearly limits its applications to large-scale problems. On the
other hand, VL-BFGS in Algorithm 3 only requires (2m+1)2 memory storage and is independent on
d. For the distributed approach in section 4.2, it requires at least 2m map-reduce step in a two-loop
recursion. Given the number of iteration as N (generally N > 100), the total number of map-reduce
steps is 2mN . Fortunately, the VL-BFGS only requires 3N map-reduce steps. In summary, VLBFGS algorithm enjoys a similar overall complexity but it is born with massive degree of parallelism.
For problem with billion scale of variables, it is the only map-reduce friendly implementation of the
three different approaches.
6
Experiment and Discussion
As demonstrated above, it is clear that VL-BFGS has a better scalability property than original LBFGS. Although it is always desirable to invent an exact algorithm that could be mathematically
proved to obtain a better scalability property, it is beneficial to demonstrate the value of larger
number of variables with an industrial application. On the other hand, for a problem with billions
of variables, there are existing practical approaches to reduce it into a smaller number of variables
and then solve it with traditional approaches designed for centralized algorithm. In this section, we
justify the value of learning large scale variables and simultaneously compare it with the hashing
approach, and finally demonstrate the scalability advantage of VL-BFGS.
6.1
Dataset and Experimental Setting
The dataset we used is from an Ads Click-through Rate (CTR) prediction problem [1] collected from
an industrial search engine. The click event (click or not) is used as the label for each instance. The
features include the terms from a query and an Ad keyword along with the contextual information
such as Ad position, session-related information and time. We collect 30 days of data and split them
into training and test set chronologically. The data from the first 20 days are used as the training
set and rest 10 days are used as test set. The total training data have about 12 billions instances and
another 6 billion in testing data. There are 1,038,934,683 features the number of non-zero features
per instance is about 100 on average. Altogether it has about 2 trillion entries in the data matrix.
6
Table 1: Relative AUC Performance over different number of variables
K
Relative AUC Performance
Baseline(K=1,038,934,683) 0.0%
K=250 millions
-0.1007388%
K=100 millions
-0.1902843%
K= 10 millions
-0.3134094%
K= 1 millions
-0.5701142%
Table 2: Relative AUC Performance over different number of Hash bits
K
Relative AUC Performance
Baseline(K=1,038,934,683) 0.0%
K=64 millions(26 bits)
-0.1063033%
K=16 millions(24 bits)
-0.2323647%
K= 4 millions(22 bits)
-0.3300788%
K= 1 millions(20 bits)
-0.5080904%
We run logistic regression training, so thus each feature corresponds to a variable. The model is
evaluated based on the testing data using Area Under ROC Curve [19], denoted as AUC. We set
the historical state length m = 10 and enforce L1[20] regularizer to avoid overfitting and achieve
sparsity. The regularizer parameter is tuned following the approach in [18].
We run the experiment in a shared cluster with tens of thousands of machines. Each machine has up
to 12 concurrent vertices. A vertex is generally a map or reduce step with an allocation of 2 cores
and 6G memory. There are more than 1000 different jobs running simultaneously but this number
also varies significantly. We split the training data into 400 partitions and allocate 400 tokens for this
job, which means this job can use up to 400 vertices at the same time. When we partition vectors to
calculate their dot products, our strategy is to allocate up to 5 million entries in a partial vector. For
example, 1 billion variables will be split into 200 partitions evenly.
We use the model trained with original 1 billion features as the baseline. All the other experiments
are compared with it. Since we are not allowed to exhibit the exact AUC number due to privacy
consideration, we report the relative change compared with the baseline. The scale of the dataset
makes any relative AUC change over 0.001% produce a p-value less than 0.01.
6.2
Value of Large Number of Variables
To reduce the number of variables in the original problem, we sort the features based on their frequency in the training data. If we plan to reduce the problem to K variables, we keep the top K
frequent features. The baseline without filtering is equivalent to K = 1, 038, 934, 683. We choose
different K values and report the relative AUC number in Table 1.
The table shows that while we reduce the number of variables, the results consistently decline significantly. When the number of variables is 1 million, the drop is more than 0.5% . This is considerably
significant for the problem. Even when we increase the number of variable up to 250 million, the
decline is still obvious and significant. This demonstrates that the large number of variables is really
needed to learn a good model and the value of learning with billion-scale of variables.
6.3
Comparison with Hashing
We follow the approach in [21][18] to calculate a new hash value for each original feature value
based on a hash function in [18]. The number of hash bits ranges from 20 to 26. Experimental
results compared with the baseline in terms of relative AUC performance are presented in Table 2
Consistently with previous results, all the hashing experiments result in degradation. For the experiment with 20 bits, the degradation is 0.5%. This is a substantial decline for this problem. When we
increase the number of bits till 26, the gap becomes smaller but still noticeable. All of these consis7
tently demonstrate that the hashing approach will sacrifice noticeable performance. It is beneficial
to train with large-scale number of raw features.
6.4
Training Time Comparison
We compare the L-BFGS in section 4.1 with the proposed VL-BFGS. To enable a larger number of
variable support for L-BFGS, we reduce the m parameter into 3. We conduct the experiments with
varying number of feature number and report their corresponding running time. We use the original
data after hashing into 1M features as the baseline and compare all the other experiments with it and
report the relative training time for same number of iterations. We run each experiment 5 times and
report their mean to cope with the variance in each run. The results with respect to different hash bits
range from 20 to 29 and the original 1B features are shown in figure 1. When the number of features
is less than 10M, the original L-BFGS has a small advantage over VL-BFGS. However, when we
continue to increase the feature number, the running time of L-BFGS grows quickly while that of
VL-BFGS increases slowly. On the other hand, when we increase the feature number to 512M, the
L-BFGS fails with an out-of-memory exception, while VL-BFGS can easily scale to 1B features.All
of these clearly show the scalability advantage of VL-BFGS over traditional L-BFGS.
Figure 1: Training time over feature number.
7
Conclusion
We have presented a new vector-free exact L-BFGS updating procedure called VL-BFGS. As opposed to original L-BFGS algorithm in map-reduce, the core two-loop recursion in VL-BFGS is
independent on the number of variables. This enables it to be easily parallelized in map-reduce
and scale up to billions of variables. We present its mathematical equivalence to original L-BFGS,
show its scalability advantage over traditional L-BFGS in map-reduce with a great degree of parallelism, and perform experiments to demonstrate the value of large-scale learning with billions of
variables using VL-BFGS. Although we emphasis the implementation on map-reduce in this paper,
VL-BFGS can be straightforwardly utilized by other distributed frameworks to avoid their centralized problem and scale up their algorithms. In short, VL-BFGS is highly beneficial for machine
learning algorithms relying on L-BFGS to scale up to another order of magnitude.
8
References
[1] T. Graepel, J.Q. Candela, T. Borchert, and R. Herbrich. Web-Scale Bayesian Click-Through
Rate Prediction for Sponsored Search Advertising in Microsofts Bing Search Engine. In International Conference on Machine Learning, pages 13?20. Citeseer, 2010.
[2] Jeffrey Dean, G Corrado, Rajat Monga, Kai Chen, and Matthieu Devin. Large Scale Distributed
Deep Networks. Advances in Neural Information Processing Systems 25, pages 1232?1240,
2012.
[3] Jeffrey Dean and Sanjay Ghemawat. MapReduce : Simplified Data Processing on Large Clusters. Communications of the ACM, 51(1):1?13, 2008.
[4] DC Liu and Jorge Nocedal. On the limited memory BFGS method for large scale optimization.
Mathematical programming, 45(1-3):503?528, 1989.
[5] J Nocedal and S J Wright. Numerical Optimization, volume 43 of Springer Series in Operations Research. Springer, 1999.
[6] C Zhu, RH Byrd, P Lu, and J Nocedal. Algorithm 778: L-BFGS-B: Fortran subroutines for
large-scale bound-constrained optimization. ACM Transactions on Mathematical Software, 23,
pages 550?560, 1997.
[7] Stephen G. Nash and Ariela Sofer. Block truncated-Newton methods for parallel optimization.
Mathematical Programming, 45(1-3):529?546, 1989.
[8] Jorge Nocedal. Updating quasi-Newton matrices with limited storage, 1980.
[9] DF Shanno. On broyden-fletcher-goldfarb-shanno method. Journal of Optimization Theory
and Applications, 1985.
[10] N Schraudolph, J Yu, and S G?unter. A stochastic quasi-Newton method for online convex
optimization. Journal of Machine Learning Research, pages 436?443, 2007.
[11] H Daum?e III. Notes on CG and LM-BFGS optimization of logistic regression. 2004.
[12] Y Low, J Gonzalez, and A Kyrola. Graphlab: A new framework for parallel machine learning.
Uncertainty in Artificial Intelligence, 2010.
[13] Cheng-Tao Chu, Sang Kyun Kim, Yi-An Lin, YuanYuan Yu, Gary Bradski, Andrew Y. Ng,
and Kunle Olukotun. Map-Reduce for Machine Learning on Multicore. In Advances in Neural
Information Processing Systems 19, pages 281?288. MIT Press, 2007.
[14] S Boyd, N Parikh, and E Chu. Distributed optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends in in Machine Learning,
(3):1?122, 2011.
[15] J Langford, AJ Smola, and M Zinkevich. Slow learners are fast. Advances in Neural Information Processing Systems 22, pages 2331?2339, 2009.
[16] C Teo, Le.Q, A Smola, and SVN Vishwanathan. A scalable modular convex solver for regularized risk minimization. ACM SIGKDD Conference on Knowledge Discovery and Data Mining,
2007.
[17] S Gopal and Y Yang. Distributed training of Large-scale Logistic models. Proceedings of the
30th International Conference on Machine Learning, 28:287?297, 2013.
[18] Alekh Agarwal, Oliveier Chapelle, Miroslav Dud??k, and John Langford. A Reliable Effective
Terascale Linear Learning System. Journal of Machine Learning Research, 15:1111?1133,
2014.
[19] CX Ling, J Huang, and H Zhang. AUC: a statistically consistent and more discriminating
measure than accuracy. IJCAI, pages 329?341, 2003.
[20] Galen Andrew and Jianfeng Gao. Scalable training of l1-regularized log-linear models. Proceedings of the 24th International Conference on Machine Learning, pages 33?40, 2007.
[21] K Weinberger, A Dasgupta, J Langford, Smola.A, and J Attenberg. Feature hashing for large
scale multitask learning. International Conference on Machine Learning, 2009.
9
| 5333 |@word multitask:1 illustrating:1 version:1 decomposition:2 citeseer:1 initial:2 born:1 liu:1 series:1 tuned:1 existing:1 current:3 com:1 contextual:1 si:9 yet:2 chu:2 john:1 devin:1 numerical:1 partition:9 enables:3 designed:2 drop:1 update:18 sponsored:1 hash:5 intelligence:1 xk:13 core:6 short:1 node:1 herbrich:1 launching:1 zhang:1 mathematical:6 along:1 prove:1 overhead:2 privacy:1 x0:1 sacrifice:1 expected:1 frequently:1 examine:2 multi:1 relying:2 byrd:1 little:1 considering:1 increasing:1 becomes:2 solver:1 moreover:2 substantially:1 impractical:2 commodity:3 every:2 friendly:1 demonstrates:1 before:3 negligible:1 limit:4 consequence:1 era:1 despite:1 parallelize:3 becoming:1 plus:1 emphasis:1 studied:1 equivalence:3 collect:1 limited:4 bi:2 adoption:1 range:2 statistically:1 practical:2 yj:3 testing:2 practice:1 block:2 implement:2 procedure:20 area:1 significantly:2 boyd:1 word:2 pre:3 unfeasible:1 scheduling:1 storage:8 context:1 applying:1 risk:1 equivalent:7 map:51 demonstrated:1 dean:2 zinkevich:1 straightforward:2 starting:1 convex:2 formalized:1 assigns:1 pure:1 matthieu:1 m2:3 dominate:1 handle:1 updated:1 trigger:1 massive:4 exact:3 programming:2 trend:2 wolfe:2 expensive:1 particularly:1 updating:4 continues:2 utilized:1 wang:1 solved:2 reexamine:1 thousand:4 calculate:13 connected:1 keyword:1 yk:10 mentioned:2 intuition:1 environment:14 substantial:1 complexity:6 nash:1 motivate:2 trained:1 solving:2 creates:1 efficiency:1 learner:1 compactly:1 easily:4 various:5 represented:1 regularizer:2 train:1 fast:1 effective:3 describe:1 query:1 artificial:1 tell:1 aggregate:1 jianfeng:1 modular:1 larger:2 solve:3 kai:1 final:3 online:2 timeout:1 advantage:5 propose:2 product:26 maximal:1 frequent:1 loop:23 till:1 achieve:1 yuanyuan:1 weizhu:1 scalability:8 billion:23 ijcai:1 cluster:8 requirement:2 produce:5 spent:1 depending:1 andrew:2 multicore:1 noticeable:2 job:6 dividing:1 implemented:2 come:1 direction:15 stochastic:1 enable:1 require:5 galen:1 really:1 decompose:1 summation:2 mathematically:4 strictly:1 wright:1 great:2 fletcher:1 algorithmic:3 bj:5 b2m:4 lm:1 early:1 estimation:1 label:1 teo:1 concurrent:1 create:1 minimization:1 mit:1 clearly:4 always:1 gopal:1 zhou:1 avoid:2 varying:1 focus:2 consistently:2 kyrola:1 greatly:2 industrial:3 sigkdd:1 contrast:1 baseline:7 cg:1 kim:1 dependent:2 vl:19 multicores:1 quasi:3 subroutine:1 tao:1 overall:4 among:1 denoted:1 plan:1 constrained:1 special:1 ng:1 represents:2 yu:2 report:5 few:1 simultaneously:2 delayed:1 floating:1 replaced:1 jeffrey:2 microsoft:2 huge:7 centralized:10 bradski:1 highly:2 mining:1 multiply:1 analyzed:1 compilation:1 nowadays:2 closer:1 partial:3 necessary:1 experience:1 unter:1 modest:1 unless:1 conduct:1 re:2 minimal:1 miroslav:1 instance:5 industry:1 classify:1 obstacle:1 earlier:1 cost:5 vertex:3 entry:4 too:4 stored:2 straightforwardly:2 varies:1 considerably:1 thanks:1 fundamental:1 international:4 shanno:2 discriminating:1 quickly:1 sofer:1 ctr:3 settled:1 opposed:1 choose:1 slowly:1 huang:1 sang:1 bfgs:78 singleton:1 b2:1 coefficient:7 satisfy:1 ad:5 depends:1 break:1 lot:1 view:1 candela:1 portion:1 start:1 sort:1 maintains:1 parallel:5 complicated:1 accuracy:1 variance:1 efficiently:1 raw:1 bayesian:1 lu:1 advertising:1 worth:1 processor:2 history:1 deploying:1 parallelizable:1 frequency:1 involved:1 obvious:1 associated:1 dataset:4 proved:1 popular:1 intrinsically:1 knowledge:2 improves:1 graepel:1 formalize:3 carefully:2 hashing:6 day:3 follow:2 execute:1 evaluated:1 furthermore:1 just:1 stage:1 smola:3 langford:3 hand:3 eqn:5 web:1 glance:1 logistic:3 aj:1 grows:1 building:1 multiplier:1 alternating:1 dud:1 goldfarb:1 deal:1 during:4 auc:10 maintained:1 mpi:1 allowable:1 outline:1 demonstrate:8 l1:2 bring:1 consideration:4 recently:1 parikh:1 common:2 volume:1 million:15 theirs:2 significant:5 broyden:1 session:1 dot:26 reliability:1 chapelle:1 alekh:1 base:10 closest:1 recent:1 discard:1 store:7 server:2 continue:2 success:1 jorge:2 yi:8 fortunately:2 parallelized:3 converge:1 corrado:1 stephen:1 multiple:3 desirable:1 reduces:1 infer:1 calculation:13 schraudolph:1 lin:1 devised:1 kunle:1 impact:1 prediction:4 scalable:3 basic:1 involving:6 invent:1 regression:2 df:1 iteration:11 represent:2 ssii:2 monga:1 agarwal:1 proposal:1 addition:2 separately:1 extra:1 rest:1 elegant:1 allreduce:2 effectiveness:1 integer:1 call:1 leverage:1 noting:1 yang:1 split:4 easy:1 iii:1 variety:1 opposite:1 click:4 reduce:57 idea:1 decline:3 svn:1 bottleneck:1 whether:1 allocate:2 gb:2 reuse:1 invariable:3 hessian:4 proceed:1 deep:3 generally:4 useful:1 clear:1 involve:1 amount:2 ten:3 hardware:2 category:3 simplest:1 p2m:2 per:4 diverse:1 write:1 dasgupta:1 nevertheless:1 changing:2 neither:1 nocedal:4 graph:1 chronologically:1 olukotun:1 run:8 package:2 inverse:1 uncertainty:1 reasonable:1 gonzalez:1 scaling:1 bit:9 bound:1 followed:1 cheng:1 adapted:1 constraint:1 vishwanathan:1 software:2 extremely:1 according:1 combination:4 precompute:1 across:1 beneficial:3 smaller:2 happens:1 invariant:1 bing:1 loose:1 needed:1 bjm:2 fortran:1 end:8 studying:1 operation:12 apply:2 enforce:1 attenberg:1 save:1 alternative:1 weinberger:1 altogether:3 original:16 top:1 running:4 include:1 maintaining:1 newton:4 calculating:1 daum:1 conquer:2 unchanged:1 bl:2 objective:3 initializes:1 strategy:1 md:7 traditional:5 exhibit:1 gradient:15 consumption:3 evenly:1 collected:1 reason:1 length:5 unfortunately:1 gk:4 negative:1 stated:1 implementation:15 design:1 unknown:1 perform:6 observation:5 datasets:1 truncated:1 kyun:1 ever:2 communication:1 dc:1 parallelizing:1 introduced:2 bk:2 pair:1 extensive:1 connection:1 engine:3 zhenghao:1 parallelism:4 sanjay:1 sparsity:1 challenge:3 built:4 reliable:1 memory:17 power:1 critical:1 overlap:1 difficulty:1 examination:1 suitable:1 event:1 regularized:2 recursion:19 mn:2 zhu:1 naive:2 comply:1 mapreduce:3 discovery:1 multiplication:6 relative:9 law:1 fully:2 limitation:3 allocation:1 filtering:1 foundation:1 integrate:1 degree:3 verification:1 sufficient:1 consistent:1 terascale:1 tiny:3 share:1 pi:3 production:2 summary:1 token:1 last:1 free:6 enjoys:1 guide:1 wide:1 distributed:15 curve:1 dimension:2 world:1 avoids:1 numeric:1 collection:2 simplified:1 bm:8 historical:7 cope:1 transaction:1 approximate:3 compact:2 implicitly:1 keep:1 graphlab:2 global:2 overfitting:1 b1:1 xi:1 alternatively:1 search:4 decade:1 sk:8 table:5 learn:1 excellent:2 borchert:1 meanwhile:2 pk:3 dense:1 main:1 rh:1 big:2 whole:2 ling:1 nothing:1 allowed:1 join:1 roc:1 deployed:1 slow:1 precision:1 formalization:1 position:5 fails:1 candidate:1 lie:2 ghemawat:1 list:1 effectively:2 magnitude:2 execution:2 commutative:1 demand:1 chen:2 gap:1 cx:1 explore:2 lbfgs:2 gao:1 prevents:1 scalar:13 springer:2 corresponds:1 gary:1 relies:1 trillion:2 acm:3 shared:3 admm:1 replace:2 feasible:2 change:2 typical:1 specifically:1 except:1 justify:1 degradation:2 called:3 total:2 experimental:2 exception:1 support:1 rajat:1 |
4,787 | 5,334 | Recovery of Coherent Data via Low-Rank
Dictionary Pursuit
Ping Li
Department of Statistics and Biostatistics
Department of Computer Science
Rutgers University
Piscataway, NJ 08854, USA
[email protected]
Guangcan Liu
Department of Statistics and Biostatistics
Department of Computer Science
Rutgers University
Piscataway, NJ 08854, USA
[email protected]
Abstract
The recently established RPCA [4] method provides a convenient way to restore
low-rank matrices from grossly corrupted observations. While elegant in theory
and powerful in reality, RPCA is not an ultimate solution to the low-rank matrix recovery problem. Indeed, its performance may not be perfect even when
data are strictly low-rank. This is because RPCA ignores clustering structures of
the data which are ubiquitous in applications. As the number of cluster grows,
the coherence of data keeps increasing, and accordingly, the recovery performance of RPCA degrades. We show that the challenges raised by coherent data
(i.e., data with high coherence) could be alleviated by Low-Rank Representation
(LRR) [13], provided that the dictionary in LRR is configured appropriately. More
precisely, we mathematically prove that if the dictionary itself is low-rank then
LRR is immune to the coherence parameter which increases with the underlying
cluster number. This provides an elementary principle for dealing with coherent
data and naturally leads to a practical algorithm for obtaining proper dictionaries
in unsupervised environments. Experiments on randomly generated matrices and
real motion sequences verify our claims. See the full paper at arXiv:1404.4032.
1 Introduction
Nowadays our data are often high-dimensional, massive and full of gross errors (e.g., corruptions,
outliers and missing measurements). In the presence of gross errors, the classical Principal Component Analysis (PCA) method, which is probably the most widely used tool for data analysis and
dimensionality reduction, becomes brittle ? A single gross error could render the estimate produced
by PCA arbitrarily far from the desired estimate. As a consequence, it is crucial to develop new statistical tools for robustifying PCA. A variety of methods have been proposed and explored in the
literature over several decades, e.g., [2, 3, 4, 8, 9, 10, 11, 12, 24, 13, 16, 19, 25]. One of the most exciting methods is probably the so-called RPCA (Robust Principal Component Analysis) method [4],
which was built upon the exploration of the following low-rank matrix recovery problem:
Problem 1 (Low-Rank Matrix Recovery) Suppose we have a data matrix X ? Rm?n and we
know it can be decomposed as
X = L 0 + S0 ,
m?n
(1.1)
where L0 ? R
is a low-rank matrix each column of which is a data point drawn from some
low-dimensional subspace, and S0 ? Rm?n is a sparse matrix supported on ? ? {1, ? ? ? , m} ?
{1, ? ? ? , n}. Except these mild restrictions, both components are arbitrary. The rank of L0 is unknown, the support set ? (i.e., the locations of the nonzero entries of S0 ) and its cardinality (i.e.,
the amount of the nonzero entries of S0 ) are unknown either. In particular, the magnitudes of the
nonzero entries in S0 may be arbitrarily large. Given X, can we recover both L0 and S0 , in a
scalable and exact fashion?
1
cluster 1
cluster 2
Figure 1: Exemplifying the extra structures of low-rank data. Each entry of the data matrix is a grade
that a user assigns to a movie. It is often the case that such data are low-rank, as there exist wide
correlations among the grades that different users assign to the same movie. Also, such data could
own some clustering structure, since the preferences of the same type of users are more similar to
each other than to those with different gender, personality, culture and education background. In
summary, such data (1) are often low-rank and (2) exhibit some clustering structure.
The theory of RPCA tells us that, very generally, when the low-rank matrix L0 is meanwhile incoherent (i.e., with low coherence), both the low-rank and the sparse matrices can be exactly recovered
by using the following convex, potentially scalable program:
min kLk? + ?kSk1 , s.t.
L,S
X = L + S,
(1.2)
where k ? k? is the nuclear norm [7] of a matrix, k ? k1 denotes the ?1 norm of a matrix seen as
a long vector, and ? > 0 is a parameter. Besides of its elegance in theory, RPCA also has good
empirical performance in many practical areas, e.g., image processing [26], computer vision [18],
radar imaging [1], magnetic resonance imaging [17], etc.
While complete in theory and powerful in reality, RPCA cannot be an ultimate solution to the lowrank matrix recovery Problem 1. Indeed, the method might not produce perfect recovery even when
L0 is strictly low-rank. This is because RPCA captures only the low-rankness property, which
however is not the only property of our data, but essentially ignores the extra structures (beyond
low-rankness) widely existing in data: Given the low-rankness constraint that the data points (i.e.,
columns vectors of L0 ) locate on a low-dimensional subspace, it is unnecessary for the data points
to locate on the subspace uniformly at random and it is quite normal that the data may have some
extra structures, which specify in more detail how the data points locate on the subspace. Figure 1
demonstrates a typical example of extra structures; that is, the clustering structures which are ubiquitous in modern applications. Whenever the data are exhibiting some clustering structures, RPCA
is no longer a method of perfection. Because, as will be shown in this paper, while the rank of L0 is
fixed and the underlying cluster number goes large, the coherence of L0 keeps heightening and thus,
arguably, the performance of RPCA drops.
To better handle coherent data (i.e., the cases where L0 has large coherence parameters), a seemingly straightforward idea is to avoid the coherence parameters of L0 . However, as explained in [4],
the coherence parameters are indeed necessary (if there is no additional condition assumed on the
data). This paper shall further indicate that the coherence parameters are related in nature to some
extra structures intrinsically existing in L0 and therefore cannot be discarded simply. Interestingly,
we show that it is possible to avoid the coherence parameters by using some additional conditions,
which are easy to obey in supervised environment and can also be approximately achieved in unsupervised environment. Our study is based on the following convex program termed Low-Rank
Representation (LRR) [13]:
min kZk? + ?kSk1 , s.t.
Z,S
X = AZ + S,
(1.3)
where A ? Rm?d is a size-d dictionary matrix constructed in advance1, and ? > 0 is a parameter. In
order for LRR to avoid the coherence parameters which increase with the cluster number underlying
1
It is not crucial to determine the exact value of d. Suppose Z ? is the optimal solution with respect to Z.
Then LRR uses AZ ? to restore L0 . LRR falls back to RPCA when A = I (identity matrix). Furthermore, it can
be proved that the recovery produced by LRR is the same as RPCA whenever the dictionary A is orthogonal.
2
L0 , we prove that it is sufficient to construct in advance a dictionary A which is low-rank by itself.
This gives a generic prescription to defend the possible infections raised by coherent data, providing
an elementary criteria for learning the dictionary matrix A. Subsequently, we propose a simple and
effective algorithm that utilizes the output of RPCA to construct the dictionary in LRR. Our extensive experiments demonstrated on randomly generated matrices and motion data show promising
results. In summary, the contributions of this paper include the following:
? For the first time, this paper studies the problem of recovering low-rank, and coherent (or
less incoherent as equal) matrices from their corrupted versions. We investigate the physical
regime where coherent data arise. For example, the widely existing clustering structures
may lead to coherent data. We prove some basic theories for resolving the problem, and
also establish a practical algorithm that outperforms RPCA in our experimental study.
? Our studies help reveal the physical meaning of coherence, which is now standard and
widely used in various literatures, e.g., [2, 3, 4, 25, 15]. We show that the coherence
parameters are not ?assumptions? for a proof, but rather some excellent quantities that
relate in nature to the extra structures (beyond low-rankness) intrinsically existing in L0 .
? This paper provides insights regarding the LRR model proposed by [13]. While the special
case of A = X has been extensively studied, the LRR model (1.3) with general dictionaries
is not fully understood yet. We show that LRR (1.3) equipped with proper dictionaries
could well handle coherent data.
? The idea of replacing L with AZ is essentially related to the spirit of matrix factorization
which has been explored for long, e.g., [20, 23]. In that sense, the explorations of this paper
help to understand why factorization techniques are useful.
2 Summary of Main Notations
Capital letters such as M are used to represent matrices, and accordingly, [M ]ij denotes its (i, j)th
entry. Letters U , V , ? and their variants (complements, subscripts, etc.) are reserved for left singular
vectors, right singular vectors and support set, respectively. We shall abuse the notation U (resp. V )
to denote the linear space spanned by the columns of U (resp. V ), i.e., the column space (resp. row
space). The projection onto the column space U , is denoted by PU and given by PU (M ) = U U T M ,
and similarly for the row space PV (M ) = M V V T . We shall also abuse the notation ? to denote
the linear space of matrices supported on ?. Then P? and P?? respectively denote the projections
onto ? and ?c such that P? + P?? = I, where I is the identity operator. The symbol (?)+ denotes
T
the Moore-Penrose pseudoinverse of a matrix: M + = VM ??1
M UM for a matrix M with Singular
2
T
Value Decomposition (SVD) UM ?M VM .
Six different matrix norms are used in this paper. The first three norms are functions of the singular
values: 1) The operator norm (i.e., the largest singular value) denoted by kM k, 2) the Frobenius
norm (i.e., square root of the sum of squared singular values) denoted by kM kF , and 3) the nuclear
norm (i.e., the sum of singular values) denoted byP
kM k? . The other three are the ?1 , ?? (i.e.,
sup-norm) and ?2,? norms of a matrix: kM k1 =
i,j |[M ]ij |, kM k? = maxi,j {|[M ]ij |} and
qP
2
kM k2,? = maxj {
i [M ]ij }, respectively.
The Greek letter ? and its variants (e.g., subscripts and superscripts) are reserved for the coherence
parameters of a matrix. We shall also reserve two lower case letters, m and n, to respectively denote
the data dimension and the number of data points, and we use the following two symbols throughout
this paper:
n1 = max(m, n)
and n2 = min(m, n).
3 On the Recovery of Coherent Data
In this section, we shall firstly investigate the physical regime that raises coherent (or less incoherent) data, and then discuss the problem of recovering coherent data from corrupted observations,
providing some basic principles and an algorithm for resolving the problem.
2
In this paper, SVD always refers to skinny SVD. For a rank-r matrix M ? Rm?n , its SVD is of the form
T
UM ?M VM
, with UM ? Rm?r , ?M ? Rr?r and VM ? Rn?r .
3
3.1 Coherence Parameters and Their Properties
As the rank function cannot fully capture all characteristics of L0 , it is necessary to define some
quantities to measure the effects of various extra structures (beyond low-rankness) such as the clustering structure as demonstrated in Figure 1. The coherence parameters defined in [3, 4] are excellent
exemplars of such quantities.
3.1.1 Coherence Parameters: ?1 , ?2 , ?3
For an m ? n matrix L0 with rank r0 and SVD L0 = U0 ?0 V0T , some important properties can
be characterized by three coherence parameters, denoted as ?1 , ?2 and ?3 , respectively. The first
coherence parameter, 1 ? ?1 (L0 ) ? m, which characterizes the column space identified by U0 , is
defined as
m
?1 (L0 ) =
max kU T ei k22 ,
(3.4)
r0 1?i?m 0
where ei denotes the ith standard basis. The second coherence parameter, 1 ? ?2 (L0 ) ? n, which
characterizes the row space identified by V0 , is defined as
?2 (L0 ) =
n
max kV T ej k22 .
r0 1?j?n 0
(3.5)
The third coherence parameter, 1 ? ?3 (L0 ) ? mn, which characterizes the joint space identified
by U0 V0T , is defined as
?3 (L0 ) =
mn
mn
(kU0 V0T k? )2 =
max(|hU0T ei , V0T ej i|)2 .
r0
r0 i,j
(3.6)
The analysis in RPCA [4] merges the above three parameters into a single one: ?(L0 ) =
max{?1 (L0 ), ?2 (L0 ), ?3 (L0 )}. As will be seen later, the behaviors of those three coherence parameters are different from each other, and hence it is more adequate to consider them individually.
3.1.2 ?2 -phenomenon
According to the analysis in [4], the success condition (regarding L0 ) of RPCA is
rank (L0 ) ?
cr n 2
,
?(L0 )(log n1 )2
(3.7)
where ?(L0 ) = max{?1 (L0 ), ?2 (L0 ), ?3 (L0 )} and cr > 0 is some numerical constant. Thus,
RPCA will be less successful when the coherence parameters are considerably larger. In this subsection, we shall show that the widely existing clustering structure can enlarge the coherence parameters
and, accordingly, downgrades the performance of RPCA.
Given the restriction that rank (L0 ) = r0 , the data points (i.e., column vectors of L0 ) are unnecessarily sampled from a r0 -dimensional subspace uniformly at random. A more realistic interpretation
is to consider the data points as samples from the union of k number of subspaces (i.e., clusters),
and the sum of those multiple subspaces together has a dimension r0 . That is to say, there are
multiple ?small? subspaces inside one r0 -dimensional ?large? subspace, as exemplified in Figure 1.
Whenever the low-rank matrix L0 is meanwhile exhibiting such clustering behaviors, the second
coherence parameter ?2 (L0 ) (and so ?3 (L0 )) will increase with the number of clusters underlying
L0 , as shown in Figure 2. When the coherence is heightening, (3.7) suggests that the performance
of RPCA will drop, as verified in Figure 2(d). Note here that the variation of ?3 is mainly due
to the variation of the row space, which is characterized by ?2 . We call the phenomena shown in
Figure 2(b)?(d) as the ??2 -phenomenon?. Readers can also refer to the full paper to see why the
second coherence parameter increases with the cluster number underlying L0 .
Interestingly, one may have noticed that ?1 is invariant to the variation of the clustering number, as
can be seen from Figure 2(a). This is because the clustering behavior of the data points can only
affect the row space, while ?1 is defined on the column space. Yet, if the row vectors of L0 also
own some clustering structure, ?1 could be large as well. Such kind of data can exist widely in text
documents and we leave this as future work.
4
(a)
?3
?
2
?1
40
2
20
1
1
2
5 10 20 50
#cluster
0
0.3
recover error
3
0.5
(d)
60
4
1
0
(c)
(b)
1.5
1
2
0
5 10 20 50
1
2
5 10 20 50
#cluster
#cluster
0.2
0.1
0
1
2
5 10 20 50
#cluster
Figure 2: Exploring the influence of the cluster number, using randomly generated matrices. The
size and rank of L0 are fixed to be 500 ? 500 and 100, respectively. The underlying cluster number
varies from 1 to 50. For the recovery experiments, S0 is fixed as a sparse matrix with 13% nonzero
entries. (a) The first coherence parameter ?1 (L0 ) vs cluster number. (b) ?2 (L0 ) vs cluster number.
(c) ?3 (L0 ) vs cluster number. (d) Recover error (produced by RPCA) vs cluster number. The
numbers shown in these figure are averaged from 100 random trials. The recover error is computed
? 0 ? L0 kF /kL0 kF , where L
? 0 denotes an estimate of L0 .
as kL
3.2 Avoiding ?2 by LRR
The ?2 -phenomenon implies that the second coherence parameter ?2 is related in nature to some
intrinsic structures of L0 and thus cannot be eschewed without using additional conditions. In the
following, we shall figure out under what conditions the second coherence parameter ?2 (and ?3 )
can be avoided such that LRR could well handle coherent data.
Main Result: We show that, when the dictionary A itself is low-rank, LRR is able to avoid ?2 .
Namely, the following theorem is proved without using ?2 . See the full paper for a detailed proof.
Theorem 1 (Noiseless) Let A ? Rm?d with SVD A = UA ?A VAT be a column-wisely unit-normed
(i.e., kAei k2 = 1, ?i) dictionary matrix which satisfies PUA (U0 ) = U0 (i.e., U0 is a subspace of
UA ). For any 0 < ? < 0.5 and some numerical constant ca > 1, if
rank (L0 ) ? rank (A) ?
? 2 n2
ca ?1 (A) log n1
and |?| ? (0.5 ? ?)mn,
(3.8)
?10
then
? with probability at least 1 ? n1 , the optimal solution to the LRR problem (1.3) with ? =
1/ n1 is unique and exact, in a sense that
Z ? = A+ L0
and S ? = S0 ,
where (Z ? , S ? ) is the optimal solution to (1.3).
It is worth noting that the restriction rank (L0 ) ? O(n2 / log n1 ) is looser than that of PRCA3 , which
requires rank (L0 ) ? O(n2 /(log n1 )2 ). The requirement of column-wisely unit-normed
? dictionary
(i.e., kAei k2 = 1, ?i) is purely for complying the parameter estimate of ? = 1/ n1 , which is
consistent with RPCA. The condition PUA (U0 ) = U0 , i.e., U0 is a subspace of UA , is indispensable
if we ask for exact recovery, because PUA (U0 ) = U0 is implied by the equality AZ ? = L0 . This
necessary condition, together with the low-rankness condition, provides an elementary criterion for
learning the dictionary matrix A in LRR. Figure 3 presents an example, which further confirms our
main result; that is, LRR is able to avoid ?2 as long as U0 ? UA and A is low-rank. It is also
worth noting that it is unnecessary for A to satisfy UA = U0 , and that LRR is actually tolerant to the
?errors? possibly existing in the dictionary.
The program (1.3) is designed for the case where the uncorrupted observations are noiseless. In
reality this assumption is often not true and all entries of X can be contaminated by a small amount
of noises, i.e., X = L0 + S0 + N , where N is a matrix of dense Gaussian noises. In this case, the
formula of LRR (1.3) need be modified to
min kZk? + ?kSk1 , s.t.
Z,S
3
kX ? AZ ? SkF ? ?,
(3.9)
In terms of exact recovery, O(n2 / log n1 ) is probably the ?finest? bound one could accomplish in theory.
5
AZ*
*
S
0.2
recover error
X
0.1
0
1 5
10 15 20
rank(A)
Figure 3: Exemplifying that LRR can void ?2 . In this experiment, L0 is a 200 ? 200 rank-1 matrix
with one column being 1 (i.e., a vector of all ones) and everything else being zero. Thus, ?1 (L0 ) = 1
and ?2 (L0 ) = 200. The dictionary is set as A = [1, W ], where W is a 200 ? p random Gaussian
matrix (with varying p). As long as rank (A) = p + 1 ? 10, LRR with ? = 0.08 can exactly recover
L0 from a grossly corrupted observation matrix X.
where ? is a parameter that measures the noise level of data. In the experiments of this paper,
we consistently set ? = 10?6 kXkF . In the presence of dense noises, the latent matrices, L0 and
S0 , cannot be exactly restored. Yet we have the following theorem to guarantee the near recovery
property of the solution produced by the program (3.9):
Theorem 2 (Noisy) Suppose kX ? L0 ? S0 kF ? ?. Let A ? Rm?d with SVD A = UA ?A VAT be a
column-wisely unit-normed dictionary matrix which satisfies PUA (U0 ) = U0 (i.e., U0 is a subspace
of UA ). For any 0 < ? < 0.35 and some numerical constant ca > 1, if
rank (L0 ) ? rank (A) ?
? 2 n2
ca ?1 (A) log n1
and |?| ? (0.35 ? ?)mn,
(3.10)
?
then with probability at least 1 ? n?10
, any solution (Z ??
, S ? ) to (3.9) with ? = 1/ n1?gives a near
1
recovery to (L0 , S0 ), in a sense that kAZ ? ? L0 kF ? 8 mn? and kS ? ? S0 kF ? (8 mn + 2)?.
3.3 An Unsupervised Algorithm for Matrix Recovery
To handle coherent (equivalently, less incoherent) data, Theorem 1 suggests that the dictionary matrix A should be low-rank and satisfy U0 ? UA . In certain supervised environment, this might not be
difficult as one could potentially use clear, well processed training data to construct the dictionary. In
an unsupervised environment, however, it will be challenging to identify a low-rank dictionary that
obeys U0 ? UA . Note that U0 ? UA can be viewed as supervision information (if A is low-rank).
In this paper, we will introduce a heuristic algorithm that can work distinctly better than RPCA in
an unsupervised environment. As can be seen from (3.7), RPCA is actually not brittle with respect
to coherent data (although its performance is depressed). Based on this observation, we propose
a simple algorithm, as summarized in Algorithm 1, to achieve a solid improvement over RPCA.
Our idea is straightforward: We first obtain an estimate of L0 by using RPCA and then utilize the
estimate to construct the dictionary matrix A in LRR. The post-processing steps (Step 2 and Step 3)
that slightly modify the solution of RPCA is to encourage well-conditioned dictionary, which is the
circumstance favoring LRR.
Whenever the recovery produced by RPCA is already exact, the claim in Theorem 1 gives that the
recovery produced by our Algorithm 1 is exact as well. That is to say, in terms of exactly recovering
L0 from a given X, the success probability of our Algorithm 1 is greater than or equal to that of
RPCA. From the computational perspective, Algorithm 1 does not really double the work of RPCA,
although there are two convex programs in our algorithm. In fact, according to our simulations,
usually the computational time of Algorithm 1 is merely about 1.2 times as much as RPCA. The
reason is that, as has been explored by [13], the complexity of solving the LRR problem (1.3) is
O(n2 rA ) (assuming m = n), which is much lower than that of RPCA (which requires O(n3 ))
provided that the obtained dictionary matrix A is fairly low-rank (i.e., rA is small).
One may have noticed that the procedure of Algorithm 1 could be made iterative, i.e., one can
? ? as a new estimate of L0 and use it to further update the dictionary matrix A, and so
consider AZ
on. Nevertheless, we empirically find that such an iterative procedure often converges within two
iterations. Hence, for the sake of simplicity, we do not consider iterative strategies in this paper.
6
Algorithm 1 Matrix Recovery
input: Observed data matrix X ? Rm?n .
adjustable parameter: ?.
? 0 by optimizing the RPCA problem (1.2) with ? = 1/?n1 .
1. Solve for L
? 0 by
2. Estimate the rank of L
r?0 = #{i : ?i > 10?3 ?1 },
? 0.
where ?1 , ?2 , ? ? ? , ?n2 are the singular values of L
? 0 by using the rank-?
? 0 . That is,
3. Form L
r0 approximation of L
? 0 = arg min kL ? L
? 0 k2 , s.t. rank (L) ? r?0 ,
L
F
L
which is solved by SVD.
? 0 by normalizing the column vectors of L
? 0:
4. Construct a dictionary A? from L
? :,i =
[A]
? 0 ]:,i
[L
, i = 1, ? ? ? , n,
? 0 ]:,i k2
k[L
where [?]:,i denotes the ith column of a matrix.
?
5. Solve for Z ? by optimizing the LRR problem (1.3) with A = A? and ? = 1/ n1 .
? ?.
output: AZ
4 Experiments
4.1 Results on Randomly Generated Matrices
We first verify the effectiveness of our Algorithm 1 on randomly generated matrices. We generate
a collection of 200 ? 1000 data matrices according to the model of X = P?? (L0 ) + P? (S0 ):
? is a support set chosen at random; L0 is created by sampling 200 data points from each of 5
randomly generated subspaces; S0 consists of random values from Bernoulli ?1. The dimension of
each subspace varies from 1 to 20 with step size 1, and thus the rank of L0 varies from 5 to 100 with
step size 5. The fraction |?|/(mn) varies from 2.5% to 50% with step size 2.5%. For each pair of
rank and support size (r0 , |?|), we run 10 trials, resulting in a total of 4000 (20 ? 20 ? 10) trials.
RPCA
Algorithm 1
32
22
12
2
0.1 0.2 0.3 0.4 0.5
rank(L0)/n2
42
corruption (%)
corruption (%)
corruption (%)
50
42
32
22
12
2
0.1 0.2 0.3 0.4 0.5
rank(L0)/n2
40
RPCA
Algorithm 1
30
20
10
0.1 0.2 0.3 0.4 0.5
rank(L0)/n2
Figure 4:?Algorithm 1 vs RPCA for the task of recovering randomly generated matrices, both using
? = 1/ n1 . A curve shown in the third subfigure is the boundary for a method to be successful
? The recovery is successful for any pair (r0 /n2 , |?|/(mn)) that locates below the curve. Here, a
? 0 ? L0 kF < 0.05kL0kF , where L
? 0 denotes an estimate of L0 .
success means kL
?
Figure 4 compares our Algorithm 1 to RPCA, both using ? = 1/ n1 . It can be seen that, using the
learned dictionary matrix, Algorithm 1 works distinctly better than RPCA. In fact, the success area
(i.e., the area of the white region) of our algorithm is 47% wider than that of RPCA! We should also
mention that it is possible for RPCA to be exactly successful on coherent (or less incoherent) data,
provided that the rank of L0 is low enough and/or S0 is sparse enough. Our algorithm in general
improves RPCA when L0 is moderately low-rank and/or S0 is moderately sparse.
7
4.2 Results on Corrupted Motion Sequences
We now present our experiment with 11 additional sequences attached to the Hopkins155 [21]
database. In those sequences, about 10% of the entries in the data matrix of trajectories are unobserved (i.e., missed) due to vision occlusion. We replace each missed entry with a number from
Bernoulli ?1, resulting in a collection of corrupted trajectory matrices for evaluating the effectiveness of matrix recovery algorithms. We perform subspace clustering on both the corrupted trajectory
matrices and the recovered versions, and use the clustering error rates produced by existing subspace
clustering methods as the evaluation metrics. We consider three state-of-the-art subspace clustering
methods: Shape Interaction Matrix (SIM) [5], Low-Rank Representation with A = X [14] (which
is referred to as ?LRRx?) and Sparse Subspace Clustering (SSC) [6].
Table 1: Clustering error rates (%) on 11 corrupted motion sequences.
SIM
RPCA + SIM
Algorithm 1 + SIM
LRRx
RPCA + LRRx
Algorithm 1 + LRRx
SSC
RPCA + SSC
Algorithm 1 + SSC
Mean
29.19
14.82
8.74
21.38
10.70
7.09
22.81
9.50
5.74
Median
27.77
8.38
3.09
22.00
3.05
3.06
20.78
2.13
1.85
Maximum
45.82
45.78
42.61
56.96
46.25
32.33
58.24
50.32
27.84
Minimum
12.45
0.97
0.23
0.58
0.20
0.22
1.55
0.61
0.20
Std.
11.74
16.23
12.95
17.10
15.63
10.59
18.46
16.17
8.52
Time (sec.)
0.07
9.96
11.64
1.80
10.75
12.11
3.18
12.51
13.11
Table 1 shows the error rates of various algorithms. Without the preprocessing of matrix recovery,
all the subspace clustering methods fail to accurately categorize the trajectories of motion objects,
producing error rates higher than 20%. This illustrates that it is important for motion segmentation
to correct
?the gross corruptions possibly existing in the data matrix of trajectories. By using RPCA
(? = 1/ n1 ) to correct the corruptions, the clustering performances of all considered methods are
improved dramatically. For example, the error rate of ?
SSC is reduced from 22.81% to 9.50%. By
choosing an appropriate dictionary for LRR (? = 1/ n1 ), the error rates can be reduced again,
from 9.50% to 5.74%, which is a 40% relative improvement. These results verify the effectiveness
of our dictionary learning strategy in realistic environments.
5 Conclusion and Future Work
We have studied the problem of disentangling the low-rank and sparse components in a given data
matrix. Whenever the low-rank component exhibits clustering structures, the state-of-the-art RPCA
method could be less successful. This is because RPCA prefers incoherent data, which however may
be inconsistent with data in the real world. When the number of clusters becomes large, the second
and third coherence parameters enlarge and hence the performance of RPCA could be depressed. We
have showed that the challenges arising from coherent (equivalently, less incoherent) data could be
effectively alleviated by learning a suitable dictionary under the LRR framework. Namely, when the
dictionary matrix is low-rank and contains information about the ground truth matrix, LRR can be
immune to the coherence parameters that increase with the underlying cluster number. Furthermore,
we have established a practical algorithm that outperforms RPCA in our extensive experiments.
The problem of recovering coherent data essentially concerns the robustness issues of the Generalized PCA (GPCA) [22] problem. Although the classic GPCA problem has been explored for several
decades, robust GPCA is new and has not been well studied. The approach proposed in this paper is in a sense preliminary, and it is possible to develop other effective methods for learning the
dictionary matrix in LRR and for handling coherent data. We leave these as future work.
Acknowledgement
Guangcan Liu was a Postdoctoral Researcher supported by NSF-DMS0808864, NSF-SES1131848,
NSF-EAGER1249316, AFOSR-FA9550-13-1-0137, and ONR-N00014-13-1-0764. Ping Li is also
partially supported by NSF-III1360971 and NSF-BIGDATA1419210.
8
References
[1] Liliana Borcea, Thomas Callaghan, and George Papanicolaou. Synthetic aperture radar imaging and
motion estimation via robust principle component analysis. Arxiv, 2012.
[2] Emmanuel Cand`es and Yaniv Plan. Matrix completion with noise. In IEEE Proceeding, volume 98, pages
925?936, 2010.
[3] Emmanuel Cand`es and Benjamin Recht. Exact matrix completion via convex optimization. Foundations
of Computational Mathematics, 9(6):717?772, 2009.
[4] Emmanuel J. Cand`es, Xiaodong Li, Yi Ma, and John Wright. Robust principal component analysis?
Journal of the ACM, 58(3):1?37, 2011.
[5] Joao Costeira and Takeo Kanade. A multibody factorization method for independently moving objects.
International Journal of Computer Vision, 29(3):159?179, 1998.
[6] E. Elhamifar and R. Vidal. Sparse subspace clustering. In IEEE Conference on Computer Vision and
Pattern Recognition, volume 2, pages 2790?2797, 2009.
[7] M. Fazel. Matrix rank minimization with applications. PhD thesis, 2002.
[8] Martin Fischler and Robert Bolles. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Communications of the ACM, 24(6):381?395,
1981.
[9] R. Gnanadesikan and J. R. Kettenring. Robust estimates, residuals, and outlier detection with multiresponse data. Biometrics, 28(1):81?124, 1972.
[10] D. Gross. Recovering low-rank matrices from few coefficients in any basis. IEEE Transactions on
Information Theory, 57(3):1548?1566, 2011.
[11] Qifa Ke and Takeo Kanade. Robust l1 norm factorization in the presence of outliers and missing data
by alternative convex programming. In IEEE Conference on Computer Vision and Pattern Recognition,
pages 739?746, 2005.
[12] Fernando De la Torre and Michael J. Black. A framework for robust subspace learning. International
Journal of Computer Vision, 54(1-3):117?142, 2003.
[13] Guangcan Liu, Zhouchen Lin, Shuicheng Yan, Ju Sun, Yong Yu, and Yi Ma. Robust recovery of subspace
structures by low-rank representation. IEEE Transactions on Pattern Analysis and Machine Intelligence,
35(1):171?184, 2013.
[14] Guangcan Liu, Zhouchen Lin, and Yong Yu. Robust subspace segmentation by low-rank representation.
In International Conference on Machine Learning, pages 663?670, 2010.
[15] Guangcan Liu, Huan Xu, and Shuicheng Yan. Exact subspace segmentation and outlier detection by
low-rank representation. Journal of Machine Learning Research - Proceedings Track, 22:703?711, 2012.
[16] Rahul Mazumder, Trevor Hastie, and Robert Tibshirani. Spectral regularization algorithms for learning
large incomplete matrices. Journal of Machine Learning Research, 11:2287?2322, 2010.
[17] Ricardo Otazo, Emmanuel Cand`es, and Daniel K. Sodickson. Low-rank and sparse matrix decomposition
for accelerated dynamic mri with separation of background and dynamic components. Arxiv, 2012.
[18] YiGang Peng, Arvind Ganesh, John Wright, Wenli Xu, and Yi Ma. Rasl: Robust alignment by sparse
and low-rank decomposition for linearly correlated images. IEEE Transactions on Pattern Analysis and
Machine Intelligence, 34(11):2233?2246, 2012.
[19] Mahdi Soltanolkotabi, Ehsan Elhamifar, and Emmanuel Candes.
Robust subspace clustering.
arXiv:1301.2603, 2013.
[20] Nathan Srebro and Tommi Jaakkola. Generalization error bounds for collaborative prediction with lowrank matrices. In Neural Information Processing Systems, pages 5?27, 2005.
[21] Roberto Tron and Rene Vidal. A benchmark for the comparison of 3-d motion segmentation algorithms.
In IEEE Conference on Computer Vision and Pattern Recognition, pages 1?8, 2007.
[22] Rene Vidal, Yi Ma, and S. Sastry. Generalized Principal Component Analysis. Springer Verlag, 2012.
[23] Markus Weimer, Alexandros Karatzoglou, Quoc V. Le, and Alex J. Smola. Cofi rank - maximum margin
matrix factorization for collaborative ranking. In Neural Information Processing Systems, 2007.
[24] Huan Xu, Constantine Caramanis, and Shie Mannor. Outlier-robust pca: The high-dimensional case.
IEEE Transactions on Information Theory, 59(1):546?572, 2013.
[25] Huan Xu, Constantine Caramanis, and Sujay Sanghavi. Robust pca via outlier pursuit. In Neural Information Processing Systems, 2010.
[26] Zhengdong Zhang, Arvind Ganesh, Xiao Liang, and Yi Ma. Tilt: Transform invariant low-rank textures.
International Journal of Computer Vision, 99(1):1?24, 2012.
9
| 5334 |@word mild:1 trial:3 version:2 complying:1 mri:1 norm:10 km:6 confirms:1 simulation:1 shuicheng:2 decomposition:3 mention:1 iii1360971:1 solid:1 klk:1 reduction:1 liu:5 contains:1 daniel:1 document:1 interestingly:2 outperforms:2 existing:8 ksk1:3 recovered:2 yet:3 finest:1 john:2 takeo:2 numerical:3 realistic:2 shape:1 drop:2 designed:1 update:1 v:5 intelligence:2 accordingly:3 ith:2 fa9550:1 alexandros:1 provides:4 gpca:3 mannor:1 location:1 preference:1 firstly:1 zhang:1 constructed:1 prove:3 consists:1 fitting:1 inside:1 introduce:1 peng:1 ra:2 indeed:3 behavior:3 cand:4 grade:2 decomposed:1 equipped:1 cardinality:1 increasing:1 becomes:2 provided:3 ua:10 underlying:7 notation:3 joao:1 multibody:1 biostatistics:2 what:1 kind:1 unobserved:1 nj:2 guarantee:1 exactly:5 um:4 rm:8 k2:5 demonstrates:1 unit:3 producing:1 arguably:1 understood:1 modify:1 consequence:1 subscript:2 approximately:1 abuse:2 might:2 black:1 studied:3 k:1 suggests:2 challenging:1 factorization:5 lrr:30 averaged:1 obeys:1 fazel:1 practical:4 unique:1 union:1 procedure:2 area:3 empirical:1 yan:2 convenient:1 alleviated:2 projection:2 refers:1 cannot:5 onto:2 operator:2 influence:1 restriction:3 demonstrated:2 missing:2 go:1 straightforward:2 normed:3 convex:5 independently:1 ke:1 simplicity:1 recovery:22 assigns:1 insight:1 nuclear:2 spanned:1 classic:1 handle:4 variation:3 resp:3 suppose:3 massive:1 exact:9 user:3 qifa:1 us:1 programming:1 recognition:3 std:1 database:1 observed:1 solved:1 capture:2 region:1 sun:1 gross:5 benjamin:1 environment:7 complexity:1 moderately:2 fischler:1 dynamic:2 radar:2 raise:1 solving:1 purely:1 upon:1 basis:2 joint:1 various:3 caramanis:2 effective:2 tell:1 choosing:1 quite:1 heuristic:1 widely:6 larger:1 solve:2 say:2 statistic:2 transform:1 itself:3 noisy:1 superscript:1 seemingly:1 sequence:5 rr:1 propose:2 interaction:1 achieve:1 frobenius:1 kv:1 az:8 cluster:21 requirement:1 double:1 yaniv:1 produce:1 perfect:2 leave:2 converges:1 object:2 help:2 wider:1 develop:2 completion:2 exemplar:1 ij:4 lowrank:2 sim:4 recovering:6 indicate:1 implies:1 exhibiting:2 tommi:1 greek:1 correct:2 torre:1 subsequently:1 exploration:2 karatzoglou:1 everything:1 education:1 assign:1 generalization:1 really:1 preliminary:1 elementary:3 mathematically:1 strictly:2 exploring:1 considered:1 ground:1 normal:1 wright:2 claim:2 reserve:1 dictionary:32 estimation:1 rpca:49 gnanadesikan:1 individually:1 largest:1 tool:2 minimization:1 always:1 gaussian:2 modified:1 rather:1 avoid:5 ej:2 cr:2 varying:1 jaakkola:1 l0:77 kaz:1 improvement:2 consistently:1 rank:65 bernoulli:2 mainly:1 cartography:1 defend:1 sense:4 favoring:1 arg:1 among:1 issue:1 denoted:5 resonance:1 raised:2 special:1 fairly:1 art:2 plan:1 equal:2 construct:5 enlarge:2 sampling:1 unnecessarily:1 yu:2 unsupervised:5 future:3 contaminated:1 sanghavi:1 few:1 modern:1 randomly:7 maxj:1 skinny:1 occlusion:1 n1:17 detection:2 investigate:2 evaluation:1 multiresponse:1 alignment:1 nowadays:1 encourage:1 necessary:3 culture:1 huan:3 orthogonal:1 skf:1 biometrics:1 incomplete:1 desired:1 subfigure:1 column:14 kxkf:1 entry:9 successful:5 byp:1 varies:4 corrupted:8 accomplish:1 considerably:1 synthetic:1 recht:1 ju:1 international:4 vm:4 michael:1 together:2 thesis:1 squared:1 eschewed:1 again:1 possibly:2 ssc:5 ricardo:1 li:3 de:1 summarized:1 sec:1 coefficient:1 configured:1 satisfy:2 ranking:1 later:1 root:1 kl0:1 sup:1 characterizes:3 recover:6 hopkins155:1 candes:1 guangcan:5 contribution:1 collaborative:2 square:1 reserved:2 characteristic:1 identify:1 zhengdong:1 accurately:1 produced:7 trajectory:5 worth:2 researcher:1 corruption:6 ping:2 whenever:5 infection:1 trevor:1 grossly:2 naturally:1 elegance:1 proof:2 sampled:1 proved:2 intrinsically:2 ask:1 subsection:1 dimensionality:1 ubiquitous:2 improves:1 segmentation:4 actually:2 back:1 higher:1 supervised:2 costeira:1 specify:1 improved:1 rahul:1 borcea:1 furthermore:2 smola:1 correlation:1 replacing:1 ei:3 ganesh:2 reveal:1 grows:1 xiaodong:1 usa:2 effect:1 verify:3 k22:2 true:1 hence:3 equality:1 regularization:1 nonzero:4 moore:1 white:1 criterion:2 generalized:2 complete:1 bolles:1 tron:1 motion:8 l1:1 image:3 meaning:1 recently:1 physical:3 qp:1 empirically:1 attached:1 tilt:1 volume:2 interpretation:1 measurement:1 refer:1 rene:2 sujay:1 sastry:1 mathematics:1 similarly:1 zhouchen:2 depressed:2 soltanolkotabi:1 immune:2 moving:1 longer:1 supervision:1 v0:1 etc:2 pu:2 own:2 showed:1 perspective:1 optimizing:2 constantine:2 termed:1 indispensable:1 certain:1 n00014:1 verlag:1 onr:1 arbitrarily:2 success:4 yi:5 uncorrupted:1 seen:5 minimum:1 additional:4 greater:1 george:1 r0:12 determine:1 paradigm:1 fernando:1 u0:19 resolving:2 full:4 multiple:2 characterized:2 arvind:2 long:4 lin:2 prescription:1 post:1 locates:1 vat:2 prediction:1 scalable:2 basic:2 variant:2 vision:8 metric:1 essentially:3 rutgers:4 arxiv:4 noiseless:2 represent:1 circumstance:1 iteration:1 achieved:1 background:2 void:1 singular:8 else:1 median:1 crucial:2 appropriately:1 extra:7 probably:3 elegant:1 shie:1 inconsistent:1 spirit:1 effectiveness:3 call:1 near:2 presence:3 noting:2 easy:1 enough:2 automated:1 variety:1 affect:1 hastie:1 identified:3 idea:3 regarding:2 papanicolaou:1 six:1 pca:6 ultimate:2 render:1 prefers:1 adequate:1 dramatically:1 generally:1 useful:1 detailed:1 clear:1 amount:2 extensively:1 processed:1 reduced:2 generate:1 exist:2 wisely:3 nsf:5 arising:1 track:1 tibshirani:1 shall:7 nevertheless:1 drawn:1 capital:1 verified:1 utilize:1 kettenring:1 imaging:3 merely:1 fraction:1 sum:3 run:1 letter:4 powerful:2 throughout:1 reader:1 looser:1 utilizes:1 missed:2 separation:1 coherence:32 bound:2 precisely:1 constraint:1 alex:1 exemplifying:2 n3:1 sake:1 yong:2 markus:1 nathan:1 robustifying:1 min:5 cofi:1 martin:1 department:4 according:3 piscataway:2 slightly:1 gcliu:1 quoc:1 yigang:1 bigdata1419210:1 outlier:6 explained:1 invariant:2 discus:1 fail:1 know:1 pursuit:2 vidal:3 obey:1 generic:1 appropriate:1 magnetic:1 spectral:1 alternative:1 robustness:1 thomas:1 personality:1 denotes:7 clustering:23 include:1 k1:2 emmanuel:5 establish:1 classical:1 implied:1 noticed:2 pingli:1 quantity:3 already:1 restored:1 degrades:1 strategy:2 exhibit:2 subspace:25 consensus:1 reason:1 assuming:1 besides:1 providing:2 equivalently:2 difficult:1 disentangling:1 liang:1 robert:2 potentially:2 relate:1 proper:2 unknown:2 adjustable:1 perform:1 observation:5 discarded:1 benchmark:1 communication:1 locate:3 rn:1 arbitrary:1 complement:1 namely:2 pair:2 kl:3 extensive:2 coherent:19 merges:1 learned:1 established:2 beyond:3 able:2 usually:1 exemplified:1 below:1 ku0:1 pattern:5 regime:2 challenge:2 program:5 built:1 max:6 suitable:1 restore:2 residual:1 mn:9 movie:2 created:1 incoherent:7 perfection:1 roberto:1 text:1 literature:2 acknowledgement:1 kf:7 relative:1 afosr:1 fully:2 brittle:2 srebro:1 foundation:1 sufficient:1 consistent:1 s0:17 xiao:1 principle:3 exciting:1 rasl:1 row:6 summary:3 supported:4 understand:1 wide:1 fall:1 v0t:4 sparse:10 distinctly:2 kzk:2 dimension:3 curve:2 pua:4 boundary:1 evaluating:1 world:1 ignores:2 made:1 collection:2 preprocessing:1 avoided:1 far:1 transaction:4 aperture:1 keep:2 dealing:1 pseudoinverse:1 tolerant:1 unnecessary:2 assumed:1 postdoctoral:1 latent:1 iterative:3 decade:2 why:2 reality:3 table:2 kanade:2 nature:3 promising:1 robust:13 ku:1 ca:4 correlated:1 obtaining:1 mazumder:1 ehsan:1 excellent:2 meanwhile:2 main:3 dense:2 linearly:1 weimer:1 noise:5 arise:1 n2:12 xu:4 referred:1 fashion:1 pv:1 mahdi:1 third:3 theorem:6 formula:1 symbol:2 explored:4 maxi:1 normalizing:1 concern:1 intrinsic:1 effectively:1 texture:1 callaghan:1 magnitude:1 phd:1 conditioned:1 illustrates:1 elhamifar:2 kx:2 rankness:6 margin:1 simply:1 penrose:1 partially:1 springer:1 gender:1 truth:1 satisfies:2 acm:2 ma:5 identity:2 viewed:1 replace:1 typical:1 except:1 uniformly:2 principal:4 called:1 total:1 experimental:1 svd:8 e:4 la:1 support:4 categorize:1 accelerated:1 phenomenon:4 avoiding:1 handling:1 |
4,788 | 5,335 | Scalable Methods for Nonnegative Matrix
Factorizations of Near-separable Tall-and-skinny
Matrices
Jason D. Lee
ICME
Stanford University
Stanford, CA
[email protected]
Austin R. Benson
ICME
Stanford University
Stanford, CA
[email protected]
Bartek Rajwa
Bindley Biosciences Center
Purdue University
West Lafeyette, IN
[email protected]
David F. Gleich
Computer Science Department
Purdue University
West Lafeyette, IN
[email protected]
Abstract
Numerous algorithms are used for nonnegative matrix factorization under the assumption that the matrix is nearly separable. In this paper, we show how to
make these algorithms scalable for data matrices that have many more rows than
columns, so-called ?tall-and-skinny matrices.? One key component to these improved methods is an orthogonal matrix transformation that preserves the separability of the NMF problem. Our final methods need to read the data matrix only
once and are suitable for streaming, multi-core, and MapReduce architectures.
We demonstrate the efficacy of these algorithms on terabyte-sized matrices from
scientific computing and bioinformatics.
1
Nonnegative matrix factorizations at scale
A nonnegative matrix factorization (NMF) for an m ? n matrix X with real-valued, nonnegative
entries is
X = WH
(1)
where W is m ? r, H is r ? n, r < min(m, n), and both factors have nonnegative entries. While
there are already standard dimension reduction techniques for general matrices such as the singular
value decomposition, the advantage of NMF is in interpretability of the data. A common example is
facial image decomposition [17]. If the columns of X are pixels of a facial image, the columns of W
may be facial features such as eyes or ears, and the coefficients in H represent the intensity of these
features. For this reason, among a host of other reasons, NMF is used in a broad range of applications
including graph clustering [21], protein sequence motif discovery [20], and hyperspectral unmixing
[18].
An important property of matrices in these applications and other massive scientific data sets is that
they have many more rows than columns (m n). For example, this matrix structure is common
in big data applications with hundreds of millions of samples and a small set of features?see,
e.g., Section 4.2 for a bioinformatics application where the data matrix has 1.6 billion rows and 25
columns. We call matrices with many more rows than columns tall-and-skinny. The number of
columns of these matrices is small, so there is no problem storing or manipulating them. Our use
1
of NMF is then to uncover the hidden structure in the data rather than for dimension reduction or
compression.
In this paper, we present scalable and computationally efficient NMF algorithms for tall-and-skinny
matrices as prior work has not taken advantage of this structure for large-scale factorizations. The
advantages of our method are: we preserve the geometry of the problem, we only read the data
matrix once, and we can test several different nonnegative ranks (r) with negligible cost. Furthermore, we show that these methods can be implemented in parallel (Section 3) to handle large data
sets. In Section 2.3, we present a new dimension reduction technique using orthogonal transformations. These transformations are particularly effective for tall-and-skinny matrices and lead to
algorithms that only need to read the data matrix once. We compare this method with a Gaussian
projection technique from the hyperspectral unmixing community [5, 7]. We test our algorithms on
data sets from two scientific applications, heat transfer simulations and flow cytometry, in Section 4.
Our new dimension reduction technique outperforms Gaussian projections on these data sets. In
the remainder of the introduction, we review the state of the art for computing non-negative matrix
factorizations.
1.1
Separable NMF
We first turn to the issue of how to practically compute the factorization in Equation (1). Unfortunately, for a fixed non-negative rank r, finding the factors W and H for which the residual kX ? WHk
is minimized is NP-complete [26]. To make the problem tractable, we make assumptions about the
data. In particular, we require a separability condition on the matrix. A nonnegative matrix X is
separable if
X = X(:, K)H,
where K is an index set with |K| = r and X(:, K) is Matlab notation for the matrix X restricted to
the columns indexed by K. Since the coefficients of H are nonnegative, all columns of X live in
the conical hull of the ?extreme? columns indexed by K. The idea of separability was developed
by Donoho and Stodden [15], and recent work has produced tractable NMF algorithms by assuming
that X almost satisfies a separability condition [3, 6].
A matrix X is noisy r-separable or near-separable if X = X(:, K)H + N, where N is a noise matrix
whose entries are small. Near-separability means that all data points approximately live in the
conical hull of the extreme columns. The algorithms for near-separable NMF are typically based on
convex geometry (see Section 2.1) and can be described by the same two-step approach:
1. Determine the extreme columns, indexed by K, and let W = X(:, K).
2. Solve H = arg minY?Rr?n
kX ? WYk.
+
The bulk of the literature is focused on the first step. In Section 3, we show how to implement both
steps in a single pass over the data and provide the details of a MapReduce implementation. We
note that separability (or near-separability) is a severe and restrictive assumption. The tradeoff is
that our algorithms are extremely scalable and provably correct under this assumption. In big data
applications, scalability is at a premium, and this provides some justification for using separability
as a tool for exploratory data analysis. Furthermore, our experiments on real scientific data sets in
Section 4 under the separability assumption lead to new insights.
1.2
Alternative NMF algorithms and related work
There are several approaches to solving Equation (1) that do not assume the separability condition. These algorithms typically employ block coordinate descent, optimizing over W and H while
keeping one factor fixed. Examples include the seminal work by Lee and Seung [23], alternating
least squares [10], and fast projection-based least squares [19]. Some of these methods are used in
MapReduce architectures at scale [24].
Alternating methods require updating the entire factor W or H after each optimization step. When
one of the factors is large, repeated updates can be prohibitively expensive. The problem is exacerbated in Hadoop MapReduce, where intermediate results are written to disk. In addition, alternating
methods can take an intolerable number of iterations to converge. Regardless of the approach or
computing platform, the algorithms are too slow when the matrices cannot fit in main memory In
2
contrast, we show in Sections 2 and 3 that the separability assumption leads to algorithms that do
not require updates to large matrices. This approach is scalable for large tall-and-skinny matrices in
big data problems.
2
Algorithms and dimension reduction for near-separable NMF
There are several popular algorithms for near-separable NMF, and they are motivated by convex
geometry. The goal of this section is to show that when X is tall-and-skinny, we can apply dimension
reduction techniques so that established algorithms can execute on n ? n matrices, rather than the
original m ? n. Our new dimension reduction technique in Section 2.3 is also motivated by convex
geometry. In Section 3, we leverage the dimension reduction into scalable algorithms.
2.1
Geometric algorithms
There are two geometric strategies typically employed for near-separable
P NMF. The first deals with
conical hulls. A cone C ? Rm is a non-empty convex set with C = { i ?i xi | ?i ? R+ , xi ? Rm }.
The xi are generating vectors. In separable NMF,
X = X(:, K)H
implies that all columns of X lie in the cone generated by the columns indexed by K. For any k ? K,
{?X(:, k) | ? ? R+ } is an extreme ray of this cone, In other words, the set of columns indexed by
K are the set of extreme rays of the cone. The goal of the XRAY algorithm [22] is to find these
extreme rays (i.e., to find K). In particular, the greedy variant of XRAY selects the maximum
column norm arg max j kRT X(:, j)k2 /kX(:, j)k2 , where R is a residual matrix that gets updated with
each new extreme column.
The second approach deals with convex hulls, where the columns of X are `1 -normalized. If D is a
diagonal matrix with Dii = kX(:, i)k1 and X is separable, then
?
XD?1 = X(:, K)D(K, K)?1 D(K, K)HD?1 = (XD?1 )(:, K)H.
Thus, XD?1 is also separable (in fact, this holds for any nonsingular diagonal matrix D). Since the
columns are `1 -normalized, the columns of H? have non-negative entries and sum to one. In other
words, all columns of XD?1 are in the convex hull of the columns indexed by K. The problem of
determining K is reduced to finding the extreme points of a convex hull. Popular approaches in the
context of NMF include the Successive Projection Algorithm (SPA, [2]) and its generalization [16].
Another alternative, based on linear programming, is Hott Topixx [6]. Other geometric approaches
had good heuristic performance [9, 25] before the more recent theoretical work. As an example of
the particulars of one such method, SPA, which we will use in Section 4, finds extreme points by
computing arg max j kR(:, j)k22 , where R is a residual matrix related to the data matrix X.
In any algorithm, we call the columns indexed by K extreme columns. The next two subsections are
devoted to dimension reduction techniques for finding the extreme columns in the case when X is
tall-and-skinny.
2.2
Gaussian projection
A common dimension reduction technique is random Gaussian projections, and the idea has been
used in hyperspectral unmixing problems [5]. In the hyperspectral unmixing literature, the separability is referred to as the pure-pixel assumption, and the random projections are motivated by
convex geometry [7]. In particular, given a matrix G ? Rm?k with Gaussian i.i.d. entries, the extreme
columns of X are taken as the extreme columns of GT X, which is of dimension k ? n. Recent work
shows that when X is nearly r-separable and k = O(r log r), then all of the extreme columns are
found with high probability [13].
2.3
Orthogonal transformations
Our new alternative dimension reduction technique is also motivated by convex geometry. Consider
a cone C ? Rm and a nonsingular m ? m matrix M. It is easily shown that x is an extreme ray of C
3
if and only if Mx is an extreme ray of MC = {Mz | z ? C}. Similarly, for any convex set, invertible
transformations preserve extreme points.
We take advantage of these facts by applying specific orthogonal transformations as the nonsingular
? T be the full QR factorization and singular value decomposition
matrix M. Let X = QR? and X = U ?V
(SVD) of X, so that Q and U are m ? m orthogonal (and hence nonsingular) matrices. Then
!
!
R
?V T
T
T
, U X=
Q X=
,
0
0
where R and ? are the top n ? n blocks of R? and ?? and 0 is an (m ? n) ? n matrix of zeroes. The zero
rows provide no information on which columns of QT X or U T X are extreme rays or extreme points.
Thus, we can restrict ourselves to finding the extreme columns of R and ?V T . These matrices are
n ? n, and we have significantly reduced the dimension of the problem. In fact, if X = X(:, K)H is a
separable representation, we immediately have separated representations for R and ?V T :
R = R(:, K)H,
?V T = ?V T (:, K)H.
We note that, although any invertible transformation preserves extreme columns, many transformations will destroy the geometric structure of the data. However, orthogonal transformations are
either rotations or reflections, and they preserve the data?s geometry. Also, although QT and U T are
m ? m, we will only apply them implicitly (see Section 3.1), i.e., these matrices are never formed or
computed.
This dimension reduction technique is exact when X is r-separable, and the results will be the same
for orthogonal transformations QT and U T . This is a consequence of the transformed data having
the same separability as the original data. The SPA and XRAY algorithms briefly described in
Section 2.1 only depend on computing column 2-norms, which are preserved under orthogonal
transformations. For these algorithms, applying QT or U T preserves the column 2-norms of the data,
and the selected extreme columns are the same. However, other NMF algorithms do not possess this
invariance. For this reason, we present both of the orthogonal transformations.
Finally, we highlight an important benefit of this dimension reduction technique. In many applications, the data is noisy and the separation rank (r in Equation (1)) is not known a priori. In
Section 2.4, we show that the H factor can be computed in the small dimension. Thus, it is viable to
try several different values of the separation rank and pick the best one. This idea is extremely useful
for the applications presented in Section 4, where we do not have a good estimate of the separability
of the data.
2.4
Computing H
Selecting the extreme columns indexed by K completes one half of the NMF factorization in Equation (1). How do we compute H? We want H = arg minY?Rr?n
kX ? X(:, K)Yk2 for some norm.
+
Choosing the Frobenius norm results in a set of n nonnegative least squares (NNLS) problems:
H(:, i) = arg minr kX(:, K)y ? X(:, i)k22 ,
y?R+
i = 1, . . . , n.
? Then H(:, i) is computed by finding y ? Rr+ that
Let X = QR? with R the upper n ? n block of R.
minimizes
kX(:, K)y ? X(:, i)k22 = kQT (X(:, K)y ? X(:, i)) k22 = kR(:, K)y ? R(:, i)k22
Thus, we can solve the NNLS problem with matrices of size n ? n. After computing just the small
R factor from the QR factorization, we can compute the entire nonnegative matrix factorization by
working with matrices of size n ? n. Analogous results hold for the SVD, where we replace Q by
U, the left singular vectors. In Section 3, we show that these computations are simple and scalable.
Since m n, computations on O(n2 ) data are fast, even in serial. Finally, note that we can also
compute the residual in this reduced space, i.e.:
min kX(:, K)y ? X(:, i)k22 = minn kR(:, K)y ? R(:, i)k22 .
y?Rn+
y?R+
This simple fact is significant in practice. When there are several candidate sets of extreme columns
K, the residual error for each set can be computed quickly. In Section 4, we compute many residual
errors for different sets K in order to choose an optimal separation rank.
4
We have now shown how to use dimension reduction techniques for tall-and-skinny matrix data in
near-separable NMF algorithms. Following the same strategy as many NMF algorithms, we first
compute extreme columns and then solve for the coefficient matrix H. Fortunately, once the upfront
cost of the orthogonal transformation is complete, both steps can be computed using O(n2 ) data.
3
Implementation
Remarkably, when the matrix is tall-and-skinny, we only need to read the data matrix once. The
reads can be performed in parallel, and computing platforms such as MapReduce, Spark, distributed
memory MPI, and GPUs can all achieve optimal parallel communication. For our implementation,
we use Hadoop MapReduce for convenience.1 While all of the algorithms use sophisticated computation, these routines are only ever invoked with matrices of size n ? n. Furthermore, the local
memory requirements of these algorithms are only O(n2 ). Thus, we get extremely scalable implementations. We note that, using MapReduce, computing GT X for the Gaussian projection technique
is a simple variation of standard methods to compute X T X [4].
3.1
TSQR and R-SVD
The thin QR factorization of an m ? n real-valued matrix X with m > n is X = QR where Q is an
m ? n orthogonal matrix and R is an n ? n upper triangular matrix. This is precisely the factorization
we need in Section 2. For our purposes, QT is applied implicitly, and we only need to compute
R. When m n, communication-optimal algorithms for computing the factorization are referred
to as TSQR [14]. Implementations and specializations of the TSQR ideas are available in several
environments, including MapReduce [4, 11], distributed memory MPI [14], and GPUs [1]. All of
these methods avoid computing X T X and hence are numerically stable.
The thin SVD used in Section 2.3 is a small extension of the thin QR factorization. The thin SVD is
X = U?V T , where U is m ? n and orthogonal, ? is diagonal with decreasing, nonnegative diagonal
entries, and V is n?n and orthogonal. Let X = QR be the thin QR factorization of X and R = UR ?V T
be the SVD of R. Then X = (QUR )?V T = U?V T . The matrix U = QUR is m ? n and orthogonal, so
this is the thin SVD of X. The dimension of R is n ? n, so computing its SVD takes O(n3 ) floating
point operations (flops), a trivial cost when n is small. When m n, this method for computing the
SVD is called the R-SVD [8]. Both TSQR and R-SVD require O(mn2 ) flops. However, the dominant
cost is data I/O, and TSQR only reads the data matrix once.
3.2
Column normalization
The convex hull algorithms from Section 2.1 and the Gaussian projection algorithm from Section 2.2
require the columns of the data matrix X to be normalized. A naive implementation of the column
normalization in a MapReduce environment is: (1) read X and compute the column norms; (2) read
X, normalize the columns, and write the normalized data to disk; (3) use TSQR on the normalized
matrix. This requires reading the data matrix twice and writing O(mn) data to disk once just to
normalize the columns. The better approach is a single step: use TSQR on the unnormalized data X
and simultaneously compute the column norms. If D is the diagonal matrix of column norms, then
X = QR ? XD?1 = Q(RD?1 ).
The matrix R? = RD?1 is upper triangular, so QR? is the thin QR factorization of the columnnormalized data. This approach reads the data once and only writes O(n2 ) data. The same idea
applies to Gaussian projection since GT (XD?1 ) = (GT X)D?1 . Thus, our algorithms only need to
read the data matrix once in all cases. (We refer to the algorithm output as selecting the columns
and computing the matrix H, which is typically what is used in practice. Retrieving the entries from
the columns of A from K does require a subsequent pass.)
4
Applications
In this section, we test our dimension reduction technique on massive scientific data sets. The data
are nonnegative, but we do not know a priori that the data is separable. Experiments on synthetic
1
The code is available at https://github.com/arbenson/mrnmf.
5
data sets are provided in an online version of this paper and show that our algorithms are effective
and correct on near-separable data sets.2
All experiments were conducted on a 10-node, 40-core MapReduce cluster. Each node has 6 2-TB
disks, 24 GB of RAM, and a single Intel Core i7-960 3.2 GHz processor. They are connected via
Gigabit ethernet. We test the following three algorithms: (1) dimension reduction with the SVD
followed by SPA; (2) Dimension reduction with the SVD followed by the greedy variant of the
XRAY algorithm; (3) Gaussian projection (GP) as described in Section 2.2. We note that the greedy
variant of XRAY is not exact in the separable case but works well in practice [22].
Using our dimension reduction technique, all three algorithms require reading the data only once.
The algorithms were selected to be a representative set of the approaches in the literature, and we
will refer to the three algorithms as SPA, XRAY, and GP. As discussed in Section 2.3, the choice of
QR or SVD does not matter for these algorithms (although it may matter for other NMF algorithms).
Thus, we only consider the SVD transformation in the subsequent numerical experiments.
4.1
Heat transfer simulation
The heat transfer simulation data contains the simulated heat in a high-conductivity stainless steel
block with a low-conductivity foam bubble inserted in the block [12].3 Each column of the matrix
corresponds to simulation results for a foam bubble of a different radius. Several simulations for random foam bubble locations are included in a column. Each row corresponds to a three-dimensional
spatial coordinate, a time step, and a bubble location. An entry of the matrix is the temperature of
the block at a single spatial location, time step, bubble location, and bubble radius. The matrix is
constructed such that columns near 64 have far more variability in the data ? this is then responsible
for additional ?rank-like? structure. Thus, we would intuitively expect the NMF algorithms to select
additional columns closer to the end of the matrix. (And indeed, this is what we will see shortly.) In
total, the matrix has approximately 4.9 billion rows and 64 columns and occupies a little more than
2 TB on the Hadoop Distributed File System (HDFS).
The left plot of Figure 1 shows the relative error for varying separation ranks. The relative error is
defined as kX ? X(:, K)Hk2F /kXk2F . Even a small separation rank (r = 4) results in a small residual.
SPA has the smallest residuals, and XRAY and GP are comparable. An advantage of our projection
method is that we can quickly test many values of r. For the heat transfer simulation data, we choose
r = 10 for further experiments. This value is near an ?elbow? in the residual plot for the GP curve.
We note that the original SPA and XRAY algorithms would achieve the same reconstruction error
if applied to the entire data set. Our dimension reduction technique allows us to accelerate these
established methods for this large problem.
The middle plot of Figure 1 shows the columns selected by each algorithm. Columns 5 through
30 are not extreme in any algorithm. Both SPA and GP select at least one column in indices one
through four. Columns 41 through 64 have the highest density of extreme columns for all algorithms.
Although the extreme columns are different for the algorithms, the coefficient matrix H exhibits
remarkably similar characteristics in all cases. Figure 2 visualizes the matrix H for each algorithm.
Each non-extreme column is expressed as a conic combination of only two extreme columns. In
general, the two extreme columns corresponding to column i are j1 = arg max{ j ? K | j < i} and
arg min{ j ? K | j > i}. In other words, a non-extreme column is a conic combination of the two
extreme columns that ?sandwich? it in the data matrix. Furthermore, when the index i is closer to
j1 , the coefficient for j1 is larger and the coefficient for j2 is smaller. This phenomenon is illustrated
in the right plot of Figure 1.
4.2
Flow cytometry
The flow cytometry (FC) data represent abundances of fluorescent molecules labeling antibodies
that bind to specific targets on the surface of blood cells.4 The phenotype and function of individual
cells can be identified by decoding these label combinations. The analyzed data set contains measurements of 40,000 single cells. The measurement fluorescence intensity conveying the abundance
2
http://arxiv.org/abs/1402.6964.
The heat transfer simulation data is available at https://www.opensciencedatacloud.org.
4
The FC data is available at https://github.com/arbenson/mrnmf/tree/master/data.
3
6
Figure 1: (Left) Relative error in the separable factorization as a function of separation rank (r)
for the heat transfer simulation data. Our dimension reduction technique lets us test all values of
r quickly. (Middle) The first 10 extreme columns selected by SPA, XRAY, and GP. We choose 10
columns as there is an ?elbow? in the GP curve there (left plot). The columns with larger indices
are more extreme, but the algorithms still select different columns. (Right) Values of H(K ?1 (1), j)
and H(K ?1 (34), j) computed by SPA for j = 2, . . . , 33, where K ?1 (1) and K ?1 (34) are the indices
of the extreme columns 1 and 34 in W (X = WH). Columns 2 through 33 of X are roughly convex
combinations of columns 1 and 34, and are not selected as extreme columns by SPA. As j increases,
H(K ?1 (1), j) decreases and H(K ?1 (34), j) increases.
Figure 2: Coefficient matrix H for SPA, XRAY, and GP for the heat transfer simulation data when
r = 10. In all cases, the non-extreme columns are conic combinations of two of the selected columns,
i.e., each column in H has at most two non-zero values. Specifically, the non-extreme columns are
conic combinations of the two extreme columns that ?sandwich? them in the matrix. See the right
plot of Figure 1 for a closer look at the coefficients.
information were collected at five different bands corresponding to the FITC, PE, ECD, PC5, and
PC7 fluorescent labels tagging antibodies against CD4, CD8, CD19, CD45, and CD3 epitopes.
The measurements are represented as the data matrix A of size 40, 000 ? 5. Our interest in the presented analysis was to study pairwise interactions in the data (cell vs. cell, and marker vs. marker).
Thus, we are interested in the matrix X = A ? A, the Kronecker product of A with itself. Each row
of X corresponds to a pair of cells and each column to a pair of marker abundance values. X has
dimension 40, 0002 ? 52 and occupies 345 GB on HDFS.
The left plot of Figure 3 shows the residuals for the three algorithms applied to the FC data for
varying values of the separation rank. In contrast to the heat transfer simulation data, the relative
errors are quite large for small r. In fact, SPA has large relative error until nearly all columns are
selected (r = 22). XRAY has the smallest residual for any value of r. The right plot of Figure 3
shows the columns selected when r = 16. XRAY and GP only disagree on one column. SPA
chooses different columns, which is not surprising given the relative residual error. Interestingly, the
columns involving the second marker defining the phenotype (columns 2, 6, 7, 8, 9, 10, 12, 17, 22)
are underrepresented in all the choices. This suggests that the information provided by the second
marker may be redundant. In biological terms, it may indicate that the phenotypes of the individual
cells can be inferred from a smaller number of markers. Consequently, this opens a possibility that in
modified experimental conditions, the FC researchers may omit this particular label, and still be able
to recover the complete phenotypic information. Owing to the preliminary nature of these studies,
a more in-depth analysis involving multiple similar blood samples would be desirable in order to
confirm this hypothesis.
7
Figure 3: (Left) Relative error in the separable factorization as a function of nonnegative rank (r)
for the flow cytometry data. (Right) The first 16 extreme columns selected by SPA, XRAY, and GP.
We choose 16 columns since the XRAY and GP curve levels for larger r (left plot).
Figure 4: Coefficient matrix H for SPA, XRAY, and GP for the flow cytometry data when r =
16. The coefficients tend to be clustered near the diagonal. This is remarkably different to the
coefficients for the heat transfer simulation data in Figure 2.
Finally, Figure 4 shows the coefficient matrix H. The coefficients are larger on the diagonal, which
means that the non-extreme columns are composed of nearby extreme columns in the matrix.
5
Discussion
We have shown how to compute nonnegative matrix factorizations at scale for near-separable talland-skinny matrices. Our main tool was TSQR, and our algorithms only needed to read the data
matrix once. By reducing the dimension of the problem, we can easily compute the efficacy of
factorizations for several values of the separation rank r. With these tools, we have computed the
largest separable nonnegative matrix factorizations to date. Furthermore, our algorithms provide
new insights into massive scientific data sets. The coefficient matrix H exposed structure in the
results of heat transfer simulations. Extreme column selection in flow cytometry showed that one
of the labels used in measurements may be redundant. In future work, we would like to analyze
additional large-scale scientific data sets. We also plan to test additional NMF algorithms.
The practical limits of our algorithm are imposed by the tall-and-skinny requirement where we
assume that it is easy to manipulate n ? n matrices. The synthetic examples we explored used up
to 200 columns, and regimes up to 5000 columns have been explored in prior work [11]. A rough
rule of thumb is that our implementations should be possible as long as an n ? n matrix fits into
main memory. This means that implementations based on our work will scale up to 30, 000 columns
on machines with more than 8 GB of memory; although at this point communication begins to
dominate. Solving these problems with more columns is a challenging opportunity for the future.
Acknowledgments
ARB and JDL are supported by an Office of Technology Licensing Stanford Graduate Fellowship.
JDL is also supported by a NSF Graduate Research Fellowship. DFG is supported by NSF CAREER
award CCF-1149756. BR is supported by NIH grant 1R21EB015707-01.
8
References
[1] M. Anderson, G. Ballard, J. Demmel, and K. Keutzer. Communication-avoiding QR decomposition for
GPUs. In IPDPS, pages 48?58, 2011.
[2] M. Ara?ujo et al. The successive projections algorithm for variable selection in spectroscopic multicomponent analysis. Chemometrics and Intelligent Laboratory Systems, 57(2):65?73, 2001.
[3] S. Arora, R. Ge, R. Kannan, and A. Moitra. Computing a nonnegative matrix factorization?provably. In
Proceedings of the 44th symposium on Theory of Computing, pages 145?162. ACM, 2012.
[4] A. R. Benson, D. F. Gleich, and J. Demmel. Direct QR factorizations for tall-and-skinny matrices in
MapReduce architectures. In 2013 IEEE International Conference on Big Data, pages 264?272, 2013.
[5] J. M. Bioucas-Dias and A. Plaza. An overview on hyperspectral unmixing: geometrical, statistical, and
sparse regression based approaches. In IGARSS, pages 1135?1138, 2011.
[6] V. Bittorf, B. Recht, C. Re, and J. A. Tropp. Factoring nonnegative matrices with linear programs. In
NIPS, pages 1223?1231, 2012.
[7] J. W. Boardman et al. Automating spectral unmixing of aviris data using convex geometry concepts. In
4th Annu. JPL Airborne Geoscience Workshop, volume 1, pages 11?14. JPL Publication 93?26, 1993.
[8] T. F. Chan. An improved algorithm for computing the singular value decomposition. ACM Trans. Math.
Softw., 8(1):72?83, Mar. 1982.
[9] M. T. Chu and M. M. Lin. Low-dimensional polytope approximation and its applications to nonnegative
matrix factorization. SIAM Journal on Scientific Computing, 30(3):1131?1155, 2008.
[10] A. Cichocki and R. Zdunek. Regularized alternating least squares algorithms for non-negative matrix/tensor factorization. In Advances in Neural Networks?ISNN 2007, pages 793?802. Springer, 2007.
[11] P. G. Constantine and D. F. Gleich. Tall and skinny QR factorizations in MapReduce architectures. In
Second international workshop on MapReduce and its applications, pages 43?50. ACM, 2011.
[12] P. G. Constantine, D. F. Gleich, Y. Hou, and J. Templeton. Model reduction with MapReduce-enabled tall
and skinny singular value decomposition. SIAM J. Sci. Comput., Accepted:To appear, 2014.
[13] A. Damle and Y. Sun. Random projections for non-negative matrix factorization. arXiv:1405.4275, 2014.
[14] J. Demmel, L. Grigori, M. Hoemmen, and J. Langou. Communication-optimal parallel and sequential
QR and LU factorizations. SIAM J. Sci. Comp., 34, Feb. 2012.
[15] D. Donoho and V. Stodden. When does non-negative matrix factorization give a correct decomposition
into parts? In NIPS, 2003.
[16] N. Gillis and S. Vavasis. Fast and robust recursive algorithms for separable nonnegative matrix factorization. Pattern Analysis and Machine Intelligence, IEEE Transactions on, PP(99):1?1, 2013.
[17] D. Guillamet and J. Vitri`a. Non-negative matrix factorization for face recognition. In Topics in Artificial
Intelligence, pages 336?344. Springer, 2002.
[18] S. Jia and Y. Qian. Constrained nonnegative matrix factorization for hyperspectral unmixing. Geoscience
and Remote Sensing, IEEE Transactions on, 47(1):161?173, 2009.
[19] D. Kim, S. Sra, and I. S. Dhillon. Fast projection-based methods for the least squares nonnegative matrix
approximation problem. Statistical Analysis and Data Mining, 1(1):38?51, 2008.
[20] W. Kim, B. Chen, J. Kim, Y. Pan, and H. Park. Sparse nonnegative matrix factorization for protein
sequence motif discovery. Expert Systems with Applications, 38(10):13198?13207, 2011.
[21] D. Kuang, H. Park, and C. H. Ding. Symmetric nonnegative matrix factorization for graph clustering. In
SDM, volume 12, pages 106?117, 2012.
[22] A. Kumar, V. Sindhwani, and P. Kambadur. Fast conical hull algorithms for near-separable non-negative
matrix factorization. In ICML, 2013.
[23] D. D. Lee and H. S. Seung. Algorithms for non-negative matrix factorization. In NIPS, 2000.
[24] C. Liu, H.-C. Yang, J. Fan, L.-W. He, and Y.-M. Wang. Distributed nonnegative matrix factorization for
web-scale dyadic data analysis on mapreduce. In WWW, pages 681?690. ACM, 2010.
[25] C. Thurau, K. Kersting, and C. Bauckhage. Yes we can: simplex volume maximization for descriptive
web-scale matrix factorization. In CIKM, pages 1785?1788. ACM, 2010.
[26] S. Vavasis. On the complexity of nonnegative matrix factorization. SIAM Journal on Optimization,
20(3):1364?1377, 2010.
9
| 5335 |@word version:1 middle:2 briefly:1 compression:1 norm:8 disk:4 open:1 simulation:12 decomposition:7 pick:1 reduction:21 liu:1 contains:2 efficacy:2 selecting:2 interestingly:1 outperforms:1 com:2 surprising:1 chu:1 written:1 hou:1 subsequent:2 numerical:1 j1:3 plot:9 update:2 v:2 greedy:3 selected:9 half:1 intelligence:2 core:3 provides:1 math:1 node:2 location:4 successive:2 bittorf:1 org:2 five:1 constructed:1 direct:1 symposium:1 viable:1 retrieving:1 ray:6 pairwise:1 tagging:1 indeed:1 roughly:1 gigabit:1 multi:1 ara:1 decreasing:1 little:1 boardman:1 elbow:2 provided:2 begin:1 notation:1 what:2 minimizes:1 developed:1 finding:5 transformation:14 xd:6 prohibitively:1 rm:4 k2:2 conductivity:2 omit:1 grant:1 appear:1 bioucas:1 before:1 negligible:1 local:1 bind:1 limit:1 consequence:1 wyk:1 approximately:2 twice:1 cd3:1 suggests:1 challenging:1 kxk2f:1 aviris:1 factorization:40 range:1 graduate:2 practical:1 responsible:1 acknowledgment:1 practice:3 block:6 implement:1 recursive:1 writes:1 ipdps:1 krt:1 significantly:1 projection:15 word:3 protein:2 get:2 cannot:1 convenience:1 selection:2 context:1 live:2 seminal:1 applying:2 writing:1 www:2 imposed:1 center:1 regardless:1 convex:13 focused:1 underrepresented:1 spark:1 immediately:1 pure:1 qian:1 insight:2 rule:1 dominate:1 hd:1 enabled:1 handle:1 exploratory:1 coordinate:2 justification:1 nnls:2 updated:1 analogous:1 variation:1 target:1 massive:3 exact:2 programming:1 hypothesis:1 expensive:1 particularly:1 updating:1 recognition:1 inserted:1 guillamet:1 ding:1 wang:1 connected:1 sun:1 remote:1 decrease:1 mz:1 highest:1 environment:2 complexity:1 miny:2 seung:2 rajwa:1 depend:1 solving:2 exposed:1 easily:2 accelerate:1 represented:1 separated:1 heat:11 fast:5 effective:2 mn2:1 demmel:3 artificial:1 labeling:1 choosing:1 whose:1 heuristic:1 stanford:7 valued:2 solve:3 larger:4 quite:1 triangular:2 vitri:1 gp:12 noisy:2 itself:1 final:1 online:1 advantage:5 rr:3 sequence:2 sdm:1 descriptive:1 reconstruction:1 interaction:1 product:1 remainder:1 j2:1 date:1 achieve:2 frobenius:1 normalize:2 scalability:1 qr:17 chemometrics:1 billion:2 empty:1 requirement:2 cluster:1 unmixing:7 generating:1 tall:14 arb:1 qt:5 arbenson:3 exacerbated:1 implemented:1 implies:1 ethernet:1 indicate:1 radius:2 correct:3 owing:1 hull:8 occupies:2 dii:1 require:7 generalization:1 clustered:1 preliminary:1 spectroscopic:1 biological:1 extension:1 hold:2 practically:1 thurau:1 smallest:2 purpose:1 label:4 fluorescence:1 largest:1 tool:3 rough:1 gaussian:9 modified:1 rather:2 avoid:1 kersting:1 varying:2 office:1 publication:1 epitope:1 rank:12 contrast:2 kim:3 motif:2 factoring:1 streaming:1 typically:4 entire:3 hidden:1 manipulating:1 transformed:1 selects:1 interested:1 provably:2 pixel:2 issue:1 among:1 arg:7 priori:2 plan:1 art:1 platform:2 spatial:2 constrained:1 once:11 never:1 having:1 softw:1 broad:1 look:1 park:2 nearly:3 thin:7 icml:1 future:2 minimized:1 np:1 simplex:1 intelligent:1 employ:1 composed:1 preserve:6 simultaneously:1 stainless:1 individual:2 dfg:1 floating:1 geometry:8 skinny:15 ourselves:1 sandwich:2 xray:15 ab:1 interest:1 possibility:1 mining:1 dgleich:1 severe:1 analyzed:1 extreme:44 devoted:1 closer:3 facial:3 orthogonal:14 indexed:8 tree:1 re:1 theoretical:1 column:93 maximization:1 cost:4 minr:1 entry:8 hundred:1 kuang:1 conducted:1 too:1 synthetic:2 chooses:1 recht:1 density:1 international:2 siam:4 automating:1 lee:3 decoding:1 invertible:2 quickly:3 ear:1 moitra:1 choose:4 expert:1 coefficient:14 matter:2 performed:1 try:1 jason:1 analyze:1 recover:1 parallel:4 jia:1 square:5 formed:1 characteristic:1 nonsingular:4 conveying:1 foam:3 yes:1 thumb:1 produced:1 mc:1 lu:1 comp:1 researcher:1 processor:1 visualizes:1 against:1 pp:1 bioscience:1 popular:2 wh:2 subsection:1 gleich:4 routine:1 uncover:1 sophisticated:1 improved:2 execute:1 mar:1 anderson:1 furthermore:5 just:2 until:1 working:1 tropp:1 web:2 marker:6 icme:2 scientific:8 k22:7 normalized:5 concept:1 ccf:1 hence:2 read:11 alternating:4 laboratory:1 dhillon:1 symmetric:1 illustrated:1 deal:2 mpi:2 unnormalized:1 complete:3 demonstrate:1 reflection:1 temperature:1 geometrical:1 image:2 invoked:1 nih:1 common:3 rotation:1 overview:1 hk2f:1 volume:3 million:1 jdl17:1 discussed:1 he:1 numerically:1 significant:1 refer:2 measurement:4 rd:2 similarly:1 had:1 stable:1 yk2:1 surface:1 gt:4 feb:1 dominant:1 recent:3 showed:1 chan:1 optimizing:1 constantine:2 cd4:1 fortunately:1 additional:4 terabyte:1 employed:1 determine:1 converge:1 redundant:2 full:1 multiple:1 desirable:1 long:1 lin:1 host:1 serial:1 manipulate:1 award:1 scalable:8 variant:3 hdfs:2 involving:2 regression:1 arxiv:2 iteration:1 represent:2 normalization:2 cell:7 preserved:1 addition:1 want:1 remarkably:3 fellowship:2 completes:1 singular:5 hott:1 airborne:1 posse:1 file:1 tend:1 flow:6 call:2 intolerable:1 near:15 leverage:1 yang:1 intermediate:1 easy:1 gillis:1 fit:2 architecture:4 restrict:1 identified:1 idea:5 tradeoff:1 br:1 i7:1 motivated:4 specialization:1 gb:3 matlab:1 useful:1 stodden:2 band:1 reduced:3 http:4 vavasis:2 nsf:2 upfront:1 cikm:1 bulk:1 write:1 key:1 four:1 blood:2 phenotypic:1 destroy:1 graph:2 ram:1 cone:5 sum:1 master:1 almost:1 separation:8 antibody:2 keutzer:1 spa:16 comparable:1 followed:2 conical:4 fan:1 nonnegative:26 plaza:1 precisely:1 kronecker:1 n3:1 nearby:1 min:3 extremely:3 kumar:1 separable:26 gpus:3 department:1 combination:6 smaller:2 pan:1 separability:14 ur:1 templeton:1 benson:2 intuitively:1 restricted:1 taken:2 computationally:1 equation:4 turn:1 needed:1 know:1 ge:1 tractable:2 end:1 dia:1 available:4 operation:1 apply:2 spectral:1 alternative:3 shortly:1 original:3 top:1 clustering:2 include:2 ecd:1 opportunity:1 restrictive:1 k1:1 tensor:1 already:1 strategy:2 diagonal:7 exhibit:1 mx:1 simulated:1 sci:2 topic:1 polytope:1 collected:1 trivial:1 reason:3 kannan:1 assuming:1 code:1 index:5 minn:1 kambadur:1 unfortunately:1 negative:9 steel:1 implementation:8 upper:3 disagree:1 purdue:4 descent:1 flop:2 defining:1 communication:5 ever:1 variability:1 rn:1 cytometry:6 community:1 intensity:2 nmf:22 inferred:1 david:1 pair:2 established:2 nip:3 trans:1 able:1 pattern:1 regime:1 reading:2 tb:2 program:1 kqt:1 interpretability:1 including:2 memory:6 max:3 suitable:1 regularized:1 residual:12 mn:1 fitc:1 github:2 technology:1 eye:1 numerous:1 conic:4 arora:1 bubble:6 naive:1 cichocki:1 prior:2 review:1 mapreduce:15 discovery:2 literature:3 geometric:4 determining:1 relative:7 expect:1 highlight:1 licensing:1 fluorescent:2 storing:1 austin:1 row:8 supported:4 keeping:1 face:1 sparse:2 benefit:1 distributed:4 ghz:1 dimension:26 curve:3 depth:1 far:1 transaction:2 implicitly:2 confirm:1 xi:3 ballard:1 nature:1 transfer:10 ca:2 sra:1 molecule:1 hadoop:3 career:1 robust:1 whk:1 main:3 big:4 noise:1 n2:4 jdl:2 repeated:1 dyadic:1 west:2 referred:2 intel:1 representative:1 slow:1 comput:1 lie:1 candidate:1 pe:1 abundance:3 annu:1 specific:2 zdunek:1 explored:2 sensing:1 jpl:2 workshop:2 sequential:1 kr:3 hyperspectral:6 kx:9 chen:1 phenotype:3 fc:4 expressed:1 geoscience:2 sindhwani:1 applies:1 springer:2 bauckhage:1 corresponds:3 satisfies:1 acm:5 sized:1 goal:2 donoho:2 consequently:1 replace:1 included:1 specifically:1 reducing:1 called:2 total:1 pas:2 invariance:1 svd:15 premium:1 experimental:1 accepted:1 select:3 bioinformatics:2 phenomenon:1 avoiding:1 |
4,789 | 5,336 | Analog Memories in a Balanced Rate-Based
Network of E-I Neurons
Dylan Festa
[email protected]
Guillaume Hennequin
[email protected]
M?at?e Lengyel
[email protected]
Computational & Biological Learning Lab, Department of Engineering
University of Cambridge, UK
Abstract
The persistent and graded activity often observed in cortical circuits is sometimes seen as a signature of autoassociative retrieval of memories stored earlier
in synaptic efficacies. However, despite decades of theoretical work on the subject, the mechanisms that support the storage and retrieval of memories remain
unclear. Previous proposals concerning the dynamics of memory networks have
fallen short of incorporating some key physiological constraints in a unified way.
Specifically, some models violate Dale?s law (i.e. allow neurons to be both excitatory and inhibitory), while some others restrict the representation of memories to
a binary format, or induce recall states in which some neurons fire at rates close
to saturation. We propose a novel control-theoretic framework to build functioning attractor networks that satisfy a set of relevant physiological constraints. We
directly optimize networks of excitatory and inhibitory neurons to force sets of
arbitrary analog patterns to become stable fixed points of the dynamics. The resulting networks operate in the balanced regime, are robust to corruptions of the
memory cue as well as to ongoing noise, and incidentally explain the reduction
of trial-to-trial variability following stimulus onset that is ubiquitously observed
in sensory and motor cortices. Our results constitute a step forward in our understanding of the neural substrate of memory.
1
Introduction
Memories are thought to be encoded in the joint, persistent activity of groups of neurons. According
to this view, memories are embedded via long-lasting modifications of the synaptic connections
between neurons (storage) such that partial or noisy initialization of the network activity drives
the collective dynamics of the neurons into the corresponding memory state (recall) [1]. Models of
memory circuits following these principles abound in the theoretical neuroscience literature, but few
respect some of the most fundamental properties of brain networks, including: i) the separation of
neurons into distinct classes of excitatory (E) and inhibitory (I) cells ? known as Dale?s law ?, ii) the
presence of recurrent and sparse synaptic connections, iii) the possibility for each neuron to sustain
graded levels of activity in different memories, iv) the firing of action potentials at reasonably low
rates, and v) a dynamic balance of E and I inputs.
In the original Hopfield network [1], connectivity must be symmetrical, which violates Dale?s law.
Moreover, just as in much of the work following it up, memories are encoded in binary neuronal
responses and so converge towards effectively binary recall states even if the recall dynamics formally uses graded activities [2]. Subsequent work considered non-binary pattern distributions [3, 4],
and derived high theoretical capacity limits for them, but those capacities proved difficult ? if not
impossible ? to realise in practice [5, 6], and the network dynamics therein did not explicitly model
inhibitory neurons thus implicitly assuming instantaneous inhibitory feedback. More recent work
1
a
c
b
exc. (prescribed distribution)
20 Hz
memories
5 Hz
inh. (optimized distribution)
0
exc. neurons
inh. neurons
10
20
30
firing rate [Hz]
Figure 1: (a) Examples of analog patterns of excitatory neuronal activities, drawn from a log-normal
distribution. In all our training experiments, network parameters were optimized to stabilize a set
of such analog patterns and the baseline, uniform activity state (top row). For ease of visualization,
only 30 of the 100 excitatory neurons are shown. (b) Optimized values of the inhibitory (auxiliary)
neuronal firing rates for 5 of 30 learned memories (corresponding to those in panel a). Only 30 of
the 50 auxiliary neurons are shown. (c) Empirical distributions of firing rates across neurons and
memory patterns, for each population.
incorporated Dale?s law, and described neurons using the more realistic, leaky integrate-and-fire
(LIF) neuron model [7]. However, the stability of the recall states still relied critically on the saturating behavior of the LIF input-output transfer function at high rates. Although it was later shown
that dynamic feedback inhibition can stabilize relatively low firing rates in subpopulations of more
tightly connected neurons [8, 9], inhibitory feedback in these models is global, and calibrated for a
single stereotypical level of excitation for all memories, implying effectively binary memories again.
Finally, spatially connected networks are able to sustain graded activity patterns (spatial ?bumps?),
but make strong assumptions about the spatial structure of both the connectivity and the memory
patterns, and are sensitive to ongoing noise (e.g. [10, 11]). Ref. [12] provides a rare example of
spike timing-based graded memory network, but it again did not contain inhibitory units.
Here we propose a general control-theoretic framework that overcomes all of the above limitations
with minimal additional assumptions. We formalize memory storage as implying two conditions:
that the desired activity states be fixed points of the dynamics, and that the dynamics be stable
around those fixed points. We directly optimize the network parameters, including the synaptic
connectivity, to satisfy both conditions for a collection of arbitrary, graded memory patterns (Fig. 1).
The fixed point condition is achieved by minimizing the time derivative of the neural activity, such
that ideally it reaches zero, at each of the desired attractor states. Stability, however, is more difficult
to achieve because the fixed-point constraints tend to create strong positive feedback loops in the
recurrent circuitry, and direct measures of dynamical stability (eg. the spectral abscissa) do not admit
efficient, gradient-based optimization. Thus, we use recently developed methods from robust control
theory, namely the minimization of the Smoothed Spectral Abscissa (SSA, [13, 14]) to perform
robust stability optimization. To satisfy biological constraints, we parametrize the networks that we
optimize such that they have realistic firing rate dynamics and their connectivities obey Dale?s law.
We show that despite these constraints the resulting networks perform memory recall that is robust
to noise in both the recall cue and the ongoing dynamics, and is stabilized through a tight dynamic
balance of excitation and inhibition. This novel way of constructing structurally realistic memory
networks should open new routes to the understanding of memory and its neural substrate.
2
Methods
We study a network of n = nE (excitatory) +nI (inhibitory) neurons. The activity of neuron i is
represented by a single scalar potential vi , which is converted into a firing rate ri via a thresholdquadratic gain function (e.g. [15]):
2
?vi
if vi > 0
:=
ri = g(vi )
(1)
0
otherwise
2
We set ? to 0.04, such that g(vi ) spans a few tens of Hz when vi spans a few tens of mV, as
experimentally observed in cortical areas (e.g. cat?s V1 [16]). The instantaneous state of the system
can be expressed as a vector v(t) := (v1 (t), . . . , vn (t)). We denote the activity of the excitatory or
inhibitory subpopulation by vexc and vinh , respectively. The recurrent interactions between neurons
are governed by a synaptic weight matrix W, in which the sign of each element Wij depends on
the nature (excitatory or inhibitory) of the presynaptic neuron j. We enforce Dale?s law via a reparameterization of the synaptic weights:
+1
if j ? nE
(2)
Wij = sj log(1 + exp ?ij ) with sj =
?1
otherwise
where the ?ij ?s are free, unconstrained parameters. (We do not allow for autapses, i.e. we fix Wii =
0). The network dynamics are thus given by:
?i
n
X
dvi
= ?vi +
Wij g(vj ) + hi ,
dt
j=1
(3)
where ?i is the membrane time constant, and hi is a constant external input, independent of the
memory we wish to recall.
It is worth noting that, since the gain function g(vi ) defined in Eq (1) has no upper saturation,
recurrent interactions can easily result in runaway excitation and firing rates growing unbounded.
However, our optimization algorithm will naturally seek stable solutions, in which firing rates are
kept within a limited range due to a fine dynamic balance of excitation and inhibition [14].
Optimizing network parameters to embed attractor memories
We are going to build and study networks that have a desired set of analog activity patterns as stable
?
}?=1,...,m be a set of m target analog patterns (Fig. 1),
fixed points of their dynamics. Let {vexc
defined in the space of excitatory neuronal activity (potentials). For a given pattern ?, the inhibitory
?
neurons will be free to adjust their steady state firing rates vinh
to whatever pattern proves to be
optimal to maintain stability. In other words, we think of the activity of inhibitory neurons as
?auxiliary? variables.
>
>
? >
?
A given activity pattern v? ? (vexc
, vinh
) is a stable fixed point of the network dynamics if, and
only if, it satisfies the following two conditions:
dv
=0
and
? (J? ) < 0
(4)
dt
?
v=v
?
:= Wij g 0 (vj? ) ? ?ij (Kronecker?s
where J? is the Jacobian matrix of the dynamics in Eq. 3, i.e. Jij
?
delta), and ?(J ) denotes the spectral abscissa (SA), defined as the largest real part in the eigenvalue
spectrum of J? . The first condition makes v? a fixed point of the dynamics, while the second
condition makes that fixed point asymptotically stable with respect to small local perturbations.
Note that the width of the basin of attraction is not captured by the SA.
The two conditions in Eq. 4 depend on a set of network parameters that we will allow ourselves
to optimize. These are all the synaptic weight parameters (?ij , i 6= j), as well as the values of the
?
inhibitory neurons? firing rates in each attractor (vinh
, ? = 1, . . . , m). Thus, we may adjust a total
of n(n ? 1) + nI m parameters.
Pn
Using Eq. 3, the first condition in Eq. 4 can be rewritten as vi? ? j=1 Wij g(vj? ) ? hi = 0.
Despite this equation being linear in the synaptic weights, the re-parameterization of Eq. 2 makes
?
it nonlinear in ?, and it is in any case nonlinear in vinh
. We will therefore seek to satisfy this
2
condition by minimizing k dv/dt|v=v? k , which quantifies how fast the potentials drift away when
initialized in the desired attractor state v? . When it is zero, v? is a fixed point of the dynamics. Our
optimization procedure (see below) may not be able to set this term to exactly zero, especially as we
try to store a large number of memories, but in practice we find it becomes small enough that the
Jacobian-based stability criterion remains valid.
Meeting the stability condition (second condition in Eq. 4) turns out to be more involved. The SA
is, in general, a non-smooth function of the matrix elements and is therefore difficult to minimize.
3
A more suitable stability measure has been introduced recently in the context of robust control
theory [13, 14], called the Smoothed Spectral Abscissa (SSA), which we will use here and denote
by ?
? ? (J? ). The SSA, defined for some smoothness parameter ? > 0, is a differentiable relaxation of
the SA, with the properties ?(J? ) < ?
? ? (J? ) and lim??0 ?
? ? (J? ) = ?(J? ). Therefore, the criterion
?
? ? (J? ) ? 0 implies ?(J? ) < 0, and can therefore be used as an indication of local stability.
Both the SSA and its gradient are straightforward to evaluate numerically, making it amenable to
minimization through gradient descent. Note that the SSA depends on the Jacobian matrix elements
?
?
{Jij
}, which in turn depend both on the connectivity parameters {?ij } and on vinh
. Note also that
the parameter ? > 0 controls how tightly the SSA hugs the SA. Small values make it a tight upper
bound, with increasingly ill-behaved gradients. Large values imply more smoothness, but may no
longer guarantee that the SSA has a negative minimum even though the SA might have one. In our
system of n = 150 neurons we found ? = 0.01 to yield a good compromise. In the general case the
distance between SA and SSA grows linearly with the number of dimensions.To keep it invariant, ?
should be scaled accordingly. We therefore used the following heuristic rule ? = 0.01 ? 150/n.
We summarize the above objective into a global cost function by lumping together the fixed point
and stability conditions, summing over the entire set of m target memory patterns, and adding an L2
penalty term on the synaptic weights to regularize:
!
2
m
1 X 1
?F
dv
2
?
?
? ({?ij }, {vinh }) :=
+ ?s ?
? ? (J ) + 2 kWkF .
(5)
m ?=1 n
dt
v=v?
n
where kWk2F is the squared Frobenius norm of W, i.e. the sum of its squared elements, and the
parameters ?s and ?F control the relative importance of each component of the objective function.
We set them heuristically (Table 1). We used a variant of the low-storage BFGS algorithm included
in the open source library NLopt [17] to minimize ?.
Choice of initial parameters and attractors
The synaptic weights are initially drawn randomly from a Gamma distribution with a shape factor of
2 and a mean that depends only on the type of pre- and post-synaptic population. The mean synaptic
weights of the four synapse types were computed using a mean-field reduction of the full network
?=1
in which all
to meet the condition that the network initially exhibits a stable baseline state vexc
excitatory firing rates equal rbaseline = 5 Hz (Table 1, and Supplementary Material). This baseline state was included in every set of m target attractors that we used and was thus stable from
?
}?=2,...,m were generated
the beginning, by construction. For the remaining target patterns, {vexc
?1
by inverting (using g ) firing rates that were sampled from a log-normal distribution with a mean
matching the baseline firing rate, rbaseline (Fig. 1a) and a variance of 5 Hz. This log-normal distribution was chosen to roughly capture the skewed and heavy-tailed nature of firing rate distributions
?
observed in vivo (see e.g. for a review [18]). The inhibitory potentials in the memory states, {vinh
},
?1
were initialized to the baseline, g (5 Hz), and were subsequently used as free parameters by the
learning algorithm (cf. above; see also Fig. 1b).
3
Results
Example of successful storage
Figure 2 shows an example of stability optimization: in this specific run we used 150 neurons to embed 30 graded attractors (examples of which where shown in Fig. 1), yielding a storage capacity of
0.2. Other parameters are listed in Table 1. Gradient descent gradually reduces each of the attractorspecific sub-objectives in Eq. 5, namely the SSA, the SA, and the potential velocities kdv/dtk2 in
each target state (Fig. 2). After convergence, the SSA has become negative for all desired states,
indicating stability. Note, however, that kdv/dtk after convergence is small but non-zero in each
of the target memories. Thus, strictly speaking, the target patterns haven?t become fixed points of
the dynamics, but only slow points from which the system will eventually drift away. In practice
though, we found that stability was robust enough that an exact, stable fixed point had in fact been
created very near each target pattern. This is detailed below.
4
? ? (J? )i?
h?
m = 50
b
?
0
?0.5
?1
m = 30
10?2
D
E
v? (?)
2
SA / SSA
a
h? (J? )i?
10?4
0
0
20
40
60
time (hours)
20
40
60
time (hours)
Figure 2: (a) Decrease of the SA (solid line) and of the SSA (dotted line) during learning in systems
with 30 (purple) and 50 attractors (orange). Thick lines show averages across attractors, flanking
lines show the corresponding standard deviations. The x-axis marks the actual duration of the run of
the learning algorithm. (b) Euclidean norm of the velocity at the fixed point during learning. Lines
and colors as in a. Note the logarithmic y-axis.
Table 1: Parameter settings
nE
nI
m
100
50
30
?E
?I
rbaseline
20 ms
10 ms
5 Hz
?s
?F
0.02
0.001
Memory recall performance and robustness
For recall, we initialize neuronal activities at a noisy version of one of the target patterns, and study
the subsequent evolution of the network state. The network performs well if its dynamics clean up
the noise and home in on the target pattern (autoassociative behavior) and if it achieves this robustly
even in the face of large amounts of noise.
Initial cues are chosen to be linear combinations of the form r(t = 0) = ? ?r + (1 ? ?) r? , where r?
is the memory we intend to recall and ?r is an independent random vector with the same lognormal
statistics used to generate the memory patterns themselves. The parameter ? regulates the noise
level: ? = 0 sets the network activity directly in the desired attractor, while ? = 1 initializes it with
completely random values.
The deviation of the momentary network state r(t) ? g(v(t)) from the target pattern r? ? g(v? )
is measured in terms of the squared Euclidean distance, further normalized by the expected squared
distance between r? and a random pattern drawn from the same distribution (log-normal in our
case). Formally:
krexc (t) ? r?exc k2
d? (t) :=
.
(6)
hk?rexc ? r?exc k2 i?r
Figure 3a shows the temporal evolution of d? (t) on a few sample recall trials, for two different noise
levels ?. For ? = 0.5, recalls are always successful, as the network state converges to the right target
pattern on each trial. For ? = 0.75, the network activity occasionally settles in another, well distinct
attractor.
We used the convention that a trial is deemed successful if the distance d? (t) falls below 0.001. (A
? 3 Hz deviation from the target in only one of the 100 exc. neurons, with all other 99 neurons
behaving perfectly, would be sufficient to cross this threshold and fail the test.) We further measure
performance as the probability of successful recall, which we estimated from many independent
trials with different realizations of the noise ?r in the initial condition (Figure 3b). The network
performance is also compared to an ?ideal observer? [6] that has direct access to all the stored
memories (rather than just their reflection in the synaptic weights) and simply returns that pattern
in the training set {r? } to which the initial cue is closest (Fig. 3b). Thus, as an upper bound on
performance, the ideal observer only produces a wrong recall when the added noise brings the
initial state closer to an attractor that is different from the target. Remarkably, our network dynamics
5
2
? = 0.50
? = 0.75
d? (t )
1.5
1
0.5
0
0
0.1
t (s)
b
probability of success
(a)
a
0.2
network
ideal
memories
baseline
1
0.8
0.6
0.4
0.2
0
0
0.2
0.4 0.6
d? (t0 )
0.8
1
Figure 3: (a) Example recall trials for a single memory r? , which is presented to the network at
time t = 0 in a corrupted version that is different on every trial, for two different values of the
noise level ? (colors). Shown here is the temporal evolution of the momentary distance between the
vector of excitatory firing rates rexc (t) and the memory pattern r?exc . Different lines correspond to
different trials. (b) Fraction of trials that converged onto the correct attractor (final distance d? (t =
?) < 0.001, cf. text) as a function of the normalized distance between the initial condition and the
desired attractor, d? (t = 0). Thick lines show medians across attractors, flanking thin lines show
the 25th and 75th percentiles. The performance of the baseline state is shown separately (orange).
The dashed lines show the performance of an ?ideal observer?, always selecting the memory closest
to the initial condition, for the same trials.
(continuous lines) and the ideal observer (dashed lines) have comparable performances. When trying
to recall the uniform pattern of baseline activity, the performance appears much better (orange line)
both for the ideal observer and the network. This is simply because the random vectors used to
perturb the system have a high probability of lying closer to the mean of the log normal distribution
(that is, the baseline state) than to any other memory pattern. Moreover, the network was initialized
prior to learning with the baseline as the single global attractor, and this might account for the
additional tendency of the network (solid orange line) to fall on such state, as compared to the ideal
observer (dotted orange line).
Only a few strong synaptic weights contribute to memory recall
Synaptic weights after learning (Fig. 4a) are sparse: their distribution shows the characteristic peak
near zero and the long tail observed in real cortical circuits [19, 20] (Fig. 4b). This sparseness cannot
be accounted for by the L2 norm regularizer in the cost function (Eq. 5) as it does not promote
sparsity as an L1 term would. Thus, the observed sparsity in the trained network must be a genuine
consequence of having optimized the connectivity for robust stability.
If we assume that weights |Wij | ? 0.01 correspond to functionally silent synapses, then the trained
network contains 52% of silent excitatory synapses and 46% of silent inhibitory ones (Fig. 4c). We
wondered if those weak, ?silent? synapses are necessary for stability of memory recall, or could be
removed altogether without affecting performance. To test that, we clipped those synapses {|Wij | <
0.01} to zero, and computed recall performance again (Fig. 4d). This clipping turns out to slightly
shift the position of the attractors in state space, so we increased the distance threshold that defines
a successful recall trial to 0.08. The test reveals that one of the attractors loses stability, reducing
the average performance. However the remaining 29 attractors are robust to this removal of weak
synapses and show near-equal recall performance as above. This demonstrates that small weights,
though numerous, are not necessary for competent recall performance.
Balanced state
As a result of the connection weight distributions and robust stability, the trained network produces
a regime in which excitation and inhibition balance each other, precisely tuning each neuron to
its
attractor.
as hexc
i (t) =
Pntarget frequency in each
PnExcitatory and inhibitory inputs are defined
inh
exc
bW
c
r
(t)
and
h
(t)
=
b?
W
c
r
(t)
so
that
the
difference
h
(t)
?
hinh
ij
+
j
ij
+
j
i
i
i (t)
j=1
j=1
corresponds to the total recurrent input, i.e. the second term on the r.h.s. of Eq. 3.
6
c
1
a
exc.
150
inh.
b
15
5
0
exc.
inh.
1
postsynaptic
0.5
10?4 10?2 100
weight
0.1
0
-1
-5
-15
1
success rate
-0.1
1
?10
Wij
150
presynaptic
?5
d
1
0.75
0.5
0.25
0
0
5
weight
10
clipped
full
0
0.5
1
starting distance from attr.
Figure 4: (a) Synaptic weight matrix after learning. Note the logarithmic color scale. (b) Distribution of the excitatory (red) and inhibitory (blue) weights. (c) Cumulative weight distribution of
absolute weight values. Gray line marks the 0.01 threshold we use to defined ?silent? synapses. (d)
Performance of the network after clipping the weights below 0.01 to zero (black, median with 25th
and 75th percentiles), compared to the performance of the unperturbed network redrawn from Fig. 3
(purple).
hkexc (t ), hkinh (t )
40
20
0
k =3
k = 72
k = 101
b 60
hkinh (t? )
k =3
k = 72
k = 101
a
40
k = 15
c
20
0
0
20
40 60
t (ms)
0
20
40
hkexc (t? )
60
0
0.5
1
correlation
Figure 5: (a) Dynamics of the excitatory and inhibitory inputs during a memory recall trial, for
three sample neurons. (b) Scatter plot of steady-state excitatory versus inhibitory inputs. Each dot
corresponds to a different memory pattern, and several neurons are shown in different colors. (c)
Histogram of E and I input correlations across all memories for each neuron (for example, one value
binned in this histogram would be the correlation between all green dots in b).
inh
Figure 5a shows the evolution of hexc
i (t) and hi (t) during a recall trial for one of the stored random
attractors, for 3 different neurons. Neuron 3 has rate target of 9Hz, well above average, therefore its
excitation is much higher than inhibition. Neuron 72 has a steady state firing rate of 2 Hz, below
average: its inhibitory input is greater than the excitatory one, and firing is driven by the external
current. Finally, neuron 101 is inhibitory and has a target rate 0, and indeed its inhibitory input
is large enough to overwhelm the combined effects of the external and recurrent excitatory inputs.
Notably, in all these cases, both E and I input currents are fairly large but cancel each other to leave
something smaller, either positive or negative.
Figure 5b shows the E vs. I inputs at steady-state across all the embedded attractors, for various
neurons plotted in different colors. These E and I inputs tend to be correlated across attractors for
every single neuron (dots in Fig. 5 tend to hug the identity line), with relative differences fine-tuned
to yield the desired firing rates. These across-attractors E/I correlations are summarized in Fig. 5c
as a histogram over neurons.
Robustness to ongoing noise and reduction of across-trial variability following recall onset
Finally, to probe the system under more realistic dynamics, we added time-varying, Gaussian white
noise such that, in an excitatory neuron free from network interactions, the potential would fluctuate
7
nearest
a
others
b
hstd [vi (t )]ii
d? (t )
2
1
3
2
1
0
0
0
0.2
0.4 0.6
t (s)
0.8
0
1
0.2
0.4 0.6
t (s)
0.8
1
Figure 6: (a) Normalized distance calculated according to Eq. 6 between the network activity and
each of the attractors (targeted attractor: green line; others: orange lines) during a noisy recall
episode. (b) Trial-to-trial variability, expressed as the standard deviation of a neuron?s activity across
multiple repetitions with random initial conditions. At time t = 0.5 s the network receives a pulse
in the direction of one target attractor (? = 2). Gray lines are for single neurons; the black line is
an average over the population.
with standard deviation 0.33. Figure 6a shows the momentary distance d? (t) of the network state
from the attractor closest to the initial cue (green), and for all other attractors (orange), during a
recall trial. It is clear that the system revolves around the desired attractor, performing successful
recall despite the ongoing noise. In a second experiment, we ran many trials in which the initialization at time t = 0 was random, while the same spatially patterned stimulation ? aligned onto a
chosen attractor ? is given to the network in each trial at time t = 0.5 sec. Figure 6b shows the standard deviation of the internal state of a neuron across trials, averaged across the neural population.
Following stimulus onset, neurons are always pushed towards the target attractor, and this greatly
reduces trial-by-trial variability, compared to the initial spontaneous regime in which the neurons
would fluctuate around any of the activity levels corresponding to its assigned attractors. Interestingly, such stimulus-induced variability reduction has been observed very broadly across sensory
and motor cortical areas [21]. This extends previous work, e.g. [22] and [23], showing variability
reduction in a multiple-attractor scenario with effectively binary patterns, to the case of patterns with
graded activities.
4
Discussion
We have provided a proof of concept that a model cortical networks of E and I neurons can embed
multiple analog memories as stable fixed-points of their dynamics. Memories are stable in the face
of ongoing noise and corruption of the recall cues. Neuronal activities do not saturate, and indeed,
our single-neuron model did not explicitly incorporate an upper saturation mechanism: dynamic
feedback inhibition, precisely matched to the level of excitation incurred by each attractor, ensures
that each neuron can fire at a relatively low rate during recall. As a result, excitation and inhibition
are tightly balanced.
We have used a rate-based formulation of the circuit dynamics, which raises the question of the
applicability of our method to understanding spiking memory networks. Once the connectivity
in the rate model is generated and optimized, it could still be used in a spiking model, provided
the gain function we have used here matches that of the single spiking neurons. In this respect,
the gain function we have used here is likely an appropriate choice: in physiological conditions,
cortical neurons have input-output gain functions that are well approximated by a rectified powerlaw function over their entire dynamic range [24, 25, 26].
An important question for future research is how local synaptic learning rules can achieve the stabilization objective that we have approached here from an optimal, algorithmic viewpoint. Inhibitory
synaptic plasticity is a promising candidate, as it has already been shown to enable self-regulation of
the spontaneous, baseline activity regime, and also to promote the stable storage of binary memory
patterns [27]. More work is required in this direction.
Acknowledgements. This work was supported by the Wellcome Trust (GH, ML), the European
Union Seventh Framework Programme (FP7/20072013) under grant agreement no. 269921 (BrainScaleS) (DF, ML), and the Swiss National Science Foundation (GH).
8
References
[1] Hopfield J. Neural networks and physical systems with emergent collective computational abilities, Proceedings of the national academy of sciences 79:2554, 1982.
[2] Hopfield J. Neurons with graded response have collective computational properties like those of two-state
neurons, Proceedings of the national academy of sciences 81:3088, 1984.
[3] Treves A. Graded-response neurons and information encodings in autoassociative memories, Phys. Rev.
A 42:2418, 1990.
[4] Treves A, Rolls ET. What determines the capacity of autoassociative memories in the brain?, Network:
Computation in Neural Systems 2:371, 1991.
[5] Battaglia FP, Treves A. Stable and rapid recurrent processing in realistic autoassociative memories, Neural
Comput 10:431, 1998.
[6] Lengyel M, Dayan P. Rate- and phase-coded autoassociative memory, In Advances in Neural Information
Processing Systems 17, 769, Cambridge, MA, 2005. MIT Press.
[7] Amit D, Brunel N. Dynamics of a recurrent network of spiking neurons before and following learning,
Network: Computation in Neural Systems 8:373, 1997.
[8] Latham P, Nirenberg S. Computing and stability in cortical networks, Neural computation 16:1385, 2004.
[9] Roudi Y, Latham PE. A balanced memory network, PLoS Computational Biology 3:e141, 2007.
[10] Ben-Yishai R, et al. Theory of orientation tuning in visual cortex, Proc. Natl. Acad. Sci. USA 92:3844,
1995.
[11] Goldberg JA, et al. Patterns of ongoing activity and the functional architecture of the primary visual
cortex, Neuron 42:489, 2004.
[12] Lengyel M, et al. Matching storage and recall: hippocampal spike timing?dependent plasticity and phase
response curves, Nature Neuroscience 8:1677, 2005.
[13] Vanbiervliet J, et al. The smoothed spectral abscissa for robust stability optimization, SIAM Journal on
Optimization 20:156, 2009.
[14] Hennequin G, et al. Optimal control of transient dynamics in balanced networks supports generation of
complex movements, Neuron 82:1394, 2014.
[15] Ahmadian Y, et al. Analysis of the stabilized supralinear network, Neural Comput. 25:1994, 2013.
[16] Anderson JS, et al. The contribution of noise to contrast invariance of orientation tuning in cat visual
cortex, Science 290:1968, 2000.
[17] Johnson SG. The NLopt nonlinear-optimization package, http://ab-initio.mit.edu/nlopt .
[18] Roxin A, et al. On the distribution of firing rates in networks of cortical neurons, The Journal of Neuroscience 31:16217, 2011.
[19] Song S, et al. Highly nonrandom features of synaptic connectivity in local cortical circuits, PLoS Biol 3:
e68, 2005.
[20] Lefort S, et al. The excitatory neuronal network of the C2 barrel column in mouse primary somatosensory
cortex, Neuron 61:301 , 2009.
[21] Churchland MM, et al. Stimulus onset quenches neural variability: a widespread cortical phenomenon,
Nat Neurosci 13:369, 2010.
[22] Litwin-Kumar A, Doiron B. Slow dynamics and high variability in balanced cortical networks with clustered connections, Nat Neurosci 15:1498, 2012.
[23] Deco G, Hugues E. Neural network mechanisms underlying stimulus driven variability reduction, PLoS
computational biology 8:e1002395, 2012.
[24] Priebe NJ, Ferster D. Direction selectivity of excitation and inhibition in simple cells of the cat primary
visual cortex, Neuron 45:133, 2005.
[25] Priebe NJ, Ferster D. Mechanisms underlying cross-orientation suppression in cat visual cortex, Nat Neurosci 9:552, 2006.
[26] Finn IM, et al. The emergence of contrast-invariant orientation tuning in simple cells of cat visual cortex,
Neuron 54:137, 2007.
[27] Vogels TP, et al. Inhibitory plasticity balances excitation and inhibition in sensory pathways and memory
networks, Science 334:1569, 2011.
9
| 5336 |@word trial:23 version:2 dtk:1 norm:3 open:2 heuristically:1 seek:2 pulse:1 eng:1 solid:2 reduction:6 initial:10 contains:1 efficacy:1 selecting:1 tuned:1 interestingly:1 current:2 scatter:1 must:2 subsequent:2 realistic:5 plasticity:3 shape:1 motor:2 plot:1 v:1 implying:2 cue:6 parameterization:1 accordingly:1 beginning:1 short:1 provides:1 contribute:1 unbounded:1 c2:1 direct:2 become:3 persistent:2 pathway:1 notably:1 indeed:2 expected:1 roughly:1 abscissa:5 themselves:1 growing:1 rapid:1 brain:2 behavior:2 actual:1 hugues:1 abound:1 becomes:1 provided:2 moreover:2 matched:1 circuit:5 panel:1 barrel:1 underlying:2 what:1 developed:1 unified:1 nj:2 guarantee:1 temporal:2 nonrandom:1 every:3 exactly:1 wrong:1 scaled:1 k2:2 uk:4 control:7 unit:1 whatever:1 grant:1 demonstrates:1 positive:2 before:1 engineering:1 timing:2 local:4 limit:1 consequence:1 acad:1 despite:4 encoding:1 meet:1 firing:20 might:2 black:2 initialization:2 therein:1 ease:1 limited:1 revolves:1 patterned:1 range:2 averaged:1 practice:3 union:1 swiss:1 procedure:1 area:2 empirical:1 thought:1 matching:2 word:1 induce:1 subpopulation:2 pre:1 onto:2 close:1 cannot:1 storage:8 context:1 impossible:1 optimize:4 straightforward:1 starting:1 duration:1 wondered:1 lumping:1 powerlaw:1 attr:1 rule:2 stereotypical:1 attraction:1 hennequin:2 regularize:1 biol:1 reparameterization:1 population:4 stability:19 target:18 construction:1 spontaneous:2 exact:1 substrate:2 us:1 goldberg:1 agreement:1 element:4 velocity:2 approximated:1 observed:7 capture:1 ensures:1 connected:2 episode:1 plo:3 decrease:1 removed:1 movement:1 ran:1 balanced:7 ideally:1 cam:3 dynamic:31 signature:1 trained:3 depend:2 tight:2 raise:1 compromise:1 nlopt:3 churchland:1 completely:1 easily:1 joint:1 hopfield:3 emergent:1 represented:1 cat:5 various:1 regularizer:1 distinct:2 fast:1 ahmadian:1 approached:1 encoded:2 heuristic:1 supplementary:1 otherwise:2 nirenberg:1 ability:1 statistic:1 think:1 emergence:1 noisy:3 final:1 eigenvalue:1 differentiable:1 indication:1 propose:2 interaction:3 jij:2 relevant:1 loop:1 realization:1 aligned:1 achieve:2 academy:2 frobenius:1 convergence:2 produce:2 incidentally:1 converges:1 leave:1 ben:1 recurrent:8 ac:3 measured:1 nearest:1 ij:8 sa:10 eq:11 strong:3 auxiliary:3 implies:1 somatosensory:1 convention:1 direction:3 thick:2 correct:1 subsequently:1 redrawn:1 stabilization:1 enable:1 runaway:1 settle:1 violates:1 material:1 transient:1 ja:1 fix:1 clustered:1 biological:2 im:1 strictly:1 initio:1 lying:1 around:3 considered:1 mm:1 normal:5 exp:1 algorithmic:1 bump:1 circuitry:1 achieves:1 battaglia:1 proc:1 sensitive:1 largest:1 repetition:1 create:1 autapses:1 minimization:2 mit:2 always:3 gaussian:1 rather:1 pn:1 fluctuate:2 varying:1 derived:1 hk:1 greatly:1 contrast:2 suppression:1 baseline:11 dayan:1 dependent:1 entire:2 initially:2 wij:8 going:1 ill:1 orientation:4 spatial:2 lif:2 orange:7 initialize:1 fairly:1 field:1 equal:2 genuine:1 having:1 once:1 biology:2 kdv:2 cancel:1 thin:1 promote:2 future:1 others:3 stimulus:5 haven:1 few:5 randomly:1 gamma:1 tightly:3 national:3 phase:2 ourselves:1 bw:1 fire:3 attractor:36 maintain:1 ab:1 possibility:1 highly:1 adjust:2 yielding:1 natl:1 yishai:1 amenable:1 closer:2 partial:1 necessary:2 iv:1 euclidean:2 initialized:3 desired:9 re:1 plotted:1 theoretical:3 minimal:1 increased:1 column:1 earlier:1 tp:1 clipping:2 cost:2 applicability:1 deviation:6 rare:1 uniform:2 successful:6 seventh:1 johnson:1 stored:3 e68:1 corrupted:1 calibrated:1 combined:1 fundamental:1 peak:1 siam:1 together:1 mouse:1 connectivity:8 again:3 squared:4 deco:1 admit:1 external:3 derivative:1 return:1 account:1 potential:7 converted:1 bfgs:1 summarized:1 stabilize:2 sec:1 satisfy:4 explicitly:2 mv:1 onset:4 vi:10 depends:3 later:1 view:1 try:1 lab:1 observer:6 red:1 relied:1 vivo:1 vinh:8 contribution:1 minimize:2 purple:2 ni:3 roll:1 variance:1 characteristic:1 yield:2 correspond:2 weak:2 fallen:1 critically:1 worth:1 drive:1 corruption:2 lengyel:4 rectified:1 converged:1 explain:1 synapsis:6 reach:1 phys:1 synaptic:19 realise:1 frequency:1 involved:1 naturally:1 proof:1 gain:5 sampled:1 proved:1 recall:32 lim:1 color:5 formalize:1 appears:1 higher:1 dt:4 sustain:2 response:4 synapse:1 formulation:1 though:3 anderson:1 just:2 correlation:4 receives:1 trust:1 nonlinear:3 widespread:1 defines:1 brings:1 gray:2 behaved:1 vogels:1 grows:1 usa:1 effect:1 contain:1 normalized:3 functioning:1 concept:1 evolution:4 assigned:1 spatially:2 eg:1 white:1 skewed:1 self:1 width:1 during:7 excitation:10 steady:4 percentile:2 criterion:2 m:3 trying:1 hippocampal:1 theoretic:2 latham:2 performs:1 l1:1 reflection:1 gh:2 instantaneous:2 novel:2 recently:2 stimulation:1 spiking:4 physical:1 regulates:1 functional:1 analog:7 tail:1 numerically:1 functionally:1 cambridge:2 smoothness:2 tuning:4 unconstrained:1 had:1 dot:3 stable:13 access:1 cortex:8 longer:1 inhibition:9 behaving:1 something:1 j:1 closest:3 recent:1 roudi:1 optimizing:1 driven:2 scenario:1 route:1 store:1 occasionally:1 selectivity:1 binary:7 success:2 ubiquitously:1 meeting:1 seen:1 captured:1 additional:2 minimum:1 greater:1 converge:1 dashed:2 ii:2 violate:1 full:2 multiple:3 reduces:2 smooth:1 match:1 cross:2 long:2 retrieval:2 concerning:1 post:1 coded:1 variant:1 df:1 histogram:3 sometimes:1 achieved:1 cell:3 proposal:1 affecting:1 remarkably:1 fine:2 separately:1 median:2 source:1 operate:1 subject:1 hz:11 tend:3 kwkf:1 induced:1 near:3 presence:1 noting:1 ideal:7 iii:1 enough:3 architecture:1 restrict:1 perfectly:1 silent:5 shift:1 t0:1 penalty:1 song:1 speaking:1 constitute:1 action:1 autoassociative:6 detailed:1 listed:1 clear:1 amount:1 ten:2 generate:1 http:1 inhibitory:25 stabilized:2 dotted:2 sign:1 neuroscience:3 delta:1 estimated:1 blue:1 broadly:1 group:1 key:1 four:1 threshold:3 drawn:3 clean:1 kept:1 v1:2 asymptotically:1 relaxation:1 fraction:1 sum:1 run:2 package:1 clipped:2 extends:1 vn:1 separation:1 home:1 comparable:1 pushed:1 bound:2 hi:4 activity:27 binned:1 constraint:5 kronecker:1 precisely:2 ri:2 prescribed:1 span:2 kumar:1 performing:1 relatively:2 format:1 department:1 according:2 combination:1 membrane:1 remain:1 across:12 increasingly:1 slightly:1 postsynaptic:1 smaller:1 rev:1 modification:1 making:1 lasting:1 dv:3 invariant:2 gradually:1 flanking:2 wellcome:1 equation:1 visualization:1 remains:1 overwhelm:1 turn:3 eventually:1 mechanism:4 fail:1 fp7:1 finn:1 parametrize:1 wii:1 rewritten:1 probe:1 obey:1 away:2 spectral:5 enforce:1 appropriate:1 lefort:1 robustly:1 robustness:2 altogether:1 original:1 top:1 denotes:1 remaining:2 cf:2 perturb:1 build:2 graded:10 prof:1 especially:1 amit:1 objective:4 intend:1 initializes:1 added:2 spike:2 question:2 already:1 primary:3 unclear:1 exhibit:1 gradient:5 distance:11 sci:1 capacity:4 exc:9 presynaptic:2 assuming:1 balance:5 minimizing:2 difficult:3 regulation:1 negative:3 priebe:2 collective:3 perform:2 upper:4 neuron:61 descent:2 variability:9 incorporated:1 inh:6 perturbation:1 smoothed:3 arbitrary:2 drift:2 treves:3 introduced:1 inverting:1 namely:2 required:1 connection:4 optimized:5 learned:1 hug:2 hour:2 able:2 dynamical:1 pattern:32 below:5 regime:4 sparsity:2 summarize:1 fp:1 saturation:3 including:2 memory:55 green:3 suitable:1 force:1 imply:1 ne:3 library:1 numerous:1 axis:2 created:1 deemed:1 text:1 review:1 sg:1 understanding:3 literature:1 l2:2 prior:1 removal:1 relative:2 law:6 embedded:2 acknowledgement:1 generation:1 limitation:1 versus:1 foundation:1 integrate:1 incurred:1 basin:1 sufficient:1 principle:1 dvi:1 viewpoint:1 heavy:1 row:1 excitatory:19 accounted:1 supported:1 free:4 allow:3 fall:2 face:2 lognormal:1 absolute:1 sparse:2 leaky:1 feedback:5 dimension:1 cortical:11 valid:1 cumulative:1 calculated:1 curve:1 dale:6 sensory:3 forward:1 collection:1 vexc:5 programme:1 sj:2 supralinear:1 implicitly:1 overcomes:1 keep:1 ml:2 global:3 reveals:1 summing:1 symmetrical:1 spectrum:1 continuous:1 decade:1 quantifies:1 tailed:1 table:4 quenches:1 promising:1 nature:3 reasonably:1 robust:10 transfer:1 european:1 complex:1 constructing:1 vj:3 did:3 linearly:1 neurosci:3 kwk2f:1 noise:15 ref:1 competent:1 neuronal:7 fig:14 slow:2 structurally:1 sub:1 momentary:3 wish:1 position:1 doiron:1 dylan:1 candidate:1 governed:1 comput:2 pe:1 jacobian:3 saturate:1 embed:3 specific:1 showing:1 unperturbed:1 physiological:3 ssa:12 incorporating:1 adding:1 effectively:3 importance:1 nat:3 sparseness:1 logarithmic:2 simply:2 likely:1 visual:6 expressed:2 saturating:1 scalar:1 brunel:1 corresponds:2 loses:1 satisfies:1 determines:1 ma:1 identity:1 targeted:1 towards:2 ferster:2 brainscales:1 experimentally:1 included:2 specifically:1 reducing:1 total:2 called:1 invariance:1 tendency:1 indicating:1 formally:2 guillaume:1 internal:1 support:2 mark:2 ongoing:7 incorporate:1 evaluate:1 phenomenon:1 correlated:1 |
4,790 | 5,337 | Feedforward Learning of Mixture Models
Matthew Lawlor?
Applied Math
Yale University
New Haven, CT 06520
[email protected]
Steven W. Zucker
Computer Science
Yale University
New Haven, CT 06520
[email protected]
Abstract
We develop a biologically-plausible learning rule that provably converges to the
class means of general mixture models. This rule generalizes the classical BCM
neural rule within a tensor framework, substantially increasing the generality of
the learning problem it solves. It achieves this by incorporating triplets of samples
from the mixtures, which provides a novel information processing interpretation
to spike-timing-dependent plasticity. We provide both proofs of convergence, and
a close fit to experimental data on STDP.
1
Introduction
Spectral tensor methods and tensor decomposition are emerging themes in machine learning, but
they remain global rather than ?online.? While incremental (online) learning can be useful for applications, it is essential for neurobiology. Error back propagation does operate incrementally, but
its neurobiological relevance remains a question for debate. We introduce a triplet learning rule
for mixture distributions based on a tensor formulation of the BCM biological learning rule. It is
implemented in a feedforward fashion, removing the need for backpropagation of error signals.
The triplet requirement is natural biologically. Informally imagine your eyes microsaccading during
a fixation, so that a tiny image fragment is ?sampled? repeatedly until the next fixation. Viewed
from visual cortex, edge selective neurons will fire repeatedly. Importantly, they exhibit strong
statistical dependencies due to the geometry of objects and their relationships in the world. ?Hidden?
information such as edge curvatures, the presence of textures, and lighting discontinuities all affect
the probability distribution of firing rates among orientation selective neurons, leading to complex
statistical interdependencies between neurons.
Latent variable models are powerful tools in this context. They formalize the idea that highly coupled
random variables can be simply explained by a small number of hidden causes. Conditioned on these
causes, the input distribution should be simple. For example, while the joint distribution of edges in
a small patch of a scene might be quite complex, the distribution conditioned on the presence of a
curved object at a particular location might be comparatively simple [14]. The specific question is
whether brains can learn these mixture models, and how.
Example: Imagine a stimulus space of K inputs. These could be images of edges at particular
orientations, or audio tones at K frequencies. These stimuli are fed into a network of n LinearNonlinear Poisson (LNP) spiking neurons. Let rij denote the firing rate of neuron i to stimulus j.
Assuming the stimuli are drawn independently with probability ?k , then the number of spikes d in
an interval where a single stimulus is shown is distributed according to a mixture model.
X
P (d) =
?k Pk (d)
k
1
Now at Google Inc.
1
where Pk (d) is a vector of independent Poisson distributions, and the rate parameter of the ith
component is rik . We seek a filter that responds (in expectation) to one and only one stimulus. To
do this, we must learn a set of weights that are orthogonal to all but one of the vectors of rates r ?j .
Each rate vector corresponds to the mean of one of the mixtures. Our problem is thus to learn the
means of mixtures. We will demonstrate that this can be done non-parametrically over a broad class
of firing patterns, not just Poisson spiking neurons.
Although fitting mixture models can be exponentially hard, under a certain multiview assumption,
non-parametric estimation of mixture means can be done by tensor decomposition [2][1]. This
multiview assumption requires access to at least 3 independent copies of the samples; i.e., multiple
samples drawn from the same mixture component. For the LNP example above, this multiview
assumption requires only that we have access to the number of spikes in three disjoint intervals,
while the stimulus remains constant. After these intervals, the stimulus is free to change ? in vision,
say, after a saccade ? after which point another sample triple is taken.
Our main result is that, with a slight modification of classical Bienenstock-Cooper-Munro [5] synaptic update rule a neuron can perform a tensor decomposition of the input data. By incorporating the
interactions between input triplets, our online learning rule can provably learn the mixture means
under an extremely broad class of mixture distributions and noise models. (The classical BCM
learning rule will not converge properly in the presence of noise.) Specifically we show how the
classical BCM neuron performs gradient ascent in a tensor objective function, when the data consists of discrete input vectors, and how our modified rule converges when the data are drawn from a
general mixture model.
The multiview requirement has an intriguing implication for neuroscience. Since spikes arrive in
waves, and spike trains matter for learning [9], our model suggests that the waves of spikes arriving during adjacent epochs in time provide multiple samples of a given stimulus. This provides an
unusual information processing interpretation for the functional role of spike trains. To realize it
fully, we point out that classical BCM can be implemented via spike timing dependent plasticity
[17][10][6][18]. However, most of these approaches require much stronger distributional assumptions on the input data (generally Poisson), or learn a much simpler decomposition of the data than
our algorithm. Other, Bayesian methods [16], require the computation of a posterior distribution
which requires an implausible normalization step. Our learning rule successfully avoids these issues, and has provable guarantees of convergence to the true mixture means. At the end of this
paper we show how our rule predicts pair and triple spike timing dependent plasticity data.
2
Tensor Notation
Let ? denote the tensor product. We denote application of a k-tensor to k vectors by T (w1 , ..., wk ),
so in the simple case where T = v 1 ? ... ? v k ,
Y
T (w1 , ..., wk ) =
hv j , wj i
j
We further denote the application of a k-tensor to k matrices by T (M1 , ..., Mk ) where
X
T (M1 , ..., Mk )i1 ,...,ik =
Tj1 ,...,jk [M1 ]j1 ,i1 ...[Mk ]jk ,ik
j1 ,...,jk
Thus if T is a symmetric 2-tensor, T (M1 , M2 ) = M1T T M2 with ordinary matrix multiplication.
Similarly, T (v 1 , v 2 ) = v T1 T v 2 .
We say that T has an orthogonal tensor decomposition if
X
T =
?k v k ? v k ? ... ? v k and hv i , v j i = ?ij
k
3
Connection Between BCM Neuron and Tensor Decompositions
The BCM learning rule was introduced in 1982 in part to correct failings of the classical Hebbian
learning rule [5]. The Hebbian learning rule [11] is one of the simplest and oldest learning rules. It
2
posits that the selectivity of a neuron to input i, mt (i) is increased in proportion to the post-synaptic
activity of that neuron ct = hmt?1 , dt i, where m is a vector of synaptic weights.
mt ? mt?1 = ?t ct dt
This learning rule will become increasingly correlated with its input. As formulated this rule does
not converge for most input, as kmk ? ?. In addition, in the presence of multiple inputs Hebbian
learning rule will always converge to an ?average? of the inputs, rather than becoming selective to
one particular input. It is possible to choose a normalization of m such that m will converge to
the first eigenvector of the input data. The BCM rule tries to correct for the lack of selectivity, and
for the stabilization problems. Like the Hebbian learning rule, it always updates its weights in the
direction of the input, however it also has a sliding threshold that controls the magnitude and sign of
this weight update.
The original formulation of the BCM rule is as follows: Let c be the post-synaptic firing rate, d ? RN
be the vector of presynaptic firing rates, and m be the vector of synaptic weights. Then the BCM
synaptic modification rule is
c = hm, di
? = ?(c, ?)d
m
? is a non-linear function of the firing rate, and ? is a sliding threshold that increases as a superlinear
function of the average firing rate.
There are many different formulations of the BCM rule. The primary features that are required are
?(c, ?) is convex in c, ?(0, ?) = 0, ?(?, ?) = 0, and ? is a super-linear function of E[c].
These properties guarantee that the BCM learning rule will not grow without bound. There have
been many variants of this rule. One of the most theoretically well analyzed variants is the Intrator
and Cooper model [12], which has the following form for ? and ?.
?(c, ?) = c(c ? ?) with ? = E[c2 ]
Definition 3.1 (BCM Update Rule). With the Intrator and Cooper definition, the BCM rule is defined
as
mt = mt?1 + ?t ct (ct ? ?t?1 )dt
(1)
2
where
ct = hmt?1
P
P, dt i2 and ? = E[c ]. ?t is a sequence of positive step sizes with the property that
?
?
?
and
<
?
?
t t
t t
The traditional application of this rule is a system where the input d is drawn from linearly independent vectors {d1 , ..., dk } with probabilities ?1 , ..., ?K , with K = N , the dimension of the
space.
These choices are quite convenient because they lead to the following objective function formulation
of the synaptic update rule.
R(m) =
i 1 h
i2
1 h
3
2
E hm, di ? E hm, di
3
4
Thus,
h
i
2
2
?R = E hm, di d ? E[hm, di ] hm, di d
= E[c(c ? ?)d]
= E[?(c, ?)d]
So in expectation, the BCM rule performs a gradient ascent in R(m). For random, discrete input
this rule would then be a form of stochastic gradient ascent.
With this model, we observe that the objective function can be rewritten in tensor notation. Note
that this input model can be seen as a kind of degenerate mixture model.
3
This objective function can be written as a tensor objective function, by noting the following:
X
T =
?k dk ? dk ? dk
k
M=
X
k
R(m) =
?k dk ? dk
1
1
T (m, m, m) ? M (m, m)2
3
4
(2)
For completeness, we present a proof that the stable points of the expected BCM update are selective
for only one of the data vectors.
? = 0. Let ck = hm, dk i and ?k =
The stable points of the expected update occur when E[m]
?(ck , ?). Let c = [c1 , . . . , cK ]T and ? = [?1 , . . . , ?K ]T .
DT = [d1 | ? ? ? |dK ]
P = diag(?)
Theorem 3.2. (Intrator 1992) Let K = N , let each dk be linearly independent, and let ?k > 0 and
? = ?R occur when
distinct. Then stable points (in the sense of Lyapunov) of the expected update m
c = ?k?1 ek or m = ?k?1 D?1 ek . ek is the unit basis vector, so there is activity for only one stimuli.
P
? = DT P ? which is 0 only when ? = 0. Note ? = k ?k c2k . ?k = 0 if ck = 0 or
Proof. E[m]
ck = ?. Let S+ = {k : ck 6= 0}, and S? = {k : ck = 0}. Then for all k ? S+ , ck = ?S+
?
??1
X
X
?S+ ? ?S2 +
?i = 0
?S+ = ?
?i ?
k?S+
k?S+
Therefore the solutions of the BCM learning rule are c = 1S+ ?S+ , for all subsets S+ ? {1, . . . , K}.
We now need to check which solutions are stable. The stable points (in the sense of Lyapunov) are
points where the matrix
?
?? ?c
??
?E[m]
T
T
=D P
=D P
D
H=
?m
?c ?m
?c
is negative semidefinite.
Let S be an index set S ? {1, . . . , n}. We will use the following notation for the diagonal matrix
IS :
1 i?S
(IS )ii =
(3)
0 i?
/S
So IS + IS c = I, and ei eTi = I{i}
a quick calculation shows
??i
= ?S+ IS+ ? ?S+ IS? ? 2?S2 + diag(?) 1S+ 1TS+
?cj
This is negative semidefinite iff A = IS+ ? 2?S+ diag(?) 1S+ 1TS+ is negative semidefinite.
Assuming a non-degeneracy of the probabilities ?, and assume |S+ | > 1. Let j = arg mink?S+ ?k .
Then ?S+ ?j < 12 so A is not negative semi-definite. However, if |S+ | = 1 then A = ?IS+ so the
stable points occur when c = ?1i ei
The triplet version of BCM can be viewed as a modification of the classical BCM rule which allows
it to converge in the presence of zero-mean noise. This indicates that the stable solutions of this
learning rule are selective for only one data vector, dk .
Building off of the work of [2] we will use this characterization of the objective function to build a
triplet BCM update rule which will converge for general mixtures, not just discrete data points.
4
hm1,di
Noise sensitivity of m after 10e6 steps
14
12
10
Triplet Rule
BCM
3
8
6
d1
4
2
0
?10
?5
0
5
10
15
20
km ? m0 k
2
m1
25
hm2,di
m2
22
1
20
18
d2
16
14
12
10
0
8
6
10?2
4
2
0
?10
?5
0
5
10
15
20
100
101
Noise ?
(a) Geometry of stable solutions. Each stable
solution is selective in expectation for a single
mixture. Note that the classical BCM rule will
not converge to these values in the presence of
noise.
4
10?1
25
(b) Noise response of triplet BCM update
rule vs BCM update. Input data was a mixture of Gaussians with standard deviation ?.
The selectivity of the triplet BCM rule remains unchanged in the presence of noise.
Triplet BCM Rule
We now show that by modifying the update rule to incorporate information from triplets of input
vectors, the generality of the input data can be dramatically increased. Our new BCM rule will learn
selectivity for arbitrary mixture distributions, and learn weights which in expectation are selective
for only one mixture component. Assume that
X
P (d) =
?k Pk (d)
k
where EPk [d] = dk . For example, the data could be a mixture of axis-aligned Gaussians, a mixture
of independent Poisson variables, or mixtures of independent Bernoulli random variables to name a
few. We also require EPk [kdk2 ] < ?. We emphasize that we do not require our data to come from
any parametric distribution.
We interpret k to be a latent variable that signals the hidden cause of the underlying input distribution, with distribution Pk . Critically, we assume that the hidden variable k changes slowly compared
to the inter-spike period of the neuron. In particular, we need at least 3 samples from each Pk . This
corresponds to the multi-view assumption of [2]. A particularly relevant model meeting this assumption is that of spike counts in disjoint intervals under a Poisson process, with a discrete, time
varying rate parameter. For the purpose of this paper, we assume the number of mixed distributions,
k, is equal to the number of dimensions, n, however it is possible to relax this to k < n.
Let {d1 , d2 , d3 } be a triplet of independent copies from some Pk (d), i.e. each are drawn from
the same mixture. It is critical to note that if {d1 , d2 , d3 } are not drawn from the same class, this
update will not converge to the global maximum. Numerical experiments show this assumption can
be violated somewhat without severe changes to the fixed points of the algorithm. Our sample
i
is
i
then a sequence of triplets, each triplet drawn from the same latent distribution. Let c = d , m .
With these independent triples, we note that the tensors T and M from equation (2) can be written
as moments of the independent triplets
T = E[d1 ? d2 ? d3 ]
M = E[d1 ? d2 ]
1
1
R(m) = T (m, m, m) ? M (m, m)2
3
4
This is precisely the same objective function optimized by the classical BCM update, with the conditional means of the mixture distributions taking the place of discrete data points. With access to
independent triplets, selectivity for significantly richer input distributions can be learned.
5
As with classical BCM, we can perform gradient ascent in this objective function which leads to the
expected update
E[?R] = E[c1 c2 d3 + (c1 d2 + c2 d1 )(c3 ? 2?)]
where ? = E[c1 c2 ]. This update is rather complicated, and couples pre and post synaptic firing rates
across multiple time intervals. Since each ci and di are identically distributed, this expectation is
equal to
E[c2 (c3 ? ?)d1 ]
which suggests a much simpler update. This ordering was chosen to match the spike timing dependency of synaptic modification. This update depends on the presynaptic input, and the postsynaptic
excitation in two disjoint time periods.
Definition 4.1 (Full-rank Triplet BCM). The full-rank Triplet BCM update rule is:
mt = ?(mt?1 + ?t ?(c2 , c3 , ?t?1 )d1 )
(4)
P
P 2
2 3
2 3
where ?(c , c , ?) = c (c ? ?), the step size ?t obeys t ?t ? ?, and t ?t < ?. ? is a
projection into an arbitrary large compact ball, which is needed for technical reasons to guarantee
convergence.
5
Stochastic Approximation
Having found the stable points of the expected update for BCM and triplet BCM, we now turn to
a proof of convergence for the stochastic update generated by mixture models. For this, we turn to
results from the theory of stochastic approximation.
We will decompose our update into two parts, the expected update, and the (random) deviation.
This deviation will be a L2 bounded martingale, while the expected update will be a ODE with the
previously calculated stable points. Since the expected update is the gradient of a objective function
R, the Lyapunov functions required for the stability analysis are simply this objective function.
The decomposition of the triplet BCM stochastic process is as follows:
mt ? mt?1 = ?t ?(c2t , c3t , ?t?1 )d1
= ?n E[?(c2 , c3 , ?t?1 )d1 ] + ?n ?(c2 , c3 , ?t?1 )d1 ? E[?(c2 , c3 , ?t?1 )d1 ]
= ?t h(mt ) ? ?t ?t
Here, h(mt ) is the deterministic expected update, and ?t is a martingale. All our expectations are
taken with respect to triplets of input data. The decomposition for classical BCM is similar.
This is the Doob decomposition [8] of the sequence. Using a theorem of Delyon [7], we will show
that our triplet BCM algorithm will converge with probability 1 to the stable points of the expected
update. As was shown previously, these stable points are selective for one and only one mixture
component in expectation.
Theorem 5.1. For the full rank case, the projected update converges w.p. 1 to the zeros of ??
Proof. See supplementary material, or an extended discussion in a forthcoming arXiv preprint [13].
6
Triplet BCM Explains STDP Up to Spike Triplets
Biophysically synaptic efficiency in the brain is more closely modeled by spike timing dependent
plasticity (STDP). It depends precisely on the interval between pre- and post-synaptic spikes. Initial
research on spike pairs [15, 3] showed that a presynaptic spike followed in close succession by
a postsynaptic spike tended to strengthen a synapse, while the reverse timing weakened it. Later
work on natural spike chains [9], triplets of spikes [4, 19], and quadruplets have shown interaction
effects beyond pairs. Most closely to ours, recent work by Pfister and Gerstner [17] suggested that
a synaptic modification function depending only on spike triplets is sufficient to explain all current
experimental data. Furthermore, their rule resembles a BCM learning rule when the pre- and postsynaptic firing distributions are independent Poisson.
6
We now demonstrate that our learning rule can model both the pairwise and triplet results from
Pfister and Gerstner using a smaller number of free parameters and without the introduction of
hidden leaky timing variables. Instead, we work directly with the pre- and post-synaptic voltages,
and model the natural voltage decay during the falling phase of an action potential. Our (four)
free variables are the voltage decay, which we set within reasonable biological limits; a bin width,
controlling the distance between spiking triplet periods; ?, our sliding voltage threshold; and an
overall multiplicative constant. We emphasize that our model was not designed to fit these data; it
was designed to learn selectivity for the multi-view mixture task. Spike timing dependence falls out
as a natural consequence of our multi-view assumption.
Change in EPSC Amplitude (%)
100
50
0
?50
?100
?80
?60
?40
?20
0
20
Spike Timing (ms)
40
60
80
100
Figure 2: Fit of triplet BCM learning rule to synaptic strength STDP curve from [3]. Data points
were recreated from [3] . Spike timing measures the time between post synaptic and presynaptic
spikes, tpost ? tpre . A positive time means the presynaptic spike was followed by a postsynaptic
spike.
We first model hippocampus data from Mu-ming Poo [3], who applied repeated electrical stimulation to the pre- and post-synaptic neurons in a pairing protocol within which the relative timing of
the two spike chains was varied. After repeated stimulation at a fixed timing offset, the change in
synaptic strength (postsynaptic current) was measured.
We take the average voltage in triplet intervals to be the measure of pre- and post-synaptic activity,
and consider a one-dimensional version of our synaptic update:
?m = Ac2 (c3 ? ?)d1
(5)
2
3
where c and c are the postsynaptic voltage averaged over the second and third time bins, and d1
is the presynaptic voltage averaged over the first time bin. We assume our pre and post synaptic
voltages are governed by the differential equation:
dV
= ?? V
(6)
dt
such that, if t = sk where sk is the kth spike, V (t) ? 1. That is, the voltage is set to 1 at each spike
time before decaying again.
Let Vpre be the presynaptic voltage trace, and Vpost be the postsynaptic voltage trace. They are
determined by the timing of pre- and post-synaptic spikes, which occur at r1 , r2 , . . . , rn for the
presynaptic spikes, and o1 , o2 , . . . om for the postsynaptic spikes.
To model the pairwise experiments, we let ri = r0 + iT where T = 1000ms, a large time constant.
Then oi = ri + ?t where ?t is the spike timing. Let ?b be the size of the bins. That is to say,
Z t+ ?2b
Z t+ ?2b
1
0
0
2
d (t) =
Vpre (t + ?b )dt
c (t) =
Vpost (t0 )dt0
?b
2
?
t+ 2b
t?
c3 (t) =
Z
t?
?b
2
t?
Vpost (t0 ? ?b )dt0
?b
2
Vpost (t) = Vpre (t ? ?t )
Then the overall synaptic modification is given by
Z
Ac2 (t)(c3 (t) ? ?)d1 (t)dt
t
7
We fit A, ? , ?, and the bin size of integration. Recall that the sliding threshold, ? is a function of the
expected firing rate of the neuron. Therefore we would not expect it to be a fixed constant. Instead,
it should vary slowly over a time period much longer than the data sampling period. For the purpose
of these experiments it would be at an unknown level that depends on the history of neural activity.
See figure 2 for the fit for Mu-ming Poo?s synaptic modification data.
Froemke and Dan also investigated higher order spike chains, and found that two spikes in short
succession did not simply multiply in their effects. This would be the expected result if the spike
timing dependence treated each pair in a triplet as an independent event. Instead, they found that a
presynaptic spike followed by two postsynaptic spikes resulted in significantly less excitation than
expected if the two pairs were treated as independent events. They posited that repeated spikes
interacted suppressively, and fit a model based on that hypothesis. They performed two triplet experiments with pre- pre-post triplets, and pre-post-post triplets. Results of their experiment along
with the predictions based on our model are presented in figure 3.
Figure 3: Measured excitation and inhibition for spike triplets from Froemke and Dan are demarcated in circles and triangles. A red circle or triangle indicates excitation, while a blue circle or
triangle indicates inhibition. The predicted results from our model are indicated by the background
color. Numerical results for our model, with boundaries for the Froemke and Dan model are reproduced.
Left figure is pairs of presynaptic spikes, and a single post-synaptic spike. The right figure is pairs of
postsynaptic spikes, and a presynaptic spike. For each figure, t1 measures the time between the first
paired spike with the singleton spike, with the convention that each t is positive if the presynaptic
spike happens before the post synaptic spike. See paired STDP experiments for our spiking model.
For the top figure, ? = .65, our bin width was 2ms, and our spike voltage decay rate ? = 8ms. For
the right figure ? = .45. Red is excitatory, blue is inhibitory, white is no modification. A positive t
indicates a presynaptic spike occurred before a postsynaptic spike.
7
Conclusion
We introduced a modified formulation of the classical BCM neural update rule. This update rule
drives the synaptic weights toward the components of a tensor decomposition of the input data.
By further modifying the update to incorporate information from triplets of input data, this tensor decomposition can learn the mixture means for a broad class of mixture distributions. Unlike
other methods to fit mixture models, we incorporate a multiview assumption that allows us to learn
asymptotically exact mixture means, rather than local maxima of a similarity measure. This is in
stark contrast to EM and other gradient ascent based methods, which have limited guarantees about
the quality of their results. Conceptually our model suggests a different view of spike waves during
adjacent time epochs: they provide multiple independent samples of the presynaptic ?image.?
Due to size constraints, this abstract has has skipped some details, particularly in the experimental
sections. More detailed explanations will be provided in future publications.
Research supported by NSF, NIH, The Paul Allen Foundation, and The Simons Foundation.
8
References
[1] Animashree Anandkumar, Dean P Foster, Daniel Hsu, Sham M Kakade, and Yi-Kai Liu. Two
svds suffice: Spectral decompositions for probabilistic topic modeling and latent dirichlet allocation. CoRR, abs/1204.6703, 1, 2012.
[2] Animashree Anandkumar, Rong Ge, Daniel Hsu, Sham M Kakade, and Matus Telgarsky. Tensor decompositions for learning latent variable models. arXiv preprint arXiv:1210.7559, 2012.
[3] Guo-qiang Bi and Mu-ming Poo. Synaptic modifications in cultured hippocampal neurons:
dependence on spike timing, synaptic strength, and postsynaptic cell type. The Journal of
Neuroscience, 18(24):10464?10472, 1998.
[4] Guo-Qiang Bi and Huai-Xing Wang. Temporal asymmetry in spike timing-dependent synaptic
plasticity. Physiology & behavior, 77(4):551?555, 2002.
[5] Elie L Bienenstock, Leon N Cooper, and Paul W Munro. Theory for the development of neuron
selectivity: orientation specificity and binocular interaction in visual cortex. The Journal of
Neuroscience, 2(1):32?48, 1982.
[6] Natalia Caporale and Yang Dan. Spike timing-dependent plasticity: a hebbian learning rule.
Annual Review Neuroscience, 31:25?46, 2008.
[7] Bernard Delyon. General results on the convergence of stochastic algorithms. Automatic
Control, IEEE Transactions on, 41(9):1245?1255, 1996.
[8] Joseph L Doob. Stochastic processes, volume 101. New York Wiley, 1953.
[9] Robert C Froemke and Yang Dan. Spike-timing-dependent synaptic modification induced by
natural spike trains. Nature, 416(6879):433?438, 2002.
[10] Julijana Gjorgjieva, Claudia Clopath, Juliette Audet, and Jean-Pascal Pfister. A triplet
spike-timing?dependent plasticity model generalizes the bienenstock?cooper?munro rule to
higher-order spatiotemporal correlations. Proceedings of the National Academy of Sciences,
108(48):19383?19388, 2011.
[11] DO Hebb. The organization of behavior; a neuropsychological theory. 1949.
[12] Nathan Intrator and Leon N Cooper. Objective function formulation of the bcm theory of visual
cortical plasticity: Statistical connections, stability conditions. Neural Networks, 5(1):3?17,
1992.
[13] Matthew Lawlor and Steven S. W. Zucker. An online algorithm for learning selectivity to
mixture means. arXiv preprint, 2014.
[14] Matthew Lawlor and Steven W Zucker. Third-order edge statistics: Contour continuation,
curvature, and cortical connections. In Advances in Neural Information Processing Systems,
pages 1763?1771, 2013.
[15] WB Levy and O Steward. Temporal contiguity requirements for long-term associative potentiation/depression in the hippocampus. Neuroscience, 8(4):791?797, 1983.
[16] Bernhard Nessler, Michael Pfeiffer, and Wolfgang Maass. Stdp enables spiking neurons to
detect hidden causes of their inputs. In Advances in neural information processing systems,
pages 1357?1365, 2009.
[17] Jean-Pascal Pfister and Wulfram Gerstner. Triplets of spikes in a model of spike timingdependent plasticity. The Journal of neuroscience, 26(38):9673?9682, 2006.
[18] Sen Song, Kenneth D Miller, and Larry F Abbott. Competitive hebbian learning through spiketiming-dependent synaptic plasticity. Nature neuroscience, 3(9):919?926, 2000.
[19] Huai-Xing Wang, Richard C Gerkin, David W Nauen, and Guo-Qiang Bi. Coactivation and
timing-dependent integration of synaptic potentiation and depression. Nature neuroscience,
8(2):187?193, 2005.
9
| 5337 |@word version:2 hippocampus:2 stronger:1 proportion:1 km:1 d2:6 seek:1 decomposition:13 moment:1 initial:1 liu:1 fragment:1 daniel:2 ours:1 demarcated:1 o2:1 kmk:1 current:2 com:1 gmail:1 intriguing:1 must:1 written:2 realize:1 numerical:2 j1:2 plasticity:10 m1t:1 enables:1 designed:2 update:32 v:1 tone:1 oldest:1 ith:1 short:1 provides:2 math:1 completeness:1 location:1 characterization:1 simpler:2 along:1 c2:9 become:1 differential:1 ik:2 pairing:1 consists:1 fixation:2 vpre:3 fitting:1 kdk2:1 dan:5 introduce:1 pairwise:2 theoretically:1 inter:1 expected:13 behavior:2 multi:3 brain:2 ming:3 increasing:1 provided:1 notation:3 underlying:1 bounded:1 suffice:1 kind:1 substantially:1 emerging:1 eigenvector:1 contiguity:1 guarantee:4 temporal:2 control:2 unit:1 t1:2 positive:4 before:3 timing:21 local:1 limit:1 consequence:1 firing:10 becoming:1 might:2 weakened:1 resembles:1 suggests:3 limited:1 bi:3 coactivation:1 obeys:1 averaged:2 elie:1 neuropsychological:1 definite:1 backpropagation:1 significantly:2 physiology:1 convenient:1 projection:1 pre:11 specificity:1 close:2 superlinear:1 context:1 nessler:1 deterministic:1 quick:1 dean:1 poo:3 independently:1 convex:1 m2:3 rule:51 importantly:1 stability:2 imagine:2 controlling:1 cultured:1 strengthen:1 exact:1 hypothesis:1 jk:3 particularly:2 distributional:1 predicts:1 steven:3 role:1 preprint:3 rij:1 hv:2 electrical:1 epsc:1 svds:1 wj:1 wang:2 ordering:1 mu:3 suppressively:1 efficiency:1 basis:1 triangle:3 joint:1 train:3 distinct:1 gjorgjieva:1 quite:2 richer:1 supplementary:1 plausible:1 dt0:2 say:3 relax:1 kai:1 jean:2 statistic:1 online:4 reproduced:1 associative:1 sequence:3 sen:1 interaction:3 product:1 aligned:1 relevant:1 tj1:1 iff:1 degenerate:1 academy:1 convergence:5 interacted:1 requirement:3 r1:1 asymmetry:1 natalia:1 incremental:1 converges:3 telgarsky:1 object:2 depending:1 develop:1 measured:2 ij:1 solves:1 strong:1 implemented:2 c:1 predicted:1 come:1 convention:1 lyapunov:3 direction:1 posit:1 closely:2 correct:2 filter:1 stochastic:7 modifying:2 stabilization:1 larry:1 material:1 bin:6 explains:1 require:4 potentiation:2 decompose:1 biological:2 rong:1 stdp:6 matus:1 matthew:3 m0:1 achieves:1 vary:1 purpose:2 failing:1 estimation:1 vpost:4 successfully:1 tool:1 eti:1 always:2 super:1 modified:2 rather:4 ck:8 varying:1 voltage:12 publication:1 properly:1 rank:3 bernoulli:1 check:1 indicates:4 contrast:1 skipped:1 sense:2 detect:1 dependent:10 hidden:6 bienenstock:3 selective:8 doob:2 i1:2 provably:2 issue:1 among:1 orientation:3 pascal:2 overall:2 arg:1 development:1 integration:2 equal:2 having:1 sampling:1 qiang:3 broad:3 future:1 stimulus:10 haven:2 few:1 ac2:2 richard:1 resulted:1 national:1 geometry:2 phase:1 fire:1 ab:1 organization:1 highly:1 multiply:1 severe:1 mixture:34 analyzed:1 semidefinite:3 chain:3 implication:1 edge:5 hm2:1 orthogonal:2 circle:3 mk:3 increased:2 modeling:1 lawlor:3 wb:1 ordinary:1 deviation:3 parametrically:1 subset:1 dependency:2 spatiotemporal:1 sensitivity:1 probabilistic:1 off:1 recreated:1 michael:1 w1:2 again:1 choose:1 slowly:2 ek:3 leading:1 stark:1 potential:1 singleton:1 wk:2 inc:1 matter:1 depends:3 later:1 try:1 view:4 multiplicative:1 performed:1 wolfgang:1 red:2 wave:3 decaying:1 xing:2 complicated:1 competitive:1 simon:1 om:1 oi:1 who:1 succession:2 miller:1 conceptually:1 bayesian:1 biophysically:1 critically:1 epk:2 lighting:1 drive:1 history:1 explain:1 implausible:1 tended:1 synaptic:32 definition:3 frequency:1 proof:5 di:9 degeneracy:1 sampled:1 couple:1 hsu:2 animashree:2 recall:1 color:1 cj:1 formalize:1 amplitude:1 back:1 higher:2 dt:9 response:1 synapse:1 formulation:6 done:2 juliette:1 generality:2 furthermore:1 just:2 binocular:1 c2k:1 until:1 correlation:1 ei:2 propagation:1 incrementally:1 google:1 lack:1 quality:1 indicated:1 building:1 name:1 effect:2 true:1 symmetric:1 maass:1 i2:2 white:1 adjacent:2 during:4 quadruplet:1 width:2 excitation:4 claudia:1 timingdependent:1 m:4 hippocampal:1 multiview:5 demonstrate:2 performs:2 allen:1 image:3 novel:1 nih:1 functional:1 spiking:5 mt:11 stimulation:2 exponentially:1 volume:1 interpretation:2 slight:1 m1:5 interpret:1 occurred:1 automatic:1 gerkin:1 similarly:1 access:3 zucker:4 cortex:2 stable:13 longer:1 inhibition:2 similarity:1 curvature:2 posterior:1 showed:1 recent:1 reverse:1 selectivity:8 certain:1 steward:1 meeting:1 yi:1 lnp:2 seen:1 somewhat:1 r0:1 converge:9 period:5 signal:2 ii:1 sliding:4 multiple:5 interdependency:1 semi:1 full:3 sham:2 hebbian:6 technical:1 match:1 calculation:1 long:1 posited:1 post:15 paired:2 prediction:1 variant:2 vision:1 expectation:7 poisson:7 arxiv:4 normalization:2 c2t:1 cell:1 c1:4 addition:1 background:1 linearnonlinear:1 ode:1 interval:7 grow:1 operate:1 unlike:1 ascent:5 induced:1 anandkumar:2 presence:7 noting:1 feedforward:2 yang:2 identically:1 affect:1 fit:7 forthcoming:1 idea:1 t0:2 whether:1 caporale:1 munro:3 clopath:1 song:1 york:1 cause:4 repeatedly:2 action:1 depression:2 dramatically:1 useful:1 generally:1 detailed:1 informally:1 simplest:1 continuation:1 nsf:1 inhibitory:1 sign:1 neuroscience:8 disjoint:3 blue:2 discrete:5 hmt:2 four:1 threshold:4 falling:1 drawn:7 d3:4 audet:1 abbott:1 kenneth:1 asymptotically:1 powerful:1 arrive:1 place:1 reasonable:1 patch:1 bound:1 ct:7 followed:3 yale:3 annual:1 activity:4 strength:3 occur:4 precisely:2 constraint:1 your:1 scene:1 ri:2 nathan:1 extremely:1 leon:2 according:1 ball:1 remain:1 across:1 increasingly:1 postsynaptic:12 smaller:1 em:1 kakade:2 joseph:1 biologically:2 modification:10 happens:1 explained:1 dv:1 taken:2 equation:2 remains:3 previously:2 turn:2 count:1 needed:1 ge:1 fed:1 end:1 unusual:1 generalizes:2 gaussians:2 rewritten:1 observe:1 intrator:4 spectral:2 original:1 top:1 dirichlet:1 spiketiming:1 build:1 classical:12 comparatively:1 unchanged:1 tensor:20 objective:11 question:2 spike:61 parametric:2 primary:1 dependence:3 responds:1 traditional:1 diagonal:1 exhibit:1 gradient:6 kth:1 distance:1 topic:1 presynaptic:14 reason:1 provable:1 toward:1 assuming:2 o1:1 index:1 relationship:1 modeled:1 robert:1 debate:1 trace:2 negative:4 mink:1 unknown:1 perform:2 neuron:17 curved:1 t:2 neurobiology:1 extended:1 rn:2 varied:1 arbitrary:2 introduced:2 david:1 pair:7 required:2 c3:9 connection:3 optimized:1 bcm:41 learned:1 discontinuity:1 tpre:1 beyond:1 suggested:1 pattern:1 explanation:1 critical:1 event:2 natural:5 treated:2 pfeiffer:1 eye:1 axis:1 hm:7 coupled:1 epoch:2 review:1 l2:1 multiplication:1 relative:1 fully:1 expect:1 mixed:1 allocation:1 triple:3 foundation:2 rik:1 sufficient:1 foster:1 tiny:1 excitatory:1 supported:1 copy:2 free:3 arriving:1 fall:1 taking:1 leaky:1 distributed:2 curve:1 dimension:2 calculated:1 world:1 avoids:1 boundary:1 cortical:2 contour:1 tpost:1 projected:1 transaction:1 emphasize:2 compact:1 neurobiological:1 bernhard:1 global:2 latent:5 triplet:38 sk:2 learn:10 nature:3 gerstner:3 complex:2 froemke:4 investigated:1 protocol:1 diag:3 did:1 pk:6 main:1 linearly:2 s2:2 noise:8 paul:2 repeated:3 fashion:1 martingale:2 cooper:6 wiley:1 hebb:1 theme:1 governed:1 levy:1 third:2 removing:1 theorem:3 specific:1 offset:1 dk:11 decay:3 r2:1 incorporating:2 essential:1 corr:1 ci:1 nauen:1 texture:1 magnitude:1 delyon:2 conditioned:2 simply:3 visual:3 saccade:1 corresponds:2 conditional:1 viewed:2 formulated:1 hm1:1 hard:1 change:5 wulfram:1 specifically:1 determined:1 pfister:4 bernard:1 experimental:3 e6:1 guo:3 relevance:1 violated:1 incorporate:3 audio:1 d1:17 correlated:1 |
4,791 | 5,338 | A Bayesian model for identifying hierarchically
organised states in neural population activity
Patrick Putzky1,2,3 , Florian Franzen1,2,3 , Giacomo Bassetto1,3 , Jakob H. Macke1,3
1
Max Planck Institute for Biological Cybernetics, T?
ubingen
2
Graduate Training Centre of Neuroscience, University of T?
ubingen
3
Bernstein Center for Computational Neuroscience, T?
ubingen
[email protected], [email protected]
[email protected], [email protected]
Abstract
Neural population activity in cortical circuits is not solely driven by external inputs, but is also modulated by endogenous states which vary on
multiple time-scales. To understand information processing in cortical circuits, we need to understand the statistical structure of internal states
and their interaction with sensory inputs. Here, we present a statistical
model for extracting hierarchically organised neural population states from
multi-channel recordings of neural spiking activity. Population states are
modelled using a hidden Markov decision tree with state-dependent tuning
parameters and a generalised linear observation model. We present a variational Bayesian inference algorithm for estimating the posterior distribution
over parameters from neural population recordings. On simulated data, we
show that we can identify the underlying sequence of population states and
reconstruct the ground truth parameters. Using population recordings from
visual cortex, we find that a model with two levels of population states outperforms both a one-state and a two-state generalised linear model. Finally,
we find that modelling of state-dependence also improves the accuracy with
which sensory stimuli can be decoded from the population response.
1
Introduction
It has long been recognised that the firing properties of cortical neurons are not constant
over time, but that neural systems can exhibit multiple distinct firing regimes. For example,
cortical circuits can be in a ?synchronised? state during slow-wave sleep, exhibiting synchronised fluctuations of neural excitability [1] or in a ?desynchronised? state in which firing is
irregular. Neural activity in anaesthetised animals exhibits distinct states which lead to
widespread modulations of neural firing rates and contribute to cross-neural correlations
[2]. Changes in network state can be brought about through the influence of inter-area
interactions [3] and affect communication between cortical and subcortical structures [4].
Given the strong impact of cortical states on neural firing [3, 5, 4], an understanding of the
interplay between internal states and external stimuli is essential for understanding how populations of cortical neurons collectively process information. Multi-cell recording techniques
allow to record neural activity from dozens or even hundreds of neurons simultaneously,
making it possible to identify the signatures of underlying states by fitting appropriate
statistical models to neural population activity.
It is thought that the state-dependence of neocortical circuits is not well described using a
global bi-modal state. Instead, the structure of cortical states is more accurately described
1
Figure 1: Illustration of the model. A) Generative model. At time t, the cortical state st
is determined using a Hidden Markov Decision Tree (HMDT) and depends on the previous
state st?1 , population activity yt?1 and on the current stimulus xt . In our simulations,
we assumed that the first split of the tree determined whether to transition into an up
or down-state. Up-states contained transient periods of high firing across the population
(up-high) as well as sustained periods of irregular firing (up-low). Each cortical state is
then associated with different spike-generation dynamics, modelling state-dependence of
firing properties such as ?burstiness?. B) State-transition probabilities depend on the treestructure. Transition matrices are depicted as Hinton diagrams where each block represents
a probability and each column sums to 1. Each row corresponds to the possible future state
st (see colour), and each column to the current state.
(1) A model in which transition-probabilities in the first level of the tree (up/down) are
biased towards the up-state (green squares are bigger than gray ones), and weakly depend on
the previous state st?1 . In this example, both high/low phases are equally likely within upstates (second level of tree, depicted in second column) and do not depend on the previous
state (all orange/red squares have same size). The resulting 3 ? 3 matrix of transition
probabilities across all states can be calculated from the transition-probabilities in the tree.
(2) Changing the properties of the second-level node only leads to a local change in the
transition matrix: It affects the proportion between the orange/red states, but leaves the
green state unchanged.
using multiple states which vary both between and within brain regions [6]. In addition,
the ?state? of a neural population can vary across multiple time scales from milliseconds to
seconds or more [6]: For example, cortical recordings can switch between up- and downphases. During an up-phase cortical activity can exhibit ?volleys? of synchronised activity
[7]?sometimes referred to as population bursts?which can be modelled as transient states.
These observations suggest that the structure of cortical states could be captured by a
hierarchical organisation in which each state can give rise to multiple temporally nested
?sub-states?. This structure naturally yields a binary tree: States can be divided into subclasses, with states further down the tree operating at faster time-scales determined by
their parent node. We hypothesise that other cortical states also exhibit similar hierarchical
structure. Our goal here is to provide a statistical model which can identify cortical states
and their hierarchical organisation from recordings of population activity. As a running
example of such a hierarchical organisation we use a model in which the population exhibits
synchronised population bursts during up-states, but not during down-states. This system
is modelled using a first level of state (up/down), and for which the up-state is further
divided into two states (transient high-firing events and normal firing, see 1A).
We present an inhomogeneous hidden Markov model (HMM) [8] to model the temporal
dynamics of state-transitions [9, 10]. Our approach is most closely related to [10], who
developed a state-dependent generalised linear model [11] in which both the tuning prop2
erties and state-transitions can be modelled to depend on external covariates. However,
our formulation also allows for hierarchically organised state-structures. In addition, previous population models based on discrete latent states [10, 12] used point-estimation for
parameter learning. In contrast, we present algorithms for full Bayesian inference over the
parameters of our model, making it possible to identify states in smaller or noisier data
[13]. This is important for neural population recordings which are typically characterised
by short recording times relative to the dimensionality of the data and by high variability.
In addition, estimates of posterior distributions are important for visualising uncertainty
and for optimising experimental paradigms with active-learning methods [14, 15].
2
Methods
We use a hidden Markov decision tree (HMDT) [16] to model hierarchically organised states
with binary splits and a generalised linear observation model (GLM). An HMDT combines
the properties of a hidden Markov model (to model temporal structure) with a hierarchical
mixture of experts (HME, to model a hierarchy of latent states) [17]. In general the hierarchical approach can represent richer dependence of states on external covariates, analogous
to the difference between multi-class logistic regression and multi-class binary decision trees.
For example, a two-level binary tree can separate four point clouds situated at the corners
of a square whereas a 4-class multinomial regression cannot. We use Bayesian logistic regression [18] to model transition gates and emissions. In the following, we describe the model
structure and propose a variational algorithm [8, 19] for inferring its parameters.
2.1
Hierarchical hidden Markov model for multivariate binary data
We consider discrete time-series data of multivariate binary1 neural spiking events yt ?
{0, 1}C where C is the number of cells. We assume that neural spiking can be influenced
by (observed) covariates xt ? RD . The covariates xt could represent external stimuli,
spiking history of neurons or other measures such as the total population spike count. In
our analyses below, we assume that correlations across neurons arise only from the joint
coupling to the population state, and we do not include couplings between neurons as is
sometimes done with GLMs [11]. Dependence of neural firing on internal states is modelled
by including a 1-of-K latent state vector st , where K is the number of latent states. The
emission probabilities for the observable vector yt (i.e. the probability of spiking for each
neuron) are thus given by
C
K Y
s(i)
Y
(c) (c)
(c) t
,
(1)
p yt |xt , ?i
p (yt |xt , st , ?) =
i=1 c=1
where ? is a set of model parameters. We allow the external covariate xt to be different for
each neuron c.
To model temporal dynamics over st , we use a hidden Markov model (HMM) [10], where
the state transitions take the form
(j)
K Y
K
s(i)
Y
t st?1
(i) (j)
p (st |st?1 , xt , ?) =
p st |st?1 , xt , ?
,
(2)
i=1 j=1
where ? is a set of parameters of the transition model. The model allows state-transitions
to be dependent on an external input xt ? this can e.g. be used to model state-transitions
caused by stimulation of subcortical structures involved in controlling cortical states [20].
Moving beyond this standard input output HMM formulation [21], we introduce hierarchically organised auxiliary latent variables zt which represent the current state st through a
binary tree. Using HME terminology, we refer to the nodes representing zt as ?gates?. Each
of the K leaves of the tree (or, equivalently, each path through the tree) corresponds to one
of the K entries of st and we can thus represent st in the form
L
A(l,k)
A(l,k)
Y
L
R
(k)
(l)
(l)
st =
zt
1 ? zt
,
(3)
l=1
1
All derivations below can be generalised to model the emission probabilities by any kind of
generalised linear model.
3
where AL and AR are adjacency matrices which indicate whether state k is in the left or
right branch of gate l, respectively (see [19]). Using this representation, st is deterministic
given zt which significantly simplifies the inference process. The auxiliary latent variables
(l)
zt are Bernoulli random variables and we chose their conditional probability distribution
to be
(l)
(l)
(l)
p(zt = 1|xt , st?1 , vl ) = ? v>
u
.
(4)
t
l
Here, ?(?) is the logistic sigmoid, vl are the parameters of the l-th gate and ut represents
a concatenation of the previous state st?1 , the input xt (which could for example represent
population firing rate, time in trial or an external stimulus) and a constant term of unit
(l)
value to model the prior probability of z0 = 1. This parametrisation significantly reduces
the number of parameters used for the transition probabilities as compared to [10]. To
enforce stronger temporal locality and less jumping between states we could also reduce
this probability to be conditioned only on previous activations of a sub-tree of the HMDT
instead of all population states.
2.2
Learning & Inference
For posterior inference over the model parameters we would need to infer the joint distribution over all stochastic variables conditioned on X,
p (Y, S, ?, ?, ?, ?|X) =p (Y|S, X, ?) p (S|X, ?) p (?|?) p (?) p (?|?) p (?)
(5)
where Y is the set of yt ?s, ? and ? are the sets of parameters for the emission and gating
distributions, respectively, and ? and ? are the hyperparameters for the parameter priors.
Since there is no closed form solution for this distribution, we use a variational approximation
[8]. We assume that the posterior factorises as
q (S, ?, ?, ?, ?) =q (S) q (?) q (?) q (?) q (?)
(6)
L
K Y
C
Y
Y
(c)
(c)
q (? l ) q (? l ) ,
(7)
=q (S)
q ? k q ?k
k=1 c=1
l=1
and find the variational approximation to the posterior over parameters, q (S, ?, ?, ?, ?),
by optimising the variational lower bound L(q) to the evidence
X ZZZZ
p (Y, S, ?, ?, ?, ?|X)
L(q) :=
q (S, ?, ?, ?, ?) ln
d?d?d?d?
(8)
q (S, ?, ?, ?, ?)
S
X ZZZZ
? ln
p (Y, S, ?, ?, ?, ?|X) d?d?d?d? = ln p (Y|X) .
(9)
S
We use variational Expectation-Maximisation (VBEM) to perform alternating updates on
the posterior over latent state variables and the posterior over model parameters. To infer
the posterior over latent variables (i.e. responsibilities), we use a modified forward-backward
algorithm as proposed in [22] (see also [8]). In order to perform the forward and backward
steps, they propose the use of subnormalised probabilities of the form
i
h
(i) (j)
(i) (j)
(10)
p? st |st?1 , xt , ? := exp E? ln p st |st?1 , xt , ?
p? (yt |xt , ?i ) := exp (E?i [ln p (yt |xt , ?i )])
(11)
for the state-transition probabilities and emission probabilities. Since all relevant probabilities in our model are over discrete variables, it would be straightforward to normalise those
probabilities, but we found that normalisation did not noticeably change results.
With the approximations from above, the forward probability can thus be written as
K
X
1
(i)
(i)
(j)
(i) (j)
? st
=
p? yt |st , xt , ?
? st?1 p? st |st?1 , xt , ? ,
(12)
C?t
j=1
(i)
(i)
where ?(st ) is the probability-mass of state st given previous time steps and C?t is a
normalisation constant. Similar to the forward step, the backward recursion takes the form
4
K
1 X (j)
(i)
(j)
(j)
(i)
? st
=
?t st+1 p? yt+1 |st+1 , xt+1 , ? p? st+1 |st , xt , ? .
C?t
(13)
j=1
Using the forward and backward equation steps we can infer the state posteriors [8]. Given
the state posteriors, the logarithm of the approximate parameter posterior for each of the
nodes takes the form
T
X
(n)
(n) (n)
ln q ? (? n ) =
?t ln p ?t |xt , ? n , (. . . ) + E? n [ln p (? n |? n )] + const.
(14)
t=1
where ? n are the parameters of the n-th node and p (? n |? n ) is the prior over the param(n)
eters. Here, ?t is the posterior responsibility or estimated influence of node n on the tth
(n)
observation and ?t denotes the expected output (known for state nodes) of node n (see
supplement for details). This equation also holds for a tree structure with multinomial gates
and for non-binary emission models such as Poisson and linear models. The above equations
are valid for maximum likelihood inference, except that all parameter priors are removed,
and the expectations of log-likelihoods reduce to log-likelihoods We use logistic regression
for all emission probabilities and gates, and a local variational approximation to the logistic
sigmoid as presented in [18].
As parameter priors we use anisotropic Gaussians with individual Gamma priors on each
diagonal entry of the precision matrix. With this prior structure we can perform automatic
relevance determination [23]. We chose shape parameter a0 =1 ? 10?2 and rate parameter
b0 = 1 ? 10?4 , leading to a broad Gamma hyperprior [19]. In many applications, it will
be reasonable to assume that neurons in close-by states of the tree show similar response
characteristics (similar parameters). The hierarchical organisation of the model yields a
natural structure for hierarchical priors which can encourage parameter similarity2 .
2.3
Details of simulated and neurophysiological data
To assess and illustrate our model, we simulated a population recording with trials of 3 s
length (20 neurons, 10 ms time bins). As illustrated in Fig. 1 A, we modelled one low-firingrate down state (down, base firing rate 0.5 Hz) and two up states (up-low and up-high, with
base firing rates of 5, and 50 Hz respectively). The root node switched between up and
down states, whereas a second node controlled transitions between the two types of upstates. Up-high states only occurred transiently, modelling synchronised bouts of activity.
In the down state, neurons have a 10 ms refractory period, during up states they exhibit
bursting activity. Transitions from down to up go mainly via up-high to up-low, while downtransitions go from up-low to down; stimulation increases the probability of being in one of
the up states. A pulse-stimulus occurred at time 1 s of each trial. Each model was fit on a
set of 20 trials and evaluated on a different test set of 20 trials. For each training set, 24
random parameter initialisations were drawn and the one with highest evidence was chosen
for evaluation. State predictions were evaluated using the Viterbi algorithm [24, Ch. 13].
We analysed a recording from visual cortex (V1) of an anaesthetised macaque [2]. The
data-set consisted of 1600 presentations of drifting gratings (16 directions, 100 trials each),
each lasting 2 s. Experimental details are described in [2]. For each trial, we kept a segment
of 500 ms before and after a stimulus presentation, resulting in trials of length 3 s each. We
binned and binarised spike trains in 50 ms bins. Additional spikes (present in (5.45 ? 1.56) %
of bins) were discarded by the binarisation procedure. We chose the representation of
the stimulus to be the outer product of the two vectors [1, sin(?), cos(?)], where ? is
the phase of the grating, and [1, sin(?), cos(?), sin(2?), cos(2?)] for the direction ? of the
grating. This resulted in a 15 dimensional stimulus-parametrisation, and made it possible to
represent tuning-curves with orientation and direction selectivity, as well as modulation of
firing rates by stimulus phase. The only gate input was chosen to be an indicator function
with unit value during stimulus presentation and zero value otherwise. Post-spike filters
were parametrised using five cubic b-splines for the last 10 bins with a bin width of 50 ms.
2
See supplement for an example of how this could be implemented with Gaussian priors.
5
Figure 2: Performance of the model on simulated data. A) Example rasters sampled using ground truth (GT) parameters, colors indicate sequence of underlying population
states. B) For the sample from (A), the state-sequence decoded with our variational Bayes
(VB) method matches the decoded sequence using GT parameters. C) Comparison of statedecoding performance using GT parameters, VB and maximum likelihood (ML) learning
(Wilcoxon ranksum, * p < 0.05; *** p 0.001). D) Model performance quantified using
per-data-point log-likelihood difference between estimated and GT-model on test-set. Our
VB method outperforms ML (Wilcoxon ranksum, *** p 0.001), and both models considerably outperform a 1-state GLM (not shown). E) Estimated post-spike filters match
the GT values well (depicted are the filters from one of the cross-validated models). F)
Comparison of the autocorrelation of the ground truth data and samples drawn from the
VB fit as in (E). G) GT (top) and VB estimated (bottom) transition matrices in absence
(left) or presence (right) of a stimulus.
3
3.1
Results
Results on simulated data
To illustrate our model and to evaluate the estimation procedure on data with known ground
truth, we used a simulated population recording of 20 neurons by sampling from our model
(details in Methods, see Fig. 2 A). In this simulation, the up-state had much higher firing
rates than the down-state. It was therefore possible to decode the underlying states from
the population spike trains with high accuracy (Fig. 2 B). For the VB method, we used
the posterior mean over parameters for state-inference. In addition, we compared both of
these approaches to state-decoding based on a model estimated using maximum likelihood
learning. All three models showed similar performance, but the decoding advantage of the
3-state VB model was statistically significant (using pairwise comparisons, Fig. 2 C).
We also directly evaluated performance of the VB and ML methods for parameter estimation
by calculating the log-likelihood of the data on held-out test-data, and found that our VB
method performed significantly better than the ML method (Fig. 2 D). Finally, we also
compared the estimated post-spike filters (Fig. 2 E), auto-correlation functions (Fig. 2 F)
and state-transition matrices (Fig. 2 G) and found an excellent agreement between the GT
parameters and the estimates returned by VB.
To test whether the VB method is able to determine the correct model complexity, we
fit an over-parameterised model with 3 layers and potentially 8 states to the simulation
data. The best model fit from 200 random restarts (lower bound of ?2.24 ? 104 , no crossvalidation, results not shown) only used 3 out of the 8 possible states (the other 5 states
had a probability of less than 0.5 %). Therefore, in this example, the best lower bound is
achieved by a model with correct, and low, complexity.
6
C 50
*** *** *** *** ***
0
-1
40
30
20
10
-2
90
180 270
direction (deg)
360
0
0
360
0
0
1000
2000
time (ms)
1.5
1
0.5
0
0
G
10
10
10
250
500
empirical ITIs (ms)
750
population rate (%)
F
sampeled ITIs
events per trial
E
90
180 270
direction (deg)
0.2
0
-2
-4
-2
10
empirical ITIs
10
0
1S
1S
2S
3S
1
0.5
0
0
1000
2000
time (ms)
50
250
time (ms)
500
H 50
40
30
20
10
0
PR
d
le
co
up
5
0
L
3S
1S
PR
le
d
10
v
1.5
modulation
5
0.4
iv
15
spikes (hz)
10
iii
0.6
population rate (%)
0
co
up
M
0.2
ii
15
spikes (hz)
p?(spike)
0.4
0
1S-GLM
empirical
i
pt(spike|?=67.5?)
sampled
D0.6
1S
0
2S
2000
1S
1000
3S
0
L
-500
M
?loglikelihood
B
1S
down
up low-rate
accuracy (%)
up high-rate
A
0
5
10
number of spikes (per bin)
40
30
20
10
0
0
5
10
number of spikes (per bin)
Figure 3: Results for population recordings from V1. A) Raster plot of population
response to a drifting grating with orientation 67.5? . Arrows indicate stimulus onset and
offset, colours show the most likely state sequence inferred with the 3-state variational Bayes
(3S-VB) model. B) Cross-validated log-likelihoods per trial, relative to the 3S-VB model.
C) Stimulus decoding performance, in percentage of correctly decoded stimuli (16 discrete
stimuli, chance level 6.25 %), using maximum-likelihood decoding.
D) Tuning properties of an example neuron. i) Orientation tuning calculated from the
tuning-parameters of 3S-VB (red, orange, green) or 1-state GLM (purple). iii) Temporal
component of tuning parameters. ii) Orientation tuning measured from sampled data of the
estimated model, each line representing one state. Note that the firing rate also depends
on state-transitions and post-spike filters. iv) Peri-stimulus time-histograms (PSTHs) estimated from samples of the estimated models. v) Post-spike filters for each state, and
comparison with 1-state GLM (purple). E) Distributions of times spent in each state, i.e.
inter-transition intervals (ITIs), estimated from the empirical data using 3S-VB. F) Comparison between distribution of ITIs in samples from model 3S-VB and in the Viterbi-decoded
path (from E).
G) Histogram of population rates (i.e. number of synchronous spikes across the population in each 50 ms bin) for 3S-VB (blue), 1S (purple), and data (gray). H) Histograms of
population rate for each state.
3.2
Results on neurophysiological recordings
We analysed a neural population recording from V1 to determine whether we could successfully identify cortical states by decoding the activity of the neural population, and whether
accounting for state-dependence resulted in a more accurate statistical model of neural firing.
While neurons generally responded robustly to the stimulus (3 D), firing rates were strongly
modulated by internal states [2] (Fig. 3 A). We fit different models to data, and found that
our 3-state model estimated with VB resulted in better cross-validation performance than
either the 3-state model estimated with ML, the 2-state model or a 1-state GLM (i.e. a
GLM without cross-neural couplings, Fig. 3 B). In addition we fit a fully coupled GLM
(with cross-history terms as in [11, 13]), as well as one in which the total population count
was used as a history feature using VB. These models were intermediate between the 1-state
GLM and the 2-state model, i.e. both worse than the 3-state one. A ?flat? 3-states model
with a single multinomial gate estimated with ML performed similarly to the hierarchical
3S-ML model. This is to be expected, as any differences in expressive power between the
two models will only become substantial for a different choice of xt or larger models.
7
We also evaluated the ability of different models to decode the stimulus, (i.e. the direction
of the presented grating) from population spike trains. We evaluated the likelihood of each
population spike train for each of the 16 stimulus directions, and decoded the stimulus which
yielded the highest likelihood. The 3-state VB model shows best decoding performance
among all tested models (3 C), and all models with state-dependence (3-state VB, 3-state
ML, 2-state) outperformed the 1-state GLM. We sampled from the estimated 3S-VB model
to evaluate to what extent the model captures the tuning properties of neurons (Fig. 3 D(ii
& iv)). The example neuron shows strong modulation of base firing rate dependent on the
population state, but not a qualitative change of the tuning properties (Fig. 3 D i-iv). The
down-state post-spike filter (Fig. 3 D v) exhibits a small oscillatory component which is not
present in the post-spike filters of the other states or the 1-state GLM.
Investigation of inter-transition-interval (ITI) distributions from the data (after Viterbidecoding) shows heavy tails (Fig. 3 E). Comparison of ITI-distribution estimated from the
empirical data and from sampled data (3S-VB) show good agreement, apart from small
deficiencies of the model to capture the heavy tails of the empirical ITI distribution (Fig.
3 F). Finally, population rates (i.e. total number of spikes across the population) are often
used as a summary-measure for characterizing cortical states [6]. We found that the distribution of population rates in the data was well matched by the distribution estimated
from our model (Fig. 3 G) with the three states having markedly different population rate
distributions (Fig. 3 H). Although a 1-state GLM also captured the tuning-properties of
this neuron (Fig. 3 D) it failed to recover the distribution of population rates (Fig. 3 G).
4
Discussion
We presented a statistical method for extracting cortical states from multi-cell recordings of
spiking activity. Our model is based on a ?state-dependent? GLM [10] in which the states are
organised hierarchically and evolve over time according to a hidden Markov model. Whether,
and in which situations, the best descriptions of cortical states are multi-dimensional, discrete or continuous [25, 2] is an open question [6], and models like the one presented here
will help shed light on these questions. We showed that the use of variational inference
methods makes it possible to estimate the posterior over parameters. Bayesian inference
provides better model performance on limited data [13], uncertainty information, and is
also an important building block for active learning approaches [14]. Finally, it can be used
to determine the best model complexity: For example, one could start inference with a
model containing only one state and iteratively add states (as in divisive clustering) until
the variational bound stops increasing.
Cortical states can have a substantial impact on the firing and coding properties of cortical
neurons [6] and interact with inter-area communication [4, 3]. Therefore, a better understanding of the interplay between cortical states and sensory information, and the role of
cortical states in gating information in local cortical circuits will be indispensable for our
understanding of how populations of neurons collectively process information. Advances in
experimental technology enable us to record neural activity in large populations of neurons
distributed across brain areas. This makes it possible to empirically study how cortical
states vary across the brain, to identify pathways which influence state, and ultimately to
understand their role in neural coding and computation. The combination of such data with
statistical methods for identifying the organisation of cortical states holds great promise for
making progress on understanding state-dependent information processing in the brain.
Acknowledgements
We are grateful to the authors of [2] for sharing their data (toliaslab.org/publications/eckeret-al-2014/) and to Alexander Ecker, William McGhee, Marcel Nonnenmacher and David
Janssen for comments on the manuscript. This work was funded by the German Federal Ministry of Education and Research (BMBF; FKZ: 01GQ1002, Bernstein Center T?
ubingen) and
the Max Planck Society. Supplementary details and code are available at www.mackelab.org.
8
References
[1] M. Steriade and R. W. McCarley, Brain Control of Wakefulness and Sleep. Kluwer Academic/plemum publishers, 2005.
[2] A. S. Ecker, P. Berens, R. J. Cotton, M. Subramaniyan, G. H. Denfield, C. R. Cadwell, S. M.
Smirnakis, M. Bethge, and A. S. Tolias, ?State dependence of noise correlations in macaque
primary visual cortex,? Neuron, vol. 82, no. 1, 2014.
[3] E. Zagha, A. E. Casale, R. N. S. Sachdev, M. J. McGinley, and D. A. McCormick, ?Motor
cortex feedback influences sensory processing by modulating network state,? Neuron, vol. 79,
no. 3, 2013.
[4] N. K. Logothetis, O. Eschenko, Y. Murayama, M. Augath, T. Steudel, H. C. Evrard,
M. Besserve, and A. Oeltermann, ?Hippocampal-cortical interaction during periods of subcortical silence,? Nature, vol. 491, no. 7425, 2012.
[5] T. Bezdudnaya, M. Cano, Y. Bereshpolova, C. R. Stoelzel, J.-M. Alonso, and H. A. Swadlow,
?Thalamic burst mode and inattention in the awake LGNd,? Neuron, vol. 49, no. 3, 2006.
[6] K. D. Harris and A. Thiele, ?Cortical state and attention,? Nature reviews. Neuroscience,
vol. 12, no. 9, 2011.
[7] M. A. Kisley and G. L. Gerstein, ?Trial-to-Trial Variability and State-Dependent Modulation
of Auditory-Evoked Responses in Cortex,? J. Neurosci., vol. 19, no. 23, 1999.
[8] M. J. Beal, ?Variational algorithms for approximate bayesian inference,? 2003.
[9] L. M. Jones, A. Fontanini, B. F. Sadacca, P. Miller, and D. B. Katz, ?Natural stimuli evoke
dynamic sequences of states in sensory cortical ensembles,? PNAS, vol. 104, no. 47, 2007.
[10] S. Escola, A. Fontanini, D. Katz, and L. Paninski, ?Hidden Markov models for the stimulusresponse relationships of multistate neural systems,? Neural Computation, vol. 23, no. 5, 2011.
[11] L. Paninski, J. Pillow, and J. Lewi, ?Statistical models for neural encoding, decoding, and
optimal stimulus design,? Progress in Brain Research, vol. 165, 2007.
[12] Z. Chen, S. Vijayan, R. Barbieri, M. A. Wilson, and E. N. Brown, ?Discrete- and continuoustime probabilistic models and algorithms for inferring neuronal UP and DOWN states,? Neural
Computation, vol. 21, no. 7, 2009.
[13] S. Gerwinn, J. H. Macke, and M. Bethge, ?Bayesian inference for generalized linear models
for spiking neurons,? Frontiers in Computational Neuroscience, vol. 4, no. 12, 2010.
[14] J. Lewi, R. Butera, and L. Paninski, ?Sequential optimal design of neurophysiology experiments,? Neural Computation, vol. 21, no. 3, 2009.
[15] B. Shababo, B. Paige, A. Pakman, and L. Paninski, ?Bayesian inference and online experimental design for mapping neural microcircuits,? in Advances in Neural Information Processing
Systems 26, pp. 1304?1312, Curran Associates, Inc., 2013.
[16] M. I. Jordan, Z. Ghahramani, and L. K. Saul, ?Hidden markov decision trees,? in Advances
in Neural Information Processing Systems 9, pp. 501?507, MIT Press, 1997.
[17] M. I. Jordan and R. A. Jacobs, ?Hierarchical Mixtures of Experts and the EM Algorithm,?
Neural Computation, vol. 6, no. 2, 1994.
[18] T. S. Jaakkola and M. I. Jordan, ?A variational approach to Bayesian logistic regression models
and their extensions,? 1996.
[19] C. M. Bishop and M. Svenskn, ?Bayesian hierarchical mixtures of experts,? in Proceedings of
the Nineteenth Conference on Uncertainty in Artificial Intelligence, UAI?03, (San Francisco,
CA, USA), pp. 57?64, Morgan Kaufmann Publishers Inc., 2003.
[20] G. Aston-Jones and J. D. Cohen, ?An integrative theory of locus coeruleus-norepinephrine
function: Adaptive gain and optimal performance,? in Annual Review of Neuroscience, vol. 28,
pp. 403?450, Annual Reviews, 2005.
[21] Y. Bengio and P. Frasconi, ?An input output hmm architecture,? in Advances in Neural
Information Processing Systems 7, pp. 427?434, MIT Press, 1995.
[22] D. J. C. MacKay, ?Ensemble Learning for Hidden Markov Models,? tech. rep., Cavendish
Laboratory, University of Cambridge, 1997.
[23] D. J. C. MacKay, ?Bayesian Non-linear Modeling for the Prediction Competition,? ASHRAE
Transactions, vol. 100, no. 2, pp. 1053?1062, 1994.
[24] C. M. Bishop, Pattern Recognition and Machine Learning. Information science and statistics,
New York: Springer, 2006.
[25] J. H. Macke, L. Buesing, J. P. Cunningham, B. M. Yu, K. V. Shenoy, and M. Sahani, ?Empirical models of spiking in neural populations,? in Advances in Neural Information Processing
Systems, vol. 24, Curran Associates, Inc., 2011.
9
| 5338 |@word neurophysiology:1 trial:12 proportion:1 stronger:1 open:1 integrative:1 simulation:3 pulse:1 accounting:1 jacob:1 series:1 initialisation:1 outperforms:2 current:3 com:1 analysed:2 activation:1 gmail:1 written:1 shape:1 motor:1 plot:1 update:1 generative:1 leaf:2 intelligence:1 shababo:1 anaesthetised:2 short:1 record:2 provides:1 contribute:1 node:10 org:2 five:1 burst:3 become:1 qualitative:1 sustained:1 fitting:1 combine:1 pathway:1 autocorrelation:1 introduce:1 pairwise:1 inter:4 expected:2 mpg:3 multi:6 brain:6 param:1 increasing:1 estimating:1 underlying:4 matched:1 circuit:5 mass:1 what:1 kind:1 developed:1 temporal:5 besserve:1 subclass:1 smirnakis:1 shed:1 control:1 unit:2 planck:2 shenoy:1 generalised:6 before:1 local:3 encoding:1 barbieri:1 solely:1 firing:21 fluctuation:1 modulation:5 path:2 chose:3 bursting:1 quantified:1 evoked:1 co:5 limited:1 graduate:1 bi:1 statistically:1 maximisation:1 block:2 lewi:2 procedure:2 area:3 empirical:7 thought:1 significantly:3 suggest:1 cannot:1 desynchronised:1 close:1 inattention:1 influence:4 www:1 deterministic:1 ecker:2 center:2 yt:10 straightforward:1 go:2 attention:1 identifying:2 population:49 cavendish:1 analogous:1 hierarchy:1 controlling:1 pt:1 decode:2 logothetis:1 curran:2 agreement:2 associate:2 recognition:1 observed:1 cloud:1 bottom:1 role:2 visualising:1 capture:2 region:1 removed:1 highest:2 thiele:1 burstiness:1 substantial:2 complexity:3 covariates:4 dynamic:4 ultimately:1 signature:1 depend:4 weakly:1 segment:1 grateful:1 joint:2 derivation:1 train:4 distinct:2 describe:1 artificial:1 treestructure:1 macke1:1 richer:1 larger:1 supplementary:1 nineteenth:1 loglikelihood:1 reconstruct:1 otherwise:1 ability:1 statistic:1 online:1 beal:1 interplay:2 sequence:6 advantage:1 multistate:1 propose:2 interaction:3 product:1 steriade:1 relevant:1 wakefulness:1 murayama:1 description:1 competition:1 crossvalidation:1 parent:1 spent:1 coupling:3 illustrate:2 help:1 measured:1 b0:1 progress:2 grating:5 strong:2 auxiliary:2 implemented:1 marcel:1 indicate:3 exhibiting:1 direction:7 inhomogeneous:1 closely:1 correct:2 filter:8 stochastic:1 transient:3 enable:1 adjacency:1 noticeably:1 bin:8 education:1 investigation:1 biological:1 frontier:1 extension:1 hold:2 ground:4 normal:1 exp:2 great:1 viterbi:2 mapping:1 vary:4 estimation:3 outperformed:1 modulating:1 successfully:1 federal:1 brought:1 mit:2 gaussian:1 modified:1 wilson:1 publication:1 jaakkola:1 validated:2 emission:7 modelling:3 bernoulli:1 likelihood:11 mainly:1 tech:1 contrast:1 inference:13 dependent:7 vl:2 typically:1 a0:1 cunningham:1 hidden:11 among:1 orientation:4 mackelab:1 animal:1 orange:3 mackay:2 having:1 frasconi:1 sampling:1 optimising:2 represents:2 broad:1 jones:2 yu:1 future:1 stimulus:23 transiently:1 spline:1 simultaneously:1 gamma:2 resulted:3 individual:1 phase:4 william:1 continuoustime:1 normalisation:2 evaluation:1 mixture:3 light:1 parametrised:1 held:1 accurate:1 encourage:1 jumping:1 tree:18 hmdt:4 iv:4 logarithm:1 hyperprior:1 ashrae:1 column:3 vbem:1 modeling:1 mccarley:1 ar:1 hypothesise:1 entry:2 hundred:1 stimulusresponse:1 giacomo:2 considerably:1 st:35 peri:1 nonnenmacher:1 probabilistic:1 decoding:7 bethge:2 parametrisation:2 containing:1 worse:1 external:8 corner:1 expert:3 macke:2 leading:1 ranksum:2 hme:2 de:3 coding:2 inc:3 caused:1 depends:2 onset:1 performed:2 root:1 cadwell:1 endogenous:1 closed:1 responsibility:2 red:3 wave:1 recover:1 bayes:2 thalamic:1 start:1 ass:1 square:3 purple:3 accuracy:3 responded:1 kaufmann:1 who:1 characteristic:1 miller:1 yield:2 identify:6 ensemble:2 modelled:6 bayesian:11 buesing:1 accurately:1 eters:1 cybernetics:1 history:3 oscillatory:1 influenced:1 sharing:1 raster:2 pp:6 involved:1 naturally:1 associated:1 sampled:5 stop:1 auditory:1 gain:1 color:1 ut:1 improves:1 dimensionality:1 manuscript:1 higher:1 restarts:1 response:4 modal:1 formulation:2 done:1 evaluated:5 strongly:1 microcircuit:1 mcghee:1 parameterised:1 correlation:4 glms:1 until:1 expressive:1 widespread:1 logistic:6 mode:1 gray:2 building:1 usa:1 consisted:1 brown:1 excitability:1 alternating:1 iteratively:1 butera:1 laboratory:1 illustrated:1 sin:3 during:7 width:1 m:10 generalized:1 hippocampal:1 recognised:1 neocortical:1 cano:1 variational:13 sigmoid:2 psths:1 multinomial:3 spiking:8 stimulation:2 empirically:1 cohen:1 refractory:1 anisotropic:1 tail:2 occurred:2 kluwer:1 katz:2 refer:1 significant:1 cambridge:1 tuning:11 rd:1 automatic:1 similarly:1 centre:1 had:2 funded:1 moving:1 cortex:5 operating:1 gt:7 patrick:2 base:3 wilcoxon:2 add:1 posterior:14 multivariate:2 showed:2 driven:1 apart:1 selectivity:1 indispensable:1 ubingen:4 gerwinn:1 binary:7 rep:1 captured:2 ministry:1 additional:1 morgan:1 florian:2 determine:3 paradigm:1 period:4 ii:3 branch:1 multiple:5 full:1 pnas:1 reduces:1 infer:3 d0:1 faster:1 determination:1 match:2 cross:6 long:1 academic:1 pakman:1 divided:2 post:7 equally:1 bigger:1 controlled:1 impact:2 prediction:2 regression:5 expectation:2 poisson:1 histogram:3 sometimes:2 represent:6 achieved:1 cell:3 irregular:2 addition:5 whereas:2 interval:2 diagram:1 publisher:2 biased:1 binarised:1 markedly:1 comment:1 recording:15 hz:4 jordan:3 extracting:2 presence:1 bernstein:2 split:2 iii:2 intermediate:1 bengio:1 switch:1 affect:2 fit:6 architecture:1 fkz:1 reduce:2 simplifies:1 synchronous:1 whether:6 colour:2 returned:1 paige:1 york:1 generally:1 situated:1 tth:1 outperform:1 percentage:1 millisecond:1 neuroscience:5 estimated:16 per:5 correctly:1 sadacca:1 blue:1 discrete:6 promise:1 vol:16 four:1 terminology:1 drawn:2 changing:1 kept:1 backward:4 v1:3 sum:1 uncertainty:3 reasonable:1 gerstein:1 decision:5 vb:23 steudel:1 bound:4 layer:1 sleep:2 yielded:1 activity:15 annual:2 binned:1 deficiency:1 awake:1 flat:1 according:1 combination:1 across:8 smaller:1 em:1 making:3 lasting:1 pr:2 glm:13 ln:8 equation:3 count:2 german:1 locus:1 available:1 gaussians:1 hierarchical:12 appropriate:1 enforce:1 robustly:1 gate:8 drifting:2 denotes:1 running:1 include:1 top:1 clustering:1 const:1 calculating:1 ghahramani:1 society:1 unchanged:1 question:2 spike:22 primary:1 dependence:8 diagonal:1 exhibit:7 separate:1 simulated:6 concatenation:1 hmm:4 normalise:1 outer:1 alonso:1 extent:1 tuebingen:3 length:2 code:1 relationship:1 illustration:1 equivalently:1 potentially:1 rise:1 design:3 zt:7 perform:3 mccormick:1 observation:4 neuron:24 markov:11 discarded:1 iti:3 denfield:1 situation:1 hinton:1 communication:2 variability:2 jakob:2 inferred:1 david:1 cotton:1 bout:1 macaque:2 beyond:1 able:1 below:2 pattern:1 regime:1 max:2 green:3 including:1 power:1 event:3 natural:2 indicator:1 recursion:1 representing:2 technology:1 factorises:1 aston:1 temporally:1 auto:1 coupled:1 sahani:1 prior:9 understanding:5 acknowledgement:1 review:3 evolve:1 relative:2 fully:1 generation:1 organised:6 subcortical:3 validation:1 switched:1 gq1002:1 itis:5 heavy:2 row:1 summary:1 last:1 silence:1 fontanini:2 allow:2 understand:3 institute:1 saul:1 characterizing:1 distributed:1 curve:1 calculated:2 cortical:30 transition:23 valid:1 feedback:1 pillow:1 sensory:5 forward:5 made:1 author:1 san:1 adaptive:1 transaction:1 approximate:2 observable:1 evoke:1 ml:8 global:1 active:2 deg:2 uai:1 assumed:1 francisco:1 tolias:1 continuous:1 latent:8 norepinephrine:1 channel:1 nature:2 ca:1 interact:1 excellent:1 berens:1 did:1 hierarchically:6 neurosci:1 arrow:1 noise:1 arise:1 hyperparameters:1 neuronal:1 fig:19 referred:1 cubic:1 slow:1 bmbf:1 precision:1 sub:2 inferring:2 decoded:6 volley:1 erties:1 dozen:1 down:15 z0:1 xt:21 covariate:1 bishop:2 gating:2 offset:1 coeruleus:1 organisation:5 evidence:2 essential:1 janssen:1 sequential:1 supplement:2 zagha:1 conditioned:2 chen:1 locality:1 depicted:3 paninski:4 likely:2 neurophysiological:2 visual:3 failed:1 contained:1 collectively:2 springer:1 ch:1 corresponds:2 truth:4 nested:1 chance:1 harris:1 conditional:1 goal:1 presentation:3 towards:1 absence:1 change:4 determined:3 characterised:1 except:1 total:3 experimental:4 zzzz:2 divisive:1 internal:4 modulated:2 synchronised:5 noisier:1 relevance:1 alexander:1 evaluate:2 tested:1 |
4,792 | 5,339 | Clustered factor analysis of multineuronal spike data
Lars Buesing1 , Timothy A. Machado1,2 , John P. Cunningham1 and Liam Paninski1
1
Department of Statistics, Center for Theoretical Neuroscience
& Grossman Center for the Statistics of Mind
2
Howard Hughes Medical Institute & Department of Neuroscience
Columbia University, New York, NY
{lars,cunningham,liam}@stat.columbia.edu
Abstract
High-dimensional, simultaneous recordings of neural spiking activity are often
explored, analyzed and visualized with the help of latent variable or factor models. Such models are however ill-equipped to extract structure beyond shared,
distributed aspects of firing activity across multiple cells. Here, we extend unstructured factor models by proposing a model that discovers subpopulations or
groups of cells from the pool of recorded neurons. The model combines aspects
of mixture of factor analyzer models for capturing clustering structure, and aspects of latent dynamical system models for capturing temporal dependencies. In
the resulting model, we infer the subpopulations and the latent factors from data
using variational inference and model parameters are estimated by Expectation
Maximization (EM). We also address the crucial problem of initializing parameters for EM by extending a sparse subspace clustering algorithm to integer-valued
spike count observations. We illustrate the merits of the proposed model by applying it to calcium-imaging data from spinal cord neurons, and we show that it
uncovers meaningful clustering structure in the data.
1
Introduction
Recent progress in large-scale techniques for recording neural activity has made it possible to study
the joint firing statistics of 102 up to 105 cells at single-neuron resolution. Such data sets grant
unprecedented insight into the temporal and spatial structure of neural activity and will hopefully
lead to an improved understanding of neural coding and computation.
These recording techniques have spurred the development of statistical analysis tools which help to
make accessible the information contained in simultaneously recorded activity time-series. Amongst
these tools, latent variable models prove to be particularly useful for analyzing such data sets [1,
2, 3, 4]. They aim to capture shared structure in activity across different neurons and therefore
provide valuable summary statistics of high-dimensional data that can be used for exploratory data
analysis as well as for visualization purposes. The majority of latent variable models, however,
being relatively general purpose tools, are not designed to extract additional structure from the data.
This leads to latent variables that can be hard to interpret biologically. Furthermore, additional
information from other sources, such as spatial structure or genetic cell type information, cannot be
readily integrated into these models.
An approach to leveraging simultaneous activity recordings that is complementary to applying unstructured factor models, is to infer detailed circuit properties from the data. By modelling the
detailed interactions between neurons in a local micro-circuit, multiple tools aim at inferring the
existence, type, and strength of synaptic connections between neurons [5, 6]. In spite of algorithmic
progress [7], the feasibility of this approach has only been demonstrated in circuits of up to three
1
neurons [8], as large scale data with ground truth connectivity is currently only rarely available.
This lack of validation data sets also makes it difficult to asses the impact of model mismatch and
unobserved, highly-correlated noise sources (?common input?).
Here, we propose a statistical tool for analyzing multi-cell recordings that offers a middle ground
between unstructured latent variable models and models for inferring detailed network connectivity.
The basic goal of the model is to cluster neurons into groups based on their joint activity statistics.
Clustering is a ubiquitous and valuable tool in statistics and machine learning as it often yields
interpretable structure (a partition of the data), and is of particular relevance in neuroscience because
neurons often can be categorized into distinct groups based on their morphology, physiology, genetic
identity or stimulus-response properties. In many experimental setups, side-information allowing for
a reliable supervised partitioning of the recorded neurons is not available. Hence, the main goal of
the paper is to develop a method for clustering neurons based on their activity recordings.
We model the firing time-series of a cluster of neurons using latent factors, assuming that different
clusters are described by disjoint sets of factors. The resulting model is similar to a mixture of factor
analyzers [9, 10] with Poisson observations, where each mixture component describes a subpopulation of neurons. In contrast to a mixture of factor analyzers model which assumes independent
factors, we put a Markovian prior over the factors, capturing temporal dependencies of neural activity as well as interactions between different clusters over time. The resulting model, which we
call mixture of Poisson linear dynamical systems (mixPLDS) model, is able to capture more structure using the cluster assignments compared to latent variable models previously applied to neural
recordings, while at the same time still providing low-dimensional latent trajectories for each cluster
for exploratory data analysis and visualization. In contrast to the lack of connectivity ground truth
for neurons from large-scale recordings, there are indeed large-scale activity recordings available
that exhibit rich and biologically interpretable clustering structure, allowing for a validation of the
mixPLDS model in practice.
2
2.1
Mixture of Poisson linear dynamical systems for modelling neural
subpopulations
Model definition
Let ykt denote the observed spike count of neuron k = 1, . . . , K in time-bin t = 1, . . . , T . For
the mixture of Poisson linear dynamical systems (mixPLDS) model, we assume that each neuron k
belongs to exactly one of M groups (subpopulations, clusters), indicated by the discrete (categorical)
variable sk ? {1, . . . , M }. The sk are modelled as i.i.d.:
p(s)
=
K
Y
p(sk ) =
K
Y
Disc(sk |?0 ),
(1)
k=1
k=1
where ?0 := (?10 , . . . , ?M
0 ) are the natural parameters of the categorical distribution. In the remainder of the paper we use the convention that the group-index m = 1, . . . , M is written as superscript.
dm
The activity of each subpopulation m at time t is modeled by a latent variable xm
. We
t ? R
assume that these latent variables (we will also call them factors) are jointly normal and we model
interactions between different groups by a linear dynamical system (LDS) prior:
? 1?
? 11
?? 1 ?
xt?1
xt
A
? ? ? A1M
? .. ?
?
?
? ..
.
.. ? ? ... ?
xt = ? . ? = Axt?1 + ?t = ? .
(2)
? + ?t ,
xM
t
AM 1
m
???
AM M
xM
t?1
l
where the block-matrices Aml ? Rd ?d capture the interactions between groups m and l. The
innovations ?t are i.i.d. from N (0, Q) and the starting distribution is given by x1 ? N (?1 , Q1 ). If
neuron k belongs to group m, i.e. sk = m, we model its activity ykt at time t as Poisson distributed
spike count with a log-rate given by an affine combination of the factors of group m:
zkt | sk = m
ykt | zkt , sk
m m
= Ck:
xt
? Poisson(exp(zkt + bk )),
2
(3)
(4)
m
where b ? RK captures the baseline of the firing rates. We denote with C m ? RK?d the
m
group loading matrix with rows Ck:
for neurons k in group m and fill in the remaining rows
with 0s for all neurons not in group m. We concatenate these into the total loading matrix
PM
C := (C 1 ? ? ? C M ) ? RK?d , where d := m=1 dm is the total latent dimension. If the neurons are sorted with respect to their group membership, then the total loading C has block-diagonal
structure. Further, we denote with yk: := (yk,1 ? ? ? yk,T ) the activity time series of neuron k and
m
m
1?T
use an analogous notation for xm
for n = 1, . . . , dm . The model
n := (xn,1 ? ? ? xn,T ) ? R
parameters are ? := (A, Q, Q1 , ?1 , C, b); we consider the hyper-parameters ?0 to be given and
fixed.
For known clusters s, the mixPLDS model can be regarded as a special case of the Poisson linear
dynamical system (PLDS) model [3], where the loading C is block-diagonal. For unknown group
memberships s, the mixPLDS model defined above is similar to a mixture of factor analyzers (e.g.
see [9, 10]) with Poisson observations over neurons k = 1, . . . , K. In the mixPLDS model however,
we do not restrict the factors of the mixture components to be independent but allow for interactions
over time which are modeled by a LDS.
2.2
Variational inference and parameter estimation for the mixPLDS model
When applying the mixPLDS model to data y, we are interested in inferring the group memberships
s and the latent trajectories x as well as estimating the parameters ?. For known parameters ?,
the posterior p(x, s|y, ?) (even in the special case of a single mixture component M = 1) is not
available in closed form and needs approximating. Here we propose to approximate the posterior
using variational inference with the following factorization assumption:
p(x, s|y, ?) ? q(x)q(s).
(5)
We further restrict q(x) to be a normal distribution q(x) = N (x|m, V ) with
Q mean m and covariance
V . Under the assumption (5), q(s) further factorizes into the product k q(sk ) where q(sk ) is a
categorical distribution with natural parameters ?k = (?1k , . . . , ?M
k ). The variational parameters
m, V and ? = (?1 , . . . , ?K ) are obtained by maximizing the variational lower bound of the log
marginal likelihood log p(y|?):
L(m, V, ?, ?)
=
K
X
1
log |V | ? tr[??1 V ] ? (m ? ?)> ??1 (m ? ?) +
DKL [q(sk )kp(sk )]
2
k=1
+
M X
K X
T
X
m
m
?km (ykt hm
kt ? exp(hkt + ?kt /2)) + const
(6)
m=1 k=1 t=1
m
hm
t := C mt + b,
m
m>
?m
),
t := diag(C Vt C
?km ? exp(?m
k ),
where Vt = Covq(x) [xt ] and ? ? RdT , ? ? RdT ?dT are the mean and covariance of the LDS
prior over x. The first two terms in (6) are the Kullback-Leibler divergence between the prior
p(x, s) = p(x)p(s) and its approximation q(x)q(s), penalizing a variational posterior that is far
away from the prior. The third term in (6) is given by the expected log-likelihood of the data,
promoting a posterior approximation that explains the observed data well. We optimize L in a
coordinate ascent manner, i.e. we hold ? fixed and optimize jointly over m, V and vice versa. A
naive implementation of the optimization of L over {m, V } is prohibitively costly for data sets with
large T , as the posterior covariance V has O((dT )2 ) elements and has to be optimized over the set
of semi-definite matrices. Instead of solving this large program, we apply a method proposed in
[11], where the authors show that Gaussian variational inference for latent Gaussian models with
Poisson observations can be solved more efficiently using the dual problem. We generalize their
approach to the mixture of Poisson observation model (3) considered here, and we also leverage the
Markovian structure of the LDS prior to speed up computations (see below). In the supplementary
material, we derive this approach to inference in the mixPLDS model in detail. The optimization
over ? is available in closed form and is also given in the supplementary material. We iterate updates
over m, V and ?. In practice, this method converges very quickly, often requiring only two or three
iterations to reach a reasonable convergence criterion.
The most computationally intensive part of the proposed variational inference method is the update
of m, V . Using properties of the LDS prior (i.e. the prior precision ??1 is block-tri-diagonal),
3
we can show that evaluation of L, its dual and the gradient of the latter all cost O(KT d + T d3 ),
which is the same complexity as Kalman smoothing in a LDS with Gaussian observations or a
single iteration of Laplace inference over x. While having the same cost as Laplace approximation,
variational inference has the advantage of a non-deceasing variational lower bound L, which can be
used for monitoring convergence as well as for model comparison.
We can also get estimates for the model parameters by maximizing the lower bound L over ?. To
this end, we interleave updates of ? and m, V with maximizations over ?. The latter corresponds to
standard parameter updates in a LDS model with Poisson observations and are discussed e.g. in [3].
This procedure implements variational Expectation Maximization (VEM) in the mixPLDS model.
2.3
Initialization by Poisson subspace clustering
In principle, for a given number of groups M with given dimensions d1 , . . . , dM one can estimate
the parameters of the mixPLDS using VEM as described above. In practice we find however that
this yields poor results without having reasonable initial membership assignments s, i.e. reasonable
initial values for the variational parameters ?. Furthermore, VEM requires the a priori specification
of the latent dimensions d1 , . . . , dM . Here we show that a simple extension to an existing subspace
clustering algorithm provides, given the number of groups M , a sufficiently accurate initializer for
? and allows for an informed choice for the dimensions d1 , . . . , dM .
We first illustrate the connection of the mixPLDS model to the subspace clustering problem (for
a review of the latter see e.g. [12]). Assume that we observe the log-rates zkt defined in equation
(3) directly; we denote the corresponding data matrix as Z ? RK?T . For unknown loading C, the
m
row Zk: lies on a dm -dimensional subspace spanned by the ?basis-trajectories? xm
1,: , . . . , xdm ,: , if
neuron k is in group m. If s and x are unobserved, we only know that the rows of Z lie on a union
of M subspaces of dimensions d1 , . . . , dm in an ambient space of dimension T . Reconstructing the
subspaces and the subspace assignments is known as a subspace clustering problem and connections
to mixtures of factor analyzers have been pointed out in [13]. The authors of [13] propose to solve
the subspace clustering problem by the means of the following sparse regression problem:
1
min
kZ ? W Zk2F + ?kW k1
(7)
2
W ?RK?K
s.t.
diag(W ) = 0.
This optimization can be interpreted as trying to reconstruct each row Zk: by the remaining rows
Z\k: using sparse reconstruction weights W . Intuitively, a point on a subspace can be reconstructed
using the fewest reconstruction weights by points on the same subspace, i.e. Wkl = 0 if k and l lie
on different subspaces. The symmetrized, sign-less weights |W | + |W |> are then interpreted as the
adjacency matrix of a graph and spectral clustering, with a user defined number of clusters M , is
applied to obtain a subspace clustering solution. In the noise-free case (and taking ? ? 0 in eqn 7),
under linear independence assumptions on the subspaces, [13] shows that this procedure recovers
the correct subspace assignments.
If the matrix Z is not observed directly but only through the observation model (3), the subspace
clustering approach does not directly apply. The observed data Y generated from the model (3)
is corrupted by Poisson noise and furthermore the non-linear link function transforms the union of
subspaces into a union of manifolds. We can circumvent these problems using the simple observation that not only Z but also the rows Ck: of the loading matrix C lie on a union of subspaces of
dimensions d1 , . . . , dm (where the ambient space has dimension d). This can be easily seen from the
block-diagonal structure of C (if the neurons are sorted by their true cluster assignments) mentioned
in section 2.1. Hence we can use an estimate C? of the loading C as input to the subspace clustering
optimization (7). In order to get an initial estimate C? we can use a variety of dimensionality reduction methods with exp-Poisson observations, e.g. exponential family PCA [14], a nuclear norm
based method [15], subspace identification methods [16] and EM-based PLDS learning [16]; here
we use the nuclear norm based method [15] for reasons that will become obvious below. Because of
the non-identifiability of latent factor models, these methods only yield an estimate of C ? D with an
unknown, invertible transformation D ? Rd?d . Nevertheless, the rows of C ?D still lie on a union of
subspaces (which are however not axis-aligned anymore as is the case for C), and therefore the cluster assignments can still be recovered. Given these cluster assignments, we can get initial estimates
of the non-zero rows of C m by applying nuclear norm minimization to the individual clusters. This
4
method also returns a singular value spectrum associated with each subspace, which can be used to
determine the dimension dm . One can specify e.g. a threshold ?min , and determine the dimension
dm as the number of singular values > ?min .
2.4
The full parameter estimation algorithm
We briefly summarize the proposed parameter estimation algorithm for the mixPLDS model. The
procedure requires the user to define the number of groups M . This choice can either be informed
by biological prior knowledge or one can use standard model selection methods, such as crossvalidation on the variational approximation of the marginal likelihood. We first get an initial estimate C? of the total loading matrix by nuclear-norm-penalized Poisson dimensionality reduction.
Then, subspace clustering on C? yields initial group assignments. Based on these assignments, for
each cluster we estimate the group dimension dm and the group loading C? m . Keeping the cluster
assignments fixed, we do a few VEM steps in the mixPLDS model with an initial estimation for the
loading matrix given by (C? 1 , . . . , C? M ). This last step provides reasonable initial parameters for the
parameters A, Q, Q1 , ?1 of the dynamical system prior. Finally, we do full VEM iterations in the
mixPLDS model to refine the initial parameters. We monitor the increase of the variational lower
bound L and use its increments in a termination criterion for the VEM iterations.
2.5
Non-negativity constraints on the loading C
Each component m of the mixPLDS model, representing a subpopulation of neurons, can be a
very flexible model by itself (depending on the latent dimension dm ). This flexibility can in some
situations lead to counter-intuitive clustering results. Consider the following example. Let half of
the recorded neurons oscillate in phase and the remaining neurons oscillate with a phase shift of
? relative to the first half. Depending on the context, we might be interested in clustering the first
and second half of the neurons into separate groups reflecting oscillation phase. The mixPLDS
model could however end up putting all neurons into a single cluster, by modelling them with one
oscillating latent factor that has positive loadings on the first half of neurons and negative on the
second half (or vice versa). We can prevent this behavior, by imposing element-wise non-negativity
constraints on the loading matrix C, denoted as C ? 0 (and by simultaneously constraining the
latent dimensions of each group). The constraints guarantee that the influence of each factor on its
group has the same sign across all neurons. The suitability of these constraints strongly depends on
the biological context. In the application of the mixPLDS model in section 3.2, we found them to
be essential for obtaining meaningful results.
We modify the subspace clustering initialization to respect the constraints C ? 0 in the following way. Instead of solving the unconstrained reconstruction problem (7) with respect to W , we
add non-negativity constraints W ? 0. These sign constraints restrict the points that can be reconstructed from a given set of points to the convex cone of these points (instead of the subspace
containing these points). Hence, under these assumptions, all data points in a cluster can be approximately reconstructed by a (non-negative) convex combination of some ?time-series basis?. We
empirically observed that this yields initial loading matrix estimates with only very few negative
elements (after possible row-wise sign inversions). For the full mixPLDS model we enforce C ? 0
by the reparametrization C = exp(?) and doing VEM updates on ?.
3
3.1
Experiments
Artificial data
Here we validate the parameter estimation procedure for the mixPLDS model on artificial data. We
generate 35 random ground truth mixPLDS models with M = 3, d1 = d2 = d3 = 2 and 20 observed
neurons per cluster. We sampled from each ground truth model a data set consisting of 4 i.i.d. trials
with T = 250 times steps each. Ground truth parameters were generated such that the resulting data
was sparse (12% of the bins non-empty). We compared the ability of different clustering methods
to recover the 3 clusters from each data set. We report the results in fig. 1A in terms of the fraction
of misclassified neurons (class labels were determined by majority vote in each cluster). We applied
K-Means with careful initialization of the cluster centers [17] to the data. For K-Means, we pre5
B
0.5
0.5
freq. of misclassification
freq. of misclassification
A
0
Kmeans
specCl
subCl
PsubCl
mixPLDS
0
0
0.1 0.2 0.3 0.4
assignment uncertainty
Figure 1: Finding clusters of neurons in artificial data. A: Performance of different clustering
algorithms, reported in terms of frequency of misclassified neurons, on artificial data sampled from
ground truth mixPLDs models. Red bars indicate medians and blue boxes the 25% and 75% percentiles. Standard clustering methods (data plotted in black) such as K-Means, spectral clustering
(?specCl?), and subspace clustering (?subCl?) are substantially outperformed by the two methods
proposed here (data plotted in red). Poisson subspace clustering (?PsubCl?) yielded accurate initial
cluster estimates that were significantly improved by application of the full mixPLDs model. B:
Misclassification rate as a function of the cluster assignment uncertainty for the mixPLDS model.
This shows that the posterior over cluster assignments returned by the mixPLDS model is well calibrated, as neurons with low assignment uncertainty as rarely misclassified.
processed the data in a standard way by smoothing (Gaussian kernel, standard deviation 10 timesteps), mean-centering and scaling (such that each dimension k = 1, . . . , K has variance 1). We
found K-Means yielded reasonable clusters when all populations are one-dimensional (i.e. ?m dm =
1, data not shown) but it fails when clustering multi-dimensional groups of neurons. An alternative
approach is to cluster the cross-correlation matrix of neurons (computed from pre-processed data as
above) with standard spectral clustering [18]. We found that this approach works well when all the
factors have small variances, as in this case the link function of the observation model is only mildly
non-linear. However, with growing variances of the factors (larger dynamic ranges of neurons)
spectral clustering performance quickly degrades. Standard sparse subspace clustering [13] on the
spike trains (pre-processed as above) yielded very similar results to spectral clustering. We found
our novel Poisson subspace clustering algorithm proposed in section 2.3 to robustly outperform the
other approaches, as long as reasonable amounts of data were available (roughly T > 100 for the
above system). The mixPLDS model initialized with the Poisson subspace clustering consistently
yielded the best results, as it is able to integrate information over time and denoise the observations.
One advantage of the mixPLDS model is that it not only returns cluster assignments for neurons
but also provides a measure of uncertainty over these assignments. However, variational inference
tends to return over-confident posteriors in general and the factorization approximation (5) might
yield posterior uncertainty that is uninformative. To show that the variational posterior uncertainty
is well-calibrated we computed the entropy of the posterior cluster assignment q(sk ) for all neurons
as a measure for assignment uncertainty. We binned the neurons according to their assignment
uncertainty and report the misclassification rate for each bin in fig. 1B. 89% of the neurons have low
posterior uncertainty and reside in the first bin having a low misclassification rate of ? 0.1, whereas
few neurons (5%) have an assignment uncertainty larger than 0.3 nats and they are misclassified
with a rate of ? 0.4.
3.2
Calcium imaging of spinal cord neurons
We tested the mixPLDS model on calcium imaging data obtained from an in vitro, neonatal mouse
spinal cord that expressed the calcium indicator GCaMP3 in all motor neurons. When an isolated
spinal cord is tonically excited by a cocktail of rhythmogenic drugs (5 ?M NMDA, 10 ?M 5-HT,
50 ?M DA), motor neurons begin to fire rhythmically. In this network state, spatially clustered ensembles of motor neurons fire in phase with each other [19]. Since multiple ensembles that have
distinct phase tunings can be visualized in a single imaging field, this data represents a convenient
6
B
1
1
sorted neuron #
unsorted
neuron #
A
1
70
sorted
neuron #
factors
cluster 1
70
1
4
latent dim
C
factors
cluster 2
70
sorted
neuron #
1
70
1
500
frames
Figure 2: Application of the mixPLDS model to recordings from spinal cord neurons. A, top
panel: 500 frames of input data to the mixPLDS model. Middle panel: Same data as in upper panel,
but rows are sorted by mixPLDS clusters and factor loading. Inferred latent factors (red: cluster 1,
blue: cluster 2, solid: factor 1, dashed: factor 2) are also shown. Bottom panel: Inferred (smoothed)
firing rates. B: Loading matrix C of the mixPLDS model showing how factors 1,2 of cluster 1 and
factors 3,4 of cluster 2 influence the neurons. C: Preferred phases shown as a function of (sorted)
neuron index and colored by posterior probability of belonging to cluster 1. Clearly visible are two
clusters as well as an (approximately) increasing ordering within a cluster.
setting for testing our algorithm. The data (90 second long movies) were acquired at 15 Hz from
a custom two-photon microscope equipped with a resonant scanner (downsampled from 60 Hz to
boost SNR). The frequency of the rhythmic activity was typically 0.2 Hz. In addition, aggregate motor neuron activity was simultaneously acquired with each movie using a suction electrode attached
to a ventral root. This electrophysiology recording (referred to here as ephys-trace) was used as an
external phase reference point to compute phase tuning curves for imaged neurons, which we used
to validate our mixPLDS results.
A deconvolution algorithm [20] was applied to the recorded calcium time-series to estimate the
spiking activity of 70 motor neurons. The output of the deconvolution, a 70 ? 1140 (neurons ?
frames) matrix of posterior expected number of spikes, was used as input to the mixPLDS model.
The non-empty bins of the the first 500 out of the 1140 frames of input data (thresholded at 0.1)
are shown in fig. 2A (upper panel). We used a mixPLDS model with M = 2 groups with two
latent dimensions each, i.e. d1 = d2 = 2. We imposed the non-negativity constraints C ? 0 on the
loading matrix; these were found to be crucial for finding a meaningful clustering of the neurons,
as discussed above. The mixPLDS clustering reveals two groups with strongly periodic but phaseshifted population activities, as can be seen from the inferred latent factors shown in fig. 2A (middle
panel, factors of cluster 1 shown in red, factors of cluster 2 in blue). For each cluster, the model
learned a stronger (higher variance) latent factor (solid line) and a weaker one (dashed line); we
interpret the former as capturing the main activity structure in a cluster and the latter as describing
deviations. Based on the estimated mixPLDS model, we sorted the neurons for visualization into
two clusters according to their most likely cluster assignment argmaxsk =1,2 q(sk ). Within each
cluster, we sorted the neurons according to the ratio of the loading coefficient onto the stronger
factor over the loading onto the weaker factor. Re-plotting the spike-raster with this sorting in fig.
2A (middle panel) reveals interesting structure. First, it shows that the initial choice of two clusters
was well justified for this data set. Second, the sorting reveals that the majority of neurons tend to
7
fire at a preferred phase relative to the oscillation cycle, and the mixPLDS-based sorting corresponds
to an increasing ordering of preferred phases. Fig. 2B shows the loading matrix C of the mixPLDS,
which is found to be approximately block-diagonal.
On this data set we also have the opportunity to validate the unsupervised clustering by taking into
account the simultaneously recorded ephys-trace. We computed for each neuron a phase tuning
curve based on the ephys-trace history of the last 80 times steps (estimated via L2 regularized generalized linear model estimation, with an exp-Poisson observation model). For each neuron, we
extracted the peak location of this phase tuning curve, which we call the preferred phase. Fig. 2C
shows these preferred phases as a function of (sorted) neuron index, revealing that the two clusters
found by the mixPLDS model coincide well with the two modes of the bi-model distribution of preferred phases. Furthermore, within each cluster, the preferred phases are (approximately) increasing,
showing that the mixPLDS-sorting of neurons reflects the phase-relation of the neurons to the global,
oscillatory ephys-trace. We emphasize that the latter was not used for fitting the mixPLDS; i.e., this
constitutes an independent validation of our results.
We conclude that the mixPLDS model successfully uncovered clustering structure from the recordings that can be validated using the side information from electrophysiological tuning, and furthermore allowed for a meaningful sorting within each cluster capturing neural response properties. In
addition, the mixPLDS model leverages the temporal structure in recordings, automatically optimizing for the temporal smoothness level and revealing the main time-constants in the data (in the above
data set 1.8 and 6.5 sec) as well as main oscillation frequencies (0.2 and 0.45Hz). Furthermore, either the latent trajectories or the inferred firing rates shown in fig. 2A can be used as smoothed
proxies for their corresponding population activities for subsequent analyses.
4
Discussion
One can generalize the mixPLDS model in several ways. Here we assumed that, given the latent factors, all neurons fire independently. This is presumably a good assumption if the recorded neurons
are spatially distant, but it might break down if neurons are densely sampled from a local population
and have strong, monosynaptic connections. This more general case can be accounted for by incorporating direct interaction terms between neurons into the observation model in the spirit of coupled
GLMs (see [21]); inference and parameter learning are still tractable in this model using VEM. Furthermore, in addition to the activity recordings, one might have access to other covariates that are
informative about the clustering structure of the population, such as cell location, genetic markers,
or cell morphology. We can add such data as additional observations into the mixPLDS model to
facilitate clustering of the cells. An especially relevant example are stimulus-response properties of
cells. We can add a mixture model over receptive-field parameters using the cluster assignments s.
This extension would provide a clustering of neurons based on their joint activity statistics (such as
shared trial-to-trial variability) as well as on their receptive field properties.
We presented three technical contributions, that we expect to be useful outside the context of the
mixPLDS model. First, we proposed a simple extension of the sparse subspace clustering algorithm
to Poisson observations. We showed that if the dimension of the union of subspaces is much smaller
than the ambient dimension, our method substantially outperforms other approaches. Second, we
introduced a version of subspace clustering with non-negativity constraints on the reconstruction
weights, which therefore clusters points into convex cones. We expect this variant to be particularly
useful when clustering activity traces of cells, allowing for separating anti-phasic oscillations. Third,
we applied the dual variational inference approach of [11] to a model with a Markovian prior and
with mixtures of Poisson observations. The resulting inference method proved itself numerically
robust, and we expect it to be a valuable tool for analyzing time-series of sparse count variables.
Acknowledgements This work was supported by Simons Foundation (SCGB#325171 and
SCGB#325233), Grossman Center at Columbia University, and Gatsby Charitable Trust as well
as grants MURI W911NF-12-1-0594 from the ARO, vN00014-14-1-0243 from the ONR, W91NF14-1-0269 from DARPA and an NSF CAREER award (L.P.).
8
References
[1] Anne C Smith and Emery N Brown. Estimating a state-space model from point process observations.
Neural Computation, 15(5):965?991, 2003.
[2] Lauren M Jones, Alfredo Fontanini, Brian F Sadacca, Paul Miller, and Donald B Katz. Natural stimuli
evoke dynamic sequences of states in sensory cortical ensembles. Proceedings of the National Academy
of Sciences, 104(47):18772?18777, 2007.
[3] Jakob H Macke, Lars Buesing, John P Cunningham, M Yu Byron, Krishna V Shenoy, and Maneesh
Sahani. Empirical models of spiking in neural populations. In NIPS, pages 1350?1358, 2011.
[4] Byron M Yu, John P Cunningham, Gopal Santhanam, Stephen I Ryu, Krishna V Shenoy, and Maneesh
Sahani. Gaussian-process factor analysis for low-dimensional single-trial analysis of neural population
activity. In NIPS, pages 1881?1888, 2008.
[5] Murat Okatan, Matthew A Wilson, and Emery N Brown. Analyzing functional connectivity using a
network likelihood model of ensemble neural spiking activity. Neural Computation, 17(9):1927?1961,
2005.
[6] Yuriy Mishchenko, Joshua T Vogelstein, Liam Paninski, et al. A Bayesian approach for inferring neuronal
connectivity from calcium fluorescent imaging data. The Annals of Applied Statistics, 5(2B):1229?1261,
2011.
[7] Suraj Keshri, Eftychios Pnevmatikakis, Ari Pakman, Ben Shababo, and Liam Paninski. A shotgun
sampling solution for the common input problem in neural connectivity inference. arXiv preprint
arXiv:1309.3724, 2013.
[8] Felipe Gerhard, Tilman Kispersky, Gabrielle J Gutierrez, Eve Marder, Mark Kramer, and Uri Eden. Successful reconstruction of a physiological circuit with known connectivity from spiking activity alone.
PLoS computational biology, 9(7):e1003138, 2013.
[9] Michael E Tipping and Christopher M Bishop. Mixtures of probabilistic principal component analyzers.
Neural Computation, 11(2):443?482, 1999.
[10] Zoubin Ghahramani, Geoffrey E Hinton, et al. The EM algorithm for mixtures of factor analyzers. Technical report, Technical Report CRG-TR-96-1, University of Toronto, 1996.
[11] Mohammad Emtiyaz Khan, Aleksandr Aravkin, Michael Friedlander, and Matthias Seeger. Fast dual
variational inference for non-conjugate latent gaussian models. In Proceedings of The 30th International
Conference on Machine Learning, pages 951?959, 2013.
[12] Ren?e Vidal. A tutorial on subspace clustering. IEEE Signal Processing Magazine, 28(2):52?68, 2010.
[13] Ehsan Elhamifar and Ren?e Vidal. Sparse subspace clustering: Algorithm, theory, and applications. Pattern
Analysis and Machine Intelligence, IEEE Transactions on, 35(11):2765?2781, Nov 2013.
[14] Michael Collins, Sanjoy Dasgupta, and Robert E Schapire. A generalization of principal component
analysis to the exponential family. In NIPS, volume 13, page 23, 2001.
[15] David Pfau, Eftychios A Pnevmatikakis, and Liam Paninski. Robust learning of low-dimensional dynamics from large neural ensembles. In NIPS, pages 2391?2399, 2013.
[16] Lars Buesing, Jakob H Macke, and Maneesh Sahani. Spectral learning of linear dynamics from
generalised-linear observations with application to neural population data. In NIPS, pages 1691?1699,
2012.
[17] David Arthur and Sergei Vassilvitskii. k-means++: The advantages of careful seeding. In Proceedings
of the eighteenth annual ACM-SIAM symposium on Discrete algorithms, pages 1027?1035. Society for
Industrial and Applied Mathematics, 2007.
[18] Andrew Y Ng, Michael I Jordan, and Yair Weiss. On spectral clustering1 analysis and an algorithm. Proceedings of Advances in Neural Information Processing Systems. Cambridge, MA: MIT Press, 14:849?
856, 2001.
[19] Timothy A. Machado, Eftychios Pnevmatikakis, Liam Paninski, Thomas M. Jessell, and Andrew Miri.
Functional organization of spinal motor neurons revealed by ensemble imaging. In 79th Cold Spring
Harbor Symposium on Quantitative Biology Cognition, 2014.
[20] E. A. Pnevmatikakis, Y. Gao, D. Soudry, D. Pfau, C. Lacefield, K. Poskanzer, R. Bruno, R. Yuste, and
L. Paninski. A structured matrix factorization framework for large scale calcium imaging data analysis.
ArXiv e-prints, September 2014.
[21] Jonathan W Pillow, Jonathon Shlens, Liam Paninski, Alexander Sher, Alan M Litke, EJ Chichilnisky, and
Eero P Simoncelli. Spatio-temporal correlations and visual signalling in a complete neuronal population.
Nature, 454(7207):995?999, 2008.
9
| 5339 |@word trial:4 briefly:1 middle:4 interleave:1 loading:20 norm:4 inversion:1 stronger:2 version:1 termination:1 km:2 d2:2 uncovers:1 covariance:3 excited:1 q1:3 tr:2 solid:2 reduction:2 initial:12 series:6 uncovered:1 genetic:3 outperforms:1 existing:1 recovered:1 anne:1 written:1 readily:1 john:3 sergei:1 concatenate:1 partition:1 visible:1 subsequent:1 distant:1 informative:1 motor:6 seeding:1 designed:1 interpretable:2 update:5 alone:1 half:5 intelligence:1 signalling:1 shababo:1 smith:1 colored:1 provides:3 location:2 toronto:1 direct:1 become:1 symposium:2 prove:1 combine:1 fitting:1 manner:1 acquired:2 expected:2 indeed:1 roughly:1 behavior:1 growing:1 multi:2 morphology:2 automatically:1 equipped:2 increasing:3 begin:1 estimating:2 notation:1 monosynaptic:1 circuit:4 panel:7 interpreted:2 substantially:2 proposing:1 informed:2 unobserved:2 transformation:1 finding:2 guarantee:1 temporal:6 a1m:1 quantitative:1 exactly:1 axt:1 prohibitively:1 partitioning:1 medical:1 grant:2 okatan:1 shenoy:2 positive:1 generalised:1 local:2 modify:1 tends:1 soudry:1 aleksandr:1 analyzing:4 firing:6 approximately:4 might:4 black:1 initialization:3 factorization:3 liam:7 range:1 bi:1 testing:1 hughes:1 practice:3 block:6 definite:1 implement:1 union:6 procedure:4 cold:1 empirical:1 drug:1 maneesh:3 physiology:1 significantly:1 convenient:1 revealing:2 pre:2 subpopulation:7 downsampled:1 spite:1 donald:1 zoubin:1 get:4 cannot:1 onto:2 selection:1 put:1 context:3 applying:4 influence:2 optimize:2 imposed:1 demonstrated:1 center:4 maximizing:2 eighteenth:1 starting:1 independently:1 convex:3 resolution:1 unstructured:3 insight:1 d1:7 regarded:1 fill:1 spanned:1 nuclear:4 shlens:1 population:9 exploratory:2 coordinate:1 increment:1 analogous:1 laplace:2 annals:1 gerhard:1 user:2 magazine:1 element:3 particularly:2 muri:1 observed:6 bottom:1 preprint:1 initializing:1 paninski1:1 capture:4 solved:1 cord:5 cycle:1 ordering:2 counter:1 plo:1 valuable:3 yk:3 mentioned:1 complexity:1 nats:1 covariates:1 dynamic:4 solving:2 basis:2 easily:1 joint:3 darpa:1 fewest:1 train:1 distinct:2 fast:1 kp:1 artificial:4 aggregate:1 hyper:1 outside:1 supplementary:2 valued:1 solve:1 larger:2 reconstruct:1 ability:1 statistic:8 jointly:2 itself:2 superscript:1 covq:1 advantage:3 sequence:1 unprecedented:1 matthias:1 propose:3 reconstruction:5 interaction:6 product:1 poskanzer:1 remainder:1 aro:1 aligned:1 relevant:1 flexibility:1 academy:1 intuitive:1 validate:3 lauren:1 crossvalidation:1 convergence:2 felipe:1 cluster:52 empty:2 extending:1 electrode:1 oscillating:1 hkt:1 converges:1 emery:2 ben:1 help:2 illustrate:2 develop:1 derive:1 stat:1 depending:2 andrew:2 progress:2 strong:1 indicate:1 convention:1 aravkin:1 aml:1 correct:1 lars:4 zkt:4 jonathon:1 material:2 bin:5 explains:1 adjacency:1 clustered:2 generalization:1 suitability:1 biological:2 brian:1 crg:1 extension:3 hold:1 scanner:1 sufficiently:1 considered:1 ground:7 normal:2 exp:6 presumably:1 cognition:1 algorithmic:1 matthew:1 ventral:1 purpose:2 tonically:1 estimation:6 outperformed:1 label:1 currently:1 pnevmatikakis:4 vice:2 gutierrez:1 successfully:1 tool:7 reflects:1 minimization:1 mit:1 clearly:1 gaussian:6 gopal:1 aim:2 ck:3 ej:1 factorizes:1 wilson:1 validated:1 consistently:1 modelling:3 likelihood:4 contrast:2 seeger:1 litke:1 baseline:1 am:2 industrial:1 dim:1 inference:14 membership:4 integrated:1 typically:1 cunningham:3 relation:1 misclassified:4 interested:2 dual:4 ill:1 flexible:1 denoted:1 priori:1 development:1 smoothing:2 spatial:2 special:2 marginal:2 field:3 having:3 ng:1 sampling:1 biology:2 kw:1 represents:1 jones:1 unsupervised:1 constitutes:1 yu:2 report:4 stimulus:3 micro:1 few:3 simultaneously:4 divergence:1 densely:1 individual:1 national:1 phase:17 consisting:1 fire:4 organization:1 highly:1 custom:1 evaluation:1 analyzed:1 mixture:16 kt:3 accurate:2 ambient:3 arthur:1 initialized:1 re:1 plotted:2 isolated:1 theoretical:1 markovian:3 w911nf:1 assignment:22 maximization:3 cost:2 deviation:2 snr:1 successful:1 reported:1 dependency:2 corrupted:1 periodic:1 calibrated:2 confident:1 peak:1 international:1 siam:1 accessible:1 probabilistic:1 pool:1 invertible:1 michael:4 quickly:2 mouse:1 connectivity:7 recorded:7 initializer:1 containing:1 external:1 macke:2 grossman:2 return:3 account:1 photon:1 coding:1 sec:1 coefficient:1 depends:1 root:1 break:1 closed:2 doing:1 red:4 recover:1 reparametrization:1 identifiability:1 simon:1 vem:8 contribution:1 ass:1 variance:4 efficiently:1 ensemble:6 yield:6 miller:1 emtiyaz:1 generalize:2 modelled:1 lds:7 identification:1 buesing:2 bayesian:1 disc:1 ren:2 trajectory:4 monitoring:1 history:1 simultaneous:2 oscillatory:1 reach:1 synaptic:1 rdt:2 definition:1 centering:1 raster:1 frequency:3 obvious:1 dm:14 associated:1 recovers:1 sampled:3 proved:1 knowledge:1 dimensionality:2 ubiquitous:1 electrophysiological:1 nmda:1 reflecting:1 higher:1 dt:2 supervised:1 tipping:1 response:3 improved:2 specify:1 wei:1 box:1 strongly:2 furthermore:7 correlation:2 glms:1 eqn:1 trust:1 christopher:1 marker:1 lack:2 hopefully:1 mode:1 indicated:1 cunningham1:1 facilitate:1 requiring:1 true:1 brown:2 former:1 hence:3 spatially:2 imaged:1 leibler:1 freq:2 percentile:1 criterion:2 generalized:1 trying:1 clustering1:1 alfredo:1 mohammad:1 complete:1 variational:18 wise:2 discovers:1 novel:1 ari:1 common:2 machado:1 functional:2 spiking:5 mt:1 empirically:1 spinal:6 vitro:1 attached:1 volume:1 extend:1 discussed:2 katz:1 interpret:2 numerically:1 versa:2 imposing:1 cambridge:1 smoothness:1 rd:2 unconstrained:1 tuning:5 pm:1 mathematics:1 pointed:1 analyzer:7 bruno:1 specification:1 access:1 add:3 posterior:13 recent:1 showed:1 optimizing:1 belongs:2 onr:1 vt:2 unsorted:1 joshua:1 seen:2 krishna:2 additional:3 determine:2 signal:1 dashed:2 semi:1 multiple:3 full:4 simoncelli:1 infer:2 vogelstein:1 stephen:1 alan:1 technical:3 pakman:1 offer:1 cross:1 long:2 award:1 dkl:1 feasibility:1 impact:1 variant:1 basic:1 regression:1 expectation:2 poisson:21 arxiv:3 iteration:4 kernel:1 cell:10 microscope:1 justified:1 whereas:1 uninformative:1 addition:3 singular:2 source:2 median:1 crucial:2 ascent:1 tri:1 recording:14 hz:4 tend:1 byron:2 leveraging:1 spirit:1 jordan:1 integer:1 call:3 eve:1 leverage:2 constraining:1 revealed:1 iterate:1 independence:1 variety:1 timesteps:1 harbor:1 restrict:3 rhythmically:1 intensive:1 shift:1 eftychios:3 vassilvitskii:1 pca:1 shotgun:1 multineuronal:1 returned:1 york:1 oscillate:2 cocktail:1 useful:3 detailed:3 transforms:1 amount:1 visualized:2 processed:3 generate:1 schapire:1 outperform:1 nsf:1 tutorial:1 sign:4 neuroscience:3 estimated:3 disjoint:1 per:1 sadacca:1 blue:3 discrete:2 dasgupta:1 santhanam:1 group:28 putting:1 nevertheless:1 threshold:1 monitor:1 eden:1 d3:2 prevent:1 penalizing:1 thresholded:1 ht:1 imaging:7 graph:1 fraction:1 cone:2 uncertainty:10 family:2 reasonable:6 resonant:1 oscillation:4 scaling:1 capturing:5 bound:4 refine:1 yielded:4 activity:26 annual:1 strength:1 binned:1 marder:1 constraint:9 aspect:3 speed:1 min:3 spring:1 relatively:1 department:2 miri:1 according:3 structured:1 combination:2 poor:1 belonging:1 conjugate:1 across:3 describes:1 em:4 reconstructing:1 smaller:1 biologically:2 intuitively:1 computationally:1 equation:1 visualization:3 previously:1 describing:1 count:4 phasic:1 mind:1 merit:1 know:1 tractable:1 end:2 available:6 plds:2 vidal:2 promoting:1 apply:2 observe:1 away:1 spectral:7 enforce:1 anymore:1 robustly:1 alternative:1 yair:1 symmetrized:1 lacefield:1 existence:1 thomas:1 assumes:1 clustering:44 spurred:1 remaining:3 top:1 opportunity:1 const:1 neonatal:1 k1:1 especially:1 ghahramani:1 approximating:1 society:1 print:1 spike:7 degrades:1 costly:1 receptive:2 diagonal:5 exhibit:1 amongst:1 gradient:1 subspace:36 september:1 link:2 separate:1 separating:1 majority:3 manifold:1 reason:1 assuming:1 kalman:1 index:3 modeled:2 providing:1 ratio:1 innovation:1 difficult:1 setup:1 keshri:1 robert:1 trace:5 negative:3 implementation:1 calcium:7 murat:1 unknown:3 zk2f:1 allowing:3 upper:2 neuron:74 observation:19 howard:1 anti:1 tilman:1 situation:1 hinton:1 variability:1 wkl:1 frame:4 jakob:2 smoothed:2 inferred:4 bk:1 introduced:1 david:2 chichilnisky:1 khan:1 connection:4 optimized:1 suraj:1 pfau:2 learned:1 ryu:1 boost:1 nip:5 address:1 beyond:1 able:2 bar:1 dynamical:7 below:2 mismatch:1 xm:5 pattern:1 summarize:1 program:1 reliable:1 misclassification:5 natural:3 circumvent:1 regularized:1 indicator:1 representing:1 movie:2 axis:1 categorical:3 hm:2 sher:1 extract:2 columbia:3 naive:1 negativity:5 coupled:1 sahani:3 prior:11 understanding:1 review:1 l2:1 acknowledgement:1 friedlander:1 relative:2 expect:3 interesting:1 yuste:1 fluorescent:1 geoffrey:1 jessell:1 validation:3 foundation:1 integrate:1 affine:1 proxy:1 principle:1 plotting:1 charitable:1 row:11 summary:1 penalized:1 accounted:1 last:2 free:1 keeping:1 supported:1 yuriy:1 fontanini:1 side:2 allow:1 weaker:2 institute:1 taking:2 rhythmic:1 sparse:8 distributed:2 curve:3 dimension:17 xn:2 cortical:1 pillow:1 rich:1 kz:1 sensory:1 author:2 made:1 reside:1 coincide:1 far:1 transaction:1 reconstructed:3 approximate:1 emphasize:1 nov:1 preferred:7 kullback:1 evoke:1 global:1 reveals:3 scgb:2 conclude:1 assumed:1 eero:1 spatio:1 spectrum:1 latent:28 sk:13 nature:1 suction:1 zk:2 robust:2 career:1 obtaining:1 ehsan:1 diag:2 da:1 main:4 noise:3 paul:1 denoise:1 mishchenko:1 allowed:1 complementary:1 categorized:1 x1:1 neuronal:2 fig:8 referred:1 gatsby:1 ny:1 precision:1 fails:1 inferring:4 exponential:2 lie:5 third:2 rk:5 down:1 xt:5 bishop:1 showing:2 explored:1 physiological:1 deconvolution:2 essential:1 incorporating:1 elhamifar:1 uri:1 mildly:1 sorting:5 entropy:1 electrophysiology:1 timothy:2 paninski:6 likely:1 gao:1 visual:1 expressed:1 contained:1 ykt:4 corresponds:2 truth:6 extracted:1 acm:1 ma:1 goal:2 identity:1 sorted:10 kmeans:1 careful:2 kramer:1 shared:3 hard:1 determined:1 principal:2 total:4 sanjoy:1 experimental:1 vote:1 meaningful:4 rarely:2 mark:1 latter:5 collins:1 jonathan:1 relevance:1 alexander:1 tested:1 correlated:1 |
4,793 | 534 | Learning to Make Coherent Predictions in
Domains with Discontinuities
Suzanna Becker and Geoffrey E. Hinton
Department of Computer Science, University of Toronto
Toronto, Ontario, Canada M5S 1A4
Abstract
We have previously described an unsupervised learning procedure that
discovers spatially coherent propertit>_<; of the world by maximizing the information that parameters extracted from different parts of the sensory
input convey about some common underlying cause. When given random
dot stereograms of curved surfaces, this procedure learns to extract surface depth because that is the property that is coherent across space. It
also learns how to interpolate the depth at one location from the depths
at nearby locations (Becker and Hint.oll. 1992). 1n this paper, we propose two new models which handle surfaces with discontinuities. The first
model attempts to detect cases of discontinuities and reject them. The
second model develops a mixture of expert interpolators. It learns to detect the locations of discontinuities and to invoke specialized, asymmetric
interpolators that do not cross the discontinuities .
1
Introd uction
Standard backpropagation is implausible as a model of perceptual learning because
it requires an external teacher to specify the desired output of the network. We
have shown (Becker and Hinton, 1992) how the external teacher can be replaced
by internally derived teaching signals. These signals are generated by using the
assumption that different parts of the perceptual input have common causes in
the external world. Small modules that look at separate but related parts of the
perceptual input discover these common causes by striving to produce outputs that
agree with each other (see Figure 1 a). The modules may look at different modalities
(e.g. vision and touch), or the same modality at different times (e.g. the consecutive
2-D views of a rotating 3-D object), or even spatially adjacent parts of the same
image. In previous work, we showed that when our learning procedure is applied
372
Learning to Make Coherent Predictions in Domains with Discontinuities
to adjacent patches of 2-dimensional images, it allows a neural network that has no
prior knowledge of the third dimension to discover depth in random dot stereograms
of curved surfaces. A more general version of the method allows the network to
discover the best way of interpolating the depth at one location from the depths
at nearby locations. We first summarize this earlier work, and then introduce
two new models which allow coherent predictions to be made in the presence of
discontinuities.
a)
left
rightm~m~
patch A
patch B
Figure 1: a) Two modules that receive input from corresponding parts of stereo
images. The first module receives input from stereo patch A, consisting of a horizontal strip from the left image (striped) and a corresponding strip from the right
image (hatched). The second module receives input from an adjacent stereo patch
B . The modules try to make their outputs, d a and db, convey as much information as possible about some underlying signal (i. e., the depth) which is common to
both patches. b) The architecture of the interpolating network, consisting of multiple
copies of modules like those in a) plus a layer of interpolating units. The network
tries to maximize the information that the locally extracted parameter de and the
contextually predicted parameter de convey about some common underlying signal.
We actually used 10 modules and the central 6 modules tried to maximize agreement
between their outputs and contextually predicted values. We used weight averaging
to constrain the interpolating function to be identical for all modules.
2
Learning spatially coherent features in images
The simplest way to get the outputs of two modules to agree is to use the squared
difference between the outputs as a cost function, and to adjust the weights in each
module so as to minimize this cost. Unfortunately, this usually causes each module
to produce the same constant output that is unaffected by the input to the module
and therefore conveys no information about it. What we want is for the outputs
of two modules to agree closely (i.e. to have a small expected squared difference)
relative to how much they both vary as the input is varied. When this happens, the
two modules must be responding to something that is common to their two inputs.
In the special case when the outputs, d a , db, of the two modules are scalars, a good
373
374
Becker and Hinton
measure of agreement is:
(1)
where V is the variance over the training cases. If d a and db are both versions
of the same underlying Gaussian signal that have been corrupted by independent
Gaussian noise, it can be shown that I is the mutual information between the
underlying signal and the average of d a and db. By maximizing I we force the two
modules to extract as pure a version as possible of the underlying common signal.
2.1
The basic stereo net
We have shown how this principle can be applied to a multi-layer network that learns
to extract depth from random dot stereograms (Becker and Hinton, 1992). Each
network module received input from a patch of a left image and a corresponding
patch of a right image, as shown in Figure 1 a). Adjacent modules received input
from adjacent stereo image patches, and learned to extract depth by trying to
maximize agreement between their outputs. The real-valued depth (relative to the
plane of fixation) of each patch of the surface gives rise to a disparity between
features in the left and right images; since that disparity is the only property that
is coherent across each stereo image, the output units of modules were able to learn
to accurately detect relative depth.
2.2
The interpolating net
The basic stereo net uses a very simple model of coherence in which an underlying
parameter at one location is assumed to be approximately equal to the parameter at
a neighbouring location. This model is fine for the depth of fronto-parallel surfaces
but it is far from the best model of slanted or curved surfaces. Fortunately, we can
use a far more general model of coherence in which the parameter at one location
is assumed to be an unknown linear function of the parameters at nearby locations.
The particular linear function that is appropriate can be learned by the network.
We used a network of the type shown in Figure 1 b). The depth computed locally
by a module, dc, was compared with the depth predicted by a linear combination de
of the outputs of nearby modules, and the network tried to maximize the agreement
between de and de.
The contextual prediction, dc, was produced by computing a weighted sum of
the outputs of two adjacent modules on either side. The interpolating weights
used in this sum, and all other weights in the network, were adjusted so as to
maximize agreement between locally computed and contextually predicted depths.
To speed the learning, we first trained the lower layers of the network as before, so that agreement was maximized between neighbouring locally computed
outputs. This made it easier to learn good interpolating weights. When the
network was trained on stereograms of cubic surfaces, it learned interpolating
weights of -0.147,0.675,0.656, -0.131 (Becker and Hinton, 1992). Given noise
free estimates of local depth, the optimal linear interpolator for a cubic surfa.ce
is -0.167,0.667,0.667, -0.167.
Learning to Make Coherent Predictions in Domains with Discontinuities
3
Throwing out discontinuities
If the surface is continuous, the depth at one patch can be accurately predicted from
the depths of two patches on either side. If, however, the training data contains cases
in which there are depth discontinuities (see figure 2) the interpolator will also try
to model these cases and this will contribute considerable noise to the interpolating
weights and to the depth estimates. One way of reducing this noise is to treat the
discontinuity cases as outliers and to throw them out. Rather than making a hard
decision about whether a case is an outlier, we make a soft decision by using a
mixture model. For each training case, the network compares the locally extracted
depth, dc, with the depth predicted from the nearby context, de. It assumes that
de - de is drawn from a zero-mean Gaussian if it is a continuity case and from a
uniform distribution if it is a discontinuity case. It can then estimate the probability
of a continuity case:
--------
Spline
curve
Left
Image
I 1
l
Right
Image
"I
I
II I II \
I I
I
II I
III
1111
Iii
II ill I
til
IIII ,i I I \
I,
I"'I
II
I I ,,\
II
I
1'1
111
II
III II
I
Figure 2: Top: A curved surface strip with a discontinuity created by fitting 2
cubic splines through randomly chosen control points, 25 pixels apart, separated by
a depth discontinuity. Feature points are randomly scattered on each spline with an
average of 0.22 features per pixel. Bottom: A stereo pair of "intensity" images
of the surface strip formed by taking two different projections of the feature points,
filtering them through a gaussian, and sampling the filtered projections at evenly
spaced sample points. The sample values in corresponding patches of the two images
are used as the inputs to a module. The depth of the surface for a particular zmage
region is directly related to the disparity between corresponding features in the left
and right patch. Disparity ranges continuously from -1 to + 1 image pixels. Each
stereo image was 120 pixels wide and divided into 10 receptive fields 10 pixels wide
and separated by 2 pixel gaps, as input for the networks shown in figure 1. The
receptive field of an interpolating unit spanned 58 image pixels, and discontinuities
were randomly located a minimum of 40 pixels apart, so only rarely would more
than one discontinuity lie within an interpolator's receptive field.
375
376
Becker and Hinton
(2)
where N is a gaussian, and kdi3eont is a constant representing a uniform density.
1
We can now optimize the average information de and de transmit about their common cause. We assume that no information is transmitted in discontinuity cases,
so the average information depends on the probability of continuity and on the
variance of de + de and de - de measured only in the continuity cases.
(3)
We tried several variations of this mixture approach. The network is quite good at
rejecting the discontinuity cases, but this leads to only a modest improvement in
the performance of the interpolator. In cases where there is a depth discontinuity
between d a and db or between dd and de the interpolator works moderately well
because the weights on d a or de are small. Because of the term Peont in equation
3 there is pressure to include these cases as continuity cases, so they probably
contribute noise to the interpolating weights. In the next section we show how to
avoid making a forced choice between rejecting these cases or treating them just
like all the other continuity cases.
4
Learning a mixture of expert interpolators
The presence of a depth discontinuity somewhere within a strip of five adjacent
patches does not entirely eliminate the coherence of depth across these patches. It
just restricts the range over which this coherence operates. So instead of throwing
out cases that contain a discontinuity, the network could try to develop a number
of different, specialized interpolators each of which captures the particular type of
coherence that remains in the presence of a discontinuity at a particular location.
If, for example, there is a depth discontinuity between de and de, an extrapolator
with weights of -1.0, +2.0,0, would be an appropriate predictor of de .
?
Figure 3 shows the system of five expert interpolators that we used for predicting
de from the neighboring depths. To allow the system to invoke the appropriate
interpolator, each expert has its own "controller" which must learn to detect the
presence of a discontinuity at a particular location (or the absence of a discontinuity in the case of the interpolator for pure continuity cases). The outputs of the
controllers are normalized, as shown in figure 3, so that they form a probability distribution. We can think of these normalized outputs as the probability with which
the system selects a particular expert. The controllers get to see all five local depth
estimates and most of them learn to detect particular depth discontinuities by using
large weights of opposite sign on the local depth estimates of neighboring patches.
lWe empirically select a good (fixed) value of kdiseont, and we choose a starting value
of Veont{de - de) (some proportion of the initial variance of de - de), and gradually shrink
it during learning.
Learning to Make Coherent Predictions in Domains with Discontinuities
expert 1
expert 2
expert 3
de , I
PI
de ,2
P2
de ,3
P3
Xl
Normalization
Pi
=
J
expert 4
expert 5
X2
controller 2
X3
controller 3
e x ,2
I: e
controller 1
x J? 2
de ,4
P4
X4
de ,5
P5
X5
controller 4
controller 5
Figure 3: The architecture of the mixture of interpolators and discontinuzty detec.
tors . We actually used a larger modular network and equality constraints between
modules, as described in figure 1 b), with 6 copies of the architecture shown here .
Each copy received input from different but overlapping parts of the input.
Figure 4 shows the weights learned by the experts and by their controllers. As
expected, there is one interpolator (the top one) that is appropriate for continuity
cases and four other interpolators that are appropriate for the four different locations of a discontinuity. In interpreting the weights of the controllers it is important
to remember that a controller which produces a small X value for a particular case
may nevertheless assign high probability to its expert if all the other controllers
produce even smaller x values.
4.1
The learning procedure
In the example presented here, we first trained the network shown in figure 1b)
on images with discontinuities. We then used the outputs of the depth extracting
layer, d a , ... ,de as the inputs to the expert interpolators and their controllers. The
system learned a set of expert interpolators without backpropagating derivatives all
the way down to the weights of the local depth extracting modules. So the local
depth estimates d a , ... ,de did not change as the interpolators were learned .
To train the system we used an unsupervised version of the competing experts
algorithm described by Jacobs, Jordan, Nowlan and Hinton (1991) . The output of
the ith expert, de,i, is treated as the mean of a Gaussian distribution with variance 0- 2
and the normalized output of each controller, Pi , is treated as the mixing proportion
of that Gaussian. So, for each training case, the outputs of the experts and their
controllers define a probability distribution that is a mixture of Gaussians . The aim
377
378
Becker and Hinton
,a)
b)
Mean output vs. distance
to nearest discontinuity
Interpolator Discontinuity
weights
detector weights
1.00
iiiiiI
iiiii2
~
0.95
~T-
..
"
iI' I ,,,""
. \ , ,, ,~,,
, I
I i , ?, ,, \
,~ : I ,, ,, , \,
/~ ; iI ,
,
? ~ 1
I,
:, ,, ,\
.? I
l ~, \'
\
I'
0 .90
0 .15
0.10
0.75
YU.-4- uair -
.
0.70
0 .6.5
.:'
:
.
0 .60
0.55
0.50
0.45
0.40
ti.,
:;
I
~
I'
0.10
02S
020
I
,-~:'. -f
.....' ,
,
I
0.15
\
,
\
1
,
I
I
I
I
I
,
i
!
I '
I'
i I
\'
I
"I,
I
OAS
,
~
. ..
\
I
0.10
:'
\?
\
I
"
I~ :i'I'i ""
I: ' ! ," ,
I .
:\ ,
I
, ,
i
I ,,
/
i ~
O.lS
j'
I
0.00
.oAS
pudol
?60.00
.4()00
-20.00
0 .00
20.00
40.00
60.00
Figure 4: a) Typical weights lear~ed by the five competing interpolators and corresponding five discontinuity detectors. Positive weights are shown in white, and
negative weights in black. b) The mean probabilities computed by each discontinuity
detector are plotted against the the distance from the center of the units' receptive
field to the nearest discontinuity. The probabilistic outputs are averaged over an
ensemble of 1000 test cases. If the nearest discontinuity is beyond ? thirty pixels,
it is outside the units' receptive field and the case is therefore a continuity example.
of the learning is to maximize the log probability density of the desired output, de,
under this mixture of Gaussians distribution. For a particular training case this log
probability is given by :
'"
log P( de) = log L.,; Pi
.
I
1
to=
v2~u
exp
((d
-
e
ei )2)
-d
2 2 '
u
(4)
By taking derivatives of this objective function we can simultaneously learn the
weights in the experts and in the controllers. For the results shown here, the
nework was trained for 30 conjugate gradient iterations on a set of 1000 random
dot stereograms with discontinuities.
The rationale for the use of a variance ratio in equation 1 is to prevent the variances
of d a and db collapsing to zero. Because the local estimates d 1 , ... , d s did not change
as the system learned the expert interpolators, it was possible to use (de - dc ,d 2 in
the objective function without worrying about the possibility that the variance of
de across cases would collapse to zero during the learning . Ideally we would like to
Learning (0 Make Coherent Predictions in Domains with Discontinuities
refine the weights of the local depth estimators to maximize their agreement with
the contextually predicted depths produced by the mixture of expert interpolators.
One way to do this would be to generalize equation 3 to handle a mixture of expert
interpolators:
(5)
Alternatively we could modify equation 4 by normalizing the difference (de - de.i )2
by the actual variance of dc, though this makes the derivatives considerably more
complicated.
5
Discussion
The competing controllers in figure 3 explicitly represent which regularity applies in
a particular region. The outputs of the controllers for nearby regions may themselves
exhibit coherence at a larger spatial scale, so the same learning technique could be
applied recursively. In 2-D images this should allow the continuity of depth edges
to be discovered.
The approach presented here should be applicable to other domains which contain
a mixture of alternative local regularities aCl?OSS space or time. For example, a l?igiJ
shape causes a linear constraint between the locations of its parts in an image, so if
there are many possible shapes, there are many alternative local regularities (Zemel
and Hinton, 1991).
Our learning procedure differs from methods that try to capture as much information as possible about the input (Linsker, 1988; Atick and Redlich, 1990) because
we ignore information in the input that is not coherent across space.
Acknowledgements
This research was funded by grants from NSERC and the Ontario Information Technology Research Centre. Hinton is Noranda fellow of the Canadian Institute for Advanced
Research. Thanks to John Bridle and Steve Nowlan for helpful discussions.
References
Atick, J. J. and Redlich, A. N. (1990). Towards a theory of early visual processing.
Technical Report IASSNS-HEP-90j10, Institute for Advanced Study, Princeton.
Becker, S. and Hinton, G. E. (1992). A self-organizing neural network that discovers
surfaces in random-dot stereograms. January 1992 Nature.
Jacobs, R. A., Jordan, M. 1., Nowlan, S. J., and Hinton, G. E. (1991). Adaptive mixtures
of local experts. Neural Computation, 3(1).
Linsker, R. (1988). Self-organization in a perceptual network. IEEE Computer, March,
21:105-117.
Zemel, R. S. and Hinton, G. E. (1991). Discovering viewpoint-invariant relationships that
characterize objects. In Advances In Neural Information Processing Systems 3, pages
299-305. Morgan Kaufmann Publishers.
379
| 534 |@word version:4 proportion:2 tried:3 jacob:2 pressure:1 recursively:1 initial:1 contains:1 disparity:4 contextual:1 nowlan:3 must:2 slanted:1 john:1 shape:2 treating:1 v:1 discovering:1 plane:1 ith:1 filtered:1 contribute:2 toronto:2 location:13 five:5 fixation:1 fitting:1 introduce:1 expected:2 os:1 themselves:1 multi:1 actual:1 discover:3 underlying:7 what:1 nework:1 remember:1 fellow:1 ti:1 control:1 unit:5 internally:1 grant:1 before:1 positive:1 local:10 treat:1 modify:1 approximately:1 black:1 plus:1 acl:1 oas:2 contextually:4 collapse:1 range:2 averaged:1 thirty:1 differs:1 backpropagation:1 x3:1 procedure:5 reject:1 projection:2 get:2 context:1 optimize:1 center:1 maximizing:2 starting:1 l:1 pure:2 suzanna:1 estimator:1 spanned:1 handle:2 variation:1 transmit:1 neighbouring:2 us:1 agreement:7 located:1 asymmetric:1 bottom:1 module:27 p5:1 capture:2 region:3 stereograms:6 moderately:1 ideally:1 trained:4 interpolators:14 train:1 separated:2 forced:1 zemel:2 outside:1 quite:1 modular:1 larger:2 valued:1 think:1 net:3 propose:1 p4:1 neighboring:2 organizing:1 mixing:1 ontario:2 regularity:3 produce:4 object:2 develop:1 measured:1 nearest:3 received:3 p2:1 throw:1 predicted:7 closely:1 assign:1 adjusted:1 exp:1 j10:1 tor:1 vary:1 consecutive:1 early:1 applicable:1 weighted:1 gaussian:7 aim:1 rather:1 avoid:1 derived:1 improvement:1 detect:5 helpful:1 eliminate:1 selects:1 pixel:9 ill:1 spatial:1 special:1 mutual:1 equal:1 field:5 sampling:1 identical:1 x4:1 look:2 unsupervised:2 yu:1 linsker:2 report:1 spline:3 develops:1 hint:1 randomly:3 simultaneously:1 interpolate:1 replaced:1 consisting:2 attempt:1 organization:1 possibility:1 adjust:1 mixture:11 edge:1 modest:1 rotating:1 desired:2 plotted:1 fronto:1 lwe:1 earlier:1 soft:1 hep:1 cost:2 uniform:2 predictor:1 characterize:1 teacher:2 corrupted:1 considerably:1 thanks:1 density:2 probabilistic:1 invoke:2 continuously:1 squared:2 central:1 choose:1 collapsing:1 external:3 expert:22 derivative:3 til:1 de:38 explicitly:1 depends:1 view:1 try:5 parallel:1 complicated:1 minimize:1 formed:1 variance:8 kaufmann:1 maximized:1 spaced:1 ensemble:1 generalize:1 accurately:2 produced:2 rejecting:2 m5s:1 unaffected:1 detector:3 implausible:1 strip:5 ed:1 against:1 conveys:1 bridle:1 knowledge:1 actually:2 steve:1 specify:1 shrink:1 though:1 just:2 atick:2 receives:2 horizontal:1 touch:1 ei:1 overlapping:1 continuity:10 contain:2 normalized:3 equality:1 spatially:3 white:1 adjacent:7 x5:1 during:2 self:2 backpropagating:1 trying:1 interpreting:1 image:21 discovers:2 common:8 specialized:2 empirically:1 teaching:1 centre:1 dot:5 funded:1 surface:13 something:1 own:1 showed:1 apart:2 transmitted:1 minimum:1 fortunately:1 morgan:1 maximize:7 signal:7 ii:10 multiple:1 technical:1 cross:1 divided:1 prediction:7 basic:2 controller:18 vision:1 iteration:1 represent:1 normalization:1 receive:1 want:1 fine:1 iiii:1 modality:2 publisher:1 probably:1 db:6 jordan:2 extracting:2 presence:4 canadian:1 iii:3 architecture:3 competing:3 opposite:1 whether:1 introd:1 becker:9 stereo:9 cause:6 locally:5 simplest:1 restricts:1 sign:1 per:1 four:2 nevertheless:1 interpolator:9 drawn:1 prevent:1 ce:1 iiiiii:1 worrying:1 sum:2 patch:17 p3:1 coherence:6 decision:2 entirely:1 layer:4 refine:1 constraint:2 throwing:2 constrain:1 striped:1 x2:1 nearby:6 speed:1 department:1 combination:1 march:1 conjugate:1 across:5 smaller:1 making:2 happens:1 outlier:2 gradually:1 invariant:1 equation:4 agree:3 previously:1 remains:1 gaussians:2 v2:1 appropriate:5 alternative:2 responding:1 assumes:1 top:2 include:1 a4:1 somewhere:1 objective:2 receptive:5 exhibit:1 gradient:1 distance:2 separate:1 evenly:1 relationship:1 ratio:1 unfortunately:1 negative:1 rise:1 unknown:1 curved:4 january:1 hinton:13 dc:5 discovered:1 varied:1 canada:1 intensity:1 pair:1 coherent:11 learned:7 discontinuity:37 able:1 beyond:1 usually:1 summarize:1 treated:2 force:1 predicting:1 advanced:2 representing:1 technology:1 created:1 extract:4 prior:1 acknowledgement:1 relative:3 rationale:1 filtering:1 geoffrey:1 principle:1 dd:1 viewpoint:1 pi:4 copy:3 free:1 side:2 allow:3 institute:2 wide:2 taking:2 curve:1 depth:38 dimension:1 world:2 sensory:1 made:2 adaptive:1 far:2 hatched:1 ignore:1 assumed:2 noranda:1 surfa:1 alternatively:1 continuous:1 learn:5 nature:1 interpolating:11 domain:6 did:2 noise:5 convey:3 redlich:2 scattered:1 cubic:3 xl:1 lie:1 perceptual:4 third:1 learns:4 down:1 striving:1 oll:1 normalizing:1 uction:1 gap:1 easier:1 visual:1 nserc:1 scalar:1 applies:1 extracted:3 lear:1 towards:1 absence:1 considerable:1 hard:1 change:2 typical:1 reducing:1 operates:1 averaging:1 rarely:1 select:1 princeton:1 |
4,794 | 5,340 | Design Principles of the Hippocampal Cognitive Map
Kimberly L. Stachenfeld1 , Matthew M. Botvinick1 , and Samuel J. Gershman2
Princeton Neuroscience Institute and Department of Psychology, Princeton University
2
Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology
[email protected], [email protected], [email protected]
1
Abstract
Hippocampal place fields have been shown to reflect behaviorally relevant aspects
of space. For instance, place fields tend to be skewed along commonly traveled
directions, they cluster around rewarded locations, and they are constrained by the
geometric structure of the environment. We hypothesize a set of design principles
for the hippocampal cognitive map that explain how place fields represent space
in a way that facilitates navigation and reinforcement learning. In particular, we
suggest that place fields encode not just information about the current location,
but also predictions about future locations under the current transition distribution. Under this model, a variety of place field phenomena arise naturally from
the structure of rewards, barriers, and directional biases as reflected in the transition policy. Furthermore, we demonstrate that this representation of space can
support efficient reinforcement learning. We also propose that grid cells compute
the eigendecomposition of place fields in part because is useful for segmenting an
enclosure along natural boundaries. When applied recursively, this segmentation
can be used to discover a hierarchical decomposition of space. Thus, grid cells
might be involved in computing subgoals for hierarchical reinforcement learning.
1
Introduction
A cognitive map, as originally conceived by Tolman [46], is a geometric representation of the environment that can support sophisticated navigational behavior. Tolman was led to this hypothesis
by the observation that rats can acquire knowledge about the spatial structure of a maze even in the
absence of direct reinforcement (latent learning; [46]). Subsequent work has sought to formalize the
representational content of the cognitive map [13], the algorithms that operate on it [33, 35], and its
neural implementation [34, 27]. Much of this work was galvanized by the discovery of place cells
in the hippocampus [34], which selectively respond when an animal is in a particular location, thus
supporting the notion that the brain contains an explicit map of space. The later discovery of grid
cells in the entorhinal cortex [16], which respond periodically over the entire environment, indicated
a possible neural substrate for encoding metric information about space.
Metric information is very useful if one considers the problem of spatial navigation to be computing the shortest path from a starting point to a goal. A mechanism that accumulates a record of
displacements can easily compute the shortest path back to the origin, a technique known as path
integration. Considerable empirical evidence supports the idea that animals use this technique for
navigation [13]. Many authors have proposed theories of how grid cells and place cells can be used
to carry out the necessary computations [27].
However, the navigational problems faced by humans and animals are inextricably tied to the more
general problem of reward maximization, which cannot be reduced to the problem of finding the
shortest path between two points. This raises the question: does the brain employ the same machinery for spatial navigation and reinforcement learning (RL)? A number of authors have suggested
how RL mechanisms can support spatial learning, where spatial representations (e.g., place cells or
1
grid cells), serve as the input to the learning system [11, 15]. In contrast to the view that spatial representation is extrinsic to the RL system, we pursue the idea that the brain?s spatial representations
are designed to support RL. In particular, we show how spatial representations resembling place
cells and grid cells emerge as the solution to the problem of optimizing spatial representation in the
service of RL.
We first review the formal definition of the RL problem, along with several algorithmic solutions.
Special attention is paid to the successor representation (SR) [6], which enables a computationally
convenient decomposition of value functions. We then show how the successor representation naturally comes to represent place cells when applied to spatial domains. The eigendecomposition of
the successor representation reveals properties of an environment?s spectral graph structure, which
is particularly useful for discovering hierarchical decompositions of space. We demonstrate that the
eigenvectors resemble grid cells, and suggest that one function of the entorhinal cortex may be to
encode a compressed representation of space that aids hierarchical RL [3].
2
The reinforcement learning problem
Here we consider the problem of RL in a Markov decision process, which consists of the following
elements: a set of states S, a set of actions A, a transition distribution P (s0 |s, a) specifying the
probability of transitioning to state s0 ? S from state s ? S after taking action a ? A, a reward
function R(s) specifying the expected reward in state s, and a discount factor ? ? [0, 1]. An agent
chooses actions according to a policy ?(a|s) and collects rewards as it moves through the state space.
The standard RL problem
P? is to choose a policy that maximizes the value (expected discounted future
return), V (s) = E? [ t=0 ? t R(st ) | s0 = s]. Our focus here is on policy evaluation (computing V ).
In our simulations we feed the agent the optimal policy; in the Supplementary Materials we discuss
algorithms for policy improvement. To simplify notation,
P we assume implicit dependence on ? and
define the state transition matrix T , where T (s, s0 ) = a ?(a|s)P (s0 |s, a).
Most work on RL has focused on two classes of algorithms for policy evaluation: ?model-free?
algorithms that estimate V directly from sample paths, and ?model-based? algorithms that estimate
T and R from sample paths and then compute V by some form of dynamic programming or tree
search [44, 5]. However, there exists a third class that has received less attention. As shown by
Dayan [6], the value function can be decomposed into the inner product of the reward function with
the SR, denoted by M :
P
V (s) = s0 M (s, s0 )R(s0 ),
M = (I ? ?T )?1
(1)
where I denotes the identity matrix. The SR encodes the expected discounted future occupancy of
state s0 along a trajectory initiated in state s:
P?
M (s, s0 ) = E [ t=0 ? t I{st = s0 } | s0 = s] ,
(2)
where I{?} = 1 if its argument is true, and 0 otherwise.
The SR obeys a recursion analogous to the Bellman equation for value functions:
P
M (s, j) = I{s = j} + ? s0 T (s, s0 )M (s0 , j).
(3)
This recursion can be harnessed to derive a temporal difference learning algorithm for incrementally
? of the SR [6, 14]. After observing a transition s ? s0 , the estimate is
updating an estimate M
updated according to:
h
i
? (s, j) ? M
? (s, j) + ? I{s = j} + ? M
? (s0 , j) ? M
? (s, j) ,
M
(4)
where ? is a learning rate (unless specified otherwise, ? = 0.1 in our simulations). The SR combines
some of the advantages of model-free and model-based algorithms: like model-free algorithms,
policy evaluation is computationally efficient, but at the same time the SR provides some of the same
flexibility as model-based algorithms. As we illustrate later, an agent using the SR will be sensitive
to distal changes in reward, whereas a model-free agent will be insensitive to these changes.
3
The successor representation and place cells
In this section, we explore the neural implications of using the SR for policy evaluation: if the brain
encoded the SR, what would the receptive fields of the encoding population look like, and what
2
Empty Room
1.8
1.8
Single Barrier
c
e
1.8
1.8
2.1
1.2
5.6
1.9
1.8
1.8
1.2
1.3
1.2
1.4
1.8
f
1.6
1.3
1.3
Reward (+)
d
5.6
1.2
1.9
1.8
1.8
b
Multiple Rooms
No Reward
a
1.4
Discounted expected visiations (SR)
Figure 1: SR place fields. Top two rows show place fields without reward, bottom two show
retrospective place fields with reward (marked by +). Maximum firing rate (a.u.) indicated for each
plot. (a, b) Empty room. (c, d) Single barrier. (e, f) Multiple rooms.
Direction Selectivity
4
0.2
0.4
0.6
0.8
1
3
Figure 2: Direction selectivity along a
track. Direction selectivity arises in SR
place fields when the probability p?
of transitioning in the preferred left-toright direction along a linear track is
greater than the probability p? of transitioning in the non-preferred direction.
The legend shows the ratio of p? to p?
for each simulation.
2
1
0
100
150
200
250
300
Distance along Track
350
400
would the population look like at any point in time? This question is most easily addressed in spatial
domains, where states index spatial locations (see Supplementary Materials for simulation details).
For an open field with uniformly distributed rewards we assume a random walk policy, and the
resulting SR for a particular location is an approximately symmetric, gradually decaying halo around
that location (Fig. 1a)?the canonical description of a hippocampal place cell. In order for the
population to encode the expected visitations to each state in the domain from the current starting
state (i.e. a row of M ), each receptive field corresponds to a column of the SR matrix. This allows
the current state?s value to be computed by taking the dot product of its population vector with the
reward vector. The receptive field (i.e. column of M ) will encode the discounted expected number
of times that state was visited for each starting state, and will therefore skew in the direction of the
states that likely preceded the current state.
More interesting predictions can be made when we examine the effects of obstacles and direction
preference that shape the transition structure. For instance, when barriers are inserted into the environment, the probability of transitioning across these obstacles will go to zero. SR place fields
are therefore constrained by environmental geometry, and the receptive field will be discontinuous
across barriers (Fig. 1c,e). Consistent with this idea, experiments have shown that place fields become distorted around barriers [32, 40]. When an animal has been trained to travel in a preferred
direction along a linear track, we expect the response of place fields to become skewed opposite the
direction of travel (Fig. 2), a result that has been observed experimentally [28, 29].
Another way to alter the transition policy is by introducing a goal, which induces a tendency to move
in the direction that maximizes reward. Under these conditions, we expect firing fields centered near
rewarded locations to expand to include the surrounding locations and to increase their firing rate,
as has been observed experimentally [10, 21]. Meanwhile, we expect the majority of place fields
3
Percentage of Neurons Firing
f
a4
2
0
0.4
Depth
b
0.2
0
Distance around annular track
Figure 3: Reward clustering in annular maze. (a) Histogram of number of cells firing above
baseline at each displacement around an annular track. (b) Heat map of number of firing cells at
each location on unwrapped annular maze. Reward is centered on track. Baseline firing rate set to
10% maximum.
late detour
early detour
no detour
a
b
c
d
1.25
1.15
1.49
1.60
1.15
1.49
2.36
1.08
1.49
Firing Fields
Value
Figure 4: Tolman detour task. The starting location is at the bottom of the maze where the
three paths meet, and the reward is at the top. Barriers are shown as black horizontal lines. Three
conditions are shown: No detour, early detour, and late detour. (a, b, c) SR place fields centered near
and far from detours. Maximum firing rate (a.u.) indicated by each plot. (d) Value function.
that encode non-rewarded states to skew slightly away from the reward. Under certain settings
for what firing rate constitutes baseline (see Supplementary Materials), the spread of the rewarded
locations? fields compensates for the skew of surrounding fields away from the reward, and we
observe ?clustering? around rewarded locations (Fig. 3), as has been observed experimentally in the
annular water maze task [18]. This parameterization sensitivity may explain why goal-related firing
is not observed in all tasks [25, 24, 41].
As another illustration of the model?s response to barriers, we simulated place fields in a version
of the Tolman detour task [46], as described in [1]. Rats are trained to move from the start to the
rewarded location. At some point, an ?early? or a ?late? transparent barrier is placed in the maze
so that the rat must take a detour. For the early barrier, a short detour is available, and for the later
barrier, the only detour is a longer one. Place fields near the detour are more strongly affected than
places far away from the detour (Fig. 4a,b,c), consistent with experimental findings [1]. Fig. 4d
shows the value function in each of these detour conditions.
4
Behavioral predictions: distance estimation and latent learning
In this section, we examine some of the behavioral consequences of using the SR for RL. We first
show that the SR anticipates biases in distance estimation induced by semi-permeable boundaries.
We then explore the ability of the SR to support latent learning in contextual fear conditioning.
4
b
75
4
50
3
2
25
1
0
0
a 18
16
14
0.5
0
1
Permeability
SR Distance
b
Lesion
Control
12
10
6
4
2
0
1
2
Preexposure Duration (steps)
3
x 10
5
Figure 6: Context preexposure facilitation
effect. (a) Simulated conditioned response
(CR) to the context following one-trial contextual fear conditioning, shown as a function of
preexposure duration. The CR was approximated as the negative value summed over the
environment. The ?Lesion? corresponds to
agents with hippocampal damage, simulated by
setting the SR learning rate to 0.01. The ?Control? group has a learning rate of 0.1. (b) value
for a single location after preexposure in a control agent. (c) same as (b) in a lesioned agent.
0
?0.2
?0.4
?0.6
?0.8
c
8
0
Control
0
?0.1
Value
Conditioned Response
Figure 5: Distance estimates. (a) Increase in
the perceived distance between two points on
opposite sides of a semipermeable boundary
(marked with + and ? in 5b) as a function of
barrier permeability. (b) Perceived distance between destination (market with +) and all other
locations in the space (barrier permeability =
0.05).
5
Value
Distance (% Increase)
a
?0.2
Lesion
?0.3
Stevens and Coupe [43] reported that people overestimated the distance between two locations when
they were separated by a boundary (e.g., a state or country line). This bias was hypothesized to arise
from a hierarchical organization of space (see also [17]). We showp(Fig. 5) how distance estimates
(using the Euclidean distance between SR state representations, (M (s0 ) ? M (s))2 , as a proxy
for the perceived distance between s and s0 ) between points in different regions of the environment
are altered when an enclosure is divided by a soft (semi-permeable) boundary. We see that as the
permeability of the barrier decreases (making the boundary harder), the percent increase in perceived
distance between locations increases without bound. This gives rise to a discontinuity in perceived
travel time at the soft boundary. Interestingly, the hippocampus is directly involved in distance
estimation [31], suggesting the hippocampal cognitive map as a neural substrate for distance biases
(although a direct link has yet to be established).
The context preexposure facilitation effect refers to the finding that placing an animal inside a conditioning chamber prior to shocking it facilitates the acquisition of contextual fear [9]. In essence, this
is a form of latent learning [46]. The facilitation effect is thought to arise from the development of a
conjunctive representation of the context in the hippocampus, though areas outside the hippocampus
may also develop a conjunctive representation in the absence of the hippocampus, albeit less efficiently [48]. The SR provides a somewhat different interpretation: over the course of preexposure,
the hippocampus develops a predictive representation of the context, such that subsequent learning
is rapidly propagated across space. Fig. 6 shows a simulation of this process and how it accounts
for the facilitation effect. We simulated hippocampal lesions by reducing the SR learning rate from
0.1 to 0.01, resulting in a more punctate SR following preexposure and a reduced facilitation effect.
5
Eigendecomposition of the successor representation: hierarchical
decomposition and grid cells
Reinforcement learning and navigation can often be made more efficient by decomposing the environment hierarchically. For example, the options framework [45] utilizes a set of subgoals to divide
and conquer a complex learning environment. Recent experimental work suggests that the brain may
exploit a similar strategy [3, 36, 8]. A key problem, however, is discovering useful subgoals; while
progress has been made on this problem in machine learning, we still know very little about how the
brain solves it (but see [37]). In this section, we show how the eigendecomposition of the SR can
be used to discover subgoals. The resulting eigenvectors strikingly resemble grid cells observed in
entorhinal cortex.
5
a
Open Room
b
Single Barrier
c
Multiple Rooms
Figure 7: Eigendecomposition of the SR. Each panel shows the same 20 eigenvectors randomly
sampled from the top 100 (excluding the constant first eigenvector) for the environmental geometries
shown in Fig. 1 (no reward). (a) Empty room. (b) Single barrier. (c) Multiple rooms.
Eigendecomposition
Figure 8: Eigendecomposition of the SR in a
hairpin maze. Since the walls of the maze effectively elongate a dimension of travel (the track
of the maze), the low frequency eigenvectors resemble one-dimensional sinusoids that have been
folded to match the space. Meanwhile, the low
frequency eigenvectors exhibit the compartmentalization shown by [7].
A number of authors have used graph partitioning techniques to discover subgoals [30, 39]. These
approaches cluster states according to their community membership (a community is defined as a
highly interconnected set of nodes with relatively few outgoing edges). Transition points between
communities (bottleneck states) are then used as subgoals. One important graph partitioning technique, used by [39] to find subgoals, is the normalized cuts algorithm [38], which recursively thresholds the second smallest eigenvector (the Fiedler vector) of the normalized graph Laplacian to obtain
a graph partition. Given an undirected graph with symmetric weight matrix W , the graph Laplacian
is given by L = D ? W . The normalized graph Laplacian
is given by L = I ? D?1/2 W D?1/2 ,
P
where D is a diagonal degree matrix with D(s, s) = s0 W (s, s0 ). When states are projected onto
the second eigenvector, they are pulled along orthogonal dimensions according to their community
membership. Locations in distinct regions but close in Euclidean distance ? for instance, nearby
points on opposite sides of a boundary ? will be represented as distant in the eigenspace.
The normalized graph Laplacian is closely related to the SR [26]. Under a random walk policy,
the transition matrix is given by T = D?1 W . If ? is an eigenvector of the random walk?s graph
Laplacian I?T , then D1/2 ? is an eigenvector of the normalized graph Laplacian. The corresponding
eigenvector for the discounted Laplacian, I ? ?T , is ??. Since the matrix inverse preserves the
eigenvectors, the normalized graph Laplacian has the same eigenvectors as the SR, M = (I??T )?1 ,
scaled by ?D?1/2 . These spectral eigenvectors can be approximated by slow feature analysis [42].
Applying hierarchical slow feature analysis to streams of simulated visual inputs produces feature
representations that resemble hippocampal receptive fields [12].
A number of representative SR eigenvectors are shown in Fig. 7, for three different room topologies.
The higher frequency eigenvectors display the latticing characteristic of grid cells [16]. The eigendecomposition is often discontinuous at barriers, and in many cases different rooms are represented
by independent sinusoids. Fig. 8 shows the eigendecomposition for a hairpin maze. The eigenvectors resemble folded up one-dimensional sinusoids, and high frequency eigenvectors appear as
repeating phase-locked ?submaps? with firing selective to a subset of hallways, much like the grid
cells observed by Derdikman and Moser [7].
In the multiple rooms environment, visual inspection reveals that the SR eigenvector with the second
smallest eigenvalue (the Fiedler vector) divides the enclosure along the vertical barrier: the left half
is almost entirely blue and the right half almost entirely red, with a smooth but steep transition
at the doorway (Fig. 9a). As discussed above, this second eigenvector can therefore be used to
segment the enclosure along the vertical boundary. Applying this segmentation recursively, as in
the normalized cuts algorithm, produces a hierarchical decomposition of the environment (Figure
6
Segmentation
b
Figure 9: Segmentation using normalized cuts.
(a) The results of segmentation by thresholding
the second eigenvector of the multiple rooms environment in Fig. 1. Dotted lines indicate the
segment boundaries. (b, c) Eigenvector segmentation applied recursively to fully parse the enclosure into the four rooms.
a
c
First Level
Second Level
9b,c). By identifying useful subgoals from the environmental topology, this decomposition can be
exploited by hierarchical learning algorithms [3, 37].
One might reasonably question why the brain should represent high frequency eigenvectors (like
grid cells) if only the low frequency eigenvectors are useful for hierarchical decomposition. Spectral
features also serve as generally useful representations [26, 22], and high frequency components are
important for representing detail in the value function. The progressive increase in grid cell spacing
along the dorsal-ventral axis of the entorhinal cortex may function as a multi-scale representation
that supports both fine and coarse detail [2].
6
Discussion
We have shown how many empirically observed properties of spatial representation in the brain,
such as changes in place fields induced by manipulations of environmental geometry and reward,
can be explained by a predictive representation of the environment. This predictive representation
is intimately tied to the problem of RL: in a certain sense, it is the optimal representation of space
for the purpose of computing value functions, since it reduces value computation to a simple matrix
multiplication [6]. Moreover, this optimality principle is closely connected to ideas from manifold
learning and spectral graph theory [26]. Our work thus sheds new computational light on Tolman?s
cognitive map [46].
Our work is connected to several lines of previous work. Most relevant is Gustafson and Daw
[15], who showed how topologically-sensitive spatial representations recapitulate many aspects of
place cells and grid cells that are difficult to reconcile with a purely Euclidean representation of
space. They also showed how encoding topological structure greatly aids reinforcement learning in
complex spatial environments. Earlier work by Foster and colleagues [11] also used place cells as
features for RL, although the spatial representation did not explicitly encode topological structure.
While these theoretical precedents highlight the importance of spatial representation, they leave
open the deeper question of why particular representations are better than others. We showed that
the SR naturally encodes topological structure in a format that enables efficient RL.
Spectral graph theory provides insight into the topological structure encoded by the SR. In particular,
we showed that eigenvectors of the SR can be used to discover a hierarchical decomposition of the
environment for use in hierarchical RL. These eigenvectors may also be useful as a representational
basis for RL, encoding multi-scale spatial structure in the value function. Spectral analysis has
frequently been invoked as a computational motivation for entorhinal grid cells (e.g., [23]). The
fact that any function can be reconstructed by sums of sinusoids suggested that the entorhinal cortex
implements a kind of Fourier transform of space, and that place cells are the result of reconstructing
spatial signals from their spectral decomposition. Two problems face this interpretation. Fist, recent
evidence suggests that the emergence of place cells does not depend on grid cell input [4, 47].
Second, and more importantly for our purposes, Fourier analysis is not the right mathematical tool
when dealing with spatial representation in a topologically structured environment, since we do not
expect functions to be smooth over boundaries in the environment. This is precisely the purpose of
spectral graph theory: the eigenvectors of the graph Laplacian encode the smoothest approximation
of a function that respects the graph topology [26].
Recent work has elucidated connections between models of episodic memory and the SR. Specifically, in [14] it was shown that the SR is closely related to the Temporal Context Model (TCM)
of episodic memory [20]. The core idea of TCM is that items are bound to their temporal context
(a running average of recently experienced items), and the currently active temporal context is used
7
to cue retrieval of other items, which in turn cause their temporal context to be retrieved. The SR
can be seen as encoding a set of item-context associations. The connection to episodic memory is
especially interesting given the crucial mnemonic role played by the hippocampus and entorhinal
cortex in episodic memory. Howard and colleagues [19] have laid out a detailed mapping between
TCM and the medial temporal lobe (including entorhinal and hippocampal regions).
An important question for future work concerns how biologically plausible mechanisms can implement the computations posited in our paper. We described a simple error-driven updating rule for
learning the SR, and in the Supplementary Materials we derive a stochastic gradient learning rule
that also uses a simple error-driven update. Considerable attention has been devoted to the implementation of error-driven learning rules in the brain, so we expect that these learning rules can be
implemented in a biologically plausible manner.
References
[1] A. Alvernhe, E. Save, and B. Poucet. Local remapping of place cell firing in the tolman detour task.
European Journal of Neuroscience, 33:1696?1705, 2011.
[2] H. T. Blair, A. C. Welday, and K. Zhang. Scale-invariant memory representations emerge from moire
interference between grid fields that produce theta oscillations: a computational model. The Journal of
Neuroscience, 27:3211?3229, 2007.
[3] M. M. Botvinick, Y. Niv, and A. C. Barto. Hierarchically organized behavior and its neural foundations:
A reinforcement learning perspective. Cognition, 113:262?280, 2009.
[4] M. P. Brandon, J. Koenig, J. K. Leutgeb, and S. Leutgeb. New and distinct hippocampal place codes are
generated in a new environment during septal inactivation. Neuron, 82:789?796, 2014.
[5] N. D. Daw, Y. Niv, and P. Dayan. Uncertainty-based competition between prefrontal and dorsolateral
striatal systems for behavioral control. Nature Neuroscience, 8:1704?1711, 2005.
[6] P. Dayan. Improving generalization for temporal difference learning: The successor representation. Neural Computation, 5:613?624, 1993.
[7] D. Derdikman, J. R. Whitlock, A. Tsao, M. Fyhn, T. Hafting, M.-B. Moser, and E. I. Moser. Fragmentation
of grid cell maps in a multicompartment environment. Nature Neuroscience, 12:1325?1332, 2009.
[8] C. Diuk, K. Tsai, J. Wallis, M. Botvinick, and Y. Niv. Hierarchical learning induces two simultaneous, but
separable, prediction errors in human basal ganglia. The Journal of Neuroscience, 33:5797?5805, 2013.
[9] M. S. Fanselow. From contextual fear to a dynamic view of memory systems. Trends in Cognitive
Sciences, 14:7?15, 2010.
[10] A. Fenton, L. Zinyuk, and J. Bures. Place cell discharge along search and goal-directed trajectories.
European Journal of Neuroscience, 12:3450, 2001.
[11] D. Foster, R. Morris, and P. Dayan. A model of hippocampally dependent navigation, using the temporal
difference learning rule. Hippocampus, 10:1?16, 2000.
[12] M. Franzius, H. Sprekeler, and L. Wiskott. Slowness and sparseness lead to place, head-direction, and
spatial-view cells. PLoS Computational Biology, 3:3287?3302, 2007.
[13] C. R. Gallistel. The Organization of Learning. The MIT Press, 1990.
[14] S. J. Gershman, C. D. Moore, M. T. Todd, K. A. Norman, and P. B. Sederberg. The successor representation and temporal context. Neural Computation, 24:1553?1568, 2012.
[15] N. J. Gustafson and N. D. Daw. Grid cells, place cells, and geodesic generalization for spatial reinforcement learning. PLoS Computational Biology, 7:e1002235, 2011.
[16] T. Hafting, M. Fyhn, S. Molden, M.-B. Moser, and E. I. Moser. Microstructure of a spatial map in the
entorhinal cortex. Nature, 436:801?806, 2005.
[17] S. C. Hirtle and J. Jonides. Evidence of hierarchies in cognitive maps. Memory & Cognition, 13:208?217,
1985.
[18] S. A. Hollup, S. Molden, J. G. Donnett, M. B. Moser, and E. I. Moser. Accumulation of hippocampal
place fields at the goal location in an annular watermaze task. Journal of Neuroscience, 21:1635?1644,
2001.
[19] M. W. Howard, M. S. Fotedar, A. V. Datey, and M. E. Hasselmo. The temporal context model in spatial
navigation and relational learning: toward a common explanation of medial temporal lobe function across
domains. Psychological Review, 112:75?116, 2005.
[20] M. W. Howard and M. J. Kahana. A distributed representation of temporal context. Journal of Mathematical Psychology, 46:269?299, 2002.
8
[21] T. Kobayashi, A. Tran, H. Nishijo, T. Ono, and G. Matsumoto. Contribution of hippocampal place cell
activity to learning and formation of goal-directed navigation in rats. Neuroscience, 117:1025?35, 2003.
[22] G. Konidaris, S. Osentoski, and P. S. Thomas. Value function approximation in reinforcement learning
using the Fourier basis. In AAAI, 2011.
[23] J. Krupic, N. Burgess, and J. O?oeefe. Neural representations of location composed of spatially periodic
bands. Science, 337:853?857, 2012.
[24] P. Lenck-Santini, R. Muller, E. Save, and B. Poucet. Relationships between place cell firing fields and
navigational decisions by rats. The Journal of Neuroscience, 22:9035?47, 2002.
[25] P. Lenck-Santini, E. Save, and B. Poucet. Place-cell firing does not depend on the direction of turn in a
y-maze alternation task. European Journal of Neuroscience, 13(5):1055?8, 2001.
[26] S. Mahadevan. Learning representation and control in markov decision processes: New frontiers. Foundations and Trends in Machine Learning, 1:403?565, 2009.
[27] B. L. McNaughton, F. P. Battaglia, O. Jensen, E. I. Moser, and M.-B. Moser. Path integration and the
neural basis of the ?cognitive map?. Nature Reviews Neuroscience, 7:663?678, 2006.
[28] M. R. Mehta, C. A. Barnes, and B. L. McNaughton. Experience-dependent, asymmetric expansion of
hippocampal place fields. Proceedings of the National Academy of Sciences, 94:8918?8921, 1997.
[29] M. R. Mehta, M. C. Quirk, and M. A. Wilson. Experience-dependent asymmetric shape of hippocampal
receptive fields. Neuron, 25:707?715, 2000.
[30] I. Menache, S. Mannor, and N. Shimkin. Q-cut?dynamic discovery of sub-goals in reinforcement learning. In European Conference on Machine Learning, pages 295?306. Springer, 2002.
[31] L. K. Morgan, S. P. MacEvoy, G. K. Aguirre, and R. A. Epstein. Distances between real-world locations
are represented in the human hippocampus. The Journal of Neuroscience, 31:1238?1245, 2011.
[32] R. U. Muller and J. L. Kubie. The effects of changes in the environment on the spatial firing of hippocampal complex-spike cells. The Journal of Neuroscience, 7:1951?1968, 1987.
[33] R. U. Muller, M. Stead, and J. Pach. The hippocampus as a cognitive graph. The Journal of General
Physiology, 107:663?694, 1996.
[34] J. O?Keefe and L. Nadel. The Hippocampus as a Cognitive Map. Clarendon Press Oxford, 1978.
[35] A. K. Reid and J. R. Staddon. A dynamic route finder for the cognitive map. Psychological Review,
105:585?601, 1998.
[36] J. J. Ribas-Fernandes, A. Solway, C. Diuk, J. T. McGuire, A. G. Barto, Y. Niv, and M. M. Botvinick. A
neural signature of hierarchical reinforcement learning. Neuron, 71:370?379, 2011.
[37] A. C. Schapiro, T. T. Rogers, N. I. Cordova, N. B. Turk-Browne, and M. M. Botvinick. Neural representations of events arise from temporal community structure. Nature Neuroscience, 16:486492, 2013.
[38] J. Shi and J. Malik. Normalized cuts and image segmentation. Pattern Analysis and Machine Intelligence,
IEEE Transactions on, 22:888?905, 2000.
? S?ims?ek, A. P. Wolfe, and A. G. Barto. Identifying useful subgoals in reinforcement learning by local
[39] O.
graph partitioning. In Proceedings of the 22nd International Conference on Machine Learning, pages
816?823. ACM, 2005.
[40] W. E. Skaggs and B. L. McNaughton. Spatial firing properties of hippocampal ca1 populations in an
environment containing two visually identical regions. The Journal of Neuroscience, 18:8455?8466,
1998.
[41] A. Speakman and J. O?Keefe. Hippocampal complex spike cells do not change their place fields if the
goal is moved within a cue controlled environment. European Journal of Neuroscience, 2:544?5, 1990.
[42] H. Sprekeler. On the relation of slow feature analysis and laplacian eigenmaps. Neural computation,
23:3287?3302, 2011.
[43] A. Stevens and P. Coupe. Distortions in judged spatial relations. Cognitive Psychology, 10:422 ? 437,
1978.
[44] R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction. MIT press, 1998.
[45] R. S. Sutton, D. Precup, and S. Singh. Between MDPs and semi-MDPs: A framework for temporal
abstraction in reinforcement learning. Artificial Intelligence, 112:181?211, 1999.
[46] E. C. Tolman. Cognitive maps in rats and men. Psychological Review, 55:189?208, 1948.
[47] T. J. Wills, F. Cacucci, N. Burgess, and J. O?Keefe. Development of the hippocampal cognitive map in
preweanling rats. Science, 328:1573?1576, 2010.
[48] B. J. Wiltgen, M. J. Sanders, S. G. Anagnostaras, J. R. Sage, and M. S. Fanselow. Context fear learning
in the absence of the hippocampus. The Journal of Neuroscience, 26:5484?5491, 2006.
9
| 5340 |@word trial:1 version:1 hippocampus:12 nd:1 open:3 mehta:2 simulation:5 lobe:2 decomposition:9 recapitulate:1 diuk:2 paid:1 harder:1 recursively:4 carry:1 contains:1 interestingly:1 current:5 contextual:4 yet:1 conjunctive:2 must:1 subsequent:2 periodically:1 partition:1 distant:1 shape:2 enables:2 fyhn:2 hypothesize:1 designed:1 plot:2 medial:2 update:1 half:2 discovering:2 cue:2 item:4 parameterization:1 intelligence:2 inspection:1 hallway:1 schapiro:1 sederberg:1 short:1 core:1 record:1 cacucci:1 provides:3 coarse:1 node:1 location:22 preference:1 mannor:1 zhang:1 mathematical:2 along:13 direct:2 become:2 gallistel:1 consists:1 combine:1 behavioral:3 inside:1 manner:1 market:1 expected:6 behavior:2 examine:2 frequently:1 multi:2 brain:10 bellman:1 discounted:5 decomposed:1 little:1 unwrapped:1 discover:4 notation:1 moreover:1 maximizes:2 panel:1 eigenspace:1 remapping:1 what:3 derdikman:2 kind:1 pursue:1 eigenvector:10 ca1:1 finding:3 temporal:14 shed:1 botvinick:4 scaled:1 control:6 partitioning:3 compartmentalization:1 appear:1 reid:1 segmenting:1 service:1 kobayashi:1 local:2 todd:1 consequence:1 sutton:2 encoding:5 accumulates:1 punctate:1 initiated:1 permeable:2 meet:1 path:8 firing:17 approximately:1 oxford:1 might:2 black:1 specifying:2 collect:1 suggests:2 locked:1 speakman:1 obeys:1 directed:2 implement:2 kubie:1 displacement:2 episodic:4 area:1 empirical:1 thought:1 physiology:1 convenient:1 enclosure:5 refers:1 suggest:2 cannot:1 onto:1 close:1 judged:1 context:14 applying:2 accumulation:1 map:16 shi:1 resembling:1 go:1 attention:3 starting:4 duration:2 focused:1 identifying:2 insight:1 rule:5 hafting:2 importantly:1 facilitation:5 population:5 notion:1 mcnaughton:3 analogous:1 updated:1 discharge:1 hierarchy:1 substrate:2 programming:1 us:1 hypothesis:1 origin:1 element:1 trend:2 approximated:2 particularly:1 updating:2 osentoski:1 wolfe:1 asymmetric:2 cut:5 bottom:2 inserted:1 observed:7 role:1 region:4 connected:2 plo:2 decrease:1 environment:22 reward:21 lesioned:1 dynamic:4 geodesic:1 signature:1 trained:2 raise:1 depend:2 segment:2 singh:1 predictive:3 serve:2 purely:1 basis:3 strikingly:1 easily:2 represented:3 surrounding:2 separated:1 heat:1 fiedler:2 distinct:2 artificial:1 formation:1 outside:1 encoded:2 supplementary:4 plausible:2 distortion:1 otherwise:2 compressed:1 compensates:1 ability:1 stead:1 transform:1 emergence:1 advantage:1 eigenvalue:1 propose:1 tran:1 interconnected:1 product:2 relevant:2 rapidly:1 flexibility:1 representational:2 academy:1 description:1 moved:1 competition:1 cluster:2 empty:3 produce:3 leave:1 derive:2 illustrate:1 develop:1 quirk:1 received:1 progress:1 solves:1 implemented:1 sprekeler:2 resemble:5 come:1 indicate:1 blair:1 direction:13 stevens:2 discontinuous:2 closely:3 stochastic:1 centered:3 human:3 successor:7 material:4 rogers:1 transparent:1 generalization:2 wall:1 niv:4 microstructure:1 frontier:1 brandon:1 around:6 visually:1 algorithmic:1 mapping:1 cognition:2 matthew:1 ventral:1 sought:1 early:4 smallest:2 purpose:3 perceived:5 whitlock:1 battaglia:1 travel:4 estimation:3 currently:1 visited:1 sensitive:2 hasselmo:1 tool:1 mit:3 behaviorally:1 inactivation:1 tolman:7 cr:2 barto:4 wilson:1 encode:7 focus:1 improvement:1 greatly:1 contrast:1 baseline:3 sense:1 dayan:4 dependent:3 membership:2 abstraction:1 entire:1 relation:2 expand:1 selective:1 denoted:1 development:2 animal:5 matthewb:1 constrained:2 spatial:27 integration:2 special:1 field:34 summed:1 biology:2 placing:1 progressive:1 look:2 identical:1 constitutes:1 alter:1 future:4 hippocampally:1 others:1 simplify:1 develops:1 employ:1 few:1 randomly:1 composed:1 preserve:1 national:1 geometry:3 phase:1 organization:2 moire:1 highly:1 evaluation:4 navigation:8 light:1 devoted:1 implication:1 edge:1 necessary:1 experience:2 machinery:1 unless:1 tree:1 orthogonal:1 detour:16 permeability:4 walk:3 euclidean:3 divide:2 theoretical:1 psychological:3 instance:3 column:2 soft:2 obstacle:2 earlier:1 maximization:1 introducing:1 subset:1 eigenmaps:1 reported:1 periodic:1 anticipates:1 chooses:1 st:2 international:1 sensitivity:1 moser:9 overestimated:1 destination:1 precup:1 reflect:1 aaai:1 containing:1 choose:1 prefrontal:1 cognitive:16 ek:1 return:1 fenton:1 suggesting:1 account:1 explicitly:1 stream:1 later:3 view:3 observing:1 red:1 start:1 decaying:1 option:1 poucet:3 contribution:1 characteristic:1 efficiently:1 who:1 directional:1 trajectory:2 explain:2 simultaneous:1 donnett:1 definition:1 konidaris:1 acquisition:1 frequency:7 involved:2 colleague:2 shimkin:1 turk:1 naturally:3 propagated:1 sampled:1 massachusetts:1 knowledge:1 segmentation:7 formalize:1 organized:1 sophisticated:1 back:1 feed:1 clarendon:1 originally:1 higher:1 reflected:1 response:4 though:1 strongly:1 furthermore:1 just:1 implicit:1 koenig:1 horizontal:1 parse:1 incrementally:1 epstein:1 indicated:3 effect:7 hypothesized:1 normalized:9 true:1 bures:1 norman:1 molden:2 sinusoid:4 spatially:1 symmetric:2 moore:1 solway:1 distal:1 skewed:2 during:1 essence:1 samuel:1 rat:7 hippocampal:18 demonstrate:2 percent:1 image:1 invoked:1 recently:1 common:1 preceded:1 rl:16 empirically:1 harnessed:1 insensitive:1 conditioning:3 subgoals:9 discussed:1 interpretation:2 association:1 ims:1 grid:19 dot:1 cortex:7 longer:1 recent:3 showed:4 retrieved:1 optimizing:1 perspective:1 driven:3 rewarded:6 manipulation:1 selectivity:3 certain:2 slowness:1 route:1 alternation:1 santini:2 exploited:1 muller:3 seen:1 morgan:1 greater:1 somewhat:1 shortest:3 signal:1 semi:3 fist:1 multiple:6 reduces:1 smooth:2 annular:6 match:1 mnemonic:1 retrieval:1 posited:1 divided:1 finder:1 laplacian:10 controlled:1 prediction:4 nadel:1 metric:2 histogram:1 represent:3 cell:40 whereas:1 fine:1 spacing:1 addressed:1 country:1 crucial:1 operate:1 sr:40 induced:2 tend:1 facilitates:2 undirected:1 legend:1 pach:1 near:3 gustafson:2 mahadevan:1 sander:1 variety:1 browne:1 psychology:3 burgess:2 skaggs:1 topology:3 opposite:3 inner:1 idea:5 leutgeb:2 bottleneck:1 tcm:3 retrospective:1 cause:1 action:3 useful:9 generally:1 detailed:1 eigenvectors:17 staddon:1 repeating:1 discount:1 band:1 morris:1 induces:2 reduced:2 percentage:1 canonical:1 dotted:1 neuroscience:18 conceived:1 extrinsic:1 track:8 blue:1 affected:1 visitation:1 group:1 key:1 four:1 threshold:1 basal:1 hairpin:2 graph:19 sum:1 inverse:1 uncertainty:1 respond:2 distorted:1 topologically:2 place:41 almost:2 laid:1 utilizes:1 oscillation:1 decision:3 toright:1 dorsolateral:1 entirely:2 bound:2 played:1 display:1 topological:4 barnes:1 elucidated:1 activity:1 precisely:1 encodes:2 nearby:1 aspect:2 fourier:3 argument:1 optimality:1 separable:1 relatively:1 format:1 department:2 structured:1 according:4 kahana:1 across:4 slightly:1 reconstructing:1 intimately:1 preexposure:7 making:1 biologically:2 explained:1 gradually:1 invariant:1 interference:1 computationally:2 equation:1 discus:1 skew:3 mechanism:3 turn:2 know:1 franzius:1 available:1 decomposing:1 elongate:1 observe:1 hierarchical:14 away:3 spectral:8 fernandes:1 chamber:1 save:3 thomas:1 denotes:1 top:3 include:1 clustering:2 running:1 a4:1 datey:1 exploit:1 especially:1 conquer:1 multicompartment:1 mcguire:1 move:3 malik:1 question:5 spike:2 receptive:6 damage:1 dependence:1 strategy:1 diagonal:1 exhibit:1 gradient:1 distance:18 link:1 simulated:5 majority:1 manifold:1 considers:1 water:1 toward:1 code:1 index:1 relationship:1 illustration:1 ratio:1 acquire:1 difficult:1 steep:1 striatal:1 hirtle:1 menache:1 negative:1 rise:1 sage:1 design:2 implementation:2 policy:12 vertical:2 observation:1 neuron:4 markov:2 matsumoto:1 howard:3 supporting:1 relational:1 halo:1 excluding:1 head:1 watermaze:1 community:5 specified:1 connection:2 established:1 daw:3 discontinuity:1 suggested:2 pattern:1 navigational:3 including:1 memory:7 explanation:1 event:1 natural:1 recursion:2 representing:1 occupancy:1 altered:1 technology:1 theta:1 mdps:2 axis:1 traveled:1 faced:1 geometric:2 discovery:3 review:5 prior:1 multiplication:1 precedent:1 fully:1 expect:5 highlight:1 interesting:2 men:1 gershman:1 eigendecomposition:9 foundation:2 agent:7 degree:1 consistent:2 s0:21 proxy:1 principle:3 thresholding:1 foster:2 wiskott:1 row:2 course:1 placed:1 free:4 sjgershm:1 formal:1 bias:4 side:2 pulled:1 institute:2 deeper:1 taking:2 barrier:18 emerge:2 face:1 distributed:2 boundary:11 depth:1 dimension:2 transition:10 world:1 maze:11 author:3 commonly:1 reinforcement:16 made:3 projected:1 far:2 transaction:1 reconstructed:1 preferred:3 dealing:1 active:1 reveals:2 doorway:1 search:2 latent:4 why:3 nature:5 reasonably:1 improving:1 expansion:1 complex:4 meanwhile:2 european:5 domain:4 did:1 spread:1 hierarchically:2 motivation:1 reconcile:1 arise:4 lesion:4 fig:13 representative:1 slow:3 aid:2 experienced:1 sub:1 explicit:1 smoothest:1 tied:2 third:1 late:3 transitioning:4 ono:1 jensen:1 evidence:3 concern:1 exists:1 aguirre:1 albeit:1 fanselow:2 effectively:1 importance:1 fragmentation:1 keefe:3 entorhinal:9 conditioned:2 sparseness:1 led:1 explore:2 likely:1 ganglion:1 visual:2 fear:5 springer:1 corresponds:2 environmental:4 acm:1 goal:8 identity:1 marked:2 kimberly:1 tsao:1 room:13 absence:3 content:1 considerable:2 change:5 experimentally:3 folded:2 specifically:1 uniformly:1 reducing:1 wallis:1 tendency:1 experimental:2 selectively:1 support:7 people:1 arises:1 dorsal:1 tsai:1 outgoing:1 princeton:4 d1:1 phenomenon:1 |
4,795 | 5,341 | Scalable Inference for Neuronal Connectivity from
Calcium Imaging
Alyson K. Fletcher
Sundeep Rangan
Abstract
Fluorescent calcium imaging provides a potentially powerful tool for inferring
connectivity in neural circuits with up to thousands of neurons. However, a key
challenge in using calcium imaging for connectivity detection is that current systems often have a temporal response and frame rate that can be orders of magnitude slower than the underlying neural spiking process. Bayesian inference methods based on expectation-maximization (EM) have been proposed to overcome
these limitations, but are often computationally demanding since the E-step in the
EM procedure typically involves state estimation for a high-dimensional nonlinear dynamical system. In this work, we propose a computationally fast method
for the state estimation based on a hybrid of loopy belief propagation and approximate message passing (AMP). The key insight is that a neural system as viewed
through calcium imaging can be factorized into simple scalar dynamical systems
for each neuron with linear interconnections between the neurons. Using the structure, the updates in the proposed hybrid AMP methodology can be computed by a
set of one-dimensional state estimation procedures and linear transforms with the
connectivity matrix. This yields a computationally scalable method for inferring
connectivity of large neural circuits. Simulations of the method on realistic neural
networks demonstrate good accuracy with computation times that are potentially
significantly faster than current approaches based on Markov Chain Monte Carlo
methods.
1
Introduction
Determining connectivity in populations of neurons is fundamental to understanding neural computation and function. In recent years, calcium imaging has emerged as a promising technique for
measuring synaptic activity and mapping neural micro-circuits [1?4]. Fluorescent calcium-sensitive
dyes and genetically-encoded calcium indicators can be loaded into neurons, which can then be imaged for spiking activity either in vivo or in vitro. Current methods enable imaging populations of
hundreds to thousands of neurons with very high spatial resolution. Using two-photon microscopy,
imaging can also be localized to specific depths and cortical layers [5]. Calcium imaging also has
the potential to be combined with optogenetic stimulation techniques such as in [6].
However, inferring neural connectivity from calcium imaging remains a mathematically and computationally challenging problem. Unlike anatomical methods, calcium imaging does not directly
measure connections. Instead, connections must be inferred indirectly from statistical relationships
between spike activities of different neurons. In addition, the measurements of the spikes from calcium imaging are indirect and noisy. Most importantly, the imaging introduces significant temporal
blurring of the spike times: the typical time constants for the decay of the fluorescent calcium concentration, [Ca2+ ], can be on the order of a second ? orders of magnitude slower than the spike rates
and inter-neuron dynamics. Moreover, the calcium imaging frame rate remains relatively slow ?
often less than 100 Hz. Hence, determining connectivity typically requires super-resolution of spike
times within the frame period.
1
To overcome these challenges, the recent work [7] proposed a Bayesian inference method to estimate functional connectivity from calcium imaging in a systematic manner. Unlike ?model-free?
approaches such as in [8], the method in [7] assumed a detailed functional model of the neural dynamics with unknown parameters including a connectivity weight matrix W. The model parameters
including the connectivity matrix can then be estimated via a standard EM procedure [9]. While the
method is general, one of the challenges in implementing it is the computational complexity. As
we discuss below, the E-step in the EM procedure essentially requires estimating the distributions
of hidden states in a nonlinear dynamical system whose state dimension grows linearly with the
number of neurons. Since exact computation of these densities grows exponentially in the state dimension, [7] uses an approximate method based on blockwise Gibbs sampling where each block of
variables consists of the hidden states associated with one neuron. Since the variables within a block
are described as a low-dimensional dynamical system, the updates of the densities for the Gibbs
sampling can be computed efficiently via a standard particle filter [10, 11]. However, simulations of
the method show that the mixing between blocks can still take considerable time to converge.
This paper provides a novel method that can potentially significantly improve the computation time
of the state estimation. The key insight is to recognize that a high-dimensional neural system can be
?factorized? into simple, scalar dynamical systems for each neuron with linear interactions between
the neurons. As described below, we assume a standard leaky integrate-and-fire model for each
neuron [12] and a first-order AR process for the calcium imaging [13]. Under this model, the
dynamics of N neurons can be described by 2N systems, each with a scalar (i.e. one-dimensional)
state. The coupling between the systems will be linear as described by the connectivity matrix
W. Using this factorization, approximate state estimation can then be efficiently performed via
approximations of loopy belief propagation (BP) [14]. Specifically, we show that the loopy BP
updates at each of the factor nodes associated with the integrate-and-fire and calcium imaging can
be performed via a scalar standard forward?backward filter. For the updates associated with the
linear transform W, we use recently-developed approximate message passing (AMP) methods.
AMP was originally proposed in [15] for problems in compressed sensing. Similar to expectation
propagation [16], AMP methods use Gaussian and quadratic approximations of loopy BP but with
further simplifications that leverage the linear interactions. AMP was used for neural mapping from
multi-neuron excitation and neural receptive field estimation in [17, 18]. Here, we use a so-called
hybrid AMP technique proposed in [19] that combines AMP updates across the linear coupling
terms with standard loopy BP updates on the remainder of the system. When applied to the neural
system, we show that the estimation updates become remarkably simple: For a system with N
neurons, each iteration involves running 2N forward?backward scalar state estimation algorithms,
along with multiplications by W and WT at each time step. The practical complexity scales as
O(N T ) where T is the number of time steps. We demonstrate that the method can be significantly
faster than the blockwise Gibbs sampling proposed in [7], with similar accuracy.
2
System Model
We consider a recurrent network of N spontaneously firing neurons. All dynamics are approximated
in discrete time with some time step ?, with a typical value ? = 1 ms. Importantly, this time step
is typically smaller than the calcium imaging period, so the model captures the dynamics between
observations. Time bins are indexed by k = 0, . . . , T ?1, where T is the number of time bins so that
T ? is the total observation time in seconds. Each neuron i generates a sequence of spikes (action
potentials) indicated by random variables ski taking values 0 or 1 to represent whether there was a
spike in time bin k or not. It is assumed that the discretization step ? is sufficiently small such that
there is at most one action potential from a neuron in any one time bin. The spikes are generated
via a standard leaky integrate-and-fire (LIF) model [12] where the (single compartment) membrane
voltage vik of each neuron i and its corresponding spike output sequence ski evolve as
v?ik+1 = (1 ? ?IF )vik + qik + dkvi ,
qik =
N
X
Wij sk??
+ bIF,i ,
j
dkvi ? N (0, ?IF ),
(1)
j=1
and
(vik+1 , sk+1
i )
=
(?
vik , 0)
(0, 1)
2
if vik < ?,
if v?ik ? ?,
(2)
where ?IF is a time constant for the integration leakage; ? is the threshold potential at which the
neurons spikes; bIF,i is a constant bias term; qik is the increase in the membrane potential from the
pre-synaptic spikes from other neurons and dkvi is a noise term including both thermal noise and
currents from other neurons that are outside the observation window. The voltage has been scaled
so that the reset voltage is zero. The parameter ? is the integer delay (in units of the time step
?) between the spike in one neuron and the increase in the membrane voltage in the post-synaptic
neuron. An implicit assumption in this model is the post-synaptic current arrives in a single time bin
with a fixed delay.
To determine functional connectivity, the key parameter to estimate will be the matrix W of the
weighting terms Wij in (1). Each parameter Wij represents the increase in the membrane voltage in
neuron i due to the current triggered from a spike in neuron j. The connectivity weight Wij will be
zero whenever neuron j has no connection to neuron i. Thus, determining W will determine which
neurons are connected to one another and the strengths of those connections.
For the calcium imaging, we use a standard model [7], where the concentration of fluorescent Calcium has a fast initial rise upon an action potential followed by a slow exponential decay. Specifically, we let zik = [Ca2+ ]k be the concentration of fluorescent Calcium in neuron i in time bin k and
assume it evolves as first-order auto-regressive AR(1) model,
zik+1 = (1 ? ?CA,i )zik + ski ,
(3)
where ?CA is the Calcium time constant. The observed net fluorescence level is then given by a
noisy version of zik ,
yik = aCA,i zik + bCA,i + dkyi , dkyi ? N (0, ?y ),
(4)
where aCA,i and bCA,i are constants and dyi is white Gaussian noise with variance ?y . Nonlinearities
such as saturation described in [13] can also be modeled.
As mentioned in the Introduction, a key challenge in calcium imaging is the relatively slow frame
rate which has the effect of subsampling of the fluorescence. To model the subsampling, we
let IF denote the set of time indices k on which we observe Fik . We will assume that fluorescence values are observed once every TF time steps for some integer period TF so that
IF = {0, TF , 2TF , . . . , KTF } where K is the number of Calcium image frames.
3
3.1
Parameter Estimation via Message Passing
Problem Formulation
Let ? be set of all the unknown parameters,
? = {W, ?IF , ?CA , ?IF , bIF,i , ?CA , aCA,i , bCA,i , i = 1, . . . , N },
(5)
which includes the connectivity matrix, time constants and various variances and bias terms. Estimating the parameter set ? will provide an estimate of the connectivity matrix W, which is our main
goal.
To estimate ?, we consider a regularized maximum likelihood (ML) estimate
?b = arg max L(y|?) + ?(?),
L(y|?) = ? log p(y|?),
(6)
?
where y is the set of observed values; L(y|?) is the negative log likelihood of y given the parameters
? and ?(?) is some regularization function. For the calcium imaging problem, the observations y
are the observed fluorescence values across all the neurons,
y = {y1 , . . . , yN } , yi = yik , k ? IF ,
(7)
where yi is the set of fluorescence values from neuron i, and, as mentioned above, IF is the set of
time indices k on which the fluorescence is sampled.
The regularization function ?(?) can be used to impose constraints or priors on the parameters. In
this work, we will assume a simple regularizer that only constrains the connectivity matrix W,
X
?(?) = ?kWk1 , kWk1 :=
|Wij |,
(8)
ij
3
where ? is a positive constant. The `1 regularizer is a standard convex function used to encourage
sparsity [20], which we know in this case must be valid since most neurons are not connected to one
another.
3.2
EM Estimation
Exact computation of ?b in (6) is generally intractable, since the observed fluorescence values y
depend on the unknown parameters ? through a large set of hidden variables. Similar to [7], we thus
use a standard EM procedure [9]. To apply the EM procedure to the calcium imaging problem, let
x be the set of hidden variables,
x = {v, z, q, s} ,
(9)
where v are the membrane voltages of the neurons, z the calcium concentrations, s the spike outputs
and q the linearly combined spike inputs. For any of these variables, we will use the subscript i (e.g.
vi ) to denote the values of the variables of a particular neuron i across all time steps and superscript
k (e.g. vk ) to denote the values across all neurons at a particular time step k. Thus, for the membrane
voltage
k
v = vik , vk = v1k , . . . , vN
, vi = vi0 , . . . , viT?1 .
The EM procedure alternately estimates distributions on the hidden variables x given the current
parameter estimate for ? (the E-step); and then updates the estimates for parameter vector ? given
the current distribution on the hidden variables x (the M-step).
? E-Step: Given parameter estimates ?b` , estimate
P (x|y, ?b` ),
(10)
which is the posterior distribution of the hidden variables x given the observations y and
current parameter estimate ?b` .
? M-step Update the parameter estimate via the minimization,
h
i
?b`+1 = arg min E L(x, y|?)|?b` + ?(?),
(11)
?
where L(x, y|?) is the joint negative log likelihood,
L(x, y|?) = ? log p(x, y|?).
(12)
In (11) the expectation is with respect to the distribution found in (10) and ?(?) is the
parameter regularization function.
The next two sections will describe how we approximately perform each of these steps.
3.3
E-Step estimation via Approximate Message Passing
For the calcium imaging problem, the challenging step of the EM procedure is the E-step, since
the hidden variables x to be estimated are the states and outputs of a high-dimensional nonlinear
dynamical system. Under the model in Section 2, a system with N neurons will require N states
for the membrane voltages vik and N states for the bound Ca concentration levels zik , resulting in
a total state dimension of 2N . The E-step for this system is essentially a state estimation problem,
and exact inference of the states of a general nonlinear dynamical system grows exponentially in the
state dimension. Hence, exact computation of the posterior distribution (10) for the system will be
intractable even for a moderately sized network.
As described in the Introduction, we thus use an approximate messaging passing method that exploits the separable structure of the system. For the remainder of this section, we will assume the
parameters ? in (5) are fixed to the current parameter estimate ?b` . Then, under the assumptions of
Section 2, the joint probability distribution function of the variables can be written in a factorized
form,
P (x, y) = P (q, v, s, z, y) =
N
T?1
Y
1 Y
1{qk =Wsk } ?iIF (qi , vi , si )?iCA (si , zi , yi ),
Z
i=1
k=0
4
(13)
input currents
qi
membrane voltage
Ca2+ concentration
vi
zi
?iIF (qi , vi , si )
Integrate-and-fire
dynamics
spike outputs
si
?iCA (si , zi , yi )
Ca imaging
dynamics
observed
fluorescence
yi
Neuron i, i = 1, . . . , N
qk = Wsk
Connectivity
between neurons
Time step k, k = 0, . . . , T ?1
Figure 1: Factor graph plate representation of the system where the spike dynamics are described
by the factor node ?iIF (qi , vi , si ) and the calcium image dynamics are represented via the factor
node ?iCA (si , zi , yi ). The high-dimensional dynamical system is described as 2N scalar dynamical
systems (2 for each neuron) with linear interconnections, qk = Wsk between the neurons. A
computational efficient approximation of loopy BP [19] is applied to this graph for approximate
Bayesian inference required in the E-step of the EM algorithm.
where Z is a normalization constant; ?iIF (qi , vi , si ) is the potential function relating the summed
spike inputs qi to the membrane voltages vi and spike outputs si ; ?iCA (si , zi , yi ) relates the spike
outputs si to the bound calcium concentrations zi and observed fluorescence values yi ; and the term
1{qk =Wsk } indicates that the distribution is to be restricted to the set satisfying the linear constraints
qk = Wsk across all time steps k.
As in standard loopy BP [14], we represent the distribution (13) in a factor graph as shown in
Fig. 1. Now, for the E-step, we need to compute the marginals of the posterior distribution p(x|y)
from the joint distribution (13). Using the factor graph representation, loopy BP iteratively updates
estimates of these marginal posterior distributions using a message passing procedure, where the
estimates of the distributions (called beliefs) are passed between the variable and factor nodes in
the graph. In general, the computationally challenging component of loopy BP is the updates on
the belief messages at the factor nodes. However, using the factorized structure in Fig. 1 along
with approximate message passing (AMP) simplifications as described in [19], these updates can be
computed easily.
Details are given in the full paper [21], but the basic procedure for the factor node updates and the
reasons why these computations are simple can be summarized as follows. At a high level, the factor
graph structure in Fig. 1 partitions the 2N -dimensional nonlinear dynamical system into N scalar
systems associated with each membrane voltage vik and an additional N scalar systems associated
with each calcium concentration level zik . The only coupling between these systems is through the
linear relationships qk = Wsk . As shown in Appendix ??, on each of the scalar systems, the factor
node updates required by loopy BP essentially reduces to a state estimation problem for this system.
Since the state space of this system is scalar (i.e. one-dimensional), we can discretize the state space
well with a small number of points ? in the experiments below we use L = 20 points per dimension.
Once discretized, the state estimation can be performed via a standard forward?backward algorithm.
If there are T time steps, the algorithm will have a computational cost of O(T L2 ) per scalar system.
Hence, all the factor node updates across all the 2N scalar systems has total complexity O(N T L2 ).
For the factor nodes associated with the linear constraints qk = Wsk , we use the AMP approximations [19]. In this approximation, the messages for the transform outputs qik are approximated as
Gaussians which is, at least heuristically, justified since the they are outputs of a linear transform of
a large number of variables, ski . In the AMP algorithm, the belief updates for the variables qk and
sk can then be computed simply by linear transformations of W and WT . Since W represents a
connectivity matrix, it is generally sparse. If each row of W has d non-zero values, multiplication
5
by W and WT will be O(N d). Performing the multiplications across all time steps results in a total
complexity of O(N T d).
Thus, the total complexity of the proposed E-step estimation method is O(N T L2 + N T d) per loopy
BP iteration. We typically use a small number of loopy BP iterations per EM update (in fact, in the
experiments below, we found reasonable performance with one loopy BP update per EM update).
In summary, we see that while the overall neural system is high-dimensional, it has a linear + scalar
structure. Under the assumption of the bounded connectivity d, this structure enables an approximate
inference strategy that scales linearly with the number of neurons N and time steps T . Moreover,
the updates in different scalar systems can be computed separately allowing a readily parallelizable
implementation.
3.4
Approximate M-step Optimization
The M-step (11) is computationally relatively simple. All the parameters in ? in (5) have a linear
relationship between the components of the variables in the vector x in (9). For example, the parameters aCA,i and bCA,i appear in the fluorescence output equation (4). Since the noise dkyi in this
equation is Gaussian, the negative log likelihood (12) is given by
L(x, y|?) =
1 X k
T
(yi ? aCA,i zik ? bCA,i )2 + log(?yi ) + other terms,
2?yi
2
k?IF
where ?other terms? depend on parameters other than aCA,i and bCA,i . The expectation
E(L(x, y|?)|?b` ) will then depend only on the mean and variance of the variables yik and zik , which
are provided by the E-step estimation. Thus, the M-step optimization in (11) can be computed via a
simple least-squares problem. Using the linear relation (1), a similar method can be used for ?IF,i
and bIF,i , and the linear relation (3) can be used to estimate the calcium time constant ?CA .
To estimate the connectivity matrix W, let rk = qk ? Wsk so that the constraints in (13) is equivalent to the condition that rk = 0. Thus, the term containing W in the expectation of the negative log
likelihood E(L(x, y|?)|?b` ) is given by the negative log probability density of rk evaluated at zero.
In general, this density will be a complex function of W and difficult to minimize. So, we approxb and b
imate the density as follows: Let q
s be the expectation of the variables q and s given by the
bk ? Wb
E-step. Hence, the expectation of rk is q
sk . As a simple approximation, we will then assume
that the variables rik are Gaussian, independent and having some constant variance ? 2 . Under this
simplifying assumption, the M-step optimization of W with the `1 regularizer (8) reduces to
c = arg min 1
W
2
W
T?1
X
kb
qk ? Wb
sk k2 + ? 2 ?kWk1 ,
(14)
k=0
For a given value of ? 2 ?, the optimization (14) is a standard LASSO optimization [22] which can be
evaluated efficiently via a number of convex programming methods. In this work, in each M-step,
we adjust the regularization parameter ? 2 ? to obtain a desired fixed sparsity level in the solution W.
3.5
Initial Estimation via Sparse Regression
Since the EM algorithm cannot be guaranteed to converge a global maxima, it is important to pick
the initial parameter estimates carefully. The time constants and noise levels for the calcium image
can be extracted from the second-order statistics of fluorescence values and simple thresholding can
provide a coarse estimate of the spike rate.
The key challenge is to obtain a good estimate for the connectivity matrix W. For each neuron i, we
first make an initial estimate of the spike probabilities P (ski = 1|yi ) from the observed fluorescence
values yi , assuming some i.i.d. prior of the form P (sti ) = ??, where ? is the estimated average spike
rate per second. This estimation can be solved with the filtering method in [13] and is also equivalent
to the method we use for the factor node updates. We can then threshold these probabilities to make
a hard initial decision on each spike: ski = 0 or 1. We then propose to estimate W from the spikes
as follows. Fix a neuron i and let wi be the vector of weights Wij , j = 1, . . . , N . Under the
assumption that the initial spike sequence ski is exactly correct, it is shown in the full paper [21], that
6
Parameter
Number of neurons, N
Connection sparsity
Value
100
10% with random connections. All connections are excitatory
with the non-zero weights Wij being exponentially distributed.
10 Hz
1 ms
10 sec (10,000 time steps)
20 ms
2 time steps = 2 ms
Produced from two unobserved neurons.
500 ms
Set to 20 dB SNR
100 Hz
Mean firing rate per neuron
Simulation time step, ?
Total simulation time, T ?
Integration time constant, ?IF
Conduction delay, ?
Integration noise, dkvi
Ca time constant, ?CA
Fluorescence noise, ?CA
Ca frame rate , 1/TF
Table 1: Parameters for the Ca image simulation.
Figure 2: Typical network simulation
trace. Top panel: Spike traces for
the 100 neuron simulated network.
Bottom panel: Calcium image fluorescence levels. Due to the random network topology, neurons often
fire together, significantly complicating connectivity detection. Also, as
seen in the lower panel, the slow decay of the fluorescent calcium blurs
the spikes in the calcium image.
a regularized maximum likelihood estimate of wi and bias term bIF,i is given by
b i , bbIF,i ) = arg min
(w
wi ,bIF,i
T?1
X
Lik (uTk wi + cik bIF,i ? ?, ski ) + ?
N
X
|Wij |,
(15)
j=1
k=0
where Lik is a probit loss function and the vector uk and scalar cik can be determined from the
spike estimates. The optimization (15) is precisely a standard probit regression used in sparse linear
classification [23]. This form arises due to the nature of the leaky integrate-and-fire model (1) and
(2). Thus, assuming the initial spike sequences are estimated reasonably accurately, one can obtain
good initial estimates for the weights Wij and bias terms bIF,i by solving a standard classification
problem.
4
Numerical Example
The method was tested using realistic network parameters, as shown in Table 1, similar to those
found in neurons networks within a cortical column [24]. Similar parameters are used in [7]. The
network consisted of 100 neurons with each neuron randomly connected to 10% of the other neurons. The non-zero weights Wij were drawn from an exponential distribution. As a simplification,
all weights were positive (i.e. the neurons were excitatory ? there were no inhibitory neurons in the
simulation). A typical random matrix W generated in this manner would not in general result in a
stable system. To stabilize the system, we followed the procedure in [8] where the system is simulated multiple times. After each simulation, the rows of the matrix W were adjusted up or down to
increase or decrease the spike rate until all neurons spiked at a desired target rate. In this case, we
assumed a desired average spike rate of 10 Hz.
7
Figure 3: Weight estimation accuracy. Left: Normalized mean-squared error as a function of the
iteration number. Right: Scatter plot of the true and estimated weights.
From the parameters in Table 1, we can immediately see the challenges in the estimation. Most
importantly, the calcium imaging time constant ?CA is set for 500 ms. Since the average neurons
spike rate is assumed to be 10 Hz, several spikes will typically appear within a single time constant.
Moreover, both the integration time constant and inter-neuron conduction time are much smaller
than the
A typical simulation of the network after the stabilization is shown in Fig. 2. Observe that due to
the random connectivity, spiking in one neuron can rapidly cause the entire network to fire. This
appears as the vertical bright stripes in the lower panel of Fig. 2. This synchronization makes the
connectivity detection difficult to detect under temporal blurring of Ca imaging since it is hard to
determine which neuron is causing which neuron to fire. Thus, the random matrix is a particularly
challenging test case.
The results of the estimation are shown in Fig. 3. The left panel shows the relative mean squared
error defined as
P
cij |2
min? ij |Wij ? ?W
P
relative MSE =
,
(16)
2
ij |Wij |
cij is the estimate for the weight Wij . The minimization over all ? is performed since the
where W
method can only estimate the weights up to a constant scaling. The relative MSE is plotted as a
function of the EM iteration, where we have performed only a single loopy BP iteration for each
EM iteration. We see that after only 30 iterations we obtain a relative MSE of 7% ? a number at
least comparable to earlier results in [7], but with significantly less computation. The right panel
cij against the true weights Wij .
shows a scatter plot of the estimated weights W
5
Conclusions
We have presented a scalable method for inferring connectivity in neural systems from calcium
imaging. The method is based on factorizing the systems into scalar dynamical systems with linear
connections. Once in this form, state estimation ? the key computationally challenging component
of the EM estimation ? is tractable via approximating message passing methods. The key next step
in the work is to test the methods on real data and also provide more comprehensive computational
comparisons against current techniques such as [7].
References
[1] R. Y. Tsien, ?Fluorescent probes of cell signaling,? Ann. Rev. Neurosci., vol. 12, no. 1, pp. 227?253, 1989.
[2] K. Ohki, S. Chung, Y. H. Ch?ng, P. Kara, and R. C. Reid, ?Functional imaging with cellular resolution
reveals precise micro-architecture in visual cortex,? Nature, vol. 433, no. 7026, pp. 597?603, 2005.
[3] J. Soriano, M. R. Mart??nez, T. Tlusty, and E. Moses, ?Development of input connections in neural cultures,? Proc. Nat. Acad. Sci., vol. 105, no. 37, pp. 13 758?13 763, 2008.
[4] C. Stosiek, O. Garaschuk, K. Holthoff, and A. Konnerth, ?In vivo two-photon calcium imaging of neuronal networks,? Proc. Nat. Acad. Sci., vol. 100, no. 12, pp. 7319?7324, 2003.
8
[5] K. Svoboda and R. Yasuda, ?Principles of two-photon excitation microscopy and its applications to neuroscience,? Neuron, vol. 50, no. 6, pp. 823?839, 2006.
[6] O. Yizhar, L. E. Fenno, T. J. Davidson, M. Mogri, and K. Deisseroth, ?Optogenetics in neural systems,?
Neuron, vol. 71, no. 1, pp. 9?34, 2011.
[7] Y. Mishchenko, J. T. Vogelstein, and L. Paninski, ?A Bayesian approach for inferring neuronal connectivity from calcium fluorescent imaging data,? Ann. Appl. Stat., vol. 5, no. 2B, pp. 1229?1261, Feb. 2011.
[8] O. Stetter, D. Battaglia, J. Soriano, and T. Geisel, ?Model-free reconstruction of excitatory neuronal
connectivity from calcium imaging signals,? PLoS Computational Biology, vol. 8, no. 8, p. e1002653,
2012.
[9] A. Dempster, N. M. Laird, and D. B. Rubin, ?Maximum-likelihood from incomplete data via the EM
algorithm,? J. Roy. Statist. Soc., vol. 39, pp. 1?17, 1977.
[10] A. Doucet, S. Godsill, and C. Andrieu, ?On sequential Monte Carlo sampling methods for Bayesian
filtering,? Statistics and Computing, vol. 10, no. 3, pp. 197?208, 2000.
[11] A. Doucet and A. M. Johansen, ?A tutorial on particle filtering and smoothing: Fifteen years later,?
Handbook of Nonlinear Filtering, vol. 12, pp. 656?704, 2009.
[12] P. Dayan and L. F. Abbott, Theoretical Neuroscience. Computational and Mathematical Modeling of
Neural Systems. MIT Press, 2001.
[13] J. T. Vogelstein, B. O. Watson, A. M. Packer, R. Yuste, B. Jedynak, and L. Paninski, ?Spike inference
from calcium imaging using sequential monte carlo methods,? Biophysical J., vol. 97, no. 2, pp. 636?655,
2009.
[14] M. J. Wainwright and M. I. Jordan, ?Graphical models, exponential families, and variational inference,?
Found. Trends Mach. Learn., vol. 1, 2008.
[15] D. L. Donoho, A. Maleki, and A. Montanari, ?Message-passing algorithms for compressed sensing,?
Proc. Nat. Acad. Sci., vol. 106, no. 45, pp. 18 914?18 919, Nov. 2009.
[16] T. P. Minka, ?A family of algorithms for approximate Bayesian inference,? Ph.D. dissertation, Massachusetts Institute of Technology, Cambridge, MA, 2001.
[17] A. K. Fletcher, S. Rangan, L. Varshney, and A. Bhargava, ?Neural reconstruction with approximate message passing (NeuRAMP),? in Proc. Neural Information Process. Syst., Granada, Spain, Dec. 2011.
[18] U. S. Kamilov, S. Rangan, A. K. Fletcher, and M. Unser, ?Approximate message passing with consistent
parameter estimation and applications to sparse learning,? in Proc. NIPS, Lake Tahoe, NV, Dec. 2012.
[19] S. Rangan, A. K. Fletcher, V. K. Goyal, and P. Schniter, ?Hybrid generalized approximation message
passing with applications to structured sparsity,? in Proc. IEEE Int. Symp. Inform. Theory, Cambridge,
MA, Jul. 2012, pp. 1241?1245.
[20] D. L. Donoho, ?Compressed sensing,? IEEE Trans. Inform. Theory, vol. 52, no. 4, pp. 1289?1306, Apr.
2006.
[21] A. K. Fletcher and S. Rangan, ?Scalable inference for neuronal connectivity from calcium imaging,?
arXiv:1409.0289, Sep. 2014.
[22] R. Tibshirani, ?Regression shrinkage and selection via the lasso,? J. Royal Stat. Soc., Ser. B, vol. 58, no. 1,
pp. 267?288, 1996.
[23] C. M. Bishop, Pattern Recognition and Machine Learning, ser. Information Science and Statistics. New
York, NY: Springer, 2006.
[24] R. Sayer, M. Friedlander, and S. Redman, ?The time course and amplitude of EPSPs evoked at synapses
between pairs of CA3/CA1 neurons in the hippocampal slice,? J. Neuroscience, vol. 10, no. 3, pp. 826?
836, 1990.
9
| 5341 |@word version:1 heuristically:1 simulation:9 simplifying:1 pick:1 fifteen:1 deisseroth:1 initial:8 amp:11 current:12 discretization:1 si:11 scatter:2 must:2 readily:1 written:1 numerical:1 realistic:2 partition:1 blur:1 enables:1 plot:2 update:21 zik:9 dissertation:1 regressive:1 provides:2 coarse:1 node:10 tahoe:1 mathematical:1 along:2 become:1 ik:2 consists:1 combine:1 symp:1 manner:2 inter:2 ica:4 multi:1 bif:8 discretized:1 window:1 provided:1 estimating:2 underlying:1 moreover:3 circuit:3 factorized:4 bounded:1 panel:6 spain:1 developed:1 ca1:1 unobserved:1 transformation:1 temporal:3 every:1 exactly:1 scaled:1 k2:1 uk:1 ser:2 unit:1 yn:1 appear:2 reid:1 positive:2 approxb:1 acad:3 mach:1 subscript:1 firing:2 approximately:1 evoked:1 challenging:5 appl:1 factorization:1 ktf:1 jedynak:1 practical:1 spontaneously:1 block:3 goyal:1 signaling:1 procedure:11 significantly:5 pre:1 cannot:1 selection:1 equivalent:2 vit:1 convex:2 resolution:3 immediately:1 fik:1 insight:2 importantly:3 population:2 target:1 svoboda:1 exact:4 programming:1 us:1 trend:1 roy:1 approximated:2 satisfying:1 particularly:1 recognition:1 stripe:1 holthoff:1 observed:8 bottom:1 solved:1 capture:1 thousand:2 connected:3 plo:1 decrease:1 mentioned:2 dempster:1 complexity:5 constrains:1 moderately:1 dynamic:9 depend:3 solving:1 upon:1 blurring:2 alyson:1 easily:1 joint:3 indirect:1 sep:1 various:1 represented:1 e1002653:1 regularizer:3 fast:2 describe:1 monte:3 outside:1 whose:1 emerged:1 encoded:1 interconnection:2 compressed:3 optogenetics:1 statistic:3 transform:3 noisy:2 laird:1 superscript:1 sequence:4 triggered:1 biophysical:1 net:1 propose:2 reconstruction:2 interaction:2 reset:1 remainder:2 causing:1 rapidly:1 mixing:1 redman:1 coupling:3 recurrent:1 stat:2 ij:3 soc:2 geisel:1 epsps:1 involves:2 correct:1 filter:2 kb:1 stabilization:1 enable:1 implementing:1 bin:6 require:1 fix:1 mathematically:1 adjusted:1 sufficiently:1 fletcher:5 mapping:2 battaglia:1 estimation:24 proc:6 sensitive:1 fluorescence:14 tf:5 tool:1 minimization:2 mit:1 gaussian:4 super:1 shrinkage:1 voltage:11 vk:2 likelihood:7 indicates:1 detect:1 inference:10 dayan:1 typically:5 entire:1 hidden:8 relation:2 wij:14 arg:4 overall:1 classification:2 development:1 smoothing:1 spatial:1 lif:1 integration:4 summed:1 marginal:1 field:1 once:3 having:1 ng:1 sampling:4 iif:4 represents:2 biology:1 micro:2 randomly:1 packer:1 recognize:1 comprehensive:1 sundeep:1 fire:8 detection:3 message:13 adjust:1 introduces:1 arrives:1 chain:1 dyi:1 konnerth:1 schniter:1 encourage:1 culture:1 vi0:1 indexed:1 incomplete:1 desired:3 plotted:1 theoretical:1 column:1 optogenetic:1 wb:2 earlier:1 modeling:1 ar:2 measuring:1 maximization:1 loopy:14 cost:1 ca3:1 snr:1 hundred:1 delay:3 conduction:2 wsk:8 combined:2 density:5 fundamental:1 systematic:1 together:1 connectivity:29 squared:2 containing:1 messaging:1 chung:1 yasuda:1 syst:1 potential:7 photon:3 nonlinearities:1 summarized:1 sec:1 includes:1 stabilize:1 int:1 vi:8 performed:5 later:1 aca:6 jul:1 vivo:2 minimize:1 square:1 compartment:1 accuracy:3 bright:1 loaded:1 variance:4 efficiently:3 qk:10 yield:1 bayesian:6 accurately:1 produced:1 carlo:3 synapsis:1 parallelizable:1 inform:2 whenever:1 synaptic:4 against:2 pp:16 minka:1 associated:6 sampled:1 massachusetts:1 amplitude:1 carefully:1 cik:2 appears:1 originally:1 methodology:1 response:1 formulation:1 evaluated:2 implicit:1 until:1 nonlinear:6 propagation:3 indicated:1 grows:3 effect:1 consisted:1 normalized:1 true:2 andrieu:1 hence:4 regularization:4 maleki:1 imaged:1 iteratively:1 white:1 excitation:2 m:6 generalized:1 hippocampal:1 plate:1 demonstrate:2 image:6 variational:1 novel:1 recently:1 stimulation:1 spiking:3 vitro:1 functional:4 exponentially:3 relating:1 marginals:1 measurement:1 significant:1 cambridge:2 gibbs:3 particle:2 stable:1 cortex:1 feb:1 posterior:4 recent:2 dye:1 neuramp:1 watson:1 kwk1:3 kamilov:1 yi:13 seen:1 additional:1 impose:1 utk:1 converge:2 determine:3 period:3 signal:1 vogelstein:2 relates:1 full:2 lik:2 multiple:1 reduces:2 ohki:1 faster:2 post:2 qi:6 scalable:4 basic:1 regression:3 essentially:3 expectation:7 arxiv:1 iteration:8 represent:2 normalization:1 microscopy:2 qik:4 cell:1 justified:1 addition:1 remarkably:1 separately:1 dec:2 unlike:2 nv:1 hz:5 db:1 jordan:1 integer:2 leverage:1 zi:6 architecture:1 lasso:2 topology:1 vik:8 soriano:2 whether:1 passed:1 passing:12 cause:1 york:1 action:3 yik:3 generally:2 detailed:1 transforms:1 statist:1 ph:1 inhibitory:1 tutorial:1 moses:1 estimated:6 neuroscience:3 per:7 tibshirani:1 anatomical:1 discrete:1 vol:17 key:8 threshold:2 drawn:1 abbott:1 backward:3 imaging:32 graph:6 year:2 sti:1 powerful:1 ca2:3 family:2 reasonable:1 vn:1 lake:1 decision:1 appendix:1 scaling:1 comparable:1 layer:1 bound:2 followed:2 simplification:3 guaranteed:1 quadratic:1 activity:3 strength:1 constraint:4 precisely:1 rangan:5 bp:13 generates:1 min:4 performing:1 separable:1 relatively:3 structured:1 membrane:10 across:7 smaller:2 em:17 wi:4 evolves:1 rev:1 restricted:1 spiked:1 computationally:7 equation:2 remains:2 imate:1 discus:1 know:1 tractable:1 gaussians:1 apply:1 observe:2 probe:1 indirectly:1 slower:2 top:1 running:1 subsampling:2 graphical:1 exploit:1 approximating:1 leakage:1 spike:35 receptive:1 concentration:8 strategy:1 simulated:2 sci:3 cellular:1 reason:1 assuming:2 modeled:1 relationship:3 index:2 difficult:2 cij:3 potentially:3 blockwise:2 trace:2 negative:5 rise:1 godsill:1 implementation:1 calcium:42 ski:8 unknown:3 perform:1 allowing:1 discretize:1 vertical:1 neuron:65 observation:5 markov:1 thermal:1 precise:1 frame:6 y1:1 inferred:1 bk:1 pair:1 required:2 connection:9 johansen:1 alternately:1 nip:1 trans:1 dynamical:11 below:4 pattern:1 sparsity:4 challenge:6 genetically:1 saturation:1 including:3 max:1 royal:1 belief:5 wainwright:1 demanding:1 hybrid:4 regularized:2 indicator:1 bhargava:1 improve:1 technology:1 auto:1 bca:6 understanding:1 prior:2 l2:3 friedlander:1 multiplication:3 determining:3 evolve:1 relative:4 synchronization:1 probit:2 loss:1 stetter:1 yuste:1 limitation:1 fluorescent:8 filtering:4 localized:1 integrate:5 rik:1 consistent:1 rubin:1 thresholding:1 principle:1 granada:1 row:2 excitatory:3 summary:1 course:1 free:2 bias:4 institute:1 taking:1 leaky:3 sparse:4 distributed:1 slice:1 overcome:2 depth:1 cortical:2 dimension:5 valid:1 complicating:1 forward:3 approximate:13 nov:1 varshney:1 ml:1 global:1 doucet:2 reveals:1 handbook:1 assumed:4 davidson:1 factorizing:1 sk:5 why:1 table:3 promising:1 learn:1 nature:2 reasonably:1 ca:14 mse:3 complex:1 apr:1 main:1 montanari:1 linearly:3 neurosci:1 noise:7 kara:1 mishchenko:1 neuronal:5 fig:6 slow:4 ny:1 inferring:5 exponential:3 weighting:1 rk:4 down:1 specific:1 bishop:1 sensing:3 unser:1 decay:3 intractable:2 sequential:2 v1k:1 magnitude:2 nat:3 tsien:1 simply:1 nez:1 paninski:2 visual:1 scalar:16 springer:1 ch:1 extracted:1 mart:1 ma:2 viewed:1 goal:1 sized:1 ann:2 donoho:2 considerable:1 hard:2 typical:5 specifically:2 determined:1 wt:3 called:2 total:6 arises:1 tested:1 |
4,796 | 5,342 | Sparse space-time deconvolution
for Calcium image analysis
Ferran Diego
Fred A. Hamprecht
Heidelberg Collaboratory for Image Processing (HCI)
Interdisciplinary Center for Scientific Computing (IWR)
University of Heidelberg, Heidelberg 69115, Germany
{ferran.diego,fred.hamprecht}@iwr.uni-heidelberg.de
Abstract
We describe a unified formulation and algorithm to find an extremely sparse representation for Calcium image sequences in terms of cell locations, cell shapes,
spike timings and impulse responses. Solution of a single optimization problem
yields cell segmentations and activity estimates that are on par with the state of
the art, without the need for heuristic pre- or postprocessing. Experiments on real
and synthetic data demonstrate the viability of the proposed method.
1
Introduction
A detailed understanding of brain function is a still-elusive grand challenge. Experimental evidence
is collected mainly by electrophysiology and ?Calcium imaging?. In the former, multi-electrode
array recordings allow the detailed study of hundreds neurons, while field potentials reveal the collective action of dozens or hundreds of neurons. The more recent Calcium imaging, on the other
hand, is a fluorescent microscopy technique that allows the concurrent monitoring of the individual actions of thousands of neurons at the same time. While its temporal resolution is limited by
the chemistry of the employed fluorescent markers, its great information content makes Calcium
imaging an experimental technique of first importance in the study of neural processing, both in
vitro [16, 6] and in vivo [5, 7]. However, the acquired image sequences are large, and in laboratory
practice the analysis remains a semi-manual, tedious and subjective task.
Calcium image sequences reveal the activity of neural tissue over time. Whenever a neuron fires,
its fluorescence signal first increases and then decays in a characteristic time course. Evolutionary
and energetic constraints on the brain guarantee that, in most cases, neural activity is sparse in
both space (only a fraction of neurons fire at a given instant) and time (most neurons fire only
intermittently). The problem setting can be formalized as follows: given an image sequence as
input, the desired output is (i) a set of cells1 and (ii) a set of time points at which these cells were
triggered. We here propose an unsupervised learning formulation and algorithm that leverages the
known structure of the data to produce the sparsest representations published to date, and allow for
meaningful automated analysis.
1.1
Prior Art
Standard laboratory practice is to delineate each cell manually by a polygon, and then integrate their
fluorescence response over the polygon, for each point in time. The result is a set of time series, one
per cell.
1
Optical sectioning by techniques such as confocal or two-photon microscopy implies that we see only parts
of a neuron, such as a slice through its cell body or a dendrite, in an image plane. For brevity, we simply refer
to these as ?cells? in the following.
1
a) Matrix factorization [13, 15, 4, 3, 12]
b) Convolutional sparse coding [8, 25, 20, 17, 14]
Figure 1: Sketch of selected previous work. Left: Decomposition of an image sequence into a sum
of K components. Each component is given by the Cartesian product of a spatial component or
basis image Dk and its temporal evolution uk . In this article, we represent such Cartesian products
by the convolution of multidimensional arrays. Right: Description of a single image in terms of a
sum of latent feature maps Dk convolved with filters Hk
Given that the fluorescence signal impulse response to a stimulus is stereotypic, these time series
can then be deconvolved to obtain a sparse temporal representation for each cell using nonnegative
sparse deconvolution [24, 5, 10].
The problem of automatically identifying the cells has received much less attention, possibly due to
the following difficulties [16, 23]: i) low signal-to-noise ratio (SNR); ii) large variation in luminance
and contrast; iii) heterogeneous background; iv) partial occlusion and v) pulsations due to heartbeat
or breathing in live animals. Existing work either hard-codes prior knowledge on the appearance of
specific cell types [16, 22] or uses supervised learning (pixel and object level classification, [23]) or
unsupervised learning (convolutional sparse block coding, [14]).
Closest in spirit to our work are attempts to simultaneously segment the cells and estimate their time
courses. This is accomplished by matrix factorization techniques such as independent component
analysis [13], nonnegative matrix factorization [12] and (hierarchical) dictionary learning [4, 3].
However, none of the above give results that are truly sparse in time; and all of the above have to go
to some lengths to obtain reasonable cell segmentations: [13, 4] recur to heuristic post-processing,
while [3] invokes structured sparsity inducing norms and [15] use an additional clustering step as
initialization. All these extra steps are necessary to assure that each spatial component represents
exactly one cell.
In terms of mathematical modeling, we build on recent advances and experiments in convolutional
sparse coding such as [8, 25, 20, 17, 14]. Ref. [21] already applies convolutional sparse coding to
video, but achieves sparsity only in space and not in time (where only pairs of frames are used to
learn latent representations). Refs. [19, 18] consider time series which they deconvolve, without
however using (or indeed needing, given their data) a sparse spatial representation.
1.2
Contributions
Summarizing prior work, we see three strands: i) Fully automated methods that require an extrinsic cell segmentation, but can find a truly2 sparse representation of the temporal activity. ii) Fully
automated methods that can detect and segment cells, but do not estimate time courses in the same
framework. iii) Techniques that both segment cells and estimate their time courses. Unfortunately,
existing techniques either produce temporal representations that are not truly sparse [12, 4, 3] or do
not offer a unified formulation of segmentation and activity detection that succeeds without extraneous clustering steps [15].
In response, we offer the first unified formulation in terms of a single optimization problem: its
solution simultaneously yields all cells along with their actions over time. The representation of
activity is truly sparse, ideally in terms of a single nonzero coefficient for each distinct action of a
cell. This is accomplished by sparse space-time deconvolution. Given a motion-corrected sequence
of Calcium images, it estimates i) locations of cells and ii) their activity along with iii) typical cell
shapes and iv) typical impulse responses. Taken together, these ingredients afford the sparsest, and
thus hopefully most interpretable, representation of the raw data. In addition, our joint formulation
2
We distinguish a sparse representation, in which the estimated time course of a cell has many zeros; and a
?truly sparse? representation in which a single action of a cell is ideally represented in terms of a single nonzero
coefficient.
2
Figure 2: Summary of sparse space-time deconvolution. Top: Unified formulation in terms of a
single optimization problem. Bottom: Illustration on tiny subset of data. Left: raw data. The
fluorescence level to be estimated is heavily degraded by Poisson shot noise that is unavoidable
at the requisite short exposure times. Middle: smoothed raw data. Right: approximation of the
data in terms of a Cartesian product of estimated cell shapes and temporal activities. Each temporal
activity is further decomposed as a convolution of estimated impulse responses and very few nonzero
coefficients.
allows to estimate a nonuniform and temporally variable background. Experiments on difficult
artificial and real-world data show the viability of the proposed formulation.
Notation Boldface symbols describe multidimensional arrays. We define A ? B as a convolution of
multidimensional arrays A and mirror(B), with the result trimmed to the dimensions of A. Here,
the ?mirror? operation flips a multidimensional array along every dimension. A ~ B is the full
convolution result of multidimensional arrays A and mirror(B). These definitions are analogous to
the ?convn? command in matlab with shape arguments ?same? and ?full?, respectively. k ? k0 counts
the number of non-zero coeficients, and k ? kF is the Frobenius norm.
2
2.1
Sparse space-time deconvolution (SSTD)
No background subtraction
An illustration of the proposed formulation is given in Fig. 2 and our notation is summarized in
Table. 1. We seek to explain image sequence X in terms of up to K cells and their activity over time.
In so doing, all cells are assumed to have exactly one (Eq. 1.1) of J << K possible appearances,
and to reside at a unique location (Eq. 1.1). These cell locations are encoded in K latent binary
feature maps. The activity of each cell is further decomposed in terms of a convolution of impulses
(giving the precise onset of each burst) with exactly one of L << K types of impulse responses.
A single cell may ?use? different impulse responses at different times, but just one type at any one
time ((Eq. 1.2).
All of the above is achieved by solving the following optimization problem:
?
?
K
J
X
X
?
min X ?
Dk,j ? Hj ? ~
D,H,f ,s
j=1
k=1
3
!2
sk,l ? fl
l=1
L
X
F
(1)
Constraint
P
Pj kDk,j k0 ? 1, ?k
such that
l kst,k,l k0 ? 1, ?k, t
kHj k2F ? 1, ?j
kfl k22 ? 1, ?l
(1.1)
(1.2)
(1.3)
(1.4)
Interpretation
at most one location and appearance per component
only one type of activation at each time per cell
prevent cell appearance from becoming large
prevent impulse filter from becoming large
Here, the optimization is with respect to the cell detection maps D, cell appearances H, activity
patterns or impulse responses f as well as ?truly sparse? activity indicator vectors s. This optimization is subject to the two constraints mentioned earlier plus upper bounds on the norm of the learned
filters.
The user needs to select the following parameters: an upper bound K on the number of cells as
well as the size in pixels H of their matched filters / convolution kernels H; upper bounds J on
the number of different appearances and L on the number of different activity patterns that cells
may have; as well as the number of coefficients F that the learned impulse responses may have.
Considering that we propose a method for both cell detection and sparse time course estimation,
the number of six user-adjustable parameters compares favourably to previous work. Methods that
decouple these steps typically need more parameters altogether, and the heuristics that prior work
on joint optimization uses also have a large number of (implicit) parameters.
PK
PK PJ
While many other approximations such as k=1 Dk ~ sk ? fk or k=1 j=1 Dk,j ? Hj ~ sk,j ? fj
are conceivable and may make sense in other applications areas, the proposed formulation is the
most parsimonious of its kind. Indeed, it uses a small pool of J shapes and L firing patterns, which
can be combined freely, to represent all cells and their activities. It is owing to this fact that we dub
the method sparse spatio-temporal deconvolution (SSTD).
2.2
SSTD with background subtraction
In actual experiments, the observed fluorescence level is a sum of the signal of interest plus a nuisance background signal. This background is typically nonuniform in the spatial domain and, while
it can be modeled as constant over time [15, 24], is often also observed to vary over time, prompting
robust local normalization as a preprocessing step [7, 4].
Here, we generalize the formulation from (1) to correct for a background that is assumed to be
spatially smooth and time-varying. In more detail, we model the background in terms of the direct
?N ?1
1?1?T
product Bs ~ bt of a spatial component Bs ? RM
and a time series bt ? R+
. Insights
+
into the physics and biology of Calcium imaging suggest that (except for saturation regimes characterized by high neuron firing rates), it is reasonable to assume that the normalized quantity (observed
fluorescence minus background) divided by background, typically dubbed ?F/F0 , is linearly related
to the intracellular Calcium concentration [24, 10]. In keeping with this notion, we now propose our
final model, viz.
?
2
?
?
?
!
K
J
L
X
X
X
s
t?
s
t
?
?
?
min
X?
Dk,j ? Hj ~
sk,l ? fl ? B ~ b B ~ b
D,H,f ,s,Bs ,bt
j=1
k=1
l=1
F
+ ?kBs kT V such that (1.1) ? (1.4), Bs > 0, bt > 0
(2)
with ?? denoting an elementwise division. Note that the optimization now also runs over the
spatial and temporal components of the background, with the total variation (TV) regularization
term3 enforcing spatial smoothness of the spatial background component [2].
In addition to the previously defined parameters, the user also needs to select parameter ? which
determines the smoothness of the background estimate.
2.3
Optimization
The optimization problem in (2) is convex in either the spatial or the temporal filters H, f alone when
keeping all other unknowns fixed; but it is nonconvex in general. In our experiments, we use a block
coordinate descent strategy [1, Section 2.7] that iteratively optimizes one group of variables while
3
TV measures the sum of the absolute values of the spatial gradient.
4
Symbol
?N ?T
X ? RM
+
K ? N+
J ? N+
Hj ? RH?H?1
+
Dk,j ? {0, 1}M ?N ?1
L ? N+
fl ? R1?1?F
+
sk,l ? R1?1?T
+
Definition
image sequence of length T , each image is M ? N
number of cells
number of distinct cell appearances
j th cell appearance / spatial filter / matched filter of size H ? H
indicator matrix of the k th cell for the j th cell appearance
number of distinct impulse responses / activity patterns
lth impulse response of length F
indicator vector of the k th spike train for the lth impulse response
Table 1: Notation
fixing all others (see supplementary material for details). The nonconvex l0 -norm constraints require
that cell centroids D and spike trains s are estimated by techniques such as convolutional matching
pursuit [20]; while the spatio-temporal filters can be learned using simpler gradient descent [25],
K-SVD [20] or simple algebraic expressions.
All unknowns are initialized with standard Gaussian noise truncated to nonnegative values. The
limiting number of cells K can be set to a generous upper bound of the expected true number
because spatial components without activity are automatically set to zero during optimization.
3
Experimental Setup
This section describes the data and algorithms used for experiments and benchmarks.
3.1
Inferring Spike Trains
The following methods assume that cell segmentation has already been performed by some means,
and that the fluorescence signal of individual pixels has been summed up for each cell and every time
step. They can hence concentrate exclusively on the estimation of a ?truly sparse? representation of
the respective activities in terms of a ?spike train?.
Data We follow [24, 5] in generating 1100 sequences consisting of one-sided exponential decays
with a constant amplitude of 1 and decay rate ? = 1/2s, sampled at 30f ps with firing rates ranging
uniformly from 1 to 10Hz and different Gaussian noise levels ? ? [0.1, 0.6].
Fast non-negative deconvolution (FAST) [24] uses a one-sided exponential decay as parametric
model for the impulse response by invoking a first-order autoregressive process. Given that our
artificial data is free of a nuisance background signal, we disregard its ability to also model such
background. The sole remaining parameter, the rate of the exponential decay, can be fit using maximum likelihood estimation or a method-of-moments approach [15].
Peeling [5] finds spikes by means of a greedy approach that iteratively removes one impulse response
at a time from the residual fluorescence signal. Importantly, this stereotypical transient must be
manually defined a priori.
Sparse temporal deconvolution (STD) with a single impulse response is a special case of this work
for given nonoverlapping cell segmentations and L = 1; and it is also a special case of [14]. The
impulse response can be specified beforehand (amounting to sparse coding), or learned from the
data (that is, performing dictionary learning on time-series data).
3.2
Segmenting Cells and Estimating Activities
Data Following the procedure described in [4, 12, 13], we have created 80 synthetic sequences
with a duration of 15s each at a frame rate of 30f ps with image sizes M = N = 512 pixels.
The cells are randomly selected from 36 cell shapes extracted from real data, and are randomly
located in different locations with a maximum spatial overlap of 30%. Each cell fires according to
a dependent Poisson process, and its activation pattern follows a one-sided exponential decay with
5
a scale selected uniform randomly between 500 and 800ms. The average number of active cells
per frame varies from 1 to 10. Finally, the data has been distorted by additive white Gaussian noise
with a relative amplitude (max. intensity ? mean intensity)/?noise ? {3, 5, 7, 10, 12, 15, 17, 20}.
By construction, the identity, location and activity patterns of all cells are known. The supplemental
material shows an example with its corresponding inferred neural activity.
Real-world data comes from two-photon microscopy of mouse motor cortex recorded in vivo [7]
which has been motion-corrected. These sequences allow us to conduct qualitative experiments.
ADINA [4] relies on dictionary learning [11] to find both spatial components and their time courses.
Both have many zero coefficients, but are not ?truly sparse? in the sense of this paper. The method
comes with a heuristic post-processing to separate coactivated cells into distinct spatial components.
NMF+ADINA uses non-negative matrix factorization to infer both the spatial and temporal primitives of an image sequence as in [12, 15]. In contrast to [15] which uses a k-means clustering of
highly confident spike vectors to provide a good initialization in the search for spatial components,
we couple NMF with the postprocessing of ADINA.
CSBC+SC combines convolutional sparse block coding [14] based on a single still image (obtained
from the temporal mean or median image, or a maximum intensity projection across time) with
temporal sparse coding.
CSBC+STD combines convolutional sparse block coding [14] based on a single still image (obtained from the temporal mean or median image, or a maximum intensity projection across time)
with the proposed sparse temporal deconvolution in Sect. 3.1.
SSTD is the method described here. We used J = L = 2, K = 200, F = 200 and H = 31, 15 for
the artificial and real data, respectively.
4
4.1
Results
Inferring spike trains
To quantify the accuracy of activity detection, we first threshold the estimated activities and then
compute, by summing over each step in every time series, the number of true and false negatives
and positives. For a fair comparison, the thresholds were adjusted separately for each method to give
optimal accuracy. Sensitivity, precision and accuracy computed from the above implicitly measure
both the quality of the segmentation and the quality of the activity estimation. An additional measure, SPIKE distance [9], emphasizes any temporal deviations between the true and estimated spike
location in a truly sparse representation.
Fig. 3 shows that, unsurprisingly, best results are obtained when methods use the true impulse response rather than learning it from the data. This finding does not carry over to real data, where a
?true? impulse response is typically not known. Given the true impulse response, both FAST and
STD fare better than Peeling, showing that a greedy algorithm is faster but gives somewhat worse
results. Even when learning the impulse response, FAST and STD are no worse than Peeling. When
learning the parameters, FAST has an advantage over STD on this artificial data because FAST already uses the correct parametric form of the impulse response that was used to generate the data
and only needs to learn a single parameter; while STD learns a more general but nonparametric
activity model with many degrees of freedom.
The great spread of all quality measures results from the wide range of noise levels used, and the
overall deficiencies in accuracy attest to the difficulty of these simulated data sets.
4.2
Segmenting Cells and Inferring spike trains
Fig. 4 shows that all the methods from Sect. 3.2 reach respectable and comparable performance in
the task of identifying neural activity from non-trivial synthetic image sequences.
CSBC+SC reaches the highest sensitivity while SSTD has the greatest precision. SSTD apparently
achieves comparable performance to the other methods without the need for a heuristic pre- or
postprocessing. Multiple random initializations lead to similar learned filters (results not shown),
6
FAST (fixed param.)
FAST (learned param.)
Peeling (fixed param.)
STD (fixed param.)
STD (learned param.)
00
20 20 40
60
40
8060
Sensitivity
(%)
Sensitivity
(%)
100
80
1000
20
40
60
Precision (%)
100 0
80
20
40
60
Accuracy (%)
80
100 0
0.1
0.2
SPIKE distance
0.3
0.4
Figure 3: Sensitivity, precision, accuracy (higher is better) and SPIKE distance (lower is better) of
different methods for spike train estimation. Methods that need to learn the activation pattern perform worse than those using the true (but generally unknown) activation pattern and its parameters.
FAST is at an advantage here because it happens to use the very impulse response that was used in
generating the data.
so the optimization problem seems to be well-posed. The price to pay for the elegance of a unified
formulation is a much higher computational cost of this more involved optimization. Again, the
spread of sensitivities, precisions and accuracies results from the range of noise levels used in the
simulations. The plots suggest that SSTD may have fewer ?catastrophic failure? cases, but an even
larger set of sequences will be required to verify this tendency.
ADINA
NNMF+ADINA
CSBC+SC
CSBC+STD
SSTD
50
50
60 60
70
70 80
8090
Sensitivity
(%) (%)
Sensitivity
90
100
100 50
60
70
80
Precision (%)
90
100 50
60
70
80
Accuracy (%)
90
100
Figure 4: Quality of cell detection and and the estimation of their activities. SSTD does as well as
the competing methods that rely on heuristic pre- or post-processing.
Real Sequences: Qualitative results are shown in Fig. 5. SSTD is able to distinguish both cells with
spatial overlap and with high temporal correlation. It compensates large variations in luminance
and contrast, and can discriminate between different types of cells. Exploiting truly sparse but
independent representations in both the spatial and the temporal domain allows to infer plausible
neural activity and, at the same time, reduce the noise in the underlying Calcium image sequence.
5
Discussion
The proposed SSTD combines the decomposition of the data into low-rank components with the
finding of a convolutional sparse representation for each of those components. The formalism allows
exploiting sparseness and the repetitive motifs that are so characteristic of biological data. Users
need to choose the number and size of filters that indirectly determine the number of cell types
found and their activation patterns.
As shown in Fig. 5, the approach gives credible interpretations of raw data in terms of an extremely
sparse and hence parsimonious representation.
The decomposition of a spacetime volume into a Cartesian product of spatial shapes and their time
courses is only possible when cells do not move over time. This assumption holds for in vitro
experiments, and can often be satisfied by good fixation in in vivo experiments, but is not universally
valid. Correcting for motions in a generalized unified framework is an interesting direction for future
work. The experiments in section 4.1 suggest that it may also be worthwhile to investigate the use
of more parametric forms for the impulse response instead of the completely unbiased variant used
here.
7
70
48
64
56
62
2
65
29
135
142
150
88
54
66
69
41
137
18
15
37
144
103
76
75
59 20
38
131
33
3
141
30
63
121
36
96 71
125
61 44 43
106
72
107
74 136
1
58
17
60
42
62
66
68
87
6
55
43
53
16
13
91
3
99
51
0.05
17
39
57
90
0
0
20
40
Frames
60
80
25 36
2
100
0.1
18
64
61
69
32
0.2
0.15
3510
89
27
95
148
94
26
32
79
47
12
100 65
5
81
63
47
0.05
16
31
98
23
76
96
93
0.25
75
38
28
1567
54
21
14
44
55
34
9
134
56
11
4
33
84
79
37
20
34
78
77
92
86
49
88
30
0.1
52
59
5
127
149
46
8
9
91
85
7
11683
14
67
132 23
87 78
6
7
28
40
1
109
112
77
92
84
0.15
12
147
73
80
45
24
0.3
50
71
83
48
39
13
51
81
8
24
146
73
42
152
97
58
29 72
22
filter 1
filter 2
filter 3
19
70
41
90
97
27 45
85
117
49
110
115 111
22
94
11
138
31
119
126
108113 102
40
99
35 100
105
133
50
52
80
60
128
74
104
26
82 46
10 129
4
filter 1
filter 2
145
143
53
151
114
124
101
93
122 140
120
0.2
118
130
86
19
68
25
89
57
123
95
21
139
98
0
82
0
20
40
Frames
60
80
100
1
1
11
21
11
31
21
41
31
61
Cell number
Cell number
51
71
81
91
41
51
61
101
71
111
121
81
131
91
141
100
151
0
50
100
150
Time (s)
200
250
1
0
50
100
0
50
100
150
200
250
150
200
250
Time (s)
1
11
11
21
31
21
41
31
61
Cell number
Cell number
51
71
81
91
41
51
61
101
71
111
121
81
131
91
141
151
152
100
0
50
100
150
Time (s)
200
250
Time (s)
Figure 5: Qualitative results on two real data sets. The data on the left column shows mostly cell
bodies, while the data on the right shows both cell bodies (large) and dendrites (small). For each
data set, the top left shows an average projection of the relative fluorescence change across time with
cell centroids D (black dots) and contours of segmented cells, and the top right shows the learned
impulse responses. In the middle, the fluorescence levels integrated over the segmented cells are
shown in random colors. The bottom shows by means of small disks the location, type and strength
of the impulses that summarize all the data shown in the middle. Together with the cell shapes, the
impulses from part of the ?truly sparse? representation that we propose. When convolving these
spikes with the impulse responses from the top right insets, we obtain the time courses shown in
random colors.
Such advances will further help making Calcium imaging an enabling tool for the neurosciences.
8
References
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[23]
[24]
[25]
D. P. Bertsekas. Nonlinear Programming. Athena Scientific, 1999.
A. Chambolle. An algorithm for total variation minimization and applications, 2004.
F. Diego and F. A. Hamprecht. Learning multi-level sparse representations. In NIPS. 2013.
F. Diego, S. Reichinnek, M. Both, and F. A. Hamprecht. Automated identification of neuronal activity
from calcium imaging by sparse dictionary learning. ISBI 2013. Proceedings, pages 1058?1061, 2013.
B. F. Grewe, D. Langer, H. Kasper, B. M. Kampa, and F. Helmchen. High-speed in vivo calcium imaging
reveals neuronal network activity with near-millisecond precision. Nat Meth, 7(5):399?405, May 2010.
C. Grienberger and A. Konnerth. Neuron, volume 73, chapter Imaging Calcium in Neurons, pages 862?
885. Cell Press,, Mar 2012.
D. Huber, D. A. Gutnisky, S. Peron, D. H. O/?Connor, J. S. Wiegert, L. Tian, T. G. Oertner, L. L. Looger,
and K. Svoboda. Multiple dynamic representations in the motor cortex during sensorimotor learning.
Nature, 484(7395):473?478, Apr 2012.
K. Kavukcuoglu, P. Sermanet, Y. Boureau, K. Gregor, M. Mathieu, and Y. LeCun. Learning convolutional
feature hierachies for visual recognition. In NIPS, 2010.
T. Kreuz, D. Chicharro, C. Houghton, R. G. Andrzejak, and F. Mormann. Monitoring spike train synchrony. Journal of Neurophysiology, 2012.
H. Luetcke, F. Gerhard, F. Zenke, W. Gerstner, and F. Helmchen. Inference of neuronal network spike
dynamics and topology from calcium imaging data. Frontiers in Neural Circuits, 7(201), 2013.
J. Mairal, F. Bach, J. Ponce, and G. Sapiro. Online Learning for Matrix Factorization and Sparse Coding.
Journal of Machine Learning Research, 2010.
R. Maruyama, K. Maeda, H. Moroda, I. Kato, M. Inoue, H. Miyakawa, and T. Aonishi. Detecting cells
using non-negative matrix factorization on calcium imaging data. Neural Networks, 55(0):11 ? 19, 2014.
E. A. Mukamel, A. Nimmerjahn, and M. J. Schnitzer. Automated analysis of cellular signals from largescale calcium imaging data. Neuron, 2009.
M. Pachitariu, A. M. Packer, N. Pettit, H. Dalgleish, M. Hausser, and M. Sahani. Extracting regions of
interest from biological images with convolutional sparse block coding. In NIPS. 2013.
E. A. Pnevmatikakis and L. Paninski. Sparse nonnegative deconvolution for compressive calcium imaging: algorithms and phase transitions. In NIPS. 2013.
S. Reichinnek, A. von Kameke, A. M. Hagenston, E. Freitag, F. C. Roth, H. Bading, M. T. Hasan,
A. Draguhn, and M. Both. Reliable optical detection of coherent neuronal activity in fast oscillating
networks in vitro. NeuroImage, 60(1), 2012.
R. Rigamonti, A. Sironi, V. Lepetit, and P. Fua. Learning separable filters. In Conference on Computer
Vision and Pattern Recognition, 2013.
M. N. Schmidt and M. M?rup. Nonnegative matrix factor 2-D deconvolution for blind single channel
source separation. In ICA, 2006.
P. Smaragdis. Non-negative matrix factor deconvolution; extraction of multiple sound sources from monophonic inputs. In ICA, pages 494?499, 2004.
A. Szlam, K. Kavukcuoglu, and Y. LeCun. Convolutional matching pursuit and dictionary training. Computer Research Repository (arXiv), 2010.
G. W. Taylor, R. Fergus, Y. Lecun, and C. Bregler. Convolutional learning of spatio-temporal features,
2010.
J. Tomek, O. Novak, and J. Syka. Two-photon processor and seneca: a freely available software package
to process data from two-photon calcium imaging at speeds down to several milliseconds per frame. J
Neurophysiol, 110, 2013.
I. Valmianski, A. Y. Shih, J. D. Driscoll, D. W. Matthews, Y. Freund, and D. Kleinfeld. Automatic identification of fluorescently labeled brain cells for rapid functional imaging. Journal of Neurophysiology,
2010.
J. T. Vogelstein, A. M. Packer, T. A. Machado, T. Sippy, B. Babadi, R. Yuste, and L. Paninski. Fast
non-negative deconvolution for spike train inference from population calcium imaging. Journal of Neurophysiology, 2010.
M. Zeiler, D. Krishnan, G. Taylor, and R. Fergus. Deconvolutional networks. In CVPR, 2010.
9
| 5342 |@word neurophysiology:3 repository:1 middle:3 norm:4 seems:1 disk:1 tedious:1 seek:1 simulation:1 decomposition:3 invoking:1 minus:1 shot:1 schnitzer:1 lepetit:1 carry:1 moment:1 series:6 exclusively:1 denoting:1 deconvolutional:1 subjective:1 existing:2 activation:5 must:1 additive:1 shape:8 motor:2 remove:1 plot:1 interpretable:1 alone:1 greedy:2 selected:3 fewer:1 plane:1 short:1 detecting:1 location:9 simpler:1 mathematical:1 along:3 burst:1 direct:1 novak:1 qualitative:3 freitag:1 hci:1 fixation:1 combine:3 acquired:1 huber:1 ica:2 expected:1 indeed:2 rapid:1 multi:2 brain:3 decomposed:2 automatically:2 actual:1 param:5 considering:1 estimating:1 notation:3 matched:2 underlying:1 circuit:1 kind:1 compressive:1 supplemental:1 unified:6 finding:2 dubbed:1 grienberger:1 guarantee:1 temporal:21 sapiro:1 every:3 multidimensional:5 exactly:3 rm:2 uk:1 szlam:1 bertsekas:1 segmenting:2 positive:1 timing:1 local:1 kampa:1 firing:3 becoming:2 black:1 plus:2 initialization:3 kasper:1 limited:1 factorization:6 range:2 tian:1 unique:1 lecun:3 practice:2 block:5 procedure:1 area:1 matching:2 projection:3 pre:3 suggest:3 deconvolve:1 live:1 map:3 center:1 roth:1 elusive:1 go:1 attention:1 exposure:1 duration:1 convex:1 primitive:1 resolution:1 formalized:1 identifying:2 correcting:1 insight:1 stereotypical:1 array:6 importantly:1 population:1 notion:1 variation:4 coordinate:1 analogous:1 limiting:1 diego:4 construction:1 heavily:1 user:4 svoboda:1 programming:1 gerhard:1 us:7 assure:1 recognition:2 located:1 std:9 houghton:1 labeled:1 bottom:2 observed:3 thousand:1 region:1 sect:2 highest:1 mentioned:1 rup:1 ideally:2 dynamic:2 solving:1 segment:3 division:1 heartbeat:1 basis:1 completely:1 neurophysiol:1 joint:2 k0:3 polygon:2 represented:1 chapter:1 looger:1 train:9 distinct:4 fast:11 describe:2 artificial:4 sc:3 heuristic:6 encoded:1 supplementary:1 posed:1 larger:1 plausible:1 cvpr:1 compensates:1 ability:1 final:1 online:1 sequence:16 triggered:1 advantage:2 propose:4 product:5 pulsation:1 kato:1 date:1 description:1 inducing:1 frobenius:1 amounting:1 exploiting:2 electrode:1 p:2 r1:2 produce:2 generating:2 oscillating:1 object:1 help:1 fixing:1 sole:1 received:1 eq:3 implies:1 come:2 quantify:1 concentrate:1 direction:1 correct:2 owing:1 filter:16 kb:1 adina:5 transient:1 nnmf:1 material:2 require:2 pettit:1 biological:2 adjusted:1 frontier:1 bregler:1 hold:1 great:2 matthew:1 dictionary:5 achieves:2 vary:1 generous:1 estimation:6 fluorescence:10 concurrent:1 pnevmatikakis:1 helmchen:2 ferran:2 tool:1 minimization:1 gaussian:3 rather:1 collaboratory:1 hj:4 varying:1 command:1 l0:1 viz:1 ponce:1 tomek:1 rank:1 likelihood:1 mainly:1 hk:1 contrast:3 centroid:2 summarizing:1 detect:1 sense:2 inference:2 dependent:1 motif:1 typically:4 bt:4 integrated:1 germany:1 pixel:4 overall:1 classification:1 extraneous:1 priori:1 animal:1 art:2 spatial:20 summed:1 special:2 field:1 oertner:1 extraction:1 manually:2 biology:1 represents:1 unsupervised:2 k2f:1 future:1 others:1 stimulus:1 few:1 randomly:3 simultaneously:2 packer:2 individual:2 phase:1 occlusion:1 consisting:1 fire:4 attempt:1 freedom:1 detection:6 interest:2 wiegert:1 highly:1 investigate:1 truly:10 hamprecht:4 kt:1 beforehand:1 konnerth:1 partial:1 necessary:1 respective:1 conduct:1 iv:2 taylor:2 initialized:1 desired:1 deconvolved:1 formalism:1 modeling:1 earlier:1 column:1 coeficients:1 respectable:1 cost:1 deviation:1 subset:1 snr:1 hundred:2 uniform:1 varies:1 synthetic:3 combined:1 confident:1 grand:1 sensitivity:8 recur:1 interdisciplinary:1 physic:1 pool:1 together:2 mouse:1 again:1 von:1 unavoidable:1 recorded:1 satisfied:1 choose:1 possibly:1 worse:3 convolving:1 prompting:1 potential:1 photon:4 de:1 nonoverlapping:1 chemistry:1 coding:10 summarized:1 coefficient:5 onset:1 blind:1 performed:1 doing:1 apparently:1 attest:1 dalgleish:1 synchrony:1 vivo:4 contribution:1 degraded:1 convolutional:12 kfl:1 characteristic:2 accuracy:8 yield:2 generalize:1 raw:4 identification:2 kavukcuoglu:2 emphasizes:1 dub:1 none:1 monitoring:2 published:1 tissue:1 processor:1 explain:1 reach:2 manual:1 whenever:1 definition:2 failure:1 sensorimotor:1 involved:1 elegance:1 couple:1 sampled:1 maruyama:1 knowledge:1 color:2 credible:1 segmentation:7 amplitude:2 higher:2 supervised:1 follow:1 response:26 fua:1 formulation:11 delineate:1 chambolle:1 mar:1 just:1 implicit:1 correlation:1 hand:1 sketch:1 favourably:1 nonlinear:1 marker:1 hopefully:1 kleinfeld:1 quality:4 reveal:2 impulse:28 scientific:2 k22:1 normalized:1 unbiased:1 true:7 verify:1 former:1 fluorescently:1 evolution:1 regularization:1 spatially:1 hence:2 laboratory:2 nonzero:3 iteratively:2 white:1 during:2 nuisance:2 m:1 generalized:1 nimmerjahn:1 demonstrate:1 motion:3 fj:1 postprocessing:3 image:23 ranging:1 intermittently:1 functional:1 machado:1 vitro:3 volume:2 interpretation:2 fare:1 elementwise:1 refer:1 mormann:1 connor:1 smoothness:2 automatic:1 fk:1 dot:1 f0:1 cortex:2 closest:1 recent:2 optimizes:1 nonconvex:2 binary:1 accomplished:2 additional:2 somewhat:1 employed:1 freely:2 subtraction:2 determine:1 signal:9 semi:1 ii:4 multiple:3 full:2 needing:1 infer:2 sound:1 vogelstein:1 smooth:1 segmented:2 faster:1 characterized:1 offer:2 bach:1 divided:1 post:3 variant:1 heterogeneous:1 vision:1 poisson:2 arxiv:1 repetitive:1 represent:2 kernel:1 normalization:1 microscopy:3 cell:73 achieved:1 background:15 addition:2 separately:1 median:2 source:2 hasan:1 extra:1 recording:1 subject:1 hz:1 spirit:1 extracting:1 near:1 leverage:1 iii:3 viability:2 automated:5 krishnan:1 fit:1 competing:1 topology:1 reduce:1 six:1 expression:1 trimmed:1 energetic:1 algebraic:1 afford:1 action:5 matlab:1 generally:1 detailed:2 nonparametric:1 generate:1 millisecond:2 estimated:7 extrinsic:1 per:5 neuroscience:1 group:1 shih:1 threshold:2 prevent:2 pj:2 luminance:2 imaging:15 fraction:1 sum:4 langer:1 run:1 package:1 distorted:1 reasonable:2 separation:1 parsimonious:2 comparable:2 fl:3 bound:4 pay:1 distinguish:2 spacetime:1 smaragdis:1 babadi:1 nonnegative:5 activity:31 strength:1 constraint:4 deficiency:1 software:1 speed:2 argument:1 extremely:2 min:2 performing:1 separable:1 optical:2 structured:1 tv:2 according:1 describes:1 across:3 b:4 happens:1 making:1 sided:3 taken:1 remains:1 previously:1 count:1 draguhn:1 flip:1 pursuit:2 operation:1 available:1 pachitariu:1 hierarchical:1 worthwhile:1 indirectly:1 schmidt:1 altogether:1 convolved:1 top:4 clustering:3 remaining:1 zeiler:1 instant:1 giving:1 invokes:1 build:1 gregor:1 move:1 already:3 quantity:1 spike:18 strategy:1 concentration:1 parametric:3 term3:1 evolutionary:1 conceivable:1 gradient:2 distance:3 separate:1 simulated:1 athena:1 collected:1 cellular:1 trivial:1 boldface:1 enforcing:1 driscoll:1 code:1 length:3 kst:1 modeled:1 illustration:2 ratio:1 sermanet:1 difficult:1 unfortunately:1 setup:1 mostly:1 negative:6 calcium:20 collective:1 adjustable:1 unknown:3 perform:1 upper:4 neuron:11 convolution:6 benchmark:1 enabling:1 descent:2 truncated:1 precise:1 frame:6 nonuniform:2 smoothed:1 intensity:4 nmf:2 inferred:1 pair:1 required:1 specified:1 hausser:1 learned:8 coherent:1 nip:4 able:1 pattern:10 maeda:1 breathing:1 sparsity:2 challenge:1 regime:1 summarize:1 saturation:1 max:1 reliable:1 video:1 greatest:1 overlap:2 difficulty:2 rely:1 zenke:1 indicator:3 largescale:1 residual:1 meth:1 inoue:1 temporally:1 mathieu:1 confocal:1 created:1 grewe:1 sahani:1 prior:4 understanding:1 kf:1 relative:2 unsurprisingly:1 freund:1 fully:2 par:1 interesting:1 yuste:1 fluorescent:2 ingredient:1 isbi:1 integrate:1 degree:1 article:1 tiny:1 course:9 summary:1 keeping:2 free:1 allow:3 wide:1 reichinnek:2 absolute:1 sparse:40 andrzejak:1 slice:1 dimension:2 fred:2 world:2 valid:1 autoregressive:1 contour:1 reside:1 kdk:1 transition:1 preprocessing:1 universally:1 uni:1 implicitly:1 active:1 reveals:1 mairal:1 summing:1 assumed:2 spatio:3 fergus:2 search:1 latent:3 sk:5 table:2 channel:1 learn:3 nature:1 robust:1 dendrite:2 heidelberg:4 gerstner:1 gutnisky:1 domain:2 monophonic:1 pk:2 spread:2 apr:1 linearly:1 intracellular:1 rh:1 noise:9 fair:1 ref:2 body:3 neuronal:4 fig:5 precision:7 neuroimage:1 inferring:3 sparsest:2 exponential:4 learns:1 peeling:4 dozen:1 down:1 specific:1 inset:1 showing:1 symbol:2 coactivated:1 decay:6 dk:7 evidence:1 deconvolution:13 false:1 importance:1 iwr:2 mirror:3 mukamel:1 nat:1 cartesian:4 sparseness:1 boureau:1 electrophysiology:1 simply:1 appearance:9 peron:1 paninski:2 visual:1 strand:1 applies:1 khj:1 determines:1 relies:1 extracted:1 hierachies:1 lth:2 identity:1 chicharro:1 price:1 sippy:1 content:1 hard:1 change:1 typical:2 except:1 corrected:2 uniformly:1 rigamonti:1 decouple:1 stereotypic:1 total:2 discriminate:1 sectioning:1 experimental:3 svd:1 succeeds:1 disregard:1 meaningful:1 catastrophic:1 tendency:1 select:2 brevity:1 requisite:1 |
4,797 | 5,343 | Spatio-temporal Representations of Uncertainty in
Spiking Neural Networks
Sophie Deneve
Group for Neural Theory, ENS Paris
Rue d?Ulm, 29, Paris, France
[email protected]
Cristina Savin
IST Austria
Klosterneuburg, A-3400, Austria
[email protected]
Abstract
It has been long argued that, because of inherent ambiguity and noise, the brain
needs to represent uncertainty in the form of probability distributions. The neural encoding of such distributions remains however highly controversial. Here we
present a novel circuit model for representing multidimensional real-valued distributions using a spike based spatio-temporal code. Our model combines the computational advantages of the currently competing models for probabilistic codes
and exhibits realistic neural responses along a variety of classic measures. Furthermore, the model highlights the challenges associated with interpreting neural
activity in relation to behavioral uncertainty and points to alternative populationlevel approaches for the experimental validation of distributed representations.
Core brain computations, such as sensory perception, have been successfully characterized as probabilistic inference, whereby sensory stimuli are interpreted in terms of the objects or features that
gave rise to them [1, 2]. The tenet of this Bayesian framework is the idea that the brain represents uncertainty about the world in the form of probability distributions. While this notion seems
supported by behavioural evidence, the neural underpinnings of probabilistic computation remain
highly debated [1, 2]. Different proposals offer different trade-offs between flexibility, i.e. the class
of distributions they can represent, and speed, i.e. how fast can the uncertainty be read out from the
neural activity. Given these two dimensions, we can divide existing models in two main classes.
The first set, which we will refer to as spatial codes, distributes information about the distribution
across neurons; the activity of different neurons reflects different values of an underlying random
variable (alternatively, it can be viewed as encoding parameters of the underlying distribution [1,
2]). Linear probabilistic population codes (PPCs) are a popular instance of this class, whereby
the log-probability of a random variable can be linearly decoded from the responses of neurons
tuned to different values of that variable [3]. This encoding scheme has the advantage of speed, as
uncertainty can be decoded in a neurally plausible way from the quasi-instantaneous neural activity,
and reproduces aspects of the experimental data. However, these benefits come at the price of
flexibility: the class of distributions that the network can represent needs to be highly restricted,
otherwise the network size scales exponentially with the number of variables [1].
This limitation has lead to a second class of models, which we will refer to as temporal codes.These
use stochastic network dynamics to sample from the target distribution [4, 1]. Existing models
from this class assume that the activity of each neuron encodes a different random variable; the
network explores the state space such that the time spent in any particular state is proportional to its
probability under the distribution [4]. This representation is exact in the limit of infinite samples.
It has several important computational advantages (e.g. easy marginalization, parameter learning,
linear scaling of network size with the number of dimensions) and further accounts for trial-totrial variability in neural responses [1]. These benefits come at the cost of sampling time: a fair
representation of the underlying distribution requires pooling over several samples, i.e. integrating
neural activity over time. Some have argued that this feature makes sampling unfeasibly slow [2].
1
Here we show that it is possible to construct spatio-temporal codes that combine the best of both
worlds. The core idea is that the network activity evolves through recurrent dynamics such that
samples from the posterior distribution can be linearly decoded from the (quasi-)instantaneous neural responses. This distributed representation allows several independent samples to be encoded
simultaneously, thus enabling a fast representation of uncertainty that improves over time. Computationally, our model inherits all the benefits of a sampling-based representation, while overcoming potential shortcomings of classic temporal codes. We explored the general implications of the
new coding scheme for a simple inference problem and found that the network reproduces many
properties of biological neurons, such as tuning, variability, co-variability and their modulation by
uncertainty. Nonetheless, these single or pairwise measures provided limited information about the
underlying distribution represented by the circuit. In the context of our model, these results argue for
using decoding as tool for validating distributed probabilistic codes, an approach which we illustrate
with a simple example.
1
A distributed spatio-temporal representation of uncertainty
The main idea of the representation is simple: we want to approximate a real-valued D-dimensional
distribution P(x) by samples generated by K independent chains implementing Markov Chain
Monte Carlo (MCMC) sampling [5], y(t) = {yk (t)}k=1...K , with yk ? P(x) (Fig. 1). To this aim,
we encode the stochastic trajectory of the chains in a population of N spiking neurons (N > KD),
such that y(t) is linearly decodable from the neural responses. In particular, we adapt a recently
proposed coding scheme for representing time-varying signals [6] and construct stochastic neural
dynamics such that samples from the target distribution can be obtained by a linear mapping of the
spikes convolved with an epsp-like exponential kernel (Fig. 1a):
y
?(t) = ? ? r(t)
(1)
? (t) denotes the decoded state of the K MCMC chains at time t (of size D ? K), ? is the
where y
decoding matrix1 and r is the low-pass version of the spikes o, ?V r?i = ?ri + oi .
To facilitate the presentation of the model, we start by constructing recurrent dynamics for sampling
a single MCMC chain, which we then generalise to the multi-chain scenario. Based on these network
dynamics, we implement probabilistic inference in a linear Gaussian mixture, which we use in
Section 2 to investigate the neural implications of the code.
Distributed MCMC sampling
As a starting point, consider the computational task of representing an arbitrary temporal trajectory
(the gray line in Fig. 1b) as the linear combination of the responses of a set of neurons (one can think
of this as an analog-to-digital conversion of sorts). If the decoding weights of each neuron points in
a different direction (colour coded), then the trajectory could be efficiently reconstructed by adding
the proper weight vectors (the local derivative of the trajectory) at just the right moment. Indeed,
recent work has shown how to construct network dynamics enabling the network to track a trajectory
as closely as possible [6]. To achieve this, neurons use a greedy strategy: each neuron monitors
the current prediction error (the difference between the trajectory and its linear decoding from the
spikes) and spikes only when its weight vector points in the right direction. When the decoding
weights of several neurons point the same way (as in Fig. 1a), they compete to represent the signal
via recurrent inhibition:2 from the perspective of the decoder, it does not matter which of these
neurons spikes next, so the actual population responses depend on the previous spike history, initial
conditions and intrinsic neural noise.3 As a result, spikes are highly irregular and look ?random?
(with Poisson-like statistics), even when representing a constant signal. While competition is an
important driving force for the network, neurons can also act cooperatively ? when the change in the
signal is larger than the contribution of a single decoding vector, then several neurons need to spike
together to represent the signal (e.g. response to the step in Fig. 1a).
1
The decoding matrix can be arbitrary.
This competition makes spike correlations extremely weak in general [7].
3
When N D there is a strong degeneracy in the map between neural responses and the signal, such that
several different spike sequences yield the same decoded signal. In absence of internal noise, the encoding is
nonetheless deterministic despite apparent variability.
2
2
Figure 1: Overview of the model. a. We assume a linear decoder, where the estimated signal y?
is obtained as a weighted sum of neural responses (exponential kernel, blue). b. When the signal
is multidimensional, different neurons are responsible for encoding different directions along the
target trajectory (gray). c. Alternative network architectures: in the externally-driven version the
target trajectory is given as an external input, whereas in the self-generated case it is computed via
slow recurrent connections (green arrow); the input s is used during inference, when sampling from
P(x|s). d. Encoding an example MCMC trajectory in the externally-driven mode. Light colours
show ground truth; dark colours the decoded signal. e. Single-chain samples from a multivariate
distribution (shown as colormap) decoded from a spiking network; trajectory subsampled by a factor
of 10 for visibility. e. Decoded samples using 5 chains (colors) and a fifth of the time in e.
2
Formally, the network dynamics minimise the squared reconstruction error, (y ? y
?) , under certain
constraints on mean firing rate which ensure the representation is distributed (see Suppl. Info.).
The resulting network consists of spiking neurons with simple leaky-integrate-and-fire dynamics,
? denotes the temporal derivative of V, the binary vector o denotes
? = ? 1 V ? Wo + I, where V
V
?v
the spikes, oi (t) = ? iff Vi (t) > ?i , ?v is the membrane time constant (same as that of the decoder),
P
the neural threshold is ?i = j ?2ij + ? and the recurrent connections, W = ?T ? + ? ? I, can
be learned by STDP [8], where ? is a free parameter controlling neural sparseness. The membrane
potential of each neuron tracks the component of the reconstruction error along the direction of its
decoding weights. As a consequence, the network is balanced (because the dynamics aim to bring
the reconstruction error to zero) and membrane potentials are correlated, particularly in pairs of
neurons with similar decoding weights [7] (see Fig. 2c).
In the traditional form, which we refer to as the ?externally-driven? network (Fig. 1c), information
? In
about the target trajectory is provided as an external input to the neurons: I = ?T ? (1/?v y + y).
our particular case, this input implements a particular kind of MCMC sampling (Langevin). Briefly,
the sampler involves stochastic dynamics driven by the gradient of log P (y), with additive Gaussian
noise [5] (see Suppl.Info. for implementation details). Hence, the external input is stochastic I =
?T ? (1/?v y + F (y) + ), where F (y) = ? log P(y), and is D-dimensional white independent
Gaussian noise. Using our network dynamics, we can encode the MCMC trajectory with high
precision (Fig. 1d). Importantly, because of the distributed representation, the integration window
of the decoder does not restrict the frequency content of the signal. The network can represent
signals that change faster than the membrane time constant (Fig. 1a, d).
To construct a viable biological implementation of this network, we need to embed the sampling
dynamics within the circuit (?self-generated? architecture in Fig. 1c). We achieved this by approximating the current I using the decoded signal y
? instead of y. This results in a second recurrent input
? + F (?
to the neurons, ?I = ?T ? (1/?v y
y) + ). While this is an approximation, we found it does not
affect sampling quality in the parameter regime when the encoding scheme itself works well (see
example dynamics in Fig. 1e).
3
Such dynamics can be derived for any distribution from the broad class of product-of-(exponentialfamily) experts [9], with no restrictions on D; for simplicity and to ease visualisation, here we focus
on the multivariate Gaussian case and restrict the simulations to bivariate distributions (D = 2). For
a Gaussian distribution with mean ? and covariance ?, the resulting membrane potential dynamics
are linear:4
?V
1
= ? V ? Wfast o + Wslow r + D + ?T
(2)
?t
?v
where o denotes the spikes, r is a low-passed version of the spikes. The connections Wfast
correspond to the recurrent
dynamics derived above, while the slow5 connections, Wslow =
T
?1
1
1
?T ??1 ? correspond to
I??
? (e.g. NMDA currents) and the drift term D = ?slow
?slow ? ?
the deterministic component of the MCMC dynamics6 and is white independent Gaussian noise
(implemented for instance by a small chaotic subnetwork appropriately connected to the principal
neurons). In summary, relatively simple leaky integrate-and-fire neurons with appropriate recurrent
connectivity are sufficient for implementing Langevin sampling from a Gaussian distribution in a
distributed code. More complex distributions will likely involve nonlinearities in the slow connections (possibly computed in the dendrites) [10].
Multi-chain encoding: instantaneous representation of uncertainty
The earliest proposal for sampling-based neural representations of uncertainty suggested distributing
samples either across neurons or across time [4]. Nonetheless, all realisations of neural sampling use
the second solution. The reason is simple: when equating the activity of individual neurons (either
voltage or firing rate) to individual random variables, it is relatively straightforward to construct neural dynamics implementing MCMC sampling. It is less clear what kind of neural dynamics would
generate samples in several neurons at a time. One naive solution would be to construct several networks that each sample from the same distribution in parallel. This however seems to unavoidably
entail a ?copy-pasting? of all recurrent connections across different circuits, which is biologically
unrealistic. Our distributed representation, in which neurons jointly encode the sampling trajectory,
provides a potential solution to this problem. In particular, it allows several chains to be embedded
in a single network.
To extend the dynamics to a multi-chain scenario, we imagine an auxiliary probability distribution
over K random variables. We want each to correspond to one chain, so we take them to be independent and identically distributed according to P(x). Since the sampling dynamics derived above do
not restrict the dimensionality of the underlying distribution, we can use them to sample from this
D ? K-dimensional distribution instead. For the example
of a multivariate normal, for instance, we
would now sample from another Gaussian, P x?K , with mean ??K (K repetitions of ?) and covariance ??K , a block-diagonal matrix, obtained by K repetitions of ?. In general, the multi-chain
trajectory can be viewed as just another instance of MCMC sampling, where the encoding scheme
guarantees that the signals across different chains remain independent. What may change, however,
is the interpretability of neural responses in relation to the underlying encoded variable. We show
that under mild assumptions on the decoding matrix ?, the main features of single and pairwise
responses are preserved (see below and Suppl.Info. Sec.4).
Fig. 1f shows an example run for multi-chain sampling from a bivariate Gaussian. In a fifth of the
time used in the single-chain scenario (Fig. 1e), the network dynamics achieves a similar spread
across the state space, allowing for a quick estimation of uncertainty (see also Suppl.Info. 2). For a
certain precision of encoding (determined by the size of the decoding weights ?) and neural sparseness level, N scales linearly with the dimensionality of the state space D and the number of simultaneously encoded chains K. Thus, our representation provides a convenient trade-off between the
network size and the speed of the underlying computation. When N is fixed, faster sampling requires either a penalty on precision, or increased firing rates (N D). Overall, the coding scheme
allows for a linear trade-off between speed and resources (either neurons or spikes).
Since F (x) = ??1 (x ? ?), this results in a stochastic generalisation of the dynamics in [7].
?Slow? marks the fact that the term depends on the low-passed neural output r, rather than o.
6
Learning the connections goes beyond the scope of this paper; it seems parameter learning can be achieved
using the plasticity rules derived for the temporal code, if these are local (not shown).
4
5
4
2
Neural implications
To investigate the experimental implications of our coding scheme, we assumed the posterior distribution is centred around a stimulus-specific mean (a set of S = 12 values, equidistantly distributed
on a circle of radius 1 around the origin, see black dots in Fig. 3a), with a stimulus independent
covariance parametrizing the uncertainty about x. This kind of posterior arises e.g. as a result of
inference in a linear Gaussian mixture (since the focus here is not on a specific probabilistic model
of the circuit function, we keep the computation very basic, see Suppl. Info. for details). It allows us
quantify the general properties of distributed sampling in terms of classic measures (tuning curves,
Fano factors, FF, cross-correlogram, CCG, and spike count correlations, rsc ) and how these change
with uncertainty.
Since we found that, under mild assumptions for the decoding matrix ?, the results are qualitatively
similar in a single vs. a multi-chain scenario (see Suppl. Info.), and to facilitate the explanation, the
results reported in the main text used K = 1.
Figure 2: Our model recapitulates several known features of cortical responses. a. Mean firing rates
as a function of stimulus, for all neurons (N = 37); color reflects the phase of ?i (right). b. The
network is in an asynchronous state. Left: example spike raster. Right: Fano factor distribution. c.
Within-trial correlations in membrane potential for pairs of neurons as a function of the similarity
of their decoding weights. d. Spike count correlations (averaged across stimuli) as a function of
the neurons? tuning similarity. Right: distribution of rsc , with mean in magenta. e We use crosscorrelograms (CCG) to asses spike synchrony. Left: CCG for an example neuron. Middle: Area
under the peak ?10ms (between the dashed vertical bars) for all neuron pairs for 3 example stimuli;
neurons ordered by ?i phase. Right: the area under CCG peak as a function of tuning similarity.
a. The neural dynamics are consistent with a wide range of experimental observations
First, we measured the mean firing rate of the neurons for each stimulus (averaged across 50 trials, each 1s long). We found that individual neurons show selectivity to stimulus orientations, with
bell-shaped tuning curves, reminiscent of e.g. the orientation-tuning of V1 neurons (Fig. 2a). The inhomogeneity in the scale of the responses across the population is a reflection of the inhomogeneities
in the decoding matrix ?.7
7
The phase of the decoding weights was sampled uniformly around the circle, with an amplitude drawn
uniformly from the interval [0.005; 0.025].
5
Neural responses were asynchronous, with irregular firing (Fig. 2b), consistent with experimental
observations [11, 12]. To quantify neural variability, we estimated the Fano factors, measured as the
ratio between the variance and the mean of the spike counts in different trials, F Fi = ?f2i /?fi . We
found that the Fano factor distribution was centered around 1, a signature of Poisson variability. This
observation suggests that the sampling dynamics preserve the main features of the distributed code
described in Ref. [6]. Unlike the basic model, however, here neural variability arises both because
of indeterminacies, due to distributed coding, and because of ?true? stochasticity, owed to sampling.
The contribution of the latter, which is characteristic of our version, will depend on the underlying
distribution represented: when the distribution is highly peaked, the deterministic component of the
MCMC dynamics dominates, while the noise plays an increasingly important role the broader the
distribution.
At the level of the membrane potential, both sources of variability introduce correlations between
neurons with similar tuning (Fig. 2c), as seen experimentally [13]: the first because the reconstruction error acts as a shared latent cause, the second because the stochastic component ?which was
independent in the y space? is mapped through ?T in a distributed representation (see Eq. 2). While
the membrane correlations introduced by the first disappear at the level of the spikes [7], the addition
of the stochastic component turns out to have important consequences for the spike correlations both
on the fast time scale, measured by CCG, and for the across-trial spike count covariability, measured
by the noise correlations, rsc .
Fig. 2e shows the CCG of an example pair of neurons, with similar tuning; their activity synchronizes on the time scale of few milliseconds. In more detail, our CCG measure was normalised by
first computing the raw cross-correlogram (averaged across trials) and then subtracting a baseline
obtained as the CCG of shuffled data, where the responses of each neuron come from a different
trial. The raw cross-correlogram for a time delay, ? , CCG(? ) was computed as the Pearsons correlation of the neural responses, shifted in time time by ? .8 At the level of the population, the amount
of synchrony (measured as the area under the CCG peak ?10ms) was strongly modulated by the
input (Fig. 2e, middle), with synchrony most prominent in pairs of neurons that aligned with the
stimulus (not shown). This is consistent with the idea that synchrony is stimulus-specific [14, 15].
We also measured spike count correlation (the Pearsons correlation coefficient of spike counts
recorded in different trials for the same stimulus) and found they depend on the selectivity of the
neurons, with positive correlations for pairs of neurons with similar tuning (Fig. 2d), as seen in experiments [16]. The overall distribution was broad, with a small positive mean (Fig. 2d), as in recent
reports [11, 12]. Taken together, these results suggest that our model qualitatively recapitulates the
basic features of cortical neural responses.
b. Uncertainty modulates neural variability and covariability
We have seen that sampling introduces spike correlations, not seen when encoding a deterministic
dynamical system [7]. Since stochasticity seems to be key for these effects, this suggests uncertainty should significantly modulate pairwise correlations. To confirm this prediction, we varied the
covariance structure of the underlying distribution for the same circuit (Fig. 3a; the low variance condition corresponds to baseline measures reported above) and repeated all previous measurements.
We found that changes in uncertainty leave neuronal tuning invariant (Fig. 3b, not surprisingly since
the mean firing rates reflect the posterior mean). Nonetheless, increasing uncertainty had significant
effects on neural variability and co-variability.
Fano factors increased for broader distributions (Fig. 3b), congruent with the common observation
of the stimulus quenching response variability in experiments [17]. Second, we found a slower
component in the CCG, which increased with uncertainty (Fig. 3e), as in the data [15]. Lastly, the
dependence of different spike correlation measures on neural co-tuning increased with uncertainty
(Fig. 3c, d). In particular, neurons with similar stimulus preferences increased their synchrony
and spike-count correlations with increasing uncertainty, consistent with the stimulus quenching
response co-variability in neural data and increases in correlations at low contrast [17, 16].
Although we see a significant modulation of (co-)variability with changes in uncentainty, these measures provide limited information about the underlying distribution represented in the network. They
can be used to detect changes in the overall spread of the distribution, i.e. the high vs. low-variance
8
While this is not the most common expression for the CCG; we found it reliably detects synchronous firing
across neurons; spikes discretised in 2ms bins.
6
Figure 3: The effects of uncertainty on neural responses. a. Overview of different experimental conditions, posterior mean centred on different stimuli (black dots) with stimulus independent
covariance shown for 4 conditions. b. Left: Tuning curves for an example neuron, for different conditions. Right: firing rate in the low variance vs. all other conditions, summary across all neurons;
dots correspond to different neuron-stimulus pairs. c. Fano factor distribution for high-variance
condition (compare Fig.2b). d. Area under CCG peak ?10ms as a function of the tuning similarity
of the neurons, for different uncertainty conditions (colours as in b). e. Complete CCG, averaged
across 10 neurons with similar tuning while sampling from independent bivariate Gaussians with
different s.d. (0.1 for ?high variance?). f. Spike count correlations (averaged across stimuli) as a
function of the tuning similarity of the neurons, for different uncertainty conditions.
condition look different at the level of pairwise neural responses. However, they cannot discriminate
between distributions with similar spread, but very different dependency structure, e.g. between the
correlated and anti-correlated condition (Fig. 3d, f; also true for FF and the slow component of the
CCG, not shown). For this, we need to look at the population level.
a experimental setup
b
same condition (lowVar)
c
across condition(highVar)
across condition (Corr)
stimuli
neuron
S stimuli
(repeated trials)
1
2
3
4
5
estimate
true trajectory
estimate
estimate
Figure 4: A decoding approach to study the encoding of uncertainty. a. In a low-variability condition
we record neural responses for several repetitions of different stimuli (black dots); We estimated the
decoding matrix by linear regression and used it to project the activity of the population in individual
trials. b. The decoder captures well the underlying dynamics in a trial; ground-truth in black. c.
? can be used to visualise the structure of the underlying distribution in other
The same decoder ?
conditions. Note the method is robust to a misalignment in initial conditions (red trace).
c. Decoding can be used to assess neural representations of uncertainty
Since in a distributed representation single-neuron or pairwise measures tell us little about the dependency structure of the represented random variables, alternative methods need to be devised for
investigating the underlying computation performed by the circuit. The representational framework
proposed here suggests that linear decoding may be used for this purpose. In particular, we can
record neural responses for a variety of stimuli and reverse-engineer the map between spikes and
the relevant latent variables (or, if the assumed generative model is linear as here, the stimuli themselves). We can use the low-variance condition to get a reasonable estimate of the decoding matrix,
? (since the underlying sampling dynamics are close to the posterior mean) and then use the de?
coder for visualising the trajectory of the network while varying uncertainty. As an illustration, we
7
use simple linear regression of the stimuli s as a function of the neuron firing rates, scaled by ?v .9
Although the recovered decoding weights are imperfect and the initial conditions unknown, the pro? captures the main features of the underlying
jections of the neural responses in single trials along ?
sampler, both in the low-variance and in other conditions (Fig. 4b, c).
3
Discussion
How populations of neurons encode probability distributions in a central question for Bayesian approaches to understanding neural computation. While previous work has shown that spiking neural
networks could represent a probability over single real-valued variables [18], or the joint probability
of many binary random variables [19], the representation of complex multi-dimensional real-valued
distributions10 remains less clear [1, 2]. Here we have proposed a new spatio-temporal code for
representing such distributions quickly and flexibly. Our model relies on network dynamics which
approximate the target distribution by several MCMC chains, encoded in the spiking neural activity
such that the samples can be linearly decoded from the quasi-instantaneous neural responses. Unlike previous sampling-based codes [19], our model does not require a one-to-one correspondence
between random variables and neurons. This separation between computation and representation is
critical for the increased speed, as it allows multiple chains to be realistically embedded in the same
circuit, while preserving all the computational benefits of sampling. Furthermore, it makes the encoding robust to neural damage, which seems important when representing behaviourally-relevant
variables, e.g. in higher cortical areas. These benefits come at the cost of a linear increase in the
number of neurons with K, providing a convenient trade-off between speed and neural resources.
The speedup due to increases in network size is orthogonal to potential improvements in sampling
efficiency achieved by more sophisticated MCMC dynamics, e.g. relying on oscillations [21] or nonnormal stochastic dynamics [22], suggesting that distributed sampling could be made even faster by
combining the two approaches.
The distributed coding scheme has important consequences for interpreting neural responses: since
knowledge about the underlying distribution is spread across the population, the activity of single
cells does not reflect the underlying computation in any obvious way. In particular, although the
network did reproduce various properties of single neuron and pairs of neuron responses seen experimentally, we found that their modulation with uncertainty provides relatively limited information
about the underlying probabilistic computation. Changes in the overall spread (entropy) of the posterior are reflected in changes in variability (Fano factors) and covariability (synchrony on the ms
timescale and spike-count correlations across trials) of neural responses across the population, as
seen in the data. Since these features arise due to the interaction between sampling and distributed
coding, the model further predicts that the degree of correlations between a pair of neurons should
depend on their functional similarity, and that the degree of this modulation should be affected by
uncertainty. Nonetheless, the distributed representation occludes the structure of the underlying
distribution (e.g. correlations between random variables), something which would have been immediately apparent in a one-to-one sampling code.
Our results reinforce the idea that population, rather than single-cell, responses are key to understanding cortical computation, and points to linear decoding as a potential analysis tool for investigating probabilistic computation in a distributed code. In particular, we have shown that we can
train a linear decoder on spiking data and use it to reveal the underlying sampling dynamics in different conditions. While ours is a simple toy example, where we assume that we can record from
all the neurons in the population, the fact that the signal is low-dimensional relative to the number
of neurons gives hope that it should be possible to adapt more sophisticated machine learning techniques [23] for decoding the underlying trajectory traced by a neural circuit in realistic settings. If
this could be done reliability on data, then the analysis of probabilistic neural computation would
no longer be restricted to regions for which we have good ideas about the mathematical form of the
underlying distribution, but could be applied to any cortical circuit of interest.11 Thus, our coding
scheme opens exciting avenues for multiunit data analysis.
9
This requires knowledge of ?v and, in a multi-chain scenario, a grouping of neural responses by chain
preference. Proxies for which neurons should be decoded together are discussed in Suppl.Info. Sec.4.
10
Such distribution arise in many models of probabilistic inference in the brain, e.g. [20].
11
The critical requirement is to know (some of) the variables represented in the circuit, up to a linear map.
8
References
[1] Fiser, J., Berkes, P., Orb?an, G. & Lengyel, M. Statistically optimal perception and learning:
from behavior to neural representations. Trends in Cognitive Sciences 14, 119?130 (2010).
[2] Pouget, A., Beck, J.M., Ma, W.J. & Latham, P.E. Probabilistic brains: knowns and unknowns.
Nature Neuroscience 16, 1170?1178 (2013).
[3] Pouget, A., Zhang, K., Deneve, S. & Latham, P.E. Statistically efficient estimation using
population coding. Neural computation 10, 373?401 (1998).
[4] Hoyer, P.O. & Hyvarinen, A. Interpreting neural response variability as Monte Carlo sampling
of the posterior. Advances in neural information processing systems, 293?300 (2003).
[5] Neal, R. MCMC using Hamiltonian dynamics. Handbook of Markov Chain Monte Carlo 54,
113?162 (2010).
[6] Boerlin, M. & Deneve, S. Spike-based population coding and working memory. PLoS Computational Biology 7, e1001080 (2011).
[7] Boerlin, M., Machens, C.K. & Den`eve, S. Predictive coding of dynamical variables in balanced
spiking networks. PLoS Computational Biology (2013).
[8] Bourdoukan, R., Barrett, D., Machens, C. & Deneve, S. Learning optimal spike-based representations. Advances in neural information processing systems, 2294?2302 (2012).
[9] Hinton, G.E. Training products of experts by minimizing contrastive divergence. Neural
computation 14, 1771?1800 (2002).
[10] Savin, C., Dayan, P. & Lengyel, M. Correlations strike back (again): the case of associative
memory retrieval. in Advances in Neural Information Processing Systems 26 (eds. Burges, C.,
Bottou, L., Welling, M., Ghahramani, Z. & Weinberger, K.) 288?296 (2013).
[11] Renart, A. et al. The asynchronous state in cortical circuits. Science 327, 587?590 (2010).
[12] Ecker, A.S. et al. Decorrelated neuronal firing in cortical microcircuits. Science 327, 584?587
(2010).
[13] Yu, J. & Ferster, D. Functional coupling from simple to complex cells in the visually driven
cortical circuit. Journal of Neuroscience 33, 18855?18866 (2013).
[14] Ohiorhenuan, I.E. et al. Sparse coding and high-order correlations in fine-scale cortical networks. Nature 466, 617?621 (2010).
[15] Kohn, A. & Smith, M.A. Stimulus dependence of neuronal correlation in primary visual cortex
of the macaque. Journal of Neuroscience 25, 3661?3673 (2005).
[16] Smith, M.A. & Kohn, A. Spatial and temporal scales of neuronal correlation in primary visual
cortex. Journal of Neuroscience 28, 12591?12603 (2008).
[17] Churchland, M.M. et al. Stimulus onset quenches neural variability: a widespread cortical
phenomenon. Nature Neuroscience 13, 369?378 (2010).
[18] Zemel, R.S., Dayan, P. & Pouget, A. Probabilistic interpretation of population codes. Neural
computation 10, 403?430 (1998).
[19] Buesing, L., Bill, J., Nessler, B. & Maass, W. Neural dynamics as sampling: A model for
stochastic computation in recurrent networks of spiking neurons. PLoS Computational Biology
7, e1002211 (2011).
[20] Karklin, Y. & Lewicki, M. A hierarchical bayesian model for learning nonlinear statistical
regularities in nonstationary natural signals. Neural computation 17, 397?423 (2005).
[21] Savin, C., Dayan, P. & Lengyel, M. Optimal recall from bounded metaplastic synapses: predicting functional adaptations in hippocampal area CA3. PLoS Computational Biology 10,
e1003489 (2014).
[22] Hennequin, G., Aitchison, L. & Lengyel, M. Fast sampling for Bayesian inference in neural
circuits. arXiv preprint arXiv:1404.3521 (2014).
[23] Macke, J.H. et al. Empirical models of spiking in neural populations. Advances in neural
information processing systems 24, 1350?1358 (2011).
9
| 5343 |@word mild:2 trial:13 version:4 middle:2 briefly:1 seems:5 open:1 simulation:1 covariance:5 contrastive:1 moment:1 initial:3 cristina:1 tuned:1 ours:1 existing:2 current:3 recovered:1 reminiscent:1 tenet:1 realistic:2 additive:1 plasticity:1 occludes:1 visibility:1 v:3 greedy:1 generative:1 hamiltonian:1 core:2 smith:2 record:3 matrix1:1 provides:3 preference:2 zhang:1 mathematical:1 along:4 viable:1 consists:1 combine:2 behavioral:1 introduce:1 pairwise:5 indeed:1 behavior:1 themselves:1 multi:8 brain:5 detects:1 relying:1 actual:1 little:1 window:1 increasing:2 provided:2 project:1 underlying:22 bounded:1 circuit:14 nonnormal:1 coder:1 what:2 kind:3 interpreted:1 pasting:1 guarantee:1 temporal:11 multidimensional:2 act:2 colormap:1 scaled:1 positive:2 local:2 limit:1 consequence:3 despite:1 encoding:13 firing:11 modulation:4 black:4 equating:1 suggests:3 co:5 ease:1 limited:3 range:1 statistically:2 averaged:5 responsible:1 block:1 implement:2 chaotic:1 area:6 empirical:1 bell:1 significantly:1 convenient:2 integrating:1 ppcs:1 suggest:1 get:1 cannot:1 close:1 context:1 nessler:1 restriction:1 bill:1 map:3 deterministic:4 quick:1 ecker:1 straightforward:1 go:1 starting:1 flexibly:1 simplicity:1 immediately:1 pouget:3 rule:1 importantly:1 hennequin:1 bourdoukan:1 classic:3 population:16 notion:1 target:6 controlling:1 imagine:1 ulm:1 exact:1 play:1 machens:2 origin:1 trend:1 particularly:1 predicts:1 role:1 preprint:1 visualising:1 capture:2 region:1 connected:1 plo:4 trade:4 yk:2 balanced:2 dynamic:33 signature:1 depend:4 churchland:1 predictive:1 exponentialfamily:1 efficiency:1 misalignment:1 joint:1 represented:5 various:1 train:1 fast:4 shortcoming:1 unfeasibly:1 monte:3 zemel:1 tell:1 apparent:2 encoded:4 larger:1 valued:4 plausible:1 otherwise:1 statistic:1 think:1 jointly:1 itself:1 inhomogeneity:2 timescale:1 associative:1 advantage:3 sequence:1 reconstruction:4 subtracting:1 interaction:1 product:2 epsp:1 fr:1 adaptation:1 aligned:1 relevant:2 unavoidably:1 combining:1 iff:1 flexibility:2 achieve:1 representational:1 realistically:1 competition:2 regularity:1 requirement:1 congruent:1 klosterneuburg:1 leave:1 object:1 spent:1 illustrate:1 recurrent:10 ac:1 coupling:1 measured:6 ij:1 indeterminacy:1 eq:1 strong:1 implemented:1 auxiliary:1 involves:1 come:4 quantify:2 orb:1 direction:4 radius:1 closely:1 stochastic:10 centered:1 implementing:3 bin:1 argued:2 require:1 biological:2 cooperatively:1 around:4 ground:2 stdp:1 normal:1 visually:1 mapping:1 scope:1 quenching:2 driving:1 achieves:1 boerlin:2 purpose:1 estimation:2 currently:1 repetition:3 successfully:1 tool:2 reflects:2 weighted:1 hope:1 offs:1 gaussian:10 aim:2 rather:2 varying:2 voltage:1 broader:2 covariability:3 earliest:1 encode:4 derived:4 inherits:1 focus:2 improvement:1 contrast:1 baseline:2 detect:1 inference:7 dayan:3 relation:2 visualisation:1 reproduce:1 quasi:3 france:1 csavin:1 overall:4 orientation:2 spatial:2 integration:1 construct:6 shaped:1 sampling:35 biology:4 represents:1 broad:2 look:3 yu:1 peaked:1 report:1 stimulus:26 inherent:1 realisation:1 few:1 ohiorhenuan:1 decodable:1 simultaneously:2 preserve:1 divergence:1 individual:4 subsampled:1 beck:1 phase:3 metaplastic:1 fire:2 interest:1 highly:5 investigate:2 introduces:1 mixture:2 light:1 chain:23 implication:4 underpinnings:1 orthogonal:1 divide:1 circle:2 rsc:3 instance:4 increased:6 cost:2 ca3:1 delay:1 reported:2 dependency:2 explores:1 peak:4 probabilistic:13 off:3 decoding:23 together:3 quickly:1 connectivity:1 again:1 squared:1 ambiguity:1 recorded:1 reflect:2 central:1 possibly:1 external:3 cognitive:1 expert:2 derivative:2 macke:1 toy:1 account:1 potential:9 nonlinearities:1 de:1 suggesting:1 centred:2 coding:12 sec:2 coefficient:1 matter:1 vi:1 depends:1 onset:1 performed:1 red:1 start:1 sort:1 parallel:1 synchrony:6 equidistantly:1 contribution:2 ass:2 oi:2 variance:8 characteristic:1 efficiently:1 yield:1 correspond:4 weak:1 bayesian:4 raw:2 buesing:1 carlo:3 trajectory:17 lengyel:4 history:1 synapsis:1 decorrelated:1 ed:1 raster:1 nonetheless:5 frequency:1 obvious:1 associated:1 degeneracy:1 sampled:1 popular:1 austria:2 color:2 knowledge:2 improves:1 dimensionality:2 recall:1 nmda:1 amplitude:1 sophisticated:2 back:1 higher:1 reflected:1 response:32 done:1 microcircuit:1 strongly:1 furthermore:2 just:2 lastly:1 fiser:1 correlation:25 synchronizes:1 working:1 e1001080:1 nonlinear:1 widespread:1 mode:1 quality:1 gray:2 reveal:1 facilitate:2 effect:3 true:3 hence:1 shuffled:1 read:1 maass:1 neal:1 white:2 during:1 self:2 whereby:2 m:5 prominent:1 hippocampal:1 complete:1 latham:2 interpreting:3 bring:1 reflection:1 pro:1 instantaneous:4 novel:1 recently:1 fi:2 common:2 functional:3 spiking:10 overview:2 exponentially:1 visualise:1 analog:1 extend:1 discussed:1 interpretation:1 refer:3 measurement:1 significant:2 tuning:15 knowns:1 fano:7 stochasticity:2 had:1 dot:4 reliability:1 entail:1 similarity:6 longer:1 inhibition:1 cortex:2 berkes:1 something:1 posterior:8 multivariate:3 recent:2 perspective:1 driven:5 reverse:1 scenario:5 selectivity:2 certain:2 binary:2 seen:6 preserving:1 highvar:1 strike:1 signal:16 dashed:1 neurally:1 multiple:1 faster:3 characterized:1 adapt:2 offer:1 long:2 cross:3 retrieval:1 devised:1 coded:1 prediction:2 basic:3 regression:2 renart:1 poisson:2 arxiv:2 represent:7 kernel:2 suppl:7 achieved:3 cell:3 irregular:2 proposal:2 whereas:1 want:2 preserved:1 addition:1 interval:1 fine:1 source:1 appropriately:1 unlike:2 pooling:1 validating:1 nonstationary:1 eve:1 easy:1 identically:1 variety:2 marginalization:1 affect:1 gave:1 architecture:2 competing:1 restrict:3 imperfect:1 idea:6 avenue:1 minimise:1 synchronous:1 expression:1 kohn:2 colour:4 passed:2 distributing:1 wo:1 penalty:1 cause:1 clear:2 involve:1 recapitulates:2 amount:1 dark:1 generate:1 millisecond:1 shifted:1 estimated:3 neuroscience:5 track:2 blue:1 aitchison:1 affected:1 group:1 ist:2 key:2 threshold:1 monitor:1 drawn:1 traced:1 v1:1 deneve:5 sum:1 compete:1 run:1 uncertainty:29 reasonable:1 separation:1 oscillation:1 scaling:1 correspondence:1 activity:12 constraint:1 ri:1 encodes:1 aspect:1 speed:6 extremely:1 savin:3 relatively:3 speedup:1 according:1 combination:1 kd:1 membrane:8 remain:2 across:20 increasingly:1 evolves:1 biologically:1 den:1 restricted:2 invariant:1 taken:1 computationally:1 behavioural:1 resource:2 remains:2 turn:1 count:9 know:1 gaussians:1 hierarchical:1 appropriate:1 alternative:3 weinberger:1 slower:1 convolved:1 denotes:4 ensure:1 ghahramani:1 approximating:1 disappear:1 question:1 spike:34 strategy:1 damage:1 dependence:2 primary:2 traditional:1 diagonal:1 exhibit:1 gradient:1 subnetwork:1 hoyer:1 mapped:1 reinforce:1 decoder:7 f2i:1 argue:1 reason:1 code:17 illustration:1 ratio:1 providing:1 minimizing:1 setup:1 multiunit:1 info:7 trace:1 rise:1 implementation:2 reliably:1 proper:1 unknown:2 allowing:1 conversion:1 vertical:1 neuron:63 observation:4 markov:2 enabling:2 parametrizing:1 anti:1 langevin:2 hinton:1 variability:18 varied:1 arbitrary:2 drift:1 overcoming:1 introduced:1 pair:9 paris:2 discretised:1 connection:7 learned:1 macaque:1 beyond:1 suggested:1 bar:1 below:1 perception:2 dynamical:2 regime:1 challenge:1 green:1 interpretability:1 explanation:1 memory:2 unrealistic:1 critical:2 natural:1 force:1 predicting:1 karklin:1 representing:6 scheme:9 naive:1 text:1 understanding:2 relative:1 embedded:2 highlight:1 limitation:1 proportional:1 validation:1 digital:1 integrate:2 degree:2 controversial:1 sufficient:1 consistent:4 proxy:1 exciting:1 summary:2 supported:1 surprisingly:1 free:1 copy:1 asynchronous:3 normalised:1 burges:1 generalise:1 wide:1 fifth:2 leaky:2 sparse:1 distributed:21 benefit:5 curve:3 dimension:2 cortical:10 world:2 sensory:2 qualitatively:2 made:1 hyvarinen:1 welling:1 reconstructed:1 approximate:2 keep:1 confirm:1 reproduces:2 investigating:2 handbook:1 assumed:2 spatio:5 alternatively:1 latent:2 quenches:1 nature:3 robust:2 dendrite:1 bottou:1 complex:3 constructing:1 rue:1 behaviourally:1 did:1 main:6 spread:5 linearly:5 arrow:1 noise:8 arise:2 fair:1 ref:1 repeated:2 neuronal:4 fig:29 en:2 ff:2 slow:7 precision:3 decoded:11 exponential:2 debated:1 externally:3 magenta:1 embed:1 specific:3 explored:1 barrett:1 evidence:1 bivariate:3 intrinsic:1 dominates:1 grouping:1 adding:1 corr:1 modulates:1 ccg:15 sparseness:2 crosscorrelograms:1 entropy:1 likely:1 visual:2 correlogram:3 ordered:1 lewicki:1 corresponds:1 truth:2 relies:1 ma:1 modulate:1 viewed:2 presentation:1 ferster:1 price:1 absence:1 content:1 change:9 experimentally:2 shared:1 infinite:1 determined:1 generalisation:1 sophie:2 sampler:2 uniformly:2 distributes:1 principal:1 engineer:1 pas:1 discriminate:1 experimental:7 formally:1 owed:1 internal:1 mark:1 latter:1 arises:2 modulated:1 lowvar:1 mcmc:14 phenomenon:1 correlated:3 |
4,798 | 5,344 | Conditional Random Field Autoencoders
for Unsupervised Structured Prediction
Waleed Ammar
Chris Dyer
Noah A. Smith
School of Computer Science
Carnegie Mellon University
Pittsburgh, PA 15213, USA
{wammar,cdyer,nasmith}@cs.cmu.edu
Abstract
We introduce a framework for unsupervised learning of structured predictors with
overlapping, global features. Each input?s latent representation is predicted conditional on the observed data using a feature-rich conditional random field (CRF).
Then a reconstruction of the input is (re)generated, conditional on the latent structure, using a generative model which factorizes similarly to the CRF. The autoencoder formulation enables efficient exact inference without resorting to unrealistic
independence assumptions or restricting the kinds of features that can be used.
We illustrate connections to traditional autoencoders, posterior regularization, and
multi-view learning. We then show competitive results with instantiations of the
framework for two canonical tasks in natural language processing: part-of-speech
induction and bitext word alignment, and show that training the proposed model
can be substantially more efficient than a comparable feature-rich baseline.
1
Introduction
Conditional random fields [24] are used to model structure in numerous problem domains, including natural language processing (NLP), computational biology, and computer vision. They enable
efficient inference while incorporating rich features that capture useful domain-specific insights. Despite their ubiquity in supervised settings, CRFs?and, crucially, the insights about effective feature
sets obtained by developing them?play less of a role in unsupervised structure learning, a problem which traditionally requires jointly modeling observations and the latent structures of interest.
For unsupervised structured prediction problems, less powerful models with stronger independence
assumptions are standard.1 This state of affairs is suboptimal in at least three ways: (i) adhering
to inconvenient independence assumptions when designing features is limiting?we contend that
effective feature engineering is a crucial mechanism for incorporating inductive bias in unsupervised learning problems; (ii) features and their weights have different semantics in joint and conditional models (see ?3.1); and (iii) modeling the generation of high-dimensional observable data with
feature-rich models is computationally challenging, requiring expensive marginal inference in the
inner loop of iterative parameter estimation algorithms (see ?3.1).
Our approach leverages the power and flexibility of CRFs in unsupervised learning without sacrificing their attractive computational properties or changing the semantics of well-understood feature
sets. Our approach replaces the standard joint model of observed data and latent structure with a twolayer conditional random field autoencoder that first generates latent structure with a CRF (conditional on the observed data) and then (re)generates the observations conditional on just the predicted
structure. For the reconstruction model, we use distributions which offer closed-form maximum
1
For example, a first-order hidden Markov model requires that yi ? xi+1 | yi+1 for a latent sequence
y = hy1 , y2 , . . .i generating x = hx1 , x2 , . . .i, while a first-order CRF allows yi to directly depend on xi+1 .
1
Extension: partial reconstruction. In our running POS example, the reconstruction model
p? (?
xi | yi ) defines a distribution over words given tags. Because word distributions are heavytailed, estimating such a distribution reliably is quite challenging. Our solution is to define a function ? : X ? X? such that |X? | ? |X |, and let x
?i = ?(xi ) be a deterministic transformation of the
original structured observation. We can add indirect supervision by defining ? such that it represents
observed information relevant to the latent structure of interest. For example, we found reconstructing Brown clusters [5] of tokens instead of their surface forms to improve POS induction. Other
possible reconstructions include word embeddings, morphological and spelling features of words.
More general graphs. We presented the CRF autoencoder in terms of sequential Markovian assumptions for ease of exposition; however, this framework can be used to model arbitrary hidden
structures. For example, instantiations of this model can be used for unsupervised learning of parse
trees [21], semantic role labels [42], and coreference resolution [35] (in NLP), motif structures [1]
in computational biology, and object recognition [46] in computer vision. The requirements for
applying the CRF autoencoder model are:
? An encoding discriminative model defining p? (y | x, ?). The encoder may be any model family
where supervised learning from hx, yi pairs is efficient.
? A reconstruction model that defines p? (?
x | y, ?) such that inference over y given hx, x
?i is
efficient.
? The independencies among y | x, x
? are not strictly weaker than those among y | x.
2.1
Learning & Inference
Model parameters are selected to maximize the regularized conditional log likelihood of reconstructed observations x
? given the structured observation x:
P
P
x | y)
(2)
??(?, ?) = R1 (?) + R2 (?) + (x,?x)?T log y p? (y | x) ? p? (?
We apply block coordinate descent, alternating between maximizing with respect to the CRF parameters (?-step) and the reconstruction parameters (?-step). Each ?-step applies one or two iterations
of a gradient-based convex optimizer.5 The ?-step applies one or two iterations of EM [10], with a
closed-form solution in the M-step in each EM iteration. The independence assumptions among y
make the marginal inference required in both steps straightforward; we omit details for space.
In the experiments below, we apply a squared L2 regularizer for the CRF parameters ?, and a
symmetric Dirichlet prior for categorical parameters ?.
The asymptotic runtime complexity of each block coordinate descent iteration, assuming the firstorder Markov dependencies in Fig. 2 (right), is:
(3)
O |?| + |?| + |T | ? |x|max ? |Y|max ? (|Y|max ? |Fyi?1 ,yi | + |Fx,yi |)
where Fyi?1 ,yi are the active ?label bigram? features used in hyi?1 , yi i factors, Fx,yi are the active
emission-like features used in hx, yi i factors. |x|max is the maximum length of an observation
sequence. |Y|max is the maximum cardinality6 of the set of possible assignments of yi .
After learning the ? and ? parameters of the CRF autoencoder, test-time predictions are made using maximum a posteriori estimation, conditioning on both observations and reconstructions, i.e.,
? MAP = arg maxy p?,? (y | x, x
y
?).
3
Connections To Previous Work
This work relates to several strands of work in unsupervised learning. Two broad types of models
have been explored that support unsupervised learning with flexible feature representations. Both are
5
We experimented with AdaGrad [12] and L-BFGS. When using AdaGrad, we accummulate the gradient
vectors across block coordinate ascent iterations.
6
In POS induction, |Y| is a constant, the number of syntactic classes which we configure to 12 in our experiments. In word alignment, |Y| is the size of the source sentence plus one, therefore |Y|max is the maximum
length of a source sentence in the bitext corpus.
4
fully generative models that define joint distributions over x and y. We discuss these ?undirected?
and ?directed? alternatives next, then turn to less closely related methods.
3.1
Existing Alternatives for Unsupervised Learning with Features
Undirected models. A Markov random field (MRF) encodes the joint distribution through local
potential functions parameterized using features. Such models ?normalize globally,? requiring during training the calculation of a partition function summing over all possible inputs and outputs. In
our notation:
X X
exp ?? g
?(x, y)
(4)
Z(?) =
x?X ? y?Y |x|
where g
? collects all the local factorization by cliques of the graph, for clarity. The key difficulty
is in the summation over all possible observations. Approximations have been proposed, including
contrastive estimation, which sums over subsets of X ? [38, 43] (applied variously to POS learning
by Haghighi and Klein [18] and word alignment by Dyer et al. [14]) and noise contrastive estimation
[30].
Directed models. The directed alternative avoids the global partition function by factorizing the
joint distribution in terms of locally normalized conditional probabilities, which are parameterized
in terms of features. For unsupervised sequence labeling, the model was called a ?feature HMM?
by Berg-Kirkpatrick et al. [3]. The local emission probabilities p(xi | yi ) in a first-order HMM for
POS tagging are reparameterized as follows (again, using notation close to ours):
p? (xi | yi ) = P
exp ?? g(xi , yi )
?
x?X exp ? g(x, yi )
(5)
The features relating hidden to observed variables must be local within the factors implied by the
directed graph. We show below that this locality restriction excludes features that are useful (?A.1).
Put in these terms, the proposed autoencoding model is a hybrid directed-undirected model.
Asymptotic Runtime Complexity of Inference. The models just described cannot condition on
arbitrary amounts of x without increasing inference costs. Despite the strong independence assumptions of those models, the computational complexity of inference required for learning with CRF
autoencoders is better (?2.1).
Consider learning the parameters of an undirected model by maximizing likelihood of the observed
data. Computing the gradient for a training instance x requires time
O |?| + |T | ? |x| ? |Y| ? (|Y| ? |Fyi?1 ,yi |+|X | ? |Fxi ,yi |) ,
where Fxi ?yi are the emission-like features used in an arbitrary assignment of xi and yi . When the
multiplicative factor |X | is large, inference is slow compared to CRF autoencoders.
Inference in directed models is faster than in undirected models, but still slower than CRF autoencoder models. In directed models [3], each iteration requires time
O |?| + |T | ? |x| ? |Y| ? (|Y| ? |Fyi?1 ,yi | + |Fxi ,yi |)+|? ? | ? max(|Fyi?1 ,yi |, |FX ,yi |) ,
where Fxi ,yi are the active emission features used in an arbitrary assignment of xi and yi , FX ,yi
is the union of all emission features used with an arbitrary assignment of yi , and ? ? are the local
emission and transition probabilities. When |X | is large, the last term |? ? |?max(|Fyi?1 ,yi |, |FX ,yi |)
can be prohibitively large.
3.2
Other Related Work
The proposed CRF autoencoder is more distantly related to several important ideas in less-thansupervised learning.
5
Autoencoders and other ?predict self? methods. Our framework borrows its general structure,
Fig. 2 (left), as well as its name, from neural network autoencoders. The goal of neural autoencoders
has been to learn feature representations that improve generalization in otherwise supervised learning problems [44, 8, 39]. In contrast, the goal of CRF autoencoders is to learn specific interpretable
regularities of interest.7 It is not clear how neural autoencoders could be used to learn the latent
structures that CRF autoencoders learn, without providing supervised training examples. Stoyanov
et al. [40] presented a related approach for discriminative graphical model learning, including features and latent variables, based on backpropagation, which could be used to instantiate the CRF
autoencoder.
Daum?e III [9] introduced a reduction of an unsupervised problem instance to a series of singlevariable supervised classifications. The first series of these construct a latent structure y given the
entire x, then the second series reconstruct the input. The approach can make use of any supervised
learner; if feature-based probabilistic models were used, a |X | summation (akin to Eq. 5) would
be required. On unsupervised POS induction, this approach performed on par with the undirected
model of Smith and Eisner [38].
Minka [29] proposed cascading a generative model and a discriminative model, where class labels
(to be predicted at test time) are marginalized out in the generative part first, and then (re)generated
in the discriminative part. In CRF autoencoders, observations (available at test time) are conditioned
on in the discriminative part first, and then (re)generated in the generative part.
Posterior regularization. Introduced by Ganchev et al. [16], posterior regularization is an effective method for specifying constraint on the posterior distributions of the latent variables of interest;
a similar idea was proposed independently by Bellare et al. [2]. For example, in POS induction,
every sentence might be expected to contain at least one verb. This is imposed as a soft constraint,
i.e., a feature whose expected value under the model?s posterior is constrained. Such expectation
constraints are specified directly by the domain-aware model designer.8 The approach was applied
to unsupervised POS induction, word alignment, and parsing. Although posterior regularization was
applied to directed feature-less generative models, the idea is orthogonal to the model family and
can be used to add more inductive bias for training CRF autoencoder models.
4
Evaluation
We evaluate the effectiveness of CRF autoencoders for learning from unlabeled examples in POS
induction and word alignment. We defer the detailed experimental setup to Appendix A.
Part-of-Speech Induction Results. Fig. 3 compares predictions of the CRF autoencoder model
in seven languages to those of a featurized first-order HMM model [3] and a standard (feature-less)
first-order HMM, using V-measure [37] (higher is better). First, note the large gap between both
feature-rich models on the one hand, and the feature-less HMM model on the other hand. Second,
note that CRF autoencoders outperform featurized HMMs in all languages, except Italian, with an
average relative improvement of 12%.
These results provide empirical evidence that feature engineering is an important source of inductive
bias for unsupervised structured prediction problems. In particular, we found that using Brown
cluster reconstructions and specifying features which span multiple words significantly improve the
performance. Refer to Appendix A for more analysis.
Bitext Word Alignment Results. First, we consider an intrinsic evaluation on a Czech-English
dataset of manual alignments, measuring the alignment error rate (AER; [32]). We also perform an
7
This is possible in CRF autoencoders due to the interdependencies among variables in the hidden structure
and the manually specified feature templates which capture the relationship between observations and their
hidden structures.
8
In a semi-supervised setting, when some labeled examples of the hidden structure are available, Druck
and McCallum [11] used labeled examples to estimate desirable expected values. We leave semi-supervised
applications of CRF autoencoders to future work; see also Suzuki and Isozaki [41].
6
0.6
0.4
0.3
0.0
0.1
0.2
V?measure
0.5
Standard HMM
Featurized HMM
CRF autoencoder
Arabic
Basque
Danish
Greek
Hungarian
Italian
Turkish
Average
Figure 3: V-measure [37] of induced parts of speech in seven languages. The CRF autoencoder with
features spanning multiple words and with Brown cluster reconstructions achieves the best results in
all languages but Italian, closely followed by the feature-rich HMM of Berg-Kirkpatrick et al. [3].
The standard multinomial HMM consistently ranks last.
direction
forward
reverse
symmetric
fast align
model 4
auto
27.7
25.9
25.2
31.5
24.1
22.2
27.5
21.1
19.5
pair
cs-en
ur-en
zh-en
fast align
model 4
auto
15.2?0.3
20.0?0.6
56.9?1.6
15.3?0.1
20.1?0.6
56.7?1.6
15.5?0.1
20.8?0.5
56.1?1.7
Table 1: Left: AER results (%) for Czech-English word alignment. Lower values are better. . Right:
Bleu translation quality scores (%) for Czech-English, Urdu-English and Chinese-English. Higher
values are better. .
extrinsic evaluation of translation quality in three language pairs, using case-insensitive Bleu [33] of
a machine translation system (cdec9 [13]) built using the word alignment predictions of each model.
AER for variants of each model (forward, reverse, and symmetrized) are shown in Table 1 (left).
Our model significantly outperforms both baselines. Bleu scores on the three language pairs are
shown in Table 1; alignments obtained with our CRF autoencoder model improve translation quality
of the Czech-English and Urdu-English translation systems, but not of Chinese-English. This is unsurprising, given that Chinese orthography does not use letters, so that source-language spelling and
morphology features our model incorporates introduce only noise here. Better feature engineering,
or more data, is called for.
We have argued that the feature-rich CRF autoencoder will scale better than its feature-rich alternatives. Fig. 5 (in Appendix A.2) shows the average per-sentence inference runtime for the CRF
autoencoder compared to exact inference in an MRF [14] with a similar feature set, as a function of
the number of sentences in the corpus. For CRF autoencoders, the average inference runtime grows
slightly due to the increased number of parameters, while it grows substantially with vocabulary size
in MRF models [14].10
5
Conclusion
We have presented a general and scalable framework to learn from unlabeled examples for structured
prediction. The technique allows features with global scope in observed variables with favorable
asymptotic inference runtime. We achieve this by embedding a CRF as the encoding model in the
9
http://www.cdec-decoder.org/
We only compare runtime, instead of alignment quality, because retraining the MRF model with exact
inference was too expensive.
10
7
input layer of an autoencoder, and reconstructing a transformation of the input at the output layer
using simple categorical distributions. The key advantages of the proposed model are scalability and
modeling flexibility. We applied the model to POS induction and bitext word alignment, obtaining
results that are competitive with the state of the art on both tasks.
Acknowledgments
We thank Brendan O?Connor, Dani Yogatama, Jeffrey Flanigan, Manaal Faruqui, Nathan Schneider,
Phil Blunsom and the anonymous reviewers for helpful suggestions. We also thank Taylor BergKirkpatrick for providing his implementation of the POS induction baseline, and Phil Blunsom for
sharing POS induction evaluation scripts. This work was sponsored by the U.S. Army Research
Laboratory and the U.S. Army Research Office under contract/grant number W911NF-10-1-0533.
The statements made herein are solely the responsibility of the authors.
References
[1] T. L. Bailey and C. Elkan. Unsupervised learning of multiple motifs in biopolymers using
expectation maximization. Machine learning, 1995.
[2] K. Bellare, G. Druck, and A. McCallum. Alternating projections for learning with expectation
constraints. In Proc. of UAI, 2009.
[3] T. Berg-Kirkpatrick, A. Bouchard-C?ot?e, J. DeNero, and D. Klein. Painless unsupervised learning with features. In Proc. of NAACL, 2010.
[4] P. Blunsom and T. Cohn. Discriminative word alignment with conditional random fields. In
Proc. of Proceedings of ACL, 2006.
[5] P. F. Brown, P. V. deSouza, R. L. Mercer, V. J. D. Pietra, and J. C. Lai. Class-based n-gram
models of natural language. Computational Linguistics, 1992.
[6] P. F. Brown, V. J. D. Pietra, S. A. D. Pietra, and R. L. Mercer. The mathematics of statistical
machine translation: parameter estimation. In Computational Linguistics, 1993.
[7] S. Buchholz and E. Marsi. CoNLL-X shared task on multilingual dependency parsing. In
CoNLL-X, 2006.
[8] R. Collobert and J. Weston. A unified architecture for natural language processing: Deep
neural networks with multitask learning. In Proc. of ICML, 2008.
[9] H. Daum?e III. Unsupervised search-based structured prediction. In Proceedings of the 26th
Annual International Conference on Machine Learning, pages 209?216. ACM, 2009.
[10] A. P. Dempster, N. M. Laird, D. B. Rubin, et al. Maximum likelihood from incomplete data
via the em algorithm. Journal of the Royal statistical Society, 39(1):1?38, 1977.
[11] G. Druck and A. McCallum. High-performance semi-supervised learning using discriminatively constrained generative models. In Proc. of ICML, 2010.
[12] J. Duchi, E. Hazan, and Y. Singer. Adaptive subgradient methods for online learning and
stochastic optimization. JMLR, 2011.
[13] C. Dyer, A. Lopez, J. Ganitkevitch, J. Weese, F. Ture, P. Blunsom, H. Setiawan, V. Eidelman,
and P. Resnik. cdec: A decoder, alignment, and learning framework for finite-state and contextfree translation models. In Proc. of ACL, 2010.
[14] C. Dyer, J. Clark, A. Lavie, and N. A. Smith. Unsupervised word alignment with arbitrary
features. In Proc. of ACL-HLT, 2011.
[15] C. Dyer, V. Chahuneau, and N. A. Smith. A simple, fast, and effective reparameterization of
IBM Model 2. In Proc. of NAACL, 2013.
[16] K. Ganchev, J. Grac?a, J. Gillenwater, and B. Taskar. Posterior regularization for structured
latent variable models. Journal of Machine Learning Research, 11:2001?2049, 2010.
[17] Q. Gao and S. Vogel. Parallel implementations of word alignment tool. In In Proc. of the ACL
workshop, 2008.
[18] A. Haghighi and D. Klein. Prototype-driven learning for sequence models. In Proc. of NAACLHLT, 2006.
[19] F. Jelinek. Statistical Methods for Speech Recognition. MIT Press, 1997.
8
[20] M. Johnson. Why doesn?t EM find good HMM POS-taggers? In Proc. of EMNLP, 2007.
[21] D. Klein and C. D. Manning. Corpus-based induction of syntactic structure: Models of dependency and constituency. In Proc. of ACL, 2004.
[22] P. Koehn. Statistical Machine Translation. Cambridge, 2010.
[23] P. Koehn, F. J. Och, and D. Marcu. Statistical phrase-based translation. In Proc. of NAACL,
2003.
[24] J. Lafferty, A. McCallum, and F. Pereira. Conditional random fields: Probabilistic models for
segmenting and labeling sequence data. In Proc. of ICML, 2001.
[25] P. Liang. Semi-supervised learning for natural language. In Thesis, MIT, 2005.
[26] C.-C. Lin, W. Ammar, C. Dyer, and L. Levin. The cmu submission for the shared task on language identification in code-switched data. In First Workshop on Computational Approaches
to Code Switching at EMNLP, 2014.
[27] A. V. Lukashin and M. Borodovsky. Genemark. hmm: new solutions for gene finding. Nucleic
acids research, 26(4):1107?1115, 1998.
[28] B. Merialdo. Tagging english text with a probabilistic model. In Comp. Ling., 1994.
[29] T. Minka. Discriminative models, not discriminative training. Technical report, Technical
Report MSR-TR-2005-144, Microsoft Research, 2005.
[30] A. Mnih and Y. W. Teh. A fast and simple algorithm for training neural probabilistic language
models. In Proc. of ICML, 2012.
[31] J. Nivre, J. Hall, S. Kubler, R. McDonald, J. Nilsson, S. Riedel, and D. Yuret. The CoNLL
2007 shared task on dependency parsing. In Proc. of CoNLL, 2007.
[32] F. Och and H. Ney. A systematic comparison of various statistical alignment models. Computational Linguistics, 2003.
[33] K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu. Bleu: a method for automatic evaluation of
machine translation. In Proc. of ACL, 2002.
[34] S. Petrov, D. Das, and R. McDonald. A universal part-of-speech tagset. In Proc. of LREC,
May 2012.
[35] H. Poon and P. Domingos. Joint unsupervised coreference resolution with Markov logic. In
Proc. of EMNLP, 2008.
[36] S. Reddy and S. Waxmonsky. Substring-based transliteration with conditional random fields.
In Proc. of the Named Entities Workshop, 2009.
[37] A. Rosenberg and J. Hirschberg. V-measure: A conditional entropy-based external cluster
evaluation measure. In EMNLP-CoNLL, 2007.
[38] N. A. Smith and J. Eisner. Contrastive estimation: Training log-linear models on unlabeled
data. In Proc. of ACL, 2005.
[39] R. Socher, C. D. Manning, and A. Y. Ng. Learning continuous phrase representations and
syntactic parsing with recursive neural networks. In NIPS workshop, 2010.
[40] V. Stoyanov, A. Ropson, and J. Eisner. Empirical risk minimization of graphical model parameters given approximate inference, decoding, and model structure. In Proc. of AISTATS,
2011.
[41] J. Suzuki and H. Isozaki. Semi-supervised sequential labeling and segmentation using gigaword scale unlabeled data. In Proc. of ACL, 2008.
[42] R. Swier and S. Stevenson. Unsupervised semantic role labelling. In Proc. of EMNLP, 2004.
[43] D. Vickrey, C. C. Lin, and D. Koller. Non-local contrastive objectives. In Proc. of ICML, 2010.
[44] P. Vincent, H. Larochelle, Y. Bengio, and P.-A. Manzagol. Extracting and composing robust
features with denoising autoencoders. In Proc. of ICML, 2008.
[45] S. Vogel, H. Ney, and C. Tillmann. Hmm-based word alignment in statistical translation. In
Proc. of COLING, 1996.
[46] M. Weber, M. Welling, and P. Perona. Unsupervised learning of models for recognition. 2000.
[47] J. Yamato, J. Ohya, and K. Ishii. Recognizing human action in time-sequential images using
hidden Markov model. In Proc. of CVPR, pages 379?385. IEEE, 1992.
9
| 5344 |@word multitask:1 arabic:1 msr:1 bigram:1 stronger:1 retraining:1 crucially:1 contrastive:4 twolayer:1 tr:1 reduction:1 series:3 score:2 ours:1 outperforms:1 existing:1 must:1 parsing:4 partition:2 enables:1 interpretable:1 sponsored:1 generative:7 selected:1 instantiate:1 mccallum:4 affair:1 smith:5 org:1 tagger:1 lopez:1 introduce:2 tagging:2 expected:3 multi:1 morphology:1 globally:1 increasing:1 estimating:1 notation:2 kind:1 substantially:2 unified:1 finding:1 transformation:2 every:1 firstorder:1 runtime:6 prohibitively:1 grant:1 omit:1 och:2 kubler:1 segmenting:1 engineering:3 understood:1 local:6 switching:1 despite:2 encoding:2 solely:1 might:1 plus:1 blunsom:4 acl:8 collect:1 challenging:2 specifying:2 ease:1 factorization:1 hmms:1 directed:8 acknowledgment:1 merialdo:1 union:1 block:3 recursive:1 backpropagation:1 turkish:1 universal:1 empirical:2 significantly:2 projection:1 word:19 hx1:1 cannot:1 close:1 unlabeled:4 faruqui:1 put:1 risk:1 applying:1 restriction:1 www:1 deterministic:1 map:1 imposed:1 crfs:2 maximizing:2 straightforward:1 phil:2 reviewer:1 independently:1 convex:1 resolution:2 adhering:1 tillmann:1 insight:2 yuret:1 cascading:1 ropson:1 his:1 reparameterization:1 embedding:1 traditionally:1 coordinate:3 fx:5 limiting:1 weese:1 play:1 exact:3 designing:1 domingo:1 elkan:1 pa:1 fyi:6 expensive:2 recognition:3 marcu:1 submission:1 labeled:2 observed:7 role:3 taskar:1 capture:2 morphological:1 dempster:1 complexity:3 depend:1 coreference:2 learner:1 po:13 joint:6 indirect:1 various:1 regularizer:1 fast:4 effective:4 labeling:3 quite:1 whose:1 koehn:2 cvpr:1 otherwise:1 reconstruct:1 encoder:1 ward:1 syntactic:3 jointly:1 laird:1 online:1 autoencoding:1 sequence:5 advantage:1 manaal:1 reconstruction:10 relevant:1 loop:1 poon:1 flexibility:2 achieve:1 papineni:1 normalize:1 scalability:1 cluster:4 requirement:1 r1:1 regularity:1 generating:1 leave:1 object:1 illustrate:1 school:1 eq:1 strong:1 c:2 predicted:3 hungarian:1 larochelle:1 direction:1 greek:1 denero:1 closely:2 stochastic:1 human:1 enable:1 argued:1 hx:3 generalization:1 anonymous:1 summation:2 extension:1 strictly:1 hall:1 exp:3 scope:1 predict:1 optimizer:1 achieves:1 heavytailed:1 estimation:6 favorable:1 proc:28 label:3 ganchev:2 tool:1 grac:1 minimization:1 dani:1 mit:2 ohya:1 factorizes:1 rosenberg:1 office:1 emission:6 improvement:1 consistently:1 rank:1 likelihood:3 contrast:1 brendan:1 ishii:1 baseline:3 posteriori:1 inference:17 helpful:1 motif:2 entire:1 hidden:7 italian:3 koller:1 perona:1 semantics:2 arg:1 among:4 flexible:1 classification:1 constrained:2 art:1 marginal:2 field:8 construct:1 aware:1 ng:1 manually:1 biology:2 represents:1 broad:1 unsupervised:22 icml:6 distantly:1 future:1 report:2 variously:1 pietra:3 jeffrey:1 microsoft:1 interest:4 mnih:1 evaluation:6 alignment:19 kirkpatrick:3 configure:1 partial:1 orthogonal:1 tree:1 incomplete:1 taylor:1 re:4 inconvenient:1 sacrificing:1 instance:2 increased:1 modeling:3 soft:1 markovian:1 w911nf:1 measuring:1 assignment:4 maximization:1 phrase:2 cost:1 subset:1 predictor:1 recognizing:1 levin:1 johnson:1 marsi:1 too:1 unsurprising:1 dependency:4 international:1 probabilistic:4 contract:1 systematic:1 decoding:1 druck:3 squared:1 again:1 thesis:1 emnlp:5 external:1 potential:1 stevenson:1 bfgs:1 collobert:1 multiplicative:1 view:1 performed:1 closed:2 script:1 responsibility:1 hazan:1 competitive:2 parallel:1 bouchard:1 defer:1 acid:1 identification:1 vincent:1 substring:1 comp:1 manual:1 sharing:1 danish:1 hlt:1 petrov:1 minka:2 dataset:1 segmentation:1 higher:2 nivre:1 nasmith:1 supervised:11 formulation:1 just:2 autoencoders:17 hand:2 parse:1 cohn:1 overlapping:1 defines:2 quality:4 grows:2 usa:1 name:1 naacl:3 requiring:2 y2:1 brown:5 normalized:1 inductive:3 regularization:5 contain:1 alternating:2 symmetric:2 laboratory:1 semantic:2 vickrey:1 attractive:1 during:1 self:1 crf:30 mcdonald:2 duchi:1 cdec:2 weber:1 image:1 hy1:1 multinomial:1 conditioning:1 insensitive:1 relating:1 mellon:1 refer:1 connor:1 cambridge:1 bitext:4 automatic:1 resorting:1 mathematics:1 similarly:1 gillenwater:1 language:14 supervision:1 surface:1 add:2 align:2 posterior:7 driven:1 reverse:2 yi:30 isozaki:2 schneider:1 urdu:2 maximize:1 hyi:1 transliteration:1 ii:1 relates:1 multiple:3 interdependency:1 semi:5 stoyanov:2 desirable:1 technical:2 faster:1 calculation:1 offer:1 lin:2 lai:1 prediction:8 mrf:4 variant:1 scalable:1 vision:2 cmu:2 expectation:3 iteration:6 orthography:1 source:4 crucial:1 haghighi:2 ot:1 vogel:2 ascent:1 induced:1 undirected:6 incorporates:1 lafferty:1 effectiveness:1 extracting:1 leverage:1 iii:3 embeddings:1 ture:1 bengio:1 independence:5 architecture:1 suboptimal:1 inner:1 idea:3 prototype:1 akin:1 speech:5 action:1 deep:1 useful:2 clear:1 detailed:1 amount:1 locally:1 bellare:2 constituency:1 http:1 outperform:1 canonical:1 designer:1 extrinsic:1 per:1 klein:4 gigaword:1 carnegie:1 independency:1 key:2 changing:1 clarity:1 graph:3 excludes:1 subgradient:1 sum:1 parameterized:2 powerful:1 letter:1 named:1 family:2 appendix:3 conll:5 comparable:1 layer:2 lrec:1 followed:1 replaces:1 annual:1 aer:3 noah:1 constraint:4 riedel:1 x2:1 encodes:1 tag:1 generates:2 nathan:1 span:1 structured:9 developing:1 manning:2 lavie:1 across:1 slightly:1 reconstructing:2 em:4 featurized:3 ur:1 nilsson:1 maxy:1 yogatama:1 computationally:1 reddy:1 lukashin:1 discus:1 turn:1 mechanism:1 singer:1 dyer:6 available:2 apply:2 fxi:4 ubiquity:1 bailey:1 ney:2 alternative:4 symmetrized:1 slower:1 original:1 running:1 nlp:2 include:1 dirichlet:1 graphical:2 linguistics:3 marginalized:1 daum:2 eisner:3 chinese:3 society:1 implied:1 objective:1 spelling:2 traditional:1 gradient:3 thank:2 entity:1 hmm:12 decoder:2 chris:1 seven:2 spanning:1 induction:12 bleu:4 assuming:1 length:2 code:2 relationship:1 manzagol:1 providing:2 liang:1 setup:1 statement:1 implementation:2 reliably:1 contend:1 perform:1 teh:1 observation:10 nucleic:1 markov:5 finite:1 descent:2 reparameterized:1 defining:2 biopolymers:1 arbitrary:6 verb:1 introduced:2 pair:4 required:3 specified:2 connection:2 sentence:5 czech:4 herein:1 nip:1 below:2 buchholz:1 built:1 including:3 max:8 royal:1 unrealistic:1 power:1 natural:5 difficulty:1 regularized:1 hybrid:1 zhu:1 improve:4 numerous:1 categorical:2 autoencoder:16 auto:2 text:1 prior:1 l2:1 ammar:2 zh:1 adagrad:2 asymptotic:3 relative:1 fully:1 par:1 discriminatively:1 generation:1 suggestion:1 borrows:1 clark:1 switched:1 mercer:2 rubin:1 roukos:1 translation:11 ibm:1 token:1 last:2 english:9 bias:3 weaker:1 template:1 jelinek:1 vocabulary:1 transition:1 avoids:1 rich:8 gram:1 doesn:1 forward:2 made:2 suzuki:2 author:1 adaptive:1 welling:1 reconstructed:1 approximate:1 observable:1 multilingual:1 gene:1 clique:1 logic:1 global:3 active:3 instantiation:2 uai:1 desouza:1 corpus:3 pittsburgh:1 summing:1 xi:9 discriminative:8 factorizing:1 search:1 latent:12 iterative:1 continuous:1 why:1 table:3 learn:5 contextfree:1 robust:1 composing:1 obtaining:1 basque:1 domain:3 da:1 aistats:1 noise:2 ling:1 fig:4 singlevariable:1 en:3 resnik:1 slow:1 pereira:1 jmlr:1 coling:1 specific:2 r2:1 explored:1 experimented:1 evidence:1 incorporating:2 intrinsic:1 workshop:4 restricting:1 sequential:3 socher:1 labelling:1 conditioned:1 painless:1 gap:1 locality:1 entropy:1 army:2 gao:1 strand:1 applies:2 acm:1 weston:1 conditional:15 goal:2 exposition:1 shared:3 except:1 denoising:1 called:2 experimental:1 berg:3 support:1 evaluate:1 tagset:1 |
4,799 | 5,345 | Learning Generative Models with Visual Attention
Yichuan Tang, Nitish Srivastava, Ruslan Salakhutdinov
Department of Computer Science
University of Toronto
Toronto, Ontario, Canada
{tang,nitish,rsalakhu}@cs.toronto.edu
Abstract
Attention has long been proposed by psychologists to be important for efficiently
dealing with the massive amounts of sensory stimulus in the neocortex. Inspired
by the attention models in visual neuroscience and the need for object-centered
data for generative models, we propose a deep-learning based generative framework using attention. The attentional mechanism propagates signals from the
region of interest in a scene to an aligned canonical representation for generative modeling. By ignoring scene background clutter, the generative model can
concentrate its resources on the object of interest. A convolutional neural net is
employed to provide good initializations during posterior inference which uses
Hamiltonian Monte Carlo. Upon learning images of faces, our model can robustly
attend to the face region of novel test subjects. More importantly, our model can
learn generative models of new faces from a novel dataset of large images where
the face locations are not known.
1
Introduction
Building rich generative models that are capable of extracting useful, high-level latent representations from high-dimensional sensory input lies at the core of solving many AI-related tasks, including object recognition, speech perception and language understanding. These models capture
underlying structure in data by defining flexible probability distributions over high-dimensional data
as part of a complex, partially observed system. Some of the successful generative models that
are able to discover meaningful high-level latent representations include the Boltzmann Machine
family of models: Restricted Boltzmann Machines, Deep Belief Nets [1], and Deep Boltzmann Machines [2]. Mixture models, such as Mixtures of Factor Analyzers [3] and Mixtures of Gaussians,
have also been used for modeling natural image patches [4]. More recently, denoising auto-encoders
have been proposed as a way to model the transition operator that has the same invariant distribution
as the data generating distribution [5].
Generative models have an advantage over discriminative models when part of the images are occluded or missing. Occlusions are very common in realistic settings and have been largely ignored
in recent literature on deep learning. In addition, prior knowledge can be easily incorporated in
generative models in the forms of structured latent variables, such as lighting and deformable parts.
However, the enormous amount of content in high-resolution images makes generative learning difficult [6, 7]. Therefore, generative models have found most success in learning to model small
patches of natural images and objects: Zoran and Weiss [4] learned a mixture of Gaussians model
over 8?8 image patches; Salakhutdinov and Hinton [2] used 64?64 centered and uncluttered stereo
images of toy objects on a clear background; Tang et al. [8] used 24?24 images of centered and
cropped faces. The fact that these models require curated training data limits their applicability on
using the (virtually) unlimited unlabeled data.
In this paper, we propose a framework to infer the region of interest in a big image for generative modeling. This will allow us to learn a generative model of faces on a very large dataset of
(unlabeled) images containing faces. Our framework is able to dynamically route the relevant information to the generative model and can ignore the background clutter. The need to dynamically and
selectively route information is also present in the biological brain. Plethora of evidence points to
1
the presence of attention in the visual cortex [9, 10]. Recently, in visual neuroscience, attention has
been shown to exist not only in extrastriate areas, but also all the way down to V1 [11].
Attention as a form of routing was originally proposed by Anderson and Van Essen [12] and then
extended by Olshausen et al. [13]. Dynamic routing has been hypothesized as providing a way for
achieving shift and size invariance in the visual cortex [14, 15]. Tsotsos et al. [16] proposed a model
combining search and attention called the Selective Tuning model. Larochelle and Hinton [17] proposed a way of using third-order Boltzmann Machines to combine information gathered from many
foveal glimpses. Their model chooses where to look next to find locations that are most informative
of the object class. Reichert et al. [18] proposed a hierarchical model to show that certain aspects of
covert object-based attention can be modeled by Deep Boltzmann Machines. Several other related
models attempt to learn where to look for objects [19, 20] and for video based tracking [21]. Inspired
by Olshausen et al. [13], we use 2D similarity transformations to implement the scaling, rotation,
and shift operation required for routing. Our main motivation is to enable the learning of generative
models in big images where the location of the object of interest is unknown a-priori.
2
Gaussian Restricted Boltzmann Machines
Before we describe our model, we briefly review the Gaussian Restricted Boltzmann Machine
(GRBM) [22], as it will serve as the building block for our attention-based model. GRBMs are
a type of Markov Random Field model that has a bipartite structure with real-valued visible variables v ? RD connected to binary stochastic hidden variables h ? {0, 1}H . The energy of the joint
configuration {v, h} of the Gaussian RBM is defined as follows:
X
1 X (vi ? bi )2 X
?
c
h
?
Wij vi hj ,
(1)
EGRBM (v, h; ?) =
j
j
2 i
?i2
j
ij
where ? = {W, b, c, ?}
P are the model parameters. The marginal distribution over the visible vector
1
v is P (v; ?) = Z(?)
h exp (?E(v, h; ?)) and the corresponding conditional distributions take
the following form:
X
p(hj = 1|v) = 1/ 1 + exp(?
Wij vi ? cj ) ,
(2)
i
p(vi |h)
= N (vi ; ?i , ?i2 ), where ?i = bi + ?i2
X
Wij hj .
(3)
j
Observe that conditioned on the states of the hidden variables (Eq. 3), each visible unit is modeled
by a Gaussian distribution, whose mean is shifted by the weighted combination of the hidden unit
activations. Unlike directed models, an RBM?s conditional distribution over hidden nodes is factorial
and can be easily computed.
We can also add a binary RBM on top of the learned GRBM by treating the inferred h as the
?visible? layer together with a second hidden layer h2 . This results in a 2-layer Gaussian Deep
Belief Network (GDBN) [1] that is a more powerful model of v.
Specifically, in a GDBN model, p(h1 , h2 ) is modeled by the energy function of the 2nd-layer RBM,
while p(v1 |h1 ) is given by Eq. 3. Efficient inference can be performed using the greedy approach
of [1] by treating each DBN layer as a separate RBM model. GDBNs have been applied to various
tasks, including image classification, video action and speech recognition [6, 23, 24, 25].
3
The Model
Let I be a high resolution image of a scene, e.g. a 256?256 image. We want to use attention to
propagate regions of interest from I up to a canonical representation. For example, in order to learn
a model of faces, the canonical representation could be a 24?24 aligned and cropped frontal face
image. Let v ? RD represent this low resolution canonical image. In this work, we focus on a Deep
Belief Network1 to model v.
This is illustrated in the diagrams of Fig. 1. The left panel displays the model of Olshausen et.al. [13],
whereas the right panel shows a graphical diagram of our proposed generative model with an attentional mechanism. Here, h1 and h2 represent the latent hidden variables of the DBN model, and
1
Other generative models can also be used with our attention framework.
2
2d similarity
transformation
Our model
Olshausen et al. 93
Figure 1: Left: The Shifter Circuit, a well-known neuroscience model for visual attention [13]; Right: The
proposed model uses 2D similarity transformations from geometry and a Gaussian DBN to model canonical
face images. Associative memory corresponds to the DBN, object-centered frame correspond to the visible
layer and the attentional mechanism is modeled by 2D similarity transformations.
4x, 4y, 4?, 4s (position, rotation, and scale) are the parameters of the 2D similarity transformation.
The 2D similarity transformation is used to rotate, scale, and translate the canonical image v onto the
canvas that we denote by I. Let p = [x y]T be a pixel coordinate (e.g. [0, 0] or [0, 1]) of the canonical
image v. Let {p} be the set of all coordinates of v. For example, if v is 24?24, then {p} ranges
from [0, 0] to [23, 23]. Let the ?gaze? variables u ? R4 ? [4x, 4y, 4?, 4s] be the parameter
of the Similarity transformation. In order to simplify derivations and to make transformations be
linear w.r.t. the transformation parameters, we can equivalently redefine u = [a, b, 4x, 4y],
where a = s sin(?) ? 1 and b = s cos(?) (see [26] for details). We further define a function
w := w(p, u) ? p0 as the transformation function to warp points p to p0 :
h 0 i h
ih
i h
i
1+a
?b
x
4x
x
p0 ,
=
+
.
(4)
0
y
b
1+a
y
4y
We use the notation I({p}) to denote the bilinear interpolation of I at coordinates {p} with antialiasing. Let x(u) be the extracted low-resolution image at warped locations p0 :
x(u) , I(w({p}, u)).
(5)
Intuitively, x(u) is a patch extracted from I according to the shift, rotation and scale parameters
of u, as shown in Fig. 1, right panel. It is this patch of data that we seek to model generatively. Note
that the dimensionality of x(u) is equal to the cardinality of {p}, where {p} denotes the set of pixel
coordinates of the canonical image v. Unlike standard generative learning tasks, the data x(u) is
not static but changes with the latent variables u. Given v and u, we model the top-down generative
process over2 x with a Gaussian distribution having a diagonal covariance matrix ? 2 I:
1 X (xi (u) ? vi )2
p(x|v, u, I) ? exp ?
.
(6)
2 i
?i2
The fact that we do not seek to model the rest of the regions/pixels of I is by design. By using 2D
similarity transformation to mimic attention, we can discard the complex background of the scene
and let the generative model focus on the object of interest. The proposed generative model takes
the following form:
p(x, v, u|I) = p(x|v, u, I)p(v)p(u),
(7)
where for p(u) we use a flat prior that is constant for all u, and p(v) is defined by a 2-layer Gaussian
Deep Belief Network. The conditional p(x|v, u, I) is given by a Gaussian distribution as in Eq. 6.
To simplify the inference procedure, p(x|v, u, I) and the GDBN model of v, p(v), will share the
same noise parameters ?i .
2
We will often omit dependence of x on u for clarity of presentation.
3
4
Inference
While the generative equations in the last section are straightforward and intuitive, inference in these
models is typically intractable due to the complicated energy landscape of the posterior. During
inference, we wish to compute the distribution over the gaze variables u and canonical object v given
the big image I. Unlike in standard RBMs and DBNs, there are no simplifying factorial assumptions
about the conditional distribution of the latent variable u. Having a 2D similarity transformation is
reminiscent of third-order Boltzmann machines with u performing top-down multiplicative gating
of the connections between v and I. It is well known that inference in these higher-order models is
rather complicated.
One way to perform inference in our model is to resort to Gibbs sampling by computing the set of
alternating conditional posteriors: The conditional distribution over the canonical image v takes the
following form:
? + x(u)
(8)
; ?2 ,
p(v|u, h1 , I) = N
2
P
where ?i = bi + ?i2 j Wij h1j is the top-down influence of the DBN. Note that if we know the
gaze variable u and the first layer of hidden variables h1 , then v is simply defined by a Gaussian
distribution, where the mean is given by the average of the top-down influence and bottom-up information from x. The conditional distributions over h1 and h2 given v are given by the standard
DBN inference equations [1]. The conditional posterior over the gaze variables u is given by:
p(x|u, v)p(u)
p(u|x, v) =
,
p(x|v)
1 X (xi (u) ? vi )2
log p(u|x, v) ? log p(x|u, v) + log p(u) =
+ const.
(9)
2 i
?i2
Using Bayes? rule, the unnormalized log probability of p(u|x, v) is defined in Eq. 9. We stress that
this equation is atypical in that the random variable of interest u actually affects the conditioning
variable x (see Eq. 5) We can explore the gaze variables using Hamiltonian Monte Carlo (HMC)
algorithm [27, 28]. Intuitively, conditioned on the canonical object v that our model has in ?mind?,
HMC searches over the entire image I to find a region x with a good match to v.
If the goal is only to find the MAP estimate of p(u|x, v), then we may want to use second-order
methods for optimizing u. This would be equivalent to the Lucas-Kanade framework in computer
vision, developed for image alignment [29]. However, HMC has the advantage of being a proper
MCMC sampler that satisfies detailed balance and fits nicely with our probabilistic framework.
The HMC algorithm first specifies the Hamiltonian over the position variables u and auxiliary
momentum variables r: H(u, r) = U (u) + K(r), where the potential function is defined by
2
P
P
i)
U (u) = 21 i (xi (u)?v
and the kinetic energy function is given by K(r) = 12 i ri2 . The dy?i2
namics of the system is defined by:
?u
?r
?H
= r,
=?
(10)
?t
?t
?u
?H
(x(u) ? v) ?x(u)
=
,
(11)
?u
?2
?u
?x
?x
?w({p}, u) X
?xi
?w(pi , u)
=
=
.
(12)
?u
?w({p}, u)
?u
?w(p
,
u)
?u
i
i
Observe that Eq. 12 decomposes into sums over single coordinate positions pi = [x y]T . Let us
denote p0 i = w(pi , u) to be the coordinate pi warped by u. For the first term on the RHS of Eq. 12,
?xi
= ?I(p0 i ), (dimension 1 by 2 )
(13)
?w(pi , u)
where ?I(p0 i ) denotes the sampling of the gradient images of I at the warped location pi . For the
second term on the RHS of Eq. 12, we note that we can re-write Eq. 4 as:
" a #
h 0 i h
i
h
i
x
x ?y 1 0
b
x
=
+
,
(14)
0
y
y x 0 1
4x
y
4y
4
giving us
?w(pi , u) h x
=
y
?u
?y
x
1
0
i
0
.
1
(15)
HMC simulates the discretized system by performing leap-frog updates of u and r using Eq. 10.
Additional hyperparameters that need to be specified include the step size , number of leap-frog
steps, and the mass of the variables (see [28] for details).
4.1 Approximate Inference
HMC essentially performs gradient descent with momentum,
therefore it is prone to getting stuck at local optimums. This
is especially a problem for our task of finding the best transformation parameters. While the posterior over u should be
unimodal near the optimum, many local minima exist away
from the global optimum. For example, in Fig. 2(a), the big
image I is enclosed by the blue box, and the canonical image
v is enclosed by the green box. The current setting of u aligns
together the wrong eyes. However, it is hard to move the green
box to the left due to the local optima created by the dark intensities of the eye. Resampling the momentum variable every
iteration in HMC does not help significantly because we are
modeling real-valued images using a Gaussian distribution as
the residual, leading to quadratic costs in the difference between x(u) and v (see Eq. 9). This makes the energy barriers
between modes extremely high.
(a)
Average
A
B
To alleviate this problem we need to find good initializations
of u. We use a Convolutional Network (ConvNet) to perform efficient approximate inference, resulting in good initial
(b)
guesses. Specifically, given v, u and I, we predict the change Figure 2: (a) HMC can easily get
in u that will lead to the maximum log p(u|x, v). In other stuck at local optima. (b) Importance
words, instead of using the gradient field for updating u, we of modeling p(u|v, I). Best in color.
learn a ConvNet to output a better vector field in the space
of u. We used a fairly standard ConvNet architecture and the standard stochastic gradient descent
learning procedure.
We note that standard feedforward face detectors seek to model p(u|I), while completely ignoring
the canonical face v. In contrast, here we take v into account as well. The ConvNet is used to initialize u for the HMC algorithm. This is important in a proper generative model because conditioning
on v is appealing when multiple faces are present in the scene. Fig. 2(b) is a hypothesized Euclidean
space of v, where the black manifold represents canonical faces and the blue manifold represents
cropped faces x(u). The blue manifold has a low intrinsic dimensionality of 4, spanned by u. At A
and B, the blue comes close to black manifold. This means that there are at least two modes in the
posterior over u. By conditioning on v, we can narrow the posterior to a single mode, depending on
whom we want to focus our attention. We demonstrate this exact capability in Sec. 6.3.
Fig. 3 demonstrates the iterative process of how approximate inference works in our model. Specifically, based on u, the ConvNet takes a window patch around x(u) (72?72) and v (24?24) as input,
and predicts the output [4x, 4y, 4?, 4s]. In step 2, u is updated accordingly, followed by step 3
of alternating Gibbs updates of v and h, as discussed in Sec. 4. The process is repeated. For the
details of the ConvNet see the supplementary materials.
5
Learning
While inference in our framework localizes objects of interest and is akin to object detection, it is not
the main objective. Our motivation is not to compete with state-of-the-art object detectors but rather
propose a probabilistic generative framework capable of generative modeling of objects which are
at unknown locations in big images. This is because labels are expensive to obtain and are often not
available for images in an unconstrained environment.
To learn generatively without labels we propose a simple Monte Carlo based ExpectationMaximization algorithm. This algorithm is an unbiased estimator of the maximum likelihood objec5
1 Gibbs step
ConvNet
Step 1
ConvNet
Step 4
Step 3
Step 2
Figure 3: Inference process: u in step 1 is randomly initialized. The average v and the extracted x(u) form
the input to a ConvNet for approximate inference, giving a new u. The new u is used to sample p(v|I, u, h).
In step 3, one step of Gibbs sampling of the GDBN is performed. Step 4 repeats the approximate inference
using the updated v and x(u).
V
X
1
2
3
Inference steps
4
5
6
HMC
Figure 4: Example of an inference step. v is 24?24, x is 72?72. Approximate inference quickly finds a
good initialization for u, while HMC provides further adjustments. Intermediate inference steps on the right
are subsampled from 10 actual iterations.
tive. During the E-step, we use the Gibbs sampling algorithm developed in Sec. 4 to draw samples
from the posterior over the latent gaze variables u, the canonical variables v, and the hidden variables h1 , h2 of a Gaussian DBN model. During the M-step, we can update the weights of the
Gaussian DBN by using the posterior samples as its training data. In addition, we can update the
parameters of the ConvNet that performs approximate inference. Due to the fact that the first E-step
requires a good inference algorithm, we need to pretrain the ConvNet using labeled gaze data as
part of a bootstrap process. Obtaining training data for this initial phase is not a problem as we can
jitter/rotate/scale to create data. In Sec. 6.2, we demonstrate the ability to learn a good generative
model of face images from the CMU Multi-PIE dataset.
6
Experiments
We used two face datasets in our experiments. The first dataset is a frontal face dataset, called
the Caltech Faces from 1999, collected by Markus Weber. In this dataset, there are 450 faces of 27
unique individuals under different lighting conditions, expressions, and backgrounds. We downsampled the images from their native 896 by 692 by a factor of 2. The dataset also contains manually
labeled eyes and mouth coordinates, which will serve as the gaze labels. We also used the CMU
Multi-PIE dataset [30], which contains 337 subjects, captured under 15 viewpoints and 19 illumination conditions in four recording sessions for a total of more than 750,000 images. We demonstrate our model?s ability to perform approximate inference, to learn without labels, and to perform
identity-based attention given an image with two people.
6.1 Approximate inference
We first investigate the critical inference algorithm of p(u|v, I) on the Caltech Faces dataset. We
run 4 steps of approximate inference detailed in Sec. 4.1 and diagrammed in Fig. 3, followed by
three iterations of 20 leap-frog steps of HMC. Since we do not initially know the correct v, we
initialize v to be the average face across all subjects.
Fig. 4 shows the image of v and x during inference for a test subject. The initial gaze box is colored
yellow on the left. Subsequent gaze updates progress from yellow to blue. Once ConvNet-based
approximate inference gives a good initialization, starting from step 5, five iterations of 20 leap-frog
steps of HMC are used to sample from the the posterior.
Fig. 5 shows the quantitative results of Intersection over Union (IOU) of the ground truth face box
and the inferred face box. The results show that inference is very robust to initialization and requires
6
Accuracy of Approximate Inference
Accuracy of Approximate Inference
Accuracy Improvements
1.1
0.3
Average IOU Improvements
1
0.2
0.9
0.8
0.7
0.6
0.5
0
Trials with IOU > 0.5
Average IOU
20
40
60
0.8
Accuracy
Accuracy
Accuracy
1
0.6
0.4
Trials with IOU > 0.5
Average IOU
0.2
80
100
0
0
Initial Pixel Offset
5
10
0.1
0
?0.1
15
# of Inference Steps
?0.2
0
20
40
60
80
100
Initial Pixel Offset
(a)
(b)
(c)
Figure 5: (a) Accuracy as a function of gaze initialization (pixel offset). Blue curve is the percentage success
of at least 50% IOU. Red curve is the average IOU. (b) Accuracy as a function of the number of approximate
inference steps when initializing 50 pixels away. (c) Accuracy improvements of HMC as a function of gaze
initializations.
(a) DBN trained on Caltech
(b) DBN updated with Multi-PIE
Figure 6: Left: Samples from a 2-layer DBN trained on Caltech. Right: samples from an updated DBN after
training on CMU Multi-PIE without labels. Samples highlighted in green are similar to faces from CMU.
only a few steps of approximate inference to converge. HMC clearly improves model performance,
resulting in an IOU increase of about 5% for localization. This is impressive given that none of
the test subjects were part of the training and the background is different from backgrounds in the
training set.
We also compared our inference algorithm to
the template matching in the task of face deOur method OpenCV NCC template
tection. We took the first 5 subjects as test
IOU > 0.5
97%
97%
93%
78%
subjects and the rest as training. We can lo# evaluations
O(c)
O(whs) O(whs) O(whs)
Table 1: Face localization accuracy. w: image width; calize with 97% accuracy 3(IOU > 0.5) ush: image height; s: image scales; c: number of inference ing our inference algorithm . In comparison,
a near state-of-the-art face detection system
steps used.
from OpenCV 2.4.9 obtains the same 97% accuracy. It uses Haar Cascades, which is a form of AdaBoost4 . Normalized Cross Correlation [31]
obtained 93% accuracy, while Euclidean distance template matching achieved an accuracy of only
78%. However, note that our algorithm looks at a constant number of windows while the other
baselines are all based on scanning windows.
6.2
Generative learning without labels
The main advantage of our
model is that it can learn on
large images of faces without localization label information (no
manual cropping required). To
demonstrate, we use both the
Caltech and the CMU faces
Table 2: Variational lower-bound estimates on the log-density of the dataset. For the CMU faces, a
Gaussian DBNs (higher is better).
subset of 2526 frontal faces with
ground truth labels are used. We split the Caltech dataset into a training and a validation set. For
the CMU faces, we first took 10% of the images as training cases for the ConvNet for approximate
inference. This is needed due to the completely different backgrounds of the Caltech and CMU
datasets. The remaining 90% of the CMU faces are split into a training and validation set. We first
trained a GDBN with 1024 h1 and 256 h2 hidden units on the Caltech training set. We also trained
nats
No CMU training CMU w/o labels CMU w/ labels
Caltech Train
617?0.4
627?0.5
569?0.6
Caltech Valid
512?1.1
503?1.8
494?1.7
CMU Train
96?0.8
499?0.1
594?0.5
CMU Valid
85?0.5
387?0.3
503?0.7
log Z?
454.6
687.8
694.2
3
u is randomly initialized at ? 30 pixels, scale range from 0.5 to 1.5.
OpenCV detection uses pretrained model from haarcascade_frontalface_default.xml, scaleFactor=1.1,
minNeighbors=3 and minSize=30.
4
7
Figure 7: Left: Conditioned on different v will result in a different 4u. Note that the initial u is exactly the
same for two trials. Right: Additional examples. The only difference between the top and bottom panels is the
conditioned v. Best viewed in color.
a ConvNet for approximate inference using the Caltech training set and 10% of the CMU training
images.
Table 2 shows the estimates of the variational lower-bounds on the average log-density (higher is
better) that the GDBN models assign to the ground-truth cropped face images from the training/test
sets under different scenarios. In the left column, the model is only trained on Caltech faces. Thus it
gives very low probabilities to the CMU faces. Indeed, GDBNs achieve a variational lower-bound of
only 85 nats per test image. In the middle column, we use our approximate inference to estimate the
location of the CMU training faces and further trained the GDBN on the newly localized faces. This
gives a dramatic increase of the model performance on the CMU Validation set5 , achieving a lowerbound of 387 nats per test image. The right column gives the best possible results if we can train
with the CMU manual localization labels. In this case, GDBNs achieve a lower-bound of 503 nats.
We used Annealed Importance Sampling (AIS) to estimate the partition function for the top-layer
RBM. Details on estimating the variational lower bound are in the supplementary materials.
Fig. 6(a) further shows samples drawn from the Caltech trained DBN, whereas Fig. 6(b) shows
samples after training with the CMU dataset using estimated u. Observe that samples in Fig. 6(b)
show a more diverse set of faces. We trained GDBNs using a greedy, layer-wise algorithm of [1].
For the top layer we use Fast Persistent Contrastive Divergence [32], which substantially improved
generative performance of GDBNs (see supplementary material for more details).
6.3 Inference with ambiguity
Our attentional mechanism can also be useful when multiple objects/faces are present in the scene.
Indeed, the posterior p(u|x, v) is conditioned on v, which means that where to attend is a function of the canonical object v the model has in ?mind? (see Fig. 2(b)). To explore this, we first
synthetically generate a dataset by concatenating together two faces from the Caltech dataset. We
then train approximate inference ConvNet as in Sec. 4.1 and test on the held-out subjects. Indeed,
as predicted, Fig. 7 shows that depending on which canonical image is conditioned, the same exact
gaze initialization leads to two very different gaze shifts. Note that this phenomenon is observed
across different scales and location of the initial gaze. For example, in Fig. 7, right-bottom panel,
the initialized yellow box is mostly on the female?s face to the left, but because the conditioned
canonical face v is that of the right male, attention is shifted to the right.
7
Conclusion
In this paper we have proposed a probabilistic graphical model framework for learning generative
models using attention. Experiments on face modeling have shown that ConvNet based approximate
inference combined with HMC sampling is sufficient to explore the complicated posterior distribution. More importantly, we can generatively learn objects of interest from novel big images. Future
work will include experimenting with faces as well as other objects in a large scene. Currently the
ConvNet approximate inference is trained in a supervised manner, but reinforcement learning could
also be used instead.
Acknowledgements
The authors gratefully acknowledge the support and generosity from Samsung, Google, and ONR
grant N00014-14-1-0232.
5
We note that we still made use of labels coming from the 10% of CMU Multi-PIE training set in order to
pretrain our ConvNet. "w/o labels" here means that no labels for the CMU Train/Valid images are given.
8
References
[1] G. E. Hinton, S. Osindero, and Y. W. Teh. A fast learning algorithm for deep belief nets. Neural Computation, 18(7):1527?1554, 2006.
[2] R. Salakhutdinov and G. Hinton. Deep Boltzmann machines. In AISTATS, 2009.
[3] Geoffrey E. Hinton, Peter Dayan, and Michael Revow. Modeling the manifolds of images of handwritten
digits. IEEE Transactions on Neural Networks, 8(1):65?74, 1997.
[4] Daniel Zoran and Yair Weiss. From learning models of natural image patches to whole image restoration.
In ICCV. IEEE, 2011.
[5] Yoshua Bengio, Li Yao, Guillaume Alain, and Pascal Vincent. Generalized denoising auto-encoders as
generative models. In Advances in Neural Information Processing Systems 26, 2013.
[6] H. Lee, R. Grosse, R. Ranganath, and A. Y. Ng. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations. In ICML, pages 609?616, 2009.
[7] Marc?Aurelio Ranzato, Joshua Susskind, Volodymyr Mnih, and Geoffrey Hinton. On Deep Generative
Models with Applications to Recognition. In CVPR, 2011.
[8] Yichuan Tang, Ruslan Salakhutdinov, and Geoffrey E. Hinton. Deep mixtures of factor analysers. In
ICML. icml.cc / Omnipress, 2012.
[9] M. I. Posner and C. D. Gilbert. Attention and primary visual cortex. Proc. of the National Academy of
Sciences, 96(6), March 1999.
[10] E. A. Buffalo, P. Fries, R. Landman, H. Liang, and R. Desimone. A backward progression of attentional
effects in the ventral stream. PNAS, 107(1):361?365, Jan. 2010.
[11] N Kanwisher and E Wojciulik. Visual attention: Insights from brain imaging. Nature Reviews Neuroscience, 1:91?100, 2000.
[12] C. H. Anderson and D. C. Van Essen. Shifter circuits: A computational strategy for dynamic aspects of
visual processing. National Academy of Sciences, 84:6297?6301, 1987.
[13] B. A. Olshausen, C. H. Anderson, and D. C. Van Essen. A neurobiological model of visual attention and
invariant pattern recognition based on dynamic routing of information. The Journal of neuroscience : the
official journal of the Society for Neuroscience, 13(11):4700?4719, 1993.
[14] Laurenz Wiskott. How does our visual system achieve shift and size invariance?, 2004.
[15] S. Chikkerur, T. Serre, C. Tan, and T. Poggio. What and where: a Bayesian inference theory of attention.
Vision Research, 50(22):2233?2247, October 2010.
[16] J. K. Tsotsos, S. M. Culhane, W. Y. K. Wai, Y. H. Lai, N. Davis, and F. Nuflo. Modeling visual-attention
via selective tuning. Artificial Intelligence, 78(1-2):507?545, October 1995.
[17] Hugo Larochelle and Geoffrey E. Hinton. Learning to combine foveal glimpses with a third-order boltzmann machine. In NIPS, pages 1243?1251. Curran Associates, Inc., 2010.
[18] D. P. Reichert, P. Seri?s, and A. J. Storkey. A hierarchical generative model of recurrent object-based
attention in the visual cortex. In ICANN (1), volume 6791, pages 18?25. Springer, 2011.
[19] B. Alexe, N. Heess, Y. W. Teh, and V. Ferrari. Searching for objects driven by context. In NIPS 2012,
December 2012.
[20] Marc?Aurelio Ranzato. On learning where to look. arXiv, arXiv:1405.5488, 2014.
[21] M. Denil, L. Bazzani, H. Larochelle, and N. de Freitas. Learning where to attend with deep architectures
for image tracking. Neural Computation, 28:2151?2184, 2012.
[22] G. E. Hinton and R. Salakhutdinov. Reducing the dimensionality of data with neural networks. Science,
313:504?507, 2006.
[23] A. Krizhevsky. Learning multiple layers of features from tiny images, 2009.
[24] Graham W. Taylor, Rob Fergus, Yann LeCun, and Christoph Bregler. Convolutional learning of spatiotemporal features. In ECCV 2010. Springer, 2010.
[25] A. Mohamed, G. Dahl, and G. Hinton. Acoustic modeling using deep belief networks. IEEE Transactions
on Audio, Speech, and Language Processing, 2011.
[26] Richard Szeliski. Computer Vision - Algorithms and Applications. Texts in Computer Science. Springer,
2011.
[27] S. Duane, A. D. Kennedy, B. J Pendleton, and D. Roweth. Hybrid Monte Carlo. Physics Letters B,
195(2):216?222, 1987.
[28] R. M. Neal. MCMC using Hamiltonian dynamics. in Handbook of Markov Chain Monte Carlo (eds S.
Brooks, A. Gelman, G. Jones, XL Meng). Chapman and Hall/CRC Press, 2010.
[29] Simon Baker and Iain Matthews. Lucas-kanade 20 years on: A unifying framework. International Journal
of Computer Vision, 56:221?255, 2002.
[30] Ralph Gross, Iain Matthews, Jeffrey F. Cohn, Takeo Kanade, and Simon Baker. Multi-pie. Image Vision
Comput., 28(5):807?813, 2010.
[31] J. P. Lewis. Fast normalized cross-correlation, 1995.
[32] T. Tieleman and G. E. Hinton. Using fast weights to improve persistent contrastive divergence. In ICML,
volume 382, page 130. ACM, 2009.
[33] R. Salakhutdinov and I. Murray. On the quantitative analysis of deep belief networks. In Proceedings of
the Intl. Conf. on Machine Learning, volume 25, 2008.
9
| 5345 |@word trial:3 middle:1 briefly:1 nd:1 seek:3 propagate:1 covariance:1 p0:7 simplifying:1 set5:1 contrastive:2 dramatic:1 extrastriate:1 initial:7 configuration:1 foveal:2 generatively:3 contains:2 daniel:1 freitas:1 current:1 activation:1 reminiscent:1 takeo:1 realistic:1 visible:5 informative:1 subsequent:1 partition:1 treating:2 update:5 resampling:1 generative:33 greedy:2 guess:1 intelligence:1 accordingly:1 hamiltonian:4 core:1 colored:1 provides:1 node:1 toronto:3 location:8 five:1 height:1 persistent:2 combine:2 redefine:1 manner:1 kanwisher:1 indeed:3 multi:6 brain:2 discretized:1 salakhutdinov:6 inspired:2 actual:1 window:3 cardinality:1 laurenz:1 discover:1 underlying:1 notation:1 panel:5 estimating:1 circuit:2 mass:1 what:1 baker:2 substantially:1 developed:2 finding:1 transformation:13 quantitative:2 every:1 exactly:1 wrong:1 demonstrates:1 unit:3 grant:1 omit:1 before:1 attend:3 local:4 limit:1 bilinear:1 meng:1 interpolation:1 black:2 initialization:8 frog:4 dynamically:2 r4:1 christoph:1 co:1 bi:3 range:2 lowerbound:1 directed:1 unique:1 lecun:1 union:1 block:1 implement:1 bootstrap:1 digit:1 procedure:2 susskind:1 jan:1 area:1 ri2:1 significantly:1 cascade:1 matching:2 word:1 downsampled:1 get:1 onto:1 unlabeled:2 close:1 operator:1 gelman:1 context:1 influence:2 gilbert:1 equivalent:1 map:1 missing:1 annealed:1 straightforward:1 attention:24 starting:1 resolution:4 rule:1 estimator:1 insight:1 importantly:2 spanned:1 iain:2 posner:1 grbm:2 searching:1 ferrari:1 coordinate:7 updated:4 dbns:2 tan:1 massive:1 exact:2 us:4 curran:1 associate:1 storkey:1 recognition:4 expensive:1 updating:1 curated:1 predicts:1 labeled:2 native:1 observed:2 bottom:3 initializing:1 capture:1 region:6 connected:1 ranzato:2 gross:1 environment:1 nats:4 occluded:1 dynamic:4 diagrammed:1 zoran:2 trained:9 solving:1 serve:2 upon:1 bipartite:1 localization:4 completely:2 easily:3 joint:1 samsung:1 various:1 derivation:1 train:5 fast:4 describe:1 culhane:1 monte:5 seri:1 artificial:1 analyser:1 pendleton:1 whose:1 supplementary:3 valued:2 cvpr:1 ability:2 highlighted:1 associative:1 advantage:3 net:3 took:2 propose:4 coming:1 aligned:2 relevant:1 combining:1 translate:1 ontario:1 deformable:1 achieve:3 academy:2 intuitive:1 getting:1 bazzani:1 optimum:5 plethora:1 cropping:1 intl:1 generating:1 object:23 help:1 depending:2 recurrent:1 ij:1 expectationmaximization:1 progress:1 eq:11 auxiliary:1 c:1 predicted:1 come:1 larochelle:3 iou:11 concentrate:1 correct:1 stochastic:2 centered:4 routing:4 enable:1 material:3 crc:1 require:1 assign:1 alleviate:1 biological:1 bregler:1 around:1 hall:1 ground:3 exp:3 alexe:1 predict:1 opencv:3 matthew:2 ventral:1 ruslan:2 proc:1 leap:4 label:14 currently:1 create:1 weighted:1 clearly:1 gaussian:14 rather:2 denil:1 hj:3 focus:3 improvement:3 likelihood:1 experimenting:1 pretrain:2 contrast:1 generosity:1 baseline:1 inference:45 dayan:1 typically:1 entire:1 initially:1 hidden:9 selective:2 wij:4 pixel:8 ralph:1 classification:1 flexible:1 pascal:1 priori:1 lucas:2 art:2 fairly:1 initialize:2 marginal:1 field:3 equal:1 once:1 having:2 nicely:1 sampling:6 manually:1 ng:1 represents:2 chapman:1 look:4 unsupervised:1 icml:4 jones:1 mimic:1 future:1 yoshua:1 stimulus:1 simplify:2 richard:1 few:1 randomly:2 divergence:2 national:2 individual:1 subsampled:1 geometry:1 occlusion:1 phase:1 jeffrey:1 attempt:1 detection:3 interest:9 essen:3 investigate:1 mnih:1 evaluation:1 alignment:1 male:1 mixture:5 held:1 chain:1 desimone:1 capable:2 glimpse:2 poggio:1 euclidean:2 taylor:1 initialized:3 re:1 roweth:1 column:3 modeling:10 grbms:1 restoration:1 applicability:1 cost:1 subset:1 krizhevsky:1 successful:1 osindero:1 encoders:2 scanning:1 spatiotemporal:1 chooses:1 combined:1 density:2 international:1 probabilistic:3 lee:1 physic:1 gaze:15 together:3 quickly:1 michael:1 yao:1 ambiguity:1 containing:1 conf:1 warped:3 resort:1 leading:1 toy:1 li:1 account:1 potential:1 volodymyr:1 de:1 sec:6 inc:1 vi:7 stream:1 performed:2 h1:8 multiplicative:1 red:1 bayes:1 complicated:3 capability:1 simon:2 accuracy:14 convolutional:4 largely:1 efficiently:1 gathered:1 correspond:1 landscape:1 yellow:3 handwritten:1 vincent:1 bayesian:1 none:1 carlo:5 lighting:2 antialiasing:1 cc:1 kennedy:1 ncc:1 detector:2 manual:2 aligns:1 wai:1 ed:1 energy:5 rbms:1 mohamed:1 rbm:6 static:1 newly:1 dataset:14 knowledge:1 color:2 dimensionality:3 improves:1 cj:1 actually:1 originally:1 higher:3 supervised:1 wei:2 improved:1 box:7 anderson:3 correlation:2 canvas:1 cohn:1 google:1 mode:3 olshausen:5 building:2 effect:1 hypothesized:2 normalized:2 unbiased:1 serre:1 alternating:2 i2:7 illustrated:1 neal:1 sin:1 during:5 width:1 davis:1 unnormalized:1 generalized:1 stress:1 demonstrate:4 covert:1 performs:2 omnipress:1 network1:1 image:55 weber:1 variational:4 novel:3 recently:2 wise:1 common:1 rotation:3 hugo:1 conditioning:3 volume:3 discussed:1 gibbs:5 ai:2 tuning:2 rd:2 dbn:13 unconstrained:1 session:1 analyzer:1 language:2 gratefully:1 cortex:4 similarity:9 impressive:1 add:1 posterior:12 recent:1 female:1 optimizing:1 driven:1 discard:1 scenario:1 route:2 certain:1 n00014:1 binary:2 success:2 onr:1 joshua:1 caltech:14 captured:1 minimum:1 additional:2 employed:1 converge:1 signal:1 multiple:3 unimodal:1 pnas:1 infer:1 uncluttered:1 ing:1 match:1 cross:2 long:1 lai:1 scalable:1 vision:5 essentially:1 cmu:22 arxiv:2 iteration:4 represent:2 achieved:1 background:8 addition:2 cropped:4 want:3 whereas:2 diagram:2 rest:2 unlike:3 subject:8 recording:1 virtually:1 simulates:1 december:1 extracting:1 near:2 presence:1 synthetically:1 feedforward:1 intermediate:1 split:2 bengio:1 affect:1 fit:1 architecture:2 shift:5 expression:1 akin:1 stereo:1 peter:1 speech:3 action:1 deep:16 ignored:1 useful:2 heess:1 clear:1 detailed:2 factorial:2 amount:2 clutter:2 dark:1 neocortex:1 generate:1 specifies:1 exist:2 percentage:1 canonical:18 shifted:2 neuroscience:6 estimated:1 per:2 blue:6 diverse:1 write:1 four:1 enormous:1 achieving:2 drawn:1 clarity:1 dahl:1 backward:1 v1:2 imaging:1 tsotsos:2 sum:1 year:1 compete:1 run:1 letter:1 powerful:1 jitter:1 family:1 yann:1 patch:7 draw:1 dy:1 scaling:1 graham:1 layer:13 bound:5 followed:2 display:1 quadratic:1 scene:7 flat:1 unlimited:1 markus:1 aspect:2 nitish:2 extremely:1 nuflo:1 performing:2 department:1 structured:1 according:1 combination:1 march:1 across:2 ush:1 appealing:1 rob:1 rsalakhu:1 psychologist:1 intuitively:2 restricted:3 invariant:2 iccv:1 resource:1 equation:3 mechanism:4 needed:1 know:2 mind:2 available:1 gaussians:2 operation:1 observe:3 hierarchical:3 away:2 progression:1 fry:1 robustly:1 yair:1 reichert:2 top:8 denotes:2 include:3 remaining:1 graphical:2 unifying:1 const:1 giving:2 especially:1 murray:1 society:1 move:1 objective:1 strategy:1 primary:1 dependence:1 diagonal:1 gradient:4 convnet:18 attentional:5 separate:1 distance:1 manifold:5 whom:1 collected:1 shifter:2 chikkerur:1 modeled:4 providing:1 balance:1 equivalently:1 difficult:1 hmc:16 pie:6 mostly:1 liang:1 october:2 design:1 proper:2 boltzmann:10 unknown:2 perform:4 teh:2 markov:2 datasets:2 acknowledge:1 descent:2 buffalo:1 defining:1 hinton:11 incorporated:1 extended:1 frame:1 canada:1 intensity:1 inferred:2 tive:1 required:2 specified:1 connection:1 acoustic:1 learned:2 narrow:1 nip:2 brook:1 able:2 perception:1 pattern:1 including:2 memory:1 video:2 belief:8 green:3 mouth:1 critical:1 natural:3 hybrid:1 haar:1 residual:1 localizes:1 improve:1 xml:1 eye:3 created:1 auto:2 text:1 prior:2 understanding:1 literature:1 review:2 acknowledgement:1 enclosed:2 geoffrey:4 localized:1 validation:3 h2:6 sufficient:1 propagates:1 wiskott:1 viewpoint:1 tiny:1 share:1 pi:7 lo:1 prone:1 eccv:1 repeat:1 last:1 alain:1 allow:1 warp:1 szeliski:1 template:3 face:46 barrier:1 van:3 curve:2 dimension:1 transition:1 valid:3 rich:1 sensory:2 stuck:2 author:1 reinforcement:1 made:1 transaction:2 ranganath:1 approximate:21 yichuan:2 ignore:1 obtains:1 neurobiological:1 dealing:1 global:1 handbook:1 discriminative:1 xi:5 fergus:1 search:2 latent:7 iterative:1 decomposes:1 table:3 kanade:3 nature:1 learn:10 robust:1 ignoring:2 obtaining:1 complex:2 marc:2 official:1 aistats:1 icann:1 main:3 rh:2 big:6 motivation:2 noise:1 hyperparameters:1 whole:1 aurelio:2 repeated:1 fig:14 grosse:1 position:3 momentum:3 wish:1 concatenating:1 xl:1 lie:1 comput:1 atypical:1 third:3 tang:4 down:5 tection:1 gating:1 h1j:1 offset:3 evidence:1 intractable:1 intrinsic:1 ih:1 importance:2 illumination:1 conditioned:7 intersection:1 simply:1 explore:3 visual:13 adjustment:1 tracking:2 partially:1 pretrained:1 springer:3 duane:1 corresponds:1 truth:3 satisfies:1 lewis:1 extracted:3 kinetic:1 tieleman:1 acm:1 conditional:8 goal:1 presentation:1 identity:1 viewed:1 revow:1 content:1 change:2 hard:1 specifically:3 reducing:1 sampler:1 denoising:2 called:2 total:1 invariance:2 meaningful:1 selectively:1 guillaume:1 people:1 support:1 rotate:2 frontal:3 mcmc:2 audio:1 phenomenon:1 srivastava:1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.